Member since
10-01-2015
3933
Posts
1150
Kudos Received
374
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3360 | 05-03-2017 05:13 PM | |
2789 | 05-02-2017 08:38 AM | |
3064 | 05-02-2017 08:13 AM | |
3002 | 04-10-2017 10:51 PM | |
1506 | 03-28-2017 02:27 AM |
03-07-2017
06:39 PM
@Saikiran Parepally please open this as a new HCC question not to hijack the current thread.
... View more
03-07-2017
05:31 PM
2 Kudos
Go to any new host on the hosts page, click the plus sign to add components and select hiveserver2
... View more
03-07-2017
04:43 PM
@rahul gulati not sure, I'm not privy to the Cloudbreak roadmap. Perhaps you'd want to open this as a new HCC question and someone from the Cloudbreak team can respond?
... View more
03-07-2017
04:31 PM
@rahul gulati no you can't add nodes to existing cluster deployed by Ambari with Cloudbreak. To add nodes to Cloudbreak provisioned cluster you can use the UI or Cloudbreak CLI for automation purposes.
... View more
03-07-2017
04:16 PM
for dev purposes I guess it's fine. True with Cloudbreak but you can grow your cluster and remove nodes from your cluster easily, additionally specifying the type of machine for next time you add a node. Either way is fine but I'm glad you already entertained idea of Cloudbreak.
... View more
03-07-2017
04:05 PM
2 Kudos
please double check the Grafana password @Sedat Kestepe looks like you're not providing username and password in your API call. You can find it in the configs section of Ambari Metrics Service Fail: Ambari Metrics Grafana data source creation failed. POST request status: 401 Unauthorized {"message":"Invalid username or password"}
... View more
03-07-2017
04:02 PM
@rahul gulati
you only need to copy OS data, once you add new nodes to your cluster, issue HDFS rebalance command. Once done, go to a node you intend to remove, hit decommision, HDFS will start moving data from that node across the cluster. Continue with next node until all old nodes are removed. I do highly recommend you add more new nodes for DataNode purposes. 3 DN is just not enough, basically you have no resiliency. The more nodes you'll have, the faster this whole project will go for you. Also look at http://sequenceiq.com/cloudbreak-docs/latest/ this will make mgmt of your cluster on cloud a lot easier, including this topic.
... View more
03-07-2017
03:36 PM
@rahul gulati no data will not be copied automatically, all new data will start balancing around the hdfs but not the existing data. You will have to call the balancer yourself via Ambari, https://docs.hortonworks.com/HDPDocuments/Ambari-2.2.1.1/bk_Ambari_Users_Guide/content/_how_to_rebalance_hdfs.html the data residing on OS has to be copied manually by your team.
... View more
03-07-2017
12:45 PM
you should add new nodes to the cluster and rebalance cluster, then decommission each node one at a time. Since you're replacing three nodes and that's all of your hdfs, it make take some time. Consider larger footprint as you know hdfs maintains replication factor of 3. You're best suited to use a tool like Cloudbreak or if you're only running ETL, discovery, data science workloads, you can try Hortonworks Data Cloud. Both can add and remove instances as well as provision new instances with new machine type easily.
... View more
03-07-2017
12:29 PM
2 Kudos
Deleting rows in HBase is a heavy operation, instead of managing deletions yourself, let HBase handle it via TTL. Basically you can set expiration on a row or alternatively cell and it will be marked as deleted once time to live expires, time is in UTC. https://hbase.apache.org/book.html#ttl Once row has a delete market it will be cleaned up by a standard compaction mechanism.
... View more