Created on 10-25-2013 09:24 AM - edited 09-16-2022 01:49 AM
Hi, I'm new to Cloudera Standard. We [will] have a 7 node cluster (5 TaskTrackers + 2 NameNodes/JobTrackers). In addition I have one VM to host the Cloudera Manager. I've ran through the installation process a few times to get familiar, and my question is how best to distribute the role assignments.
We need HDFS, HBase, MapReduce, Hive, Hue, Sqoop, and Impala. Can someone tell me if I'm on the right track with the assignments here?
My biggest questions lie with the Hive and HBase assignments across the nodes:
Any recommendations would be greatly appreciated! Thanks!
Created 10-27-2013 03:58 PM
Is this for proof of concept/discovery? This would be a tightly packed cluster you are proposing for most production environments IMHO . Our account team provides Solutions Engineering /Solutions Architecture guidance for things like this as you begin to scope out revenue generating activity with a cluster and want enterprise support.
Review our blogs discussion here on hardware sizing:
http://blog.cloudera.com/blog/2013/08/how-to-select-the-right-hardware-for-your-new-hadoop-cluster/
As a general rule of thumb, try to minimize the services co-located with the NameNodes (NN). JN's and ZK nodes are OK to co-locate with them, we recommend JN's be co-located with the NN's + 1.
You have a over-loaded mix of services here in the combination of HBase and Mapreduce services on the nodes (there will be battles for memory if you are under-sized.)
If this is a dev or non prod footprint you can get by with what I'm proposing below... HBase can take a lot of memory so you want to monitor. MR jobs are variable based on the complexity and size of what you are doing.
Secondary Name Node (SNN) as an architecture is less safe then using Name Node High Availability (NN HA). SNN does not do what you think it does. Read the hadoop operations guide book (3rd edition) to have a better sense of this.
Once you enable NN HA you end up deploying 3 zookeeper instances and jounal nodes, so based on what you are presenting you are saddling up for future outage / loss of data if this ends up in prod this way unless you are really really careful (and even then you could get hit).
This footprint viability really depends on your workload... so you might end up relocating things once you start observing activity, what I'm proposing below is, at best, a playpen configuration so you can kick the tires and check stuff out.
You are using 3 separate DB implementations with Impala (Fast SQL perf), Hive (Slower SQL but more std SQL support) and HBase (Column based DB)... Is your design really requiring all 3 (research them a little bit more, add them if it makes sense after initial deployment). HUE is usually in the mix to facilitate end user web based access too.
Realize you can move services after deployment. Note that decommissioning a DataNode takes a while to replicate data blocks off the node to the rest.
Read up on Hive and Hive2. Hive2 is for more secure implementations (plus other stuff)
Created 11-05-2013 10:32 AM
We're sorry this is happening, but realize we have a large install base that uses CM for managing and deploying clusters without issue.
Generally the problems people run into with installation are related to
The "host inspector" function examines hosts and offers the most common flaws, but that generally does not come up until you have finished cluster role assignments.
/var/log/cloudera-scm-server provides logging of the CM server issues
/var/log/clouder-scm-agent provides logging seen through the agent configuration
/var/run/cloudera-scm-agent/process/[###]-SERVICE-Instance/ provides current runtime information for deployed parcel services started by CM.
Todd
Created 05-26-2015 01:41 AM
I am going through a new install, and got stuck on the same point, ZK not starting.
I resolved it by going to ZK dashboard, then in the Actions menu, I initialized ZK, after which everything went well.
Thereafter, I went to each role/module installed and performed steps in the Action menu. Seems like Cloudera did not process through these steps on initial install, could be due to some failed step.
Thanks