Member since
03-29-2018
2
Posts
0
Kudos Received
0
Solutions
04-01-2018
06:51 PM
@bgooley, Thank you very much for your reply and advice. Now we know that CM is necessary to start the CDH because it is managed by CM. Unfortunately, our namenodes (2) and datanodes (4) hosts have been rebooted, so I cannot get back the processes managed by agents from “/var/run/cloudera-scm-agent/process”. Now I am trying to install a new CM, using a new set of hosts to simulate the situation about the existing CDH cluster, my plan (I read a post like this: https://community.cloudera.com/t5/Cloudera-Manager-Installation/cloudera-manager-database-lost/td-p/31989) is that after I install the new CM: 1. I extract the hosts and roles configuration as json file; 2. Update the json file host name as the existing CDH hosts; 3. Import the json file into CM; 4. Update all /etc/cloudera-scm-agent/config.ini server_host to the new CM; One challenge for me is that we cannot remember all the roles in each host clearly, because we installed all services (include HDFS, Hbase, Hive, Impala, Zookeeper, Spark, Hue…etc.) before, but for HDFS and Hbase should be ok since our host name is like namenode01, namenode02, datanode01, datanode02, datanode03, datanode04… Would like to seek the help that: 1. In the new install CM, if some roles config are not match the exiting CDH, any impact? Especially to the data; 2. Can I install the new CM only and point the host into the existing CDH during the install process, I am afraid the new CM will re-install the CDH and have some impact the existing ones; BTW, it would be very appreciated if you can share the way to start HDFS without CM as your mentioned before. Thanks again!
... View more
03-29-2018
09:01 PM
We are facing a critical issue that we lost our Cloudera Manager configuration DB (It is a separate Oracle DB) and Kerberos server, so we cannot start Cloudera Manager server now but I think HDSF service should be still there, we can update the config to disable use Kerberos and start HDSF service to find back our data.
When I try to start the HDSF cluster through command line manually, found that we don’t have service “hadoop-hdfs-namenode”, “hadoop-hdfs-datanode” and other “hadoop-hdfs-*” in init.d directory, we only have “cloudera-scm-agent” and “cloudera-scm-server” in init.d:
[root@inthdpname01 init.d]# ll
total 116
-rwxr-xr-x 1 root root 8594 Jul 7 2017 cloudera-scm-agent
-rwxr-xr-x 1 root root 8436 Jul 7 2017 cloudera-scm-server
-rw-r--r--. 1 root root 13948 Sep 16 2015 functions
-rwxr-xr-x 1 root root 9972 Jan 21 2012 jexec
-rwxr-xr-x. 1 root root 2989 Sep 16 2015 netconsole
-rwxr-xr-x. 1 root root 6630 Sep 16 2015 network
-rw-r--r-- 1 root root 59 May 13 2016 output
-rw-r--r--. 1 root root 1160 Nov 20 2015 README
-rwxr-xr-x. 1 root root 41724 May 4 2016 vmware-tools
Would like to seek help that:
Any way for me to start HDFS manually?
Can we re-install a new Cloudera Manager and update its config point back the existing cluster (Say, update the host, and update each agent in the old host to the new CM)?
... View more
Labels:
- Labels:
-
Cloudera Manager
-
HDFS
-
Kerberos