Member since
02-11-2015
3
Posts
0
Kudos Received
0
Solutions
01-26-2017
07:39 PM
1 Kudo
I was having same issue and was getting same error but when I was running any command through directory where I have installed CDH, I was able to run all the commands - hadoop, hdfs, spark-shell ect. e.g. if your CHD installation location is - /dat/anlt1/cld/cloudera/CDH-5.8.3-1.cdh5.8.3.p0.2/bin you can test - $ cd /dat/anlt1/cld/cloudera/CDH-5.8.3-1.cdh5.8.3.p0.2/bin [root@xyz bin]# ./hadoop and if it work then you need to set up environment variable in your Unix master server For RHEL - [root@xyz~]# echo "$PATH" /usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin:/root/bin [root@xyz~]# export PATH=$PATH:/path/to/CHD_Installation_bin_path for me it's - /dat/anlt1/cld/cloudera/CDH-5.8.3-1.cdh5.8.3.p0.2/bin [root@xyz~]# echo "$PATH" /usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin:/root/bin:/dat/anlt1/cld/cloudera/CDH-5.8.3-1.cdh5.8.3.p0.2/bin to make permanent change - $ echo "export PATH=$PATH:/dat/anlt1/cld/cloudera/CDH-5.8.3-1.cdh5.8.3.p0.2/bin" >> /etc/profile after that restart(reboot) your server.
... View more
06-19-2015
04:46 PM
If you use CM you don't need to use the Hue safety valve. Make sure the Trift Server is added and started Make sure in the Hue configuration that the HBase service and the HBase thrift service are checked Restart Hue and that's it! Romain
... View more