Support Questions
Find answers, ask questions, and share your expertise
Announcements
Alert: Welcome to the Unified Cloudera Community. Former HCC members be sure to read and learn how to activate your account here.

"Hadoop fs -ls" Produces the Local Filesystem's "ls -la"

Solved Go to solution

"Hadoop fs -ls" Produces the Local Filesystem's "ls -la"

Yesterday I configured a 5 node cluster via Cloudera Manager. I then added another node, and was about to make a "Gateway Template" to assign to it through the Add Node workflow, when my Firefox crashed. I loaded it back up and manually assigned all the gateway roles to it (HDFS, MapReduce, HBase, Hive).

 

However, when I issue "hadoop fs -ls" on the node, it basically prints out the results of "ls -la" on my local filesystem. As in, whichever folder I am in on the local filesystem, "hadoop fs -ls" will show me all the files, their permissions, and hidden files.

 

What is the fix for this?

 

Would it hurt to make another node, and instead of assigning it Gateway roles through Cloudera Manager or downloading any Cloudera Agent, to simply download the client packages for HDFS, MapReduce, Hive, HBase, Pig, Oozie, etc. and manually connect it as a "gateway" node?

 

Thank you.

1 ACCEPTED SOLUTION

Accepted Solutions

Re: "Hadoop fs -ls" Produces the Local Filesystem's "ls -la"

Hi Matthew,

 

After using the host template to assign gateway roles to your host, did you deploy client configuration? This command will update all the files in /etc/hadoop/conf, /etc/hive/conf, etc.

 

Thanks,

Darren

3 REPLIES 3

Re: "Hadoop fs -ls" Produces the Local Filesystem's "ls -la"

Hi Matthew,

 

After using the host template to assign gateway roles to your host, did you deploy client configuration? This command will update all the files in /etc/hadoop/conf, /etc/hive/conf, etc.

 

Thanks,

Darren

Highlighted

Re: "Hadoop fs -ls" Produces the Local Filesystem's "ls -la"

Expert Contributor
It means your node is not making a call to Hadoop file system It is running a local filesystem command "ls -a"...
Check your core-site.xml in hadoop conf directory for fs.default.name property and value should be your namenode hostname...

Other way to do it through CM is, make sure the hdfs gateway role is installed on the host and deploy the client configurations..

If you don't see the host in HDFS gateway, then you need to add gateway to the host and deploy client configurations...
Em Jay

Re: "Hadoop fs -ls" Produces the Local Filesystem's "ls -la"

New Contributor
Please check softlink on
/etc/alternatives/

make sure they are pointing to correct hadoop config in /etc/hadoop/conf/core-site.xml etc.

deploying client config from CM also helps to resolve all the softlinks
Don't have an account?
Coming from Hortonworks? Activate your account here