I have an issue at installation time of datanode (Ambari 2.1.2, HDP 2.2.4) , it fails because at step "Datanode start" with the error:
File "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py", line 87, in action_create raise Fail("Applying %s failed, parent directory %s doesn't exist" % (self.resource, dirname)) resource_management.core.exceptions.Fail: Applying File['/usr/hdp/current/hadoop-client/conf/hadoop-policy.xml'] failed, parent directory /usr/hdp/current/hadoop-client/conf doesn't exist
Obviously the cause of this issue already happens at step "Datanode install", because the last log message is:
... 2016-01-12 19:44:17,233 - Skipping XmlConfig['core-site.xml'] due to only_if 2016-01-12 19:44:17,233 - Can only link configs for HDP-2.3 and higher.
Hmmm, strange, HDP2.2 should be supported by recent versions of Ambari...
Is this a known issue? any workarounds, hints what to do ?
just to give you an update on this:
- yes, there was/is enough space in /usr and the dir /usr/hdp/184.108.40.206-2 has been created
- due to time pressure I decided to start again from scratch with ambari 2.0.1 (the combination which I used for earlier clusters as well), right now it looks good, also that particular host is up as a worker node
- to cleanup all the nodes I used the hostcleanup python script and some additional directory deletions (due to non-standard directories, e.g. for logs)
I will do the Ambari upgrade later on.
Thanks for all you efforts
thanks for getting back that quickly.
I am assuming that Ambari 2.1.2 is supporting HDP 2.2.4 ?!?!
Which version should I use instead ?
ok, tried doing it again after removing any hadoop_* packages on that node from a previous run, but again:
... 2016-01-12 20:29:26,813 - Skipping XmlConfig['core-site.xml'] due to only_if 2016-01-12 20:29:26,814 - Can only link configs for HDP-2.3 and higher.
$ ls -al /usr/hdp total 12 drwxr-xr-x 3 root root 4096 Jan 12 20:28 . drwxr-xr-x 16 root root 4096 Jan 12 20:28 .. drwxr-xr-x 13 root root 4096 Jan 12 20:29 220.127.116.11-2 hence, the link 'current' is missing...thereby the datanode fails at starting up
Maybe I will start completely from scratch with a fresh installation.
Which Ambari version would you recommend for setting up a fresh HDP 2.2.4 ? Yes, unfortunately I have to use that HDP version, an upgrade is planned for later this year ;)
Hi Gerd, can you check the following,
rpm -qa | grep hadoop # make sure it is installed and that /etc/hadoop/conf exists ls -la /usr/hdp/current/hadoop-client # should be a symlink to /etc/hadoop/conf, if not, create it
thanks for your answer, but it is not exactly how it should look like ;) (maybe you are referring to an older version, before adding the version number into the installation directory tree).
Currently, the linking should be:
/usr/hdp/current -> /usr/hdp/18.104.22.168-2
/usr/hdp/22.214.171.124-2/hadoop/conf -> /etc/hadoop/conf