Support Questions
Find answers, ask questions, and share your expertise

I am getting error in NFS gateway while starring HDFS

I am getting error in NFS gateway while starring HDFS

New Contributor

Below is error stact trace

Traceback (most recent call last):
  File "/var/lib/ambari-agent/cache/common-services/HDFS/2.1.0.2.0/package/scripts/nfsgateway.py", line 147, in <module>
    NFSGateway().execute()
  File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 280, in execute
    method(env)
  File "/var/lib/ambari-agent/cache/common-services/HDFS/2.1.0.2.0/package/scripts/nfsgateway.py", line 57, in start
    self.configure(env)
  File "/var/lib/ambari-agent/cache/common-services/HDFS/2.1.0.2.0/package/scripts/nfsgateway.py", line 71, in configure
    nfsgateway(action="configure")
  File "/var/lib/ambari-agent/cache/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_nfsgateway.py", line 66, in nfsgateway
    group = params.user_group,
  File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 155, in __init__
    self.env.run()
  File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 160, in run
    self.run_action(resource, action)
  File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 124, in run_action
    provider_action()
  File "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py", line 191, in action_create
    sudo.makedir(path, self.resource.mode or 0755)
  File "/usr/lib/python2.6/site-packages/resource_management/core/sudo.py", line 90, in makedir
    os.mkdir(path)
OSError: [Errno 22] Invalid argument: '/tmp/.hdfs-nfs'

stdout: /var/lib/ambari-agent/data/output-294.txt

2017-10-13 23:09:13,213 - The hadoop conf dir /usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for version 2.5.0.0-1245
2017-10-13 23:09:13,217 - Checking if need to create versioned conf dir /etc/hadoop/2.5.0.0-1245/0
2017-10-13 23:09:13,218 - call[('ambari-python-wrap', '/usr/bin/conf-select', 'create-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.0.0-1245', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False, 'stderr': -1}
2017-10-13 23:09:13,289 - call returned (1, '/etc/hadoop/2.5.0.0-1245/0 exist already', '')
2017-10-13 23:09:13,290 - checked_call[('ambari-python-wrap', '/usr/bin/conf-select', 'set-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.0.0-1245', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False}
2017-10-13 23:09:13,357 - checked_call returned (0, '')
2017-10-13 23:09:13,358 - Ensuring that hadoop has the correct symlink structure
2017-10-13 23:09:13,359 - Using hadoop conf dir: /usr/hdp/current/hadoop-client/conf
2017-10-13 23:09:13,669 - The hadoop conf dir /usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for version 2.5.0.0-1245
2017-10-13 23:09:13,670 - Checking if need to create versioned conf dir /etc/hadoop/2.5.0.0-1245/0
2017-10-13 23:09:13,670 - call[('ambari-python-wrap', '/usr/bin/conf-select', 'create-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.0.0-1245', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False, 'stderr': -1}
2017-10-13 23:09:13,758 - call returned (1, '/etc/hadoop/2.5.0.0-1245/0 exist already', '')
2017-10-13 23:09:13,758 - checked_call[('ambari-python-wrap', '/usr/bin/conf-select', 'set-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.0.0-1245', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False}
2017-10-13 23:09:13,839 - checked_call returned (0, '')
2017-10-13 23:09:13,844 - Ensuring that hadoop has the correct symlink structure
2017-10-13 23:09:13,845 - Using hadoop conf dir: /usr/hdp/current/hadoop-client/conf
2017-10-13 23:09:13,846 - Group['livy'] {}
2017-10-13 23:09:13,848 - Group['spark'] {}
2017-10-13 23:09:13,849 - Group['ranger'] {}
2017-10-13 23:09:13,849 - Group['zeppelin'] {}
2017-10-13 23:09:13,854 - Group['hadoop'] {}
2017-10-13 23:09:13,854 - Group['users'] {}
2017-10-13 23:09:13,854 - Group['knox'] {}
2017-10-13 23:09:13,855 - User['hive'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,856 - User['storm'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,857 - User['zookeeper'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,858 - User['infra-solr'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,859 - User['oozie'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['users']}
2017-10-13 23:09:13,864 - User['atlas'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,866 - User['ams'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,866 - User['falcon'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['users']}
2017-10-13 23:09:13,867 - User['ranger'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['ranger']}
2017-10-13 23:09:13,868 - User['tez'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['users']}
2017-10-13 23:09:13,869 - User['zeppelin'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,874 - User['livy'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,875 - User['spark'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,876 - User['ambari-qa'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['users']}
2017-10-13 23:09:13,877 - User['flume'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,878 - User['kafka'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,879 - User['hdfs'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,883 - User['sqoop'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,884 - User['yarn'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,885 - User['mapred'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,886 - User['hbase'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,887 - User['knox'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,889 - User['hcat'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']}
2017-10-13 23:09:13,894 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2017-10-13 23:09:13,897 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] {'not_if': '(test $(id -u ambari-qa) -gt 1000) || (false)'}
2017-10-13 23:09:13,942 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] due to not_if
2017-10-13 23:09:13,943 - Directory['/tmp/hbase-hbase'] {'owner': 'hbase', 'create_parents': True, 'mode': 0775, 'cd_access': 'a'}
2017-10-13 23:09:13,945 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555}
2017-10-13 23:09:13,946 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] {'not_if': '(test $(id -u hbase) -gt 1000) || (false)'}
2017-10-13 23:09:13,983 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] due to not_if
2017-10-13 23:09:13,983 - Group['hdfs'] {}
2017-10-13 23:09:13,984 - User['hdfs'] {'fetch_nonlocal_groups': True, 'groups': ['hadoop', 'hdfs']}
2017-10-13 23:09:13,988 - FS Type: 
2017-10-13 23:09:13,989 - Directory['/etc/hadoop'] {'mode': 0755}
2017-10-13 23:09:14,030 - File['/usr/hdp/current/hadoop-client/conf/hadoop-env.sh'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop'}
2017-10-13 23:09:14,031 - Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] {'owner': 'hdfs', 'group': 'hadoop', 'mode': 01777}
2017-10-13 23:09:14,060 - Execute[('setenforce', '0')] {'not_if': '(! which getenforce ) || (which getenforce && getenforce | grep -q Disabled)', 'sudo': True, 'only_if': 'test -f /selinux/enforce'}
2017-10-13 23:09:14,089 - Skipping Execute[('setenforce', '0')] due to not_if
2017-10-13 23:09:14,090 - Directory['/var/log/hadoop'] {'owner': 'root', 'create_parents': True, 'group': 'hadoop', 'mode': 0775, 'cd_access': 'a'}
2017-10-13 23:09:14,097 - Directory['/var/run/hadoop'] {'owner': 'root', 'create_parents': True, 'group': 'root', 'cd_access': 'a'}
2017-10-13 23:09:14,097 - Directory['/tmp/hadoop-hdfs'] {'owner': 'hdfs', 'create_parents': True, 'cd_access': 'a'}
2017-10-13 23:09:14,115 - File['/usr/hdp/current/hadoop-client/conf/commons-logging.properties'] {'content': Template('commons-logging.properties.j2'), 'owner': 'hdfs'}
2017-10-13 23:09:14,118 - File['/usr/hdp/current/hadoop-client/conf/health_check'] {'content': Template('health_check.j2'), 'owner': 'hdfs'}
2017-10-13 23:09:14,119 - File['/usr/hdp/current/hadoop-client/conf/log4j.properties'] {'content': ..., 'owner': 'hdfs', 'group': 'hadoop', 'mode': 0644}
2017-10-13 23:09:14,155 - File['/usr/hdp/current/hadoop-client/conf/hadoop-metrics2.properties'] {'content': Template('hadoop-metrics2.properties.j2'), 'owner': 'hdfs', 'group': 'hadoop'}
2017-10-13 23:09:14,156 - File['/usr/hdp/current/hadoop-client/conf/task-log4j.properties'] {'content': StaticFile('task-log4j.properties'), 'mode': 0755}
2017-10-13 23:09:14,157 - File['/usr/hdp/current/hadoop-client/conf/configuration.xsl'] {'owner': 'hdfs', 'group': 'hadoop'}
2017-10-13 23:09:14,170 - File['/etc/hadoop/conf/topology_mappings.data'] {'owner': 'hdfs', 'content': Template('topology_mappings.data.j2'), 'only_if': 'test -d /etc/hadoop/conf', 'group': 'hadoop'}
2017-10-13 23:09:14,195 - File['/etc/hadoop/conf/topology_script.py'] {'content': StaticFile('topology_script.py'), 'only_if': 'test -d /etc/hadoop/conf', 'mode': 0755}
2017-10-13 23:09:14,640 - The hadoop conf dir /usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for version 2.5.0.0-1245
2017-10-13 23:09:14,640 - Checking if need to create versioned conf dir /etc/hadoop/2.5.0.0-1245/0
2017-10-13 23:09:14,641 - call[('ambari-python-wrap', '/usr/bin/conf-select', 'create-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.0.0-1245', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False, 'stderr': -1}
2017-10-13 23:09:14,709 - call returned (1, '/etc/hadoop/2.5.0.0-1245/0 exist already', '')
2017-10-13 23:09:14,709 - checked_call[('ambari-python-wrap', '/usr/bin/conf-select', 'set-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.0.0-1245', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False}
2017-10-13 23:09:14,786 - checked_call returned (0, '')
2017-10-13 23:09:14,787 - Ensuring that hadoop has the correct symlink structure
2017-10-13 23:09:14,787 - Using hadoop conf dir: /usr/hdp/current/hadoop-client/conf
2017-10-13 23:09:14,796 - Stack Feature Version Info: stack_version=2.5, version=2.5.0.0-1245, current_cluster_version=2.5.0.0-1245 -> 2.5.0.0-1245
2017-10-13 23:09:14,799 - The hadoop conf dir /usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for version 2.5.0.0-1245
2017-10-13 23:09:14,800 - Checking if need to create versioned conf dir /etc/hadoop/2.5.0.0-1245/0
2017-10-13 23:09:14,800 - call[('ambari-python-wrap', '/usr/bin/conf-select', 'create-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.0.0-1245', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False, 'stderr': -1}
2017-10-13 23:09:14,870 - call returned (1, '/etc/hadoop/2.5.0.0-1245/0 exist already', '')
2017-10-13 23:09:14,871 - checked_call[('ambari-python-wrap', '/usr/bin/conf-select', 'set-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.0.0-1245', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False}
2017-10-13 23:09:14,938 - checked_call returned (0, '')
2017-10-13 23:09:14,939 - Ensuring that hadoop has the correct symlink structure
2017-10-13 23:09:14,939 - Using hadoop conf dir: /usr/hdp/current/hadoop-client/conf
2017-10-13 23:09:14,951 - checked_call['rpm -q --queryformat '%{version}-%{release}' hdp-select | sed -e 's/\.el[0-9]//g''] {'stderr': -1}
2017-10-13 23:09:15,068 - checked_call returned (0, '2.5.0.0-1245', '')
2017-10-13 23:09:15,079 - Directory['/etc/security/limits.d'] {'owner': 'root', 'create_parents': True, 'group': 'root'}
2017-10-13 23:09:15,100 - File['/etc/security/limits.d/hdfs.conf'] {'content': Template('hdfs.conf.j2'), 'owner': 'root', 'group': 'root', 'mode': 0644}
2017-10-13 23:09:15,101 - XmlConfig['hadoop-policy.xml'] {'owner': 'hdfs', 'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'configuration_attributes': {}, 'configurations': ...}
2017-10-13 23:09:15,127 - Generating config: /usr/hdp/current/hadoop-client/conf/hadoop-policy.xml
2017-10-13 23:09:15,128 - File['/usr/hdp/current/hadoop-client/conf/hadoop-policy.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None, 'encoding': 'UTF-8'}
2017-10-13 23:09:15,151 - XmlConfig['ssl-client.xml'] {'owner': 'hdfs', 'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'configuration_attributes': {}, 'configurations': ...}
2017-10-13 23:09:15,179 - Generating config: /usr/hdp/current/hadoop-client/conf/ssl-client.xml
2017-10-13 23:09:15,180 - File['/usr/hdp/current/hadoop-client/conf/ssl-client.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None, 'encoding': 'UTF-8'}
2017-10-13 23:09:15,197 - Directory['/usr/hdp/current/hadoop-client/conf/secure'] {'owner': 'root', 'create_parents': True, 'group': 'hadoop', 'cd_access': 'a'}
2017-10-13 23:09:15,198 - XmlConfig['ssl-client.xml'] {'owner': 'hdfs', 'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf/secure', 'configuration_attributes': {}, 'configurations': ...}
2017-10-13 23:09:15,220 - Generating config: /usr/hdp/current/hadoop-client/conf/secure/ssl-client.xml
2017-10-13 23:09:15,221 - File['/usr/hdp/current/hadoop-client/conf/secure/ssl-client.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None, 'encoding': 'UTF-8'}
2017-10-13 23:09:15,238 - XmlConfig['ssl-server.xml'] {'owner': 'hdfs', 'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'configuration_attributes': {}, 'configurations': ...}
2017-10-13 23:09:15,259 - Generating config: /usr/hdp/current/hadoop-client/conf/ssl-server.xml
2017-10-13 23:09:15,260 - File['/usr/hdp/current/hadoop-client/conf/ssl-server.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None, 'encoding': 'UTF-8'}
2017-10-13 23:09:15,278 - XmlConfig['hdfs-site.xml'] {'owner': 'hdfs', 'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'configuration_attributes': ..., 'configurations': ...}
2017-10-13 23:09:15,299 - Generating config: /usr/hdp/current/hadoop-client/conf/hdfs-site.xml
2017-10-13 23:09:15,299 - File['/usr/hdp/current/hadoop-client/conf/hdfs-site.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None, 'encoding': 'UTF-8'}
2017-10-13 23:09:15,420 - XmlConfig['core-site.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'mode': 0644, 'configuration_attributes': {'final': {'fs.defaultFS': 'true'}}, 'owner': 'hdfs', 'configurations': ...}
2017-10-13 23:09:15,441 - Generating config: /usr/hdp/current/hadoop-client/conf/core-site.xml
2017-10-13 23:09:15,441 - File['/usr/hdp/current/hadoop-client/conf/core-site.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'}
2017-10-13 23:09:15,518 - File['/usr/hdp/current/hadoop-client/conf/slaves'] {'content': Template('slaves.j2'), 'owner': 'hdfs'}
2017-10-13 23:09:15,519 - Package['lzo'] {'retry_on_repo_unavailability': False, 'retry_count': 5}
2017-10-13 23:09:16,062 - Skipping installation of existing package lzo
2017-10-13 23:09:16,063 - Package['hadoop-lzo-native'] {'retry_on_repo_unavailability': False, 'retry_count': 5}
2017-10-13 23:09:16,515 - Skipping installation of existing package hadoop-lzo-native
2017-10-13 23:09:16,515 - Package['hadooplzo_2_5_0_0_1245'] {'retry_on_repo_unavailability': False, 'retry_count': 5}
2017-10-13 23:09:16,944 - Skipping installation of existing package hadooplzo_2_5_0_0_1245
2017-10-13 23:09:16,949 - Package['hadooplzo_2_5_0_0_1245-native'] {'retry_on_repo_unavailability': False, 'retry_count': 5}
2017-10-13 23:09:17,507 - Skipping installation of existing package hadooplzo_2_5_0_0_1245-native
2017-10-13 23:09:17,507 - Directory['/tmp/.hdfs-nfs'] {'owner': 'hdfs', 'group': 'hadoop'}
2017-10-13 23:09:17,508 - Creating directory Directory['/tmp/.hdfs-nfs'] since it doesn't exist.
 

Command failed after 1 tries

1 REPLY 1

Re: I am getting error in NFS gateway while starring HDFS

Contributor