stderr: Traceback (most recent call last): File "/usr/lib/ambari-agent/lib/resource_management/libraries/providers/hdfs_resource.py", line 287, in _run_command result_dict = json.loads(out) File "/usr/lib/ambari-agent/lib/ambari_simplejson/__init__.py", line 307, in loads return _default_decoder.decode(s) File "/usr/lib/ambari-agent/lib/ambari_simplejson/decoder.py", line 335, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/usr/lib/ambari-agent/lib/ambari_simplejson/decoder.py", line 353, in raw_decode raise ValueError("No JSON object could be decoded") ValueError: No JSON object could be decoded The above exception was the cause of the following exception: Traceback (most recent call last): File "/var/lib/ambari-agent/cache/stacks/HDP/3.0/services/YARN/package/scripts/resourcemanager.py", line 261, in Resourcemanager().execute() File "/usr/lib/ambari-agent/lib/resource_management/libraries/script/script.py", line 352, in execute method(env) File "/var/lib/ambari-agent/cache/stacks/HDP/3.0/services/YARN/package/scripts/resourcemanager.py", line 142, in start self.wait_for_dfs_directories_created(params.entity_groupfs_store_dir, params.entity_groupfs_active_dir) File "/var/lib/ambari-agent/cache/stacks/HDP/3.0/services/YARN/package/scripts/resourcemanager.py", line 211, in wait_for_dfs_directories_created self.wait_for_dfs_directory_created(dir_path, ignored_dfs_dirs) File "/usr/lib/ambari-agent/lib/resource_management/libraries/functions/decorator.py", line 62, in wrapper return function(*args, **kwargs) File "/var/lib/ambari-agent/cache/stacks/HDP/3.0/services/YARN/package/scripts/resourcemanager.py", line 236, in wait_for_dfs_directory_created list_status = util.run_command(dir_path, 'GETFILESTATUS', method='GET', ignore_status_codes=['404'], assertable_result=False) File "/usr/lib/ambari-agent/lib/resource_management/libraries/providers/hdfs_resource.py", line 214, in run_command return self._run_command(*args, **kwargs) File "/usr/lib/ambari-agent/lib/resource_management/libraries/providers/hdfs_resource.py", line 295, in _run_command raise WebHDFSCallException(err_msg, result_dict) resource_management.libraries.providers.hdfs_resource.WebHDFSCallException: Execution of 'curl -sS -L -w '%{http_code}' -X GET -d '' -H 'Content-Length: 0' 'http://nodetwo:50070/webhdfs/v1/ats/done/?op=GETFILESTATUS&user.name=hdfs'' returned status_code=503. Network Error
Network Error (dns_server_failure)

Your request could not be processed because an error occurred contacting the DNS server.
The DNS server may be temporarily unavailable, or there could be a network problem.

stdout: 2019-06-21 11:12:54,904 - Stack Feature Version Info: Cluster Stack=3.1, Command Stack=None, Command Version=3.1.0.0-78 -> 3.1.0.0-78 2019-06-21 11:12:54,931 - Using hadoop conf dir: /usr/hdp/3.1.0.0-78/hadoop/conf 2019-06-21 11:12:55,150 - Stack Feature Version Info: Cluster Stack=3.1, Command Stack=None, Command Version=3.1.0.0-78 -> 3.1.0.0-78 2019-06-21 11:12:55,159 - Using hadoop conf dir: /usr/hdp/3.1.0.0-78/hadoop/conf 2019-06-21 11:12:55,161 - Group['hdfs'] {} 2019-06-21 11:12:55,162 - Group['hadoop'] {} 2019-06-21 11:12:55,162 - Group['users'] {} 2019-06-21 11:12:55,162 - User['yarn-ats'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None} 2019-06-21 11:12:55,167 - User['zookeeper'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None} 2019-06-21 11:12:55,168 - User['ams'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None} 2019-06-21 11:12:55,169 - User['ambari-qa'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop', 'users'], 'uid': None} 2019-06-21 11:12:55,169 - User['hdfs'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hdfs', 'hadoop'], 'uid': None} 2019-06-21 11:12:55,170 - User['yarn'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None} 2019-06-21 11:12:55,171 - User['mapred'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop'], 'uid': None} 2019-06-21 11:12:55,171 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555} 2019-06-21 11:12:55,172 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa 0'] {'not_if': '(test $(id -u ambari-qa) -gt 1000) || (false)'} 2019-06-21 11:12:55,180 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa 0'] due to not_if 2019-06-21 11:12:55,181 - Group['hdfs'] {} 2019-06-21 11:12:55,181 - User['hdfs'] {'fetch_nonlocal_groups': True, 'groups': ['hdfs', 'hadoop', u'hdfs']} 2019-06-21 11:12:55,181 - FS Type: HDFS 2019-06-21 11:12:55,181 - Directory['/etc/hadoop'] {'mode': 0755} 2019-06-21 11:12:55,207 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/hadoop-env.sh'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop'} 2019-06-21 11:12:55,216 - Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] {'owner': 'hdfs', 'group': 'hadoop', 'mode': 01777} 2019-06-21 11:12:55,244 - Execute[('setenforce', '0')] {'not_if': '(! which getenforce ) || (which getenforce && getenforce | grep -q Disabled)', 'sudo': True, 'only_if': 'test -f /selinux/enforce'} 2019-06-21 11:12:55,249 - Skipping Execute[('setenforce', '0')] due to not_if 2019-06-21 11:12:55,250 - Directory['/var/log/hadoop'] {'owner': 'root', 'create_parents': True, 'group': 'hadoop', 'mode': 0775, 'cd_access': 'a'} 2019-06-21 11:12:55,251 - Directory['/var/run/hadoop'] {'owner': 'root', 'create_parents': True, 'group': 'root', 'cd_access': 'a'} 2019-06-21 11:12:55,252 - Directory['/var/run/hadoop/hdfs'] {'owner': 'hdfs', 'cd_access': 'a'} 2019-06-21 11:12:55,252 - Directory['/tmp/hadoop-hdfs'] {'owner': 'hdfs', 'create_parents': True, 'cd_access': 'a'} 2019-06-21 11:12:55,256 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/commons-logging.properties'] {'content': Template('commons-logging.properties.j2'), 'owner': 'hdfs'} 2019-06-21 11:12:55,260 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/health_check'] {'content': Template('health_check.j2'), 'owner': 'hdfs'} 2019-06-21 11:12:55,271 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/log4j.properties'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop', 'mode': 0644} 2019-06-21 11:12:55,302 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/hadoop-metrics2.properties'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop'} 2019-06-21 11:12:55,302 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/task-log4j.properties'] {'content': StaticFile('task-log4j.properties'), 'mode': 0755} 2019-06-21 11:12:55,308 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/configuration.xsl'] {'owner': 'hdfs', 'group': 'hadoop'} 2019-06-21 11:12:55,318 - File['/etc/hadoop/conf/topology_mappings.data'] {'owner': 'hdfs', 'content': Template('topology_mappings.data.j2'), 'only_if': 'test -d /etc/hadoop/conf', 'group': 'hadoop', 'mode': 0644} 2019-06-21 11:12:55,322 - File['/etc/hadoop/conf/topology_script.py'] {'content': StaticFile('topology_script.py'), 'only_if': 'test -d /etc/hadoop/conf', 'mode': 0755} 2019-06-21 11:12:55,325 - Skipping unlimited key JCE policy check and setup since it is not required 2019-06-21 11:12:55,820 - Using hadoop conf dir: /usr/hdp/3.1.0.0-78/hadoop/conf 2019-06-21 11:12:55,820 - Stack Feature Version Info: Cluster Stack=3.1, Command Stack=None, Command Version=3.1.0.0-78 -> 3.1.0.0-78 2019-06-21 11:12:55,850 - Using hadoop conf dir: /usr/hdp/3.1.0.0-78/hadoop/conf 2019-06-21 11:12:55,863 - Directory['/var/log/hadoop-yarn'] {'group': 'hadoop', 'cd_access': 'a', 'create_parents': True, 'ignore_failures': True, 'mode': 0775, 'owner': 'yarn'} 2019-06-21 11:12:55,865 - Directory['/var/run/hadoop-yarn'] {'owner': 'yarn', 'create_parents': True, 'group': 'hadoop', 'cd_access': 'a'} 2019-06-21 11:12:55,865 - Directory['/var/run/hadoop-yarn/yarn'] {'owner': 'yarn', 'create_parents': True, 'group': 'hadoop', 'cd_access': 'a'} 2019-06-21 11:12:55,866 - Directory['/var/log/hadoop-yarn/yarn'] {'owner': 'yarn', 'group': 'hadoop', 'create_parents': True, 'cd_access': 'a'} 2019-06-21 11:12:55,866 - Directory['/var/run/hadoop-mapreduce'] {'owner': 'mapred', 'create_parents': True, 'group': 'hadoop', 'cd_access': 'a'} 2019-06-21 11:12:55,866 - Directory['/var/run/hadoop-mapreduce/mapred'] {'owner': 'mapred', 'create_parents': True, 'group': 'hadoop', 'cd_access': 'a'} 2019-06-21 11:12:55,867 - Directory['/var/log/hadoop-mapreduce'] {'owner': 'mapred', 'create_parents': True, 'group': 'hadoop', 'cd_access': 'a'} 2019-06-21 11:12:55,867 - Directory['/var/log/hadoop-mapreduce/mapred'] {'owner': 'mapred', 'group': 'hadoop', 'create_parents': True, 'cd_access': 'a'} 2019-06-21 11:12:55,868 - Directory['/usr/hdp/3.1.0.0-78/hadoop/conf/embedded-yarn-ats-hbase'] {'owner': 'yarn-ats', 'group': 'hadoop', 'create_parents': True, 'cd_access': 'a'} 2019-06-21 11:12:55,868 - Directory['/etc/hadoop/conf'] {'create_parents': True, 'mode': 0755, 'cd_access': 'a'} 2019-06-21 11:12:55,872 - File['/etc/hadoop/conf/yarn.exclude'] {'owner': 'yarn', 'content': Template('exclude_hosts_list.j2'), 'group': 'hadoop'} 2019-06-21 11:12:55,873 - File['/var/log/hadoop-yarn/yarn/hadoop-mapreduce.jobsummary.log'] {'owner': 'yarn', 'group': 'hadoop'} 2019-06-21 11:12:55,881 - Directory['/usr/lib/ambari-logsearch-logfeeder/conf'] {'create_parents': True, 'mode': 0755, 'cd_access': 'a'} 2019-06-21 11:12:55,881 - Generate Log Feeder config file: /usr/lib/ambari-logsearch-logfeeder/conf/input.config-yarn.json 2019-06-21 11:12:55,881 - File['/usr/lib/ambari-logsearch-logfeeder/conf/input.config-yarn.json'] {'content': Template('input.config-yarn.json.j2'), 'mode': 0644} 2019-06-21 11:12:55,882 - XmlConfig['core-site.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/3.1.0.0-78/hadoop/conf', 'mode': 0644, 'configuration_attributes': {u'final': {u'fs.defaultFS': u'true'}}, 'owner': 'hdfs', 'configurations': ...} 2019-06-21 11:12:55,893 - Generating config: /usr/hdp/3.1.0.0-78/hadoop/conf/core-site.xml 2019-06-21 11:12:55,893 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/core-site.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'} 2019-06-21 11:12:55,921 - XmlConfig['hdfs-site.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/3.1.0.0-78/hadoop/conf', 'mode': 0644, 'configuration_attributes': {u'final': {u'dfs.datanode.failed.volumes.tolerated': u'true', u'dfs.datanode.data.dir': u'true', u'dfs.namenode.http-address': u'true', u'dfs.namenode.name.dir': u'true', u'dfs.webhdfs.enabled': u'true'}}, 'owner': 'hdfs', 'configurations': ...} 2019-06-21 11:12:55,930 - Generating config: /usr/hdp/3.1.0.0-78/hadoop/conf/hdfs-site.xml 2019-06-21 11:12:55,931 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/hdfs-site.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'} 2019-06-21 11:12:55,977 - XmlConfig['mapred-site.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/3.1.0.0-78/hadoop/conf', 'mode': 0644, 'configuration_attributes': {}, 'owner': 'yarn', 'configurations': ...} 2019-06-21 11:12:55,987 - Generating config: /usr/hdp/3.1.0.0-78/hadoop/conf/mapred-site.xml 2019-06-21 11:12:55,987 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/mapred-site.xml'] {'owner': 'yarn', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'} 2019-06-21 11:12:56,029 - Changing owner for /usr/hdp/3.1.0.0-78/hadoop/conf/mapred-site.xml from 120 to yarn 2019-06-21 11:12:56,029 - XmlConfig['yarn-site.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/3.1.0.0-78/hadoop/conf', 'mode': 0644, 'configuration_attributes': {u'hidden': {u'hadoop.registry.dns.bind-port': u'true'}}, 'owner': 'yarn', 'configurations': ...} 2019-06-21 11:12:56,038 - Generating config: /usr/hdp/3.1.0.0-78/hadoop/conf/yarn-site.xml 2019-06-21 11:12:56,038 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/yarn-site.xml'] {'owner': 'yarn', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'} 2019-06-21 11:12:56,262 - XmlConfig['capacity-scheduler.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/3.1.0.0-78/hadoop/conf', 'mode': 0644, 'configuration_attributes': {}, 'owner': 'yarn', 'configurations': ...} 2019-06-21 11:12:56,276 - Generating config: /usr/hdp/3.1.0.0-78/hadoop/conf/capacity-scheduler.xml 2019-06-21 11:12:56,277 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/capacity-scheduler.xml'] {'owner': 'yarn', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'} 2019-06-21 11:12:56,295 - Changing owner for /usr/hdp/3.1.0.0-78/hadoop/conf/capacity-scheduler.xml from 1004 to yarn 2019-06-21 11:12:56,295 - XmlConfig['hbase-site.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/3.1.0.0-78/hadoop/conf/embedded-yarn-ats-hbase', 'mode': 0644, 'configuration_attributes': {}, 'owner': 'yarn-ats', 'configurations': ...} 2019-06-21 11:12:56,310 - Generating config: /usr/hdp/3.1.0.0-78/hadoop/conf/embedded-yarn-ats-hbase/hbase-site.xml 2019-06-21 11:12:56,310 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/embedded-yarn-ats-hbase/hbase-site.xml'] {'owner': 'yarn-ats', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'} 2019-06-21 11:12:56,364 - XmlConfig['resource-types.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/3.1.0.0-78/hadoop/conf', 'mode': 0644, 'configuration_attributes': {}, 'owner': 'yarn', 'configurations': {u'yarn.resource-types.yarn.io_gpu.maximum-allocation': u'8', u'yarn.resource-types': u''}} 2019-06-21 11:12:56,390 - Generating config: /usr/hdp/3.1.0.0-78/hadoop/conf/resource-types.xml 2019-06-21 11:12:56,390 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/resource-types.xml'] {'owner': 'yarn', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'} 2019-06-21 11:12:56,394 - File['/etc/security/limits.d/yarn.conf'] {'content': Template('yarn.conf.j2'), 'mode': 0644} 2019-06-21 11:12:56,405 - File['/etc/security/limits.d/mapreduce.conf'] {'content': Template('mapreduce.conf.j2'), 'mode': 0644} 2019-06-21 11:12:56,427 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/yarn-env.sh'] {'content': InlineTemplate(...), 'owner': 'yarn', 'group': 'hadoop', 'mode': 0755} 2019-06-21 11:12:56,435 - File['/usr/hdp/3.1.0.0-78/hadoop-yarn/bin/container-executor'] {'group': 'hadoop', 'mode': 02050} 2019-06-21 11:12:56,445 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/container-executor.cfg'] {'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644} 2019-06-21 11:12:56,445 - Directory['/cgroups_test/cpu'] {'group': 'hadoop', 'create_parents': True, 'mode': 0755, 'cd_access': 'a'} 2019-06-21 11:12:56,453 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/mapred-env.sh'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'mode': 0755} 2019-06-21 11:12:56,454 - Directory['/var/log/hadoop-yarn/nodemanager/recovery-state'] {'owner': 'yarn', 'group': 'hadoop', 'create_parents': True, 'mode': 0755, 'cd_access': 'a'} 2019-06-21 11:12:56,464 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/taskcontroller.cfg'] {'content': Template('taskcontroller.cfg.j2'), 'owner': 'hdfs'} 2019-06-21 11:12:56,465 - XmlConfig['mapred-site.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/3.1.0.0-78/hadoop/conf', 'mode': 0644, 'configuration_attributes': {}, 'owner': 'mapred', 'configurations': ...} 2019-06-21 11:12:56,482 - Generating config: /usr/hdp/3.1.0.0-78/hadoop/conf/mapred-site.xml 2019-06-21 11:12:56,483 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/mapred-site.xml'] {'owner': 'mapred', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'} 2019-06-21 11:12:56,533 - Changing owner for /usr/hdp/3.1.0.0-78/hadoop/conf/mapred-site.xml from 119 to mapred 2019-06-21 11:12:56,534 - XmlConfig['capacity-scheduler.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/3.1.0.0-78/hadoop/conf', 'mode': 0644, 'configuration_attributes': {}, 'owner': 'hdfs', 'configurations': ...} 2019-06-21 11:12:56,547 - Generating config: /usr/hdp/3.1.0.0-78/hadoop/conf/capacity-scheduler.xml 2019-06-21 11:12:56,548 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/capacity-scheduler.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'} 2019-06-21 11:12:56,569 - Changing owner for /usr/hdp/3.1.0.0-78/hadoop/conf/capacity-scheduler.xml from 119 to hdfs 2019-06-21 11:12:56,569 - XmlConfig['ssl-client.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/3.1.0.0-78/hadoop/conf', 'mode': 0644, 'configuration_attributes': {}, 'owner': 'hdfs', 'configurations': ...} 2019-06-21 11:12:56,585 - Generating config: /usr/hdp/3.1.0.0-78/hadoop/conf/ssl-client.xml 2019-06-21 11:12:56,585 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/ssl-client.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'} 2019-06-21 11:12:56,591 - Directory['/usr/hdp/3.1.0.0-78/hadoop/conf/secure'] {'owner': 'root', 'create_parents': True, 'group': 'hadoop', 'cd_access': 'a'} 2019-06-21 11:12:56,592 - XmlConfig['ssl-client.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/3.1.0.0-78/hadoop/conf/secure', 'mode': 0644, 'configuration_attributes': {}, 'owner': 'hdfs', 'configurations': ...} 2019-06-21 11:12:56,608 - Generating config: /usr/hdp/3.1.0.0-78/hadoop/conf/secure/ssl-client.xml 2019-06-21 11:12:56,609 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/secure/ssl-client.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'} 2019-06-21 11:12:56,617 - XmlConfig['ssl-server.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/3.1.0.0-78/hadoop/conf', 'mode': 0644, 'configuration_attributes': {}, 'owner': 'hdfs', 'configurations': ...} 2019-06-21 11:12:56,631 - Generating config: /usr/hdp/3.1.0.0-78/hadoop/conf/ssl-server.xml 2019-06-21 11:12:56,633 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/ssl-server.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'} 2019-06-21 11:12:56,640 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/ssl-client.xml.example'] {'owner': 'mapred', 'group': 'hadoop', 'mode': 0644} 2019-06-21 11:12:56,643 - File['/usr/hdp/3.1.0.0-78/hadoop/conf/ssl-server.xml.example'] {'owner': 'mapred', 'group': 'hadoop', 'mode': 0644} 2019-06-21 11:12:56,644 - Verifying DFS directories where ATS stores time line data for active and completed applications. 2019-06-21 11:12:56,644 - Verifying if DFS directory '/ats/done/' exists. 2019-06-21 11:12:56,645 - call['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET -d '"'"''"'"' -H '"'"'Content-Length: 0'"'"' '"'"'http://nodetwo:50070/webhdfs/v1/ats/done/?op=GETFILESTATUS&user.name=hdfs'"'"' 1>/tmp/tmp9ZEKLp 2>/tmp/tmpyWhGlu''] {'logoutput': None, 'quiet': False} 2019-06-21 11:13:12,343 - call returned (0, '') 2019-06-21 11:13:12,343 - get_user_call_output returned (0, u'\r\nNetwork Error\r\n\r\n\r\n\r\n
\r\n
\r\n
\r\n\r\n\r\n\r\n\r\n\r\n
\r\n\r\nNetwork Error (dns_server_failure)\r\n
\r\n
\r\n
\r\n
\r\n\r\nYour request could not be processed because an error occurred contacting the DNS server.\r\n\r\n
\r\n\r\nThe DNS server may be temporarily unavailable, or there could be a network problem.\r\n\r\n
\r\n\r\n
\r\n
\r\n
\r\n
\r\n\r\n\r\n503', u'') 2019-06-21 11:13:12,344 - Will retry 7 time(s), caught exception: Execution of 'curl -sS -L -w '%{http_code}' -X GET -d '' -H 'Content-Length: 0' 'http://nodetwo:50070/webhdfs/v1/ats/done/?op=GETFILESTATUS&user.name=hdfs'' returned status_code=503. Network Error
Network Error (dns_server_failure)

Your request could not be processed because an error occurred contacting the DNS server.
The DNS server may be temporarily unavailable, or there could be a network problem.

. Sleeping for 20 sec(s) 2019-06-21 11:13:32,363 - Verifying if DFS directory '/ats/done/' exists. 2019-06-21 11:13:32,364 - call['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET -d '"'"''"'"' -H '"'"'Content-Length: 0'"'"' '"'"'http://nodetwo:50070/webhdfs/v1/ats/done/?op=GETFILESTATUS&user.name=hdfs'"'"' 1>/tmp/tmpIrSyDr 2>/tmp/tmpPJTCnA''] {'logoutput': None, 'quiet': False} 2019-06-21 11:13:48,017 - call returned (0, '') 2019-06-21 11:13:48,017 - get_user_call_output returned (0, u'\r\nNetwork Error\r\n\r\n\r\n\r\n
\r\n
\r\n
\r\n\r\n\r\n\r\n\r\n\r\n
\r\n\r\nNetwork Error (dns_server_failure)\r\n
\r\n
\r\n
\r\n
\r\n\r\nYour request could not be processed because an error occurred contacting the DNS server.\r\n\r\n
\r\n\r\nThe DNS server may be temporarily unavailable, or there could be a network problem.\r\n\r\n
\r\n\r\n
\r\n
\r\n
\r\n
\r\n\r\n\r\n503', u'') 2019-06-21 11:13:48,033 - Will retry 6 time(s), caught exception: Execution of 'curl -sS -L -w '%{http_code}' -X GET -d '' -H 'Content-Length: 0' 'http://nodetwo:50070/webhdfs/v1/ats/done/?op=GETFILESTATUS&user.name=hdfs'' returned status_code=503. Network Error
Network Error (dns_server_failure)

Your request could not be processed because an error occurred contacting the DNS server.
The DNS server may be temporarily unavailable, or there could be a network problem.

. Sleeping for 20 sec(s) 2019-06-21 11:14:08,047 - Verifying if DFS directory '/ats/done/' exists. 2019-06-21 11:14:08,048 - call['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET -d '"'"''"'"' -H '"'"'Content-Length: 0'"'"' '"'"'http://nodetwo:50070/webhdfs/v1/ats/done/?op=GETFILESTATUS&user.name=hdfs'"'"' 1>/tmp/tmp0OSBJq 2>/tmp/tmpr6O49H''] {'logoutput': None, 'quiet': False} 2019-06-21 11:14:23,693 - call returned (0, '') 2019-06-21 11:14:23,693 - get_user_call_output returned (0, u'\r\nNetwork Error\r\n\r\n\r\n\r\n
\r\n
\r\n
\r\n\r\n\r\n\r\n\r\n\r\n
\r\n\r\nNetwork Error (dns_server_failure)\r\n
\r\n
\r\n
\r\n
\r\n\r\nYour request could not be processed because an error occurred contacting the DNS server.\r\n\r\n
\r\n\r\nThe DNS server may be temporarily unavailable, or there could be a network problem.\r\n\r\n
\r\n\r\n
\r\n
\r\n
\r\n
\r\n\r\n\r\n503', u'') 2019-06-21 11:14:23,694 - Will retry 5 time(s), caught exception: Execution of 'curl -sS -L -w '%{http_code}' -X GET -d '' -H 'Content-Length: 0' 'http://nodetwo:50070/webhdfs/v1/ats/done/?op=GETFILESTATUS&user.name=hdfs'' returned status_code=503. Network Error
Network Error (dns_server_failure)

Your request could not be processed because an error occurred contacting the DNS server.
The DNS server may be temporarily unavailable, or there could be a network problem.

. Sleeping for 20 sec(s) 2019-06-21 11:14:43,712 - Verifying if DFS directory '/ats/done/' exists. 2019-06-21 11:14:43,713 - call['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET -d '"'"''"'"' -H '"'"'Content-Length: 0'"'"' '"'"'http://nodetwo:50070/webhdfs/v1/ats/done/?op=GETFILESTATUS&user.name=hdfs'"'"' 1>/tmp/tmp6MQMkO 2>/tmp/tmpeyJQfB''] {'logoutput': None, 'quiet': False} 2019-06-21 11:14:59,329 - call returned (0, '') 2019-06-21 11:14:59,329 - get_user_call_output returned (0, u'\r\nNetwork Error\r\n\r\n\r\n\r\n
\r\n
\r\n
\r\n\r\n\r\n\r\n\r\n\r\n
\r\n\r\nNetwork Error (dns_server_failure)\r\n
\r\n
\r\n
\r\n
\r\n\r\nYour request could not be processed because an error occurred contacting the DNS server.\r\n\r\n
\r\n\r\nThe DNS server may be temporarily unavailable, or there could be a network problem.\r\n\r\n
\r\n\r\n
\r\n
\r\n
\r\n
\r\n\r\n\r\n503', u'') 2019-06-21 11:14:59,330 - Will retry 4 time(s), caught exception: Execution of 'curl -sS -L -w '%{http_code}' -X GET -d '' -H 'Content-Length: 0' 'http://nodetwo:50070/webhdfs/v1/ats/done/?op=GETFILESTATUS&user.name=hdfs'' returned status_code=503. Network Error
Network Error (dns_server_failure)

Your request could not be processed because an error occurred contacting the DNS server.
The DNS server may be temporarily unavailable, or there could be a network problem.

. Sleeping for 20 sec(s) 2019-06-21 11:15:19,350 - Verifying if DFS directory '/ats/done/' exists. 2019-06-21 11:15:19,351 - call['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET -d '"'"''"'"' -H '"'"'Content-Length: 0'"'"' '"'"'http://nodetwo:50070/webhdfs/v1/ats/done/?op=GETFILESTATUS&user.name=hdfs'"'"' 1>/tmp/tmp6hJTY2 2>/tmp/tmpm6KNzS''] {'logoutput': None, 'quiet': False} 2019-06-21 11:15:28,957 - call returned (0, '') 2019-06-21 11:15:28,958 - get_user_call_output returned (0, u'\r\nNetwork Error\r\n\r\n\r\n\r\n
\r\n
\r\n
\r\n\r\n\r\n\r\n\r\n\r\n
\r\n\r\nNetwork Error (dns_server_failure)\r\n
\r\n
\r\n
\r\n
\r\n\r\nYour request could not be processed because an error occurred contacting the DNS server.\r\n\r\n
\r\n\r\nThe DNS server may be temporarily unavailable, or there could be a network problem.\r\n\r\n
\r\n\r\n
\r\n
\r\n
\r\n
\r\n\r\n\r\n503', u'') 2019-06-21 11:15:28,958 - Will retry 3 time(s), caught exception: Execution of 'curl -sS -L -w '%{http_code}' -X GET -d '' -H 'Content-Length: 0' 'http://nodetwo:50070/webhdfs/v1/ats/done/?op=GETFILESTATUS&user.name=hdfs'' returned status_code=503. Network Error
Network Error (dns_server_failure)

Your request could not be processed because an error occurred contacting the DNS server.
The DNS server may be temporarily unavailable, or there could be a network problem.

. Sleeping for 20 sec(s) 2019-06-21 11:15:48,978 - Verifying if DFS directory '/ats/done/' exists. 2019-06-21 11:15:48,979 - call['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET -d '"'"''"'"' -H '"'"'Content-Length: 0'"'"' '"'"'http://nodetwo:50070/webhdfs/v1/ats/done/?op=GETFILESTATUS&user.name=hdfs'"'"' 1>/tmp/tmpLDHXhn 2>/tmp/tmpusMpFC''] {'logoutput': None, 'quiet': False} 2019-06-21 11:15:56,590 - call returned (0, '') 2019-06-21 11:15:56,590 - get_user_call_output returned (0, u'\r\nNetwork Error\r\n\r\n\r\n\r\n
\r\n
\r\n
\r\n\r\n\r\n\r\n\r\n\r\n
\r\n\r\nNetwork Error (dns_server_failure)\r\n
\r\n
\r\n
\r\n
\r\n\r\nYour request could not be processed because an error occurred contacting the DNS server.\r\n\r\n
\r\n\r\nThe DNS server may be temporarily unavailable, or there could be a network problem.\r\n\r\n
\r\n\r\n
\r\n
\r\n
\r\n
\r\n\r\n\r\n503', u'') 2019-06-21 11:15:56,613 - Will retry 2 time(s), caught exception: Execution of 'curl -sS -L -w '%{http_code}' -X GET -d '' -H 'Content-Length: 0' 'http://nodetwo:50070/webhdfs/v1/ats/done/?op=GETFILESTATUS&user.name=hdfs'' returned status_code=503. Network Error
Network Error (dns_server_failure)

Your request could not be processed because an error occurred contacting the DNS server.
The DNS server may be temporarily unavailable, or there could be a network problem.

. Sleeping for 20 sec(s) 2019-06-21 11:16:16,631 - Verifying if DFS directory '/ats/done/' exists. 2019-06-21 11:16:16,632 - call['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET -d '"'"''"'"' -H '"'"'Content-Length: 0'"'"' '"'"'http://nodetwo:50070/webhdfs/v1/ats/done/?op=GETFILESTATUS&user.name=hdfs'"'"' 1>/tmp/tmpbuo6FU 2>/tmp/tmplsWf13''] {'logoutput': None, 'quiet': False} 2019-06-21 11:16:26,279 - call returned (0, '') 2019-06-21 11:16:26,279 - get_user_call_output returned (0, u'\r\nNetwork Error\r\n\r\n\r\n\r\n
\r\n
\r\n
\r\n\r\n\r\n\r\n\r\n\r\n
\r\n\r\nNetwork Error (dns_server_failure)\r\n
\r\n
\r\n
\r\n
\r\n\r\nYour request could not be processed because an error occurred contacting the DNS server.\r\n\r\n
\r\n\r\nThe DNS server may be temporarily unavailable, or there could be a network problem.\r\n\r\n
\r\n\r\n
\r\n
\r\n
\r\n
\r\n\r\n\r\n503', u'') 2019-06-21 11:16:26,285 - Will retry 1 time(s), caught exception: Execution of 'curl -sS -L -w '%{http_code}' -X GET -d '' -H 'Content-Length: 0' 'http://nodetwo:50070/webhdfs/v1/ats/done/?op=GETFILESTATUS&user.name=hdfs'' returned status_code=503. Network Error
Network Error (dns_server_failure)

Your request could not be processed because an error occurred contacting the DNS server.
The DNS server may be temporarily unavailable, or there could be a network problem.

. Sleeping for 20 sec(s) 2019-06-21 11:16:46,305 - Verifying if DFS directory '/ats/done/' exists. 2019-06-21 11:16:46,306 - call['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'curl -sS -L -w '"'"'%{http_code}'"'"' -X GET -d '"'"''"'"' -H '"'"'Content-Length: 0'"'"' '"'"'http://nodetwo:50070/webhdfs/v1/ats/done/?op=GETFILESTATUS&user.name=hdfs'"'"' 1>/tmp/tmp172aHK 2>/tmp/tmpMXZXgq''] {'logoutput': None, 'quiet': False} 2019-06-21 11:16:55,899 - call returned (0, '') 2019-06-21 11:16:55,899 - get_user_call_output returned (0, u'\r\nNetwork Error\r\n\r\n\r\n\r\n
\r\n
\r\n
\r\n\r\n\r\n\r\n\r\n\r\n
\r\n\r\nNetwork Error (dns_server_failure)\r\n
\r\n
\r\n
\r\n
\r\n\r\nYour request could not be processed because an error occurred contacting the DNS server.\r\n\r\n
\r\n\r\nThe DNS server may be temporarily unavailable, or there could be a network problem.\r\n\r\n
\r\n\r\n
\r\n
\r\n
\r\n
\r\n\r\n\r\n503', u'') Command failed after 1 tries