stderr: Traceback (most recent call last): File "/var/lib/ambari-agent/cache/common-services/HDFS/2.1.0.2.0/package/scripts/namenode.py", line 420, in NameNode().execute() File "/usr/lib/python2.6/site-packages/resource_management/libraries/script/script.py", line 280, in execute method(env) File "/var/lib/ambari-agent/cache/common-services/HDFS/2.1.0.2.0/package/scripts/namenode.py", line 101, in start upgrade_suspended=params.upgrade_suspended, env=env) File "/usr/lib/python2.6/site-packages/ambari_commons/os_family_impl.py", line 89, in thunk return fn(*args, **kwargs) File "/var/lib/ambari-agent/cache/common-services/HDFS/2.1.0.2.0/package/scripts/hdfs_namenode.py", line 156, in namenode create_log_dir=True File "/var/lib/ambari-agent/cache/common-services/HDFS/2.1.0.2.0/package/scripts/utils.py", line 269, in service Execute(daemon_cmd, not_if=process_id_exists_command, environment=hadoop_env_exports) File "/usr/lib/python2.6/site-packages/resource_management/core/base.py", line 155, in __init__ self.env.run() File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 160, in run self.run_action(resource, action) File "/usr/lib/python2.6/site-packages/resource_management/core/environment.py", line 124, in run_action provider_action() File "/usr/lib/python2.6/site-packages/resource_management/core/providers/system.py", line 273, in action_run tries=self.resource.tries, try_sleep=self.resource.try_sleep) File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 70, in inner result = function(command, **kwargs) File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 92, in checked_call tries=tries, try_sleep=try_sleep) File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 140, in _call_wrapper result = _call(command, **kwargs_copy) File "/usr/lib/python2.6/site-packages/resource_management/core/shell.py", line 293, in _call raise ExecutionFailed(err_msg, code, out, err) resource_management.core.exceptions.ExecutionFailed: Execution of 'ambari-sudo.sh su hdfs -l -s /bin/bash -c 'ulimit -c unlimited ; /usr/hdp/current/hadoop-client/sbin/hadoop-daemon.sh --config /usr/hdp/current/hadoop-client/conf start namenode'' returned 1. starting namenode, logging to /var/log/hadoop/hdfs/hadoop-hdfs-namenode-slzusxxxxxxxx.sxxx.sxxxx.out stdout: 2017-11-20 09:46:42,764 - The hadoop conf dir /usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for version 2.5.0.0-1245 2017-11-20 09:46:42,765 - Checking if need to create versioned conf dir /etc/hadoop/2.5.0.0-1245/0 2017-11-20 09:46:42,765 - call[('ambari-python-wrap', '/usr/bin/conf-select', 'create-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.0.0-1245', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False, 'stderr': -1} 2017-11-20 09:46:42,794 - call returned (1, '/etc/hadoop/2.5.0.0-1245/0 exist already', '') 2017-11-20 09:46:42,794 - checked_call[('ambari-python-wrap', '/usr/bin/conf-select', 'set-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.0.0-1245', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False} 2017-11-20 09:46:42,822 - checked_call returned (0, '') 2017-11-20 09:46:42,823 - Ensuring that hadoop has the correct symlink structure 2017-11-20 09:46:42,823 - Using hadoop conf dir: /usr/hdp/current/hadoop-client/conf 2017-11-20 09:46:42,958 - The hadoop conf dir /usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for version 2.5.0.0-1245 2017-11-20 09:46:42,959 - Checking if need to create versioned conf dir /etc/hadoop/2.5.0.0-1245/0 2017-11-20 09:46:42,959 - call[('ambari-python-wrap', '/usr/bin/conf-select', 'create-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.0.0-1245', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False, 'stderr': -1} 2017-11-20 09:46:42,988 - call returned (1, '/etc/hadoop/2.5.0.0-1245/0 exist already', '') 2017-11-20 09:46:42,988 - checked_call[('ambari-python-wrap', '/usr/bin/conf-select', 'set-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.0.0-1245', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False} 2017-11-20 09:46:43,016 - checked_call returned (0, '') 2017-11-20 09:46:43,017 - Ensuring that hadoop has the correct symlink structure 2017-11-20 09:46:43,017 - Using hadoop conf dir: /usr/hdp/current/hadoop-client/conf 2017-11-20 09:46:43,019 - Group['livy'] {} 2017-11-20 09:46:43,021 - Group['spark'] {} 2017-11-20 09:46:43,021 - Group['zeppelin'] {} 2017-11-20 09:46:43,022 - Group['hadoop'] {} 2017-11-20 09:46:43,022 - Group['users'] {} 2017-11-20 09:46:43,022 - Group['knox'] {} 2017-11-20 09:46:43,022 - User['hive'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,024 - User['storm'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,025 - User['zookeeper'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,027 - User['infra-solr'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,028 - User['oozie'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['users']} 2017-11-20 09:46:43,029 - User['atlas'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,030 - User['ams'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,032 - User['falcon'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['users']} 2017-11-20 09:46:43,033 - User['tez'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['users']} 2017-11-20 09:46:43,034 - User['zeppelin'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,035 - User['accumulo'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,037 - User['mahout'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,038 - User['livy'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,039 - User['spark'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,040 - User['ambari-qa'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['users']} 2017-11-20 09:46:43,042 - User['kafka'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,043 - User['hdfs'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,044 - User['sqoop'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,045 - User['yarn'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,047 - User['mapred'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,048 - User['hbase'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,050 - User['knox'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,051 - User['hcat'] {'gid': 'hadoop', 'fetch_nonlocal_groups': True, 'groups': ['hadoop']} 2017-11-20 09:46:43,052 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555} 2017-11-20 09:46:43,054 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] {'not_if': '(test $(id -u ambari-qa) -gt 1000) || (false)'} 2017-11-20 09:46:43,063 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh ambari-qa /tmp/hadoop-ambari-qa,/tmp/hsperfdata_ambari-qa,/home/ambari-qa,/tmp/ambari-qa,/tmp/sqoop-ambari-qa'] due to not_if 2017-11-20 09:46:43,063 - Directory['/tmp/hbase-hbase'] {'owner': 'hbase', 'create_parents': True, 'mode': 0775, 'cd_access': 'a'} 2017-11-20 09:46:43,064 - File['/var/lib/ambari-agent/tmp/changeUid.sh'] {'content': StaticFile('changeToSecureUid.sh'), 'mode': 0555} 2017-11-20 09:46:43,066 - Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] {'not_if': '(test $(id -u hbase) -gt 1000) || (false)'} 2017-11-20 09:46:43,074 - Skipping Execute['/var/lib/ambari-agent/tmp/changeUid.sh hbase /home/hbase,/tmp/hbase,/usr/bin/hbase,/var/log/hbase,/tmp/hbase-hbase'] due to not_if 2017-11-20 09:46:43,074 - Group['hdfs'] {} 2017-11-20 09:46:43,075 - User['hdfs'] {'fetch_nonlocal_groups': True, 'groups': ['hadoop', 'hdfs']} 2017-11-20 09:46:43,076 - FS Type: 2017-11-20 09:46:43,076 - Directory['/etc/hadoop'] {'mode': 0755} 2017-11-20 09:46:43,096 - File['/usr/hdp/current/hadoop-client/conf/hadoop-env.sh'] {'content': InlineTemplate(...), 'owner': 'hdfs', 'group': 'hadoop'} 2017-11-20 09:46:43,097 - Directory['/var/lib/ambari-agent/tmp/hadoop_java_io_tmpdir'] {'owner': 'hdfs', 'group': 'hadoop', 'mode': 01777} 2017-11-20 09:46:43,111 - Execute[('setenforce', '0')] {'not_if': '(! which getenforce ) || (which getenforce && getenforce | grep -q Disabled)', 'sudo': True, 'only_if': 'test -f /selinux/enforce'} 2017-11-20 09:46:43,121 - Skipping Execute[('setenforce', '0')] due to not_if 2017-11-20 09:46:43,121 - Directory['/var/log/hadoop'] {'owner': 'root', 'create_parents': True, 'group': 'hadoop', 'mode': 0775, 'cd_access': 'a'} 2017-11-20 09:46:43,124 - Directory['/var/run/hadoop'] {'owner': 'root', 'create_parents': True, 'group': 'root', 'cd_access': 'a'} 2017-11-20 09:46:43,124 - Directory['/tmp/hadoop-hdfs'] {'owner': 'hdfs', 'create_parents': True, 'cd_access': 'a'} 2017-11-20 09:46:43,132 - File['/usr/hdp/current/hadoop-client/conf/commons-logging.properties'] {'content': Template('commons-logging.properties.j2'), 'owner': 'hdfs'} 2017-11-20 09:46:43,134 - File['/usr/hdp/current/hadoop-client/conf/health_check'] {'content': Template('health_check.j2'), 'owner': 'hdfs'} 2017-11-20 09:46:43,135 - File['/usr/hdp/current/hadoop-client/conf/log4j.properties'] {'content': ..., 'owner': 'hdfs', 'group': 'hadoop', 'mode': 0644} 2017-11-20 09:46:43,149 - File['/usr/hdp/current/hadoop-client/conf/hadoop-metrics2.properties'] {'content': Template('hadoop-metrics2.properties.j2'), 'owner': 'hdfs', 'group': 'hadoop'} 2017-11-20 09:46:43,150 - File['/usr/hdp/current/hadoop-client/conf/task-log4j.properties'] {'content': StaticFile('task-log4j.properties'), 'mode': 0755} 2017-11-20 09:46:43,151 - File['/usr/hdp/current/hadoop-client/conf/configuration.xsl'] {'owner': 'hdfs', 'group': 'hadoop'} 2017-11-20 09:46:43,156 - File['/etc/hadoop/conf/topology_mappings.data'] {'owner': 'hdfs', 'content': Template('topology_mappings.data.j2'), 'only_if': 'test -d /etc/hadoop/conf', 'group': 'hadoop'} 2017-11-20 09:46:43,165 - File['/etc/hadoop/conf/topology_script.py'] {'content': StaticFile('topology_script.py'), 'only_if': 'test -d /etc/hadoop/conf', 'mode': 0755} 2017-11-20 09:46:43,352 - The hadoop conf dir /usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for version 2.5.0.0-1245 2017-11-20 09:46:43,352 - Checking if need to create versioned conf dir /etc/hadoop/2.5.0.0-1245/0 2017-11-20 09:46:43,353 - call[('ambari-python-wrap', '/usr/bin/conf-select', 'create-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.0.0-1245', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False, 'stderr': -1} 2017-11-20 09:46:43,381 - call returned (1, '/etc/hadoop/2.5.0.0-1245/0 exist already', '') 2017-11-20 09:46:43,382 - checked_call[('ambari-python-wrap', '/usr/bin/conf-select', 'set-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.0.0-1245', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False} 2017-11-20 09:46:43,410 - checked_call returned (0, '') 2017-11-20 09:46:43,410 - Ensuring that hadoop has the correct symlink structure 2017-11-20 09:46:43,411 - Using hadoop conf dir: /usr/hdp/current/hadoop-client/conf 2017-11-20 09:46:43,411 - Stack Feature Version Info: stack_version=2.5, version=2.5.0.0-1245, current_cluster_version=2.5.0.0-1245 -> 2.5.0.0-1245 2017-11-20 09:46:43,414 - The hadoop conf dir /usr/hdp/current/hadoop-client/conf exists, will call conf-select on it for version 2.5.0.0-1245 2017-11-20 09:46:43,414 - Checking if need to create versioned conf dir /etc/hadoop/2.5.0.0-1245/0 2017-11-20 09:46:43,415 - call[('ambari-python-wrap', '/usr/bin/conf-select', 'create-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.0.0-1245', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False, 'stderr': -1} 2017-11-20 09:46:43,442 - call returned (1, '/etc/hadoop/2.5.0.0-1245/0 exist already', '') 2017-11-20 09:46:43,443 - checked_call[('ambari-python-wrap', '/usr/bin/conf-select', 'set-conf-dir', '--package', 'hadoop', '--stack-version', '2.5.0.0-1245', '--conf-version', '0')] {'logoutput': False, 'sudo': True, 'quiet': False} 2017-11-20 09:46:43,472 - checked_call returned (0, '') 2017-11-20 09:46:43,473 - Ensuring that hadoop has the correct symlink structure 2017-11-20 09:46:43,473 - Using hadoop conf dir: /usr/hdp/current/hadoop-client/conf 2017-11-20 09:46:43,481 - checked_call['rpm -q --queryformat '%{version}-%{release}' hdp-select | sed -e 's/\.el[0-9]//g''] {'stderr': -1} 2017-11-20 09:46:43,528 - checked_call returned (0, '2.5.0.0-1245', '') 2017-11-20 09:46:43,534 - Directory['/etc/security/limits.d'] {'owner': 'root', 'create_parents': True, 'group': 'root'} 2017-11-20 09:46:43,542 - File['/etc/security/limits.d/hdfs.conf'] {'content': Template('hdfs.conf.j2'), 'owner': 'root', 'group': 'root', 'mode': 0644} 2017-11-20 09:46:43,543 - XmlConfig['hadoop-policy.xml'] {'owner': 'hdfs', 'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'configuration_attributes': {}, 'configurations': ...} 2017-11-20 09:46:43,555 - Generating config: /usr/hdp/current/hadoop-client/conf/hadoop-policy.xml 2017-11-20 09:46:43,555 - File['/usr/hdp/current/hadoop-client/conf/hadoop-policy.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None, 'encoding': 'UTF-8'} 2017-11-20 09:46:43,566 - XmlConfig['ssl-client.xml'] {'owner': 'hdfs', 'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'configuration_attributes': {}, 'configurations': ...} 2017-11-20 09:46:43,577 - Generating config: /usr/hdp/current/hadoop-client/conf/ssl-client.xml 2017-11-20 09:46:43,577 - File['/usr/hdp/current/hadoop-client/conf/ssl-client.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None, 'encoding': 'UTF-8'} 2017-11-20 09:46:43,585 - Directory['/usr/hdp/current/hadoop-client/conf/secure'] {'owner': 'root', 'create_parents': True, 'group': 'hadoop', 'cd_access': 'a'} 2017-11-20 09:46:43,586 - XmlConfig['ssl-client.xml'] {'owner': 'hdfs', 'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf/secure', 'configuration_attributes': {}, 'configurations': ...} 2017-11-20 09:46:43,597 - Generating config: /usr/hdp/current/hadoop-client/conf/secure/ssl-client.xml 2017-11-20 09:46:43,597 - File['/usr/hdp/current/hadoop-client/conf/secure/ssl-client.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None, 'encoding': 'UTF-8'} 2017-11-20 09:46:43,605 - XmlConfig['ssl-server.xml'] {'owner': 'hdfs', 'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'configuration_attributes': {}, 'configurations': ...} 2017-11-20 09:46:43,616 - Generating config: /usr/hdp/current/hadoop-client/conf/ssl-server.xml 2017-11-20 09:46:43,616 - File['/usr/hdp/current/hadoop-client/conf/ssl-server.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None, 'encoding': 'UTF-8'} 2017-11-20 09:46:43,624 - XmlConfig['hdfs-site.xml'] {'owner': 'hdfs', 'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'configuration_attributes': {'final': {'dfs.support.append': 'true', 'dfs.datanode.data.dir': 'true', 'dfs.namenode.http-address': 'true', 'dfs.namenode.name.dir': 'true', 'dfs.webhdfs.enabled': 'true', 'dfs.datanode.failed.volumes.tolerated': 'true'}}, 'configurations': ...} 2017-11-20 09:46:43,635 - Generating config: /usr/hdp/current/hadoop-client/conf/hdfs-site.xml 2017-11-20 09:46:43,635 - File['/usr/hdp/current/hadoop-client/conf/hdfs-site.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': None, 'encoding': 'UTF-8'} 2017-11-20 09:46:43,696 - XmlConfig['core-site.xml'] {'group': 'hadoop', 'conf_dir': '/usr/hdp/current/hadoop-client/conf', 'mode': 0644, 'configuration_attributes': {'final': {'fs.defaultFS': 'true'}}, 'owner': 'hdfs', 'configurations': ...} 2017-11-20 09:46:43,707 - Generating config: /usr/hdp/current/hadoop-client/conf/core-site.xml 2017-11-20 09:46:43,707 - File['/usr/hdp/current/hadoop-client/conf/core-site.xml'] {'owner': 'hdfs', 'content': InlineTemplate(...), 'group': 'hadoop', 'mode': 0644, 'encoding': 'UTF-8'} 2017-11-20 09:46:43,738 - File['/usr/hdp/current/hadoop-client/conf/slaves'] {'content': Template('slaves.j2'), 'owner': 'hdfs'} 2017-11-20 09:46:43,740 - Directory['/hadoop/nn/namenodeprimary'] {'owner': 'hdfs', 'group': 'hadoop', 'create_parents': True, 'mode': 0755, 'cd_access': 'a'} 2017-11-20 09:46:43,741 - Called service start with upgrade_type: None 2017-11-20 09:46:43,741 - Ranger admin not installed 2017-11-20 09:46:43,743 - File['/etc/hadoop/conf/dfs.exclude'] {'owner': 'hdfs', 'content': Template('exclude_hosts_list.j2'), 'group': 'hadoop'} 2017-11-20 09:46:43,744 - Options for start command are: 2017-11-20 09:46:43,744 - Directory['/var/run/hadoop'] {'owner': 'hdfs', 'group': 'hadoop', 'mode': 0755} 2017-11-20 09:46:43,745 - Changing owner for /var/run/hadoop from 0 to hdfs 2017-11-20 09:46:43,745 - Changing group for /var/run/hadoop from 0 to hadoop 2017-11-20 09:46:43,745 - Directory['/var/run/hadoop/hdfs'] {'owner': 'hdfs', 'group': 'hadoop', 'create_parents': True} 2017-11-20 09:46:43,745 - Directory['/var/log/hadoop/hdfs'] {'owner': 'hdfs', 'group': 'hadoop', 'create_parents': True} 2017-11-20 09:46:43,746 - File['/var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid'] {'action': ['delete'], 'not_if': 'ambari-sudo.sh -H -E test -f /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid && ambari-sudo.sh -H -E pgrep -F /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid'} 2017-11-20 09:46:43,762 - Deleting File['/var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid'] 2017-11-20 09:46:43,762 - Execute['ambari-sudo.sh su hdfs -l -s /bin/bash -c 'ulimit -c unlimited ; /usr/hdp/current/hadoop-client/sbin/hadoop-daemon.sh --config /usr/hdp/current/hadoop-client/conf start namenode''] {'environment': {'HADOOP_LIBEXEC_DIR': '/usr/hdp/current/hadoop-client/libexec'}, 'not_if': 'ambari-sudo.sh -H -E test -f /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid && ambari-sudo.sh -H -E pgrep -F /var/run/hadoop/hdfs/hadoop-hdfs-namenode.pid'} 2017-11-20 09:46:47,863 - Execute['find /var/log/hadoop/hdfs -maxdepth 1 -type f -name '*' -exec echo '==> {} <==' \; -exec tail -n 40 {} \;'] {'logoutput': True, 'ignore_failures': True, 'user': 'hdfs'} ==> /var/log/hadoop/hdfs/gc.log-201711071521 <== 2017-11-07T22:53:03.741+0100: 27066.338: [GC (Allocation Failure) 2017-11-07T22:53:03.741+0100: 27066.338: [ParNew: 105691K->768K(118016K), 0.0044343 secs] 126702K->21779K(1035520K), 0.0045569 secs] [Times: user=0.02 sys=0.00, real=0.01 secs] 2017-11-08T01:17:11.533+0100: 35714.130: [GC (Allocation Failure) 2017-11-08T01:17:11.533+0100: 35714.130: [ParNew: 105728K->946K(118016K), 0.0055099 secs] 126739K->21957K(1035520K), 0.0055869 secs] [Times: user=0.01 sys=0.00, real=0.01 secs] 2017-11-08T03:52:12.627+0100: 45015.225: [GC (Allocation Failure) 2017-11-08T03:52:12.627+0100: 45015.225: [ParNew: 105906K->980K(118016K), 0.0050527 secs] 126917K->21991K(1035520K), 0.0051573 secs] [Times: user=0.02 sys=0.00, real=0.01 secs] 2017-11-08T06:31:13.610+0100: 54556.207: [GC (Allocation Failure) 2017-11-08T06:31:13.610+0100: 54556.207: [ParNew: 105940K->686K(118016K), 0.0053639 secs] 126951K->22013K(1035520K), 0.0054780 secs] [Times: user=0.02 sys=0.00, real=0.01 secs] 2017-11-08T09:09:14.524+0100: 64037.122: [GC (Allocation Failure) 2017-11-08T09:09:14.525+0100: 64037.122: [ParNew: 105646K->563K(118016K), 0.0067316 secs] 126973K->21895K(1035520K), 0.0068177 secs] [Times: user=0.01 sys=0.00, real=0.00 secs] 2017-11-08T11:46:15.567+0100: 73458.165: [GC (Allocation Failure) 2017-11-08T11:46:15.567+0100: 73458.165: [ParNew: 105523K->561K(118016K), 0.0047586 secs] 126855K->21899K(1035520K), 0.0048500 secs] [Times: user=0.01 sys=0.00, real=0.00 secs] 2017-11-08T14:04:16.319+0100: 81738.917: [GC (Allocation Failure) 2017-11-08T14:04:16.319+0100: 81738.917: [ParNew: 105521K->490K(118016K), 0.0044713 secs] 126859K->21841K(1035520K), 0.0045993 secs] [Times: user=0.01 sys=0.00, real=0.01 secs] 2017-11-08T16:17:17.146+0100: 89719.743: [GC (Allocation Failure) 2017-11-08T16:17:17.146+0100: 89719.743: [ParNew: 105450K->491K(118016K), 0.0052748 secs] 126801K->21843K(1035520K), 0.0053749 secs] [Times: user=0.01 sys=0.00, real=0.01 secs] 2017-11-08T18:35:17.911+0100: 98000.509: [GC (Allocation Failure) 2017-11-08T18:35:17.911+0100: 98000.509: [ParNew: 105451K->474K(118016K), 0.0039868 secs] 126803K->21832K(1035520K), 0.0040609 secs] [Times: user=0.01 sys=0.00, real=0.01 secs] 2017-11-08T20:31:19.368+0100: 104961.965: [GC (Allocation Failure) 2017-11-08T20:31:19.368+0100: 104961.965: [ParNew: 105434K->435K(118016K), 0.0174474 secs] 126792K->21796K(1035520K), 0.0175512 secs] [Times: user=0.06 sys=0.00, real=0.02 secs] 2017-11-08T22:47:19.375+0100: 113121.972: [GC (Allocation Failure) 2017-11-08T22:47:19.375+0100: 113121.973: [ParNew: 105395K->467K(118016K), 0.0038860 secs] 126756K->21829K(1035520K), 0.0039840 secs] [Times: user=0.01 sys=0.00, real=0.00 secs] 2017-11-09T01:06:43.490+0100: 121486.087: [GC (Allocation Failure) 2017-11-09T01:06:43.490+0100: 121486.087: [ParNew: 105427K->488K(118016K), 0.0040562 secs] 126789K->21853K(1035520K), 0.0041386 secs] [Times: user=0.01 sys=0.00, real=0.00 secs] 2017-11-09T03:23:44.382+0100: 129706.979: [GC (Allocation Failure) 2017-11-09T03:23:44.382+0100: 129706.979: [ParNew: 105448K->538K(118016K), 0.0040844 secs] 126813K->21906K(1035520K), 0.0041692 secs] [Times: user=0.01 sys=0.00, real=0.01 secs] 2017-11-09T05:40:19.416+0100: 137902.014: [GC (Allocation Failure) 2017-11-09T05:40:19.416+0100: 137902.014: [ParNew: 105498K->474K(118016K), 0.0068873 secs] 126866K->21844K(1035520K), 0.0071759 secs] [Times: user=0.02 sys=0.00, real=0.01 secs] 2017-11-09T07:58:46.104+0100: 146208.702: [GC (Allocation Failure) 2017-11-09T07:58:46.104+0100: 146208.702: [ParNew: 105434K->491K(118016K), 0.0100049 secs] 126804K->21860K(1035520K), 0.0101000 secs] [Times: user=0.03 sys=0.00, real=0.01 secs] 2017-11-09T10:14:46.979+0100: 154369.577: [GC (Allocation Failure) 2017-11-09T10:14:46.979+0100: 154369.577: [ParNew: 105451K->501K(118016K), 0.0200495 secs] 126820K->21871K(1035520K), 0.0201507 secs] [Times: user=0.06 sys=0.00, real=0.02 secs] 2017-11-09T12:32:47.740+0100: 162650.337: [GC (Allocation Failure) 2017-11-09T12:32:47.740+0100: 162650.337: [ParNew: 105461K->497K(118016K), 0.0048214 secs] 126831K->21872K(1035520K), 0.0049083 secs] [Times: user=0.01 sys=0.00, real=0.01 secs] 2017-11-09T14:29:48.434+0100: 169671.031: [GC (Allocation Failure) 2017-11-09T14:29:48.434+0100: 169671.031: [ParNew: 105457K->461K(118016K), 0.0090959 secs] 126832K->21836K(1035520K), 0.0091777 secs] [Times: user=0.03 sys=0.00, real=0.01 secs] 2017-11-09T16:24:49.262+0100: 176571.860: [GC (Allocation Failure) 2017-11-09T16:24:49.262+0100: 176571.860: [ParNew: 105421K->455K(118016K), 0.0041125 secs] 126796K->21831K(1035520K), 0.0041998 secs] [Times: user=0.01 sys=0.00, real=0.00 secs] 2017-11-09T18:21:49.884+0100: 183592.482: [GC (Allocation Failure) 2017-11-09T18:21:49.884+0100: 183592.482: [ParNew: 105415K->437K(118016K), 0.0039211 secs] 126791K->21816K(1035520K), 0.0040383 secs] [Times: user=0.02 sys=0.00, real=0.01 secs] 2017-11-09T20:18:50.489+0100: 190613.086: [GC (Allocation Failure) 2017-11-09T20:18:50.489+0100: 190613.086: [ParNew: 105397K->445K(118016K), 0.0041996 secs] 126776K->21825K(1035520K), 0.0042986 secs] [Times: user=0.01 sys=0.00, real=0.01 secs] 2017-11-09T21:51:51.107+0100: 196193.704: [GC (Allocation Failure) 2017-11-09T21:51:51.107+0100: 196193.704: [ParNew: 105405K->396K(118016K), 0.0037150 secs] 126785K->21783K(1035520K), 0.0037859 secs] [Times: user=0.01 sys=0.00, real=0.01 secs] 2017-11-09T23:48:51.694+0100: 203214.292: [GC (Allocation Failure) 2017-11-09T23:48:51.694+0100: 203214.292: [ParNew: 105356K->431K(118016K), 0.0036814 secs] 126743K->21827K(1035520K), 0.0037672 secs] [Times: user=0.01 sys=0.00, real=0.01 secs] 2017-11-10T01:23:19.394+0100: 208881.991: [GC (Allocation Failure) 2017-11-10T01:23:19.394+0100: 208881.991: [ParNew: 105391K->396K(118016K), 0.0058656 secs] 126787K->21793K(1035520K), 0.0059389 secs] [Times: user=0.02 sys=0.00, real=0.01 secs] 2017-11-10T03:19:54.958+0100: 215877.555: [GC (Allocation Failure) 2017-11-10T03:19:54.958+0100: 215877.555: [ParNew: 105356K->429K(118016K), 0.0038042 secs] 126753K->21827K(1035520K), 0.0038820 secs] [Times: user=0.01 sys=0.00, real=0.00 secs] 2017-11-10T04:51:55.571+0100: 221398.169: [GC (Allocation Failure) 2017-11-10T04:51:55.571+0100: 221398.169: [ParNew: 105389K->382K(118016K), 0.0033761 secs] 126787K->21780K(1035520K), 0.0034551 secs] [Times: user=0.01 sys=0.00, real=0.00 secs] 2017-11-10T06:48:19.397+0100: 228381.994: [GC (Allocation Failure) 2017-11-10T06:48:19.397+0100: 228381.994: [ParNew: 105342K->439K(118016K), 0.0036918 secs] 126740K->21837K(1035520K), 0.0037854 secs] [Times: user=0.01 sys=0.00, real=0.00 secs] 2017-11-10T08:44:56.809+0100: 235379.407: [GC (Allocation Failure) 2017-11-10T08:44:56.809+0100: 235379.407: [ParNew: 105399K->431K(118016K), 0.0044360 secs] 126797K->21831K(1035520K), 0.0045575 secs] [Times: user=0.01 sys=0.00, real=0.00 secs] 2017-11-10T10:17:19.370+0100: 240921.968: [GC (Allocation Failure) 2017-11-10T10:17:19.370+0100: 240921.968: [ParNew: 105391K->380K(118016K), 0.0044975 secs] 126791K->21783K(1035520K), 0.0045730 secs] [Times: user=0.01 sys=0.00, real=0.00 secs] 2017-11-10T12:13:58.125+0100: 247920.722: [GC (Allocation Failure) 2017-11-10T12:13:58.125+0100: 247920.722: [ParNew: 105340K->421K(118016K), 0.0047748 secs] 126743K->21824K(1035520K), 0.0048544 secs] [Times: user=0.01 sys=0.00, real=0.00 secs] 2017-11-10T13:48:58.592+0100: 253621.190: [GC (Allocation Failure) 2017-11-10T13:48:58.592+0100: 253621.190: [ParNew: 105381K->371K(118016K), 0.0041528 secs] 126784K->21774K(1035520K), 0.0042236 secs] [Times: user=0.01 sys=0.00, real=0.00 secs] 2017-11-10T15:43:19.414+0100: 260482.011: [GC (Allocation Failure) 2017-11-10T15:43:19.414+0100: 260482.011: [ParNew: 105331K->432K(118016K), 0.0040952 secs] 126734K->21836K(1035520K), 0.0044700 secs] [Times: user=0.01 sys=0.00, real=0.00 secs] Heap par new generation total 118016K, used 98024K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 92% used [0x00000000c0000000, 0x00000000c5f4dcb8, 0x00000000c6680000) from space 13056K, 3% used [0x00000000c7340000, 0x00000000c73ac378, 0x00000000c8000000) to space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) concurrent mark-sweep generation total 917504K, used 21404K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 25539K, capacity 25858K, committed 26236K, reserved 1073152K class space used 2851K, capacity 2939K, committed 2940K, reserved 1048576K ==> /var/log/hadoop/hdfs/hadoop-hdfs-secondarynamenode-slzusxxxxxxxx.sxxx.sxxxx.out.4 <== ulimit -a for user hdfs core file size (blocks, -c) unlimited data seg size (kbytes, -d) unlimited scheduling priority (-e) 0 file size (blocks, -f) unlimited pending signals (-i) 63685 max locked memory (kbytes, -l) 64 max memory size (kbytes, -m) unlimited open files (-n) 128000 pipe size (512 bytes, -p) 8 POSIX message queues (bytes, -q) 819200 real-time priority (-r) 0 stack size (kbytes, -s) 10240 cpu time (seconds, -t) unlimited max user processes (-u) 65536 virtual memory (kbytes, -v) unlimited file locks (-x) unlimited ==> /var/log/hadoop/hdfs/gc.log-201711101733 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7560928k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC 2017-11-10T17:33:02.701+0100: 1.547: [GC (Allocation Failure) 2017-11-10T17:33:02.701+0100: 1.547: [ParNew: 104960K->11576K(118016K), 0.0197353 secs] 104960K->11576K(1035520K), 0.0199021 secs] [Times: user=0.03 sys=0.00, real=0.02 secs] Heap par new generation total 118016K, used 115330K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 98% used [0x00000000c0000000, 0x00000000c65527e8, 0x00000000c6680000) from space 13056K, 88% used [0x00000000c7340000, 0x00000000c7e8e098, 0x00000000c8000000) to space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 21224K, capacity 21548K, committed 21884K, reserved 1069056K class space used 2547K, capacity 2652K, committed 2688K, reserved 1048576K ==> /var/log/hadoop/hdfs/gc.log-201711171719 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7254936k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin/kill-secondary-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin/kill-secondary-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin/kill-secondary-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC Heap par new generation total 118016K, used 60937K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 58% used [0x00000000c0000000, 0x00000000c3b82410, 0x00000000c6680000) from space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) to space 13056K, 0% used [0x00000000c7340000, 0x00000000c7340000, 0x00000000c8000000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 9688K, capacity 9914K, committed 10240K, reserved 1058816K class space used 1117K, capacity 1191K, committed 1280K, reserved 1048576K ==> /var/log/hadoop/hdfs/hadoop-hdfs-journalnode-slzusxxxxxxxx.sxxx.sxxxx.log <== at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:212) at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:404) at org.mortbay.io.nio.SelectChannelEndPoint.run(SelectChannelEndPoint.java:410) at org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582) Caused by: java.lang.NullPointerException at org.apache.hadoop.hdfs.qjournal.server.JournalNode.getJournalsStatus(JournalNode.java:267) at sun.reflect.GeneratedMethodAccessor97.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at sun.reflect.misc.Trampoline.invoke(MethodUtil.java:71) at sun.reflect.GeneratedMethodAccessor7.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at sun.reflect.misc.MethodUtil.invoke(MethodUtil.java:275) at com.sun.jmx.mbeanserver.ConvertingMethod.invokeWithOpenReturn(ConvertingMethod.java:193) at com.sun.jmx.mbeanserver.ConvertingMethod.invokeWithOpenReturn(ConvertingMethod.java:175) at com.sun.jmx.mbeanserver.MXBeanIntrospector.invokeM2(MXBeanIntrospector.java:117) at com.sun.jmx.mbeanserver.MXBeanIntrospector.invokeM2(MXBeanIntrospector.java:54) at com.sun.jmx.mbeanserver.MBeanIntrospector.invokeM(MBeanIntrospector.java:237) at com.sun.jmx.mbeanserver.PerInterface.getAttribute(PerInterface.java:83) at com.sun.jmx.mbeanserver.MBeanSupport.getAttribute(MBeanSupport.java:206) at com.sun.jmx.interceptor.DefaultMBeanServerInterceptor.getAttribute(DefaultMBeanServerInterceptor.java:647) ... 31 more 2017-11-20 09:46:12,354 INFO ipc.Server (Server.java:logException(2401)) - IPC Server handler 0 on 8485, call org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocol.getEditLogManifest from 10.1xx.2xx.xx:35181 Call#27092 Retry#0 java.io.IOException: Invalid directory or I/O error occurred for dir: /hadoop/qj/hdfs/journal/CPZUSD5DFS01/current at org.apache.hadoop.fs.FileUtil.listFiles(FileUtil.java:1144) at org.apache.hadoop.hdfs.server.namenode.FileJournalManager.matchEditLogs(FileJournalManager.java:273) at org.apache.hadoop.hdfs.server.namenode.FileJournalManager.getRemoteEditLogs(FileJournalManager.java:204) at org.apache.hadoop.hdfs.qjournal.server.Journal.getEditLogManifest(Journal.java:656) at org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.getEditLogManifest(JournalNodeRpcServer.java:189) at org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolServerSideTranslatorPB.getEditLogManifest(QJournalProtocolServerSideTranslatorPB.java:224) at org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos$QJournalProtocolService$2.callBlockingMethod(QJournalProtocolProtos.java:25431) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:640) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2313) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2309) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1724) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2307) ==> /var/log/hadoop/hdfs/gc.log-201711101815 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7540140k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC 2017-11-10T18:15:11.145+0100: 1.622: [GC (Allocation Failure) 2017-11-10T18:15:11.145+0100: 1.622: [ParNew: 104960K->11592K(118016K), 0.0212852 secs] 104960K->11592K(1035520K), 0.0214318 secs] [Times: user=0.04 sys=0.00, real=0.03 secs] Heap par new generation total 118016K, used 109268K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 93% used [0x00000000c0000000, 0x00000000c5f631c0, 0x00000000c6680000) from space 13056K, 88% used [0x00000000c7340000, 0x00000000c7e921b8, 0x00000000c8000000) to space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 21153K, capacity 21420K, committed 21628K, reserved 1069056K class space used 2534K, capacity 2652K, committed 2688K, reserved 1048576K ==> /var/log/hadoop/hdfs/hadoop-hdfs-namenode-slzusxxxxxxxx.sxxx.sxxxx.out.5 <== ulimit -a for user hdfs core file size (blocks, -c) unlimited data seg size (kbytes, -d) unlimited scheduling priority (-e) 0 file size (blocks, -f) unlimited pending signals (-i) 63685 max locked memory (kbytes, -l) 64 max memory size (kbytes, -m) unlimited open files (-n) 128000 pipe size (512 bytes, -p) 8 POSIX message queues (bytes, -q) 819200 real-time priority (-r) 0 stack size (kbytes, -s) 10240 cpu time (seconds, -t) unlimited max user processes (-u) 65536 virtual memory (kbytes, -v) unlimited file locks (-x) unlimited ==> /var/log/hadoop/hdfs/gc.log-201711101756 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7549060k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC 2017-11-10T17:56:52.817+0100: 1.594: [GC (Allocation Failure) 2017-11-10T17:56:52.817+0100: 1.594: [ParNew: 104960K->11577K(118016K), 0.0155550 secs] 104960K->11577K(1035520K), 0.0156764 secs] [Times: user=0.04 sys=0.00, real=0.02 secs] Heap par new generation total 118016K, used 107491K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 91% used [0x00000000c0000000, 0x00000000c5daa4b0, 0x00000000c6680000) from space 13056K, 88% used [0x00000000c7340000, 0x00000000c7e8e7c0, 0x00000000c8000000) to space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 21170K, capacity 21484K, committed 21884K, reserved 1069056K class space used 2537K, capacity 2652K, committed 2688K, reserved 1048576K ==> /var/log/hadoop/hdfs/hdfs-audit.log <== ==> /var/log/hadoop/hdfs/hadoop-hdfs-secondarynamenode-slzusxxxxxxxx.sxxx.sxxxx.out.3 <== ulimit -a for user hdfs core file size (blocks, -c) unlimited data seg size (kbytes, -d) unlimited scheduling priority (-e) 0 file size (blocks, -f) unlimited pending signals (-i) 63685 max locked memory (kbytes, -l) 64 max memory size (kbytes, -m) unlimited open files (-n) 128000 pipe size (512 bytes, -p) 8 POSIX message queues (bytes, -q) 819200 real-time priority (-r) 0 stack size (kbytes, -s) 10240 cpu time (seconds, -t) unlimited max user processes (-u) 65536 virtual memory (kbytes, -v) unlimited file locks (-x) unlimited ==> /var/log/hadoop/hdfs/hadoop-hdfs-secondarynamenode-slzusxxxxxxxx.sxxx.sxxxx.out.2 <== ulimit -a for user hdfs core file size (blocks, -c) unlimited data seg size (kbytes, -d) unlimited scheduling priority (-e) 0 file size (blocks, -f) unlimited pending signals (-i) 63685 max locked memory (kbytes, -l) 64 max memory size (kbytes, -m) unlimited open files (-n) 128000 pipe size (512 bytes, -p) 8 POSIX message queues (bytes, -q) 819200 real-time priority (-r) 0 stack size (kbytes, -s) 10240 cpu time (seconds, -t) unlimited max user processes (-u) 65536 virtual memory (kbytes, -v) unlimited file locks (-x) unlimited ==> /var/log/hadoop/hdfs/SecurityAuth.audit <== ==> /var/log/hadoop/hdfs/hadoop-hdfs-namenode-slzusxxxxxxxx.sxxx.sxxxx.log <== 2017-11-20 09:46:46,031 INFO namenode.FSNamesystem (FSNamesystem.java:initRetryCache(915)) - Retry cache on namenode is enabled 2017-11-20 09:46:46,031 INFO namenode.FSNamesystem (FSNamesystem.java:initRetryCache(923)) - Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis 2017-11-20 09:46:46,034 INFO util.GSet (LightWeightGSet.java:computeCapacity(354)) - Computing capacity for map NameNodeRetryCache 2017-11-20 09:46:46,034 INFO util.GSet (LightWeightGSet.java:computeCapacity(355)) - VM type = 64-bit 2017-11-20 09:46:46,034 INFO util.GSet (LightWeightGSet.java:computeCapacity(356)) - 0.029999999329447746% max memory 1011.3 MB = 310.7 KB 2017-11-20 09:46:46,035 INFO util.GSet (LightWeightGSet.java:computeCapacity(361)) - capacity = 2^15 = 32768 entries 2017-11-20 09:46:46,050 INFO common.Storage (Storage.java:tryLock(774)) - Lock on /hadoop/nn/namenodeprimary/in_use.lock acquired by nodename 19450@slzusxxxxxxxx.sxxx.sxxxx 2017-11-20 09:46:46,052 WARN namenode.FSNamesystem (FSNamesystem.java:loadFromDisk(692)) - Encountered exception loading fsimage java.io.IOException: NameNode is not formatted. at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:225) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:1015) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:690) at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:688) at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:752) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:992) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:976) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1686) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1754) 2017-11-20 09:46:46,056 INFO mortbay.log (Slf4jLog.java:info(67)) - Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@slzusxxxxxxxx.sxxx.sxxxx:50070 2017-11-20 09:46:46,156 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stop(211)) - Stopping NameNode metrics system... 2017-11-20 09:46:46,157 INFO impl.MetricsSinkAdapter (MetricsSinkAdapter.java:publishMetricsFromQueue(141)) - timeline thread interrupted. 2017-11-20 09:46:46,159 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stop(217)) - NameNode metrics system stopped. 2017-11-20 09:46:46,159 INFO timeline.HadoopTimelineMetricsSink (HadoopTimelineMetricsSink.java:run(416)) - Closing HadoopTimelineMetricSink. Flushing metrics to collector... 2017-11-20 09:46:46,159 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:shutdown(606)) - NameNode metrics system shutdown complete. 2017-11-20 09:46:46,159 ERROR namenode.NameNode (NameNode.java:main(1759)) - Failed to start namenode. java.io.IOException: NameNode is not formatted. at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:225) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:1015) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:690) at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:688) at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:752) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:992) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:976) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1686) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1754) 2017-11-20 09:46:46,161 INFO util.ExitUtil (ExitUtil.java:terminate(124)) - Exiting with status 1 2017-11-20 09:46:46,162 INFO namenode.NameNode (LogAdapter.java:info(47)) - SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down NameNode at slzusxxxxxxxx.sxxx.sxxxx/10.166.86.17 ************************************************************/ ==> /var/log/hadoop/hdfs/gc.log-201711171715 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7240464k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin/kill-secondary-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin/kill-secondary-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin/kill-secondary-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC Heap par new generation total 118016K, used 60977K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 58% used [0x00000000c0000000, 0x00000000c3b8c6d0, 0x00000000c6680000) from space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) to space 13056K, 0% used [0x00000000c7340000, 0x00000000c7340000, 0x00000000c8000000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 9683K, capacity 9914K, committed 10240K, reserved 1058816K class space used 1117K, capacity 1191K, committed 1280K, reserved 1048576K ==> /var/log/hadoop/hdfs/gc.log-201711101735 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7534092k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC 2017-11-10T17:35:48.050+0100: 1.505: [GC (Allocation Failure) 2017-11-10T17:35:48.050+0100: 1.505: [ParNew: 104960K->11564K(118016K), 0.0131959 secs] 104960K->11564K(1035520K), 0.0133228 secs] [Times: user=0.03 sys=0.00, real=0.01 secs] Heap par new generation total 118016K, used 109263K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 93% used [0x00000000c0000000, 0x00000000c5f68b30, 0x00000000c6680000) from space 13056K, 88% used [0x00000000c7340000, 0x00000000c7e8b138, 0x00000000c8000000) to space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 21162K, capacity 21484K, committed 21884K, reserved 1069056K class space used 2535K, capacity 2652K, committed 2688K, reserved 1048576K ==> /var/log/hadoop/hdfs/gc.log-201711171704 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7243676k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin/kill-secondary-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin/kill-secondary-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin/kill-secondary-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC Heap par new generation total 118016K, used 60959K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 58% used [0x00000000c0000000, 0x00000000c3b87de0, 0x00000000c6680000) from space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) to space 13056K, 0% used [0x00000000c7340000, 0x00000000c7340000, 0x00000000c8000000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 9682K, capacity 9898K, committed 10240K, reserved 1058816K class space used 1117K, capacity 1191K, committed 1280K, reserved 1048576K ==> /var/log/hadoop/hdfs/gc.log-201711171638 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7273648k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC 2017-11-17T16:39:01.498+0100: 1.479: [GC (Allocation Failure) 2017-11-17T16:39:01.498+0100: 1.479: [ParNew: 104960K->11565K(118016K), 0.0224702 secs] 104960K->11565K(1035520K), 0.0225916 secs] [Times: user=0.05 sys=0.01, real=0.03 secs] Heap par new generation total 118016K, used 109274K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 93% used [0x00000000c0000000, 0x00000000c5f6b778, 0x00000000c6680000) from space 13056K, 88% used [0x00000000c7340000, 0x00000000c7e8b428, 0x00000000c8000000) to space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 21174K, capacity 21484K, committed 21884K, reserved 1069056K class space used 2534K, capacity 2652K, committed 2688K, reserved 1048576K ==> /var/log/hadoop/hdfs/hadoop-hdfs-journalnode-slzusxxxxxxxx.sxxx.sxxxx.out <== ulimit -a for user hdfs core file size (blocks, -c) unlimited data seg size (kbytes, -d) unlimited scheduling priority (-e) 0 file size (blocks, -f) unlimited pending signals (-i) 63685 max locked memory (kbytes, -l) 64 max memory size (kbytes, -m) unlimited open files (-n) 128000 pipe size (512 bytes, -p) 8 POSIX message queues (bytes, -q) 819200 real-time priority (-r) 0 stack size (kbytes, -s) 10240 cpu time (seconds, -t) unlimited max user processes (-u) 65536 virtual memory (kbytes, -v) unlimited file locks (-x) unlimited ==> /var/log/hadoop/hdfs/gc.log-201711171657 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7239808k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC 2017-11-17T16:57:50.165+0100: 1.532: [GC (Allocation Failure) 2017-11-17T16:57:50.165+0100: 1.532: [ParNew: 104960K->11578K(118016K), 0.0260068 secs] 104960K->11578K(1035520K), 0.0261411 secs] [Times: user=0.07 sys=0.01, real=0.03 secs] Heap par new generation total 118016K, used 111385K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 95% used [0x00000000c0000000, 0x00000000c6177fa8, 0x00000000c6680000) from space 13056K, 88% used [0x00000000c7340000, 0x00000000c7e8e838, 0x00000000c8000000) to space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 21163K, capacity 21484K, committed 21884K, reserved 1069056K class space used 2536K, capacity 2652K, committed 2688K, reserved 1048576K ==> /var/log/hadoop/hdfs/gc.log-201711101758 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7543280k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC 2017-11-10T17:58:54.207+0100: 1.520: [GC (Allocation Failure) 2017-11-10T17:58:54.207+0100: 1.520: [ParNew: 104960K->11594K(118016K), 0.0202573 secs] 104960K->11594K(1035520K), 0.0203992 secs] [Times: user=0.05 sys=0.01, real=0.02 secs] Heap par new generation total 118016K, used 109150K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 92% used [0x00000000c0000000, 0x00000000c5f45268, 0x00000000c6680000) from space 13056K, 88% used [0x00000000c7340000, 0x00000000c7e92878, 0x00000000c8000000) to space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 21168K, capacity 21484K, committed 21884K, reserved 1069056K class space used 2536K, capacity 2652K, committed 2688K, reserved 1048576K ==> /var/log/hadoop/hdfs/gc.log-201711171643 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7269392k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC 2017-11-17T16:43:06.079+0100: 1.520: [GC (Allocation Failure) 2017-11-17T16:43:06.079+0100: 1.520: [ParNew: 104960K->11595K(118016K), 0.0317000 secs] 104960K->11595K(1035520K), 0.0318505 secs] [Times: user=0.08 sys=0.01, real=0.03 secs] Heap par new generation total 118016K, used 111398K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 95% used [0x00000000c0000000, 0x00000000c6176960, 0x00000000c6680000) from space 13056K, 88% used [0x00000000c7340000, 0x00000000c7e92f00, 0x00000000c8000000) to space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 21147K, capacity 21484K, committed 21884K, reserved 1069056K class space used 2533K, capacity 2652K, committed 2688K, reserved 1048576K ==> /var/log/hadoop/hdfs/hadoop-hdfs-namenode-slzusxxxxxxxx.sxxx.sxxxx.out.3 <== ulimit -a for user hdfs core file size (blocks, -c) unlimited data seg size (kbytes, -d) unlimited scheduling priority (-e) 0 file size (blocks, -f) unlimited pending signals (-i) 63685 max locked memory (kbytes, -l) 64 max memory size (kbytes, -m) unlimited open files (-n) 128000 pipe size (512 bytes, -p) 8 POSIX message queues (bytes, -q) 819200 real-time priority (-r) 0 stack size (kbytes, -s) 10240 cpu time (seconds, -t) unlimited max user processes (-u) 65536 virtual memory (kbytes, -v) unlimited file locks (-x) unlimited ==> /var/log/hadoop/hdfs/hadoop-hdfs-namenode-slzusxxxxxxxx.sxxx.sxxxx.out.4 <== ulimit -a for user hdfs core file size (blocks, -c) unlimited data seg size (kbytes, -d) unlimited scheduling priority (-e) 0 file size (blocks, -f) unlimited pending signals (-i) 63685 max locked memory (kbytes, -l) 64 max memory size (kbytes, -m) unlimited open files (-n) 128000 pipe size (512 bytes, -p) 8 POSIX message queues (bytes, -q) 819200 real-time priority (-r) 0 stack size (kbytes, -s) 10240 cpu time (seconds, -t) unlimited max user processes (-u) 65536 virtual memory (kbytes, -v) unlimited file locks (-x) unlimited ==> /var/log/hadoop/hdfs/gc.log-201711101826 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7516196k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC 2017-11-10T18:26:27.457+0100: 1.548: [GC (Allocation Failure) 2017-11-10T18:26:27.458+0100: 1.548: [ParNew: 104960K->11578K(118016K), 0.0134061 secs] 104960K->11578K(1035520K), 0.0135408 secs] [Times: user=0.03 sys=0.01, real=0.02 secs] Heap par new generation total 118016K, used 109285K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 93% used [0x00000000c0000000, 0x00000000c5f6ae68, 0x00000000c6680000) from space 13056K, 88% used [0x00000000c7340000, 0x00000000c7e8e990, 0x00000000c8000000) to space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 21176K, capacity 21484K, committed 21884K, reserved 1069056K class space used 2535K, capacity 2652K, committed 2688K, reserved 1048576K ==> /var/log/hadoop/hdfs/gc.log-201711101734 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7557724k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC 2017-11-10T17:34:57.686+0100: 1.540: [GC (Allocation Failure) 2017-11-10T17:34:57.686+0100: 1.540: [ParNew: 104960K->11583K(118016K), 0.0223429 secs] 104960K->11583K(1035520K), 0.0224659 secs] [Times: user=0.03 sys=0.01, real=0.02 secs] Heap par new generation total 118016K, used 109295K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 93% used [0x00000000c0000000, 0x00000000c5f6bfd0, 0x00000000c6680000) from space 13056K, 88% used [0x00000000c7340000, 0x00000000c7e8fee0, 0x00000000c8000000) to space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 21160K, capacity 21484K, committed 21884K, reserved 1069056K class space used 2535K, capacity 2652K, committed 2688K, reserved 1048576K ==> /var/log/hadoop/hdfs/gc.log-201711200946 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7033820k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC 2017-11-20T09:46:45.394+0100: 1.508: [GC (Allocation Failure) 2017-11-20T09:46:45.394+0100: 1.508: [ParNew: 104960K->11565K(118016K), 0.0357723 secs] 104960K->11565K(1035520K), 0.0359449 secs] [Times: user=0.12 sys=0.01, real=0.04 secs] Heap par new generation total 118016K, used 111364K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 95% used [0x00000000c0000000, 0x00000000c6175eb0, 0x00000000c6680000) from space 13056K, 88% used [0x00000000c7340000, 0x00000000c7e8b4f0, 0x00000000c8000000) to space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 21173K, capacity 21484K, committed 21884K, reserved 1069056K class space used 2535K, capacity 2652K, committed 2688K, reserved 1048576K ==> /var/log/hadoop/hdfs/hadoop-hdfs-namenode-slzusxxxxxxxx.sxxx.sxxxx.out.1 <== ulimit -a for user hdfs core file size (blocks, -c) unlimited data seg size (kbytes, -d) unlimited scheduling priority (-e) 0 file size (blocks, -f) unlimited pending signals (-i) 63685 max locked memory (kbytes, -l) 64 max memory size (kbytes, -m) unlimited open files (-n) 128000 pipe size (512 bytes, -p) 8 POSIX message queues (bytes, -q) 819200 real-time priority (-r) 0 stack size (kbytes, -s) 10240 cpu time (seconds, -t) unlimited max user processes (-u) 65536 virtual memory (kbytes, -v) unlimited file locks (-x) unlimited ==> /var/log/hadoop/hdfs/hadoop-hdfs-namenode-slzusxxxxxxxx.sxxx.sxxxx.out <== ulimit -a for user hdfs core file size (blocks, -c) unlimited data seg size (kbytes, -d) unlimited scheduling priority (-e) 0 file size (blocks, -f) unlimited pending signals (-i) 63685 max locked memory (kbytes, -l) 64 max memory size (kbytes, -m) unlimited open files (-n) 128000 pipe size (512 bytes, -p) 8 POSIX message queues (bytes, -q) 819200 real-time priority (-r) 0 stack size (kbytes, -s) 10240 cpu time (seconds, -t) unlimited max user processes (-u) 65536 virtual memory (kbytes, -v) unlimited file locks (-x) unlimited ==> /var/log/hadoop/hdfs/hadoop-hdfs-secondarynamenode-slzusxxxxxxxx.sxxx.sxxxx.out <== ulimit -a for user hdfs core file size (blocks, -c) unlimited data seg size (kbytes, -d) unlimited scheduling priority (-e) 0 file size (blocks, -f) unlimited pending signals (-i) 63685 max locked memory (kbytes, -l) 64 max memory size (kbytes, -m) unlimited open files (-n) 128000 pipe size (512 bytes, -p) 8 POSIX message queues (bytes, -q) 819200 real-time priority (-r) 0 stack size (kbytes, -s) 10240 cpu time (seconds, -t) unlimited max user processes (-u) 65536 virtual memory (kbytes, -v) unlimited file locks (-x) unlimited ==> /var/log/hadoop/hdfs/gc.log-201711101732 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7553800k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC 2017-11-10T17:32:50.952+0100: 1.484: [GC (Allocation Failure) 2017-11-10T17:32:50.952+0100: 1.484: [ParNew: 104960K->10155K(118016K), 0.0309917 secs] 104960K->10155K(1035520K), 0.0311515 secs] [Times: user=0.08 sys=0.01, real=0.03 secs] Heap par new generation total 118016K, used 93776K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 79% used [0x00000000c0000000, 0x00000000c51a9500, 0x00000000c6680000) from space 13056K, 77% used [0x00000000c7340000, 0x00000000c7d2aca8, 0x00000000c8000000) to space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 18357K, capacity 18620K, committed 18944K, reserved 1067008K class space used 2142K, capacity 2220K, committed 2304K, reserved 1048576K ==> /var/log/hadoop/hdfs/hadoop-hdfs-namenode-slzusxxxxxxxx.sxxx.sxxxx.out.2 <== ulimit -a for user hdfs core file size (blocks, -c) unlimited data seg size (kbytes, -d) unlimited scheduling priority (-e) 0 file size (blocks, -f) unlimited pending signals (-i) 63685 max locked memory (kbytes, -l) 64 max memory size (kbytes, -m) unlimited open files (-n) 128000 pipe size (512 bytes, -p) 8 POSIX message queues (bytes, -q) 819200 real-time priority (-r) 0 stack size (kbytes, -s) 10240 cpu time (seconds, -t) unlimited max user processes (-u) 65536 virtual memory (kbytes, -v) unlimited file locks (-x) unlimited ==> /var/log/hadoop/hdfs/gc.log-201711101757 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7538020k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC 2017-11-10T17:57:34.760+0100: 1.553: [GC (Allocation Failure) 2017-11-10T17:57:34.760+0100: 1.553: [ParNew: 104960K->11569K(118016K), 0.0448597 secs] 104960K->11569K(1035520K), 0.0449886 secs] [Times: user=0.15 sys=0.01, real=0.04 secs] Heap par new generation total 118016K, used 109058K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 92% used [0x00000000c0000000, 0x00000000c5f34248, 0x00000000c6680000) from space 13056K, 88% used [0x00000000c7340000, 0x00000000c7e8c5e0, 0x00000000c8000000) to space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 21168K, capacity 21484K, committed 21884K, reserved 1069056K class space used 2535K, capacity 2652K, committed 2688K, reserved 1048576K ==> /var/log/hadoop/hdfs/hadoop-hdfs-secondarynamenode-slzusxxxxxxxx.sxxx.sxxxx.out.5 <== ulimit -a for user hdfs core file size (blocks, -c) unlimited data seg size (kbytes, -d) unlimited scheduling priority (-e) 0 file size (blocks, -f) unlimited pending signals (-i) 63685 max locked memory (kbytes, -l) 64 max memory size (kbytes, -m) unlimited open files (-n) 128000 pipe size (512 bytes, -p) 8 POSIX message queues (bytes, -q) 819200 real-time priority (-r) 0 stack size (kbytes, -s) 10240 cpu time (seconds, -t) unlimited max user processes (-u) 65536 virtual memory (kbytes, -v) unlimited file locks (-x) unlimited ==> /var/log/hadoop/hdfs/hadoop-hdfs-secondarynamenode-slzusxxxxxxxx.sxxx.sxxxx.out.1 <== ulimit -a for user hdfs core file size (blocks, -c) unlimited data seg size (kbytes, -d) unlimited scheduling priority (-e) 0 file size (blocks, -f) unlimited pending signals (-i) 63685 max locked memory (kbytes, -l) 64 max memory size (kbytes, -m) unlimited open files (-n) 128000 pipe size (512 bytes, -p) 8 POSIX message queues (bytes, -q) 819200 real-time priority (-r) 0 stack size (kbytes, -s) 10240 cpu time (seconds, -t) unlimited max user processes (-u) 65536 virtual memory (kbytes, -v) unlimited file locks (-x) unlimited ==> /var/log/hadoop/hdfs/gc.log-201711171635 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7248752k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC 2017-11-17T16:35:54.358+0100: 1.544: [GC (Allocation Failure) 2017-11-17T16:35:54.358+0100: 1.544: [ParNew: 104960K->11585K(118016K), 0.0291982 secs] 104960K->11585K(1035520K), 0.0293301 secs] [Times: user=0.06 sys=0.01, real=0.03 secs] Heap par new generation total 118016K, used 109257K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 93% used [0x00000000c0000000, 0x00000000c5f62108, 0x00000000c6680000) from space 13056K, 88% used [0x00000000c7340000, 0x00000000c7e905f0, 0x00000000c8000000) to space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 21169K, capacity 21484K, committed 21884K, reserved 1069056K class space used 2535K, capacity 2652K, committed 2688K, reserved 1048576K ==> /var/log/hadoop/hdfs/gc.log-201711171703 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7266564k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin/kill-secondary-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin/kill-secondary-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin/kill-secondary-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC Heap par new generation total 118016K, used 60977K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 58% used [0x00000000c0000000, 0x00000000c3b8c7d0, 0x00000000c6680000) from space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) to space 13056K, 0% used [0x00000000c7340000, 0x00000000c7340000, 0x00000000c8000000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 9680K, capacity 9898K, committed 10240K, reserved 1058816K class space used 1114K, capacity 1191K, committed 1280K, reserved 1048576K ==> /var/log/hadoop/hdfs/hadoop-hdfs-secondarynamenode-slzusxxxxxxxx.sxxx.sxxxx.log <== STARTUP_MSG: Starting SecondaryNameNode STARTUP_MSG: user = hdfs STARTUP_MSG: host = slzusxxxxxxxx.sxxx.sxxxx/10.166.86.17 STARTUP_MSG: args = [] STARTUP_MSG: version = 2.7.3.2.5.0.0-1245 STARTUP_MSG: classpath = /usr/hdp/current/hadoop-client/conf:/usr/hdp/2.5.0.0-1245/hadoop/lib/asm-3.2.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/slf4j-log4j12-1.7.10.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/httpcore-4.4.4.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/ranger-yarn-plugin-shim-0.6.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-digester-1.8.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-collections-3.2.2.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jsch-0.1.42.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/zookeeper-3.4.6.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-codec-1.4.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/ranger-hdfs-plugin-shim-0.6.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/stax-api-1.0-2.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/joda-time-2.8.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/java-xmlbuilder-0.4.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/azure-storage-2.2.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/json-smart-1.1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-logging-1.1.3.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/avro-1.7.4.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/xz-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-math3-3.1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/gson-2.2.4.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/nimbus-jose-jwt-3.9.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/aws-java-sdk-core-1.10.6.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/netty-3.6.2.Final.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/log4j-1.2.17.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jsp-api-2.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jersey-json-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-compress-1.4.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/httpclient-4.5.2.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/mockito-all-1.8.5.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/servlet-api-2.5.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jaxb-impl-2.2.3-1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-io-2.4.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/hamcrest-core-1.3.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/xmlenc-0.52.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-beanutils-core-1.8.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/api-util-1.0.0-M20.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/curator-recipes-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-configuration-1.6.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-net-3.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jettison-1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/aws-java-sdk-kms-1.10.6.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/slf4j-api-1.7.10.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jackson-jaxrs-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jsr305-3.0.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jersey-server-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/curator-framework-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jcip-annotations-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jackson-annotations-2.2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/junit-4.11.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/guava-11.0.2.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-cli-1.2.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/curator-client-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/ojdbc6.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/paranamer-2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jersey-core-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-lang-2.6.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jaxb-api-2.2.2.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-beanutils-1.7.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/activation-1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jackson-xc-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/ranger-plugin-classloader-0.6.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jets3t-0.9.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jackson-databind-2.2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jackson-core-2.2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/aws-java-sdk-s3-1.10.6.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-aws-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-azure.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-nfs-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-common-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-auth-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-nfs.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-azure-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-common.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-auth.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-annotations.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-common-tests.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-common-2.7.3.2.5.0.0-1245-tests.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-annotations-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-aws.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/./:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/asm-3.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/xercesImpl-2.9.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/commons-codec-1.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/okhttp-2.4.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/commons-logging-1.1.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/netty-3.6.2.Final.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/log4j-1.2.17.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/okio-1.4.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/servlet-api-2.5.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/commons-io-2.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/commons-daemon-1.0.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/xmlenc-0.52.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/leveldbjni-all-1.8.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/netty-all-4.0.23.Final.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/jsr305-3.0.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/jersey-server-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/guava-11.0.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/commons-cli-1.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/jersey-core-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/commons-lang-2.6.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/xml-apis-1.3.04.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/.//hadoop-hdfs-2.7.3.2.5.0.0-1245-tests.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/.//hadoop-hdfs-nfs.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/.//hadoop-hdfs.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/.//hadoop-hdfs-tests.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/.//hadoop-hdfs-nfs-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/.//hadoop-hdfs-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/asm-3.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/httpcore-4.4.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/metrics-core-3.0.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-digester-1.8.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-collections-3.2.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jsch-0.1.42.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/zookeeper-3.4.6.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-codec-1.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/stax-api-1.0-2.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/java-xmlbuilder-0.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/json-smart-1.1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-logging-1.1.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/avro-1.7.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/zookeeper-3.4.6.2.5.0.0-1245-tests.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/xz-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jersey-guice-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-math3-3.1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jersey-client-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/gson-2.2.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/objenesis-2.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/nimbus-jose-jwt-3.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/netty-3.6.2.Final.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/log4j-1.2.17.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/fst-2.24.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jsp-api-2.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jersey-json-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/javax.inject-1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-compress-1.4.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/httpclient-4.5.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/servlet-api-2.5.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jaxb-impl-2.2.3-1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-io-2.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/xmlenc-0.52.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-beanutils-core-1.8.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/guice-servlet-3.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/api-util-1.0.0-M20.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/curator-recipes-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/javassist-3.18.1-GA.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-configuration-1.6.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-net-3.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jettison-1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jackson-jaxrs-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/leveldbjni-all-1.8.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jsr305-3.0.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jersey-server-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/curator-framework-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jcip-annotations-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jackson-annotations-2.2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/aopalliance-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/guava-11.0.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-cli-1.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/curator-client-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/paranamer-2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jersey-core-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-lang-2.6.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jaxb-api-2.2.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/guice-3.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-beanutils-1.7.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/activation-1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jackson-xc-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jets3t-0.9.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jackson-databind-2.2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jackson-core-2.2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-nodemanager-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-nodemanager.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-api-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-applicationhistoryservice.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-sharedcachemanager.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-client.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-resourcemanager.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-common-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-api.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-common.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-web-proxy.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-client-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-registry.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-sharedcachemanager-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-applicationhistoryservice-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-timeline-pluginstorage-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-web-proxy-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-applications-distributedshell-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-applications-distributedshell.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-tests-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-common-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-registry-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-common.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-tests.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-applications-unmanaged-am-launcher.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-applications-unmanaged-am-launcher-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-resourcemanager-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-timeline-pluginstorage.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/asm-3.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/avro-1.7.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/xz-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/jersey-guice-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/netty-3.6.2.Final.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/log4j-1.2.17.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/javax.inject-1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/commons-compress-1.4.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/commons-io-2.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/hamcrest-core-1.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/guice-servlet-3.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/leveldbjni-all-1.8.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/jersey-server-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/junit-4.11.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/aopalliance-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/paranamer-2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/jersey-core-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/guice-3.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//asm-3.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-hs-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-datajoin.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient-tests.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-app.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jackson-core-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-gridmix.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-hs-plugins-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//httpcore-4.4.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-sls-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//metrics-core-3.0.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-digester-1.8.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-collections-3.2.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jsch-0.1.42.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-examples-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//zookeeper-3.4.6.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-core.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-codec-1.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//stax-api-1.0-2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-distcp-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-common.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-shuffle-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//java-xmlbuilder-0.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//json-smart-1.1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-logging-1.1.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//avro-1.7.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//xz-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-auth-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-math3-3.1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//gson-2.2.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//protobuf-java-2.5.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//nimbus-jose-jwt-3.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-app-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//netty-3.6.2.Final.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//log4j-1.2.17.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-lang3-3.3.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jsp-api-2.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-ant-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-common-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jersey-json-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//htrace-core-3.1.0-incubating.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-auth.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-compress-1.4.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-openstack-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//httpclient-4.5.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//mockito-all-1.8.5.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//servlet-api-2.5.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-rumen-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jaxb-impl-2.2.3-1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-io-2.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-datajoin-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hamcrest-core-1.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//xmlenc-0.52.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-beanutils-core-1.8.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jetty-util-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//api-util-1.0.0-M20.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//curator-recipes-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-rumen.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-archives-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-shuffle.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-configuration-1.6.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-net-3.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-streaming.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-streaming-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient-2.7.3.2.5.0.0-1245-tests.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jettison-1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jackson-jaxrs-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-hs-plugins.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-httpclient-3.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jsr305-3.0.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jersey-server-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-gridmix-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//curator-framework-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jcip-annotations-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-openstack.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-distcp.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//snappy-java-1.0.4.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-ant.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jetty-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//junit-4.11.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//guava-11.0.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-cli-1.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-sls.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//curator-client-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//paranamer-2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jersey-core-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-lang-2.6.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-examples.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jaxb-api-2.2.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-extras-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-beanutils-1.7.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//activation-1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jackson-xc-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-archives.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-hs.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jets3t-0.9.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-core-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-extras.jar::mysql-connector-java-5.1.40.jar:mysql-connector-java.jar:/usr/hdp/2.5.0.0-1245/tez/tez-yarn-timeline-history-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-runtime-library-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-yarn-timeline-history-with-fs-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-api-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-common-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-runtime-internals-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-yarn-timeline-history-with-acls-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-tests-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-examples-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-job-analyzer-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-dag-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-yarn-timeline-cache-plugin-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-history-parser-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-mapreduce-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/metrics-core-3.1.0.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-collections-3.2.2.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-aws-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-codec-1.4.jar:/usr/hdp/2.5.0.0-1245/tez/lib/slf4j-api-1.7.5.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-math3-3.1.1.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jersey-client-1.9.jar:/usr/hdp/2.5.0.0-1245/tez/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-azure-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jsr305-2.0.3.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-yarn-server-timeline-pluginstorage-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-mapreduce-client-common-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jersey-json-1.9.jar:/usr/hdp/2.5.0.0-1245/tez/lib/servlet-api-2.5.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-io-2.4.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-yarn-server-web-proxy-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jettison-1.3.4.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-collections4-4.1.jar:/usr/hdp/2.5.0.0-1245/tez/lib/guava-11.0.2.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-cli-1.2.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-lang-2.6.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-annotations-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-mapreduce-client-core-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/conf:mysql-connector-java-5.1.40.jar:mysql-connector-java.jar:mysql-connector-java-5.1.40.jar:mysql-connector-java.jar:/usr/hdp/2.5.0.0-1245/tez/tez-yarn-timeline-history-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-runtime-library-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-yarn-timeline-history-with-fs-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-api-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-common-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-runtime-internals-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-yarn-timeline-history-with-acls-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-tests-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-examples-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-job-analyzer-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-dag-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-yarn-timeline-cache-plugin-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-history-parser-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-mapreduce-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/metrics-core-3.1.0.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-collections-3.2.2.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-aws-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-codec-1.4.jar:/usr/hdp/2.5.0.0-1245/tez/lib/slf4j-api-1.7.5.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-math3-3.1.1.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jersey-client-1.9.jar:/usr/hdp/2.5.0.0-1245/tez/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-azure-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jsr305-2.0.3.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-yarn-server-timeline-pluginstorage-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-mapreduce-client-common-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jersey-json-1.9.jar:/usr/hdp/2.5.0.0-1245/tez/lib/servlet-api-2.5.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-io-2.4.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-yarn-server-web-proxy-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jettison-1.3.4.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-collections4-4.1.jar:/usr/hdp/2.5.0.0-1245/tez/lib/guava-11.0.2.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-cli-1.2.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-lang-2.6.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-annotations-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-mapreduce-client-core-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/conf STARTUP_MSG: build = git@github.com:hortonworks/hadoop.git -r cb6e514b14fb60e9995e5ad9543315cd404b4e59; compiled by 'jenkins' on 2016-08-26T00:55Z STARTUP_MSG: java = 1.8.0_77 ************************************************************/ 2017-11-17 17:17:23,768 INFO namenode.SecondaryNameNode (LogAdapter.java:info(45)) - registered UNIX signal handlers for [TERM, HUP, INT] 2017-11-17 17:17:24,148 FATAL namenode.SecondaryNameNode (SecondaryNameNode.java:main(673)) - Failed to start secondary namenode java.io.IOException: Cannot use SecondaryNameNode in an HA cluster. The Standby Namenode will perform checkpointing. at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.(SecondaryNameNode.java:187) at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:671) 2017-11-17 17:17:24,151 INFO util.ExitUtil (ExitUtil.java:terminate(124)) - Exiting with status 1 2017-11-17 17:17:24,154 INFO namenode.SecondaryNameNode (LogAdapter.java:info(45)) - SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down SecondaryNameNode at slzusxxxxxxxx.sxxx.sxxxx/10.166.86.17 ************************************************************/ 2017-11-17 17:19:10,351 INFO namenode.SecondaryNameNode (LogAdapter.java:info(45)) - STARTUP_MSG: /************************************************************ STARTUP_MSG: Starting SecondaryNameNode STARTUP_MSG: user = hdfs STARTUP_MSG: host = slzusxxxxxxxx.sxxx.sxxxx/10.166.86.17 STARTUP_MSG: args = [] STARTUP_MSG: version = 2.7.3.2.5.0.0-1245 STARTUP_MSG: classpath = /usr/hdp/current/hadoop-client/conf:/usr/hdp/2.5.0.0-1245/hadoop/lib/asm-3.2.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/slf4j-log4j12-1.7.10.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/httpcore-4.4.4.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/ranger-yarn-plugin-shim-0.6.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-digester-1.8.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-collections-3.2.2.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jsch-0.1.42.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/zookeeper-3.4.6.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-codec-1.4.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/ranger-hdfs-plugin-shim-0.6.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/stax-api-1.0-2.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/joda-time-2.8.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/java-xmlbuilder-0.4.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/azure-storage-2.2.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/json-smart-1.1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-logging-1.1.3.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/avro-1.7.4.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/xz-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-math3-3.1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/gson-2.2.4.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/nimbus-jose-jwt-3.9.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/aws-java-sdk-core-1.10.6.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/netty-3.6.2.Final.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/log4j-1.2.17.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jsp-api-2.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jersey-json-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-compress-1.4.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/httpclient-4.5.2.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/mockito-all-1.8.5.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/servlet-api-2.5.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jaxb-impl-2.2.3-1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-io-2.4.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/hamcrest-core-1.3.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/xmlenc-0.52.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-beanutils-core-1.8.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/api-util-1.0.0-M20.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/curator-recipes-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-configuration-1.6.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-net-3.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jettison-1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/aws-java-sdk-kms-1.10.6.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/slf4j-api-1.7.10.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jackson-jaxrs-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jsr305-3.0.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jersey-server-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/curator-framework-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jcip-annotations-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jackson-annotations-2.2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/junit-4.11.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/guava-11.0.2.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-cli-1.2.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/curator-client-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/ojdbc6.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/paranamer-2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jersey-core-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-lang-2.6.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jaxb-api-2.2.2.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/commons-beanutils-1.7.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/activation-1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jackson-xc-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/ranger-plugin-classloader-0.6.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jets3t-0.9.0.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jackson-databind-2.2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/jackson-core-2.2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop/lib/aws-java-sdk-s3-1.10.6.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-aws-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-azure.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-nfs-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-common-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-auth-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-nfs.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-azure-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-common.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-auth.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-annotations.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-common-tests.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-common-2.7.3.2.5.0.0-1245-tests.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-annotations-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop/.//hadoop-aws.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/./:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/asm-3.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/xercesImpl-2.9.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/commons-codec-1.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/okhttp-2.4.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/commons-logging-1.1.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/netty-3.6.2.Final.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/log4j-1.2.17.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/okio-1.4.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/servlet-api-2.5.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/commons-io-2.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/commons-daemon-1.0.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/xmlenc-0.52.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/leveldbjni-all-1.8.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/netty-all-4.0.23.Final.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/jsr305-3.0.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/jersey-server-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/guava-11.0.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/commons-cli-1.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/jersey-core-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/commons-lang-2.6.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/lib/xml-apis-1.3.04.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/.//hadoop-hdfs-2.7.3.2.5.0.0-1245-tests.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/.//hadoop-hdfs-nfs.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/.//hadoop-hdfs.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/.//hadoop-hdfs-tests.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/.//hadoop-hdfs-nfs-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-hdfs/.//hadoop-hdfs-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/asm-3.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/httpcore-4.4.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/metrics-core-3.0.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-digester-1.8.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-collections-3.2.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jsch-0.1.42.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/zookeeper-3.4.6.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-codec-1.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/stax-api-1.0-2.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/java-xmlbuilder-0.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/json-smart-1.1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-logging-1.1.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/avro-1.7.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/zookeeper-3.4.6.2.5.0.0-1245-tests.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/xz-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jersey-guice-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-math3-3.1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jersey-client-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/gson-2.2.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/objenesis-2.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/nimbus-jose-jwt-3.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/netty-3.6.2.Final.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/log4j-1.2.17.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/fst-2.24.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jsp-api-2.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jersey-json-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/javax.inject-1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-compress-1.4.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/httpclient-4.5.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/servlet-api-2.5.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jaxb-impl-2.2.3-1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-io-2.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/xmlenc-0.52.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-beanutils-core-1.8.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/guice-servlet-3.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/api-util-1.0.0-M20.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/curator-recipes-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/javassist-3.18.1-GA.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-configuration-1.6.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-net-3.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jettison-1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jackson-jaxrs-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/leveldbjni-all-1.8.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jsr305-3.0.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jersey-server-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/curator-framework-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jcip-annotations-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jackson-annotations-2.2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/aopalliance-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/guava-11.0.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-cli-1.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/curator-client-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/paranamer-2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jersey-core-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-lang-2.6.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jaxb-api-2.2.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/guice-3.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/commons-beanutils-1.7.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/activation-1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jackson-xc-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jets3t-0.9.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jackson-databind-2.2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/lib/jackson-core-2.2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-nodemanager-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-nodemanager.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-api-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-applicationhistoryservice.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-sharedcachemanager.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-client.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-resourcemanager.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-common-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-api.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-common.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-web-proxy.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-client-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-registry.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-sharedcachemanager-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-applicationhistoryservice-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-timeline-pluginstorage-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-web-proxy-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-applications-distributedshell-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-applications-distributedshell.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-tests-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-common-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-registry-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-common.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-tests.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-applications-unmanaged-am-launcher.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-applications-unmanaged-am-launcher-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-resourcemanager-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-yarn/.//hadoop-yarn-server-timeline-pluginstorage.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/asm-3.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/avro-1.7.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/xz-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/jersey-guice-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/netty-3.6.2.Final.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/log4j-1.2.17.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/javax.inject-1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/commons-compress-1.4.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/commons-io-2.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/hamcrest-core-1.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/guice-servlet-3.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/leveldbjni-all-1.8.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/jersey-server-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/junit-4.11.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/aopalliance-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/paranamer-2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/jersey-core-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/lib/guice-3.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//asm-3.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-hs-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-datajoin.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient-tests.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-app.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jackson-core-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-gridmix.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-hs-plugins-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//httpcore-4.4.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-sls-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//metrics-core-3.0.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-digester-1.8.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-collections-3.2.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jsch-0.1.42.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-examples-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//zookeeper-3.4.6.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-core.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-codec-1.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//stax-api-1.0-2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-distcp-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-common.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-shuffle-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//java-xmlbuilder-0.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//json-smart-1.1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-logging-1.1.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//avro-1.7.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//xz-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-auth-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-math3-3.1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//gson-2.2.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//protobuf-java-2.5.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//nimbus-jose-jwt-3.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-app-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//netty-3.6.2.Final.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//log4j-1.2.17.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-lang3-3.3.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jsp-api-2.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-ant-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-common-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jersey-json-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//htrace-core-3.1.0-incubating.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-auth.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-compress-1.4.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-openstack-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//httpclient-4.5.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//mockito-all-1.8.5.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//servlet-api-2.5.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-rumen-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jaxb-impl-2.2.3-1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-io-2.4.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-datajoin-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hamcrest-core-1.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//xmlenc-0.52.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-beanutils-core-1.8.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jetty-util-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//api-util-1.0.0-M20.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//curator-recipes-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-rumen.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-archives-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-shuffle.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-configuration-1.6.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-net-3.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-streaming.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-streaming-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient-2.7.3.2.5.0.0-1245-tests.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jettison-1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jackson-jaxrs-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-hs-plugins.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-httpclient-3.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jsr305-3.0.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jersey-server-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-gridmix-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//curator-framework-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jcip-annotations-1.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-openstack.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-distcp.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//snappy-java-1.0.4.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-ant.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jetty-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//junit-4.11.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//guava-11.0.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-cli-1.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-sls.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//curator-client-2.7.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//paranamer-2.3.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jersey-core-1.9.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-lang-2.6.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-examples.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jaxb-api-2.2.2.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-extras-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//commons-beanutils-1.7.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//activation-1.1.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jackson-xc-1.9.13.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-archives.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-hs.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//jets3t-0.9.0.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-core-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient.jar:/usr/hdp/2.5.0.0-1245/hadoop-mapreduce/.//hadoop-extras.jar::mysql-connector-java-5.1.40.jar:mysql-connector-java.jar:/usr/hdp/2.5.0.0-1245/tez/tez-yarn-timeline-history-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-runtime-library-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-yarn-timeline-history-with-fs-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-api-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-common-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-runtime-internals-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-yarn-timeline-history-with-acls-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-tests-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-examples-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-job-analyzer-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-dag-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-yarn-timeline-cache-plugin-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-history-parser-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-mapreduce-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/metrics-core-3.1.0.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-collections-3.2.2.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-aws-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-codec-1.4.jar:/usr/hdp/2.5.0.0-1245/tez/lib/slf4j-api-1.7.5.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-math3-3.1.1.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jersey-client-1.9.jar:/usr/hdp/2.5.0.0-1245/tez/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-azure-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jsr305-2.0.3.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-yarn-server-timeline-pluginstorage-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-mapreduce-client-common-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jersey-json-1.9.jar:/usr/hdp/2.5.0.0-1245/tez/lib/servlet-api-2.5.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-io-2.4.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-yarn-server-web-proxy-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jettison-1.3.4.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-collections4-4.1.jar:/usr/hdp/2.5.0.0-1245/tez/lib/guava-11.0.2.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-cli-1.2.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-lang-2.6.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-annotations-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-mapreduce-client-core-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/conf:mysql-connector-java-5.1.40.jar:mysql-connector-java.jar:mysql-connector-java-5.1.40.jar:mysql-connector-java.jar:/usr/hdp/2.5.0.0-1245/tez/tez-yarn-timeline-history-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-runtime-library-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-yarn-timeline-history-with-fs-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-api-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-common-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-runtime-internals-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-yarn-timeline-history-with-acls-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-tests-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-examples-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-job-analyzer-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-dag-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-yarn-timeline-cache-plugin-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-history-parser-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/tez-mapreduce-0.7.0.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/metrics-core-3.1.0.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-collections-3.2.2.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-aws-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-codec-1.4.jar:/usr/hdp/2.5.0.0-1245/tez/lib/slf4j-api-1.7.5.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-math3-3.1.1.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jersey-client-1.9.jar:/usr/hdp/2.5.0.0-1245/tez/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-azure-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jsr305-2.0.3.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-yarn-server-timeline-pluginstorage-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-mapreduce-client-common-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jersey-json-1.9.jar:/usr/hdp/2.5.0.0-1245/tez/lib/servlet-api-2.5.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-io-2.4.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-yarn-server-web-proxy-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jettison-1.3.4.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/tez/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-collections4-4.1.jar:/usr/hdp/2.5.0.0-1245/tez/lib/guava-11.0.2.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-cli-1.2.jar:/usr/hdp/2.5.0.0-1245/tez/lib/commons-lang-2.6.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-annotations-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/lib/hadoop-mapreduce-client-core-2.7.3.2.5.0.0-1245.jar:/usr/hdp/2.5.0.0-1245/tez/conf STARTUP_MSG: build = git@github.com:hortonworks/hadoop.git -r cb6e514b14fb60e9995e5ad9543315cd404b4e59; compiled by 'jenkins' on 2016-08-26T00:55Z STARTUP_MSG: java = 1.8.0_77 ************************************************************/ 2017-11-17 17:19:10,364 INFO namenode.SecondaryNameNode (LogAdapter.java:info(45)) - registered UNIX signal handlers for [TERM, HUP, INT] 2017-11-17 17:19:10,770 FATAL namenode.SecondaryNameNode (SecondaryNameNode.java:main(673)) - Failed to start secondary namenode java.io.IOException: Cannot use SecondaryNameNode in an HA cluster. The Standby Namenode will perform checkpointing. at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.(SecondaryNameNode.java:187) at org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:671) 2017-11-17 17:19:10,773 INFO util.ExitUtil (ExitUtil.java:terminate(124)) - Exiting with status 1 2017-11-17 17:19:10,776 INFO namenode.SecondaryNameNode (LogAdapter.java:info(45)) - SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down SecondaryNameNode at slzusxxxxxxxx.sxxx.sxxxx/10.166.86.17 ************************************************************/ ==> /var/log/hadoop/hdfs/gc.log-201711171717 <== Java HotSpot(TM) 64-Bit Server VM (25.77-b03) for linux-amd64 JRE (1.8.0_77-b03), built on Mar 20 2016 22:00:46 by "java_re" with gcc 4.3.0 20080428 (Red Hat 4.3.0-8) Memory: 4k page, physical 16334056k(7234224k free), swap 8065020k(8065020k free) CommandLine flags: -XX:CMSInitiatingOccupancyFraction=70 -XX:ErrorFile=/var/log/hadoop/hdfs/hs_err_pid%p.log -XX:InitialHeapSize=1073741824 -XX:MaxHeapSize=1073741824 -XX:MaxNewSize=134217728 -XX:MaxTenuringThreshold=6 -XX:NewSize=134217728 -XX:OldPLABSize=16 -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin/kill-secondary-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin/kill-secondary-name-node" -XX:OnOutOfMemoryError="/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin/kill-secondary-name-node" -XX:ParallelGCThreads=8 -XX:+PrintGC -XX:+PrintGCDateStamps -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC Heap par new generation total 118016K, used 60929K [0x00000000c0000000, 0x00000000c8000000, 0x00000000c8000000) eden space 104960K, 58% used [0x00000000c0000000, 0x00000000c3b806b0, 0x00000000c6680000) from space 13056K, 0% used [0x00000000c6680000, 0x00000000c6680000, 0x00000000c7340000) to space 13056K, 0% used [0x00000000c7340000, 0x00000000c7340000, 0x00000000c8000000) concurrent mark-sweep generation total 917504K, used 0K [0x00000000c8000000, 0x0000000100000000, 0x0000000100000000) Metaspace used 9692K, capacity 9898K, committed 10240K, reserved 1058816K class space used 1117K, capacity 1191K, committed 1280K, reserved 1048576K Command failed after 1 tries