************************************************************/ 2018-04-04 22:30:10,050 INFO namenode.NameNode (LogAdapter.java:info(47)) - STARTUP_MSG: /************************************************************ STARTUP_MSG: Starting NameNode STARTUP_MSG: user = root STARTUP_MSG: host = rjio-UatCluster-NameNode1/10.20.1.180 STARTUP_MSG: args = [] STARTUP_MSG: version = 2.7.3.2.6.3.0-235 STARTUP_MSG: classpath = /usr/hdp/2.6.3.0-235/hadoop/conf:/usr/hdp/2.6.3.0-235/hadoop/lib/ojdbc6.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jackson-core-2.2.3.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/ranger-hdfs-plugin-shim-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/ranger-plugin-classloader-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jackson-databind-2.2.3.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/ranger-yarn-plugin-shim-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/stax-api-1.0-2.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/activation-1.1.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jcip-annotations-1.0.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jackson-jaxrs-1.9.13.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jersey-core-1.9.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jersey-json-1.9.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/api-util-1.0.0-M20.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/asm-3.2.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/xmlenc-0.52.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/avro-1.7.4.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jersey-server-1.9.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/aws-java-sdk-core-1.10.6.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jets3t-0.9.0.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/aws-java-sdk-kms-1.10.6.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jsp-api-2.1.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/aws-java-sdk-s3-1.10.6.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/azure-keyvault-core-0.8.0.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jsr305-3.0.0.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/azure-storage-5.4.0.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/junit-4.11.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/commons-beanutils-1.7.0.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jackson-xc-1.9.13.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/commons-beanutils-core-1.8.0.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/xz-1.0.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/commons-cli-1.2.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/log4j-1.2.17.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/commons-codec-1.4.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/java-xmlbuilder-0.4.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/commons-collections-3.2.2.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/mockito-all-1.8.5.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/commons-compress-1.4.1.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/commons-configuration-1.6.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/nimbus-jose-jwt-3.9.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/commons-digester-1.8.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/commons-io-2.4.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/commons-lang-2.6.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/netty-3.6.2.Final.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/commons-lang3-3.4.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/paranamer-2.3.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/commons-logging-1.1.3.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/commons-math3-3.1.1.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/commons-net-3.1.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/servlet-api-2.5.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/curator-client-2.7.1.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/slf4j-api-1.7.10.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/curator-framework-2.7.1.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/curator-recipes-2.7.1.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/gson-2.2.4.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/guava-11.0.2.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/slf4j-log4j12-1.7.10.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/hamcrest-core-1.3.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jaxb-api-2.2.2.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/httpclient-4.5.2.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/httpcore-4.4.4.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jaxb-impl-2.2.3-1.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jackson-annotations-2.2.3.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jettison-1.1.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/json-smart-1.1.1.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jetty-sslengine-6.1.26.hwx.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/zookeeper-3.4.6.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/joda-time-2.9.4.jar:/usr/hdp/2.6.3.0-235/hadoop/lib/jsch-0.1.54.jar:/usr/hdp/2.6.3.0-235/hadoop/.//azure-data-lake-store-sdk-2.1.4.jar:/usr/hdp/2.6.3.0-235/hadoop/.//hadoop-annotations-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop/.//hadoop-annotations.jar:/usr/hdp/2.6.3.0-235/hadoop/.//hadoop-auth-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop/.//hadoop-auth.jar:/usr/hdp/2.6.3.0-235/hadoop/.//hadoop-aws-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop/.//hadoop-aws.jar:/usr/hdp/2.6.3.0-235/hadoop/.//hadoop-azure-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop/.//hadoop-azure-datalake-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop/.//hadoop-azure-datalake.jar:/usr/hdp/2.6.3.0-235/hadoop/.//hadoop-azure.jar:/usr/hdp/2.6.3.0-235/hadoop/.//hadoop-common-2.7.3.2.6.3.0-235-tests.jar:/usr/hdp/2.6.3.0-235/hadoop/.//hadoop-common-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop/.//hadoop-common-tests.jar:/usr/hdp/2.6.3.0-235/hadoop/.//hadoop-common.jar:/usr/hdp/2.6.3.0-235/hadoop/.//hadoop-nfs-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop/.//hadoop-nfs.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/./:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/asm-3.2.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/commons-cli-1.2.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/commons-codec-1.4.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/commons-daemon-1.0.13.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/commons-io-2.4.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/commons-lang-2.6.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/commons-logging-1.1.3.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/guava-11.0.2.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/jackson-annotations-2.2.3.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/jackson-core-2.2.3.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/jackson-databind-2.2.3.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/jersey-core-1.9.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/jersey-server-1.9.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/jsr305-3.0.0.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/leveldbjni-all-1.8.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/log4j-1.2.17.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/netty-3.6.2.Final.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/netty-all-4.0.52.Final.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/okhttp-2.4.0.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/okio-1.4.0.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/servlet-api-2.5.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/xercesImpl-2.9.1.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/xml-apis-1.3.04.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/lib/xmlenc-0.52.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/.//hadoop-hdfs-2.7.3.2.6.3.0-235-tests.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/.//hadoop-hdfs-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/.//hadoop-hdfs-nfs-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/.//hadoop-hdfs-nfs.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/.//hadoop-hdfs-tests.jar:/usr/hdp/2.6.3.0-235/hadoop-hdfs/.//hadoop-hdfs.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/activation-1.1.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/aopalliance-1.0.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jersey-guice-1.9.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/javassist-3.18.1-GA.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jersey-json-1.9.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jersey-server-1.9.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/api-util-1.0.0-M20.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/asm-3.2.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/avro-1.7.4.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/javax.inject-1.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/azure-keyvault-core-0.8.0.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jets3t-0.9.0.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/azure-storage-5.4.0.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/log4j-1.2.17.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/commons-beanutils-1.7.0.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jaxb-api-2.2.2.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/commons-beanutils-core-1.8.0.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/commons-cli-1.2.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/leveldbjni-all-1.8.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/commons-codec-1.4.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jaxb-impl-2.2.3-1.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/commons-collections-3.2.2.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/metrics-core-3.0.1.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/commons-compress-1.4.1.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/json-smart-1.1.1.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/commons-configuration-1.6.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/netty-3.6.2.Final.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/commons-digester-1.8.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/commons-io-2.4.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/commons-lang-2.6.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/nimbus-jose-jwt-3.9.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/commons-lang3-3.4.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/objenesis-2.1.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/commons-logging-1.1.3.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/paranamer-2.3.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/commons-math3-3.1.1.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/commons-net-3.1.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/curator-client-2.7.1.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/servlet-api-2.5.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/curator-framework-2.7.1.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/curator-recipes-2.7.1.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/fst-2.24.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/gson-2.2.4.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/guava-11.0.2.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/guice-3.0.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/stax-api-1.0-2.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/guice-servlet-3.0.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jcip-annotations-1.0.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/httpclient-4.5.2.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/httpcore-4.4.4.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jersey-client-1.9.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jackson-annotations-2.2.3.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/xmlenc-0.52.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jackson-core-2.2.3.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/xz-1.0.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/zookeeper-3.4.6.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jackson-databind-2.2.3.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/zookeeper-3.4.6.2.6.3.0-235-tests.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jackson-jaxrs-1.9.13.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jersey-core-1.9.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jackson-xc-1.9.13.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/java-xmlbuilder-0.4.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jettison-1.1.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jsp-api-2.1.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jetty-sslengine-6.1.26.hwx.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jsch-0.1.54.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/lib/jsr305-3.0.0.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-api-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-api.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-applications-distributedshell-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-applications-distributedshell.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-applications-unmanaged-am-launcher-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-applications-unmanaged-am-launcher.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-client-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-client.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-common-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-common.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-registry-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-registry.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-server-applicationhistoryservice-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-server-applicationhistoryservice.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-server-common-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-server-common.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-server-nodemanager-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-server-nodemanager.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-server-resourcemanager-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-server-resourcemanager.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-server-sharedcachemanager-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-server-sharedcachemanager.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-server-tests-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-server-tests.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-server-timeline-pluginstorage-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-server-timeline-pluginstorage.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-server-web-proxy-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-yarn/.//hadoop-yarn-server-web-proxy.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/aopalliance-1.0.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/asm-3.2.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/avro-1.7.4.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/commons-compress-1.4.1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/commons-io-2.4.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/guice-3.0.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/guice-servlet-3.0.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/hamcrest-core-1.3.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/javax.inject-1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/jersey-core-1.9.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/jersey-guice-1.9.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/jersey-server-1.9.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/junit-4.11.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/leveldbjni-all-1.8.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/log4j-1.2.17.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/netty-3.6.2.Final.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/paranamer-2.3.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/lib/xz-1.0.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jaxb-api-2.2.2.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//activation-1.1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-rumen.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-client-core-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-sls-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-sls.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//api-util-1.0.0-M20.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//log4j-1.2.17.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//asm-3.2.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jaxb-impl-2.2.3-1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//avro-1.7.4.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-client-hs-plugins.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//azure-keyvault-core-0.8.0.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hamcrest-core-1.3.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//commons-beanutils-1.7.0.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//commons-beanutils-core-1.8.0.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jcip-annotations-1.0.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//commons-cli-1.2.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-streaming.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//commons-codec-1.4.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-client-hs.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//commons-collections-3.2.2.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//httpcore-4.4.4.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//commons-compress-1.4.1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient-tests.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//commons-configuration-1.6.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//htrace-core-3.1.0-incubating.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//commons-digester-1.8.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//httpclient-4.5.2.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//commons-httpclient-3.1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jersey-core-1.9.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//commons-io-2.4.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jersey-json-1.9.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//commons-lang-2.6.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jackson-core-asl-1.9.13.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//commons-lang3-3.4.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jackson-jaxrs-1.9.13.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//commons-logging-1.1.3.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//java-xmlbuilder-0.4.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//commons-math3-3.1.1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jersey-server-1.9.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//commons-net-3.1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//curator-client-2.7.1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jackson-xc-1.9.13.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//curator-framework-2.7.1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jetty-sslengine-6.1.26.hwx.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//curator-recipes-2.7.1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jets3t-0.9.0.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//gson-2.2.4.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jettison-1.1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//guava-11.0.2.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-examples.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-ant-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jetty-6.1.26.hwx.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-ant.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-client-core.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-archives-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jetty-util-6.1.26.hwx.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-archives.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-client-shuffle.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-auth-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//json-smart-1.1.1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-auth.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-client-hs-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-datajoin-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jsch-0.1.54.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-datajoin.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-streaming-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-distcp-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jsp-api-2.1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-distcp.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-openstack-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-extras-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//jsr305-3.0.0.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-extras.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-openstack.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-gridmix-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//junit-4.11.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-gridmix.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-client-hs-plugins-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-client-app-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-rumen-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-client-app.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-client-common.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-client-common-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//xz-1.0.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient-2.7.3.2.6.3.0-235-tests.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-client-shuffle-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//hadoop-mapreduce-examples-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//metrics-core-3.0.1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//mockito-all-1.8.5.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//netty-3.6.2.Final.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//nimbus-jose-jwt-3.9.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//okhttp-2.4.0.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//okio-1.4.0.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//paranamer-2.3.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//protobuf-java-2.5.0.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//servlet-api-2.5.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//snappy-java-1.0.4.1.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//stax-api-1.0-2.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//xmlenc-0.52.jar:/usr/hdp/2.6.3.0-235/hadoop-mapreduce/.//zookeeper-3.4.6.2.6.3.0-235.jar::mysql-connector-java.jar:/usr/hdp/2.6.3.0-235/tez/tez-api-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-common-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-dag-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-examples-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-history-parser-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-job-analyzer-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-mapreduce-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-runtime-internals-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-runtime-library-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-tests-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-yarn-timeline-cache-plugin-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-yarn-timeline-history-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-yarn-timeline-history-with-acls-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-yarn-timeline-history-with-fs-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/azure-data-lake-store-sdk-2.1.4.jar:/usr/hdp/2.6.3.0-235/tez/lib/commons-cli-1.2.jar:/usr/hdp/2.6.3.0-235/tez/lib/commons-codec-1.4.jar:/usr/hdp/2.6.3.0-235/tez/lib/commons-collections-3.2.2.jar:/usr/hdp/2.6.3.0-235/tez/lib/commons-collections4-4.1.jar:/usr/hdp/2.6.3.0-235/tez/lib/commons-io-2.4.jar:/usr/hdp/2.6.3.0-235/tez/lib/commons-lang-2.6.jar:/usr/hdp/2.6.3.0-235/tez/lib/commons-math3-3.1.1.jar:/usr/hdp/2.6.3.0-235/tez/lib/guava-11.0.2.jar:/usr/hdp/2.6.3.0-235/tez/lib/hadoop-annotations-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/hadoop-aws-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/hadoop-azure-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/hadoop-azure-datalake-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/hadoop-mapreduce-client-common-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/hadoop-mapreduce-client-core-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/hadoop-yarn-server-timeline-pluginstorage-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/hadoop-yarn-server-web-proxy-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/jersey-client-1.9.jar:/usr/hdp/2.6.3.0-235/tez/lib/jersey-json-1.9.jar:/usr/hdp/2.6.3.0-235/tez/lib/jettison-1.3.4.jar:/usr/hdp/2.6.3.0-235/tez/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.6.3.0-235/tez/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.6.3.0-235/tez/lib/jsr305-2.0.3.jar:/usr/hdp/2.6.3.0-235/tez/lib/metrics-core-3.1.0.jar:/usr/hdp/2.6.3.0-235/tez/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.6.3.0-235/tez/lib/servlet-api-2.5.jar:/usr/hdp/2.6.3.0-235/tez/lib/slf4j-api-1.7.5.jar:/usr/hdp/2.6.3.0-235/tez/conf:mysql-connector-java.jar:mysql-connector-java.jar:/usr/hdp/2.6.3.0-235/tez/tez-api-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-common-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-dag-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-examples-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-history-parser-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-job-analyzer-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-mapreduce-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-runtime-internals-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-runtime-library-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-tests-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-yarn-timeline-cache-plugin-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-yarn-timeline-history-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-yarn-timeline-history-with-acls-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/tez-yarn-timeline-history-with-fs-0.7.0.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/azure-data-lake-store-sdk-2.1.4.jar:/usr/hdp/2.6.3.0-235/tez/lib/commons-cli-1.2.jar:/usr/hdp/2.6.3.0-235/tez/lib/commons-codec-1.4.jar:/usr/hdp/2.6.3.0-235/tez/lib/commons-collections-3.2.2.jar:/usr/hdp/2.6.3.0-235/tez/lib/commons-collections4-4.1.jar:/usr/hdp/2.6.3.0-235/tez/lib/commons-io-2.4.jar:/usr/hdp/2.6.3.0-235/tez/lib/commons-lang-2.6.jar:/usr/hdp/2.6.3.0-235/tez/lib/commons-math3-3.1.1.jar:/usr/hdp/2.6.3.0-235/tez/lib/guava-11.0.2.jar:/usr/hdp/2.6.3.0-235/tez/lib/hadoop-annotations-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/hadoop-aws-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/hadoop-azure-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/hadoop-azure-datalake-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/hadoop-mapreduce-client-common-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/hadoop-mapreduce-client-core-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/hadoop-yarn-server-timeline-pluginstorage-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/hadoop-yarn-server-web-proxy-2.7.3.2.6.3.0-235.jar:/usr/hdp/2.6.3.0-235/tez/lib/jersey-client-1.9.jar:/usr/hdp/2.6.3.0-235/tez/lib/jersey-json-1.9.jar:/usr/hdp/2.6.3.0-235/tez/lib/jettison-1.3.4.jar:/usr/hdp/2.6.3.0-235/tez/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.6.3.0-235/tez/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.6.3.0-235/tez/lib/jsr305-2.0.3.jar:/usr/hdp/2.6.3.0-235/tez/lib/metrics-core-3.1.0.jar:/usr/hdp/2.6.3.0-235/tez/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.6.3.0-235/tez/lib/servlet-api-2.5.jar:/usr/hdp/2.6.3.0-235/tez/lib/slf4j-api-1.7.5.jar:/usr/hdp/2.6.3.0-235/tez/conf STARTUP_MSG: build = git@github.com:hortonworks/hadoop.git -r 45bfd33bba8acadfa0e6024c80981c023b28d454; compiled by 'jenkins' on 2017-10-30T02:33Z STARTUP_MSG: java = 1.8.0_112 ************************************************************/ 2018-04-04 22:30:10,059 INFO namenode.NameNode (LogAdapter.java:info(47)) - registered UNIX signal handlers for [TERM, HUP, INT] 2018-04-04 22:30:10,063 INFO namenode.NameNode (NameNode.java:createNameNode(1615)) - createNameNode [] 2018-04-04 22:30:10,318 INFO impl.MetricsConfig (MetricsConfig.java:loadFirst(112)) - loaded properties from hadoop-metrics2.properties 2018-04-04 22:30:10,493 INFO timeline.HadoopTimelineMetricsSink (HadoopTimelineMetricsSink.java:init(82)) - Initializing Timeline metrics sink. 2018-04-04 22:30:10,494 INFO timeline.HadoopTimelineMetricsSink (HadoopTimelineMetricsSink.java:init(102)) - Identified hostname = rjio-uatcluster-namenode1, serviceName = namenode 2018-04-04 22:30:10,700 WARN availability.MetricCollectorHAHelper (MetricCollectorHAHelper.java:findLiveCollectorHostsFromZNode(90)) - Unable to connect to zookeeper. org.apache.hadoop.metrics2.sink.relocated.zookeeper.KeeperException$ConnectionLossException: KeeperErrorCode = ConnectionLoss for /ambari-metrics-cluster at org.apache.hadoop.metrics2.sink.relocated.zookeeper.KeeperException.create(KeeperException.java:99) at org.apache.hadoop.metrics2.sink.relocated.zookeeper.KeeperException.create(KeeperException.java:51) at org.apache.hadoop.metrics2.sink.relocated.zookeeper.ZooKeeper.exists(ZooKeeper.java:1045) at org.apache.hadoop.metrics2.sink.relocated.zookeeper.ZooKeeper.exists(ZooKeeper.java:1073) at org.apache.hadoop.metrics2.sink.timeline.availability.MetricCollectorHAHelper.findLiveCollectorHostsFromZNode(MetricCollectorHAHelper.java:77) at org.apache.hadoop.metrics2.sink.timeline.AbstractTimelineMetricsSink.findPreferredCollectHost(AbstractTimelineMetricsSink.java:434) at org.apache.hadoop.metrics2.sink.timeline.HadoopTimelineMetricsSink.init(HadoopTimelineMetricsSink.java:120) at org.apache.hadoop.metrics2.impl.MetricsConfig.getPlugin(MetricsConfig.java:199) at org.apache.hadoop.metrics2.impl.MetricsSystemImpl.newSink(MetricsSystemImpl.java:529) at org.apache.hadoop.metrics2.impl.MetricsSystemImpl.configureSinks(MetricsSystemImpl.java:501) at org.apache.hadoop.metrics2.impl.MetricsSystemImpl.configure(MetricsSystemImpl.java:480) at org.apache.hadoop.metrics2.impl.MetricsSystemImpl.start(MetricsSystemImpl.java:189) at org.apache.hadoop.metrics2.impl.MetricsSystemImpl.init(MetricsSystemImpl.java:164) at org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.init(DefaultMetricsSystem.java:54) at org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.initialize(DefaultMetricsSystem.java:50) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1700) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1769) 2018-04-04 22:30:11,793 INFO timeline.HadoopTimelineMetricsSink (HadoopTimelineMetricsSink.java:init(128)) - No suitable collector found. 2018-04-04 22:30:11,796 INFO timeline.HadoopTimelineMetricsSink (HadoopTimelineMetricsSink.java:init(180)) - RPC port properties configured: {8020=client} 2018-04-04 22:30:11,803 INFO impl.MetricsSinkAdapter (MetricsSinkAdapter.java:start(206)) - Sink timeline started 2018-04-04 22:30:11,864 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:startTimer(376)) - Scheduled snapshot period at 10 second(s). 2018-04-04 22:30:11,864 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:start(192)) - NameNode metrics system started 2018-04-04 22:30:11,869 INFO namenode.NameNode (NameNode.java:setClientNamenodeAddress(450)) - fs.defaultFS is hdfs://ril-ha-uat 2018-04-04 22:30:11,870 INFO namenode.NameNode (NameNode.java:setClientNamenodeAddress(470)) - Clients are to use ril-ha-uat to access this namenode/service. 2018-04-04 22:30:12,204 INFO security.UserGroupInformation (UserGroupInformation.java:loginUserFromKeytab(1101)) - Login successful for user nn/rjio-uatcluster-namenode1@RJIO.CFMS.UAT using keytab file /etc/security/keytabs/nn.service.keytab 2018-04-04 22:30:12,218 INFO util.JvmPauseMonitor (JvmPauseMonitor.java:run(179)) - Starting JVM pause monitor 2018-04-04 22:30:12,228 INFO hdfs.DFSUtil (DFSUtil.java:httpServerTemplateForNNAndJN(1791)) - Starting web server as: HTTP/rjio-uatcluster-namenode1@RJIO.CFMS.UAT 2018-04-04 22:30:12,228 INFO hdfs.DFSUtil (DFSUtil.java:httpServerTemplateForNNAndJN(1803)) - Starting Web-server for hdfs at: http://rjio-UatCluster-NameNode1:50070 2018-04-04 22:30:12,277 INFO mortbay.log (Slf4jLog.java:info(67)) - Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog 2018-04-04 22:30:12,285 INFO server.AuthenticationFilter (AuthenticationFilter.java:constructSecretProvider(296)) - Unable to initialize FileSignerSecretProvider, falling back to use random secrets. 2018-04-04 22:30:12,290 INFO http.HttpRequestLog (HttpRequestLog.java:getRequestLog(80)) - Http request log for http.requests.namenode is not defined 2018-04-04 22:30:12,295 INFO http.HttpServer2 (HttpServer2.java:addGlobalFilter(791)) - Added global filter 'safety' (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter) 2018-04-04 22:30:12,298 INFO http.HttpServer2 (HttpServer2.java:addFilter(766)) - Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context hdfs 2018-04-04 22:30:12,298 INFO http.HttpServer2 (HttpServer2.java:addFilter(774)) - Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context static 2018-04-04 22:30:12,298 INFO http.HttpServer2 (HttpServer2.java:addFilter(774)) - Added filter static_user_filter (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs 2018-04-04 22:30:12,298 INFO security.HttpCrossOriginFilterInitializer (HttpCrossOriginFilterInitializer.java:initFilter(49)) - CORS filter not enabled. Please set hadoop.http.cross-origin.enabled to 'true' to enable it 2018-04-04 22:30:12,323 INFO http.HttpServer2 (NameNodeHttpServer.java:initWebHdfs(93)) - Added filter 'org.apache.hadoop.hdfs.web.AuthFilter' (class=org.apache.hadoop.hdfs.web.AuthFilter) 2018-04-04 22:30:12,324 INFO http.HttpServer2 (HttpServer2.java:addJerseyResourcePackage(690)) - addJerseyResourcePackage: packageName=org.apache.hadoop.hdfs.server.namenode.web.resources;org.apache.hadoop.hdfs.web.resources, pathSpec=/webhdfs/v1/* 2018-04-04 22:30:12,327 INFO http.HttpServer2 (HttpServer2.java:addInternalServlet(748)) - Adding Kerberos (SPNEGO) filter to getDelegationToken 2018-04-04 22:30:12,328 INFO http.HttpServer2 (HttpServer2.java:addInternalServlet(748)) - Adding Kerberos (SPNEGO) filter to renewDelegationToken 2018-04-04 22:30:12,328 INFO http.HttpServer2 (HttpServer2.java:addInternalServlet(748)) - Adding Kerberos (SPNEGO) filter to cancelDelegationToken 2018-04-04 22:30:12,329 INFO http.HttpServer2 (HttpServer2.java:addInternalServlet(748)) - Adding Kerberos (SPNEGO) filter to fsck 2018-04-04 22:30:12,330 INFO http.HttpServer2 (HttpServer2.java:addInternalServlet(748)) - Adding Kerberos (SPNEGO) filter to imagetransfer 2018-04-04 22:30:12,333 INFO http.HttpServer2 (HttpServer2.java:bindListener(989)) - Jetty bound to port 50070 2018-04-04 22:30:12,333 INFO mortbay.log (Slf4jLog.java:info(67)) - jetty-6.1.26.hwx 2018-04-04 22:30:12,495 INFO server.KerberosAuthenticationHandler (KerberosAuthenticationHandler.java:init(262)) - Login using keytab /etc/security/keytabs/spnego.service.keytab, for principal HTTP/rjio-uatcluster-namenode1@RJIO.CFMS.UAT 2018-04-04 22:30:12,498 INFO server.KerberosAuthenticationHandler (KerberosAuthenticationHandler.java:init(281)) - Map server: rjio-uatcluster-namenode1 to principal: [HTTP/rjio-uatcluster-namenode1@RJIO.CFMS.UAT], added = true 2018-04-04 22:30:12,502 INFO server.KerberosAuthenticationHandler (KerberosAuthenticationHandler.java:init(262)) - Login using keytab /etc/security/keytabs/spnego.service.keytab, for principal HTTP/rjio-uatcluster-namenode1@RJIO.CFMS.UAT 2018-04-04 22:30:12,503 INFO server.KerberosAuthenticationHandler (KerberosAuthenticationHandler.java:init(281)) - Map server: rjio-uatcluster-namenode1 to principal: [HTTP/rjio-uatcluster-namenode1@RJIO.CFMS.UAT], added = true 2018-04-04 22:30:12,508 INFO mortbay.log (Slf4jLog.java:info(67)) - Started HttpServer2$SelectChannelConnectorWithSafeStartup@rjio-UatCluster-NameNode1:50070 2018-04-04 22:30:12,534 WARN common.Util (Util.java:stringAsURI(57)) - Path /hadoop/hdfs/namenode should be specified as a URI in configuration files. Please update hdfs configuration. 2018-04-04 22:30:12,534 WARN common.Util (Util.java:stringAsURI(57)) - Path /opt/hadoop/hdfs/namenode should be specified as a URI in configuration files. Please update hdfs configuration. 2018-04-04 22:30:12,535 WARN common.Util (Util.java:stringAsURI(57)) - Path /hadoop/hdfs/namenode should be specified as a URI in configuration files. Please update hdfs configuration. 2018-04-04 22:30:12,535 WARN common.Util (Util.java:stringAsURI(57)) - Path /opt/hadoop/hdfs/namenode should be specified as a URI in configuration files. Please update hdfs configuration. 2018-04-04 22:30:12,541 WARN common.Util (Util.java:stringAsURI(57)) - Path /hadoop/hdfs/namenode should be specified as a URI in configuration files. Please update hdfs configuration. 2018-04-04 22:30:12,541 WARN common.Util (Util.java:stringAsURI(57)) - Path /opt/hadoop/hdfs/namenode should be specified as a URI in configuration files. Please update hdfs configuration. 2018-04-04 22:30:12,541 WARN common.Util (Util.java:stringAsURI(57)) - Path /hadoop/hdfs/namenode should be specified as a URI in configuration files. Please update hdfs configuration. 2018-04-04 22:30:12,542 WARN common.Util (Util.java:stringAsURI(57)) - Path /opt/hadoop/hdfs/namenode should be specified as a URI in configuration files. Please update hdfs configuration. 2018-04-04 22:30:12,548 WARN common.Storage (NNStorage.java:setRestoreFailedStorage(208)) - set restore failed storage to true 2018-04-04 22:30:12,573 INFO namenode.FSEditLog (FSEditLog.java:newInstance(225)) - Edit logging is async:false 2018-04-04 22:30:12,577 INFO namenode.FSNamesystem (FSNamesystem.java:(758)) - No KeyProvider found. 2018-04-04 22:30:12,577 INFO namenode.FSNamesystem (FSNamesystem.java:(764)) - Enabling async auditlog 2018-04-04 22:30:12,578 INFO namenode.FSNamesystem (FSNamesystem.java:(768)) - fsLock is fair:false 2018-04-04 22:30:12,608 INFO blockmanagement.HeartbeatManager (HeartbeatManager.java:(91)) - Setting heartbeat recheck interval to 30000 since dfs.namenode.stale.datanode.interval is less than dfs.namenode.heartbeat.recheck-interval 2018-04-04 22:30:12,611 INFO common.Util (Util.java:isDiskStatsEnabled(111)) - dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO profiling 2018-04-04 22:30:12,621 INFO blockmanagement.DatanodeManager (DatanodeManager.java:(274)) - dfs.block.invalidate.limit=1000 2018-04-04 22:30:12,621 INFO blockmanagement.DatanodeManager (DatanodeManager.java:(280)) - dfs.namenode.datanode.registration.ip-hostname-check=true 2018-04-04 22:30:12,623 INFO blockmanagement.BlockManager (InvalidateBlocks.java:printBlockDeletionTime(71)) - dfs.namenode.startup.delay.block.deletion.sec is set to 000:01:00:00.000 2018-04-04 22:30:12,623 INFO blockmanagement.BlockManager (InvalidateBlocks.java:printBlockDeletionTime(76)) - The block deletion will start around 2018 Apr 04 23:30:12 2018-04-04 22:30:12,625 INFO util.GSet (LightWeightGSet.java:computeCapacity(395)) - Computing capacity for map BlocksMap 2018-04-04 22:30:12,625 INFO util.GSet (LightWeightGSet.java:computeCapacity(396)) - VM type = 64-bit 2018-04-04 22:30:12,626 INFO util.GSet (LightWeightGSet.java:computeCapacity(397)) - 2.0% max memory 499.3 MB = 10.0 MB 2018-04-04 22:30:12,626 INFO util.GSet (LightWeightGSet.java:computeCapacity(402)) - capacity = 2^20 = 1048576 entries 2018-04-04 22:30:12,632 INFO blockmanagement.BlockManager (BlockManager.java:createBlockTokenSecretManager(409)) - dfs.block.access.token.enable=true 2018-04-04 22:30:12,632 INFO blockmanagement.BlockManager (BlockManager.java:createBlockTokenSecretManager(430)) - dfs.block.access.key.update.interval=600 min(s), dfs.block.access.token.lifetime=600 min(s), dfs.encrypt.data.transfer.algorithm=null 2018-04-04 22:30:12,639 INFO blockmanagement.BlockManager (BlockManager.java:(395)) - defaultReplication = 3 2018-04-04 22:30:12,639 INFO blockmanagement.BlockManager (BlockManager.java:(396)) - maxReplication = 50 2018-04-04 22:30:12,639 INFO blockmanagement.BlockManager (BlockManager.java:(397)) - minReplication = 1 2018-04-04 22:30:12,639 INFO blockmanagement.BlockManager (BlockManager.java:(398)) - maxReplicationStreams = 2 2018-04-04 22:30:12,639 INFO blockmanagement.BlockManager (BlockManager.java:(399)) - replicationRecheckInterval = 3000 2018-04-04 22:30:12,639 INFO blockmanagement.BlockManager (BlockManager.java:(400)) - encryptDataTransfer = false 2018-04-04 22:30:12,639 INFO blockmanagement.BlockManager (BlockManager.java:(401)) - maxNumBlocksToLog = 1000 2018-04-04 22:30:12,641 INFO namenode.FSNamesystem (FSNamesystem.java:(789)) - fsOwner = nn/rjio-uatcluster-namenode1@RJIO.CFMS.UAT (auth:KERBEROS) 2018-04-04 22:30:12,641 INFO namenode.FSNamesystem (FSNamesystem.java:(790)) - supergroup = hdfs 2018-04-04 22:30:12,641 INFO namenode.FSNamesystem (FSNamesystem.java:(791)) - isPermissionEnabled = true 2018-04-04 22:30:12,641 INFO namenode.FSNamesystem (FSNamesystem.java:(800)) - Determined nameservice ID: ril-ha-uat 2018-04-04 22:30:12,641 INFO namenode.FSNamesystem (FSNamesystem.java:(802)) - HA Enabled: true 2018-04-04 22:30:12,642 INFO namenode.FSNamesystem (FSNamesystem.java:(839)) - Append Enabled: true 2018-04-04 22:30:12,676 INFO util.GSet (LightWeightGSet.java:computeCapacity(395)) - Computing capacity for map INodeMap 2018-04-04 22:30:12,676 INFO util.GSet (LightWeightGSet.java:computeCapacity(396)) - VM type = 64-bit 2018-04-04 22:30:12,677 INFO util.GSet (LightWeightGSet.java:computeCapacity(397)) - 1.0% max memory 499.3 MB = 5.0 MB 2018-04-04 22:30:12,677 INFO util.GSet (LightWeightGSet.java:computeCapacity(402)) - capacity = 2^19 = 524288 entries 2018-04-04 22:30:12,677 INFO namenode.FSDirectory (FSDirectory.java:(256)) - ACLs enabled? false 2018-04-04 22:30:12,677 INFO namenode.FSDirectory (FSDirectory.java:(260)) - XAttrs enabled? true 2018-04-04 22:30:12,678 INFO namenode.FSDirectory (FSDirectory.java:(268)) - Maximum size of an xattr: 16384 2018-04-04 22:30:12,678 INFO namenode.NameNode (FSDirectory.java:(321)) - Caching file names occuring more than 10 times 2018-04-04 22:30:12,686 INFO util.GSet (LightWeightGSet.java:computeCapacity(395)) - Computing capacity for map cachedBlocks 2018-04-04 22:30:12,687 INFO util.GSet (LightWeightGSet.java:computeCapacity(396)) - VM type = 64-bit 2018-04-04 22:30:12,687 INFO util.GSet (LightWeightGSet.java:computeCapacity(397)) - 0.25% max memory 499.3 MB = 1.2 MB 2018-04-04 22:30:12,687 INFO util.GSet (LightWeightGSet.java:computeCapacity(402)) - capacity = 2^17 = 131072 entries 2018-04-04 22:30:12,689 INFO namenode.FSNamesystem (FSNamesystem.java:(5574)) - dfs.namenode.safemode.threshold-pct = 1.0 2018-04-04 22:30:12,689 INFO namenode.FSNamesystem (FSNamesystem.java:(5575)) - dfs.namenode.safemode.min.datanodes = 0 2018-04-04 22:30:12,689 INFO namenode.FSNamesystem (FSNamesystem.java:(5576)) - dfs.namenode.safemode.extension = 30000 2018-04-04 22:30:12,691 INFO metrics.TopMetrics (TopMetrics.java:logConf(76)) - NNTop conf: dfs.namenode.top.window.num.buckets = 10 2018-04-04 22:30:12,691 INFO metrics.TopMetrics (TopMetrics.java:logConf(78)) - NNTop conf: dfs.namenode.top.num.users = 10 2018-04-04 22:30:12,691 INFO metrics.TopMetrics (TopMetrics.java:logConf(80)) - NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25 2018-04-04 22:30:12,694 INFO namenode.FSNamesystem (FSNamesystem.java:initRetryCache(967)) - Retry cache on namenode is enabled 2018-04-04 22:30:12,694 INFO namenode.FSNamesystem (FSNamesystem.java:initRetryCache(975)) - Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis 2018-04-04 22:30:12,696 INFO util.GSet (LightWeightGSet.java:computeCapacity(395)) - Computing capacity for map NameNodeRetryCache 2018-04-04 22:30:12,696 INFO util.GSet (LightWeightGSet.java:computeCapacity(396)) - VM type = 64-bit 2018-04-04 22:30:12,696 INFO util.GSet (LightWeightGSet.java:computeCapacity(397)) - 0.029999999329447746% max memory 499.3 MB = 153.4 KB 2018-04-04 22:30:12,696 INFO util.GSet (LightWeightGSet.java:computeCapacity(402)) - capacity = 2^14 = 16384 entries 2018-04-04 22:30:12,706 INFO common.Storage (Storage.java:tryLock(776)) - Lock on /hadoop/hdfs/namenode/in_use.lock acquired by nodename 27933@rjio-UatCluster-NameNode1 2018-04-04 22:30:12,760 INFO common.Storage (Storage.java:tryLock(776)) - Lock on /opt/hadoop/hdfs/namenode/in_use.lock acquired by nodename 27933@rjio-UatCluster-NameNode1 2018-04-04 22:30:13,131 INFO namenode.FSImage (FSImage.java:loadFSImageFile(731)) - Planning to load image: FSImageFile(file=/hadoop/hdfs/namenode/current/fsimage_0000000000000000000, cpktTxId=0000000000000000000) 2018-04-04 22:30:13,152 INFO namenode.FSImageFormatPBINode (FSImageFormatPBINode.java:loadINodeSection(257)) - Loading 1 INodes. 2018-04-04 22:30:13,186 INFO namenode.FSImageFormatProtobuf (FSImageFormatProtobuf.java:load(184)) - Loaded FSImage in 0 seconds. 2018-04-04 22:30:13,186 INFO namenode.FSImage (FSImage.java:loadFSImage(897)) - Loaded image for txid 0 from /hadoop/hdfs/namenode/current/fsimage_0000000000000000000 2018-04-04 22:30:13,192 INFO namenode.FSImage (FSImage.java:loadEdits(835)) - Reading org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream@3301500b expecting start txid #1 2018-04-04 22:30:13,192 INFO namenode.FSImage (FSEditLogLoader.java:loadFSEdits(142)) - Start loading edits file http://rjio-UatCluster-NameNode1:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50, http://rjio-uatcluster-datanode1:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50, http://rjio-uatcluster-namenode2:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50 2018-04-04 22:30:13,197 INFO namenode.RedundantEditLogInputStream (RedundantEditLogInputStream.java:nextOp(177)) - Fast-forwarding stream 'http://rjio-UatCluster-NameNode1:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50, http://rjio-uatcluster-datanode1:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50, http://rjio-uatcluster-namenode2:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50' to transaction ID 1 2018-04-04 22:30:13,197 INFO namenode.RedundantEditLogInputStream (RedundantEditLogInputStream.java:nextOp(177)) - Fast-forwarding stream 'http://rjio-UatCluster-NameNode1:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50' to transaction ID 1 2018-04-04 22:30:13,229 ERROR namenode.EditLogInputStream (EditLogFileInputStream.java:nextOpImpl(194)) - caught exception initializing http://rjio-UatCluster-NameNode1:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50 java.io.IOException: org.apache.hadoop.security.authentication.client.AuthenticationException: Authentication failed, URL: http://rjio-UatCluster-NameNode1:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50&user.name=nn/rjio-uatcluster-namenode1@RJIO.CFMS.UAT, status: 403, message: org.apache.hadoop.security.authentication.client.AuthenticationException at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:473) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:465) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1866) at org.apache.hadoop.security.SecurityUtil.doAsUser(SecurityUtil.java:509) at org.apache.hadoop.security.SecurityUtil.doAsCurrentUser(SecurityUtil.java:503) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog.getInputStream(EditLogFileInputStream.java:464) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream.init(EditLogFileInputStream.java:141) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream.nextOpImpl(EditLogFileInputStream.java:192) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream.nextOp(EditLogFileInputStream.java:250) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.skipUntil(EditLogInputStream.java:151) at org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream.nextOp(RedundantEditLogInputStream.java:179) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.skipUntil(EditLogInputStream.java:151) at org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream.nextOp(RedundantEditLogInputStream.java:179) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadEditRecords(FSEditLogLoader.java:190) at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadFSEdits(FSEditLogLoader.java:143) at org.apache.hadoop.hdfs.server.namenode.FSImage.loadEdits(FSImage.java:838) at org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:693) at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:289) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:1073) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:723) at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:688) at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:752) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:992) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:976) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1701) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1769) Caused by: org.apache.hadoop.security.authentication.client.AuthenticationException: Authentication failed, URL: http://rjio-UatCluster-NameNode1:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50&user.name=nn/rjio-uatcluster-namenode1@RJIO.CFMS.UAT, status: 403, message: org.apache.hadoop.security.authentication.client.AuthenticationException at org.apache.hadoop.security.authentication.client.AuthenticatedURL.extractToken(AuthenticatedURL.java:281) at org.apache.hadoop.security.authentication.client.PseudoAuthenticator.authenticate(PseudoAuthenticator.java:77) at org.apache.hadoop.security.authentication.client.KerberosAuthenticator.authenticate(KerberosAuthenticator.java:212) at org.apache.hadoop.security.authentication.client.AuthenticatedURL.openConnection(AuthenticatedURL.java:216) at org.apache.hadoop.hdfs.web.URLConnectionFactory.openConnection(URLConnectionFactory.java:190) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:471) ... 30 more 2018-04-04 22:30:13,230 ERROR namenode.RedundantEditLogInputStream (RedundantEditLogInputStream.java:nextOp(222)) - Got error reading edit log input stream http://rjio-UatCluster-NameNode1:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50; failing over to edit log http://rjio-uatcluster-datanode1:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50 org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream$PrematureEOFException: got premature end-of-file at txid 0; expected file to go up to 4 at org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream.nextOp(RedundantEditLogInputStream.java:197) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.skipUntil(EditLogInputStream.java:151) at org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream.nextOp(RedundantEditLogInputStream.java:179) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadEditRecords(FSEditLogLoader.java:190) at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadFSEdits(FSEditLogLoader.java:143) at org.apache.hadoop.hdfs.server.namenode.FSImage.loadEdits(FSImage.java:838) at org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:693) at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:289) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:1073) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:723) at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:688) at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:752) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:992) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:976) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1701) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1769) 2018-04-04 22:30:13,231 INFO namenode.RedundantEditLogInputStream (RedundantEditLogInputStream.java:nextOp(177)) - Fast-forwarding stream 'http://rjio-uatcluster-datanode1:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50' to transaction ID 1 2018-04-04 22:30:13,245 ERROR namenode.EditLogInputStream (EditLogFileInputStream.java:nextOpImpl(194)) - caught exception initializing http://rjio-uatcluster-datanode1:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50 java.io.IOException: org.apache.hadoop.security.authentication.client.AuthenticationException: Authentication failed, URL: http://rjio-uatcluster-datanode1:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50&user.name=nn/rjio-uatcluster-namenode1@RJIO.CFMS.UAT, status: 403, message: org.apache.hadoop.security.authentication.client.AuthenticationException at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:473) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:465) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1866) at org.apache.hadoop.security.SecurityUtil.doAsUser(SecurityUtil.java:509) at org.apache.hadoop.security.SecurityUtil.doAsCurrentUser(SecurityUtil.java:503) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog.getInputStream(EditLogFileInputStream.java:464) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream.init(EditLogFileInputStream.java:141) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream.nextOpImpl(EditLogFileInputStream.java:192) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream.nextOp(EditLogFileInputStream.java:250) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.skipUntil(EditLogInputStream.java:151) at org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream.nextOp(RedundantEditLogInputStream.java:179) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.skipUntil(EditLogInputStream.java:151) at org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream.nextOp(RedundantEditLogInputStream.java:179) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadEditRecords(FSEditLogLoader.java:190) at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadFSEdits(FSEditLogLoader.java:143) at org.apache.hadoop.hdfs.server.namenode.FSImage.loadEdits(FSImage.java:838) at org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:693) at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:289) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:1073) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:723) at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:688) at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:752) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:992) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:976) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1701) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1769) Caused by: org.apache.hadoop.security.authentication.client.AuthenticationException: Authentication failed, URL: http://rjio-uatcluster-datanode1:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50&user.name=nn/rjio-uatcluster-namenode1@RJIO.CFMS.UAT, status: 403, message: org.apache.hadoop.security.authentication.client.AuthenticationException at org.apache.hadoop.security.authentication.client.AuthenticatedURL.extractToken(AuthenticatedURL.java:281) at org.apache.hadoop.security.authentication.client.PseudoAuthenticator.authenticate(PseudoAuthenticator.java:77) at org.apache.hadoop.security.authentication.client.KerberosAuthenticator.authenticate(KerberosAuthenticator.java:212) at org.apache.hadoop.security.authentication.client.AuthenticatedURL.openConnection(AuthenticatedURL.java:216) at org.apache.hadoop.hdfs.web.URLConnectionFactory.openConnection(URLConnectionFactory.java:190) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:471) ... 30 more 2018-04-04 22:30:13,245 ERROR namenode.RedundantEditLogInputStream (RedundantEditLogInputStream.java:nextOp(222)) - Got error reading edit log input stream http://rjio-uatcluster-datanode1:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50; failing over to edit log http://rjio-uatcluster-namenode2:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50 org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream$PrematureEOFException: got premature end-of-file at txid 0; expected file to go up to 4 at org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream.nextOp(RedundantEditLogInputStream.java:197) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.skipUntil(EditLogInputStream.java:151) at org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream.nextOp(RedundantEditLogInputStream.java:179) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadEditRecords(FSEditLogLoader.java:190) at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadFSEdits(FSEditLogLoader.java:143) at org.apache.hadoop.hdfs.server.namenode.FSImage.loadEdits(FSImage.java:838) at org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:693) at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:289) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:1073) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:723) at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:688) at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:752) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:992) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:976) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1701) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1769) 2018-04-04 22:30:13,246 INFO namenode.RedundantEditLogInputStream (RedundantEditLogInputStream.java:nextOp(177)) - Fast-forwarding stream 'http://rjio-uatcluster-namenode2:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50' to transaction ID 1 2018-04-04 22:30:13,263 ERROR namenode.EditLogInputStream (EditLogFileInputStream.java:nextOpImpl(194)) - caught exception initializing http://rjio-uatcluster-namenode2:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50 java.io.IOException: org.apache.hadoop.security.authentication.client.AuthenticationException: Authentication failed, URL: http://rjio-uatcluster-namenode2:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50&user.name=nn/rjio-uatcluster-namenode1@RJIO.CFMS.UAT, status: 403, message: org.apache.hadoop.security.authentication.client.AuthenticationException at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:473) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:465) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1866) at org.apache.hadoop.security.SecurityUtil.doAsUser(SecurityUtil.java:509) at org.apache.hadoop.security.SecurityUtil.doAsCurrentUser(SecurityUtil.java:503) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog.getInputStream(EditLogFileInputStream.java:464) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream.init(EditLogFileInputStream.java:141) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream.nextOpImpl(EditLogFileInputStream.java:192) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream.nextOp(EditLogFileInputStream.java:250) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.skipUntil(EditLogInputStream.java:151) at org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream.nextOp(RedundantEditLogInputStream.java:179) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.skipUntil(EditLogInputStream.java:151) at org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream.nextOp(RedundantEditLogInputStream.java:179) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadEditRecords(FSEditLogLoader.java:190) at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadFSEdits(FSEditLogLoader.java:143) at org.apache.hadoop.hdfs.server.namenode.FSImage.loadEdits(FSImage.java:838) at org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:693) at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:289) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:1073) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:723) at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:688) at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:752) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:992) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:976) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1701) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1769) Caused by: org.apache.hadoop.security.authentication.client.AuthenticationException: Authentication failed, URL: http://rjio-uatcluster-namenode2:8480/getJournal?jid=ril-ha-uat&segmentTxId=1&storageInfo=-63%3A77008602%3A0%3ACID-55d3d3e8-7c5b-445a-85a0-f284b4ec8a50&user.name=nn/rjio-uatcluster-namenode1@RJIO.CFMS.UAT, status: 403, message: org.apache.hadoop.security.authentication.client.AuthenticationException at org.apache.hadoop.security.authentication.client.AuthenticatedURL.extractToken(AuthenticatedURL.java:281) at org.apache.hadoop.security.authentication.client.PseudoAuthenticator.authenticate(PseudoAuthenticator.java:77) at org.apache.hadoop.security.authentication.client.KerberosAuthenticator.authenticate(KerberosAuthenticator.java:212) at org.apache.hadoop.security.authentication.client.AuthenticatedURL.openConnection(AuthenticatedURL.java:216) at org.apache.hadoop.hdfs.web.URLConnectionFactory.openConnection(URLConnectionFactory.java:190) at org.apache.hadoop.hdfs.server.namenode.EditLogFileInputStream$URLLog$1.run(EditLogFileInputStream.java:471) ... 30 more 2018-04-04 22:30:13,264 ERROR namenode.FSImage (FSEditLogLoader.java:loadEditRecords(199)) - Error replaying edit log at offset 0. Expected transaction ID was 1 org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream$PrematureEOFException: got premature end-of-file at txid 0; expected file to go up to 4 at org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream.nextOp(RedundantEditLogInputStream.java:197) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.skipUntil(EditLogInputStream.java:151) at org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream.nextOp(RedundantEditLogInputStream.java:179) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadEditRecords(FSEditLogLoader.java:190) at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadFSEdits(FSEditLogLoader.java:143) at org.apache.hadoop.hdfs.server.namenode.FSImage.loadEdits(FSImage.java:838) at org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:693) at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:289) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:1073) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:723) at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:688) at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:752) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:992) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:976) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1701) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1769) 2018-04-04 22:30:13,369 WARN namenode.FSNamesystem (FSNamesystem.java:loadFromDisk(725)) - Encountered exception loading fsimage org.apache.hadoop.hdfs.server.namenode.EditLogInputException: Error replaying edit log at offset 0. Expected transaction ID was 1 at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadEditRecords(FSEditLogLoader.java:203) at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadFSEdits(FSEditLogLoader.java:143) at org.apache.hadoop.hdfs.server.namenode.FSImage.loadEdits(FSImage.java:838) at org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:693) at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:289) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:1073) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:723) at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:688) at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:752) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:992) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:976) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1701) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1769) Caused by: org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream$PrematureEOFException: got premature end-of-file at txid 0; expected file to go up to 4 at org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream.nextOp(RedundantEditLogInputStream.java:197) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.skipUntil(EditLogInputStream.java:151) at org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream.nextOp(RedundantEditLogInputStream.java:179) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadEditRecords(FSEditLogLoader.java:190) ... 12 more 2018-04-04 22:30:13,371 INFO mortbay.log (Slf4jLog.java:info(67)) - Stopped HttpServer2$SelectChannelConnectorWithSafeStartup@rjio-UatCluster-NameNode1:50070 2018-04-04 22:30:13,472 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stop(211)) - Stopping NameNode metrics system... 2018-04-04 22:30:13,472 INFO impl.MetricsSinkAdapter (MetricsSinkAdapter.java:publishMetricsFromQueue(141)) - timeline thread interrupted. 2018-04-04 22:30:13,473 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:stop(217)) - NameNode metrics system stopped. 2018-04-04 22:30:13,474 INFO impl.MetricsSystemImpl (MetricsSystemImpl.java:shutdown(606)) - NameNode metrics system shutdown complete. 2018-04-04 22:30:13,474 ERROR namenode.NameNode (NameNode.java:main(1774)) - Failed to start namenode. org.apache.hadoop.hdfs.server.namenode.EditLogInputException: Error replaying edit log at offset 0. Expected transaction ID was 1 at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadEditRecords(FSEditLogLoader.java:203) at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadFSEdits(FSEditLogLoader.java:143) at org.apache.hadoop.hdfs.server.namenode.FSImage.loadEdits(FSImage.java:838) at org.apache.hadoop.hdfs.server.namenode.FSImage.loadFSImage(FSImage.java:693) at org.apache.hadoop.hdfs.server.namenode.FSImage.recoverTransitionRead(FSImage.java:289) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFSImage(FSNamesystem.java:1073) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.loadFromDisk(FSNamesystem.java:723) at org.apache.hadoop.hdfs.server.namenode.NameNode.loadNamesystem(NameNode.java:688) at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:752) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:992) at org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:976) at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1701) at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1769) Caused by: org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream$PrematureEOFException: got premature end-of-file at txid 0; expected file to go up to 4 at org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream.nextOp(RedundantEditLogInputStream.java:197) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.skipUntil(EditLogInputStream.java:151) at org.apache.hadoop.hdfs.server.namenode.RedundantEditLogInputStream.nextOp(RedundantEditLogInputStream.java:179) at org.apache.hadoop.hdfs.server.namenode.EditLogInputStream.readOp(EditLogInputStream.java:85) at org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.loadEditRecords(FSEditLogLoader.java:190) ... 12 more 2018-04-04 22:30:13,475 INFO util.ExitUtil (ExitUtil.java:terminate(124)) - Exiting with status 1 2018-04-04 22:30:13,476 INFO timeline.HadoopTimelineMetricsSink (AbstractTimelineMetricsSink.java:getCurrentCollectorHost(278)) - No live collector to send metrics to. Metrics to be sent will be discarded. This message will be skipped for the next 20 times. 2018-04-04 22:30:13,476 INFO namenode.NameNode (LogAdapter.java:info(47)) - SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down NameNode at rjio-UatCluster-NameNode1/10.20.1.180