2017-06-06 05:35:06,108 INFO org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer: STARTUP_MSG: /************************************************************ STARTUP_MSG: Starting ApplicationHistoryServer STARTUP_MSG: host = c2m.xdata.com/10.90.2.90 STARTUP_MSG: args = [] STARTUP_MSG: version = 2.7.1.2.4.0.0-170 STARTUP_MSG: classpath = /usr/lib/ambari-metrics-collector/activation-1.1.jar:/usr/lib/ambari-metrics-collector/commons-lang-2.5.jar:/usr/lib/ambari-metrics-collector/jackson-jaxrs-1.9.2.jar:/usr/lib/ambari-metrics-collector/jetty-6.1.26.hwx.jar:/usr/lib/ambari-metrics-collector/jersey-json-1.11.jar:/usr/lib/ambari-metrics-collector/commons-compress-1.4.1.jar:/usr/lib/ambari-metrics-collector/hadoop-annotations-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/api-util-1.0.0-M20.jar:/usr/lib/ambari-metrics-collector/aopalliance-1.0.jar:/usr/lib/ambari-metrics-collector/hadoop-yarn-registry-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/joda-time-2.7.jar:/usr/lib/ambari-metrics-collector/commons-httpclient-3.1.jar:/usr/lib/ambari-metrics-collector/commons-pool2-2.3.jar:/usr/lib/ambari-metrics-collector/commons-beanutils-1.7.0.jar:/usr/lib/ambari-metrics-collector/antlr-2.7.7.jar:/usr/lib/ambari-metrics-collector/httpclient-4.0.1.jar:/usr/lib/ambari-metrics-collector/apacheds-i18n-2.0.0-M15.jar:/usr/lib/ambari-metrics-collector/jackson-xc-1.9.13.jar:/usr/lib/ambari-metrics-collector/commons-io-2.4.jar:/usr/lib/ambari-metrics-collector/htrace-core-3.1.0-incubating.jar:/usr/lib/ambari-metrics-collector/jetty-util-6.1.26.hwx.jar:/usr/lib/ambari-metrics-collector/findbugs-annotations-1.3.9-1.jar:/usr/lib/ambari-metrics-collector/stax-api-1.0.1.jar:/usr/lib/ambari-metrics-collector/guice-servlet-3.0.jar:/usr/lib/ambari-metrics-collector/hbase-hadoop-compat-1.1.2.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/slf4j-log4j12-1.7.2.jar:/usr/lib/ambari-metrics-collector/curator-client-2.7.1.jar:/usr/lib/ambari-metrics-collector/jaxb-api-2.2.2.jar:/usr/lib/ambari-metrics-collector/netty-3.6.2.Final.jar:/usr/lib/ambari-metrics-collector/ambari-metrics-timelineservice-2.2.1.0.161.jar:/usr/lib/ambari-metrics-collector/jetty-sslengine-6.1.26.hwx.jar:/usr/lib/ambari-metrics-collector/okhttp-2.4.0.jar:/usr/lib/ambari-metrics-collector/jackson-mapper-asl-1.9.13.jar:/usr/lib/ambari-metrics-collector/commons-csv-1.0.jar:/usr/lib/ambari-metrics-collector/disruptor-3.3.0.jar:/usr/lib/ambari-metrics-collector/hbase-client-1.1.2.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/jettison-1.1.jar:/usr/lib/ambari-metrics-collector/hadoop-mapreduce-client-jobclient-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/hadoop-yarn-api-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/commons-logging-1.1.1.jar:/usr/lib/ambari-metrics-collector/servlet-api-2.5.jar:/usr/lib/ambari-metrics-collector/stringtemplate-3.2.1.jar:/usr/lib/ambari-metrics-collector/asm-3.1.jar:/usr/lib/ambari-metrics-collector/jersey-server-1.11.jar:/usr/lib/ambari-metrics-collector/curator-recipes-2.7.1.jar:/usr/lib/ambari-metrics-collector/ambari-metrics-common-2.2.1.0.161.jar:/usr/lib/ambari-metrics-collector/joni-2.1.2.jar:/usr/lib/ambari-metrics-collector/hadoop-mapreduce-client-shuffle-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/jline-2.11.jar:/usr/lib/ambari-metrics-collector/jsp-api-2.1.jar:/usr/lib/ambari-metrics-collector/hadoop-yarn-common-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/jersey-client-1.11.jar:/usr/lib/ambari-metrics-collector/servlet-api-2.5-6.1.14.jar:/usr/lib/ambari-metrics-collector/snappy-0.3.jar:/usr/lib/ambari-metrics-collector/phoenix-core-4.4.0.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/stax-api-1.0-2.jar:/usr/lib/ambari-metrics-collector/jsp-api-2.1-6.1.14.jar:/usr/lib/ambari-metrics-collector/hbase-procedure-1.1.2.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/hadoop-common-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/commons-cli-1.2.jar:/usr/lib/ambari-metrics-collector/commons-beanutils-core-1.8.0.jar:/usr/lib/ambari-metrics-collector/httpcore-4.2.5.jar:/usr/lib/ambari-metrics-collector/avro-1.7.4.jar:/usr/lib/ambari-metrics-collector/hadoop-auth-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/netty-all-4.0.23.Final.jar:/usr/lib/ambari-metrics-collector/hadoop-mapreduce-client-core-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/api-asn1-api-1.0.0-M20.jar:/usr/lib/ambari-metrics-collector/hadoop-yarn-server-nodemanager-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/jersey-guice-1.11.jar:/usr/lib/ambari-metrics-collector/jersey-core-1.11.jar:/usr/lib/ambari-metrics-collector/log4j-1.2.17.jar:/usr/lib/ambari-metrics-collector/leveldbjni-all-1.8.jar:/usr/lib/ambari-metrics-collector/xz-1.0.jar:/usr/lib/ambari-metrics-collector/jdk.tools-1.7.jar:/usr/lib/ambari-metrics-collector/slf4j-api-1.7.2.jar:/usr/lib/ambari-metrics-collector/commons-math-2.2.jar:/usr/lib/ambari-metrics-collector/zookeeper-3.4.6.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/hadoop-yarn-server-common-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/hadoop-mapreduce-client-common-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/xml-apis-1.4.01.jar:/usr/lib/ambari-metrics-collector/commons-net-3.1.jar:/usr/lib/ambari-metrics-collector/commons-digester-1.8.jar:/usr/lib/ambari-metrics-collector/commons-configuration-1.6.jar:/usr/lib/ambari-metrics-collector/jersey-servlet-1.11.jar:/usr/lib/ambari-metrics-collector/antlr-3.5.jar:/usr/lib/ambari-metrics-collector/annotations-1.3.2.jar:/usr/lib/ambari-metrics-collector/hadoop-mapreduce-client-app-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/jamon-runtime-2.3.1.jar:/usr/lib/ambari-metrics-collector/paranamer-2.3.jar:/usr/lib/ambari-metrics-collector/jackson-core-asl-1.9.9.jar:/usr/lib/ambari-metrics-collector/okio-1.4.0.jar:/usr/lib/ambari-metrics-collector/ST4-4.0.7.jar:/usr/lib/ambari-metrics-collector/jcodings-1.0.8.jar:/usr/lib/ambari-metrics-collector/commons-collections-3.2.1.jar:/usr/lib/ambari-metrics-collector/hbase-server-1.1.2.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/jackson-core-2.2.3.jar:/usr/lib/ambari-metrics-collector/commons-daemon-1.0.13.jar:/usr/lib/ambari-metrics-collector/azure-storage-2.2.0.jar:/usr/lib/ambari-metrics-collector/javax.inject-1.jar:/usr/lib/ambari-metrics-collector/commons-codec-1.7.jar:/usr/lib/ambari-metrics-collector/jets3t-0.9.0.jar:/usr/lib/ambari-metrics-collector/guice-3.0.jar:/usr/lib/ambari-metrics-collector/guava-14.0.1.jar:/usr/lib/ambari-metrics-collector/hadoop-hdfs-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/jsch-0.1.42.jar:/usr/lib/ambari-metrics-collector/jaxb-impl-2.2.3-1.jar:/usr/lib/ambari-metrics-collector/xercesImpl-2.11.0.jar:/usr/lib/ambari-metrics-collector/hbase-common-1.1.2.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/protobuf-java-2.5.0.jar:/usr/lib/ambari-metrics-collector/jsr305-2.0.1.jar:/usr/lib/ambari-metrics-collector/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/lib/ambari-metrics-collector/commons-lang3-3.3.2.jar:/usr/lib/ambari-metrics-collector/snappy-java-1.0.4.1.jar:/usr/lib/ambari-metrics-collector/java-xmlbuilder-0.4.jar:/usr/lib/ambari-metrics-collector/hbase-common-1.1.2.2.4.0.0-170-tests.jar:/usr/lib/ambari-metrics-collector/microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/lib/ambari-metrics-collector/hbase-annotations-1.1.2.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/high-scale-lib-1.1.1.jar:/usr/lib/ambari-metrics-collector/hbase-protocol-1.1.2.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/xmlenc-0.52.jar:/usr/lib/ambari-metrics-collector/hadoop-client-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/gson-2.2.4.jar:/usr/lib/ambari-metrics-collector/curator-framework-2.7.1.jar:/usr/lib/ambari-metrics-collector/hadoop-yarn-client-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/antlr-runtime-3.5.jar:/usr/lib/ambari-metrics-collector/hadoop-azure-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/commons-math3-3.4.1.jar:/usr/lib/ambari-metrics-collector/jsp-2.1-6.1.14.jar:/usr/lib/ambari-metrics-collector/hbase-hadoop2-compat-1.1.2.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/sqlline-1.1.8.jar:/usr/lib/ambari-metrics-collector/metrics-core-2.2.0.jar:/etc/ambari-metrics-collector/conf STARTUP_MSG: build = git@github.com:hortonworks/hadoop.git -r 26104d8ac833884c8776473823007f176854f2eb; compiled by 'jenkins' on 2016-02-12T20:07Z STARTUP_MSG: java = 1.8.0_60 ************************************************************/ 2017-06-06 05:35:06,115 INFO org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer: registered UNIX signal handlers for [TERM, HUP, INT] 2017-06-06 05:35:06,139 INFO org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.TimelineMetricConfiguration: Found hbase site configuration: file:/etc/ambari-metrics-collector/conf/hbase-site.xml 2017-06-06 05:35:06,139 INFO org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.TimelineMetricConfiguration: Found metric service configuration: file:/etc/ambari-metrics-collector/conf/ams-site.xml 2017-06-06 05:35:06,418 INFO org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer: Explicitly disabled application timeline store. 2017-06-06 05:35:06,419 INFO org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer: Creating metrics store. 2017-06-06 05:35:06,433 INFO org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryManagerImpl: ApplicationHistory Init 2017-06-06 05:35:06,433 INFO org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryManagerImpl: Explicitly disabled application timeline store. 2017-06-06 05:35:06,697 INFO org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor: Initializing metrics schema... 2017-06-06 05:35:06,809 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:35:06,904 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:35:06,945 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:35:07,037 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:35:07,197 WARN org.apache.hadoop.util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 2017-06-06 05:35:07,334 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x5524cca1 connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:35:07,342 INFO org.apache.zookeeper.ZooKeeper: Client environment:zookeeper.version=3.4.6-170--1, built on 02/12/2016 19:39 GMT 2017-06-06 05:35:07,342 INFO org.apache.zookeeper.ZooKeeper: Client environment:host.name=c2m.xdata.com 2017-06-06 05:35:07,342 INFO org.apache.zookeeper.ZooKeeper: Client environment:java.version=1.8.0_60 2017-06-06 05:35:07,343 INFO org.apache.zookeeper.ZooKeeper: Client environment:java.vendor=Oracle Corporation 2017-06-06 05:35:07,343 INFO org.apache.zookeeper.ZooKeeper: Client environment:java.home=/usr/jdk64/jdk1.8.0_60/jre 2017-06-06 05:35:07,343 INFO org.apache.zookeeper.ZooKeeper: Client environment:java.class.path=/usr/lib/ambari-metrics-collector/activation-1.1.jar:/usr/lib/ambari-metrics-collector/commons-lang-2.5.jar:/usr/lib/ambari-metrics-collector/jackson-jaxrs-1.9.2.jar:/usr/lib/ambari-metrics-collector/jetty-6.1.26.hwx.jar:/usr/lib/ambari-metrics-collector/jersey-json-1.11.jar:/usr/lib/ambari-metrics-collector/commons-compress-1.4.1.jar:/usr/lib/ambari-metrics-collector/hadoop-annotations-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/api-util-1.0.0-M20.jar:/usr/lib/ambari-metrics-collector/aopalliance-1.0.jar:/usr/lib/ambari-metrics-collector/hadoop-yarn-registry-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/joda-time-2.7.jar:/usr/lib/ambari-metrics-collector/commons-httpclient-3.1.jar:/usr/lib/ambari-metrics-collector/commons-pool2-2.3.jar:/usr/lib/ambari-metrics-collector/commons-beanutils-1.7.0.jar:/usr/lib/ambari-metrics-collector/antlr-2.7.7.jar:/usr/lib/ambari-metrics-collector/httpclient-4.0.1.jar:/usr/lib/ambari-metrics-collector/apacheds-i18n-2.0.0-M15.jar:/usr/lib/ambari-metrics-collector/jackson-xc-1.9.13.jar:/usr/lib/ambari-metrics-collector/commons-io-2.4.jar:/usr/lib/ambari-metrics-collector/htrace-core-3.1.0-incubating.jar:/usr/lib/ambari-metrics-collector/jetty-util-6.1.26.hwx.jar:/usr/lib/ambari-metrics-collector/findbugs-annotations-1.3.9-1.jar:/usr/lib/ambari-metrics-collector/stax-api-1.0.1.jar:/usr/lib/ambari-metrics-collector/guice-servlet-3.0.jar:/usr/lib/ambari-metrics-collector/hbase-hadoop-compat-1.1.2.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/slf4j-log4j12-1.7.2.jar:/usr/lib/ambari-metrics-collector/curator-client-2.7.1.jar:/usr/lib/ambari-metrics-collector/jaxb-api-2.2.2.jar:/usr/lib/ambari-metrics-collector/netty-3.6.2.Final.jar:/usr/lib/ambari-metrics-collector/ambari-metrics-timelineservice-2.2.1.0.161.jar:/usr/lib/ambari-metrics-collector/jetty-sslengine-6.1.26.hwx.jar:/usr/lib/ambari-metrics-collector/okhttp-2.4.0.jar:/usr/lib/ambari-metrics-collector/jackson-mapper-asl-1.9.13.jar:/usr/lib/ambari-metrics-collector/commons-csv-1.0.jar:/usr/lib/ambari-metrics-collector/disruptor-3.3.0.jar:/usr/lib/ambari-metrics-collector/hbase-client-1.1.2.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/jettison-1.1.jar:/usr/lib/ambari-metrics-collector/hadoop-mapreduce-client-jobclient-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/hadoop-yarn-api-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/commons-logging-1.1.1.jar:/usr/lib/ambari-metrics-collector/servlet-api-2.5.jar:/usr/lib/ambari-metrics-collector/stringtemplate-3.2.1.jar:/usr/lib/ambari-metrics-collector/asm-3.1.jar:/usr/lib/ambari-metrics-collector/jersey-server-1.11.jar:/usr/lib/ambari-metrics-collector/curator-recipes-2.7.1.jar:/usr/lib/ambari-metrics-collector/ambari-metrics-common-2.2.1.0.161.jar:/usr/lib/ambari-metrics-collector/joni-2.1.2.jar:/usr/lib/ambari-metrics-collector/hadoop-mapreduce-client-shuffle-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/jline-2.11.jar:/usr/lib/ambari-metrics-collector/jsp-api-2.1.jar:/usr/lib/ambari-metrics-collector/hadoop-yarn-common-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/jersey-client-1.11.jar:/usr/lib/ambari-metrics-collector/servlet-api-2.5-6.1.14.jar:/usr/lib/ambari-metrics-collector/snappy-0.3.jar:/usr/lib/ambari-metrics-collector/phoenix-core-4.4.0.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/stax-api-1.0-2.jar:/usr/lib/ambari-metrics-collector/jsp-api-2.1-6.1.14.jar:/usr/lib/ambari-metrics-collector/hbase-procedure-1.1.2.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/hadoop-common-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/commons-cli-1.2.jar:/usr/lib/ambari-metrics-collector/commons-beanutils-core-1.8.0.jar:/usr/lib/ambari-metrics-collector/httpcore-4.2.5.jar:/usr/lib/ambari-metrics-collector/avro-1.7.4.jar:/usr/lib/ambari-metrics-collector/hadoop-auth-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/netty-all-4.0.23.Final.jar:/usr/lib/ambari-metrics-collector/hadoop-mapreduce-client-core-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/api-asn1-api-1.0.0-M20.jar:/usr/lib/ambari-metrics-collector/hadoop-yarn-server-nodemanager-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/jersey-guice-1.11.jar:/usr/lib/ambari-metrics-collector/jersey-core-1.11.jar:/usr/lib/ambari-metrics-collector/log4j-1.2.17.jar:/usr/lib/ambari-metrics-collector/leveldbjni-all-1.8.jar:/usr/lib/ambari-metrics-collector/xz-1.0.jar:/usr/lib/ambari-metrics-collector/jdk.tools-1.7.jar:/usr/lib/ambari-metrics-collector/slf4j-api-1.7.2.jar:/usr/lib/ambari-metrics-collector/commons-math-2.2.jar:/usr/lib/ambari-metrics-collector/zookeeper-3.4.6.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/hadoop-yarn-server-common-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/hadoop-mapreduce-client-common-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/xml-apis-1.4.01.jar:/usr/lib/ambari-metrics-collector/commons-net-3.1.jar:/usr/lib/ambari-metrics-collector/commons-digester-1.8.jar:/usr/lib/ambari-metrics-collector/commons-configuration-1.6.jar:/usr/lib/ambari-metrics-collector/jersey-servlet-1.11.jar:/usr/lib/ambari-metrics-collector/antlr-3.5.jar:/usr/lib/ambari-metrics-collector/annotations-1.3.2.jar:/usr/lib/ambari-metrics-collector/hadoop-mapreduce-client-app-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/jamon-runtime-2.3.1.jar:/usr/lib/ambari-metrics-collector/paranamer-2.3.jar:/usr/lib/ambari-metrics-collector/jackson-core-asl-1.9.9.jar:/usr/lib/ambari-metrics-collector/okio-1.4.0.jar:/usr/lib/ambari-metrics-collector/ST4-4.0.7.jar:/usr/lib/ambari-metrics-collector/jcodings-1.0.8.jar:/usr/lib/ambari-metrics-collector/commons-collections-3.2.1.jar:/usr/lib/ambari-metrics-collector/hbase-server-1.1.2.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/jackson-core-2.2.3.jar:/usr/lib/ambari-metrics-collector/commons-daemon-1.0.13.jar:/usr/lib/ambari-metrics-collector/azure-storage-2.2.0.jar:/usr/lib/ambari-metrics-collector/javax.inject-1.jar:/usr/lib/ambari-metrics-collector/commons-codec-1.7.jar:/usr/lib/ambari-metrics-collector/jets3t-0.9.0.jar:/usr/lib/ambari-metrics-collector/guice-3.0.jar:/usr/lib/ambari-metrics-collector/guava-14.0.1.jar:/usr/lib/ambari-metrics-collector/hadoop-hdfs-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/jsch-0.1.42.jar:/usr/lib/ambari-metrics-collector/jaxb-impl-2.2.3-1.jar:/usr/lib/ambari-metrics-collector/xercesImpl-2.11.0.jar:/usr/lib/ambari-metrics-collector/hbase-common-1.1.2.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/protobuf-java-2.5.0.jar:/usr/lib/ambari-metrics-collector/jsr305-2.0.1.jar:/usr/lib/ambari-metrics-collector/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/lib/ambari-metrics-collector/commons-lang3-3.3.2.jar:/usr/lib/ambari-metrics-collector/snappy-java-1.0.4.1.jar:/usr/lib/ambari-metrics-collector/java-xmlbuilder-0.4.jar:/usr/lib/ambari-metrics-collector/hbase-common-1.1.2.2.4.0.0-170-tests.jar:/usr/lib/ambari-metrics-collector/microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/lib/ambari-metrics-collector/hbase-annotations-1.1.2.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/high-scale-lib-1.1.1.jar:/usr/lib/ambari-metrics-collector/hbase-protocol-1.1.2.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/xmlenc-0.52.jar:/usr/lib/ambari-metrics-collector/hadoop-client-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/gson-2.2.4.jar:/usr/lib/ambari-metrics-collector/curator-framework-2.7.1.jar:/usr/lib/ambari-metrics-collector/hadoop-yarn-client-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/antlr-runtime-3.5.jar:/usr/lib/ambari-metrics-collector/hadoop-azure-2.7.1.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/commons-math3-3.4.1.jar:/usr/lib/ambari-metrics-collector/jsp-2.1-6.1.14.jar:/usr/lib/ambari-metrics-collector/hbase-hadoop2-compat-1.1.2.2.4.0.0-170.jar:/usr/lib/ambari-metrics-collector/sqlline-1.1.8.jar:/usr/lib/ambari-metrics-collector/metrics-core-2.2.0.jar:/etc/ambari-metrics-collector/conf 2017-06-06 05:35:07,343 INFO org.apache.zookeeper.ZooKeeper: Client environment:java.library.path=/usr/lib/ams-hbase/lib/hadoop-native -XX:+UseConcMarkSweepGC -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCDateStamps -Xloggc:/var/log/ambari-metrics-collector/collector-gc.log-201706061335 2017-06-06 05:35:07,343 INFO org.apache.zookeeper.ZooKeeper: Client environment:java.io.tmpdir=/tmp 2017-06-06 05:35:07,343 INFO org.apache.zookeeper.ZooKeeper: Client environment:java.compiler= 2017-06-06 05:35:07,343 INFO org.apache.zookeeper.ZooKeeper: Client environment:os.name=Linux 2017-06-06 05:35:07,343 INFO org.apache.zookeeper.ZooKeeper: Client environment:os.arch=amd64 2017-06-06 05:35:07,343 INFO org.apache.zookeeper.ZooKeeper: Client environment:os.version=4.4.0-64-generic 2017-06-06 05:35:07,343 INFO org.apache.zookeeper.ZooKeeper: Client environment:user.name=ams 2017-06-06 05:35:07,344 INFO org.apache.zookeeper.ZooKeeper: Client environment:user.home=/home/ams 2017-06-06 05:35:07,344 INFO org.apache.zookeeper.ZooKeeper: Client environment:user.dir=/home/ams 2017-06-06 05:35:07,344 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0x5524cca10x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:35:07,381 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:35:07,390 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:35:07,511 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d610001, negotiated timeout = 120000 2017-06-06 05:35:07,518 INFO org.apache.hadoop.hbase.client.ZooKeeperRegistry: ClusterId read in ZooKeeper is null 2017-06-06 05:35:07,553 INFO org.apache.phoenix.metrics.Metrics: Initializing metrics system: phoenix 2017-06-06 05:35:07,588 WARN org.apache.hadoop.metrics2.impl.MetricsConfig: Cannot locate configuration: tried hadoop-metrics2-phoenix.properties,hadoop-metrics2.properties 2017-06-06 05:35:07,744 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot period at 10 second(s). 2017-06-06 05:35:07,745 INFO org.apache.hadoop.metrics2.impl.MetricsSystemImpl: phoenix metrics system started 2017-06-06 05:35:08,021 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:35:08,152 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:35:08,386 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x77a98a6a connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:35:08,386 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0x77a98a6a0x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:35:08,387 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:35:08,387 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:35:08,522 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d610003, negotiated timeout = 120000 2017-06-06 05:35:08,523 INFO org.apache.hadoop.hbase.client.ZooKeeperRegistry: ClusterId read in ZooKeeper is null 2017-06-06 05:35:08,586 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:35:28,349 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing master protocol: MasterService 2017-06-06 05:35:28,349 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d610003 2017-06-06 05:35:28,373 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d610003 closed 2017-06-06 05:35:28,373 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:35:28,386 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d610001 2017-06-06 05:35:28,409 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d610001 closed 2017-06-06 05:35:28,409 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:35:28,420 WARN org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource: Unable to connect to HBase store using Phoenix. org.apache.phoenix.schema.TableNotFoundException: ERROR 1012 (42M03): Table undefined. tableName=SYSTEM.CATALOG at org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:436) at org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:939) at org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:860) at org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1174) at org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:112) at org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1974) at org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:770) at org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:186) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:305) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:297) at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53) at org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:295) at org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1244) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1850) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:77) at org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:180) at org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:132) at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:151) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:270) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource.getConnection(DefaultPhoenixDataSource.java:69) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnection(PhoenixHBaseAccessor.java:238) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnectionRetryingOnException(PhoenixHBaseAccessor.java:216) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.initMetricSchema(PhoenixHBaseAccessor.java:258) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.initializeSubsystem(HBaseTimelineMetricStore.java:83) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.serviceInit(HBaseTimelineMetricStore.java:76) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:83) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:136) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:146) 2017-06-06 05:35:33,460 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:35:33,461 INFO org.apache.hadoop.conf.Configuration.deprecation: io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 2017-06-06 05:35:33,486 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:35:33,487 INFO org.apache.hadoop.conf.Configuration.deprecation: io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 2017-06-06 05:35:33,491 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x345e5a17 connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:35:33,491 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0x345e5a170x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:35:33,492 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:35:33,493 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:35:33,510 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d610008, negotiated timeout = 120000 2017-06-06 05:35:33,567 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x2102a4d5 connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:35:33,567 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0x2102a4d50x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:35:33,569 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:35:33,569 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:35:33,644 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d610009, negotiated timeout = 120000 2017-06-06 05:35:33,664 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing master protocol: MasterService 2017-06-06 05:35:33,664 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d610009 2017-06-06 05:35:34,162 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d610009 closed 2017-06-06 05:35:34,162 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:35:34,173 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d610008 2017-06-06 05:35:34,928 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d610008 closed 2017-06-06 05:35:34,928 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:35:34,939 WARN org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource: Unable to connect to HBase store using Phoenix. org.apache.phoenix.schema.TableNotFoundException: ERROR 1012 (42M03): Table undefined. tableName=SYSTEM.CATALOG at org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:436) at org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:939) at org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:860) at org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1174) at org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:112) at org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1974) at org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:770) at org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:186) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:305) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:297) at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53) at org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:295) at org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1244) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1850) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:77) at org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:180) at org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:132) at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:151) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:270) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource.getConnection(DefaultPhoenixDataSource.java:69) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnection(PhoenixHBaseAccessor.java:238) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnectionRetryingOnException(PhoenixHBaseAccessor.java:216) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.initMetricSchema(PhoenixHBaseAccessor.java:258) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.initializeSubsystem(HBaseTimelineMetricStore.java:83) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.serviceInit(HBaseTimelineMetricStore.java:76) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:83) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:136) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:146) 2017-06-06 05:35:44,968 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:35:44,969 INFO org.apache.hadoop.conf.Configuration.deprecation: io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 2017-06-06 05:35:44,991 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:35:44,993 INFO org.apache.hadoop.conf.Configuration.deprecation: io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 2017-06-06 05:35:44,996 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x22db8f4 connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:35:44,996 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0x22db8f40x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:35:44,997 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:35:44,998 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:35:45,847 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d61000a, negotiated timeout = 120000 2017-06-06 05:35:45,894 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x29caf222 connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:35:45,894 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0x29caf2220x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:35:45,896 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:35:45,901 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:35:46,340 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d61000b, negotiated timeout = 120000 2017-06-06 05:35:46,357 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing master protocol: MasterService 2017-06-06 05:35:46,357 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d61000b 2017-06-06 05:35:46,407 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d61000b closed 2017-06-06 05:35:46,408 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:35:46,418 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d61000a 2017-06-06 05:35:46,456 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d61000a closed 2017-06-06 05:35:46,456 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:35:46,467 WARN org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource: Unable to connect to HBase store using Phoenix. org.apache.phoenix.schema.TableNotFoundException: ERROR 1012 (42M03): Table undefined. tableName=SYSTEM.CATALOG at org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:436) at org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:939) at org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:860) at org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1174) at org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:112) at org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1974) at org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:770) at org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:186) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:305) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:297) at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53) at org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:295) at org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1244) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1850) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:77) at org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:180) at org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:132) at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:151) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:270) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource.getConnection(DefaultPhoenixDataSource.java:69) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnection(PhoenixHBaseAccessor.java:238) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnectionRetryingOnException(PhoenixHBaseAccessor.java:216) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.initMetricSchema(PhoenixHBaseAccessor.java:258) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.initializeSubsystem(HBaseTimelineMetricStore.java:83) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.serviceInit(HBaseTimelineMetricStore.java:76) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:83) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:136) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:146) 2017-06-06 05:36:06,498 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:36:06,499 INFO org.apache.hadoop.conf.Configuration.deprecation: io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 2017-06-06 05:36:06,523 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:36:06,525 INFO org.apache.hadoop.conf.Configuration.deprecation: io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 2017-06-06 05:36:06,529 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x28a2a3e7 connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:36:06,529 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0x28a2a3e70x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:36:06,530 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:36:06,531 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:36:06,990 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d61000c, negotiated timeout = 120000 2017-06-06 05:36:07,037 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0xea27e34 connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:36:07,038 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0xea27e340x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:36:07,039 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:36:07,039 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:36:07,051 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d61000d, negotiated timeout = 120000 2017-06-06 05:36:07,071 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing master protocol: MasterService 2017-06-06 05:36:07,071 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d61000d 2017-06-06 05:36:07,080 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d61000d closed 2017-06-06 05:36:07,080 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:36:07,091 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d61000c 2017-06-06 05:36:07,101 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d61000c closed 2017-06-06 05:36:07,101 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:36:07,112 WARN org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource: Unable to connect to HBase store using Phoenix. org.apache.phoenix.schema.TableNotFoundException: ERROR 1012 (42M03): Table undefined. tableName=SYSTEM.CATALOG at org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:436) at org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:939) at org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:860) at org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1174) at org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:112) at org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1974) at org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:770) at org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:186) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:305) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:297) at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53) at org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:295) at org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1244) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1850) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:77) at org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:180) at org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:132) at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:151) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:270) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource.getConnection(DefaultPhoenixDataSource.java:69) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnection(PhoenixHBaseAccessor.java:238) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnectionRetryingOnException(PhoenixHBaseAccessor.java:216) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.initMetricSchema(PhoenixHBaseAccessor.java:258) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.initializeSubsystem(HBaseTimelineMetricStore.java:83) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.serviceInit(HBaseTimelineMetricStore.java:76) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:83) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:136) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:146) 2017-06-06 05:36:47,141 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:36:47,141 INFO org.apache.hadoop.conf.Configuration.deprecation: io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 2017-06-06 05:36:47,177 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:36:47,178 INFO org.apache.hadoop.conf.Configuration.deprecation: io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 2017-06-06 05:36:47,180 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x22ee2d0 connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:36:47,181 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0x22ee2d00x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:36:47,182 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:36:47,183 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:36:47,199 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d61000e, negotiated timeout = 120000 2017-06-06 05:36:47,220 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x53bc1328 connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:36:47,220 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0x53bc13280x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:36:47,221 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:36:47,221 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:36:47,229 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d61000f, negotiated timeout = 120000 2017-06-06 05:36:47,241 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing master protocol: MasterService 2017-06-06 05:36:47,241 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d61000f 2017-06-06 05:36:47,251 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d61000f closed 2017-06-06 05:36:47,251 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:36:47,262 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d61000e 2017-06-06 05:36:47,270 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d61000e closed 2017-06-06 05:36:47,271 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:36:47,281 WARN org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource: Unable to connect to HBase store using Phoenix. org.apache.phoenix.schema.TableNotFoundException: ERROR 1012 (42M03): Table undefined. tableName=SYSTEM.CATALOG at org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:436) at org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:939) at org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:860) at org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1174) at org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:112) at org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1974) at org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:770) at org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:186) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:305) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:297) at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53) at org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:295) at org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1244) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1850) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:77) at org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:180) at org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:132) at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:151) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:270) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource.getConnection(DefaultPhoenixDataSource.java:69) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnection(PhoenixHBaseAccessor.java:238) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnectionRetryingOnException(PhoenixHBaseAccessor.java:216) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.initMetricSchema(PhoenixHBaseAccessor.java:258) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.initializeSubsystem(HBaseTimelineMetricStore.java:83) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.serviceInit(HBaseTimelineMetricStore.java:76) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:83) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:136) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:146) 2017-06-06 05:38:07,303 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:38:07,304 INFO org.apache.hadoop.conf.Configuration.deprecation: io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 2017-06-06 05:38:07,324 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:38:07,325 INFO org.apache.hadoop.conf.Configuration.deprecation: io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 2017-06-06 05:38:07,327 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0xe362c57 connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:38:07,327 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0xe362c570x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:38:07,329 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:38:07,329 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:38:07,345 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d610010, negotiated timeout = 120000 2017-06-06 05:38:07,367 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x5aa360ea connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:38:07,367 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0x5aa360ea0x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:38:07,368 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:38:07,368 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:38:07,379 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d610011, negotiated timeout = 120000 2017-06-06 05:38:07,399 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing master protocol: MasterService 2017-06-06 05:38:07,399 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d610011 2017-06-06 05:38:07,403 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d610011 closed 2017-06-06 05:38:07,403 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:38:07,414 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d610010 2017-06-06 05:38:07,420 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d610010 closed 2017-06-06 05:38:07,420 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:38:07,430 WARN org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource: Unable to connect to HBase store using Phoenix. org.apache.phoenix.schema.TableNotFoundException: ERROR 1012 (42M03): Table undefined. tableName=SYSTEM.CATALOG at org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:436) at org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:939) at org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:860) at org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1174) at org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:112) at org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1974) at org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:770) at org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:186) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:305) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:297) at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53) at org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:295) at org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1244) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1850) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:77) at org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:180) at org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:132) at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:151) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:270) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource.getConnection(DefaultPhoenixDataSource.java:69) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnection(PhoenixHBaseAccessor.java:238) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnectionRetryingOnException(PhoenixHBaseAccessor.java:216) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.initMetricSchema(PhoenixHBaseAccessor.java:258) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.initializeSubsystem(HBaseTimelineMetricStore.java:83) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.serviceInit(HBaseTimelineMetricStore.java:76) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:83) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:136) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:146) 2017-06-06 05:40:47,452 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:40:47,453 INFO org.apache.hadoop.conf.Configuration.deprecation: io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 2017-06-06 05:40:47,470 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:40:47,471 INFO org.apache.hadoop.conf.Configuration.deprecation: io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 2017-06-06 05:40:47,473 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x5987e932 connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:40:47,473 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0x5987e9320x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:40:47,474 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:40:47,475 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:40:47,818 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d610012, negotiated timeout = 120000 2017-06-06 05:40:47,842 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x438bad7c connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:40:47,843 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0x438bad7c0x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:40:47,844 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:40:47,845 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:40:47,910 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d610013, negotiated timeout = 120000 2017-06-06 05:40:47,926 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing master protocol: MasterService 2017-06-06 05:40:47,926 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d610013 2017-06-06 05:40:47,949 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d610013 closed 2017-06-06 05:40:47,949 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:40:47,960 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d610012 2017-06-06 05:40:48,057 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d610012 closed 2017-06-06 05:40:48,057 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:40:48,068 WARN org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource: Unable to connect to HBase store using Phoenix. org.apache.phoenix.schema.TableNotFoundException: ERROR 1012 (42M03): Table undefined. tableName=SYSTEM.CATALOG at org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:436) at org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:939) at org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:860) at org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1174) at org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:112) at org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1974) at org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:770) at org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:186) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:305) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:297) at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53) at org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:295) at org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1244) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1850) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:77) at org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:180) at org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:132) at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:151) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:270) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource.getConnection(DefaultPhoenixDataSource.java:69) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnection(PhoenixHBaseAccessor.java:238) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnectionRetryingOnException(PhoenixHBaseAccessor.java:216) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.initMetricSchema(PhoenixHBaseAccessor.java:258) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.initializeSubsystem(HBaseTimelineMetricStore.java:83) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.serviceInit(HBaseTimelineMetricStore.java:76) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:83) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:136) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:146) 2017-06-06 05:46:08,095 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:46:08,096 INFO org.apache.hadoop.conf.Configuration.deprecation: io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 2017-06-06 05:46:08,117 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:46:08,118 INFO org.apache.hadoop.conf.Configuration.deprecation: io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 2017-06-06 05:46:08,120 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x2472c7d8 connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:46:08,120 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0x2472c7d80x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:46:08,122 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:46:08,123 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:46:08,144 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d610014, negotiated timeout = 120000 2017-06-06 05:46:08,176 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x236ab296 connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:46:08,176 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0x236ab2960x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:46:08,178 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:46:08,179 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:46:08,195 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d610015, negotiated timeout = 120000 2017-06-06 05:46:08,215 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing master protocol: MasterService 2017-06-06 05:46:08,215 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d610015 2017-06-06 05:46:08,223 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d610015 closed 2017-06-06 05:46:08,223 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:46:08,233 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d610014 2017-06-06 05:46:08,240 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d610014 closed 2017-06-06 05:46:08,240 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:46:08,251 WARN org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource: Unable to connect to HBase store using Phoenix. org.apache.phoenix.schema.TableNotFoundException: ERROR 1012 (42M03): Table undefined. tableName=SYSTEM.CATALOG at org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:436) at org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:939) at org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:860) at org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1174) at org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:112) at org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1974) at org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:770) at org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:186) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:305) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:297) at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53) at org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:295) at org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1244) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1850) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:77) at org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:180) at org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:132) at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:151) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:270) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource.getConnection(DefaultPhoenixDataSource.java:69) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnection(PhoenixHBaseAccessor.java:238) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnectionRetryingOnException(PhoenixHBaseAccessor.java:216) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.initMetricSchema(PhoenixHBaseAccessor.java:258) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.initializeSubsystem(HBaseTimelineMetricStore.java:83) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.serviceInit(HBaseTimelineMetricStore.java:76) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:83) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:136) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:146) 2017-06-06 05:56:48,281 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:56:48,282 INFO org.apache.hadoop.conf.Configuration.deprecation: io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 2017-06-06 05:56:48,305 WARN org.apache.hadoop.hbase.io.util.HeapMemorySizeUtil: hbase.regionserver.global.memstore.upperLimit is deprecated by hbase.regionserver.global.memstore.size 2017-06-06 05:56:48,306 INFO org.apache.hadoop.conf.Configuration.deprecation: io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 2017-06-06 05:56:48,309 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0xa68df9 connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:56:48,309 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0xa68df90x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:56:48,311 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:56:48,311 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:56:48,322 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d610016, negotiated timeout = 120000 2017-06-06 05:56:48,344 INFO org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x2f6bbeb0 connecting to ZooKeeper ensemble=localhost:61181 2017-06-06 05:56:48,344 INFO org.apache.zookeeper.ZooKeeper: Initiating client connection, connectString=localhost:61181 sessionTimeout=120000 watcher=hconnection-0x2f6bbeb00x0, quorum=localhost:61181, baseZNode=/ams-hbase-unsecure 2017-06-06 05:56:48,346 INFO org.apache.zookeeper.ClientCnxn: Opening socket connection to server localhost/127.0.0.1:61181. Will not attempt to authenticate using SASL (unknown error) 2017-06-06 05:56:48,346 INFO org.apache.zookeeper.ClientCnxn: Socket connection established to localhost/127.0.0.1:61181, initiating session 2017-06-06 05:56:48,359 INFO org.apache.zookeeper.ClientCnxn: Session establishment complete on server localhost/127.0.0.1:61181, sessionid = 0x15c7be63d610017, negotiated timeout = 120000 2017-06-06 05:56:48,373 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing master protocol: MasterService 2017-06-06 05:56:48,373 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d610017 2017-06-06 05:56:48,384 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d610017 closed 2017-06-06 05:56:48,384 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:56:48,394 INFO org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x15c7be63d610016 2017-06-06 05:56:48,404 INFO org.apache.zookeeper.ZooKeeper: Session: 0x15c7be63d610016 closed 2017-06-06 05:56:48,404 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down 2017-06-06 05:56:48,415 WARN org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource: Unable to connect to HBase store using Phoenix. org.apache.phoenix.schema.TableNotFoundException: ERROR 1012 (42M03): Table undefined. tableName=SYSTEM.CATALOG at org.apache.phoenix.query.ConnectionQueryServicesImpl.getAllTableRegions(ConnectionQueryServicesImpl.java:436) at org.apache.phoenix.query.ConnectionQueryServicesImpl.checkClientServerCompatibility(ConnectionQueryServicesImpl.java:939) at org.apache.phoenix.query.ConnectionQueryServicesImpl.ensureTableCreated(ConnectionQueryServicesImpl.java:860) at org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1174) at org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:112) at org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1974) at org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:770) at org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:186) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:305) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:297) at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53) at org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:295) at org.apache.phoenix.jdbc.PhoenixStatement.executeUpdate(PhoenixStatement.java:1244) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1850) at org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:77) at org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:1819) at org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:180) at org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.connect(PhoenixEmbeddedDriver.java:132) at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:151) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:270) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.query.DefaultPhoenixDataSource.getConnection(DefaultPhoenixDataSource.java:69) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnection(PhoenixHBaseAccessor.java:238) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.getConnectionRetryingOnException(PhoenixHBaseAccessor.java:216) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.PhoenixHBaseAccessor.initMetricSchema(PhoenixHBaseAccessor.java:258) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.initializeSubsystem(HBaseTimelineMetricStore.java:83) at org.apache.hadoop.yarn.server.applicationhistoryservice.metrics.timeline.HBaseTimelineMetricStore.serviceInit(HBaseTimelineMetricStore.java:76) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.serviceInit(ApplicationHistoryServer.java:83) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.launchAppHistoryServer(ApplicationHistoryServer.java:136) at org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryServer.main(ApplicationHistoryServer.java:146)