Member since
03-09-2016
91
Posts
3
Kudos Received
1
Solution
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1157 | 10-26-2018 09:52 AM |
07-28-2017
11:44 PM
Labels:
- Labels:
-
Hortonworks Data Platform (HDP)
06-14-2017
02:58 AM
Log file: 2017-06-12 23:06:29,603 ERROR
[regionserver/sddsvrwm383.scglobaluat.aduat.com/172.25.12.67:16020]
zookeeper.RecoverableZooKeeper: ZooKeeper getChildren failed after 7 attempts 2017-06-12 23:06:29,603 WARN
[regionserver/sddsvrwm383.scglobaluat.aduat. com/172.25.12.67:16020]
zookeeper.ZKUtil: regionserver:16020-0x35c93984f940fd2,
quorum=sddsvrwm369.scglobaluat.aduat. com:2181,sddsvrwm367.scglobaluat.aduat. com:2181,sddsvrwm368.scglobaluat.aduat.com:2181,
baseZNode=/hbase-secure Unable to list children of znode
/hbase-secure/replication/rs/sddsvrwm383.scglobaluat.aduat. com,16020,1497136332052 org.apache.zookeeper.KeeperException$SessionExpiredException:
KeeperErrorCode = Session expired for
/hbase-secure/replication/rs/sddsvrwm383.scglobaluat.aduat. com,16020,1497136332052 at
org.apache.zookeeper.KeeperException.create(KeeperException.java:127) at
org.apache.zookeeper.KeeperException.create(KeeperException.java:51) at
org.apache.zookeeper.ZooKeeper.getChildren(ZooKeeper.java:1472) at
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.getChildren(RecoverableZooKeeper.java:295) at
org.apache.hadoop.hbase.zookeeper.ZKUtil.listChildrenAndWatchForNewChildren(ZKUtil.java:455) at
org.apache.hadoop.hbase.zookeeper.ZKUtil.listChildrenAndWatchThem(ZKUtil.java:483) at
org.apache.hadoop.hbase.zookeeper.ZKUtil.listChildrenBFSAndWatchThem(ZKUtil.java:1462) at
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeRecursivelyMultiOrSequential(ZKUtil.java:1384) at
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeRecursively(ZKUtil.java:1266) at
org.apache.hadoop.hbase.replication.ReplicationQueuesZKImpl.removeAllQueues(ReplicationQueuesZKImpl.java:196) at
org.apache.hadoop.hbase.replication.regionserver.ReplicationSourceManager.join(ReplicationSourceManager.java:302) at org.apache.hadoop.hbase.replication.regionserver.Replication.join(Replication.java:202) at
org.apache.hadoop.hbase.replication.regionserver.Replication.stopReplicationService(Replication.java:194) at
org.apache.hadoop.hbase.regionserver.HRegionServer.stopServiceThreads(HRegionServer.java:2163) at
org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:1090) at
java.lang.Thread.run(Thread.java:745) 2017-06-12 23:06:29,604 ERROR
[regionserver/sddsvrwm383.scglobaluat.aduat.com/172.25.12.67:16020]
zookeeper.ZooKeeperWatcher: regionserver:16020-0x35c93984f940fd2,
quorum=sddsvrwm369.scglobaluat.aduat.com:2181,sddsvrwm367.scglobaluat.aduat.com:2181,sddsvrwm368.scglobaluat.aduat.com:2181,
baseZNode=/hbase-secure Received unexpected KeeperException, re-throwing
exception org.apache.zookeeper.KeeperException$SessionExpiredException:
KeeperErrorCode = Session expired for
/hbase-secure/replication/rs/sddsvrwm383.scglobaluat.aduat. com,16020,1497136332052 at
org.apache.zookeeper.KeeperException.create(KeeperException.java:127) at
org.apache.zookeeper.KeeperException.create(KeeperException.java:51) at
org.apache.zookeeper.ZooKeeper.getChildren(ZooKeeper.java:1472) at
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.getChildren(RecoverableZooKeeper.java:295) at
org.apache.hadoop.hbase.zookeeper.ZKUtil.listChildrenAndWatchForNewChildren(ZKUtil.java:455) at org.apache.hadoop.hbase.zookeeper.ZKUtil.listChildrenAndWatchThem(ZKUtil.java:483) at
org.apache.hadoop.hbase.zookeeper.ZKUtil.listChildrenBFSAndWatchThem(ZKUtil.java:1462) at
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeRecursivelyMultiOrSequential(ZKUtil.java:1384) at
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNodeRecursively(ZKUtil.java:1266) at
org.apache.hadoop.hbase.replication.ReplicationQueuesZKImpl.removeAllQueues(ReplicationQueuesZKImpl.java:196) at
org.apache.hadoop.hbase.replication.regionserver.ReplicationSourceManager.join(ReplicationSourceManager.java:302) at
org.apache.hadoop.hbase.replication.regionserver.Replication.join(Replication.java:202) at
org.apache.hadoop.hbase.replication.regionserver.Replication.stopReplicationService(Replication.java:194) at
org.apache.hadoop.hbase.regionserver.HRegionServer.stopServiceThreads(HRegionServer.java:2163) at
org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:1090) at
java.lang.Thread.run(Thread.java:745) 2017-06-12 23:06:29,605 INFO
[regionserver/sddsvrwm383.scglobaluat.aduat.com/172.25.12.67:16020]
ipc.RpcServer: Stopping server on 16020 2017-06-12 23:06:29,606 INFO [regionserver/sddsvrwm383.scglobaluat.aduat. com/172.25.12.67:16020]
token.AuthenticationTokenSecretManager: Stopping leader election, because:
SecretManager stopping 2017-06-12 23:06:29,607 INFO
[RpcServer.listener,port=16020] ipc.RpcServer:
RpcServer.listener,port=16020: stopping 2017-06-12 23:06:29,614 INFO
[RpcServer.responder] ipc.RpcServer: RpcServer.responder: stopped 2017-06-12 23:06:29,614 INFO
[RpcServer.responder] ipc.RpcServer: RpcServer.responder: stopping 2017-06-12 23:06:29,619 WARN
[regionserver/sddsvrwm383.scglobaluat.aduat.com/172.25.12.67:16020]
zookeeper.RecoverableZooKeeper: Possibly transient ZooKeeper,
quorum=sddsvrwm369.scglobaluat.aduat. com:2181,sddsvrwm367.scglobaluat.aduat. com:2181,sddsvrwm368.scglobaluat.aduat.
com:2181,
exception=org.apache.zookeeper.KeeperException$SessionExpiredException:
KeeperErrorCode = Session expired for
/hbase-secure/rs/sddsvrwm383.scglobaluat.aduat. com,16020,1497136332052 2017-06-12 23:07:29,625 INFO
[HBase-Metrics2-1] impl.MetricsSystemImpl: Stopping HBase metrics
system... 2017-06-12 23:07:29,625 INFO
[timeline] impl.MetricsSinkAdapter: timeline thread interrupted. 2017-06-12 23:07:29,628 INFO
[HBase-Metrics2-1] impl.MetricsSystemImpl: HBase metrics system stopped. 2017-06-12 23:07:29,628 INFO
[pool-684-thread-1] timeline.HadoopTimelineMetricsSink: Closing
HadoopTimelineMetricSink. Flushing metrics to collector... 2017-06-12 23:07:30,132 INFO
[HBase-Metrics2-1] impl.MetricsConfig: loaded properties from
hadoop-metrics2-hbase.properties 2017-06-12 23:07:30,148 INFO
[HBase-Metrics2-1] timeline.HadoopTimelineMetricsSink: Initializing
Timeline metrics sink. 2017-06-12 23:07:30,148 INFO
[HBase-Metrics2-1] timeline.HadoopTimelineMetricsSink: Identified
hostname = sddsvrwm383.scglobaluat.aduat.com, serviceName = hbase 2017-06-12 23:07:30,148 INFO
[HBase-Metrics2-1] timeline.HadoopTimelineMetricsSink: Collector Uri:
http://sddsvrwm368.scglobaluat.aduat. com:6188/ws/v1/timeline/metrics 2017-06-12 23:07:30,151 INFO
[HBase-Metrics2-1] impl.MetricsSinkAdapter: Sink timeline started 2017-06-12 23:07:30,151 INFO
[HBase-Metrics2-1] impl.MetricsSystemImpl: Scheduled snapshot period at
10 second(s). 2017-06-12 23:07:30,151 INFO
[HBase-Metrics2-1] impl.MetricsSystemImpl: HBase metrics system started 2017-06-12 23:07:32,620 WARN
[regionserver/sddsvrwm383.scglobaluat.aduat. com/172.25.12.67:16020]
zookeeper.RecoverableZooKeeper: Possibly transient ZooKeeper,
quorum=sddsvrwm369.scglobaluat.aduat. com:2181,sddsvrwm367.scglobaluat.aduat. com:2181,sddsvrwm368.scglobaluat.aduat.com:2181,
exception=org.apache.zookeeper.KeeperException$SessionExpiredException: KeeperErrorCode
= Session expired for /hbase-secure/rs/sddsvrwm383.scglobaluat.aduat. com,16020,1497136332052 2017-06-12 23:08:36,621 WARN
[regionserver/sddsvrwm383.scglobaluat.aduat. com/172.25.12.67:16020]
zookeeper.RecoverableZooKeeper: Possibly transient ZooKeeper,
quorum=sddsvrwm369.scglobaluat.aduat. com:2181,sddsvrwm367.scglobaluat.aduat. com:2181,sddsvrwm368.scglobaluat.aduat.
com:2181, exception=org.apache.zookeeper.KeeperException$SessionExpiredException:
KeeperErrorCode = Session expired for
/hbase-secure/rs/sddsvrwm383.scglobaluat.aduat. com,16020,1497136332052 2017-06-12 23:08:36,621 ERROR
[regionserver/sddsvrwm383.scglobaluat.aduat. com/172.25.12.67:16020]
zookeeper.RecoverableZooKeeper: ZooKeeper delete failed after 7 attempts 2017-06-12 23:08:36,621 WARN
[regionserver/sddsvrwm383.scglobaluat.aduat. com/172.25.12.67:16020]
regionserver.HRegionServer: Failed deleting my ephemeral node org.apache.zookeeper.KeeperException$SessionExpiredException:
KeeperErrorCode = Session expired for
/hbase-secure/rs/sddsvrwm383.scglobaluat.aduat. com,16020,1497136332052 at
org.apache.zookeeper.KeeperException.create(KeeperException.java:127) at
org.apache.zookeeper.KeeperException.create(KeeperException.java:51) at
org.apache.zookeeper.ZooKeeper.delete(ZooKeeper.java:873) at
org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper.delete(RecoverableZooKeeper.java:178) at
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1222) at
org.apache.hadoop.hbase.zookeeper.ZKUtil.deleteNode(ZKUtil.java:1211) at
org.apache.hadoop.hbase.regionserver.HRegionServer.deleteMyEphemeralNode(HRegionServer.java:1427) at
org.apache.hadoop.hbase.regionserver.HRegionServer.run(HRegionServer.java:1098) at
java.lang.Thread.run(Thread.java:745) 2017-06-12 23:08:36,622 INFO
[regionserver/sddsvrwm383.scglobaluat.aduat. com/172.25.12.67:16020]
regionserver.HRegionServer: stopping server sddsvrwm383.scglobaluat.aduat. com,16020,1497136332052;
zookeeper connection closed. 2017-06-12 23:08:36,622 INFO
[regionserver/sddsvrwm383.scglobaluat.aduat. com/172.25.12.67:16020]
regionserver.HRegionServer: regionserver/sddsvrwm383.scglobaluat.aduat. com/172.25.12.67:16020
exiting 2017-06-12 23:08:36,627 ERROR [main]
regionserver.HRegionServerCommandLine: Region server exiting java.lang.RuntimeException: HRegionServer Aborted at
org.apache.hadoop.hbase.regionserver.HRegionServerCommandLine.start(HRegionServerCommandLine.java:68) at
org.apache.hadoop.hbase.regionserver.HRegionServerCommandLine.run(HRegionServerCommandLine.java:87) at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76) at
org.apache.hadoop.hbase.util.ServerCommandLine.doMain(ServerCommandLine.java:126) at
org.apache.hadoop.hbase.regionserver.HRegionServer.main(HRegionServer.java:2681) 2017-06-12 23:08:36,633 INFO
[pool-5-thread-1] provider.AuditProviderFactory: ==>
JVMShutdownHook.run() 2017-06-12 23:08:36,633 INFO
[pool-5-thread-1] provider.AuditProviderFactory: JVMShutdownHook:
Signalling async audit cleanup to start. 2017-06-12 23:08:36,633 INFO
[pool-5-thread-1] provider.AuditProviderFactory: JVMShutdownHook:
Waiting up to 30 seconds for audit cleanup to finish. 2017-06-12 23:08:36,634 INFO
[Ranger async Audit cleanup] provider.AuditProviderFactory:
RangerAsyncAuditCleanup: Starting cleanup 2017-06-12 23:08:36,635 INFO
[Ranger async Audit cleanup] destination.HDFSAuditDestination: Flush
HDFS audit logs completed..... 2017-06-12 23:08:36,635 INFO
[Ranger async Audit cleanup] queue.AuditAsyncQueue: Stop called.
name=hbaseRegional.async 2017-06-12 23:08:36,635 INFO
[Ranger async Audit cleanup] queue.AuditAsyncQueue: Interrupting
consumerThread. name=hbaseRegional.async, consumer=hbaseRegional.async.summary 2017-06-12 23:08:36,635 INFO
[Ranger async Audit cleanup] provider.AuditProviderFactory: RangerAsyncAuditCleanup:
Done cleanup 2017-06-12 23:08:36,635 INFO
[Ranger async Audit cleanup] provider.AuditProviderFactory:
RangerAsyncAuditCleanup: Waiting to audit cleanup start signal 2017-06-12 23:08:36,635 INFO
[pool-5-thread-1] provider.AuditProviderFactory: JVMShutdownHook: Audit
cleanup finished after 2 milli seconds 2017-06-12 23:08:36,635 INFO
[pool-5-thread-1] provider.AuditProviderFactory: JVMShutdownHook:
Interrupting ranger async audit cleanup thread 2017-06-12 23:08:36,635 INFO
[pool-5-thread-1] provider.AuditProviderFactory: <==
JVMShutdownHook.run() 2017-06-12 23:08:36,635 INFO
[Ranger async Audit cleanup] provider.AuditProviderFactory:
RangerAsyncAuditCleanup: Interrupted while waiting for audit startCleanup
signal! Exiting the thread... java.lang.InterruptedException at
java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:998) at
java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304) at
java.util.concurrent.Semaphore.acquire(Semaphore.java:312) at
org.apache.ranger.audit.provider.AuditProviderFactory$RangerAsyncAuditCleanup.run(AuditProviderFactory.java:487) at
java.lang.Thread.run(Thread.java:745) 2017-06-12 23:08:36,636 INFO
[org.apache.ranger.audit.queue.AuditAsyncQueue0] queue.AuditAsyncQueue:
Caught exception in consumer thread. Shutdown might be in progress 2017-06-12 23:08:36,636 INFO
[org.apache.ranger.audit.queue.AuditAsyncQueue0] queue.AuditAsyncQueue:
Exiting polling loop. name=hbaseRegional.async 2017-06-12 23:08:36,636 INFO
[org.apache.ranger.audit.queue.AuditAsyncQueue0] queue.AuditAsyncQueue:
Calling to stop consumer. name=hbaseRegional.async,
consumer.name=hbaseRegional.async.summary 2017-06-12 23:08:36,636 INFO
[org.apache.ranger.audit.queue.AuditAsyncQueue0]
queue.AuditSummaryQueue: Stop called. name=hbaseRegional.async.summary 2017-06-12 23:08:36,636 INFO
[org.apache.ranger.audit.queue.AuditAsyncQueue0]
queue.AuditSummaryQueue: Interrupting consumerThread.
name=hbaseRegional.async.summary, consumer=hbaseRegional.async.summary.batch 2017-06-12 23:08:36,636 INFO
[org.apache.ranger.audit.queue.AuditAsyncQueue0] queue.AuditAsyncQueue:
Exiting consumerThread.run() method. name=hbaseRegional.async 2017-06-12 23:08:36,636 INFO
[org.apache.ranger.audit.queue.AuditSummaryQueue0]
queue.AuditSummaryQueue: Caught exception in consumer thread. Shutdown might be
in progress 2017-06-12 23:08:36,637 INFO
[org.apache.ranger.audit.queue.AuditSummaryQueue0]
queue.AuditSummaryQueue: Exiting polling loop. name=hbaseRegional.async.summary 2017-06-12 23:08:36,637 INFO
[org.apache.ranger.audit.queue.AuditSummaryQueue0] queue.AuditSummaryQueue:
Calling to stop consumer. name=hbaseRegional.async.summary,
consumer.name=hbaseRegional.async.summary.batch 2017-06-12 23:08:36,637 INFO
[org.apache.ranger.audit.queue.AuditSummaryQueue0]
queue.AuditBatchQueue: Stop called. name=hbaseRegional.async.summary.batch 2017-06-12 23:08:36,637 INFO
[pool-5-thread-1] regionserver.ShutdownHook: Shutdown hook starting;
hbase.shutdown.hook=true;
fsShutdownHook=org.apache.hadoop.fs.FileSystem$Cache$ClientFinalizer@4f3faa70 2017-06-12 23:08:36,637 INFO
[org.apache.ranger.audit.queue.AuditSummaryQueue0]
destination.HDFSAuditDestination: Flush HDFS audit logs completed..... 2017-06-12 23:08:36,637 INFO
[org.apache.ranger.audit.queue.AuditSummaryQueue0]
queue.AuditBatchQueue: Interrupting consumerThread. name=hbaseRegional.async.summary.batch,
consumer=hbaseRegional.async.summary.batch.hdfs 2017-06-12 23:08:36,637 INFO
[org.apache.ranger.audit.queue.AuditSummaryQueue0]
queue.AuditSummaryQueue: Exiting consumerThread.run() method.
name=hbaseRegional.async.summary name=hbaseRegional.async.summary 2017-06-12 23:08:36,638 INFO
[pool-5-thread-1] regionserver.ShutdownHook: Starting fs shutdown hook
thread. 2017-06-12 23:08:36,637 INFO
[org.apache.ranger.audit.queue.AuditBatchQueue0] queue.AuditBatchQueue:
Caught exception in consumer thread. Shutdown might be in progress 2017-06-12 23:08:36,638 INFO
[org.apache.ranger.audit.queue.AuditBatchQueue0] queue.AuditBatchQueue:
Exiting consumerThread. Queue=hbaseRegional.async.summary.batch,
dest=hbaseRegional.async.summary.batch.hdfs 2017-06-12 23:08:36,638 INFO
[org.apache.ranger.audit.queue.AuditBatchQueue0] queue.AuditBatchQueue:
Calling to stop consumer. name=hbaseRegional.async.summary.batch,
consumer.name=hbaseRegional.async.summary.batch.hdfs 2017-06-12 23:08:36,651 INFO
[org.apache.ranger.audit.queue.AuditBatchQueue0]
provider.BaseAuditHandler: Audit Status Log:
name=hbaseRegional.async.summary.batch.hdfs, interval=04:18.058 minutes,
events=2, succcessCount=2, totalEvents=12, totalSuccessCount=12 2017-06-12 23:08:36,651 INFO
[org.apache.ranger.audit.queue.AuditBatchQueue0] queue.AuditFileSpool:
Stop called, queueName=hbaseRegional.async.summary.batch,
consumer=hbaseRegional.async.summary.batch.hdfs 2017-06-12 23:08:36,652 INFO
[org.apache.ranger.audit.queue.AuditBatchQueue0] queue.AuditBatchQueue:
Exiting consumerThread.run() method. name=hbaseRegional.async.summary.batch 2017-06-12 23:08:36,651 INFO
[hbaseRegional.async.summary.batch_hbaseRegional.async.summary.batch.hdfs_destWriter]
queue.AuditFileSpool: Caught exception in consumer thread. Shutdown might be in
progress 2017-06-12 23:08:36,655 INFO
[pool-5-thread-1] regionserver.ShutdownHook: Shutdown hook finished.
... View more
Labels:
- Labels:
-
Apache HBase
12-09-2016
05:50 PM
can anyone please provide a docs or link for performance tuning of HDP.
... View more
Labels:
- Labels:
-
Hortonworks Data Platform (HDP)
11-09-2016
05:17 AM
@Savanna Endicott: I have done the HA rollback, please send the link of that docs, I am also rollback HA but got some error in the following command " curl --negotiate -u root:hashmap "X-Requested-By: ambari" -i -X POST -d '{"host_components" : [{"HostRoles":{"component_name":"navideh02.hash.net"}] }' http://localhost:8080/api/v1/clusters/NHDP/hosts?Hosts/host_name=navideh02.hash.net" please correct the above command
... View more
10-26-2016
08:07 PM
Thanks a lot , find it and and solve it ?
... View more
10-26-2016
07:45 PM
It appear four to six node and at end one datanode is dead (1) details of node --------------- ---------------
... View more
10-26-2016
07:40 PM
Thanks but still not display which one is down, any other command
... View more
10-26-2016
07:25 PM
How to find which datanode is dead via terminal or ambari
... View more
Labels:
- Labels:
-
Apache Ambari
09-30-2016
10:03 AM
Dear Gerd, I have set the above line in my.cnf file and restart mysqld but have same issue [root@hdp etc]# vi my.cnf
[root@hdp etc]# sudo systemctl restart mysqld [root@hdp /]# ambari-server start Using python /usr/bin/python2
Starting ambari-server
Ambari Server running with administrator privileges.
Organizing resource files at /var/lib/ambari-server/resources...
Server PID at: /var/run/ambari-server/ambari-server.pid
Server out at: /var/log/ambari-server/ambari-server.out
Server log at: /var/log/ambari-server/ambari-server.log
Waiting for server start....................
Ambari Server 'start' completed successfully. [root@hdp /]# ambari-server status Using python /usr/bin/python2
Ambari-server status
Ambari Server not running. Stale PID File at: /var/run/ambari-server/ambari-server.pid
... View more
09-28-2016
12:09 PM
log file: ambarilog.txt
... View more