Member since
07-09-2014
8
Posts
0
Kudos Received
0
Solutions
06-01-2017
08:37 PM
Thank you, I gonna verify it.
... View more
05-26-2017
05:52 PM
Thank you, I will try it
... View more
05-26-2017
08:14 AM
This the log 2017-05-25 14:40:37,470 INFO BlockStateChange: BLOCK* addStoredBlock: blockMap updated: .79:50010 is added to blk_3006468944_1934832460{blockUCState=UNDER_CONSTRUCTION, primaryNodeIndex=-1, replicas=[ReplicaUnderConstruction[[DISK]DS-4a898333-0a0f-4197-944e-e2f499645e4c:NORMAL: .55:50010|FINALIZED], ReplicaUnderConstruction[[DISK]DS-67e857f6-31e6-4d16-b079-7eac06de1f22:NORMAL: .79:50010|FINALIZED]]} size 0 2017-05-25 14:40:37,740 INFO org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader: replaying edit log: 137515/140868 transactions completed. (98%) 2017-05-25 14:40:37,766 INFO org.apache.hadoop.hdfs.server.namenode.FSImage: Edits file http:// .local:8480/getJournal?jid=nameservice1&segmentTxId=20804872533&storageInfo=-60%3A78118620%3A1469591854075%3Acluster24, http://XXX.XXX:8480/getJournal?jid=nameservice1&segmentTxId=20804872533&storageInfo=-60%3A78118620%3A1469591854075%3Acluster24, http://XXX.XXX:8480/getJournal?jid=nameservice1&segmentTxId=20804872533&storageInfo=-60%3A78118620%3A1469591854075%3Acluster24 of size 28215334 edits # 140868 loaded in 3 seconds 2017-05-25 14:41:18,563 INFO org.apache.hadoop.http.HttpServer2: Process Thread Dump: jsp requested 292 active threads Thread 261120 (492304899@qtp-1369470652-3169): sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:186) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043) java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) java.lang.Thread.run(Thread.java:745) Thread 261102 (Logger channel (from single-thread executor) to XXX.XXX/ .4:8485): State: WAITING Blocked count: 8 Waited count: 10 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@5632f864 Stack: sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:186) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:2043) java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:442) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:1068) java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1130) java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) java.lang.Thread.run(Thread.java:745) org.apache.hadoop.hdfs.protocolPB.DatanodeProtocolServerSideTranslatorPB.sendHeartbeat(DatanodeProtocolServerSideTranslatorPB.java:114) org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos$DatanodeProtocolService$2.callBlockingMethod(DatanodeProtocolProtos.java:29186) org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617) org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1073) org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086) org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082) java.security.AccessController.doPrivileged(Native Method) javax.security.auth.Subject.doAs(Subject.java:415) org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1693) org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080) Thread 297 (IPC Server handler 106 on 8022): State: WAITING Blocked count: 5299 Waited count: 2047289 Waiting on java.util.concurrent.locks.ReentrantReadWriteLock$FairSync@35deedf3 Stack: sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:186) java.util.concurrent.locks.AbstractQueuedSynchronizer.parkAndCheckInterrupt(AbstractQueuedSynchronizer.java:834) java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireShared(AbstractQueuedSynchronizer.java:964) java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireShared(AbstractQueuedSynchronizer.java:1282) java.lang.ref.Finalizer$FinalizerThread.run(Finalizer.java:189) Thread 2 (Reference Handler): State: WAITING Blocked count: 109242 Waited count: 108895 Waiting on java.lang.ref.Reference$Lock@5f243f31 Stack: java.lang.Object.wait(Native Method) java.lang.Object.wait(Object.java:503) java.lang.ref.Reference$ReferenceHandler.run(Reference.java:133) Thread 1 (main): State: WAITING Blocked count: 9 Waited count: 6 Waiting on org.apache.hadoop.ipc.ProtobufRpcEngine$Server@20487a6c Stack: java.lang.Object.wait(Native Method) java.lang.Object.wait(Object.java:503) org.apache.hadoop.ipc.Server.join(Server.java:2508) org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.join(NameNodeRpcServer.java:443) org.apache.hadoop.hdfs.server.namenode.NameNode.join(NameNode.java:889) org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1613) 2017-05-25 14:41:27,207 INFO org.apache.hadoop.hdfs.server.namenode.ha.EditLogTailer: Loaded 140868 edits starting from txid 20804872532 2017-05-25 14:41:27,207 INFO org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: Marking all datandoes as stale 2017-05-25 14:41:27,208 INFO org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Processing 15 messages from DataNodes that were previously queued during standby state 2017-05-25 14:41:27,208 INFO org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Received an RBW replica for blk_3005346895_1933976855 on .124:50010: ignoring it, since it is complete with the same genstamp 2017-05-25 14:41:27,208 INFO BlockStateChange: BLOCK NameSystem.addToCorruptReplicasMap: blk_2995704659 added as corrupt on .148:50010 by / because block is COMPLETE and reported length 0 does not match length in block map 155346 2017-05-25 14:41:27,208 INFO BlockStateChange: BLOCK* invalidateBlock: blk_2995704659_1924067671(same as stored) on .148:50010 2017-05-25 14:41:27,208 INFO BlockStateChange: BLOCK* invalidateBlocks: postponing invalidation of blk_2995704659_1924067671(same as stored) on .148:50010 because 4 replica(s) are located on nodes with potentially out-of-date block reports 2017-05-25 14:41:27,208 INFO BlockStateChange: BLOCK NameSystem.addToCorruptReplicasMap: blk_2995704823 added as corrupt on .148:50010 by / because block is COMPLETE and reported length 0 does not match length in block map 257707 2017-05-25 14:41:27,208 INFO BlockStateChange: BLOCK* invalidateBlock: blk_2995704823_1924067835(same as stored) on .148:50010 2017-05-25 14:41:27,208 INFO BlockStateChange: BLOCK* invalidateBlocks: postponing invalidation of blk_2995704823_1924067835(same as stored) on .148:50010 because 4 replica(s) are located on nodes with potentially out-of-date block reports 2017-05-25 14:41:27,208 INFO BlockStateChange: BLOCK NameSystem.addToCorruptReplicasMap: blk_2995704820 added as corrupt on .148:50010 by / because block is COMPLETE and reported length 0 does not match length in block map 129686 2017-05-25 14:41:27,208 INFO BlockStateChange: BLOCK* invalidateBlock: blk_2995704820_1924067832(same as stored) on .148:50010 2017-05-25 14:41:27,208 INFO BlockStateChange: BLOCK* invalidateBlocks: postponing invalidation of blk_2995704820_1924067832(same as stored) on .148:50010 because 4 replica(s) are located on nodes with potentially out-of-date block reports 2017-05-25 14:41:27,208 INFO BlockStateChange: BLOCK NameSystem.addToCorruptReplicasMap: blk_2995704331 added as corrupt on .148:50010 by / because block is COMPLETE and reported length 0 does not match length in block map 3027 2017-05-25 14:41:27,208 INFO BlockStateChange: BLOCK* invalidateBlock: blk_2995704331_1924067343(same as stored) on .148:50010 2017-05-25 14:41:27,208 INFO BlockStateChange: BLOCK* invalidateBlocks: postponing invalidation of blk_2995704331_1924067343(same as stored) on .148:50010 because 4 replica(s) are located on nodes with potentially out-of-date block reports 2017-05-25 14:41:27,208 INFO BlockStateChange: BLOCK NameSystem.addToCorruptReplicasMap: blk_2995704571 added as corrupt on .148:50010 by / because block is COMPLETE and reported length 0 does not match length in block map 2969 2017-05-25 14:41:27,208 INFO BlockStateChange: BLOCK* invalidateBlock: blk_2995704571_1924067583(same as stored) on .148:50010 2017-05-25 14:41:27,208 INFO BlockStateChange: BLOCK* invalidateBlocks: postponing invalidation of blk_2995704571_1924067583(same as stored) on .148:50010 because 4 replica(s) are located on nodes with potentially out-of-date block reports 2017-05-25 14:41:27,209 INFO BlockStateChange: BLOCK NameSystem.addToCorruptReplicasMap: blk_2995704519 added as corrupt on .148:50010 by / because block is COMPLETE and reported length 0 does not match length in block map 3103 2017-05-25 14:41:27,209 INFO BlockStateChange: BLOCK* invalidateBlock: blk_2995704519_1924067531(same as stored) on .148:50010 2017-05-25 14:41:27,209 INFO BlockStateChange: BLOCK* invalidateBlocks: postponing invalidation of blk_2995704519_1924067531(same as stored) on .148:50010 because 4 replica(s) are located on nodes with potentially out-of-date block reports 2017-05-25 14:41:27,209 INFO BlockStateChange: BLOCK NameSystem.addToCorruptReplicasMap: blk_2995704285 added as corrupt on .148:50010 by / because block is COMPLETE and reported length 0 does not match length in block map 3349 2017-05-25 14:41:27,209 INFO BlockStateChange: BLOCK* invalidateBlock: blk_2995704285_1924067297(same as stored) on .148:50010 2017-05-25 14:41:27,209 INFO BlockStateChange: BLOCK* invalidateBlocks: postponing invalidation of blk_2995704285_1924067297(same as stored) on .148:50010 because 4 replica(s) are located on nodes with potentially out-of-date block reports 2017-05-25 14:41:27,209 INFO BlockStateChange: BLOCK NameSystem.addToCorruptReplicasMap: blk_2995703869 added as corrupt on .148:50010 by / because block is COMPLETE and reported length 0 does not match length in block map 518573 2017-05-25 14:41:27,209 INFO BlockStateChange: BLOCK* invalidateBlock: blk_2995703869_1924066881(same as stored) on .148:50010 2017-05-25 14:41:27,209 INFO BlockStateChange: BLOCK* invalidateBlocks: postponing invalidation of blk_2995703869_1924066881(same as stored) on .148:50010 because 4 replica(s) are located on nodes with potentially out-of-date block reports 2017-05-25 14:41:27,209 INFO BlockStateChange: BLOCK NameSystem.addToCorruptReplicasMap: blk_2995703867 added as corrupt on .148:50010 by / because block is COMPLETE and reported length 0 does not match length in block map 36769 2017-05-25 14:41:27,209 INFO BlockStateChange: BLOCK* invalidateBlock: blk_2995703867_1924066879(same as stored) on .148:50010 2017-05-25 14:41:27,209 INFO BlockStateChange: BLOCK* invalidateBlocks: postponing invalidation of blk_2995703867_1924066879(same as stored) on .148:50010 because 4 replica(s) are located on nodes with potentially out-of-date block reports 2017-05-25 14:41:27,209 INFO BlockStateChange: BLOCK NameSystem.addToCorruptReplicasMap: blk_2995701937 added as corrupt on .148:50010 by / because reported RWR replica with genstamp 1924064949 does not match COMPLETE block's genstamp in block map 1924307850 2017-05-25 14:41:27,209 INFO BlockStateChange: BLOCK* invalidateBlock: blk_2995701937_1924064949(stored=blk_2995701937_1924307850) on .148:50010 2017-05-25 14:41:27,209 INFO BlockStateChange: BLOCK* invalidateBlocks: postponing invalidation of blk_2995701937_1924064949(stored=blk_2995701937_1924307850) on .148:50010 because 4 replica(s) are located on nodes with potentially out-of-date block reports 2017-05-25 14:41:27,209 INFO org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Received an RBW replica for blk_3005532032_1934102219 on .73:50010: ignoring it, since it is complete with the same genstamp 2017-05-25 14:41:27,209 INFO org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Received an RBW replica for blk_3005613028_1934020244 on .124:50010: ignoring it, since it is complete with the same genstamp 2017-05-25 14:41:27,209 INFO org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Received an RBW replica for blk_3005491542_1933974671 on .128:50010: ignoring it, since it is complete with the same genstamp 2017-05-25 14:41:27,209 INFO org.apache.hadoop.hdfs.server.blockmanagement.BlockManager: Received an RBW replica for blk_3005491542_1933974671 on .93:50010: ignoring it, since it is complete with the same genstamp 2017-05-25 14:41:27,209 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Reprocessing replication and invalidation queues 2017-05-25 14:41:27,209 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: initializing replication queues 2017-05-25 14:41:27,209 INFO org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Will take over writing edit logs at txnid 20805013401 2017-05-25 14:41:27,305 INFO org.apache.hadoop.hdfs.server.namenode.FSEditLog: Starting log segment at 20805013401 2017-05-25 14:41:27,346 FATAL org.apache.hadoop.hdfs.server.namenode.FSEditLog: Error: starting log segment 20805013401 failed for required journal (JournalAndStream(mgr=QJM to [ , .4:8485, .5:8485, .6:8485, .14:8485], stream=null)) org.apache.hadoop.hdfs.qjournal.client.QuorumException: Got too many exceptions to achieve quorum size 3/5. 5 exceptions thrown: : IPC's epoch 46 is less than the last promised epoch 47 at org.apache.hadoop.hdfs.qjournal.server.Journal.checkRequest(Journal.java:429) at org.apache.hadoop.hdfs.qjournal.server.Journal.startLogSegment(Journal.java:513) at org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.startLogSegment(JournalNodeRpcServer.java:162) at org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolServerSideTranslatorPB.startLogSegment(QJournalProtocolServerSideTranslatorPB.java:198) at org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos$QJournalProtocolService$2.callBlockingMethod(QJournalProtocolProtos.java:25425) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1073) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1693) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080) .14:8485: IPC's epoch 46 is less than the last promised epoch 47 at org.apache.hadoop.hdfs.qjournal.server.Journal.checkRequest(Journal.java:429) at org.apache.hadoop.hdfs.qjournal.server.Journal.startLogSegment(Journal.java:513) at org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.startLogSegment(JournalNodeRpcServer.java:162) at org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolServerSideTranslatorPB.startLogSegment(QJournalProtocolServerSideTranslatorPB.java:198) at org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos$QJournalProtocolService$2.callBlockingMethod(QJournalProtocolProtos.java:25425) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1073) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1693) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080) .5:8485: IPC's epoch 46 is less than the last promised epoch 47 at org.apache.hadoop.hdfs.qjournal.server.Journal.checkRequest(Journal.java:429) at org.apache.hadoop.hdfs.qjournal.server.Journal.startLogSegment(Journal.java:513) at org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.startLogSegment(JournalNodeRpcServer.java:162) at org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolServerSideTranslatorPB.startLogSegment(QJournalProtocolServerSideTranslatorPB.java:198) at org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos$QJournalProtocolService$2.callBlockingMethod(QJournalProtocolProtos.java:25425) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1073) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1693) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080) .4:8485: IPC's epoch 46 is less than the last promised epoch 47 at org.apache.hadoop.hdfs.qjournal.server.Journal.checkRequest(Journal.java:429) at org.apache.hadoop.hdfs.qjournal.server.Journal.startLogSegment(Journal.java:513) at org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.startLogSegment(JournalNodeRpcServer.java:162) at org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolServerSideTranslatorPB.startLogSegment(QJournalProtocolServerSideTranslatorPB.java:198) at org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos$QJournalProtocolService$2.callBlockingMethod(QJournalProtocolProtos.java:25425) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1073) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1693) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080) .6:8485: IPC's epoch 46 is less than the last promised epoch 47 at org.apache.hadoop.hdfs.qjournal.server.Journal.checkRequest(Journal.java:429) at org.apache.hadoop.hdfs.qjournal.server.Journal.startLogSegment(Journal.java:513) at org.apache.hadoop.hdfs.qjournal.server.JournalNodeRpcServer.startLogSegment(JournalNodeRpcServer.java:162) at org.apache.hadoop.hdfs.qjournal.protocolPB.QJournalProtocolServerSideTranslatorPB.startLogSegment(QJournalProtocolServerSideTranslatorPB.java:198) at org.apache.hadoop.hdfs.qjournal.protocol.QJournalProtocolProtos$QJournalProtocolService$2.callBlockingMethod(QJournalProtocolProtos.java:25425) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1073) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1693) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080) at org.apache.hadoop.hdfs.qjournal.client.QuorumException.create(QuorumException.java:81) at org.apache.hadoop.hdfs.qjournal.client.QuorumCall.rethrowException(QuorumCall.java:223) at org.apache.hadoop.hdfs.qjournal.client.AsyncLoggerSet.waitForWriteQuorum(AsyncLoggerSet.java:142) at org.apache.hadoop.hdfs.qjournal.client.QuorumJournalManager.startLogSegment(QuorumJournalManager.java:403) at org.apache.hadoop.hdfs.server.namenode.JournalSet$JournalAndStream.startLogSegment(JournalSet.java:107) at org.apache.hadoop.hdfs.server.namenode.JournalSet$3.apply(JournalSet.java:222) at org.apache.hadoop.hdfs.server.namenode.JournalSet.mapJournalsAndReportErrors(JournalSet.java:393) at org.apache.hadoop.hdfs.server.namenode.JournalSet.startLogSegment(JournalSet.java:219) at org.apache.hadoop.hdfs.server.namenode.FSEditLog.startLogSegment(FSEditLog.java:1206) at org.apache.hadoop.hdfs.server.namenode.FSEditLog.openForWrite(FSEditLog.java:316) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startActiveServices(FSNamesystem.java:1248) at org.apache.hadoop.hdfs.server.namenode.NameNode$NameNodeHAContext.startActiveServices(NameNode.java:1767) at org.apache.hadoop.hdfs.server.namenode.ha.ActiveState.enterState(ActiveState.java:61) at org.apache.hadoop.hdfs.server.namenode.ha.HAState.setStateInternal(HAState.java:64) at org.apache.hadoop.hdfs.server.namenode.ha.StandbyState.setState(StandbyState.java:49) at org.apache.hadoop.hdfs.server.namenode.NameNode.transitionToActive(NameNode.java:1640) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.transitionToActive(NameNodeRpcServer.java:1375) at org.apache.hadoop.ha.protocolPB.HAServiceProtocolServerSideTranslatorPB.transitionToActive(HAServiceProtocolServerSideTranslatorPB.java:107) at org.apache.hadoop.ha.proto.HAServiceProtocolProtos$HAServiceProtocolService$2.callBlockingMethod(HAServiceProtocolProtos.java:4460) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1073) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1693) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080) 2017-05-25 14:41:27,348 INFO org.apache.hadoop.util.ExitUtil: Exiting with status 1 2017-05-25 14:41:27,375 INFO org.apache.hadoop.hdfs.server.namenode.NameNode: SHUTDOWN_MSG: /usr/java/jdk1.7.0_55/bin/java -Dproc_namenode -Xmx1000m -Dhdfs.audit.logger=INFO,RFAAUDIT -Dsecurity.audit.logger=INFO,RFAS -Djava.net.preferIPv4Stack=true -Dhadoop.log.dir=/agoda/log/hadoop-hdfs -Dhadoop.log.file=hadoop-cmf-hdfs-NAMENODE- .local.log.out -Dhadoop.home.dir=/opt/cloudera/parcels/CDH-5.8.0-1.cdh5.8.0.p0.42/lib/hadoop -Dhadoop.id.str=hdfs -Dhadoop.root.logger=INFO,RFA -Djava.library.path=/opt/cloudera/parcels/CDH-5.8.0-1.cdh5.8.0.p0.42/lib/hadoop/lib/native -Dhadoop.policy.file=hadoop-policy.xml -Djava.net.preferIPv4Stack=true -Xms137438953472 -Xmx137438953472 -XX:+UseParNewGC -XX:+UseConcMarkSweepGC -XX:-CMSConcurrentMTEnabled -XX:CMSInitiatingOccupancyFraction=70 -XX:+CMSParallelRemarkEnabled -Dcom.sun.management.jmxremote -Dcom.sun.management.jmxremote.authenticate=false -Dcom.sun.management.jmxremote.ssl=false -Dcom.sun.management.jmxremote.port=8004 -XX:OnOutOfMemoryError=/usr/lib64/cmf/service/common/killparent.sh -Dhadoop.security.logger=INFO,RFAS org.apache.hadoop.hdfs.server.namenode.NameNode
... View more
Labels:
- Labels:
-
HDFS