Created 06-15-2016 05:09 AM
Hi All,
HBase server fails to start with the error below.
A resolution suggested in another post of Hbase start issue - Deleting files under WALS path didn't help.
IF you could advise what can be done or investigated that would be a great help.
Regards,
-----------------------------
2016-06-15 15:02:38,692 FATAL [ip-172-31-25-3:16000.activeMasterManager] master.HMaster: Failed to become active master java.io.FileNotFoundException: File does not exist: /apps/hbase/data/MasterProcWALs/state-************00000001.log
2016-06-15 15:02:38,636 INFO [ip-172-31-25-3:16000.activeMasterManager] procedure.ZKProcedureUtil: Clearing all procedure znodes: /hbase-unsecure/flush-table-proc/acquired /hbase-unsecure/flush-table-proc/reached /hbase-unsecure/flush-table-proc/abort 2016-06-15 15:02:38,643 INFO [ip-172-31-25-3:16000.activeMasterManager] procedure.ZKProcedureUtil: Clearing all procedure znodes: /hbase-unsecure/online-snapshot/acquired /hbase-unsecure/online-snapshot/reached /hbase-unsecure/online-snapshot/abort 2016-06-15 15:02:38,664 INFO [ip-172-31-25-3:16000.activeMasterManager] master.MasterCoprocessorHost: System coprocessor loading is enabled 2016-06-15 15:02:38,676 INFO [ip-172-31-25-3:16000.activeMasterManager] procedure2.ProcedureExecutor: Starting procedure executor threads=5 2016-06-15 15:02:38,677 INFO [ip-172-31-25-3:16000.activeMasterManager] wal.WALProcedureStore: Starting WAL Procedure Store lease recovery 2016-06-15 15:02:38,680 INFO [ip-172-31-25-3:16000.activeMasterManager] util.FSHDFSUtils: Recovering lease on dfs file hdfs://ip-172-31-25-3.ap-southeast-2.compute.internal:8020/apps/hbase/data/MasterProcWALs/state-************00000001.log 2016-06-15 15:02:38,692 FATAL [ip-172-31-25-3:16000.activeMasterManager] master.HMaster: Failed to become active master java.io.FileNotFoundException: File does not exist: /apps/hbase/data/MasterProcWALs/state-************00000001.log at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:71) at org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:61) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLease(FSNamesystem.java:2877) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.recoverLease(NameNodeRpcServer.java:753) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.recoverLease(ClientNamenodeProtocolServerSideTranslatorPB.java:671) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:969) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2206) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2202) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1709) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2200)
Created 06-15-2016 02:50 PM
Thanks Ted and sjiang.
This issue has gone without any action - not that I know of.
I have stopped the entire cluster since the issue, and when restarted it after a while hbase came up alright. No actions taken.
I will reply to your answers(so you can be notified) the full log if the issue comes back.
Again thanks.
Created 07-15-2016 03:02 AM
Sorry, I hate to repeat this, but this issue came back again. I am not trying to fool anyone here. Sorry for the misleading clue above.
This time no change has been made but the cluster restart. Same error. Restarting doesn't solve the problem.
This issue is so intermittent so very hard to troubleshoot.