Community Articles

Find and share helpful community-sourced technical articles.
Labels (2)

A common error to see in initial installations is the following from Accumulo TabletServer logs

Caused by: org.apache.hadoop.ipc.RemoteException( File /apps/accumulo/data/wal/ could only be replicated to 0 nodes instead of minReplication (=1).  There are 3 datanode(s) running and no node(s) are excluded in this operation.
    at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(
    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getNewBlockTargets(
    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(
    at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(
    at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(
    at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(
    at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$
    at org.apache.hadoop.ipc.RPC$
    at org.apache.hadoop.ipc.Server$Handler$
    at org.apache.hadoop.ipc.Server$Handler$
    at Method)
    at org.apache.hadoop.ipc.Server$
    at org.apache.hadoop.ipc.Client.getRpcResponse(
    at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(

This exception will be printed repeatedly in the TabletServer logs as Accumulo has no other solution than to try to create its write-ahead log file again.

This exception is, indirectly, telling us multiple things about the current state:

  1. There are three Datanodes
  2. None of the Datanodes were avoided -- this means all three of them should have been able to accept the write
  3. None of the Datanodes successfully accepted the write

The most common cause of this issue is that each Datanode has a very small amount of disk space to use. When Accumulo creates its write-ahead log files, it sets a large HDFS block size (by default: 1GB). If the Datanode does not have enough free space to store 1GB of data, the allocation fails. When all of the Datanodes are in this situation, you would see the above error message.

The solution to the above problem is to provide more storage for the Datanode. Commonly, this is because HDFS is not configured to use the correct data directories or some hard drives were not mounted to the data dirs (and thus the Datanodes are using the root volume).

0 Kudos
Take a Tour of the Community
Don't have an account?
Your experience may be limited. Sign in to explore more.
Version history
Last update:
‎03-10-2017 11:26 PM
Updated by:
Top Kudoed Authors