Community Articles
Find and share helpful community-sourced technical articles
Alert: Welcome to the Unified Cloudera Community. Former HCC members be sure to read and learn how to activate your account here.
Labels (2)

A common error to see in initial installations is the following from Accumulo TabletServer logs

Caused by: org.apache.hadoop.ipc.RemoteException( File /apps/accumulo/data/wal/ could only be replicated to 0 nodes instead of minReplication (=1).  There are 3 datanode(s) running and no node(s) are excluded in this operation.
    at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(
    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getNewBlockTargets(
    at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(
    at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(
    at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(
    at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(
    at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$
    at org.apache.hadoop.ipc.RPC$
    at org.apache.hadoop.ipc.Server$Handler$
    at org.apache.hadoop.ipc.Server$Handler$
    at Method)
    at org.apache.hadoop.ipc.Server$
    at org.apache.hadoop.ipc.Client.getRpcResponse(
    at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(

This exception will be printed repeatedly in the TabletServer logs as Accumulo has no other solution than to try to create its write-ahead log file again.

This exception is, indirectly, telling us multiple things about the current state:

  1. There are three Datanodes
  2. None of the Datanodes were avoided -- this means all three of them should have been able to accept the write
  3. None of the Datanodes successfully accepted the write

The most common cause of this issue is that each Datanode has a very small amount of disk space to use. When Accumulo creates its write-ahead log files, it sets a large HDFS block size (by default: 1GB). If the Datanode does not have enough free space to store 1GB of data, the allocation fails. When all of the Datanodes are in this situation, you would see the above error message.

The solution to the above problem is to provide more storage for the Datanode. Commonly, this is because HDFS is not configured to use the correct data directories or some hard drives were not mounted to the data dirs (and thus the Datanodes are using the root volume).

0 Kudos
Don't have an account?
Coming from Hortonworks? Activate your account here
Version history
Revision #:
1 of 1
Last update:
‎03-10-2017 11:26 PM
Updated by:
Top Kudoed Authors