09-17-2018 10:59 PM
While putting data in Hbase, using the HTable.put method, I'll end up with the below exception occasionally. But the data has been actually written to Hbase when I checked the get operation for that particular rowkey.
For the same time I have searched for the logs in both HMaster and HRegionservers to identify the issue. But unable to find that.
Please help to fine tune Hbase Configurations in order to avoid InterruptedIOException
java.io.InterruptedIOException: #17209, interrupted. currentNumberOfTask=1 at org.apache.hadoop.hbase.client.AsyncProcess.waitForMaximumCurrentTasks(AsyncProcess.java:1764) at org.apache.hadoop.hbase.client.AsyncProcess.waitForMaximumCurrentTasks(AsyncProcess.java:1734) at org.apache.hadoop.hbase.client.AsyncProcess.waitForAllPreviousOpsAndReset(AsyncProcess.java:1810) at org.apache.hadoop.hbase.client.BufferedMutatorImpl.backgroundFlushCommits(BufferedMutatorImpl.java:240) at org.apache.hadoop.hbase.client.BufferedMutatorImpl.flush(BufferedMutatorImpl.java:190) at org.apache.hadoop.hbase.client.HTable.flushCommits(HTable.java:1434) at org.apache.hadoop.hbase.client.HTable.put(HTable.java:1018)
Please help to solve it
The same exception has been faced by someone . But in that thread, there is no explanation about which are configurations need to be checked in order to avoid it