Support Questions

Find answers, ask questions, and share your expertise
Announcements
Check out our newest addition to the community, the Cloudera Data Analytics (CDA) group hub.

hive and spark Job Failures because of missing blocks

New Contributor

Hello,

 

I have a problem with Spark and Hive Jobs because of missing blocks in the cluster. Cloudera Version is 5.12.0.

 

Error Message is: 

Caused by: java.io.IOException: org.apache.hadoop.hdfs.BlockMissingException: Could not obtain block: BP-1229449011-10.65.184.84-1424518888833:blk_1116890537_43225117 

 

So we have actually missing blocks in the cluster. I have created a list of all missing blocks and it's locations.

The blocks in the error message are not missing. 

 

The problem occurs with running hive queries over ssh with "hive -e". If I run the same Statement over hue it works. Further is the replication factor of the above Block  > 1.

 

What could be the cause for that issue. What could I do for beter debugging the problem?

Thank you in advance,

 

best regards 

Joern

 

 

 

2 REPLIES 2

Explorer
any update on this issue? i’m facing the same problem.

New Contributor

Hello,

 

in my case the problem was a node with a different OS. After decomissioning the node, the error was fixed.

Maybe this information could help somebody.

 

Best wishes 

Joern

Take a Tour of the Community
Don't have an account?
Your experience may be limited. Sign in to explore more.