Support Questions

Find answers, ask questions, and share your expertise
Welcome to the upgraded Community! Read this blog to see What’s New!

sqoop job getting 'Cannot allocate memory' (errno=12) error

New Contributor

When I run sqoop job to move data from oracle to hdfs(hive). I'm getting "Java HotSpot(TM) 64-Bit Server VM warning: INFO: os::commit_memory(0x00000000fef80000, 17301504, 0) failed; error='Cannot allocate memory' (errno=12)" error. Any idea what I have to do? By the way Each node has 24 GB




Can you include the sqoop command? What is the size of the data that you are transferring from Oracle database? What value did you give --m option ?

Java was not able to allocate enough memory, i.e. it's not Java's heap limit that's in the way but rather no more memory available to be given to Java by OS. Check that the machine is not running out of memory. And first clean ram or increase ram then check if again there is an out of memory error then increase heap size:

  • -Xms128m min(heap size)
  • -Xmx512m max(heap size)
  • -XX:MaxPermSize max(perm size)

Hope that helps