Created on 01-19-2016 03:52 AM - edited 09-16-2022 02:58 AM
I have Configured HDFS & HIVE on Fedora 20 (32 bit linux) on VM. HDFS and HIVE is running properly on VM, but problem occurs when trying to connect through any external tool or program like jasper or java.
PFB the error for reference-
Exception in thread "main" java.lang.OutOfMemoryError: Java heap space at org.apache.thrift.protocol.TBinaryProtocol.readStringBody(TBinaryProtocol.java:353) at org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:215) at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:69) at org.apache.hadoop.hive.service.ThriftHive$Client.recv_execute(ThriftHive.java:116) at org.apache.hadoop.hive.service.ThriftHive$Client.execute(ThriftHive.java:103) at org.apache.hadoop.hive.jdbc.HiveStatement.executeQuery(HiveStatement.java:192) at org.apache.hadoop.hive.jdbc.HiveStatement.execute(HiveStatement.java:132) at org.apache.hadoop.hive.jdbc.HiveConnection.configureConnection(HiveConnection.java:133) at org.apache.hadoop.hive.jdbc.HiveConnection.<init>(HiveConnection.java:122) at org.apache.hadoop.hive.jdbc.HiveDriver.connect(HiveDriver.java:106) at java.sql.DriverManager.getConnection(Unknown Source) at java.sql.DriverManager.getConnection(Unknown Source) at Hive_test.main(Hive_test.java:20)
Following JARS i have included-
ant-1.8.1.jar apache-httpcomponents-httpclient.jar apache-httpcomponents-httpcore.jar commons-el-1.0-javadoc.jar commons-el-1.0-sources.jar commons-el-1.0.jar commons-logging-1.1.1.jar commons-logging-api-1.0.4.jar connector-sdk-1.99.5.jar hadoop-common-2.6.0.2.2.4.2-2.jar hive-beeline-0.12.0.jar hive-cli-0.12.0.jar hive-common-0.12.0.jar hive-contrib-0.12.0.jar hive-exec-0.12.0.jar hive-exec-0.8.0.jar hive-exec-0.8.0.jar.zip hive-io-exp-core-0.6.jar.zip hive-io-exp-deps-0.6-sources.jar.zip hive-io-exp-deps-0.6.jar.zip hive-jdbc-0.12.0.jar hive-metastore-0.12.0.jar hive-service-0.12.0.jar httpclient-4.2.5.jar httpcore-4.2.4.jar jackson-core-asl-1.8.8.jar jasper-compiler-5.5.23.jar jasper-runtime-5.5.23.jar jsp-api-2.1-sources.jar jsp-api-2.1.jar junit-4.11-javadoc.jar junit-4.11-sources.jar junit-4.11.jar libfb303-0.9.0.jar libthrift-0.9.0.jar log4j-1.2.16-javadoc.jar log4j-1.2.16-sources.jar log4j-1.2.16.jar ql.jar slf4j-api-1.6.1-javadoc.jar slf4j-api-1.6.1-sources.jar slf4j-api-1.6.1.jar slf4j-log4j12-1.6.1.jar sqoop-connector-kite-1.99.5.jar
So please suggest any corrective action.
Created 01-19-2016 06:03 AM
Could you post some of you heap configurations? How much memory is available on the machine? OOM error usually means the heap configuration is not correct or their is not enough memory available on the machine.
You also might want to check the open files limit (ulimit -a), if its too low it can cause OOM errors. (see https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.3.2/bk_installing_manually_book/content/ref-729...)
Even though you might be able to run Hadoop on a 32bit system, I wouldn't recommend it. You should use a 64bit system (see http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.3.2/bk_installing_manually_book/content/meet-min...)
Created 01-19-2016 04:04 AM
System is running out of memory
Hive is trying to create tez container and system does not have enough Memory
Check the vm memory and see if you can increase it
Created 01-19-2016 06:03 AM
Could you post some of you heap configurations? How much memory is available on the machine? OOM error usually means the heap configuration is not correct or their is not enough memory available on the machine.
You also might want to check the open files limit (ulimit -a), if its too low it can cause OOM errors. (see https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.3.2/bk_installing_manually_book/content/ref-729...)
Even though you might be able to run Hadoop on a 32bit system, I wouldn't recommend it. You should use a 64bit system (see http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.3.2/bk_installing_manually_book/content/meet-min...)
Created 02-03-2016 01:43 AM
@Kumar Ratan are you still having issues with this? Can you accept best answer or provide your workaround?