Member since
06-20-2016
13
Posts
6
Kudos Received
0
Solutions
08-10-2016
09:12 AM
1 Kudo
Hi, I've seen that the technical preview of HDP 2.5 is currently available. Is there any estimation of the hdp 2.5 final release date? Thanks a lot in advance.
... View more
Labels:
- Labels:
-
Hortonworks Data Platform (HDP)
08-04-2016
06:27 AM
Thanks @vshukla.
... View more
08-03-2016
10:03 AM
1 Kudo
Hi, Is the version 2.0.0 of Spark supported in Hortonworks 2.4.x? Currently we have Spark 1.6 installed but we would need to update to 2.0. Thanks a lot in advance.
... View more
Labels:
- Labels:
-
Apache Spark
07-19-2016
07:53 AM
Hi again @rbiswas, Up to my knowledge, each time a mapper (or reducer) is created, the ApplicationMaster will request the NodeManager to allocate a new Container with mapreduce.map.memory.mb (and mapreduce.reduce.memory.mb) MBytes available. So, with my specific configuration, if three mappers are created, then, YARN will try to create three containers with 3072 MB each. Am I right? If so, what if YARN can't reserve (3*3072MB)? Will it raise a Java Heap Space Exception? Thanks in advance.
... View more
07-19-2016
06:58 AM
Hi @rbiswas, Thanks for your comment. I didn't know how to adjust that memory parameter. However, looking in hadoop-env.sh I discovered that it was set at 1024MB: export HADOOP_NAMENODE_INIT_HEAPSIZE="-Xms1024m" Unfortunately, that didn't solve the problem.
... View more
07-18-2016
08:43 AM
1 Kudo
Hi, I'm trying to configure the optimum memory configuration in YARN to implement some MR tasks in R. For the moment, I have a single node with around 40GB RAM available. I have tried different memory combinations but all of them result in Java Heap Space exceptions when trying to execute a simple MR R code (using the library plyrmr) to process a small (a few KB size) text file. The relevant memory configuration parameters I have so far (in yarn-site.xml and map-red.xml) are: yarn.scheduler.maximum-allocation-mb = 24576
yarn.scheduler.minimum-allocation-mb = 3076
yarn.app.mapreduce.am.resource.mb = 3076
mapreduce.map.java.opts = -Xmx2457m
mapreduce.map.memory.mb = 3072
mapreduce.reduce.java.opts = -Xmx4915m
mapreduce.reduce.memory.mb = 6144
Is there any other memory configuration parameter that needs to be set or adjusted? After launching the task, 2 split jobs are created and a Java Heap Space exception is raised. Looking through the YARN logs of the application that raises the exception, I stumple upon the following line after executing the launch_container.sh: exec /bin/bash -c "$JAVA_HOME/bin/java -server -XX:NewRatio=8 -Djava.net.preferIPv4Stack=true -Dhdp.version=2.4.0.0-169 -Xmx400M What are these "400 MB" of Java Space for? I have checked a lot of different configuration files but I couldn't find any parameter related to these 400MB of space. Is there any other Java parameter that needs to be set in the aforementioned list of configuration properties? The relevant log part of the MR task is: INFO mapreduce.Job: Counters: 17
Job Counters
Failed map tasks=7
Killed map tasks=1
Killed reduce tasks=1
Launched map tasks=8
Other local map tasks=6
Data-local map tasks=2
Total time spent by all maps in occupied slots (ms)=37110
Total time spent by all reduces in occupied slots (ms)=0
Total time spent by all map tasks (ms)=37110
Total time spent by all reduce tasks (ms)=0
Total vcore-seconds taken by all map tasks=37110
Total vcore-seconds taken by all reduce tasks=0
Total megabyte-seconds taken by all map tasks=114001920
Total megabyte-seconds taken by all reduce tasks=0
Map-Reduce Framework
CPU time spent (ms)=0
Physical memory (bytes) snapshot=0
Virtual memory (bytes) snapshot=0 Is there anything that I'm missing? Thanks a lot for your time.
... View more
Labels:
- Labels:
-
Apache YARN
06-28-2016
08:30 AM
Removing the sqoop-site.xml from that folder also worked for me.
... View more
06-24-2016
10:01 AM
Thanks, Artem. As I read from another question you answered, you were not able to use MySql as a metastore (https://community.hortonworks.com/questions/22152/sqoop-metastore-usage-supported-databases-backup.html). Do you think it is worth having a try with MySQL, or would you use another database? Maybe Postgres?
... View more
06-23-2016
01:53 PM
2 Kudos
EDITED: Just in case somebody ends up in this question trying to set MySQL as the Sqoop metastore, the instructions to do that are described in this post: https://community.hortonworks.com/articles/55937/using-sqoop-with-mysql-as-metastore.html
Hi,
I have a single Hortonworks node, using Sqoop to import some tables from an SQLServer database. Whenever I try to execute two Oozie workflows simultaneously (each of which performs a Sqoop import job), I get an exception related to the hsqldb metastore. 3606 [main] ERROR org.apache.sqoop.tool.JobTool - I/O error performing job operation: java.io.IOException: Exception creating SQL connection
at org.apache.sqoop.metastore.hsqldb.HsqldbJobStorage.init(HsqldbJobStorage.java:216)
at org.apache.sqoop.metastore.hsqldb.AutoHsqldbStorage.open(AutoHsqldbStorage.java:112)
at org.apache.sqoop.tool.JobTool.run(JobTool.java:274)
at org.apache.sqoop.Sqoop.run(Sqoop.java:148)
at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
at org.apache.sqoop.Sqoop.runSqoop(Sqoop.java:184)
at org.apache.sqoop.Sqoop.runTool(Sqoop.java:226)
at org.apache.sqoop.Sqoop.runTool(Sqoop.java:235)
at org.apache.sqoop.Sqoop.main(Sqoop.java:244)
at org.apache.oozie.action.hadoop.SqoopMain.runSqoopJob(SqoopMain.java:197)
at org.apache.oozie.action.hadoop.SqoopMain.run(SqoopMain.java:177)
at org.apache.oozie.action.hadoop.LauncherMain.run(LauncherMain.java:47)
at org.apache.oozie.action.hadoop.SqoopMain.main(SqoopMain.java:46)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.apache.oozie.action.hadoop.LauncherMapper.map(LauncherMapper.java:241)
at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:453)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162)
Caused by: java.sql.SQLException: The database is already in use by another process: org.hsqldb.persist.NIOLockFile@9a5ed4be[file =/home/badevelop/.sqoop/metastore.db.lck, exists=true, locked=false, valid=false, fl =null]: java.lang.Exception: checkHeartbeat(): lock file [/home/david/.sqoop/metastore.db.lck] is presumably locked by another process.
at org.hsqldb.jdbc.Util.sqlException(Unknown Source)
at org.hsqldb.jdbc.jdbcConnection.<init>(Unknown Source)
at org.hsqldb.jdbcDriver.getConnection(Unknown Source)
at org.hsqldb.jdbcDriver.connect(Unknown Source)
at java.sql.DriverManager.getConnection(DriverManager.java:664)
at java.sql.DriverManager.getConnection(DriverManager.java:247)
at org.apache.sqoop.metastore.hsqldb.HsqldbJobStorage.init(HsqldbJobStorage.java:176)
... 25 more
I assume that when one of the processes gets the .lock file in order to access to the hsqldb database, it prevents the other from getting into the metastore, raising the exception. Is this behaviour normal using the default Sqoop configuration? (I haven't modified any config property in the sqoop-site.xml file or in the metastore.db.* scripts). Since the metastore is going to be used by many Oozie workflows, I don't understand why this message seems to imply that the access to the database has to be sequential. Will changing the metastore to another database (for instance MySQL) solve the problem? Any hint would be greatly appreciated.
... View more
Labels:
- Labels:
-
Apache Oozie
-
Apache Sqoop