Member since
11-06-2017
4
Posts
0
Kudos Received
0
Solutions
11-08-2017
01:44 PM
I'd like to thank you sir!!!!! Now both the spark job and the sqoop job can run at the same time. Can you explain to me what exactly this did?
... View more
11-07-2017
07:04 PM
Unfortunately I won't be able to add another Node Manager because we only have 1 host. Adding another host is not ideal for my situation.
... View more
11-07-2017
03:58 PM
I do have a node manager, I will attach a screenshot.
... View more
11-06-2017
08:14 PM
At the moment I have a spark job (java) that will always need to be running. It doesn't need too many resources. However, whenever I run a sqoop job (MapReduce), the job is stuck as ACCEPTED: waiting for AM container to be allocated, launched and register with RM. I checked Ambari and the spark config for scheduling is FAIR. For testing, I tried to run 2 of the same spark job and it ran no problems (state is RUNNING on both). There should be enough cores and memory left for the map reduce job to run. Spark Submit command: /usr/hdp/current/spark-client/bin/spark-submit --class com.some.App --master yarn-cluster --deploy-mode cluster --num-executors 1 /path/to/file.jar "some.server:6667" "Some_App" "Some_App_Parser" "some.server"
jdbc:jtds:sqlserver://some.server:1433/HL7_Metadata
&; done
My sqoop command, I added the memory limit but it didn't help: sqoop import -D mapreduce.map.memory.mb=2048 --connect "jdbc:sqlserver://some.server\SQL2012;database=SomeDB;username=someUser;passwor =somePass" --e "SELECT SOMETHING" where \$CONDITIONS" --fields-terminated-by \002 --escaped-by \ --check-column Message_Audit_Log_Id --incremental append --last-value 1 --split-by Message_Audit_Log_Id --target-dir /target/path/
Here are some images for reference:
... View more
Labels: