Archives of Support Questions (Read Only)

This is an archived board for historical reference. Information and links may no longer be available or relevant
Announcements
This board is archived and read-only for historical reference. To ask a new question, please post a new topic on the appropriate active board.

livy2 zepplin issue

avatar
New Member

Hi all,

I tried to connect to livy2.spark thru zepplin notebook at HDP 2.6.4

I don't know what configuration should be at interpreter(please check attached image) in Zepplin.

I counldn't cocnnect to livy2.spark, I had error:

org.apache.zeppelin.livy.LivyException: The creation of session 4 is timeout within 120 seconds, appId: null, log: [stdout: , 
stderr: , Warning: Master yarn-cluster is deprecated since 2.0. Please use master "yarn" with specified deploy mode instead., 18/08/15 07:47:28 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable, 18/08/15 07:47:55 WARN DomainSocketFactory: The short-circuit local reads feature cannot be used because libhadoop cannot be loaded., 18/08/15 07:47:56 INFO RMProxy: Connecting to ResourceManager at sandbox-hdp.hortonworks.com/172.18.0.2:8032, 
YARN Diagnostics: ]
	at org.apache.zeppelin.livy.BaseLivyInterpreter.createSession(BaseLivyInterpreter.java:291)
	at org.apache.zeppelin.livy.BaseLivyInterpreter.initLivySession(BaseLivyInterpreter.java:184)
	at org.apache.zeppelin.livy.LivySharedInterpreter.open(LivySharedInterpreter.java:57)
	at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
	at org.apache.zeppelin.livy.BaseLivyInterpreter.getLivySharedInterpreter(BaseLivyInterpreter.java:165)
	at org.apache.zeppelin.livy.BaseLivyInterpreter.open(BaseLivyInterpreter.java:139)
	at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
	at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493)
	at org.apache.zeppelin.scheduler.Job.run(Job.java:175)
	at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139)
	at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
	at java.util.concurrent.FutureTask.run(FutureTask.java:266)
	at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
	at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)

 

capture.png
1 ACCEPTED SOLUTION

avatar

Hello @Maksym Shved!
Right, looking at your screenshot. I'm assuming you're doing this tests in a non-production environment. If so, then change the following parameters:

#Ambari > Spark2 > Advanced livy2-conf
livy.spark.master=yarn 
livy.impersonation.enabled=false
#Ambari > Spark2 > Advanced livy2-env
            export SPARK_HOME=/usr/hdp/current/spark2-client
            export SPARK_CONF_DIR=/etc/spark2/conf
            export JAVA_HOME={{java_home}}
            export HADOOP_CONF_DIR=/etc/hadoop/conf
            export LIVY_LOG_DIR={{livy2_log_dir}}
            export LIVY_PID_DIR={{livy2_pid_dir}}
            export LIVY_SERVER_JAVA_OPTS="-Xmx2g"

Restart the SPARK2 service and make the test again. If you're still facing issues, then go back to the Ambari and set the following.

Ambari > HDFS > Custom Core-Site
hadoop.proxyuser.root.groups=*
hadoop.proxyuser.root.hosts=*

Hope this helps.

View solution in original post

2 REPLIES 2

avatar

Hello @Maksym Shved!
Right, looking at your screenshot. I'm assuming you're doing this tests in a non-production environment. If so, then change the following parameters:

#Ambari > Spark2 > Advanced livy2-conf
livy.spark.master=yarn 
livy.impersonation.enabled=false
#Ambari > Spark2 > Advanced livy2-env
            export SPARK_HOME=/usr/hdp/current/spark2-client
            export SPARK_CONF_DIR=/etc/spark2/conf
            export JAVA_HOME={{java_home}}
            export HADOOP_CONF_DIR=/etc/hadoop/conf
            export LIVY_LOG_DIR={{livy2_log_dir}}
            export LIVY_PID_DIR={{livy2_pid_dir}}
            export LIVY_SERVER_JAVA_OPTS="-Xmx2g"

Restart the SPARK2 service and make the test again. If you're still facing issues, then go back to the Ambari and set the following.

Ambari > HDFS > Custom Core-Site
hadoop.proxyuser.root.groups=*
hadoop.proxyuser.root.hosts=*

Hope this helps.

avatar
New Member

Thanks a lot!

@Vinicius Higa Murakami, I have one small question, how I can connect to livy2 server?

I need to know some parameters to connect(look at the screenshot), but I don't know where I could find them.

87413-1.png

I'm work at SandBox 2.6.5(last version)

Please help me, if u can.

Thanks!