[root@lhdcsi02v zeppelin]# tail -1000f zeppelin-interpreter-spark2-spark-zeppelin-lhdcsi02v.production.local.log INFO [2019-04-24 13:48:28,520] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@31bd94e8{/jobs/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,521] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@607f6387{/jobs/job,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,523] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@17d9a6fd{/jobs/job/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,524] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@4a1e7a3a{/stages,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,525] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@51664ba5{/stages/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,526] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@2245c0ca{/stages/stage,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,528] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@22df4e49{/stages/stage/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,529] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@2e908948{/stages/pool,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,530] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@5db3a047{/stages/pool/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,531] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@24b8afb8{/storage,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,532] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@7c4d1060{/storage/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,533] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@56bd1c68{/storage/rdd,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,535] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@16cc2f60{/storage/rdd/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,536] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@7768876e{/environment,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,537] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@633257f3{/environment/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,538] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@5d5249a1{/executors,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,539] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@2a92ca44{/executors/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,541] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@19f27b8f{/executors/threadDump,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,542] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@680e2a8d{/executors/threadDump/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,555] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@275e696c{/static,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,556] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@46431fca{/,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,559] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@d6e2ec6{/api,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,560] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@7338c6e1{/jobs/job/kill,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,561] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@57dc5515{/stages/stage/kill,null,AVAILABLE,@Spark} INFO [2019-04-24 13:48:28,565] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Bound SparkUI to 0.0.0.0, and started at http://lhdcsi02v.production.local:4041 INFO [2019-04-24 13:48:28,606] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Added JAR file:/usr/hdp/current/zeppelin-server/interpreter/spark/zeppelin-spark_2.11-0.7.3.2.6.5.0-292.jar at spark://lhdcsi02v.production.local:35331/jars/zeppelin-spark_2.11-0.7.3.2.6.5.0-292.jar with timestamp 1556110108605 WARN [2019-04-24 13:48:28,665] ({pool-3-thread-2} Logging.scala[logWarning]:66) - Fair Scheduler configuration file not found so jobs will be scheduled in FIFO order. To use fair scheduling, configure pools in fairscheduler.xml or set spark.scheduler.allocation.file to a file that contains the configuration. INFO [2019-04-24 13:48:28,670] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Created default pool: default, schedulingMode: FIFO, minShare: 0, weight: 1 INFO [2019-04-24 13:48:30,208] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Attempting to login to the Kerberos using principal: zeppelin-pruuk_cluster@production.local and keytab: /etc/security/keytabs/zeppelin.server.kerberos.keytab INFO [2019-04-24 13:48:30,218] ({pool-3-thread-2} RMProxy.java[newProxyInstance]:125) - Connecting to ResourceManager at lhdcsi04v.production.local/10.237.14.24:8032 ERROR [2019-04-24 13:48:30,494] ({pool-3-thread-2} Logging.scala[logError]:91) - Error initializing SparkContext. org.apache.hadoop.security.AccessControlException: SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:423) at org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53) at org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:104) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getClusterMetrics(ApplicationClientProtocolPBClientImpl.java:209) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:290) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:202) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:184) at com.sun.proxy.$Proxy16.getClusterMetrics(Unknown Source) at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getYarnClusterMetrics(YarnClientImpl.java:524) at org.apache.spark.deploy.yarn.Client$$anonfun$submitApplication$1.apply(Client.scala:155) at org.apache.spark.deploy.yarn.Client$$anonfun$submitApplication$1.apply(Client.scala:155) at org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) at org.apache.spark.deploy.yarn.Client.logInfo(Client.scala:59) at org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:154) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:57) at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:164) at org.apache.spark.SparkContext.(SparkContext.scala:500) at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2493) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:930) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:921) at scala.Option.getOrElse(Option.scala:121) at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:921) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkSession(SparkInterpreter.java:318) at org.apache.zeppelin.spark.SparkInterpreter.getSparkSession(SparkInterpreter.java:189) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:788) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) at org.apache.zeppelin.scheduler.Job.run(Job.java:175) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1554) at org.apache.hadoop.ipc.Client.call(Client.java:1498) at org.apache.hadoop.ipc.Client.call(Client.java:1398) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:233) at com.sun.proxy.$Proxy15.getClusterMetrics(Unknown Source) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getClusterMetrics(ApplicationClientProtocolPBClientImpl.java:206) ... 46 more INFO [2019-04-24 13:48:30,513] ({pool-3-thread-2} AbstractConnector.java[doStop]:318) - Stopped Spark@7b7e4ceb{HTTP/1.1,[http/1.1]}{0.0.0.0:4041} INFO [2019-04-24 13:48:30,520] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Stopped Spark web UI at http://lhdcsi02v.production.local:4041 WARN [2019-04-24 13:48:30,538] ({dispatcher-event-loop-1} Logging.scala[logWarning]:66) - Attempted to request executors before the AM has registered! INFO [2019-04-24 13:48:30,544] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Stopped INFO [2019-04-24 13:48:30,558] ({dispatcher-event-loop-3} Logging.scala[logInfo]:54) - MapOutputTrackerMasterEndpoint stopped! INFO [2019-04-24 13:48:30,570] ({pool-3-thread-2} Logging.scala[logInfo]:54) - MemoryStore cleared INFO [2019-04-24 13:48:30,571] ({pool-3-thread-2} Logging.scala[logInfo]:54) - BlockManager stopped INFO [2019-04-24 13:48:30,592] ({pool-3-thread-2} Logging.scala[logInfo]:54) - BlockManagerMaster stopped WARN [2019-04-24 13:48:30,593] ({pool-3-thread-2} Logging.scala[logWarning]:66) - Stopping a MetricsSystem that is not running INFO [2019-04-24 13:48:30,602] ({dispatcher-event-loop-0} Logging.scala[logInfo]:54) - OutputCommitCoordinator stopped! INFO [2019-04-24 13:48:30,609] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Successfully stopped SparkContext ERROR [2019-04-24 13:48:30,609] ({pool-3-thread-2} Utils.java[invokeMethod]:40) - java.lang.reflect.InvocationTargetException at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkSession(SparkInterpreter.java:318) at org.apache.zeppelin.spark.SparkInterpreter.getSparkSession(SparkInterpreter.java:189) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:788) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) at org.apache.zeppelin.scheduler.Job.run(Job.java:175) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: org.apache.hadoop.security.AccessControlException: SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:423) at org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53) at org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:104) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getClusterMetrics(ApplicationClientProtocolPBClientImpl.java:209) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:290) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:202) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:184) at com.sun.proxy.$Proxy16.getClusterMetrics(Unknown Source) at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getYarnClusterMetrics(YarnClientImpl.java:524) at org.apache.spark.deploy.yarn.Client$$anonfun$submitApplication$1.apply(Client.scala:155) at org.apache.spark.deploy.yarn.Client$$anonfun$submitApplication$1.apply(Client.scala:155) at org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) at org.apache.spark.deploy.yarn.Client.logInfo(Client.scala:59) at org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:154) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:57) at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:164) at org.apache.spark.SparkContext.(SparkContext.scala:500) at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2493) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:930) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:921) at scala.Option.getOrElse(Option.scala:121) at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:921) ... 24 more Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1554) at org.apache.hadoop.ipc.Client.call(Client.java:1498) at org.apache.hadoop.ipc.Client.call(Client.java:1398) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:233) at com.sun.proxy.$Proxy15.getClusterMetrics(Unknown Source) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getClusterMetrics(ApplicationClientProtocolPBClientImpl.java:206) ... 46 more INFO [2019-04-24 13:48:30,611] ({pool-3-thread-2} SparkInterpreter.java[createSparkSession]:319) - Created Spark session with Hive support ERROR [2019-04-24 13:48:30,611] ({pool-3-thread-2} PySparkInterpreter.java[open]:165) - Error java.lang.NullPointerException at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext_2(SparkInterpreter.java:348) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:337) at org.apache.zeppelin.spark.SparkInterpreter.getSparkContext(SparkInterpreter.java:142) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:790) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) at org.apache.zeppelin.scheduler.Job.run(Job.java:175) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) ERROR [2019-04-24 13:48:30,613] ({pool-3-thread-2} Job.java[run]:188) - Job failed org.apache.zeppelin.interpreter.InterpreterException: java.lang.NullPointerException at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:166) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) at org.apache.zeppelin.scheduler.Job.run(Job.java:175) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.NullPointerException at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext_2(SparkInterpreter.java:348) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:337) at org.apache.zeppelin.spark.SparkInterpreter.getSparkContext(SparkInterpreter.java:142) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:790) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) ... 11 more INFO [2019-04-24 13:48:30,626] ({pool-3-thread-2} SchedulerFactory.java[jobFinished]:137) - Job remoteInterpretJob_1556110102859 finished by scheduler interpreter_1939533868 INFO [2019-04-24 13:51:39,675] ({pool-3-thread-5} SchedulerFactory.java[jobStarted]:131) - Job remoteInterpretJob_1556110299674 started by scheduler interpreter_1939533868 INFO [2019-04-24 13:51:39,678] ({pool-3-thread-5} PySparkInterpreter.java[createPythonScript]:109) - File /tmp/zeppelin_pyspark-2454150356972787827.py created INFO [2019-04-24 13:51:40,552] ({pool-3-thread-5} SparkInterpreter.java[createSparkSession]:276) - ------ Create new SparkSession yarn ------- WARN [2019-04-24 13:51:40,558] ({pool-3-thread-5} Logging.scala[logWarning]:66) - Another SparkContext is being constructed (or threw an exception in its constructor). This may indicate an error, since only one SparkContext may be running in this JVM (see SPARK-2243). The other SparkContext was created at: org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:921) sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) java.lang.reflect.Method.invoke(Method.java:498) org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) org.apache.zeppelin.spark.SparkInterpreter.createSparkSession(SparkInterpreter.java:318) org.apache.zeppelin.spark.SparkInterpreter.getSparkSession(SparkInterpreter.java:189) org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:788) org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) org.apache.zeppelin.scheduler.Job.run(Job.java:175) org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) java.util.concurrent.FutureTask.run(FutureTask.java:266) INFO [2019-04-24 13:51:40,559] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Running Spark version 2.3.0.2.6.5.0-292 WARN [2019-04-24 13:51:40,559] ({pool-3-thread-5} Logging.scala[logWarning]:66) - spark.master yarn-client is deprecated in Spark 2.0+, please instead use "yarn" with specified deploy mode. INFO [2019-04-24 13:51:40,560] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Submitted application: Zeppelin INFO [2019-04-24 13:51:40,561] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Changing view acls to: zeppelin INFO [2019-04-24 13:51:40,561] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Changing modify acls to: zeppelin INFO [2019-04-24 13:51:40,562] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Changing view acls groups to: INFO [2019-04-24 13:51:40,562] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Changing modify acls groups to: INFO [2019-04-24 13:51:40,562] ({pool-3-thread-5} Logging.scala[logInfo]:54) - SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(zeppelin); groups with view permissions: Set(); users with modify permissions: Set(zeppelin); groups with modify permissions: Set() INFO [2019-04-24 13:51:40,570] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Successfully started service 'sparkDriver' on port 46188. INFO [2019-04-24 13:51:40,573] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Registering MapOutputTracker INFO [2019-04-24 13:51:40,575] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Registering BlockManagerMaster INFO [2019-04-24 13:51:40,575] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information INFO [2019-04-24 13:51:40,576] ({pool-3-thread-5} Logging.scala[logInfo]:54) - BlockManagerMasterEndpoint up INFO [2019-04-24 13:51:40,577] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Created local directory at /tmp/blockmgr-b9ee8f57-80e4-420e-84ab-3a7bf9696154 INFO [2019-04-24 13:51:40,579] ({pool-3-thread-5} Logging.scala[logInfo]:54) - MemoryStore started with capacity 366.3 MB INFO [2019-04-24 13:51:40,694] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Registering OutputCommitCoordinator INFO [2019-04-24 13:51:40,704] ({pool-3-thread-5} Server.java[doStart]:346) - jetty-9.3.z-SNAPSHOT INFO [2019-04-24 13:51:40,706] ({pool-3-thread-5} Server.java[doStart]:414) - Started @201386ms WARN [2019-04-24 13:51:40,707] ({pool-3-thread-5} Logging.scala[logWarning]:66) - Service 'SparkUI' could not bind on port 4040. Attempting port 4041. INFO [2019-04-24 13:51:40,708] ({pool-3-thread-5} AbstractConnector.java[doStart]:278) - Started ServerConnector@af1c2db{HTTP/1.1,[http/1.1]}{0.0.0.0:4041} INFO [2019-04-24 13:51:40,709] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Successfully started service 'SparkUI' on port 4041. INFO [2019-04-24 13:51:40,710] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@62138440{/jobs,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,711] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@3c15b88a{/jobs/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,711] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@452764d8{/jobs/job,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,712] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@41c491bd{/jobs/job/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,713] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@3f0b2c41{/stages,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,713] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@2e4b02c2{/stages/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,714] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@5034a277{/stages/stage,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,714] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@351898b5{/stages/stage/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,715] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@5397d373{/stages/pool,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,716] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@25bc86d1{/stages/pool/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,716] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@506eaf0b{/storage,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,717] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@21424883{/storage/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,718] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@4c231b93{/storage/rdd,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,718] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@de886c8{/storage/rdd/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,719] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@28db7689{/environment,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,719] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@1b6e3541{/environment/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,720] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@4f454a3c{/executors,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,721] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@6524bda{/executors/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,721] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@754e1aa7{/executors/threadDump,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,722] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@1fe0ed72{/executors/threadDump/json,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,723] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@7a115018{/static,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,723] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@6e89d81b{/,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,724] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@6b176c4f{/api,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,725] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@36598da{/jobs/job/kill,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,726] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@230cd1ad{/stages/stage/kill,null,AVAILABLE,@Spark} INFO [2019-04-24 13:51:40,726] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Bound SparkUI to 0.0.0.0, and started at http://lhdcsi02v.production.local:4041 INFO [2019-04-24 13:51:40,753] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Added JAR file:/usr/hdp/current/zeppelin-server/interpreter/spark/zeppelin-spark_2.11-0.7.3.2.6.5.0-292.jar at spark://lhdcsi02v.production.local:46188/jars/zeppelin-spark_2.11-0.7.3.2.6.5.0-292.jar with timestamp 1556110300753 WARN [2019-04-24 13:51:40,763] ({pool-3-thread-5} Logging.scala[logWarning]:66) - Fair Scheduler configuration file not found so jobs will be scheduled in FIFO order. To use fair scheduling, configure pools in fairscheduler.xml or set spark.scheduler.allocation.file to a file that contains the configuration. INFO [2019-04-24 13:51:40,764] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Created default pool: default, schedulingMode: FIFO, minShare: 0, weight: 1 INFO [2019-04-24 13:51:40,821] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Attempting to login to the Kerberos using principal: zeppelin-pruuk_cluster@production.local and keytab: /etc/security/keytabs/zeppelin.server.kerberos.keytab INFO [2019-04-24 13:51:40,824] ({pool-3-thread-5} RMProxy.java[newProxyInstance]:125) - Connecting to ResourceManager at lhdcsi04v.production.local/10.237.14.24:8032 ERROR [2019-04-24 13:51:40,835] ({pool-3-thread-5} Logging.scala[logError]:91) - Error initializing SparkContext. org.apache.hadoop.security.AccessControlException: SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:423) at org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53) at org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:104) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getClusterMetrics(ApplicationClientProtocolPBClientImpl.java:209) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:290) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:202) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:184) at com.sun.proxy.$Proxy16.getClusterMetrics(Unknown Source) at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getYarnClusterMetrics(YarnClientImpl.java:524) at org.apache.spark.deploy.yarn.Client$$anonfun$submitApplication$1.apply(Client.scala:155) at org.apache.spark.deploy.yarn.Client$$anonfun$submitApplication$1.apply(Client.scala:155) at org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) at org.apache.spark.deploy.yarn.Client.logInfo(Client.scala:59) at org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:154) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:57) at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:164) at org.apache.spark.SparkContext.(SparkContext.scala:500) at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2493) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:930) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:921) at scala.Option.getOrElse(Option.scala:121) at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:921) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkSession(SparkInterpreter.java:318) at org.apache.zeppelin.spark.SparkInterpreter.getSparkSession(SparkInterpreter.java:189) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:788) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) at org.apache.zeppelin.scheduler.Job.run(Job.java:175) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1554) at org.apache.hadoop.ipc.Client.call(Client.java:1498) at org.apache.hadoop.ipc.Client.call(Client.java:1398) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:233) at com.sun.proxy.$Proxy15.getClusterMetrics(Unknown Source) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getClusterMetrics(ApplicationClientProtocolPBClientImpl.java:206) ... 46 more INFO [2019-04-24 13:51:40,839] ({pool-3-thread-5} AbstractConnector.java[doStop]:318) - Stopped Spark@af1c2db{HTTP/1.1,[http/1.1]}{0.0.0.0:4041} INFO [2019-04-24 13:51:40,841] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Stopped Spark web UI at http://lhdcsi02v.production.local:4041 WARN [2019-04-24 13:51:40,842] ({dispatcher-event-loop-1} Logging.scala[logWarning]:66) - Attempted to request executors before the AM has registered! INFO [2019-04-24 13:51:40,844] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Stopped INFO [2019-04-24 13:51:40,845] ({dispatcher-event-loop-2} Logging.scala[logInfo]:54) - MapOutputTrackerMasterEndpoint stopped! INFO [2019-04-24 13:51:40,849] ({pool-3-thread-5} Logging.scala[logInfo]:54) - MemoryStore cleared INFO [2019-04-24 13:51:40,850] ({pool-3-thread-5} Logging.scala[logInfo]:54) - BlockManager stopped INFO [2019-04-24 13:51:40,852] ({pool-3-thread-5} Logging.scala[logInfo]:54) - BlockManagerMaster stopped WARN [2019-04-24 13:51:40,853] ({pool-3-thread-5} Logging.scala[logWarning]:66) - Stopping a MetricsSystem that is not running INFO [2019-04-24 13:51:40,854] ({dispatcher-event-loop-0} Logging.scala[logInfo]:54) - OutputCommitCoordinator stopped! INFO [2019-04-24 13:51:40,862] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Successfully stopped SparkContext ERROR [2019-04-24 13:51:40,862] ({pool-3-thread-5} Utils.java[invokeMethod]:40) - java.lang.reflect.InvocationTargetException at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkSession(SparkInterpreter.java:318) at org.apache.zeppelin.spark.SparkInterpreter.getSparkSession(SparkInterpreter.java:189) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:788) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) at org.apache.zeppelin.scheduler.Job.run(Job.java:175) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: org.apache.hadoop.security.AccessControlException: SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:423) at org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53) at org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:104) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getClusterMetrics(ApplicationClientProtocolPBClientImpl.java:209) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:290) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:202) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:184) at com.sun.proxy.$Proxy16.getClusterMetrics(Unknown Source) at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getYarnClusterMetrics(YarnClientImpl.java:524) at org.apache.spark.deploy.yarn.Client$$anonfun$submitApplication$1.apply(Client.scala:155) at org.apache.spark.deploy.yarn.Client$$anonfun$submitApplication$1.apply(Client.scala:155) at org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) at org.apache.spark.deploy.yarn.Client.logInfo(Client.scala:59) at org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:154) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:57) at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:164) at org.apache.spark.SparkContext.(SparkContext.scala:500) at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2493) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:930) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:921) at scala.Option.getOrElse(Option.scala:121) at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:921) ... 24 more Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1554) at org.apache.hadoop.ipc.Client.call(Client.java:1498) at org.apache.hadoop.ipc.Client.call(Client.java:1398) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:233) at com.sun.proxy.$Proxy15.getClusterMetrics(Unknown Source) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getClusterMetrics(ApplicationClientProtocolPBClientImpl.java:206) ... 46 more INFO [2019-04-24 13:51:40,865] ({pool-3-thread-5} SparkInterpreter.java[createSparkSession]:319) - Created Spark session with Hive support ERROR [2019-04-24 13:51:40,865] ({pool-3-thread-5} PySparkInterpreter.java[open]:165) - Error java.lang.NullPointerException at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext_2(SparkInterpreter.java:348) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:337) at org.apache.zeppelin.spark.SparkInterpreter.getSparkContext(SparkInterpreter.java:142) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:790) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) at org.apache.zeppelin.scheduler.Job.run(Job.java:175) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) ERROR [2019-04-24 13:51:40,866] ({pool-3-thread-5} Job.java[run]:188) - Job failed org.apache.zeppelin.interpreter.InterpreterException: java.lang.NullPointerException at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:166) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) at org.apache.zeppelin.scheduler.Job.run(Job.java:175) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.NullPointerException at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext_2(SparkInterpreter.java:348) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:337) at org.apache.zeppelin.spark.SparkInterpreter.getSparkContext(SparkInterpreter.java:142) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:790) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) ... 11 more INFO [2019-04-24 13:51:40,867] ({pool-3-thread-5} SchedulerFactory.java[jobFinished]:137) - Job remoteInterpretJob_1556110299674 finished by scheduler interpreter_1939533868 INFO [2019-04-24 14:09:20,908] ({pool-2-thread-2} InterpreterGroup.java[close]:151) - Close interpreter group 2C4U48MY3_spark2:shared_process INFO [2019-04-24 14:09:23,026] ({pool-1-thread-1} Logging.scala[logInfo]:54) - Shutdown hook called INFO [2019-04-24 14:09:23,030] ({pool-1-thread-1} Logging.scala[logInfo]:54) - Deleting directory /tmp/spark-3430b33d-d2c6-4bcd-8605-01517b66f8ca INFO [2019-04-24 14:09:23,031] ({pool-1-thread-1} Logging.scala[logInfo]:54) - Deleting directory /tmp/spark-82cf09f6-673f-44a3-8faa-2c4438a626d9 INFO [2019-04-24 14:09:23,032] ({pool-1-thread-1} Logging.scala[logInfo]:54) - Deleting directory /tmp/spark-324a2e34-627b-495e-b3c4-e059d28f33d6 INFO [2019-04-24 14:09:48,600] ({Thread-4} RemoteInterpreterServer.java[run]:97) - Starting remote interpreter server on port 38930 INFO [2019-04-24 14:09:49,343] ({pool-2-thread-2} RemoteInterpreterServer.java[createInterpreter]:198) - Instantiate interpreter org.apache.zeppelin.spark.SparkInterpreter INFO [2019-04-24 14:09:49,398] ({pool-2-thread-2} RemoteInterpreterServer.java[createInterpreter]:198) - Instantiate interpreter org.apache.zeppelin.spark.SparkSqlInterpreter INFO [2019-04-24 14:09:49,404] ({pool-2-thread-2} RemoteInterpreterServer.java[createInterpreter]:198) - Instantiate interpreter org.apache.zeppelin.spark.DepInterpreter INFO [2019-04-24 14:09:49,415] ({pool-2-thread-2} RemoteInterpreterServer.java[createInterpreter]:198) - Instantiate interpreter org.apache.zeppelin.spark.PySparkInterpreter INFO [2019-04-24 14:09:49,421] ({pool-2-thread-2} RemoteInterpreterServer.java[createInterpreter]:198) - Instantiate interpreter org.apache.zeppelin.spark.SparkRInterpreter INFO [2019-04-24 14:09:49,529] ({pool-3-thread-2} SchedulerFactory.java[jobStarted]:131) - Job remoteInterpretJob_1556111389526 started by scheduler interpreter_843404039 INFO [2019-04-24 14:09:49,538] ({pool-3-thread-2} PySparkInterpreter.java[createPythonScript]:109) - File /tmp/zeppelin_pyspark-4057001016835477048.py created INFO [2019-04-24 14:09:53,947] ({pool-3-thread-2} SparkInterpreter.java[createSparkSession]:276) - ------ Create new SparkSession yarn ------- INFO [2019-04-24 14:09:54,003] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Running Spark version 2.3.0.2.6.5.0-292 WARN [2019-04-24 14:09:54,035] ({pool-3-thread-2} Logging.scala[logWarning]:66) - spark.master yarn-client is deprecated in Spark 2.0+, please instead use "yarn" with specified deploy mode. INFO [2019-04-24 14:09:54,037] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Submitted application: Zeppelin INFO [2019-04-24 14:09:54,190] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Changing view acls to: zeppelin INFO [2019-04-24 14:09:54,190] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Changing modify acls to: zeppelin INFO [2019-04-24 14:09:54,191] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Changing view acls groups to: INFO [2019-04-24 14:09:54,191] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Changing modify acls groups to: INFO [2019-04-24 14:09:54,197] ({pool-3-thread-2} Logging.scala[logInfo]:54) - SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(zeppelin); groups with view permissions: Set(); users with modify permissions: Set(zeppelin); groups with modify permissions: Set() INFO [2019-04-24 14:09:54,520] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Successfully started service 'sparkDriver' on port 37295. INFO [2019-04-24 14:09:54,556] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Registering MapOutputTracker INFO [2019-04-24 14:09:54,586] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Registering BlockManagerMaster INFO [2019-04-24 14:09:54,590] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information INFO [2019-04-24 14:09:54,591] ({pool-3-thread-2} Logging.scala[logInfo]:54) - BlockManagerMasterEndpoint up INFO [2019-04-24 14:09:54,606] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Created local directory at /tmp/blockmgr-2a62312b-26c9-4be9-8e41-edccaa42b618 INFO [2019-04-24 14:09:54,628] ({pool-3-thread-2} Logging.scala[logInfo]:54) - MemoryStore started with capacity 366.3 MB INFO [2019-04-24 14:09:54,797] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Registering OutputCommitCoordinator INFO [2019-04-24 14:09:54,914] ({pool-3-thread-2} Log.java[initialized]:192) - Logging initialized @9197ms INFO [2019-04-24 14:09:55,015] ({pool-3-thread-2} Server.java[doStart]:346) - jetty-9.3.z-SNAPSHOT INFO [2019-04-24 14:09:55,041] ({pool-3-thread-2} Server.java[doStart]:414) - Started @9323ms WARN [2019-04-24 14:09:55,065] ({pool-3-thread-2} Logging.scala[logWarning]:66) - Service 'SparkUI' could not bind on port 4040. Attempting port 4041. INFO [2019-04-24 14:09:55,073] ({pool-3-thread-2} AbstractConnector.java[doStart]:278) - Started ServerConnector@744537cc{HTTP/1.1,[http/1.1]}{0.0.0.0:4041} INFO [2019-04-24 14:09:55,074] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Successfully started service 'SparkUI' on port 4041. INFO [2019-04-24 14:09:55,123] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@4e06dc80{/jobs,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,124] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@1ef29cfd{/jobs/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,125] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@3346c9e7{/jobs/job,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,126] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@2b18bfb7{/jobs/job/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,127] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@2107cb5d{/stages,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,128] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@240b276b{/stages/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,128] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@4e4eac10{/stages/stage,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,130] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@42ca75fb{/stages/stage/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,131] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@795f6f9e{/stages/pool,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,132] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@73ab01c3{/stages/pool/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,133] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@7542b5fc{/storage,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,133] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@64b66302{/storage/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,134] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@6617d395{/storage/rdd,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,135] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@4d19738d{/storage/rdd/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,136] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@2d14c287{/environment,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,137] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@7a6d0a54{/environment/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,138] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@22bef5f7{/executors,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,139] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@24bd6f36{/executors/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,140] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@2fb681dc{/executors/threadDump,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,141] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@3d9b01d4{/executors/threadDump/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,152] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@685f4add{/static,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,153] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@7d26926b{/,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,155] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@7cf7a128{/api,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,156] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@3770a8e5{/jobs/job/kill,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,157] ({pool-3-thread-2} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@462585d0{/stages/stage/kill,null,AVAILABLE,@Spark} INFO [2019-04-24 14:09:55,160] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Bound SparkUI to 0.0.0.0, and started at http://lhdcsi02v.production.local:4041 INFO [2019-04-24 14:09:55,227] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Added JAR file:/usr/hdp/current/zeppelin-server/interpreter/spark/zeppelin-spark_2.11-0.7.3.2.6.5.0-292.jar at spark://lhdcsi02v.production.local:37295/jars/zeppelin-spark_2.11-0.7.3.2.6.5.0-292.jar with timestamp 1556111395226 WARN [2019-04-24 14:09:55,297] ({pool-3-thread-2} Logging.scala[logWarning]:66) - Fair Scheduler configuration file not found so jobs will be scheduled in FIFO order. To use fair scheduling, configure pools in fairscheduler.xml or set spark.scheduler.allocation.file to a file that contains the configuration. INFO [2019-04-24 14:09:55,303] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Created default pool: default, schedulingMode: FIFO, minShare: 0, weight: 1 INFO [2019-04-24 14:09:56,883] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Attempting to login to the Kerberos using principal: zeppelin-pruuk_cluster@production.local and keytab: /etc/security/keytabs/zeppelin.server.kerberos.keytab INFO [2019-04-24 14:09:56,894] ({pool-3-thread-2} RMProxy.java[newProxyInstance]:125) - Connecting to ResourceManager at lhdcsi04v.production.local/10.237.14.24:8032 ERROR [2019-04-24 14:09:57,171] ({pool-3-thread-2} Logging.scala[logError]:91) - Error initializing SparkContext. org.apache.hadoop.security.AccessControlException: SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:423) at org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53) at org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:104) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getClusterMetrics(ApplicationClientProtocolPBClientImpl.java:209) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:290) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:202) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:184) at com.sun.proxy.$Proxy16.getClusterMetrics(Unknown Source) at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getYarnClusterMetrics(YarnClientImpl.java:524) at org.apache.spark.deploy.yarn.Client$$anonfun$submitApplication$1.apply(Client.scala:155) at org.apache.spark.deploy.yarn.Client$$anonfun$submitApplication$1.apply(Client.scala:155) at org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) at org.apache.spark.deploy.yarn.Client.logInfo(Client.scala:59) at org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:154) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:57) at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:164) at org.apache.spark.SparkContext.(SparkContext.scala:500) at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2493) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:930) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:921) at scala.Option.getOrElse(Option.scala:121) at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:921) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkSession(SparkInterpreter.java:328) at org.apache.zeppelin.spark.SparkInterpreter.getSparkSession(SparkInterpreter.java:189) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:788) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) at org.apache.zeppelin.scheduler.Job.run(Job.java:175) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1554) at org.apache.hadoop.ipc.Client.call(Client.java:1498) at org.apache.hadoop.ipc.Client.call(Client.java:1398) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:233) at com.sun.proxy.$Proxy15.getClusterMetrics(Unknown Source) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getClusterMetrics(ApplicationClientProtocolPBClientImpl.java:206) ... 46 more INFO [2019-04-24 14:09:57,187] ({pool-3-thread-2} AbstractConnector.java[doStop]:318) - Stopped Spark@744537cc{HTTP/1.1,[http/1.1]}{0.0.0.0:4041} INFO [2019-04-24 14:09:57,191] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Stopped Spark web UI at http://lhdcsi02v.production.local:4041 WARN [2019-04-24 14:09:57,226] ({dispatcher-event-loop-1} Logging.scala[logWarning]:66) - Attempted to request executors before the AM has registered! INFO [2019-04-24 14:09:57,229] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Stopped INFO [2019-04-24 14:09:57,239] ({dispatcher-event-loop-3} Logging.scala[logInfo]:54) - MapOutputTrackerMasterEndpoint stopped! INFO [2019-04-24 14:09:57,255] ({pool-3-thread-2} Logging.scala[logInfo]:54) - MemoryStore cleared INFO [2019-04-24 14:09:57,257] ({pool-3-thread-2} Logging.scala[logInfo]:54) - BlockManager stopped INFO [2019-04-24 14:09:57,277] ({pool-3-thread-2} Logging.scala[logInfo]:54) - BlockManagerMaster stopped WARN [2019-04-24 14:09:57,278] ({pool-3-thread-2} Logging.scala[logWarning]:66) - Stopping a MetricsSystem that is not running INFO [2019-04-24 14:09:57,291] ({dispatcher-event-loop-0} Logging.scala[logInfo]:54) - OutputCommitCoordinator stopped! INFO [2019-04-24 14:09:57,299] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Successfully stopped SparkContext ERROR [2019-04-24 14:09:57,299] ({pool-3-thread-2} Utils.java[invokeMethod]:40) - java.lang.reflect.InvocationTargetException at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkSession(SparkInterpreter.java:328) at org.apache.zeppelin.spark.SparkInterpreter.getSparkSession(SparkInterpreter.java:189) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:788) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) at org.apache.zeppelin.scheduler.Job.run(Job.java:175) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: org.apache.hadoop.security.AccessControlException: SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:423) at org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53) at org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:104) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getClusterMetrics(ApplicationClientProtocolPBClientImpl.java:209) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:290) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:202) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:184) at com.sun.proxy.$Proxy16.getClusterMetrics(Unknown Source) at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getYarnClusterMetrics(YarnClientImpl.java:524) at org.apache.spark.deploy.yarn.Client$$anonfun$submitApplication$1.apply(Client.scala:155) at org.apache.spark.deploy.yarn.Client$$anonfun$submitApplication$1.apply(Client.scala:155) at org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) at org.apache.spark.deploy.yarn.Client.logInfo(Client.scala:59) at org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:154) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:57) at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:164) at org.apache.spark.SparkContext.(SparkContext.scala:500) at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2493) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:930) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:921) at scala.Option.getOrElse(Option.scala:121) at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:921) ... 24 more Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1554) at org.apache.hadoop.ipc.Client.call(Client.java:1498) at org.apache.hadoop.ipc.Client.call(Client.java:1398) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:233) at com.sun.proxy.$Proxy15.getClusterMetrics(Unknown Source) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getClusterMetrics(ApplicationClientProtocolPBClientImpl.java:206) ... 46 more INFO [2019-04-24 14:09:57,301] ({pool-3-thread-2} SparkInterpreter.java[createSparkSession]:329) - Created Spark session ERROR [2019-04-24 14:09:57,301] ({pool-3-thread-2} PySparkInterpreter.java[open]:165) - Error java.lang.NullPointerException at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext_2(SparkInterpreter.java:348) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:337) at org.apache.zeppelin.spark.SparkInterpreter.getSparkContext(SparkInterpreter.java:142) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:790) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) at org.apache.zeppelin.scheduler.Job.run(Job.java:175) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) ERROR [2019-04-24 14:09:57,302] ({pool-3-thread-2} Job.java[run]:188) - Job failed org.apache.zeppelin.interpreter.InterpreterException: java.lang.NullPointerException at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:166) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) at org.apache.zeppelin.scheduler.Job.run(Job.java:175) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.NullPointerException at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext_2(SparkInterpreter.java:348) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:337) at org.apache.zeppelin.spark.SparkInterpreter.getSparkContext(SparkInterpreter.java:142) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:790) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) ... 11 more INFO [2019-04-24 14:09:57,316] ({pool-3-thread-2} SchedulerFactory.java[jobFinished]:137) - Job remoteInterpretJob_1556111389526 finished by scheduler interpreter_843404039 INFO [2019-04-24 14:18:19,729] ({pool-2-thread-2} InterpreterGroup.java[close]:151) - Close interpreter group 2C4U48MY3_spark2:shared_process INFO [2019-04-24 14:18:21,842] ({pool-1-thread-1} Logging.scala[logInfo]:54) - Shutdown hook called INFO [2019-04-24 14:18:21,843] ({pool-1-thread-1} Logging.scala[logInfo]:54) - Deleting directory /tmp/spark-581e2784-ad50-4610-afab-2370de85853b INFO [2019-04-24 14:18:21,844] ({pool-1-thread-1} Logging.scala[logInfo]:54) - Deleting directory /tmp/spark-1a391e47-a945-402d-a80a-d66c263eacd9 INFO [2019-04-24 14:18:21,845] ({pool-1-thread-1} Logging.scala[logInfo]:54) - Deleting directory /tmp/spark-aff6230d-3aa8-439f-8e91-c4e097a7f4b6 INFO [2019-04-24 14:23:50,252] ({Thread-4} RemoteInterpreterServer.java[run]:97) - Starting remote interpreter server on port 33385 INFO [2019-04-24 14:23:51,174] ({pool-2-thread-2} RemoteInterpreterServer.java[createInterpreter]:198) - Instantiate interpreter org.apache.zeppelin.spark.SparkInterpreter INFO [2019-04-24 14:23:51,221] ({pool-2-thread-2} RemoteInterpreterServer.java[createInterpreter]:198) - Instantiate interpreter org.apache.zeppelin.spark.SparkSqlInterpreter INFO [2019-04-24 14:23:51,229] ({pool-2-thread-2} RemoteInterpreterServer.java[createInterpreter]:198) - Instantiate interpreter org.apache.zeppelin.spark.DepInterpreter INFO [2019-04-24 14:23:51,240] ({pool-2-thread-2} RemoteInterpreterServer.java[createInterpreter]:198) - Instantiate interpreter org.apache.zeppelin.spark.PySparkInterpreter INFO [2019-04-24 14:23:51,244] ({pool-2-thread-2} RemoteInterpreterServer.java[createInterpreter]:198) - Instantiate interpreter org.apache.zeppelin.spark.SparkRInterpreter INFO [2019-04-24 14:23:51,388] ({pool-3-thread-5} SchedulerFactory.java[jobStarted]:131) - Job remoteInterpretJob_1556112231383 started by scheduler interpreter_875951848 INFO [2019-04-24 14:23:51,399] ({pool-3-thread-5} PySparkInterpreter.java[createPythonScript]:109) - File /tmp/zeppelin_pyspark-6688962895170432982.py created INFO [2019-04-24 14:23:55,820] ({pool-3-thread-5} SparkInterpreter.java[createSparkSession]:276) - ------ Create new SparkSession yarn ------- INFO [2019-04-24 14:23:55,878] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Running Spark version 2.3.0.2.6.5.0-292 WARN [2019-04-24 14:23:55,910] ({pool-3-thread-5} Logging.scala[logWarning]:66) - spark.master yarn-client is deprecated in Spark 2.0+, please instead use "yarn" with specified deploy mode. INFO [2019-04-24 14:23:55,913] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Submitted application: Zeppelin INFO [2019-04-24 14:23:56,073] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Changing view acls to: zeppelin INFO [2019-04-24 14:23:56,073] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Changing modify acls to: zeppelin INFO [2019-04-24 14:23:56,074] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Changing view acls groups to: INFO [2019-04-24 14:23:56,074] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Changing modify acls groups to: INFO [2019-04-24 14:23:56,076] ({pool-3-thread-5} Logging.scala[logInfo]:54) - SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(zeppelin); groups with view permissions: Set(); users with modify permissions: Set(zeppelin); groups with modify permissions: Set() INFO [2019-04-24 14:23:56,410] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Successfully started service 'sparkDriver' on port 35872. INFO [2019-04-24 14:23:56,490] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Registering MapOutputTracker INFO [2019-04-24 14:23:56,557] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Registering BlockManagerMaster INFO [2019-04-24 14:23:56,567] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information INFO [2019-04-24 14:23:56,569] ({pool-3-thread-5} Logging.scala[logInfo]:54) - BlockManagerMasterEndpoint up INFO [2019-04-24 14:23:56,601] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Created local directory at /tmp/blockmgr-084c4d48-54ba-402c-b2f0-d78169c4da85 INFO [2019-04-24 14:23:56,640] ({pool-3-thread-5} Logging.scala[logInfo]:54) - MemoryStore started with capacity 366.3 MB INFO [2019-04-24 14:23:56,810] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Registering OutputCommitCoordinator INFO [2019-04-24 14:23:56,928] ({pool-3-thread-5} Log.java[initialized]:192) - Logging initialized @9281ms INFO [2019-04-24 14:23:57,029] ({pool-3-thread-5} Server.java[doStart]:346) - jetty-9.3.z-SNAPSHOT INFO [2019-04-24 14:23:57,068] ({pool-3-thread-5} Server.java[doStart]:414) - Started @9420ms WARN [2019-04-24 14:23:57,120] ({pool-3-thread-5} Logging.scala[logWarning]:66) - Service 'SparkUI' could not bind on port 4040. Attempting port 4041. INFO [2019-04-24 14:23:57,139] ({pool-3-thread-5} AbstractConnector.java[doStart]:278) - Started ServerConnector@5662482e{HTTP/1.1,[http/1.1]}{0.0.0.0:4041} INFO [2019-04-24 14:23:57,139] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Successfully started service 'SparkUI' on port 4041. INFO [2019-04-24 14:23:57,232] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@755717e5{/jobs,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,234] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@4f1586b{/jobs/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,236] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@52ab1573{/jobs/job,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,239] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@7b26acd1{/jobs/job/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,240] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@759a81de{/stages,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,242] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@7f100293{/stages/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,244] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@619ffca{/stages/stage,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,247] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@16d9f47c{/stages/stage/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,249] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@5ff7bec2{/stages/pool,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,251] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@2bcb96bb{/stages/pool/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,253] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@512974ec{/storage,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,255] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@88a00a1{/storage/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,256] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@6ae9aedc{/storage/rdd,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,258] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@1d609374{/storage/rdd/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,258] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@7741a20{/environment,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,259] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@5fa39d67{/environment/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,260] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@781b4701{/executors,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,261] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@61798045{/executors/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,262] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@467b2350{/executors/threadDump,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,263] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@570c65d{/executors/threadDump/json,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,273] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@27f1aabd{/static,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,274] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@4e6228eb{/,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,275] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@189fc5d0{/api,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,276] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@d7bd664{/jobs/job/kill,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,277] ({pool-3-thread-5} ContextHandler.java[doStart]:781) - Started o.s.j.s.ServletContextHandler@29e83b7d{/stages/stage/kill,null,AVAILABLE,@Spark} INFO [2019-04-24 14:23:57,280] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Bound SparkUI to 0.0.0.0, and started at http://lhdcsi02v.production.local:4041 INFO [2019-04-24 14:23:57,310] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Added JAR file:/usr/hdp/current/zeppelin-server/interpreter/spark/zeppelin-spark_2.11-0.7.3.2.6.5.0-292.jar at spark://lhdcsi02v.production.local:35872/jars/zeppelin-spark_2.11-0.7.3.2.6.5.0-292.jar with timestamp 1556112237309 WARN [2019-04-24 14:23:57,398] ({pool-3-thread-5} Logging.scala[logWarning]:66) - Fair Scheduler configuration file not found so jobs will be scheduled in FIFO order. To use fair scheduling, configure pools in fairscheduler.xml or set spark.scheduler.allocation.file to a file that contains the configuration. INFO [2019-04-24 14:23:57,408] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Created default pool: default, schedulingMode: FIFO, minShare: 0, weight: 1 INFO [2019-04-24 14:23:59,067] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Attempting to login to the Kerberos using principal: zeppelin-pruuk_cluster@production.local and keytab: /etc/security/keytabs/zeppelin.server.kerberos.keytab INFO [2019-04-24 14:23:59,078] ({pool-3-thread-5} RMProxy.java[newProxyInstance]:125) - Connecting to ResourceManager at lhdcsi04v.production.local/10.237.14.24:8032 ERROR [2019-04-24 14:23:59,354] ({pool-3-thread-5} Logging.scala[logError]:91) - Error initializing SparkContext. org.apache.hadoop.security.AccessControlException: SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:423) at org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53) at org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:104) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getClusterMetrics(ApplicationClientProtocolPBClientImpl.java:209) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:290) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:202) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:184) at com.sun.proxy.$Proxy16.getClusterMetrics(Unknown Source) at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getYarnClusterMetrics(YarnClientImpl.java:524) at org.apache.spark.deploy.yarn.Client$$anonfun$submitApplication$1.apply(Client.scala:155) at org.apache.spark.deploy.yarn.Client$$anonfun$submitApplication$1.apply(Client.scala:155) at org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) at org.apache.spark.deploy.yarn.Client.logInfo(Client.scala:59) at org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:154) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:57) at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:164) at org.apache.spark.SparkContext.(SparkContext.scala:500) at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2493) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:930) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:921) at scala.Option.getOrElse(Option.scala:121) at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:921) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkSession(SparkInterpreter.java:328) at org.apache.zeppelin.spark.SparkInterpreter.getSparkSession(SparkInterpreter.java:189) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:788) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) at org.apache.zeppelin.scheduler.Job.run(Job.java:175) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1554) at org.apache.hadoop.ipc.Client.call(Client.java:1498) at org.apache.hadoop.ipc.Client.call(Client.java:1398) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:233) at com.sun.proxy.$Proxy15.getClusterMetrics(Unknown Source) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getClusterMetrics(ApplicationClientProtocolPBClientImpl.java:206) ... 46 more INFO [2019-04-24 14:23:59,371] ({pool-3-thread-5} AbstractConnector.java[doStop]:318) - Stopped Spark@5662482e{HTTP/1.1,[http/1.1]}{0.0.0.0:4041} INFO [2019-04-24 14:23:59,374] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Stopped Spark web UI at http://lhdcsi02v.production.local:4041 WARN [2019-04-24 14:23:59,405] ({dispatcher-event-loop-1} Logging.scala[logWarning]:66) - Attempted to request executors before the AM has registered! INFO [2019-04-24 14:23:59,415] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Stopped INFO [2019-04-24 14:23:59,435] ({dispatcher-event-loop-3} Logging.scala[logInfo]:54) - MapOutputTrackerMasterEndpoint stopped! INFO [2019-04-24 14:23:59,456] ({pool-3-thread-5} Logging.scala[logInfo]:54) - MemoryStore cleared INFO [2019-04-24 14:23:59,457] ({pool-3-thread-5} Logging.scala[logInfo]:54) - BlockManager stopped INFO [2019-04-24 14:23:59,481] ({pool-3-thread-5} Logging.scala[logInfo]:54) - BlockManagerMaster stopped WARN [2019-04-24 14:23:59,482] ({pool-3-thread-5} Logging.scala[logWarning]:66) - Stopping a MetricsSystem that is not running INFO [2019-04-24 14:23:59,495] ({dispatcher-event-loop-0} Logging.scala[logInfo]:54) - OutputCommitCoordinator stopped! INFO [2019-04-24 14:23:59,514] ({pool-3-thread-5} Logging.scala[logInfo]:54) - Successfully stopped SparkContext ERROR [2019-04-24 14:23:59,514] ({pool-3-thread-5} Utils.java[invokeMethod]:40) - java.lang.reflect.InvocationTargetException at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkSession(SparkInterpreter.java:328) at org.apache.zeppelin.spark.SparkInterpreter.getSparkSession(SparkInterpreter.java:189) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:788) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) at org.apache.zeppelin.scheduler.Job.run(Job.java:175) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: org.apache.hadoop.security.AccessControlException: SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:423) at org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53) at org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:104) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getClusterMetrics(ApplicationClientProtocolPBClientImpl.java:209) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:290) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:202) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:184) at com.sun.proxy.$Proxy16.getClusterMetrics(Unknown Source) at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.getYarnClusterMetrics(YarnClientImpl.java:524) at org.apache.spark.deploy.yarn.Client$$anonfun$submitApplication$1.apply(Client.scala:155) at org.apache.spark.deploy.yarn.Client$$anonfun$submitApplication$1.apply(Client.scala:155) at org.apache.spark.internal.Logging$class.logInfo(Logging.scala:54) at org.apache.spark.deploy.yarn.Client.logInfo(Client.scala:59) at org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:154) at org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:57) at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:164) at org.apache.spark.SparkContext.(SparkContext.scala:500) at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2493) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:930) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:921) at scala.Option.getOrElse(Option.scala:121) at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:921) ... 24 more Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): SIMPLE authentication is not enabled. Available:[TOKEN, KERBEROS] at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1554) at org.apache.hadoop.ipc.Client.call(Client.java:1498) at org.apache.hadoop.ipc.Client.call(Client.java:1398) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:233) at com.sun.proxy.$Proxy15.getClusterMetrics(Unknown Source) at org.apache.hadoop.yarn.api.impl.pb.client.ApplicationClientProtocolPBClientImpl.getClusterMetrics(ApplicationClientProtocolPBClientImpl.java:206) ... 46 more INFO [2019-04-24 14:23:59,517] ({pool-3-thread-5} SparkInterpreter.java[createSparkSession]:329) - Created Spark session ERROR [2019-04-24 14:23:59,517] ({pool-3-thread-5} PySparkInterpreter.java[open]:165) - Error java.lang.NullPointerException at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext_2(SparkInterpreter.java:348) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:337) at org.apache.zeppelin.spark.SparkInterpreter.getSparkContext(SparkInterpreter.java:142) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:790) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) at org.apache.zeppelin.scheduler.Job.run(Job.java:175) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) ERROR [2019-04-24 14:23:59,519] ({pool-3-thread-5} Job.java[run]:188) - Job failed org.apache.zeppelin.interpreter.InterpreterException: java.lang.NullPointerException at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:166) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:493) at org.apache.zeppelin.scheduler.Job.run(Job.java:175) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.NullPointerException at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext_2(SparkInterpreter.java:348) at org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:337) at org.apache.zeppelin.spark.SparkInterpreter.getSparkContext(SparkInterpreter.java:142) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:790) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.spark.PySparkInterpreter.getSparkInterpreter(PySparkInterpreter.java:567) at org.apache.zeppelin.spark.PySparkInterpreter.createGatewayServerAndStartScript(PySparkInterpreter.java:210) at org.apache.zeppelin.spark.PySparkInterpreter.open(PySparkInterpreter.java:163) ... 11 more INFO [2019-04-24 14:23:59,536] ({pool-3-thread-5} SchedulerFactory.java[jobFinished]:137) - Job remoteInterpretJob_1556112231383 finished by scheduler interpreter_875951848 ^Z