<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question ERROR SparkContext: Error initializing SparkContext. in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/ERROR-SparkContext-Error-initializing-SparkContext/m-p/369869#M240580</link>
    <description>&lt;P&gt;HI all,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I'm getting the following error when trying to launch pyspark. Please help !! Thanks in advance.&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;C:\Users\abhay&amp;gt;pyspark&lt;BR /&gt;Python 3.10.3 (tags/v3.10.3:a342a49, Mar 16 2022, 13:07:40) [MSC v.1929 64 bit (AMD64)] on win32&lt;BR /&gt;Type "help", "copyright", "credits" or "license" for more information.&lt;BR /&gt;WARNING: An illegal reflective access operation has occurred&lt;BR /&gt;WARNING: Illegal reflective access by org.apache.spark.unsafe.Platform (file:/D:/demo/spark-3.2.1/jars/spark-unsafe_2.12-3.2.1.jar) to constructor java.nio.DirectByteBuffer(long,int)&lt;BR /&gt;WARNING: Please consider reporting this to the maintainers of org.apache.spark.unsafe.Platform&lt;BR /&gt;WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations&lt;BR /&gt;WARNING: All illegal access operations will be denied in a future release&lt;BR /&gt;Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties&lt;BR /&gt;Setting default log level to "WARN".&lt;BR /&gt;To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).&lt;BR /&gt;23/05/03 09:04:32 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable&lt;BR /&gt;23/05/03 09:04:34 ERROR SparkContext: Error initializing SparkContext.&lt;BR /&gt;org.apache.spark.SparkException: Invalid Spark URL: spark://HeartbeatReceiver@Abhay_Mishra:54720&lt;BR /&gt;at org.apache.spark.rpc.RpcEndpointAddress$.apply(RpcEndpointAddress.scala:66)&lt;BR /&gt;at org.apache.spark.rpc.netty.NettyRpcEnv.asyncSetupEndpointRefByURI(NettyRpcEnv.scala:140)&lt;BR /&gt;at org.apache.spark.rpc.RpcEnv.setupEndpointRefByURI(RpcEnv.scala:101)&lt;BR /&gt;at org.apache.spark.rpc.RpcEnv.setupEndpointRef(RpcEnv.scala:109)&lt;BR /&gt;at org.apache.spark.util.RpcUtils$.makeDriverRef(RpcUtils.scala:36)&lt;BR /&gt;at org.apache.spark.executor.Executor.&amp;lt;init&amp;gt;(Executor.scala:218)&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalEndpoint.&amp;lt;init&amp;gt;(LocalSchedulerBackend.scala:64)&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalSchedulerBackend.start(LocalSchedulerBackend.scala:132)&lt;BR /&gt;at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:220)&lt;BR /&gt;at org.apache.spark.SparkContext.&amp;lt;init&amp;gt;(SparkContext.scala:581)&lt;BR /&gt;at org.apache.spark.api.java.JavaSparkContext.&amp;lt;init&amp;gt;(JavaSparkContext.scala:58)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)&lt;BR /&gt;at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)&lt;BR /&gt;at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)&lt;BR /&gt;at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:247)&lt;BR /&gt;at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)&lt;BR /&gt;at py4j.Gateway.invoke(Gateway.java:238)&lt;BR /&gt;at py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)&lt;BR /&gt;at py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)&lt;BR /&gt;at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)&lt;BR /&gt;at py4j.ClientServerConnection.run(ClientServerConnection.java:106)&lt;BR /&gt;at java.base/java.lang.Thread.run(Thread.java:834)&lt;BR /&gt;23/05/03 09:04:35 ERROR Utils: Uncaught exception in thread Thread-2&lt;BR /&gt;java.lang.NullPointerException&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalSchedulerBackend.org$apache$spark$scheduler$local$LocalSchedulerBackend$$stop(LocalSchedulerBackend.scala:173)&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalSchedulerBackend.stop(LocalSchedulerBackend.scala:144)&lt;BR /&gt;at org.apache.spark.scheduler.TaskSchedulerImpl.stop(TaskSchedulerImpl.scala:927)&lt;BR /&gt;at org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:2567)&lt;BR /&gt;at org.apache.spark.SparkContext.$anonfun$stop$12(SparkContext.scala:2086)&lt;BR /&gt;at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1442)&lt;BR /&gt;at org.apache.spark.SparkContext.stop(SparkContext.scala:2086)&lt;BR /&gt;at org.apache.spark.SparkContext.&amp;lt;init&amp;gt;(SparkContext.scala:677)&lt;BR /&gt;at org.apache.spark.api.java.JavaSparkContext.&amp;lt;init&amp;gt;(JavaSparkContext.scala:58)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)&lt;BR /&gt;at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)&lt;BR /&gt;at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)&lt;BR /&gt;at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:247)&lt;BR /&gt;at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)&lt;BR /&gt;at py4j.Gateway.invoke(Gateway.java:238)&lt;BR /&gt;at py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)&lt;BR /&gt;at py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)&lt;BR /&gt;at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)&lt;BR /&gt;at py4j.ClientServerConnection.run(ClientServerConnection.java:106)&lt;BR /&gt;at java.base/java.lang.Thread.run(Thread.java:834)&lt;BR /&gt;23/05/03 09:04:35 WARN MetricsSystem: Stopping a MetricsSystem that is not running&lt;BR /&gt;23/05/03 09:04:35 WARN SparkContext: Another SparkContext is being constructed (or threw an exception in its constructor). This may indicate an error, since only one SparkContext should be running in this JVM (see SPARK-2243). The other SparkContext was created at:&lt;BR /&gt;org.apache.spark.api.java.JavaSparkContext.&amp;lt;init&amp;gt;(JavaSparkContext.scala:58)&lt;BR /&gt;java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)&lt;BR /&gt;java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)&lt;BR /&gt;java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)&lt;BR /&gt;java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)&lt;BR /&gt;py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:247)&lt;BR /&gt;py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)&lt;BR /&gt;py4j.Gateway.invoke(Gateway.java:238)&lt;BR /&gt;py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)&lt;BR /&gt;py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)&lt;BR /&gt;py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)&lt;BR /&gt;py4j.ClientServerConnection.run(ClientServerConnection.java:106)&lt;BR /&gt;java.base/java.lang.Thread.run(Thread.java:834)&lt;BR /&gt;23/05/03 09:04:35 ERROR SparkContext: Error initializing SparkContext.&lt;BR /&gt;org.apache.spark.SparkException: Invalid Spark URL: spark://HeartbeatReceiver@Abhay_Mishra:54743&lt;BR /&gt;at org.apache.spark.rpc.RpcEndpointAddress$.apply(RpcEndpointAddress.scala:66)&lt;BR /&gt;at org.apache.spark.rpc.netty.NettyRpcEnv.asyncSetupEndpointRefByURI(NettyRpcEnv.scala:140)&lt;BR /&gt;at org.apache.spark.rpc.RpcEnv.setupEndpointRefByURI(RpcEnv.scala:101)&lt;BR /&gt;at org.apache.spark.rpc.RpcEnv.setupEndpointRef(RpcEnv.scala:109)&lt;BR /&gt;at org.apache.spark.util.RpcUtils$.makeDriverRef(RpcUtils.scala:36)&lt;BR /&gt;at org.apache.spark.executor.Executor.&amp;lt;init&amp;gt;(Executor.scala:218)&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalEndpoint.&amp;lt;init&amp;gt;(LocalSchedulerBackend.scala:64)&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalSchedulerBackend.start(LocalSchedulerBackend.scala:132)&lt;BR /&gt;at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:220)&lt;BR /&gt;at org.apache.spark.SparkContext.&amp;lt;init&amp;gt;(SparkContext.scala:581)&lt;BR /&gt;at org.apache.spark.api.java.JavaSparkContext.&amp;lt;init&amp;gt;(JavaSparkContext.scala:58)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)&lt;BR /&gt;at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)&lt;BR /&gt;at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)&lt;BR /&gt;at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:247)&lt;BR /&gt;at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)&lt;BR /&gt;at py4j.Gateway.invoke(Gateway.java:238)&lt;BR /&gt;at py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)&lt;BR /&gt;at py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)&lt;BR /&gt;at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)&lt;BR /&gt;at py4j.ClientServerConnection.run(ClientServerConnection.java:106)&lt;BR /&gt;at java.base/java.lang.Thread.run(Thread.java:834)&lt;BR /&gt;23/05/03 09:04:35 ERROR Utils: Uncaught exception in thread Thread-2&lt;BR /&gt;java.lang.NullPointerException&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalSchedulerBackend.org$apache$spark$scheduler$local$LocalSchedulerBackend$$stop(LocalSchedulerBackend.scala:173)&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalSchedulerBackend.stop(LocalSchedulerBackend.scala:144)&lt;BR /&gt;at org.apache.spark.scheduler.TaskSchedulerImpl.stop(TaskSchedulerImpl.scala:927)&lt;BR /&gt;at org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:2567)&lt;BR /&gt;at org.apache.spark.SparkContext.$anonfun$stop$12(SparkContext.scala:2086)&lt;BR /&gt;at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1442)&lt;BR /&gt;at org.apache.spark.SparkContext.stop(SparkContext.scala:2086)&lt;BR /&gt;at org.apache.spark.SparkContext.&amp;lt;init&amp;gt;(SparkContext.scala:677)&lt;BR /&gt;at org.apache.spark.api.java.JavaSparkContext.&amp;lt;init&amp;gt;(JavaSparkContext.scala:58)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)&lt;BR /&gt;at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)&lt;BR /&gt;at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)&lt;BR /&gt;at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:247)&lt;BR /&gt;at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)&lt;BR /&gt;at py4j.Gateway.invoke(Gateway.java:238)&lt;BR /&gt;at py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)&lt;BR /&gt;at py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)&lt;BR /&gt;at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)&lt;BR /&gt;at py4j.ClientServerConnection.run(ClientServerConnection.java:106)&lt;BR /&gt;at java.base/java.lang.Thread.run(Thread.java:834)&lt;BR /&gt;23/05/03 09:04:35 WARN MetricsSystem: Stopping a MetricsSystem that is not running&lt;BR /&gt;D:\demo\spark-3.2.1\python\pyspark\shell.py:42: UserWarning: Failed to initialize Spark session.&lt;BR /&gt;warnings.warn("Failed to initialize Spark session.")&lt;BR /&gt;Traceback (most recent call last):&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\pyspark\shell.py", line 38, in &amp;lt;module&amp;gt;&lt;BR /&gt;spark = SparkSession._create_shell_session() # type: ignore&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\pyspark\sql\session.py", line 553, in _create_shell_session&lt;BR /&gt;return SparkSession.builder.getOrCreate()&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\pyspark\sql\session.py", line 228, in getOrCreate&lt;BR /&gt;sc = SparkContext.getOrCreate(sparkConf)&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\pyspark\context.py", line 392, in getOrCreate&lt;BR /&gt;SparkContext(conf=conf or SparkConf())&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\pyspark\context.py", line 146, in __init__&lt;BR /&gt;self._do_init(master, appName, sparkHome, pyFiles, environment, batchSize, serializer,&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\pyspark\context.py", line 209, in _do_init&lt;BR /&gt;self._jsc = jsc or self._initialize_context(self._conf._jconf)&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\pyspark\context.py", line 329, in _initialize_context&lt;BR /&gt;return self._jvm.JavaSparkContext(jconf)&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\lib\py4j-0.10.9.3-src.zip\py4j\java_gateway.py", line 1585, in __call__&lt;BR /&gt;return_value = get_return_value(&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\lib\py4j-0.10.9.3-src.zip\py4j\protocol.py", line 326, in get_return_value&lt;BR /&gt;raise Py4JJavaError(&lt;BR /&gt;py4j.protocol.Py4JJavaError: An error occurred while calling None.org.apache.spark.api.java.JavaSparkContext.&lt;BR /&gt;: org.apache.spark.SparkException: Invalid Spark URL: spark://HeartbeatReceiver@Abhay_Mishra:54743&lt;BR /&gt;at org.apache.spark.rpc.RpcEndpointAddress$.apply(RpcEndpointAddress.scala:66)&lt;BR /&gt;at org.apache.spark.rpc.netty.NettyRpcEnv.asyncSetupEndpointRefByURI(NettyRpcEnv.scala:140)&lt;BR /&gt;at org.apache.spark.rpc.RpcEnv.setupEndpointRefByURI(RpcEnv.scala:101)&lt;BR /&gt;at org.apache.spark.rpc.RpcEnv.setupEndpointRef(RpcEnv.scala:109)&lt;BR /&gt;at org.apache.spark.util.RpcUtils$.makeDriverRef(RpcUtils.scala:36)&lt;BR /&gt;at org.apache.spark.executor.Executor.&amp;lt;init&amp;gt;(Executor.scala:218)&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalEndpoint.&amp;lt;init&amp;gt;(LocalSchedulerBackend.scala:64)&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalSchedulerBackend.start(LocalSchedulerBackend.scala:132)&lt;BR /&gt;at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:220)&lt;BR /&gt;at org.apache.spark.SparkContext.&amp;lt;init&amp;gt;(SparkContext.scala:581)&lt;BR /&gt;at org.apache.spark.api.java.JavaSparkContext.&amp;lt;init&amp;gt;(JavaSparkContext.scala:58)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)&lt;BR /&gt;at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)&lt;BR /&gt;at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)&lt;BR /&gt;at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:247)&lt;BR /&gt;at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)&lt;BR /&gt;at py4j.Gateway.invoke(Gateway.java:238)&lt;BR /&gt;at py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)&lt;BR /&gt;at py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)&lt;BR /&gt;at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)&lt;BR /&gt;at py4j.ClientServerConnection.run(ClientServerConnection.java:106)&lt;BR /&gt;at java.base/java.lang.Thread.run(Thread.java:834)&lt;/P&gt;&lt;P&gt;23/05/03 09:04:35 ERROR Utils: Uncaught exception in thread shutdown-hook-0&lt;BR /&gt;java.lang.ExceptionInInitializerError&lt;BR /&gt;at org.apache.spark.executor.Executor.stop(Executor.scala:333)&lt;BR /&gt;at org.apache.spark.executor.Executor.$anonfun$stopHookReference$1(Executor.scala:76)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188)&lt;BR /&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;BR /&gt;at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188)&lt;BR /&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;BR /&gt;at scala.util.Try$.apply(Try.scala:213)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178)&lt;BR /&gt;at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)&lt;BR /&gt;at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)&lt;BR /&gt;at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)&lt;BR /&gt;at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)&lt;BR /&gt;at java.base/java.lang.Thread.run(Thread.java:834)&lt;BR /&gt;Caused by: java.lang.NullPointerException&lt;BR /&gt;at org.apache.spark.shuffle.ShuffleBlockPusher$.&amp;lt;init&amp;gt;(ShuffleBlockPusher.scala:465)&lt;BR /&gt;at org.apache.spark.shuffle.ShuffleBlockPusher$.&amp;lt;clinit&amp;gt;(ShuffleBlockPusher.scala)&lt;BR /&gt;... 16 more&lt;BR /&gt;23/05/03 09:04:35 WARN ShutdownHookManager: ShutdownHook '' failed, java.util.concurrent.ExecutionException: java.lang.ExceptionInInitializerError&lt;BR /&gt;java.util.concurrent.ExecutionException: java.lang.ExceptionInInitializerError&lt;BR /&gt;at java.base/java.util.concurrent.FutureTask.report(FutureTask.java:122)&lt;BR /&gt;at java.base/java.util.concurrent.FutureTask.get(FutureTask.java:205)&lt;BR /&gt;at org.apache.hadoop.util.ShutdownHookManager.executeShutdown(ShutdownHookManager.java:124)&lt;BR /&gt;at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:95)&lt;BR /&gt;Caused by: java.lang.ExceptionInInitializerError&lt;BR /&gt;at org.apache.spark.executor.Executor.stop(Executor.scala:333)&lt;BR /&gt;at org.apache.spark.executor.Executor.$anonfun$stopHookReference$1(Executor.scala:76)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188)&lt;BR /&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;BR /&gt;at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188)&lt;BR /&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;BR /&gt;at scala.util.Try$.apply(Try.scala:213)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178)&lt;BR /&gt;at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)&lt;BR /&gt;at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)&lt;BR /&gt;at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)&lt;BR /&gt;at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)&lt;BR /&gt;at java.base/java.lang.Thread.run(Thread.java:834)&lt;BR /&gt;Caused by: java.lang.NullPointerException&lt;BR /&gt;at org.apache.spark.shuffle.ShuffleBlockPusher$.&amp;lt;init&amp;gt;(ShuffleBlockPusher.scala:465)&lt;BR /&gt;at org.apache.spark.shuffle.ShuffleBlockPusher$.&amp;lt;clinit&amp;gt;(ShuffleBlockPusher.scala)&lt;BR /&gt;... 16 more&lt;/P&gt;&lt;P&gt;C:\Users\abhay&amp;gt;SUCCESS: The process with PID 25604 (child process of PID 29652) has been terminated.&lt;BR /&gt;SUCCESS: The process with PID 29652 (child process of PID 21240) has been terminated.&lt;BR /&gt;SUCCESS: The process with PID 21240 (child process of PID 22864) has been terminated.&lt;/P&gt;</description>
    <pubDate>Wed, 03 May 2023 03:59:24 GMT</pubDate>
    <dc:creator>Abhay_Kumar</dc:creator>
    <dc:date>2023-05-03T03:59:24Z</dc:date>
    <item>
      <title>ERROR SparkContext: Error initializing SparkContext.</title>
      <link>https://community.cloudera.com/t5/Support-Questions/ERROR-SparkContext-Error-initializing-SparkContext/m-p/369869#M240580</link>
      <description>&lt;P&gt;HI all,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I'm getting the following error when trying to launch pyspark. Please help !! Thanks in advance.&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;C:\Users\abhay&amp;gt;pyspark&lt;BR /&gt;Python 3.10.3 (tags/v3.10.3:a342a49, Mar 16 2022, 13:07:40) [MSC v.1929 64 bit (AMD64)] on win32&lt;BR /&gt;Type "help", "copyright", "credits" or "license" for more information.&lt;BR /&gt;WARNING: An illegal reflective access operation has occurred&lt;BR /&gt;WARNING: Illegal reflective access by org.apache.spark.unsafe.Platform (file:/D:/demo/spark-3.2.1/jars/spark-unsafe_2.12-3.2.1.jar) to constructor java.nio.DirectByteBuffer(long,int)&lt;BR /&gt;WARNING: Please consider reporting this to the maintainers of org.apache.spark.unsafe.Platform&lt;BR /&gt;WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operations&lt;BR /&gt;WARNING: All illegal access operations will be denied in a future release&lt;BR /&gt;Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties&lt;BR /&gt;Setting default log level to "WARN".&lt;BR /&gt;To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).&lt;BR /&gt;23/05/03 09:04:32 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable&lt;BR /&gt;23/05/03 09:04:34 ERROR SparkContext: Error initializing SparkContext.&lt;BR /&gt;org.apache.spark.SparkException: Invalid Spark URL: spark://HeartbeatReceiver@Abhay_Mishra:54720&lt;BR /&gt;at org.apache.spark.rpc.RpcEndpointAddress$.apply(RpcEndpointAddress.scala:66)&lt;BR /&gt;at org.apache.spark.rpc.netty.NettyRpcEnv.asyncSetupEndpointRefByURI(NettyRpcEnv.scala:140)&lt;BR /&gt;at org.apache.spark.rpc.RpcEnv.setupEndpointRefByURI(RpcEnv.scala:101)&lt;BR /&gt;at org.apache.spark.rpc.RpcEnv.setupEndpointRef(RpcEnv.scala:109)&lt;BR /&gt;at org.apache.spark.util.RpcUtils$.makeDriverRef(RpcUtils.scala:36)&lt;BR /&gt;at org.apache.spark.executor.Executor.&amp;lt;init&amp;gt;(Executor.scala:218)&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalEndpoint.&amp;lt;init&amp;gt;(LocalSchedulerBackend.scala:64)&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalSchedulerBackend.start(LocalSchedulerBackend.scala:132)&lt;BR /&gt;at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:220)&lt;BR /&gt;at org.apache.spark.SparkContext.&amp;lt;init&amp;gt;(SparkContext.scala:581)&lt;BR /&gt;at org.apache.spark.api.java.JavaSparkContext.&amp;lt;init&amp;gt;(JavaSparkContext.scala:58)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)&lt;BR /&gt;at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)&lt;BR /&gt;at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)&lt;BR /&gt;at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:247)&lt;BR /&gt;at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)&lt;BR /&gt;at py4j.Gateway.invoke(Gateway.java:238)&lt;BR /&gt;at py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)&lt;BR /&gt;at py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)&lt;BR /&gt;at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)&lt;BR /&gt;at py4j.ClientServerConnection.run(ClientServerConnection.java:106)&lt;BR /&gt;at java.base/java.lang.Thread.run(Thread.java:834)&lt;BR /&gt;23/05/03 09:04:35 ERROR Utils: Uncaught exception in thread Thread-2&lt;BR /&gt;java.lang.NullPointerException&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalSchedulerBackend.org$apache$spark$scheduler$local$LocalSchedulerBackend$$stop(LocalSchedulerBackend.scala:173)&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalSchedulerBackend.stop(LocalSchedulerBackend.scala:144)&lt;BR /&gt;at org.apache.spark.scheduler.TaskSchedulerImpl.stop(TaskSchedulerImpl.scala:927)&lt;BR /&gt;at org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:2567)&lt;BR /&gt;at org.apache.spark.SparkContext.$anonfun$stop$12(SparkContext.scala:2086)&lt;BR /&gt;at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1442)&lt;BR /&gt;at org.apache.spark.SparkContext.stop(SparkContext.scala:2086)&lt;BR /&gt;at org.apache.spark.SparkContext.&amp;lt;init&amp;gt;(SparkContext.scala:677)&lt;BR /&gt;at org.apache.spark.api.java.JavaSparkContext.&amp;lt;init&amp;gt;(JavaSparkContext.scala:58)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)&lt;BR /&gt;at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)&lt;BR /&gt;at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)&lt;BR /&gt;at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:247)&lt;BR /&gt;at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)&lt;BR /&gt;at py4j.Gateway.invoke(Gateway.java:238)&lt;BR /&gt;at py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)&lt;BR /&gt;at py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)&lt;BR /&gt;at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)&lt;BR /&gt;at py4j.ClientServerConnection.run(ClientServerConnection.java:106)&lt;BR /&gt;at java.base/java.lang.Thread.run(Thread.java:834)&lt;BR /&gt;23/05/03 09:04:35 WARN MetricsSystem: Stopping a MetricsSystem that is not running&lt;BR /&gt;23/05/03 09:04:35 WARN SparkContext: Another SparkContext is being constructed (or threw an exception in its constructor). This may indicate an error, since only one SparkContext should be running in this JVM (see SPARK-2243). The other SparkContext was created at:&lt;BR /&gt;org.apache.spark.api.java.JavaSparkContext.&amp;lt;init&amp;gt;(JavaSparkContext.scala:58)&lt;BR /&gt;java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)&lt;BR /&gt;java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)&lt;BR /&gt;java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)&lt;BR /&gt;java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)&lt;BR /&gt;py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:247)&lt;BR /&gt;py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)&lt;BR /&gt;py4j.Gateway.invoke(Gateway.java:238)&lt;BR /&gt;py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)&lt;BR /&gt;py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)&lt;BR /&gt;py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)&lt;BR /&gt;py4j.ClientServerConnection.run(ClientServerConnection.java:106)&lt;BR /&gt;java.base/java.lang.Thread.run(Thread.java:834)&lt;BR /&gt;23/05/03 09:04:35 ERROR SparkContext: Error initializing SparkContext.&lt;BR /&gt;org.apache.spark.SparkException: Invalid Spark URL: spark://HeartbeatReceiver@Abhay_Mishra:54743&lt;BR /&gt;at org.apache.spark.rpc.RpcEndpointAddress$.apply(RpcEndpointAddress.scala:66)&lt;BR /&gt;at org.apache.spark.rpc.netty.NettyRpcEnv.asyncSetupEndpointRefByURI(NettyRpcEnv.scala:140)&lt;BR /&gt;at org.apache.spark.rpc.RpcEnv.setupEndpointRefByURI(RpcEnv.scala:101)&lt;BR /&gt;at org.apache.spark.rpc.RpcEnv.setupEndpointRef(RpcEnv.scala:109)&lt;BR /&gt;at org.apache.spark.util.RpcUtils$.makeDriverRef(RpcUtils.scala:36)&lt;BR /&gt;at org.apache.spark.executor.Executor.&amp;lt;init&amp;gt;(Executor.scala:218)&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalEndpoint.&amp;lt;init&amp;gt;(LocalSchedulerBackend.scala:64)&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalSchedulerBackend.start(LocalSchedulerBackend.scala:132)&lt;BR /&gt;at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:220)&lt;BR /&gt;at org.apache.spark.SparkContext.&amp;lt;init&amp;gt;(SparkContext.scala:581)&lt;BR /&gt;at org.apache.spark.api.java.JavaSparkContext.&amp;lt;init&amp;gt;(JavaSparkContext.scala:58)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)&lt;BR /&gt;at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)&lt;BR /&gt;at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)&lt;BR /&gt;at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:247)&lt;BR /&gt;at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)&lt;BR /&gt;at py4j.Gateway.invoke(Gateway.java:238)&lt;BR /&gt;at py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)&lt;BR /&gt;at py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)&lt;BR /&gt;at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)&lt;BR /&gt;at py4j.ClientServerConnection.run(ClientServerConnection.java:106)&lt;BR /&gt;at java.base/java.lang.Thread.run(Thread.java:834)&lt;BR /&gt;23/05/03 09:04:35 ERROR Utils: Uncaught exception in thread Thread-2&lt;BR /&gt;java.lang.NullPointerException&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalSchedulerBackend.org$apache$spark$scheduler$local$LocalSchedulerBackend$$stop(LocalSchedulerBackend.scala:173)&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalSchedulerBackend.stop(LocalSchedulerBackend.scala:144)&lt;BR /&gt;at org.apache.spark.scheduler.TaskSchedulerImpl.stop(TaskSchedulerImpl.scala:927)&lt;BR /&gt;at org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:2567)&lt;BR /&gt;at org.apache.spark.SparkContext.$anonfun$stop$12(SparkContext.scala:2086)&lt;BR /&gt;at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1442)&lt;BR /&gt;at org.apache.spark.SparkContext.stop(SparkContext.scala:2086)&lt;BR /&gt;at org.apache.spark.SparkContext.&amp;lt;init&amp;gt;(SparkContext.scala:677)&lt;BR /&gt;at org.apache.spark.api.java.JavaSparkContext.&amp;lt;init&amp;gt;(JavaSparkContext.scala:58)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)&lt;BR /&gt;at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)&lt;BR /&gt;at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)&lt;BR /&gt;at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:247)&lt;BR /&gt;at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)&lt;BR /&gt;at py4j.Gateway.invoke(Gateway.java:238)&lt;BR /&gt;at py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)&lt;BR /&gt;at py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)&lt;BR /&gt;at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)&lt;BR /&gt;at py4j.ClientServerConnection.run(ClientServerConnection.java:106)&lt;BR /&gt;at java.base/java.lang.Thread.run(Thread.java:834)&lt;BR /&gt;23/05/03 09:04:35 WARN MetricsSystem: Stopping a MetricsSystem that is not running&lt;BR /&gt;D:\demo\spark-3.2.1\python\pyspark\shell.py:42: UserWarning: Failed to initialize Spark session.&lt;BR /&gt;warnings.warn("Failed to initialize Spark session.")&lt;BR /&gt;Traceback (most recent call last):&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\pyspark\shell.py", line 38, in &amp;lt;module&amp;gt;&lt;BR /&gt;spark = SparkSession._create_shell_session() # type: ignore&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\pyspark\sql\session.py", line 553, in _create_shell_session&lt;BR /&gt;return SparkSession.builder.getOrCreate()&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\pyspark\sql\session.py", line 228, in getOrCreate&lt;BR /&gt;sc = SparkContext.getOrCreate(sparkConf)&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\pyspark\context.py", line 392, in getOrCreate&lt;BR /&gt;SparkContext(conf=conf or SparkConf())&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\pyspark\context.py", line 146, in __init__&lt;BR /&gt;self._do_init(master, appName, sparkHome, pyFiles, environment, batchSize, serializer,&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\pyspark\context.py", line 209, in _do_init&lt;BR /&gt;self._jsc = jsc or self._initialize_context(self._conf._jconf)&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\pyspark\context.py", line 329, in _initialize_context&lt;BR /&gt;return self._jvm.JavaSparkContext(jconf)&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\lib\py4j-0.10.9.3-src.zip\py4j\java_gateway.py", line 1585, in __call__&lt;BR /&gt;return_value = get_return_value(&lt;BR /&gt;File "D:\demo\spark-3.2.1\python\lib\py4j-0.10.9.3-src.zip\py4j\protocol.py", line 326, in get_return_value&lt;BR /&gt;raise Py4JJavaError(&lt;BR /&gt;py4j.protocol.Py4JJavaError: An error occurred while calling None.org.apache.spark.api.java.JavaSparkContext.&lt;BR /&gt;: org.apache.spark.SparkException: Invalid Spark URL: spark://HeartbeatReceiver@Abhay_Mishra:54743&lt;BR /&gt;at org.apache.spark.rpc.RpcEndpointAddress$.apply(RpcEndpointAddress.scala:66)&lt;BR /&gt;at org.apache.spark.rpc.netty.NettyRpcEnv.asyncSetupEndpointRefByURI(NettyRpcEnv.scala:140)&lt;BR /&gt;at org.apache.spark.rpc.RpcEnv.setupEndpointRefByURI(RpcEnv.scala:101)&lt;BR /&gt;at org.apache.spark.rpc.RpcEnv.setupEndpointRef(RpcEnv.scala:109)&lt;BR /&gt;at org.apache.spark.util.RpcUtils$.makeDriverRef(RpcUtils.scala:36)&lt;BR /&gt;at org.apache.spark.executor.Executor.&amp;lt;init&amp;gt;(Executor.scala:218)&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalEndpoint.&amp;lt;init&amp;gt;(LocalSchedulerBackend.scala:64)&lt;BR /&gt;at org.apache.spark.scheduler.local.LocalSchedulerBackend.start(LocalSchedulerBackend.scala:132)&lt;BR /&gt;at org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:220)&lt;BR /&gt;at org.apache.spark.SparkContext.&amp;lt;init&amp;gt;(SparkContext.scala:581)&lt;BR /&gt;at org.apache.spark.api.java.JavaSparkContext.&amp;lt;init&amp;gt;(JavaSparkContext.scala:58)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)&lt;BR /&gt;at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)&lt;BR /&gt;at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)&lt;BR /&gt;at java.base/java.lang.reflect.Constructor.newInstance(Constructor.java:490)&lt;BR /&gt;at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:247)&lt;BR /&gt;at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)&lt;BR /&gt;at py4j.Gateway.invoke(Gateway.java:238)&lt;BR /&gt;at py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)&lt;BR /&gt;at py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)&lt;BR /&gt;at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)&lt;BR /&gt;at py4j.ClientServerConnection.run(ClientServerConnection.java:106)&lt;BR /&gt;at java.base/java.lang.Thread.run(Thread.java:834)&lt;/P&gt;&lt;P&gt;23/05/03 09:04:35 ERROR Utils: Uncaught exception in thread shutdown-hook-0&lt;BR /&gt;java.lang.ExceptionInInitializerError&lt;BR /&gt;at org.apache.spark.executor.Executor.stop(Executor.scala:333)&lt;BR /&gt;at org.apache.spark.executor.Executor.$anonfun$stopHookReference$1(Executor.scala:76)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188)&lt;BR /&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;BR /&gt;at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188)&lt;BR /&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;BR /&gt;at scala.util.Try$.apply(Try.scala:213)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178)&lt;BR /&gt;at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)&lt;BR /&gt;at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)&lt;BR /&gt;at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)&lt;BR /&gt;at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)&lt;BR /&gt;at java.base/java.lang.Thread.run(Thread.java:834)&lt;BR /&gt;Caused by: java.lang.NullPointerException&lt;BR /&gt;at org.apache.spark.shuffle.ShuffleBlockPusher$.&amp;lt;init&amp;gt;(ShuffleBlockPusher.scala:465)&lt;BR /&gt;at org.apache.spark.shuffle.ShuffleBlockPusher$.&amp;lt;clinit&amp;gt;(ShuffleBlockPusher.scala)&lt;BR /&gt;... 16 more&lt;BR /&gt;23/05/03 09:04:35 WARN ShutdownHookManager: ShutdownHook '' failed, java.util.concurrent.ExecutionException: java.lang.ExceptionInInitializerError&lt;BR /&gt;java.util.concurrent.ExecutionException: java.lang.ExceptionInInitializerError&lt;BR /&gt;at java.base/java.util.concurrent.FutureTask.report(FutureTask.java:122)&lt;BR /&gt;at java.base/java.util.concurrent.FutureTask.get(FutureTask.java:205)&lt;BR /&gt;at org.apache.hadoop.util.ShutdownHookManager.executeShutdown(ShutdownHookManager.java:124)&lt;BR /&gt;at org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:95)&lt;BR /&gt;Caused by: java.lang.ExceptionInInitializerError&lt;BR /&gt;at org.apache.spark.executor.Executor.stop(Executor.scala:333)&lt;BR /&gt;at org.apache.spark.executor.Executor.$anonfun$stopHookReference$1(Executor.scala:76)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188)&lt;BR /&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;BR /&gt;at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:2019)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188)&lt;BR /&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;BR /&gt;at scala.util.Try$.apply(Try.scala:213)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188)&lt;BR /&gt;at org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178)&lt;BR /&gt;at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)&lt;BR /&gt;at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)&lt;BR /&gt;at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)&lt;BR /&gt;at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)&lt;BR /&gt;at java.base/java.lang.Thread.run(Thread.java:834)&lt;BR /&gt;Caused by: java.lang.NullPointerException&lt;BR /&gt;at org.apache.spark.shuffle.ShuffleBlockPusher$.&amp;lt;init&amp;gt;(ShuffleBlockPusher.scala:465)&lt;BR /&gt;at org.apache.spark.shuffle.ShuffleBlockPusher$.&amp;lt;clinit&amp;gt;(ShuffleBlockPusher.scala)&lt;BR /&gt;... 16 more&lt;/P&gt;&lt;P&gt;C:\Users\abhay&amp;gt;SUCCESS: The process with PID 25604 (child process of PID 29652) has been terminated.&lt;BR /&gt;SUCCESS: The process with PID 29652 (child process of PID 21240) has been terminated.&lt;BR /&gt;SUCCESS: The process with PID 21240 (child process of PID 22864) has been terminated.&lt;/P&gt;</description>
      <pubDate>Wed, 03 May 2023 03:59:24 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/ERROR-SparkContext-Error-initializing-SparkContext/m-p/369869#M240580</guid>
      <dc:creator>Abhay_Kumar</dc:creator>
      <dc:date>2023-05-03T03:59:24Z</dc:date>
    </item>
    <item>
      <title>Re: ERROR SparkContext: Error initializing SparkContext.</title>
      <link>https://community.cloudera.com/t5/Support-Questions/ERROR-SparkContext-Error-initializing-SparkContext/m-p/369977#M240596</link>
      <description>&lt;P&gt;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/104898"&gt;@Abhay_Kumar&lt;/a&gt;&amp;nbsp;Welcome to the Cloudera Community!&lt;BR /&gt;&lt;BR /&gt;To help you get the best possible solution, I have tagged our Spark experts&amp;nbsp;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/28743"&gt;@Gopinath&lt;/a&gt;&amp;nbsp;and&amp;nbsp;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/40384"&gt;@smdas&lt;/a&gt;&amp;nbsp; who may be able to assist you further.&lt;BR /&gt;&lt;BR /&gt;Please keep us updated on your post, and we hope you find a satisfactory solution to your query.&lt;/P&gt;</description>
      <pubDate>Wed, 03 May 2023 17:39:54 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/ERROR-SparkContext-Error-initializing-SparkContext/m-p/369977#M240596</guid>
      <dc:creator>DianaTorres</dc:creator>
      <dc:date>2023-05-03T17:39:54Z</dc:date>
    </item>
  </channel>
</rss>

