<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Akka Error while running Spark Jobs in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18620#M2885</link>
    <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The worker logs show the following connection erros - Any idea how to resolve?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;PRE&gt;AssociationError [akka.tcp://sparkWorker@host1:7078] -&amp;gt; [akka.tcp://sparkExecutor@worker1:33912]: 
Error [Association failed with [akka.tcp://sparkExecutor@worker1:33912]] 
[akka.remote.EndpointAssociationException: Association failed with [akka.tcp://sparkExecutor@worker1:33912]
Caused by: akka.remote.transport.netty.NettyTransport$$anonfun$associate$1$$anon$2: Connection refused: worker1/10.11.11.11:33912]&lt;/PRE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Thu, 11 Sep 2014 15:22:56 GMT</pubDate>
    <dc:creator>ArunShell</dc:creator>
    <dc:date>2014-09-11T15:22:56Z</dc:date>
    <item>
      <title>Akka Error while running Spark Jobs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18602#M2883</link>
      <description>&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am submitting a Spark Streaming job using spark-submit.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;"spark-submit &amp;nbsp;--class "test.Main" --master yarn-client testjob.jar"&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;But I am facing the below errors. Please assist to resolve.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;PRE&gt;14/09/11 05:55:45 ERROR YarnClientClusterScheduler: Lost executor 4 on host1.com: remote Akka client disassociated



14/09/11 05:56:06 ERROR JobScheduler: Error running job streaming job 1410429330000 ms.0
org.apache.spark.SparkException: Job aborted due to stage failure: Task 1.0:0 failed 4 times, most recent failure: TID 3 on host host2.com failed for unknown reason
Driver stacktrace:
        at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1033)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1017)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1015)
        at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
        at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
        at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1015)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:633)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:633)
        at scala.Option.foreach(Option.scala:236)
        at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:633)
        at org.apache.spark.scheduler.DAGSchedulerEventProcessActor$$anonfun$receive$2.applyOrElse(DAGScheduler.scala:1207)
        at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498)
        at akka.actor.ActorCell.invoke(ActorCell.scala:456)
        at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237)
        at akka.dispatch.Mailbox.run(Mailbox.scala:219)
        at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386)
        at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
        at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
        at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
        at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)




Exception in thread "main" org.apache.spark.SparkException: Job aborted due to stage failure: Task 1.0:0 failed 4 times, most recent failure: TID 3 on host host3.com failed for unknown reason
Driver stacktrace:
        at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1033)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1017)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1015)
        at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
        at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
        at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1015)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:633)
        at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:633)
        at scala.Option.foreach(Option.scala:236)
        at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:633)
        at org.apache.spark.scheduler.DAGSchedulerEventProcessActor$$anonfun$receive$2.applyOrElse(DAGScheduler.scala:1207)
        at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498)
        at akka.actor.ActorCell.invoke(ActorCell.scala:456)
        at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237)
        at akka.dispatch.Mailbox.run(Mailbox.scala:219)
        at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386)
        at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
        at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
        at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
        at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)&lt;/PRE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 09:07:21 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18602#M2883</guid>
      <dc:creator>ArunShell</dc:creator>
      <dc:date>2022-09-16T09:07:21Z</dc:date>
    </item>
    <item>
      <title>Re: Akka Error while running Spark Jobs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18604#M2884</link>
      <description>&lt;P&gt;It's nothing to do with Akka per se. This says your jobs are failing. You would have to look at the logs on the workers to understand why.&lt;/P&gt;</description>
      <pubDate>Thu, 11 Sep 2014 10:13:53 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18604#M2884</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2014-09-11T10:13:53Z</dc:date>
    </item>
    <item>
      <title>Re: Akka Error while running Spark Jobs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18620#M2885</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The worker logs show the following connection erros - Any idea how to resolve?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;PRE&gt;AssociationError [akka.tcp://sparkWorker@host1:7078] -&amp;gt; [akka.tcp://sparkExecutor@worker1:33912]: 
Error [Association failed with [akka.tcp://sparkExecutor@worker1:33912]] 
[akka.remote.EndpointAssociationException: Association failed with [akka.tcp://sparkExecutor@worker1:33912]
Caused by: akka.remote.transport.netty.NettyTransport$$anonfun$associate$1$$anon$2: Connection refused: worker1/10.11.11.11:33912]&lt;/PRE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 11 Sep 2014 15:22:56 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18620#M2885</guid>
      <dc:creator>ArunShell</dc:creator>
      <dc:date>2014-09-11T15:22:56Z</dc:date>
    </item>
    <item>
      <title>Re: Akka Error while running Spark Jobs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18624#M2886</link>
      <description>&lt;P&gt;On its face it means what it says --&amp;nbsp;the master is unable to talk to the worker. I would check your firewall rules and make sure these machines can talk to each other, and on these ports. Spark picks ephemeral ports so you may have to open ranges.&lt;/P&gt;</description>
      <pubDate>Thu, 11 Sep 2014 15:54:34 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18624#M2886</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2014-09-11T15:54:34Z</dc:date>
    </item>
    <item>
      <title>Re: Akka Error while running Spark Jobs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18676#M2887</link>
      <description>&lt;P&gt;Thanks.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Please clarify the below -&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;What is the port range that I need to ask the admin team to open on each worker node?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;And what are these ports used for, Spark Workers already use the port 7078 right? Are these random ports opened for each spark job ?&lt;/P&gt;</description>
      <pubDate>Fri, 12 Sep 2014 06:38:04 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18676#M2887</guid>
      <dc:creator>ArunShell</dc:creator>
      <dc:date>2014-09-12T06:38:04Z</dc:date>
    </item>
    <item>
      <title>Re: Akka Error while running Spark Jobs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18680#M2888</link>
      <description>&lt;P&gt;Have a look at:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;A href="https://spark.apache.org/docs/latest/configuration.html#networking" target="_blank"&gt;https://spark.apache.org/docs/latest/configuration.html#networking&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I think you are interested in fixing the driver and executor ports to a fixed value, rather than let them be chosen randomly.&lt;/P&gt;&lt;P&gt;Same with the UI ports, if you're interested in those.&lt;/P&gt;</description>
      <pubDate>Fri, 12 Sep 2014 07:45:28 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18680#M2888</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2014-09-12T07:45:28Z</dc:date>
    </item>
    <item>
      <title>Re: Akka Error while running Spark Jobs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18682#M2889</link>
      <description>&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;This makes sense - thanks!&lt;/P&gt;</description>
      <pubDate>Fri, 12 Sep 2014 07:54:14 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18682#M2889</guid>
      <dc:creator>ArunShell</dc:creator>
      <dc:date>2014-09-12T07:54:14Z</dc:date>
    </item>
    <item>
      <title>Re: Akka Error while running Spark Jobs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18700#M2890</link>
      <description>&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Hi - Does it make a difference if I use a "--master yarn-client" &amp;nbsp;or &amp;nbsp;" --master &amp;nbsp;yarn-cluster" &amp;nbsp;for &amp;nbsp;this error in "spark-submit" since yarn-client uses a local driver?&lt;/P&gt;</description>
      <pubDate>Fri, 12 Sep 2014 13:35:17 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18700#M2890</guid>
      <dc:creator>ArunShell</dc:creator>
      <dc:date>2014-09-12T13:35:17Z</dc:date>
    </item>
    <item>
      <title>Re: Akka Error while running Spark Jobs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18702#M2891</link>
      <description>&lt;P&gt;It will make a difference insofar as the driver program will run either out on the cluster (yarn-cluster) or locally (yarn-client). The same issue remains -- the processes need to talk to each other on certain ports. But it affects where the driver is and that affects what machine's ports need to be open. For example, if your ports are all open within your cluster, I expect that yarn-cluster works directly.&lt;/P&gt;</description>
      <pubDate>Fri, 12 Sep 2014 13:49:01 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Akka-Error-while-running-Spark-Jobs/m-p/18702#M2891</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2014-09-12T13:49:01Z</dc:date>
    </item>
  </channel>
</rss>

