<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Spark Streaming Fails on Cluster mode ( Flume as source) in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Streaming-Fails-on-Cluster-mode-Flume-as-source/m-p/25582#M5307</link>
    <description>&lt;P&gt;but if there's a network problem , what might it be ? i have 3 nodes 1 master and 2 workers , i can submit any job using master and it appears working on the 2 workers ,, but the network word count isn't working on cluster mode , so lets assume its a network problem , what might it be ?&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Sun, 15 Mar 2015 10:59:11 GMT</pubDate>
    <dc:creator>tarekabouzeid91</dc:creator>
    <dc:date>2015-03-15T10:59:11Z</dc:date>
    <item>
      <title>Spark Streaming Fails on Cluster mode ( Flume as source)</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Streaming-Fails-on-Cluster-mode-Flume-as-source/m-p/25577#M5304</link>
      <description>&lt;P&gt;i am using spark streaming , event count example , flume as source of avro events , everything works fine when executing spark on local mode , but when i try to run the example on my cluster i got failed to bind error ,&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Command line thats working " local mode "&amp;nbsp;&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;PRE&gt;spark-submit --class "WordCount" --master local[*] --jars /opt/cloudera/parcels/CDH-5.3.1-1.cdh5.3.1.p0.5/lib/spark/lib/spark-streaming-flume_2.10-1.2.0.jar,/opt/cloudera/parcels/CDH-5.3.1-1.cdh5.3.1.p0.5/lib/flume-ng/lib/avro-ipc.jar,/opt/cloudera/parcels/CDH-5.3.1-1.cdh5.3.1.p0.5/lib/flume-ng/lib/flume-ng-sdk-1.5.0-cdh5.3.1.jar /usr/local/WordCount/target/scala-2.10/wordcount_2.10-1.0.jar node01 6789&lt;/PRE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Command line that's not working " cluster mode "&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;PRE&gt;spark-submit --class "WordCount" --master spark://node01:7077 --jars /opt/cloudera/parcels/CDH-5.3.1-1.cdh5.3.1.p0.5/lib/spark/lib/spark-streaming-flume_2.10-1.2.0.jar,/opt/cloudera/parcels/CDH-5.3.1-1.cdh5.3.1.p0.5/lib/flume-ng/lib/avro-ipc.jar,/opt/cloudera/parcels/CDH-5.3.1-1.cdh5.3.1.p0.5/lib/flume-ng/lib/flume-ng-sdk-1.5.0-cdh5.3.1.jar /usr/local/WordCount/target/scala-2.10/wordcount_2.10-1.0.jar node01 6789&lt;/PRE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Error :&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;PRE&gt;ERROR ReceiverTracker: Deregistered receiver for stream 0: Error starting receiver 0 - org.jboss.netty.channel.ChannelException: Failed to bind to: /192.168.168.94:6789
        at org.jboss.netty.bootstrap.ServerBootstrap.bind(ServerBootstrap.java:272)
        at org.apache.avro.ipc.NettyServer.&amp;lt;init&amp;gt;(NettyServer.java:106)
        at org.apache.avro.ipc.NettyServer.&amp;lt;init&amp;gt;(NettyServer.java:119)
        at org.apache.avro.ipc.NettyServer.&amp;lt;init&amp;gt;(NettyServer.java:74)
        at org.apache.avro.ipc.NettyServer.&amp;lt;init&amp;gt;(NettyServer.java:68)
        at org.apache.spark.streaming.flume.FlumeReceiver.initServer(FlumeInputDStream.scala:164)
        at org.apache.spark.streaming.flume.FlumeReceiver.onStart(FlumeInputDStream.scala:171)
        at org.apache.spark.streaming.receiver.ReceiverSupervisor.startReceiver(ReceiverSupervisor.scala:121)
        at org.apache.spark.streaming.receiver.ReceiverSupervisor.start(ReceiverSupervisor.scala:106)
        at org.apache.spark.streaming.scheduler.ReceiverTracker$ReceiverLauncher$$anonfun$8.apply(ReceiverTracker.scala:277)
        at org.apache.spark.streaming.scheduler.ReceiverTracker$ReceiverLauncher$$anonfun$8.apply(ReceiverTracker.scala:269)
        at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:1314)
        at org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:1314)
        at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:61)
        at org.apache.spark.scheduler.Task.run(Task.scala:56)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:196)
        at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
        at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
        at java.lang.Thread.run(Thread.java:744)
Caused by: java.net.BindException: Cannot assign requested address
        at sun.nio.ch.Net.bind0(Native Method)
        at sun.nio.ch.Net.bind(Net.java:444)
        at sun.nio.ch.Net.bind(Net.java:436)
        at sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:214)
        at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
        at org.jboss.netty.channel.socket.nio.NioServerBoss$RegisterTask.run(NioServerBoss.java:193)
        at org.jboss.netty.channel.socket.nio.AbstractNioSelector.processTaskQueue(AbstractNioSelector.java:366)
        at org.jboss.netty.channel.socket.nio.AbstractNioSelector.run(AbstractNioSelector.java:290)
        at org.jboss.netty.channel.socket.nio.NioServerBoss.run(NioServerBoss.java:42)
        at org.jboss.netty.util.ThreadRenamingRunnable.run(ThreadRenamingRunnable.java:108)
        at org.jboss.netty.util.internal.DeadLockProofWorker$1.run(DeadLockProofWorker.java:42)
        ... 3 more&lt;/PRE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 09:24:09 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Streaming-Fails-on-Cluster-mode-Flume-as-source/m-p/25577#M5304</guid>
      <dc:creator>tarekabouzeid91</dc:creator>
      <dc:date>2022-09-16T09:24:09Z</dc:date>
    </item>
    <item>
      <title>Re: Spark Streaming Fails on Cluster mode ( Flume as source)</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Streaming-Fails-on-Cluster-mode-Flume-as-source/m-p/25580#M5305</link>
      <description>&lt;P&gt;As you can see, the problem is that the receiver&amp;nbsp;can't bind to its assigned address. Is there any&amp;nbsp;networking-related restriction in place that would prevent this? is this the port you intended?&lt;/P&gt;</description>
      <pubDate>Sun, 15 Mar 2015 10:29:49 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Streaming-Fails-on-Cluster-mode-Flume-as-source/m-p/25580#M5305</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2015-03-15T10:29:49Z</dc:date>
    </item>
    <item>
      <title>Re: Spark Streaming Fails on Cluster mode ( Flume as source)</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Streaming-Fails-on-Cluster-mode-Flume-as-source/m-p/25581#M5306</link>
      <description>&lt;P&gt;i guess there's no network restrictions , all 3 nodes that i am using are configured to work together normally , but i read&amp;nbsp;that when spark deploy the code on a worker it starts to listen to the port , while another worker is already using listening on this port so it cause the faliure , check this&amp;nbsp;&lt;A target="_blank" href="http://apache-spark-user-list.1001560.n3.nabble.com/How-to-use-FlumeInputDStream-in-spark-cluster-td1604.html"&gt;http://apache-spark-user-list.1001560.n3.nabble.com/How-to-use-FlumeInputDStream-in-spark-cluster-td1604.html&lt;/A&gt; &amp;nbsp;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Sun, 15 Mar 2015 10:53:27 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Streaming-Fails-on-Cluster-mode-Flume-as-source/m-p/25581#M5306</guid>
      <dc:creator>tarekabouzeid91</dc:creator>
      <dc:date>2015-03-15T10:53:27Z</dc:date>
    </item>
    <item>
      <title>Re: Spark Streaming Fails on Cluster mode ( Flume as source)</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Streaming-Fails-on-Cluster-mode-Flume-as-source/m-p/25582#M5307</link>
      <description>&lt;P&gt;but if there's a network problem , what might it be ? i have 3 nodes 1 master and 2 workers , i can submit any job using master and it appears working on the 2 workers ,, but the network word count isn't working on cluster mode , so lets assume its a network problem , what might it be ?&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Sun, 15 Mar 2015 10:59:11 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Streaming-Fails-on-Cluster-mode-Flume-as-source/m-p/25582#M5307</guid>
      <dc:creator>tarekabouzeid91</dc:creator>
      <dc:date>2015-03-15T10:59:11Z</dc:date>
    </item>
    <item>
      <title>Re: Spark Streaming Fails on Cluster mode ( Flume as source)</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Streaming-Fails-on-Cluster-mode-Flume-as-source/m-p/25586#M5308</link>
      <description>&lt;P&gt;Although that thread sounds similar, I don't think it's the same thing. Failing to bind is not a failure to connect to a remote host. It means the local host didn't allow the process to listen on a port. The two most likely explanations are:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;- an old process is still listening on that port, or at least, another still-running process is&lt;/P&gt;&lt;P&gt;- you appear to be binding to a non-routable address (192.168.x.x) This might be OK but&amp;nbsp;worth double-checking&lt;/P&gt;</description>
      <pubDate>Sun, 15 Mar 2015 15:43:23 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Streaming-Fails-on-Cluster-mode-Flume-as-source/m-p/25586#M5308</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2015-03-15T15:43:23Z</dc:date>
    </item>
    <item>
      <title>Re: Spark Streaming Fails on Cluster mode ( Flume as source)</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Streaming-Fails-on-Cluster-mode-Flume-as-source/m-p/25587#M5309</link>
      <description>&lt;P&gt;i fixed that problem by making the spark listen to node02 and flume send events to node02 , that fixed the problem acutally , thanks so much for your help&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Sun, 15 Mar 2015 16:37:29 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Streaming-Fails-on-Cluster-mode-Flume-as-source/m-p/25587#M5309</guid>
      <dc:creator>tarekabouzeid91</dc:creator>
      <dc:date>2015-03-15T16:37:29Z</dc:date>
    </item>
    <item>
      <title>Re: Spark Streaming Fails on Cluster mode ( Flume as source)</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Streaming-Fails-on-Cluster-mode-Flume-as-source/m-p/34131#M5310</link>
      <description>&lt;P&gt;Hi sowen&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;We actually have the same problem of filed to bind with the internal ip-port 10.1.0.11:50321.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The question is that sometimes connect Spark Streaming with the port and sometimes didn´t it and when the did not connect sometimes in 5 -7 min that Spark Streaming is trying connect.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Do you know what is the possible cause?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;We install Spark 1.5.1 on clouderas 5.3.7 (with YARN).&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Best regards&lt;/P&gt;</description>
      <pubDate>Tue, 17 Nov 2015 09:38:08 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Streaming-Fails-on-Cluster-mode-Flume-as-source/m-p/34131#M5310</guid>
      <dc:creator>CDHJOSE</dc:creator>
      <dc:date>2015-11-17T09:38:08Z</dc:date>
    </item>
  </channel>
</rss>

