Created 02-16-2016 04:04 PM
I have centos 7.1.
On my multinode Hadoop cluster (2.3.4) I have , through Ambari, installed spark 1.5.2. I am trying to connect to sparkR from CLI and after I run sparkR I get the following error:
Error in value[[3L]](cond) : Failed to connect JVM In addition: Warning message: In socketConnection(host = hostname, port = port, server = FALSE, : localhost:9001 cannot be opened
The port (9001) is opened on the namenode (where Im running sparkR) Do you have any ideas what Im doing wrong? Ive seen this link: http://hortonworks.com/hadoop-tutorial/apache-spark-1-5-1-technical-preview-with-hdp-2-3/
and I followed also this link:
http://www.jason-french.com/blog/2013/03/11/installing-r-in-linux/
To install R on all datanodes. I appreicate your contribution.
Created 02-19-2016 06:28 PM
@Neeraj Sabharwal, @Artem Ervits
I made it work now!
Ive used Ubuntu 14.04 Trusty, installed manually Spark 1.4.1 and set up sparkR. Now, I dont know if the problem was in centos 7.2, but the installment of R was different than what Ive done earlier and from what it says here:
http://www.jason-french.com/blog/2013/03/11/installing-r-in-linux/
If you guys want, I can try the same on centos 7.2 and report. If you want, I can describe the process of preparing the environment for using sparkR. I will also try on other spark versions. We depend on R because of the research.
Let me know if there is interest.
Created 02-17-2016 12:01 PM
@Artem ErvitsI tried that as well.
Here the status command:
sudo service iptables status
And the result:
Redirecting to /bin/systemctl status iptables.service ● iptables.service Loaded: not-found (Reason: No such file or directory) Active: inactive (dead)
I tried running the iptables and firewalld commands on the datanodes and they all seem inactive (same results as above).
Centos 7.2.
Created 02-17-2016 12:38 PM
@marko we don't officially support 7.2 maybe this is one of the reasons? Try to update your host. If you have access to 7.0 or 7.1 machine?
Created 02-17-2016 12:58 PM
@Artem Ervits, @Neeraj Sabharwal Well Here is my version: CentOS Linux release 7.2.1511 (Core) The cloud services I use have, in their menu, CentOS 7.1, I think they upgraded the systems, left 7.1 on the menu but the instances are 7.2
Now im a bit upset. Do you think this could really be it? That 7.2 version is the problem?
Created 02-17-2016 12:35 PM
@marko I see that other people are hitting the same issue. I guess you did the manual install of spark now..Correct?
Can you try to do this in some other node?
Created 02-17-2016 12:38 PM
@Neeraj Sabharwal No, the spark installation was done through Ambari. I installed, accepted all the defaults and ran the test examples from the TP website.Funny thing, when installing, you get to choose 1.4.2 version of spark, while 1.5.2 is installed.
Created 02-17-2016 12:39 PM
@marko Nice!!!!!! You are way ahead 🙂
Can you try running your job from different node?
Created 02-17-2016 12:47 PM
Nothing nice about this issue, its a show stopper if you work for researchers. 🙂
What do you mean by different node? And run it how?
Created 02-17-2016 01:09 PM
Created 02-17-2016 01:20 PM
Ive manually installed 1.6.0 on the same node where Im running spark 1.5.2. Now I ll try doing it on one datanode - I ll install 1.5.2 manually. By the way, Neeraj, have you seen the CentOS 7.2 dialog Im having with Artem. Let me know if you have a take on that one.
Created 02-17-2016 01:36 PM
I have now installed 1.5.2 on one of the datanodes, I can run spark-shell, when I run sparkR I get this issue:
Launching java with spark-submit command /usr/hdp/2.3.4.0-3485/spark/bin/spark-submit "sparkr-shell" /tmp/Rtmp6iFIT6/backend_port15af7623f5dd
16/02/17 14:35:18 WARN internal.ThreadLocalRandom: Failed to generate a seed from SecureRandom within 3 seconds. Not enough entrophy?
Error in socketConnection(port = monitorPort) : cannot open the connection
In addition: Warning message: In socketConnection(port = monitorPort) : localhost:44867 cannot be opened
Pretty much the same problem.