Support Questions

Find answers, ask questions, and share your expertise
Announcements
Celebrating as our community reaches 100,000 members! Thank you!

sparkR - Error in socketConnection(port = monitorPort)

avatar
Expert Contributor

I have centos 7.1.

On my multinode Hadoop cluster (2.3.4) I have , through Ambari, installed spark 1.5.2. I am trying to connect to sparkR from CLI and after I run sparkR I get the following error:

Error in value[[3L]](cond) : Failed to connect JVM In addition: Warning message: In socketConnection(host = hostname, port = port, server = FALSE, : localhost:9001 cannot be opened

The port (9001) is opened on the namenode (where Im running sparkR) Do you have any ideas what Im doing wrong? Ive seen this link: http://hortonworks.com/hadoop-tutorial/apache-spark-1-5-1-technical-preview-with-hdp-2-3/

and I followed also this link:

http://www.jason-french.com/blog/2013/03/11/installing-r-in-linux/

To install R on all datanodes. I appreicate your contribution.

1 ACCEPTED SOLUTION

avatar
Expert Contributor
hide-solution

This problem has been solved!

Want to get a detailed solution you have to login/registered on the community

Register/Login
33 REPLIES 33

avatar
Expert Contributor

@Neeraj Sabharwal, @Artem Ervits

Im going to give it a try on Ubuntu 14.04 today.

Neeraj, you mentioned some other people are having the same problems. Can you give more information. How did they solve it?

avatar
Expert Contributor
hide-solution

This problem has been solved!

Want to get a detailed solution you have to login/registered on the community

Register/Login

avatar
Master Mentor

@marko

I am interested to know more on this

avatar
Expert Contributor

@Neeraj Sabharwal

I needed some more time to test it on spark 1.4.1 and 1.5.2. I also put it in a nice form. Here is the process Ive taken in order to get sparkR running:

https://markobigdata.wordpress.com/2016/02/22/installing-r-on-hadoop-cluster-to-run-sparkr/