Archives of Support Questions (Read Only)

This is an archived board for historical reference. Information and links may no longer be available or relevant
Announcements
This board is archived and read-only for historical reference. To ask a new question, please post a new topic on the appropriate active board.

[CDH 5.5 VirtualBox] unable to connect to Spark Master/Worker

avatar
Frequent Visitor

Hi *,

im using the newest CDH 5.5 VirtualBox Image, so far i only installed scala, sbt, netbeans and dropbox.

 

I encounter a problem running a spark application (see error below). In the Cloudera Manager every service, including spark is on green. But however if I click on the Spark bookmarks I am unable to connect to Spark Master/Worker, the History Server works and also every other bookmark.

 

The spark application i run was build in some previous CDH 5.x release and I only changed the master to local (I have no access to the cluster at my university yet) I would be happy about some help.

 

Best Regards,

Tobias

 

15/11/27 09:30:43 INFO SparkContext: Running Spark version 1.5.0-cdh5.5.0
15/11/27 09:30:43 WARN Utils: Your hostname, quickstart.cloudera resolves to a loopback address: 127.0.0.1; using 10.0.2.15 instead (on interface eth1)
15/11/27 09:30:43 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to another address
15/11/27 09:30:44 INFO SecurityManager: Changing view acls to: cloudera
15/11/27 09:30:44 INFO SecurityManager: Changing modify acls to: cloudera
15/11/27 09:30:44 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(cloudera); users with modify permissions: Set(cloudera)
15/11/27 09:30:45 INFO Slf4jLogger: Slf4jLogger started
15/11/27 09:30:45 INFO Remoting: Starting remoting
15/11/27 09:30:45 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkDriver@10.0.2.15:51812]
15/11/27 09:30:45 INFO Remoting: Remoting now listens on addresses: [akka.tcp://sparkDriver@10.0.2.15:51812]
15/11/27 09:30:45 INFO Utils: Successfully started service 'sparkDriver' on port 51812.
15/11/27 09:30:45 INFO SparkEnv: Registering MapOutputTracker
15/11/27 09:30:45 INFO SparkEnv: Registering BlockManagerMaster
15/11/27 09:30:45 INFO DiskBlockManager: Created local directory at /tmp/blockmgr-e2a8a00b-9bac-4564-a75d-cf2bd864e0be
15/11/27 09:30:45 INFO MemoryStore: MemoryStore started with capacity 883.6 MB
15/11/27 09:30:45 INFO HttpFileServer: HTTP File server directory is /tmp/spark-3e2ba8c0-103d-416a-9963-427825993dfc/httpd-66ccf5bb-e5e7-46ad-9541-5f6438a753cc
15/11/27 09:30:45 INFO HttpServer: Starting HTTP Server
15/11/27 09:30:45 INFO Server: jetty-8.y.z-SNAPSHOT
15/11/27 09:30:46 INFO AbstractConnector: Started SocketConnector@0.0.0.0:36447
15/11/27 09:30:46 INFO Utils: Successfully started service 'HTTP file server' on port 36447.
15/11/27 09:30:46 INFO SparkEnv: Registering OutputCommitCoordinator
15/11/27 09:30:46 INFO Server: jetty-8.y.z-SNAPSHOT
15/11/27 09:30:46 INFO AbstractConnector: Started SelectChannelConnector@0.0.0.0:4040
15/11/27 09:30:46 INFO Utils: Successfully started service 'SparkUI' on port 4040.
15/11/27 09:30:46 INFO SparkUI: Started SparkUI at http://10.0.2.15:4040
15/11/27 09:30:46 INFO SparkContext: Added JAR file:/home/cloudera/workspace/./S2RDF_DataSetCreator/target/scala-2.10/datasetcreator_2.10-1.0.jar at http://10.0.2.15:36447/jars/datasetcreator_2.10-1.0.jar with timestamp 1448613046392
15/11/27 09:30:46 WARN MetricsSystem: Using default name DAGScheduler for source because spark.app.id is not set.
15/11/27 09:30:46 INFO Executor: Starting executor ID driver on host localhost
15/11/27 09:30:46 INFO Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 34264.
15/11/27 09:30:46 INFO NettyBlockTransferService: Server created on 34264
15/11/27 09:30:46 INFO BlockManager: external shuffle service port = 7337
15/11/27 09:30:46 INFO BlockManagerMaster: Trying to register BlockManager
15/11/27 09:30:46 INFO BlockManagerMasterEndpoint: Registering block manager localhost:34264 with 883.6 MB RAM, BlockManagerId(driver, localhost, 34264)
15/11/27 09:30:46 INFO BlockManagerMaster: Registered BlockManager
15/11/27 09:30:47 ERROR SparkContext: Error initializing SparkContext.
org.apache.hadoop.security.AccessControlException: Permission denied: user=cloudera, access=WRITE, inode="/user/spark/applicationHistory":spark:supergroup:drwxr-xr-x
	at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.checkFsPermission(DefaultAuthorizationProvider.java:257)
	at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.check(DefaultAuthorizationProvider.java:238)
	at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.check(DefaultAuthorizationProvider.java:216)
	at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.checkPermission(DefaultAuthorizationProvider.java:145)
	at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:138)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6609)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6591)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:6543)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInternal(FSNamesystem.java:2756)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInt(FSNamesystem.java:2674)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFile(FSNamesystem.java:2559)
	at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.create(NameNodeRpcServer.java:592)
	at org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.create(AuthorizationProviderProxyClientProtocol.java:110)
	at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.create(ClientNamenodeProtocolServerSideTranslatorPB.java:395)
	at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
	at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617)
	at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1060)
	at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086)
	at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:415)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671)
	at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080)

	at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
	at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
	at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
	at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
	at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
	at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:73)
	at org.apache.hadoop.hdfs.DFSOutputStream.newStreamForCreate(DFSOutputStream.java:1872)
	at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1737)
	at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1662)
	at org.apache.hadoop.hdfs.DistributedFileSystem$6.doCall(DistributedFileSystem.java:404)
	at org.apache.hadoop.hdfs.DistributedFileSystem$6.doCall(DistributedFileSystem.java:400)
	at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
	at org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:400)
	at org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:343)
	at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:917)
	at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:898)
	at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:795)
	at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:784)
	at org.apache.spark.scheduler.EventLoggingListener.start(EventLoggingListener.scala:121)
	at org.apache.spark.SparkContext.<init>(SparkContext.scala:541)
	at dataCreator.Settings$.loadSparkContext(Settings.scala:69)
	at dataCreator.Settings$.<init>(Settings.scala:17)
	at dataCreator.Settings$.<clinit>(Settings.scala)
	at runDriver$.main(runDriver.scala:12)
	at runDriver.main(runDriver.scala)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:672)
	at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180)
	at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205)
	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:120)
	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): Permission denied: user=cloudera, access=WRITE, inode="/user/spark/applicationHistory":spark:supergroup:drwxr-xr-x
	at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.checkFsPermission(DefaultAuthorizationProvider.java:257)
	at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.check(DefaultAuthorizationProvider.java:238)
	at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.check(DefaultAuthorizationProvider.java:216)
	at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.checkPermission(DefaultAuthorizationProvider.java:145)
	at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:138)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6609)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6591)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:6543)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInternal(FSNamesystem.java:2756)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInt(FSNamesystem.java:2674)
	at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFile(FSNamesystem.java:2559)
	at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.create(NameNodeRpcServer.java:592)
	at org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.create(AuthorizationProviderProxyClientProtocol.java:110)
	at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.create(ClientNamenodeProtocolServerSideTranslatorPB.java:395)
	at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
	at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617)
	at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1060)
	at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086)
	at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:415)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671)
	at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080)

	at org.apache.hadoop.ipc.Client.call(Client.java:1472)
	at org.apache.hadoop.ipc.Client.call(Client.java:1403)
	at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:230)
	at com.sun.proxy.$Proxy18.create(Unknown Source)
	at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.create(ClientNamenodeProtocolTranslatorPB.java:295)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:252)
	at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:104)
	at com.sun.proxy.$Proxy19.create(Unknown Source)
	at org.apache.hadoop.hdfs.DFSOutputStream.newStreamForCreate(DFSOutputStream.java:1867)
	... 27 more
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/metrics/json,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/stage/kill,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/api,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/static,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/executors/threadDump/json,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/executors/threadDump,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/executors/json,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/executors,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/environment/json,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/environment,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/storage/rdd/json,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/storage/rdd,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/storage/json,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/storage,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/pool/json,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/pool,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/stage/json,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/stage,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/json,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/jobs/job/json,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/jobs/job,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/jobs/json,null}
15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/jobs,null}
15/11/27 09:30:48 INFO SparkUI: Stopped Spark web UI at http://10.0.2.15:4040
15/11/27 09:30:48 INFO DAGScheduler: Stopping DAGScheduler
15/11/27 09:30:48 WARN QueuedThreadPool: 2 threads could not be stopped
15/11/27 09:30:48 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
15/11/27 09:30:48 INFO MemoryStore: MemoryStore cleared
15/11/27 09:30:48 INFO BlockManager: BlockManager stopped
15/11/27 09:30:48 INFO BlockManagerMaster: BlockManagerMaster stopped
15/11/27 09:30:48 INFO SparkContext: Successfully stopped SparkContext

 

2 ACCEPTED SOLUTIONS

avatar
Guru
The bookmark needs to be removed, I'm afraid: as of CDH 5.5, only
Spark-on-YARN is supported, so there is no Spark Master to run. When
running the spark shell, for instance, one specify 'yarn-client' as the
master.

View solution in original post

avatar
Visitor

You will have problems starting spark-shell with any master because there are permissioning issues with the hdfs folders that spark uses.

 

To fix it do the following:

 

sudo -u hdfs hadoop fs -chmod 777 /user/spark
sudo -u spark hadoop fs -chmod 777 /user/spark/applicationHistory

 

That should do it 🙂

View solution in original post

3 REPLIES 3

avatar
Guru
The bookmark needs to be removed, I'm afraid: as of CDH 5.5, only
Spark-on-YARN is supported, so there is no Spark Master to run. When
running the spark shell, for instance, one specify 'yarn-client' as the
master.

avatar
Visitor

You will have problems starting spark-shell with any master because there are permissioning issues with the hdfs folders that spark uses.

 

To fix it do the following:

 

sudo -u hdfs hadoop fs -chmod 777 /user/spark
sudo -u spark hadoop fs -chmod 777 /user/spark/applicationHistory

 

That should do it 🙂

avatar
Contributor

For me it was solved after login using hdfs user. You can typically then run the spark-shell command without errors