Created on 11-27-2015 12:49 AM - edited 09-16-2022 02:50 AM
Hi *,
im using the newest CDH 5.5 VirtualBox Image, so far i only installed scala, sbt, netbeans and dropbox.
I encounter a problem running a spark application (see error below). In the Cloudera Manager every service, including spark is on green. But however if I click on the Spark bookmarks I am unable to connect to Spark Master/Worker, the History Server works and also every other bookmark.
The spark application i run was build in some previous CDH 5.x release and I only changed the master to local (I have no access to the cluster at my university yet) I would be happy about some help.
Best Regards,
Tobias
15/11/27 09:30:43 INFO SparkContext: Running Spark version 1.5.0-cdh5.5.0 15/11/27 09:30:43 WARN Utils: Your hostname, quickstart.cloudera resolves to a loopback address: 127.0.0.1; using 10.0.2.15 instead (on interface eth1) 15/11/27 09:30:43 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to another address 15/11/27 09:30:44 INFO SecurityManager: Changing view acls to: cloudera 15/11/27 09:30:44 INFO SecurityManager: Changing modify acls to: cloudera 15/11/27 09:30:44 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(cloudera); users with modify permissions: Set(cloudera) 15/11/27 09:30:45 INFO Slf4jLogger: Slf4jLogger started 15/11/27 09:30:45 INFO Remoting: Starting remoting 15/11/27 09:30:45 INFO Remoting: Remoting started; listening on addresses :[akka.tcp://sparkDriver@10.0.2.15:51812] 15/11/27 09:30:45 INFO Remoting: Remoting now listens on addresses: [akka.tcp://sparkDriver@10.0.2.15:51812] 15/11/27 09:30:45 INFO Utils: Successfully started service 'sparkDriver' on port 51812. 15/11/27 09:30:45 INFO SparkEnv: Registering MapOutputTracker 15/11/27 09:30:45 INFO SparkEnv: Registering BlockManagerMaster 15/11/27 09:30:45 INFO DiskBlockManager: Created local directory at /tmp/blockmgr-e2a8a00b-9bac-4564-a75d-cf2bd864e0be 15/11/27 09:30:45 INFO MemoryStore: MemoryStore started with capacity 883.6 MB 15/11/27 09:30:45 INFO HttpFileServer: HTTP File server directory is /tmp/spark-3e2ba8c0-103d-416a-9963-427825993dfc/httpd-66ccf5bb-e5e7-46ad-9541-5f6438a753cc 15/11/27 09:30:45 INFO HttpServer: Starting HTTP Server 15/11/27 09:30:45 INFO Server: jetty-8.y.z-SNAPSHOT 15/11/27 09:30:46 INFO AbstractConnector: Started SocketConnector@0.0.0.0:36447 15/11/27 09:30:46 INFO Utils: Successfully started service 'HTTP file server' on port 36447. 15/11/27 09:30:46 INFO SparkEnv: Registering OutputCommitCoordinator 15/11/27 09:30:46 INFO Server: jetty-8.y.z-SNAPSHOT 15/11/27 09:30:46 INFO AbstractConnector: Started SelectChannelConnector@0.0.0.0:4040 15/11/27 09:30:46 INFO Utils: Successfully started service 'SparkUI' on port 4040. 15/11/27 09:30:46 INFO SparkUI: Started SparkUI at http://10.0.2.15:4040 15/11/27 09:30:46 INFO SparkContext: Added JAR file:/home/cloudera/workspace/./S2RDF_DataSetCreator/target/scala-2.10/datasetcreator_2.10-1.0.jar at http://10.0.2.15:36447/jars/datasetcreator_2.10-1.0.jar with timestamp 1448613046392 15/11/27 09:30:46 WARN MetricsSystem: Using default name DAGScheduler for source because spark.app.id is not set. 15/11/27 09:30:46 INFO Executor: Starting executor ID driver on host localhost 15/11/27 09:30:46 INFO Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 34264. 15/11/27 09:30:46 INFO NettyBlockTransferService: Server created on 34264 15/11/27 09:30:46 INFO BlockManager: external shuffle service port = 7337 15/11/27 09:30:46 INFO BlockManagerMaster: Trying to register BlockManager 15/11/27 09:30:46 INFO BlockManagerMasterEndpoint: Registering block manager localhost:34264 with 883.6 MB RAM, BlockManagerId(driver, localhost, 34264) 15/11/27 09:30:46 INFO BlockManagerMaster: Registered BlockManager 15/11/27 09:30:47 ERROR SparkContext: Error initializing SparkContext. org.apache.hadoop.security.AccessControlException: Permission denied: user=cloudera, access=WRITE, inode="/user/spark/applicationHistory":spark:supergroup:drwxr-xr-x at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.checkFsPermission(DefaultAuthorizationProvider.java:257) at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.check(DefaultAuthorizationProvider.java:238) at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.check(DefaultAuthorizationProvider.java:216) at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.checkPermission(DefaultAuthorizationProvider.java:145) at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:138) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6609) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6591) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:6543) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInternal(FSNamesystem.java:2756) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInt(FSNamesystem.java:2674) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFile(FSNamesystem.java:2559) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.create(NameNodeRpcServer.java:592) at org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.create(AuthorizationProviderProxyClientProtocol.java:110) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.create(ClientNamenodeProtocolServerSideTranslatorPB.java:395) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1060) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080) at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:526) at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106) at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:73) at org.apache.hadoop.hdfs.DFSOutputStream.newStreamForCreate(DFSOutputStream.java:1872) at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1737) at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1662) at org.apache.hadoop.hdfs.DistributedFileSystem$6.doCall(DistributedFileSystem.java:404) at org.apache.hadoop.hdfs.DistributedFileSystem$6.doCall(DistributedFileSystem.java:400) at org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) at org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:400) at org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:343) at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:917) at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:898) at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:795) at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:784) at org.apache.spark.scheduler.EventLoggingListener.start(EventLoggingListener.scala:121) at org.apache.spark.SparkContext.<init>(SparkContext.scala:541) at dataCreator.Settings$.loadSparkContext(Settings.scala:69) at dataCreator.Settings$.<init>(Settings.scala:17) at dataCreator.Settings$.<clinit>(Settings.scala) at runDriver$.main(runDriver.scala:12) at runDriver.main(runDriver.scala) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:672) at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180) at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205) at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:120) at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException): Permission denied: user=cloudera, access=WRITE, inode="/user/spark/applicationHistory":spark:supergroup:drwxr-xr-x at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.checkFsPermission(DefaultAuthorizationProvider.java:257) at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.check(DefaultAuthorizationProvider.java:238) at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.check(DefaultAuthorizationProvider.java:216) at org.apache.hadoop.hdfs.server.namenode.DefaultAuthorizationProvider.checkPermission(DefaultAuthorizationProvider.java:145) at org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkPermission(FSPermissionChecker.java:138) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6609) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkPermission(FSNamesystem.java:6591) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkAncestorAccess(FSNamesystem.java:6543) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInternal(FSNamesystem.java:2756) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInt(FSNamesystem.java:2674) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFile(FSNamesystem.java:2559) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.create(NameNodeRpcServer.java:592) at org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.create(AuthorizationProviderProxyClientProtocol.java:110) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.create(ClientNamenodeProtocolServerSideTranslatorPB.java:395) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1060) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2086) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2082) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2080) at org.apache.hadoop.ipc.Client.call(Client.java:1472) at org.apache.hadoop.ipc.Client.call(Client.java:1403) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:230) at com.sun.proxy.$Proxy18.create(Unknown Source) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.create(ClientNamenodeProtocolTranslatorPB.java:295) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:252) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:104) at com.sun.proxy.$Proxy19.create(Unknown Source) at org.apache.hadoop.hdfs.DFSOutputStream.newStreamForCreate(DFSOutputStream.java:1867) ... 27 more 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/metrics/json,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/stage/kill,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/api,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/static,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/executors/threadDump/json,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/executors/threadDump,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/executors/json,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/executors,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/environment/json,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/environment,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/storage/rdd/json,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/storage/rdd,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/storage/json,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/storage,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/pool/json,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/pool,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/stage/json,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/stage,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages/json,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/stages,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/jobs/job/json,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/jobs/job,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/jobs/json,null} 15/11/27 09:30:47 INFO ContextHandler: stopped o.s.j.s.ServletContextHandler{/jobs,null} 15/11/27 09:30:48 INFO SparkUI: Stopped Spark web UI at http://10.0.2.15:4040 15/11/27 09:30:48 INFO DAGScheduler: Stopping DAGScheduler 15/11/27 09:30:48 WARN QueuedThreadPool: 2 threads could not be stopped 15/11/27 09:30:48 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped! 15/11/27 09:30:48 INFO MemoryStore: MemoryStore cleared 15/11/27 09:30:48 INFO BlockManager: BlockManager stopped 15/11/27 09:30:48 INFO BlockManagerMaster: BlockManagerMaster stopped 15/11/27 09:30:48 INFO SparkContext: Successfully stopped SparkContext
Created 11-27-2015 05:53 AM
Created on 11-28-2015 02:06 AM - edited 11-28-2015 02:08 AM
You will have problems starting spark-shell with any master because there are permissioning issues with the hdfs folders that spark uses.
To fix it do the following:
sudo -u hdfs hadoop fs -chmod 777 /user/spark
sudo -u spark hadoop fs -chmod 777 /user/spark/applicationHistory
That should do it 🙂
Created 11-27-2015 05:53 AM
Created on 11-28-2015 02:06 AM - edited 11-28-2015 02:08 AM
You will have problems starting spark-shell with any master because there are permissioning issues with the hdfs folders that spark uses.
To fix it do the following:
sudo -u hdfs hadoop fs -chmod 777 /user/spark
sudo -u spark hadoop fs -chmod 777 /user/spark/applicationHistory
That should do it 🙂
Created 01-13-2016 04:25 AM
For me it was solved after login using hdfs user. You can typically then run the spark-shell command without errors