Support Questions
Find answers, ask questions, and share your expertise
Alert: Welcome to the Unified Cloudera Community. Former HCC members be sure to read and learn how to activate your account here.

Spark Submit Job Does Not Copy Job Files To Azure HortonWorks Sandbox


Spark Submit Job Does Not Copy Job Files To Azure HortonWorks Sandbox

New Contributor


We are trying to execute a spark submit job from our local machine to run against our Azure HortonWorks Sandbox Hadoop Environment. We are experiencing a problem where the transfer of the job files to be submitted does not copy across to the HortonWorks datanode for execution which then results in the error below. Has anyone come across this or a similar issue before? Let me know if there is more information required to help with this query.

The command we are running is: spark-submit --master yarn spark-test-0.0.3-SNAPSHOT.jar

This connects to the Azure HortonWorks environment fine (yarn configs and relevant ports appear setup/forwarded ok) but fails at the point of uploading resources, the following error is thrown:

Exception in thread "main" org.apache.hadoop.ipc.RemoteException( tion): File /user/hdfs spark-shell/.sparkStaging/application_1495640151047_0030/ could only be replicated to 0 nodes instea d of minReplication (=1). There are 1 datanode(s) running and 1 node(s) are exc luded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarg et4NewBlock( at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getNewBlockTarget s( at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBloc k( at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(Nam at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTra nslatorPB.addBlock( at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$Cl ientNamenodeProtocol$2.callBlockingMethod( at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ l( at org.apache.hadoop.ipc.RPC$ at org.apache.hadoop.ipc.Server$Handler$ at org.apache.hadoop.ipc.Server$Handler$ at Method) at at at org.apache.hadoop.ipc.Server$ at at at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEng at com.sun.proxy.$Proxy12.addBlock(Unknown Source) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB. addBlock( at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl. java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces at java.lang.reflect.Method.invoke( at at at com.sun.proxy.$Proxy13.addBlock(Unknown Source) at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBl ock( at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputSt ream( at org.apache.hadoop.hdfs.DFSOutputStream$ 17/05/25 17:28:44 INFO util.ShutdownHookManager: Shutdown hook called 17/05/25 17:28:44 INFO util.ShutdownHookManager: Deleting directory C:\Temp\spar k-c1366f57-54eb-4bb6-ae11-26acf4b5b2f8

Don't have an account?
Coming from Hortonworks? Activate your account here