<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Spark2 Unable to write to HDFS (Or Local) in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Spark2-Unable-to-write-to-HDFS-Or-Local/m-p/53049#M23553</link>
    <description>&lt;P&gt;I am attempting leverage Spark2 to write a parquet file to HDFS, but am receiving the following error:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Error summary: RemoteException: File /user/cloudera/1000genome/processed/test.vds/rdd.parquet/_temporary/0/_temporary/attempt_201703311418_0001_m_000000_3/part-r-00000-c0069d7a-101f-4bf9-9dc9-22b362285b12.snappy.parquet could only be replicated to 0 nodes instead of minReplication (=1). There are 1 datanode(s) running and no node(s) are excluded in this operation.&lt;BR /&gt;at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget4NewBlock(BlockManager.java:1622)&lt;BR /&gt;at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:3351)&lt;BR /&gt;at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:683)&lt;BR /&gt;at org.apache.hadoop.hdfs.server.namenode.AuthorizationProviderProxyClientProtocol.addBlock(AuthorizationProviderProxyClientProtocol.java:214)&lt;BR /&gt;at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:495)&lt;BR /&gt;at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)&lt;BR /&gt;at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:617)&lt;BR /&gt;at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1073)&lt;BR /&gt;at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2216)&lt;BR /&gt;at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2212)&lt;BR /&gt;at java.security.AccessController.doPrivileged(Native Method)&lt;BR /&gt;at javax.security.auth.Subject.doAs(Subject.java:422)&lt;BR /&gt;at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1796)&lt;BR /&gt;at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2210)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I had initially started with the Cloudera Quickstart VM, upgraded Cloudera Manager and CDH to 5.10, and then installed Spark2. &lt;STRONG&gt;Is there something I'm missing that would prevent Spark2 from writing to HDFS?&lt;/STRONG&gt; As well, Spark2 is unable to write this file locally, instead receiving an IO Exception Error, but has no problems reading the file.&lt;BR /&gt;&lt;BR /&gt;I am calling my Spark Script with the below line:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;spark2-submit --master yarn-client tutorial.py&lt;/P&gt;</description>
    <pubDate>Fri, 16 Sep 2022 11:23:33 GMT</pubDate>
    <dc:creator>mtrepanier</dc:creator>
    <dc:date>2022-09-16T11:23:33Z</dc:date>
  </channel>
</rss>

