<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Unable to distcp from on premise hdfs to s3 in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Unable-to-distcp-from-on-premise-hdfs-to-s3/m-p/129526#M51606</link>
    <description>&lt;P&gt;I am using this &lt;A href="https://community.hortonworks.com/articles/59161/using-hadoop-credential-api-to-store-aws-secrets.html" target="_blank"&gt;https://community.hortonworks.com/articles/59161/using-hadoop-credential-api-to-store-aws-secrets.html&lt;/A&gt; as a reference, but unable to successfully distcp or even -ls.&lt;/P&gt;&lt;P&gt;[root@test232 conf]# hadoop distcp -Dhadoop.security.credential.provider.path=jceks://hdfs/aws/aws.jceks /nsswitch.conf s3a://kartik-test.s3-us-west-2.amazonaws.com &lt;/P&gt;&lt;P&gt;17/01/15 08:16:22 ERROR tools.DistCp: Invalid arguments:
java.io.IOException: Bucket kartik-test.s3-us-west-2.amazonaws.com does not exist
        at org.apache.hadoop.fs.s3a.S3AFileSystem.initialize(S3AFileSystem.java:229)
        at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2653)
        at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:92)
        at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2687)
        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2669)
        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:371)
        at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295)
        at org.apache.hadoop.tools.DistCp.setTargetPathExists(DistCp.java:224)
        at org.apache.hadoop.tools.DistCp.run(DistCp.java:116)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
        at org.apache.hadoop.tools.DistCp.main(DistCp.java:454)
Invalid arguments: Bucket kartik-test.s3-us-west-2.amazonaws.com does not exist
&lt;/P&gt;&lt;P&gt;I am now reading this: &lt;A href="https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html" target="_blank"&gt;https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html&lt;/A&gt;&lt;/P&gt;&lt;P&gt;but unable to understand what are the xml file names that need to be edited.&lt;/P&gt;</description>
    <pubDate>Sun, 15 Jan 2017 21:58:56 GMT</pubDate>
    <dc:creator>ddolecki</dc:creator>
    <dc:date>2017-01-15T21:58:56Z</dc:date>
    <item>
      <title>Unable to distcp from on premise hdfs to s3</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Unable-to-distcp-from-on-premise-hdfs-to-s3/m-p/129526#M51606</link>
      <description>&lt;P&gt;I am using this &lt;A href="https://community.hortonworks.com/articles/59161/using-hadoop-credential-api-to-store-aws-secrets.html" target="_blank"&gt;https://community.hortonworks.com/articles/59161/using-hadoop-credential-api-to-store-aws-secrets.html&lt;/A&gt; as a reference, but unable to successfully distcp or even -ls.&lt;/P&gt;&lt;P&gt;[root@test232 conf]# hadoop distcp -Dhadoop.security.credential.provider.path=jceks://hdfs/aws/aws.jceks /nsswitch.conf s3a://kartik-test.s3-us-west-2.amazonaws.com &lt;/P&gt;&lt;P&gt;17/01/15 08:16:22 ERROR tools.DistCp: Invalid arguments:
java.io.IOException: Bucket kartik-test.s3-us-west-2.amazonaws.com does not exist
        at org.apache.hadoop.fs.s3a.S3AFileSystem.initialize(S3AFileSystem.java:229)
        at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2653)
        at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:92)
        at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2687)
        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2669)
        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:371)
        at org.apache.hadoop.fs.Path.getFileSystem(Path.java:295)
        at org.apache.hadoop.tools.DistCp.setTargetPathExists(DistCp.java:224)
        at org.apache.hadoop.tools.DistCp.run(DistCp.java:116)
        at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
        at org.apache.hadoop.tools.DistCp.main(DistCp.java:454)
Invalid arguments: Bucket kartik-test.s3-us-west-2.amazonaws.com does not exist
&lt;/P&gt;&lt;P&gt;I am now reading this: &lt;A href="https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html" target="_blank"&gt;https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html&lt;/A&gt;&lt;/P&gt;&lt;P&gt;but unable to understand what are the xml file names that need to be edited.&lt;/P&gt;</description>
      <pubDate>Sun, 15 Jan 2017 21:58:56 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Unable-to-distcp-from-on-premise-hdfs-to-s3/m-p/129526#M51606</guid>
      <dc:creator>ddolecki</dc:creator>
      <dc:date>2017-01-15T21:58:56Z</dc:date>
    </item>
    <item>
      <title>Re: Unable to distcp from on premise hdfs to s3</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Unable-to-distcp-from-on-premise-hdfs-to-s3/m-p/129527#M51607</link>
      <description>&lt;P&gt;This worked: [hdfs@test232 ~]$ hadoop distcp -Dhadoop.security.credential.provider.path=jceks://hdfs/aws/aws.jceks /test s3a://kartik-test/&lt;/P&gt;&lt;P&gt;Thanks for all your help!!&lt;/P&gt;</description>
      <pubDate>Sun, 15 Jan 2017 23:49:59 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Unable-to-distcp-from-on-premise-hdfs-to-s3/m-p/129527#M51607</guid>
      <dc:creator>ddolecki</dc:creator>
      <dc:date>2017-01-15T23:49:59Z</dc:date>
    </item>
  </channel>
</rss>

