<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Options for copying Hive data to S3 in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136070#M31710</link>
    <description>&lt;P style="margin-left: 20px;"&gt;Thanks &lt;A rel="user" href="https://community.cloudera.com/users/2528/jyadav.html" nodeid="2528"&gt;@Jitendra Yadav&lt;/A&gt;&lt;/P&gt;&lt;P style="margin-left: 20px;"&gt;Is this baked into HDP, or are there Amazon-related binaries that I need in order for this to work?&lt;/P&gt;</description>
    <pubDate>Mon, 13 Jun 2016 23:30:23 GMT</pubDate>
    <dc:creator>zack_riesland</dc:creator>
    <dc:date>2016-06-13T23:30:23Z</dc:date>
    <item>
      <title>Options for copying Hive data to S3</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136068#M31708</link>
      <description>&lt;P&gt;I want to copy some data from Hive tables on our (bare metal) cluster to a S3.&lt;/P&gt;&lt;P&gt;I know that I can export the data out of HDFS to a CSV file and upload that to S3, but I'm guessing that there are better ways to accomplish this.&lt;/P&gt;&lt;P&gt;Any ideas?&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 10:24:56 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136068#M31708</guid>
      <dc:creator>zack_riesland</dc:creator>
      <dc:date>2022-09-16T10:24:56Z</dc:date>
    </item>
    <item>
      <title>Re: Options for copying Hive data to S3</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136069#M31709</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/2593/zackriesland.html" nodeid="2593"&gt;@Zack Riesland&lt;/A&gt; You can put it directly through "hdfs fs -put /tablepath  s3://bucket/hivetable. If you have partitions in hive table and you can run this command for each partition directory in concurrent mode through a small shell script just to increase the data ingestion speed.&lt;/P&gt;&lt;P&gt;And same S3 data can be used again in hive external table.&lt;/P&gt;&lt;PRE&gt;CREATE EXTERNAL TABLE mydata (key STRING, value INT)
    ROW FORMAT DELIMITED FIELDS TERMINATED BY ' '
    LOCATION 's3n://mysbucket/';&lt;/PRE&gt;</description>
      <pubDate>Mon, 13 Jun 2016 22:50:06 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136069#M31709</guid>
      <dc:creator>jyadav</dc:creator>
      <dc:date>2016-06-13T22:50:06Z</dc:date>
    </item>
    <item>
      <title>Re: Options for copying Hive data to S3</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136070#M31710</link>
      <description>&lt;P style="margin-left: 20px;"&gt;Thanks &lt;A rel="user" href="https://community.cloudera.com/users/2528/jyadav.html" nodeid="2528"&gt;@Jitendra Yadav&lt;/A&gt;&lt;/P&gt;&lt;P style="margin-left: 20px;"&gt;Is this baked into HDP, or are there Amazon-related binaries that I need in order for this to work?&lt;/P&gt;</description>
      <pubDate>Mon, 13 Jun 2016 23:30:23 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136070#M31710</guid>
      <dc:creator>zack_riesland</dc:creator>
      <dc:date>2016-06-13T23:30:23Z</dc:date>
    </item>
    <item>
      <title>Re: Options for copying Hive data to S3</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136071#M31711</link>
      <description>&lt;P&gt;Yes it is baked by HDP, we only need to make that S3 secret keys are in place.&lt;/P&gt;&lt;P&gt;see this doc. &lt;A href="https://community.hortonworks.com/articles/25578/how-to-access-data-files-stored-in-aws-s3-buckets.html" target="_blank"&gt;https://community.hortonworks.com/articles/25578/how-to-access-data-files-stored-in-aws-s3-buckets.html&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 13 Jun 2016 23:46:10 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136071#M31711</guid>
      <dc:creator>jyadav</dc:creator>
      <dc:date>2016-06-13T23:46:10Z</dc:date>
    </item>
    <item>
      <title>Re: Options for copying Hive data to S3</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136072#M31712</link>
      <description>&lt;P&gt;In addition to the above, you might also want to install and configure aws cli:&lt;/P&gt;&lt;P&gt;&lt;A href="http://docs.aws.amazon.com/cli/latest/userguide/cli-chap-getting-started.html" target="_blank"&gt;http://docs.aws.amazon.com/cli/latest/userguide/cli-chap-getting-started.html&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 14 Jun 2016 04:48:27 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136072#M31712</guid>
      <dc:creator>namaheshwari</dc:creator>
      <dc:date>2016-06-14T04:48:27Z</dc:date>
    </item>
    <item>
      <title>Re: Options for copying Hive data to S3</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136073#M31713</link>
      <description>&lt;P&gt;Hi &lt;A rel="user" href="https://community.cloudera.com/users/2593/zackriesland.html" nodeid="2593"&gt;@Zack Riesland&lt;/A&gt; please let me know if you required further info or accept this answer to close this thread.&lt;/P&gt;</description>
      <pubDate>Wed, 15 Jun 2016 05:08:17 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136073#M31713</guid>
      <dc:creator>jyadav</dc:creator>
      <dc:date>2016-06-15T05:08:17Z</dc:date>
    </item>
    <item>
      <title>Re: Options for copying Hive data to S3</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136074#M31714</link>
      <description>&lt;P&gt;I added the appropriate entries to hive and hdfs configs in ambari (as specified here &lt;A href="https://community.hortonworks.com/articles/25578/how-to-access-data-files-stored-in-aws-s3-buckets.html"&gt;https://community.hortonworks.com/articles/25578/how-to-access-data-files-stored-in-aws-s3-buckets.html&lt;/A&gt;), and gave this a try:&lt;/P&gt;&lt;P&gt;hdfs &lt;STRONG&gt;d&lt;/STRONG&gt;fs -put /user/my_user/my_hdfs_file s3://my_bucket/my_folder&lt;/P&gt;&lt;P&gt;I got the error:&lt;/P&gt;&lt;P&gt;-put: AWS Access Key ID and Secret Access Key must be specified as the username or password (respectively) of a s3 URL, or by setting the fs.s3.awsAccessKeyId or fs.s3.awsSecretAccessKey properties (respectively).&lt;/P&gt;&lt;P&gt;I noticed that the instructions mention these settings:&lt;/P&gt;&lt;P&gt;fs.s3&lt;STRONG&gt;n&lt;/STRONG&gt;.awsAccessKeyId&lt;/P&gt;&lt;P&gt;fs.s3&lt;STRONG&gt;n&lt;/STRONG&gt;.awsSecretAccessKey&lt;/P&gt;&lt;P&gt;but the error message mentions these:&lt;/P&gt;&lt;P&gt;fs.s3.awsAccessKeyId&lt;/P&gt;&lt;P&gt;fs.s3.awsSecretAccessKey&lt;/P&gt;&lt;P&gt;Once I made that change, I was able to make some progress.&lt;/P&gt;&lt;P&gt;However, I think I still need a little help.&lt;/P&gt;&lt;P&gt;In your example, you show s3://bucket/hivetable as the destination.&lt;/P&gt;&lt;P&gt;But our S3 instance doesn't have tables, just folders. When I try and point at a folder, I get an error:&lt;/P&gt;&lt;P&gt;put: /&amp;lt;folder name&amp;gt; doesn't exist&lt;/P&gt;&lt;P&gt;Do I need to use the other syntax "create external table... LOCATION 's3n://mysbucket/" to create a TABLE in S3 and then access in this way?&lt;/P&gt;&lt;P&gt;Is there a similar way to simply transfer a file from hdfs to a FOLDER in an s3 bucket?&lt;/P&gt;&lt;P&gt;cc &lt;A rel="user" href="https://community.cloudera.com/users/2528/jyadav.html" nodeid="2528"&gt;@Jitendra Yadav&lt;/A&gt;&lt;/P&gt;&lt;P&gt;Thanks!&lt;/P&gt;</description>
      <pubDate>Wed, 15 Jun 2016 05:44:27 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136074#M31714</guid>
      <dc:creator>zack_riesland</dc:creator>
      <dc:date>2016-06-15T05:44:27Z</dc:date>
    </item>
    <item>
      <title>Re: Options for copying Hive data to S3</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136075#M31715</link>
      <description>&lt;P&gt;Yes, You can also use s3n instead of s3 as mentioned in the article and make sure secretekey defined in s3n properties.&lt;/P&gt;</description>
      <pubDate>Wed, 15 Jun 2016 15:46:25 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136075#M31715</guid>
      <dc:creator>jyadav</dc:creator>
      <dc:date>2016-06-15T15:46:25Z</dc:date>
    </item>
    <item>
      <title>Re: Options for copying Hive data to S3</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136076#M31716</link>
      <description>&lt;P&gt;I didn't understand the difference between s3 and s3n.&lt;/P&gt;&lt;P&gt;This link helped: &lt;A href="http://stackoverflow.com/questions/10569455/difference-between-amazon-s3-and-s3n-in-hadoop"&gt;http://stackoverflow.com/questions/10569455/difference-between-amazon-s3-and-s3n-in-hadoop&lt;/A&gt;&lt;/P&gt;&lt;P&gt;Thanks again.&lt;/P&gt;</description>
      <pubDate>Wed, 15 Jun 2016 16:39:31 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Options-for-copying-Hive-data-to-S3/m-p/136076#M31716</guid>
      <dc:creator>zack_riesland</dc:creator>
      <dc:date>2016-06-15T16:39:31Z</dc:date>
    </item>
  </channel>
</rss>

