<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Access S3 Bucket from Spark in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Access-S3-Bucket-from-Spark/m-p/230235#M192085</link>
    <description>&lt;P&gt;
	S3A actually has an extra option to let you set per-bucket jceks files, fs.s3a.security.credential.provider.path  This takes the same values as the normal one, but lets you take advantage of the per-bucket config feature of s3a, where every bucket-specific option of fs.s3a.bucket.* is remapped to fs.s3a.* before the bucket is set up.&lt;/P&gt;&lt;P&gt;
	you should be able to add a reference to it likes so&lt;/P&gt;&lt;P&gt;
	spark.hadoop.fs.s3a.bucket.b.security.credential.provider.path hdfs:///something.jceks&lt;/P&gt;&lt;P&gt;
	Hopefully this helps. One challenge we always have with the authentication work is that we can't log it at the detail we'd like, because that would leak secrets too easily...so even when logging at debug, not enough information gets printed. Sorry&lt;/P&gt;&lt;P&gt;
	see also: &lt;A href="https://hortonworks.github.io/hdp-aws/s3-security/index.html"&gt;https://hortonworks.github.io/hdp-aws/s3-security/index.html&lt;/A&gt;&lt;/P&gt;&lt;P&gt;
	Oh, one more thing. spark-submit copies your local AWS_ environment variables over to the fs.s3a.secret,key and fs.s3a.access.key values. Try unsetting them before you submit work and see if that makes a difference&lt;/P&gt;</description>
    <pubDate>Mon, 11 Sep 2017 17:32:28 GMT</pubDate>
    <dc:creator>stevel</dc:creator>
    <dc:date>2017-09-11T17:32:28Z</dc:date>
  </channel>
</rss>

