<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Confusion in documentation : Configuring the Spark Thrift Server on a Kerberos-Enabled Cluster in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Confusion-in-documentation-Configuring-the-Spark-Thrift/m-p/154581#M28780</link>
    <description>&lt;P&gt;You are going to use hive account to run spark thrift server. So, if it is a manual install, then &lt;/P&gt;&lt;P&gt;./sbin/start-thriftserver.sh --master yarn-client --executor-memory 512m --hiveconf hive.server2.thrift.port=10015&lt;/P&gt;&lt;P&gt;will be run as user hive (with su hive) instead of user spark in secure setup. Similarly /var/run/spark and /var/log/spark should be read/write to hive. So, just seeing contents as user hive is not enough, you need to be able to write to those folders. One good easy way is to give 77x permissions on these folders. Since spark:hadoop is owner:group and hive belongs to group hadoop, it will have write access with this setup. &lt;/P&gt;</description>
    <pubDate>Thu, 26 May 2016 13:24:39 GMT</pubDate>
    <dc:creator>ravi1</dc:creator>
    <dc:date>2016-05-26T13:24:39Z</dc:date>
    <item>
      <title>Confusion in documentation : Configuring the Spark Thrift Server on a Kerberos-Enabled Cluster</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Confusion-in-documentation-Configuring-the-Spark-Thrift/m-p/154580#M28779</link>
      <description>&lt;P&gt;Guys,&lt;/P&gt;&lt;P&gt;I am referring the document : &lt;A href="http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.4.2/bk_spark-guide/content/spark-kerb-access-hive.html" target="_blank"&gt;http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.4.2/bk_spark-guide/content/spark-kerb-access-hive.html&lt;/A&gt; and got a bit confused. Wanted to check with experts who already configured spark thrift server on kerberized environment. &lt;/P&gt;&lt;P&gt;If you are installing the Spark Thrift Server on a Kerberos-secured cluster, note the following requirements:&lt;/P&gt;&lt;UL&gt;
&lt;LI&gt;The Spark Thrift Server must run in the same host as &lt;CODE&gt;HiveServer2&lt;/CODE&gt;, so that it can access the &lt;CODE&gt;hiveserver2&lt;/CODE&gt; keytab.&lt;UL&gt;&lt;LI&gt;&lt;EM&gt;OK. Install and run Spark TS on the same host as HS2. Install STS using Ambari.&lt;/EM&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Edit permissions&lt;/STRONG&gt; in &lt;CODE&gt;/var/run/spark&lt;/CODE&gt; and &lt;CODE&gt;/var/log/spark&lt;/CODE&gt; to specify read/write permissions to the Hive service account.&lt;UL&gt;&lt;LI&gt;&lt;EM&gt;Not very clear here. I see that in our cluster, we have a user spark. And I tried to do &lt;STRONG&gt;ls /var/run&lt;/STRONG&gt; and&lt;STRONG&gt; ls /var/run/spark &lt;/STRONG&gt;as spark user and as hive user (after su spark) I see the directory contents in both cases. Is it correct or am I supposed to something else because &lt;STRONG&gt;I didn't edit the permissions. What permissions to to be edited?&lt;/STRONG&gt;&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;ll /var/run&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;drwxrwxr-x 3 spark     hadoop    4096 May 17 10:47 spark
&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;
&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt; ll /var/run/spark 
&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;-rw-r--r-- 1 root  root     6 May 17 11:18 spark-root-org.apache.spark.deploy.history.HistoryServer-1.pid&lt;/LI&gt;&lt;LI&gt;ll /var/log/&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;drwxr-xr-x 2 spark     spark             4096 Mar  9 10:06 spark&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;ll /var/log/spark&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;
&lt;/EM&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;LI&gt;Use the Hive service account to start the &lt;CODE&gt;thriftserver&lt;/CODE&gt; process. &lt;UL&gt;&lt;LI&gt;&lt;EM&gt;Does it mean, I got to do kinit with hive keytab or do su hive and start the thrift server.?&lt;/EM&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;Thanks.&lt;/P&gt;</description>
      <pubDate>Wed, 18 May 2016 18:40:41 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Confusion-in-documentation-Configuring-the-Spark-Thrift/m-p/154580#M28779</guid>
      <dc:creator>smartninja723</dc:creator>
      <dc:date>2016-05-18T18:40:41Z</dc:date>
    </item>
    <item>
      <title>Re: Confusion in documentation : Configuring the Spark Thrift Server on a Kerberos-Enabled Cluster</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Confusion-in-documentation-Configuring-the-Spark-Thrift/m-p/154581#M28780</link>
      <description>&lt;P&gt;You are going to use hive account to run spark thrift server. So, if it is a manual install, then &lt;/P&gt;&lt;P&gt;./sbin/start-thriftserver.sh --master yarn-client --executor-memory 512m --hiveconf hive.server2.thrift.port=10015&lt;/P&gt;&lt;P&gt;will be run as user hive (with su hive) instead of user spark in secure setup. Similarly /var/run/spark and /var/log/spark should be read/write to hive. So, just seeing contents as user hive is not enough, you need to be able to write to those folders. One good easy way is to give 77x permissions on these folders. Since spark:hadoop is owner:group and hive belongs to group hadoop, it will have write access with this setup. &lt;/P&gt;</description>
      <pubDate>Thu, 26 May 2016 13:24:39 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Confusion-in-documentation-Configuring-the-Spark-Thrift/m-p/154581#M28780</guid>
      <dc:creator>ravi1</dc:creator>
      <dc:date>2016-05-26T13:24:39Z</dc:date>
    </item>
  </channel>
</rss>

