<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question SPARK SUBMIT issue with SPARK 2.2 in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/SPARK-SUBMIT-issue-with-SPARK-2-2/m-p/217623#M179532</link>
    <description>&lt;P&gt;We are on HDP 2.6.3 and using SPARK 2.2 and running the job using on YARN CLUSTER mode.&lt;/P&gt;&lt;P&gt;using spark-submit and the spark-env.sh contains &lt;STRONG&gt;SPARK_YARN_DIST_FILES="/etc/spark2/conf/hive-site.xml,/etc/spark2/conf/hbase-site.xml"   &lt;/STRONG&gt;but these values are not honored.&lt;/P&gt;&lt;PRE&gt;spark-submit --class com.virtuslab.sparksql.MainClass  --master yarn --deploy-mode cluster /tmp/spark-hive-test/spark_sql_under_the_hood-spark2.2.0.jar&lt;/PRE&gt;&lt;P&gt;This is trying to connect to Hive and fetch the data from a table, but it fails with table on not found in database:&lt;/P&gt;&lt;PRE&gt; diagnostics: User class threw exception: org.apache.spark.sql.catalyst.analysis.NoSuchTableException: Table or view 'xyz' not found in database 'qwerty';
         ApplicationMaster host: 121.121.121.121
         ApplicationMaster RPC port: 0
         queue: default
         start time: 1523616607943
         final status: FAILED
         tracking URL: &lt;A href="https://managenode002xxserver:8090/proxy/application_1523374609937_10224/" target="_blank"&gt;https://managenode002xxserver:8090/proxy/application_1523374609937_10224/&lt;/A&gt;
         user: abc123

Exception in thread "main" org.apache.spark.SparkException: Application application_1523374609937_10224 finished with failed status
        at org.apache.spark.deploy.yarn.Client.run(Client.scala:1187)
        at org.apache.spark.deploy.yarn.Client$.main(Client.scala:1233)
        at org.apache.spark.deploy.yarn.Client.main(Client.scala)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:497)
        at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$runMain(SparkSubmit.scala:782)
        at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180)
        at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:119)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)&lt;/PRE&gt;&lt;P&gt;The same works when we pass the --files parameter:&lt;/P&gt;&lt;PRE&gt;spark-submit --class com.virtuslab.sparksql.MainClass  --master yarn --deploy-mode cluster --files /etc/spark2/conf/hive-site.xml /tmp/spark-hive-test/spark_sql_under_the_hood-spark2.2.0.jar&lt;/PRE&gt;&lt;P&gt;Result attached.&lt;/P&gt;&lt;P&gt;Any pointers why it is not using picking up &lt;STRONG&gt;SPARK_YARN_DIST_FILES&lt;/STRONG&gt;?&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;P&gt;Venkat&lt;/P&gt;</description>
    <pubDate>Fri, 13 Apr 2018 18:01:19 GMT</pubDate>
    <dc:creator>dheer_vijji_rag</dc:creator>
    <dc:date>2018-04-13T18:01:19Z</dc:date>
  </channel>
</rss>

