<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: remote pyspark shell and spark-submit error java.lang.NoSuchFieldError: METASTORE_CLIENT_SOCKET_LIFETIME in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/remote-pyspark-shell-and-spark-submit-error-java-lang/m-p/310328#M224103</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/67146"&gt;@jagadeesan&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;thank you for your reply.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;This airflow docker node is not in Cloudera environment, it's spark, hive, hadoop and java dependencies are not managed by CDH CM node. CDH 5.15 comes with spark 2.3.0, hadoop 2.6.0, hive 1.1.0 and java8 version.&lt;/P&gt;&lt;P&gt;To submit a job to yarn, all versions must be the same (I know I've read this somewhere, I'm not going crazy yet :D). That's why we downloaded &lt;SPAN&gt;spark-2.3.0-bin-without-hadoop, hive-1.1.0 and so on.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;We even tried to build &lt;A href="https://github.com/cloudera/spark" target="_self"&gt;CDH spark&lt;/A&gt;.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;SPARK_HADOOP_VERSION=2.3.0.cloudera4 SPARK_YARN=true sbt assembly&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;But it throws error:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;[warn] 	module not found: org.apache.hadoop#hadoop-client;2.6.0-cdh5.13.3
[warn] ==== public: tried
[warn]   https://repo1.maven.org/maven2/org/apache/hadoop/hadoop-client/2.6.0-cdh5.13.3/hadoop-client-2.6.0-cdh5.13.3.pom&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Our next shot is to add this airflow node in CDH env and assign spark, hive and hdfs gateways.&amp;nbsp;Apparently, this type of node does not need additional license...&lt;/P&gt;</description>
    <pubDate>Sun, 24 Jan 2021 17:13:41 GMT</pubDate>
    <dc:creator>adrijand</dc:creator>
    <dc:date>2021-01-24T17:13:41Z</dc:date>
    <item>
      <title>remote pyspark shell and spark-submit error java.lang.NoSuchFieldError: METASTORE_CLIENT_SOCKET_LIFETIME</title>
      <link>https://community.cloudera.com/t5/Support-Questions/remote-pyspark-shell-and-spark-submit-error-java-lang/m-p/309553#M223853</link>
      <description>&lt;P&gt;Hi all,&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;we are executing pyspark and spark-submit to kerberized CDH 5.15v from remote airflow docker container not managed by CDH CM node, e.g. airflow container is not in CDH env. Versions of hive, spark and java are the same as on CDH. There is a valid kerberos ticket before executing spark-submit or pyspark.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;Python script:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="python"&gt;from pyspark.sql import SparkSession, functions as F
spark = SparkSession.builder.enableHiveSupport().appName('appName').getOrCreate()
sa_df=spark.sql("SELECT * FROM lnz_ch.lnz_cfg_codebook")&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Error is:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="java"&gt;To adjust logging level use sc.setLogLevel(newLevel). For SparkR, use setLogLevel(newLevel).
Welcome to
      ____              __
     / __/__  ___ _____/ /__
    _\ \/ _ \/ _ `/ __/  '_/
   /__ / .__/\_,_/_/ /_/\_\   version 2.3.0
      /_/

Using Python version 3.6.12 (default, Oct 13 2020 21:45:01)
SparkSession available as 'spark'.
&amp;gt;&amp;gt;&amp;gt; from pyspark.sql import SparkSession, functions as F
&amp;gt;&amp;gt;&amp;gt; spark = SparkSession.builder.enableHiveSupport().appName('appName').getOrCreate()
&amp;gt;&amp;gt;&amp;gt; sa_df=spark.sql("SELECT * FROM lnz_ch.lnz_cfg_codebook")
Traceback (most recent call last):
  File "&amp;lt;stdin&amp;gt;", line 1, in &amp;lt;module&amp;gt;
  File "/var/lib/airflow/spark/spark-2.3.0-bin-without-hadoop/python/pyspark/sql/session.py", line 708, in sql
    return DataFrame(self._jsparkSession.sql(sqlQuery), self._wrapped)
  File "/var/lib/airflow/spark/spark-2.3.0-bin-without-hadoop/python/lib/py4j-0.10.6-src.zip/py4j/java_gateway.py", line 1160, in __call__
  File "/var/lib/airflow/spark/spark-2.3.0-bin-without-hadoop/python/pyspark/sql/utils.py", line 63, in deco
    return f(*a, **kw)
  File "/var/lib/airflow/spark/spark-2.3.0-bin-without-hadoop/python/lib/py4j-0.10.6-src.zip/py4j/protocol.py", line 320, in get_return_value
py4j.protocol.Py4JJavaError: An error occurred while calling o24.sql.
: java.lang.NoSuchFieldError: METASTORE_CLIENT_SOCKET_LIFETIME
        at org.apache.spark.sql.hive.HiveUtils$.formatTimeVarsForHiveClient(HiveUtils.scala:195)
        at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:286)
        at org.apache.spark.sql.hive.HiveExternalCatalog.client$lzycompute(HiveExternalCatalog.scala:66)
        at org.apache.spark.sql.hive.HiveExternalCatalog.client(HiveExternalCatalog.scala:65)
        at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply$mcZ$sp(HiveExternalCatalog.scala:195)
        at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:195)
        at org.apache.spark.sql.hive.HiveExternalCatalog$$anonfun$databaseExists$1.apply(HiveExternalCatalog.scala:195)
        at org.apache.spark.sql.hive.HiveExternalCatalog.withClient(HiveExternalCatalog.scala:97)
        at org.apache.spark.sql.hive.HiveExternalCatalog.databaseExists(HiveExternalCatalog.scala:194)
        at org.apache.spark.sql.internal.SharedState.externalCatalog$lzycompute(SharedState.scala:114)
        at org.apache.spark.sql.internal.SharedState.externalCatalog(SharedState.scala:102)
        at org.apache.spark.sql.hive.HiveSessionStateBuilder.externalCatalog(HiveSessionStateBuilder.scala:39)
        at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog$lzycompute(HiveSessionStateBuilder.scala:54)
        at org.apache.spark.sql.hive.HiveSessionStateBuilder.catalog(HiveSessionStateBuilder.scala:52)
        at org.apache.spark.sql.hive.HiveSessionStateBuilder$$anon$1.&amp;lt;init&amp;gt;(HiveSessionStateBuilder.scala:69)
        at org.apache.spark.sql.hive.HiveSessionStateBuilder.analyzer(HiveSessionStateBuilder.scala:69)
        at org.apache.spark.sql.internal.BaseSessionStateBuilder$$anonfun$build$2.apply(BaseSessionStateBuilder.scala:293)
        at org.apache.spark.sql.internal.BaseSessionStateBuilder$$anonfun$build$2.apply(BaseSessionStateBuilder.scala:293)
        at org.apache.spark.sql.internal.SessionState.analyzer$lzycompute(SessionState.scala:79)
        at org.apache.spark.sql.internal.SessionState.analyzer(SessionState.scala:79)
        at org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:57)
        at org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:55)
        at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:47)
        at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:74)
        at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:638)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
        at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
        at py4j.Gateway.invoke(Gateway.java:282)
        at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
        at py4j.commands.CallCommand.execute(CallCommand.java:79)
        at py4j.GatewayConnection.run(GatewayConnection.java:214)
        at java.lang.Thread.run(Thread.java:748)&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Same error is retured from yarn when executing spark-submit.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Details:&lt;/P&gt;&lt;P&gt;googling this error, we assume that the versions of spark and hive in airflow container are "somehow mismatched". Error still occurs if we specify spark-submit or pyspark like this:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;spark-shell \
--jars \
/var/lib/airflow/spark/apache-hive-1.1.0-bin/lib/hive-metastore-1.1.0.jar,\
/var/lib/airflow/spark/apache-hive-1.1.0-bin/lib/hive-exec-1.1.0.jar,\
/var/lib/airflow/spark/apache-hive-1.1.0-bin/lib/hive-common-1.1.0.jar,\
/var/lib/airflow/spark/apache-hive-1.1.0-bin/lib/hive-serde-1.1.0.jar,\
/var/lib/airflow/spark/apache-hive-1.1.0-bin/lib/guava-14.0.1.jar,\
/var/lib/airflow/spark/HiveJDBC4.jar \
--conf spark.sql.hive.metastore.version=1.1.0 \
--conf spark.sql.hive.metastore.jars=/var/lib/airflow/spark/spark-2.3.0-bin-without-hadoop/jars/*&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;As you can see, we are heavily experimenting with --jars argument :D.&lt;/P&gt;&lt;P&gt;Any ideas?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank you.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Links&lt;/P&gt;&lt;P&gt;&lt;A href="https://issues.apache.org/jira/browse/SPARK-14492" target="_blank"&gt;https://issues.apache.org/jira/browse/SPARK-14492&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;A href="https://jaceklaskowski.gitbooks.io/mastering-spark-sql/content/spark-sql-properties.html" target="_blank"&gt;https://jaceklaskowski.gitbooks.io/mastering-spark-sql/content/spark-sql-properties.html&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;A href="https://jaceklaskowski.gitbooks.io/mastering-spark-sql/content/hive/" target="_blank"&gt;https://jaceklaskowski.gitbooks.io/mastering-spark-sql/content/hive/&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 13 Jan 2021 09:01:22 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/remote-pyspark-shell-and-spark-submit-error-java-lang/m-p/309553#M223853</guid>
      <dc:creator>adrijand</dc:creator>
      <dc:date>2021-01-13T09:01:22Z</dc:date>
    </item>
    <item>
      <title>Re: remote pyspark shell and spark-submit error java.lang.NoSuchFieldError: METASTORE_CLIENT_SOCKET_LIFETIME</title>
      <link>https://community.cloudera.com/t5/Support-Questions/remote-pyspark-shell-and-spark-submit-error-java-lang/m-p/310313#M224096</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/84442"&gt;@adrijand&lt;/a&gt;&lt;/P&gt;&lt;P&gt;Yeah, it seems some jar conflicts somewhere.&lt;SPAN&gt;&amp;nbsp;You are trying to load Hive 1.1.0&amp;nbsp;&lt;/SPAN&gt;&lt;SPAN&gt;classes before the ones included with Spark, and as such, they might fail to reference a Hive configuration that didn't exist in 1.1.0. like below.&lt;/SPAN&gt;&lt;/P&gt;&lt;PRE&gt;: java.lang.NoSuchFieldError: METASTORE_CLIENT_SOCKET_LIFETIME
        at org.apache.spark.sql.hive.HiveUtils$.formatTimeVarsForHiveClient(HiveUtils.scala:195)
        at org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:286)
        at org.apache.spark.sql.hive.HiveExternalCatalog.client$lzycompute(HiveExternalCatalog.scala:66)&lt;/PRE&gt;&lt;P&gt;&lt;SPAN&gt;But here in the description mentioned you are using CDH 5.15v but in your log snippets it showing Apache Spark (&lt;/SPAN&gt;spark-2.3.0-bin-without-hadoop&lt;SPAN&gt;) and Apache Hive&amp;nbsp;(apache-hive-1.1.0-bin) version which is not a pre-built package version that comes along with CDH&amp;nbsp;stack compatibility. Are you trying with&amp;nbsp;building with varying versions of Hive which you would like to connect from a remote airflow docker container?&amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Sat, 23 Jan 2021 14:41:14 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/remote-pyspark-shell-and-spark-submit-error-java-lang/m-p/310313#M224096</guid>
      <dc:creator>jagadeesan</dc:creator>
      <dc:date>2021-01-23T14:41:14Z</dc:date>
    </item>
    <item>
      <title>Re: remote pyspark shell and spark-submit error java.lang.NoSuchFieldError: METASTORE_CLIENT_SOCKET_LIFETIME</title>
      <link>https://community.cloudera.com/t5/Support-Questions/remote-pyspark-shell-and-spark-submit-error-java-lang/m-p/310328#M224103</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/67146"&gt;@jagadeesan&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;thank you for your reply.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;This airflow docker node is not in Cloudera environment, it's spark, hive, hadoop and java dependencies are not managed by CDH CM node. CDH 5.15 comes with spark 2.3.0, hadoop 2.6.0, hive 1.1.0 and java8 version.&lt;/P&gt;&lt;P&gt;To submit a job to yarn, all versions must be the same (I know I've read this somewhere, I'm not going crazy yet :D). That's why we downloaded &lt;SPAN&gt;spark-2.3.0-bin-without-hadoop, hive-1.1.0 and so on.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;We even tried to build &lt;A href="https://github.com/cloudera/spark" target="_self"&gt;CDH spark&lt;/A&gt;.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;SPARK_HADOOP_VERSION=2.3.0.cloudera4 SPARK_YARN=true sbt assembly&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;But it throws error:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;[warn] 	module not found: org.apache.hadoop#hadoop-client;2.6.0-cdh5.13.3
[warn] ==== public: tried
[warn]   https://repo1.maven.org/maven2/org/apache/hadoop/hadoop-client/2.6.0-cdh5.13.3/hadoop-client-2.6.0-cdh5.13.3.pom&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Our next shot is to add this airflow node in CDH env and assign spark, hive and hdfs gateways.&amp;nbsp;Apparently, this type of node does not need additional license...&lt;/P&gt;</description>
      <pubDate>Sun, 24 Jan 2021 17:13:41 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/remote-pyspark-shell-and-spark-submit-error-java-lang/m-p/310328#M224103</guid>
      <dc:creator>adrijand</dc:creator>
      <dc:date>2021-01-24T17:13:41Z</dc:date>
    </item>
    <item>
      <title>Re: remote pyspark shell and spark-submit error java.lang.NoSuchFieldError: METASTORE_CLIENT_SOCKET_LIFETIME</title>
      <link>https://community.cloudera.com/t5/Support-Questions/remote-pyspark-shell-and-spark-submit-error-java-lang/m-p/310971#M224388</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/84442"&gt;@adrijand&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks for your detailed explanation here. Yeah indeed, we need&amp;nbsp;&lt;SPAN&gt;all versions to be the same to avoid some&amp;nbsp;classnotfoundexception because of jar conflicts.&amp;nbsp;We encourage you to explore these and provide feedback on your experiences.&amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 04 Feb 2021 05:39:35 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/remote-pyspark-shell-and-spark-submit-error-java-lang/m-p/310971#M224388</guid>
      <dc:creator>jagadeesan</dc:creator>
      <dc:date>2021-02-04T05:39:35Z</dc:date>
    </item>
    <item>
      <title>Re: remote pyspark shell and spark-submit error java.lang.NoSuchFieldError: METASTORE_CLIENT_SOCKET_LIFETIME</title>
      <link>https://community.cloudera.com/t5/Support-Questions/remote-pyspark-shell-and-spark-submit-error-java-lang/m-p/311760#M224814</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/67146"&gt;@jagadeesan&lt;/a&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;here is the short update. After numerous failed attempts to build&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;SPARK_HADOOP_VERSION=2.3.0.cloudera4&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;we added airflow node to CDH env. That was done by installing cloudera agent and registering the node to ClouderaManager. Next step was to edit docker-compose airflow.yaml file:&lt;/P&gt;&lt;P&gt;-volumes mount&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;...
# java home
- /usr/java/jdk1.8.0_162:/usr/java/jdk1.8.0_162
# krb5.conf
- /etc/krb5.conf:/etc/krb5.conf:ro,z
# CDH bin
- /opt/cloudera/parcels:/opt/cloudera/parcels
# /etc
- /etc/hadoop:/etc/hadoop:rw,z
- /etc/spark2:/etc/spark2:rw,z
- /etc/sqoop:/etc/sqoop:rw,z
# sqoop
- /var/lib/sqoop:/var/lib/sqoop
...&lt;/LI-CODE&gt;&lt;P&gt;-env var&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;###java
JAVA_HOME=/usr/java/jdk1.8.0_162&lt;/LI-CODE&gt;&lt;P&gt;Inside the container, symlinks were missing, so:&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;###java
ln -s /usr/java/jdk1.8.0_162/jre/bin/java /etc/alternatives/java
ln -s /etc/alternatives/java /usr/bin/java
#
ln -s /usr/java/jdk1.8.0_162/bin/java-rmi.cgi /etc/alternatives/java-rmi.cgi
ln -s /etc/alternatives/java-rmi.cgi /usr/bin/java-rmi.cgi
#
ln -s /usr/java/jdk1.8.0_162/bin/javac /etc/alternatives/javac
ln -s /etc/alternatives/javac /usr/bin/javac
#
ln -s /usr/java/jdk1.8.0_162/bin/javaws /etc/alternatives/javaws
ln -s /etc/alternatives/javaws /usr/bin/javaws
#
ln -s /usr/java/jdk1.8.0_162/bin/javapackager /etc/alternatives/javapackager
ln -s /etc/alternatives/javapackager /usr/bin/javapackager
#
ln -s /usr/java/jdk1.8.0_162/bin/javap /etc/alternatives/javap
ln -s /etc/alternatives/javap /usr/bin/javap
#
ln -s /usr/java/jdk1.8.0_162/bin/javah /etc/alternatives/javah
ln -s /etc/alternatives/javah /usr/bin/javah
#
ln -s /usr/java/jdk1.8.0_162/bin/javafxpackager /etc/alternatives/javafxpackager
ln -s /etc/alternatives/javafxpackager /usr/bin/javafxpackager
#
ln -s /usr/java/jdk1.8.0_162/bin/javadoc /etc/alternatives/javadoc
ln -s /etc/alternatives/javadoc /usr/bin/javadoc

###spark2-submit
ln -s /opt/cloudera/parcels/SPARK2-2.3.0.cloudera4-1.cdh5.13.3.p0.611179/bin/spark2-submit /etc/alternatives/spark2-submit
ln -s /etc/alternatives/spark2-submit /usr/bin/spark2-submit
ln -s /etc/spark2/conf.cloudera.spark2_on_yarn /etc/alternatives/spark2-conf

###hdfs
ln -s /opt/cloudera/parcels/CDH-5.15.1-1.cdh5.15.1.p0.4/bin/hdfs /etc/alternatives/hdfs
ln -s /etc/alternatives/hdfs /usr/bin/hdfs
ln -s /etc/hadoop/conf.cloudera.yarn /etc/alternatives/hadoop-conf

###sqoop
ln -s /opt/cloudera/parcels/CDH-5.15.1-1.cdh5.15.1.p0.4/bin/sqoop /etc/alternatives/sqoop
ln -s /etc/alternatives/sqoop /usr/bin/sqoop
ln -s /opt/cloudera/parcels/CDH-5.15.1-1.cdh5.15.1.p0.4/etc/sqoop/conf.dist /etc/alternatives/sqoop-conf&lt;/LI-CODE&gt;&lt;P&gt;After this, spark2-submit works as expected.&lt;/P&gt;</description>
      <pubDate>Thu, 18 Feb 2021 19:53:23 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/remote-pyspark-shell-and-spark-submit-error-java-lang/m-p/311760#M224814</guid>
      <dc:creator>adrijand</dc:creator>
      <dc:date>2021-02-18T19:53:23Z</dc:date>
    </item>
    <item>
      <title>Re: remote pyspark shell and spark-submit error java.lang.NoSuchFieldError: METASTORE_CLIENT_SOCKET_LIFETIME</title>
      <link>https://community.cloudera.com/t5/Support-Questions/remote-pyspark-shell-and-spark-submit-error-java-lang/m-p/311968#M224910</link>
      <description>&lt;P&gt;Thanks&amp;nbsp;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/84442"&gt;@adrijand&lt;/a&gt;&amp;nbsp; for sharing your updates,&amp;nbsp;it's highly appreciated.&lt;/P&gt;</description>
      <pubDate>Tue, 23 Feb 2021 09:44:33 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/remote-pyspark-shell-and-spark-submit-error-java-lang/m-p/311968#M224910</guid>
      <dc:creator>jagadeesan</dc:creator>
      <dc:date>2021-02-23T09:44:33Z</dc:date>
    </item>
  </channel>
</rss>

