<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Submit job spark standalone emit error kerberos when enabling event log to s3. in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Submit-job-spark-standalone-emit-error-kerberos-when/m-p/412750#M253669</link>
    <description>&lt;P&gt;Hi everyone,&lt;/P&gt;&lt;P&gt;I installed Spark cluster standalone. I want to config send log job to minio. I configured spark-default.conf file as bellow.&lt;/P&gt;&lt;DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;/DIV&gt;&lt;LI-CODE lang="markup"&gt;    spark.eventLog.enabled=false
    spark.eventLog.dir=s3a://spark-logs/
    spark.history.fs.logDirectory=s3a://spark-logs/
    spark.hadoop.fs.s3a.endpoint=http://192.168.182.131:9000
    spark.hadoop.fs.s3a.access.key=admin
    spark.hadoop.fs.s3a.secret.key=admin12345
    spark.hadoop.fs.s3a.path.style.access=true" &lt;/LI-CODE&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;&lt;SPAN&gt;When I submit job&amp;nbsp;spark-submit --master spark://spark-master-svc:7077 --conf spark.jars.ivy=/tmp/.ivy2 pi.py 1. I got the error:&lt;BR /&gt;&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/DIV&gt;&lt;LI-CODE lang="markup"&gt;5/10/26 12:13:29 INFO SparkContext: Successfully stopped SparkContext
Traceback (most recent call last):
  File "/opt/bitnami/spark/examples/src/main/python/pi.py", line 32, in &amp;lt;module&amp;gt;
    .getOrCreate()
     ^^^^^^^^^^^^^
  File "/opt/bitnami/spark/python/lib/pyspark.zip/pyspark/sql/session.py", line 497, in getOrCreate
  File "/opt/bitnami/spark/python/lib/pyspark.zip/pyspark/context.py", line 515, in getOrCreate
  File "/opt/bitnami/spark/python/lib/pyspark.zip/pyspark/context.py", line 203, in __init__
  File "/opt/bitnami/spark/python/lib/pyspark.zip/pyspark/context.py", line 296, in _do_init
  File "/opt/bitnami/spark/python/lib/pyspark.zip/pyspark/context.py", line 421, in _initialize_context
  File "/opt/bitnami/spark/python/lib/py4j-0.10.9.7-src.zip/py4j/java_gateway.py", line 1587, in __call__
  File "/opt/bitnami/spark/python/lib/py4j-0.10.9.7-src.zip/py4j/protocol.py", line 326, in get_return_value
py4j.protocol.Py4JJavaError: An error occurred while calling None.org.apache.spark.api.java.JavaSparkContext.
: org.apache.hadoop.security.KerberosAuthException: failure to login: javax.security.auth.login.LoginException: java.lang.NullPointerException: invalid null input: name
        at jdk.security.auth/com.sun.security.auth.UnixPrincipal.&amp;lt;init&amp;gt;(Unknown Source)
        at jdk.security.auth/com.sun.security.auth.module.UnixLoginModule.login(Unknown Source)
        at java.base/javax.security.auth.login.LoginContext.invoke(Unknown Source)
        at java.base/javax.security.auth.login.LoginContext$4.run(Unknown Source)
        at java.base/javax.security.auth.login.LoginContext$4.run(Unknown Source)
        at java.base/java.security.AccessController.doPrivileged(Unknown Source)
        at java.base/javax.security.auth.login.LoginContext.invokePriv(Unknown Source)
        at java.base/javax.security.auth.login.LoginContext.login(Unknown Source)
        at org.apache.hadoop.security.UserGroupInformation$HadoopLoginContext.login(UserGroupInformation.java:2065)
        at org.apache.hadoop.security.UserGroupInformation.doSubjectLogin(UserGroupInformation.java:1975)
        at org.apache.hadoop.security.UserGroupInformation.createLoginUser(UserGroupInformation.java:719)
        at org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:669)
        at org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:579)
        at org.apache.hadoop.fs.FileSystem$Cache$Key.&amp;lt;init&amp;gt;(FileSystem.java:3746)
        at org.apache.hadoop.fs.FileSystem$Cache$Key.&amp;lt;init&amp;gt;(FileSystem.java:3736)
        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:3520)
        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:540)
        at org.apache.spark.util.Utils$.getHadoopFileSystem(Utils.scala:1831)
        at org.apache.spark.deploy.history.EventLogFileWriter.&amp;lt;init&amp;gt;(EventLogFileWriters.scala:60)
        at org.apache.spark.deploy.history.SingleEventLogFileWriter.&amp;lt;init&amp;gt;(EventLogFileWriters.scala:213)
        at org.apache.spark.deploy.history.EventLogFileWriter$.apply(EventLogFileWriters.scala:181)
        at org.apache.spark.scheduler.EventLoggingListener.&amp;lt;init&amp;gt;(EventLoggingListener.scala:64)
        at org.apache.spark.SparkContext.&amp;lt;init&amp;gt;(SparkContext.scala:631)
        at org.apache.spark.api.java.JavaSparkContext.&amp;lt;init&amp;gt;(JavaSparkContext.scala:58)
        at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
        at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(Unknown Source)
        at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown Source)
        at java.base/java.lang.reflect.Constructor.newInstanceWithCaller(Unknown Source)
        at java.base/java.lang.reflect.Constructor.newInstance(Unknown Source)
        at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:247)
        at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:374)
        at py4j.Gateway.invoke(Gateway.java:238)
        at py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)
        at py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)
        at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)
        at py4j.ClientServerConnection.run(ClientServerConnection.java:106)
        at java.base/java.lang.Thread.run(Unknown Source)&lt;/LI-CODE&gt;</description>
    <pubDate>Sun, 26 Oct 2025 12:21:49 GMT</pubDate>
    <dc:creator>yoonli</dc:creator>
    <dc:date>2025-10-26T12:21:49Z</dc:date>
    <item>
      <title>Submit job spark standalone emit error kerberos when enabling event log to s3.</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Submit-job-spark-standalone-emit-error-kerberos-when/m-p/412750#M253669</link>
      <description>&lt;P&gt;Hi everyone,&lt;/P&gt;&lt;P&gt;I installed Spark cluster standalone. I want to config send log job to minio. I configured spark-default.conf file as bellow.&lt;/P&gt;&lt;DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;/DIV&gt;&lt;LI-CODE lang="markup"&gt;    spark.eventLog.enabled=false
    spark.eventLog.dir=s3a://spark-logs/
    spark.history.fs.logDirectory=s3a://spark-logs/
    spark.hadoop.fs.s3a.endpoint=http://192.168.182.131:9000
    spark.hadoop.fs.s3a.access.key=admin
    spark.hadoop.fs.s3a.secret.key=admin12345
    spark.hadoop.fs.s3a.path.style.access=true" &lt;/LI-CODE&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;&lt;SPAN&gt;When I submit job&amp;nbsp;spark-submit --master spark://spark-master-svc:7077 --conf spark.jars.ivy=/tmp/.ivy2 pi.py 1. I got the error:&lt;BR /&gt;&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/DIV&gt;&lt;LI-CODE lang="markup"&gt;5/10/26 12:13:29 INFO SparkContext: Successfully stopped SparkContext
Traceback (most recent call last):
  File "/opt/bitnami/spark/examples/src/main/python/pi.py", line 32, in &amp;lt;module&amp;gt;
    .getOrCreate()
     ^^^^^^^^^^^^^
  File "/opt/bitnami/spark/python/lib/pyspark.zip/pyspark/sql/session.py", line 497, in getOrCreate
  File "/opt/bitnami/spark/python/lib/pyspark.zip/pyspark/context.py", line 515, in getOrCreate
  File "/opt/bitnami/spark/python/lib/pyspark.zip/pyspark/context.py", line 203, in __init__
  File "/opt/bitnami/spark/python/lib/pyspark.zip/pyspark/context.py", line 296, in _do_init
  File "/opt/bitnami/spark/python/lib/pyspark.zip/pyspark/context.py", line 421, in _initialize_context
  File "/opt/bitnami/spark/python/lib/py4j-0.10.9.7-src.zip/py4j/java_gateway.py", line 1587, in __call__
  File "/opt/bitnami/spark/python/lib/py4j-0.10.9.7-src.zip/py4j/protocol.py", line 326, in get_return_value
py4j.protocol.Py4JJavaError: An error occurred while calling None.org.apache.spark.api.java.JavaSparkContext.
: org.apache.hadoop.security.KerberosAuthException: failure to login: javax.security.auth.login.LoginException: java.lang.NullPointerException: invalid null input: name
        at jdk.security.auth/com.sun.security.auth.UnixPrincipal.&amp;lt;init&amp;gt;(Unknown Source)
        at jdk.security.auth/com.sun.security.auth.module.UnixLoginModule.login(Unknown Source)
        at java.base/javax.security.auth.login.LoginContext.invoke(Unknown Source)
        at java.base/javax.security.auth.login.LoginContext$4.run(Unknown Source)
        at java.base/javax.security.auth.login.LoginContext$4.run(Unknown Source)
        at java.base/java.security.AccessController.doPrivileged(Unknown Source)
        at java.base/javax.security.auth.login.LoginContext.invokePriv(Unknown Source)
        at java.base/javax.security.auth.login.LoginContext.login(Unknown Source)
        at org.apache.hadoop.security.UserGroupInformation$HadoopLoginContext.login(UserGroupInformation.java:2065)
        at org.apache.hadoop.security.UserGroupInformation.doSubjectLogin(UserGroupInformation.java:1975)
        at org.apache.hadoop.security.UserGroupInformation.createLoginUser(UserGroupInformation.java:719)
        at org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:669)
        at org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:579)
        at org.apache.hadoop.fs.FileSystem$Cache$Key.&amp;lt;init&amp;gt;(FileSystem.java:3746)
        at org.apache.hadoop.fs.FileSystem$Cache$Key.&amp;lt;init&amp;gt;(FileSystem.java:3736)
        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:3520)
        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:540)
        at org.apache.spark.util.Utils$.getHadoopFileSystem(Utils.scala:1831)
        at org.apache.spark.deploy.history.EventLogFileWriter.&amp;lt;init&amp;gt;(EventLogFileWriters.scala:60)
        at org.apache.spark.deploy.history.SingleEventLogFileWriter.&amp;lt;init&amp;gt;(EventLogFileWriters.scala:213)
        at org.apache.spark.deploy.history.EventLogFileWriter$.apply(EventLogFileWriters.scala:181)
        at org.apache.spark.scheduler.EventLoggingListener.&amp;lt;init&amp;gt;(EventLoggingListener.scala:64)
        at org.apache.spark.SparkContext.&amp;lt;init&amp;gt;(SparkContext.scala:631)
        at org.apache.spark.api.java.JavaSparkContext.&amp;lt;init&amp;gt;(JavaSparkContext.scala:58)
        at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
        at java.base/jdk.internal.reflect.NativeConstructorAccessorImpl.newInstance(Unknown Source)
        at java.base/jdk.internal.reflect.DelegatingConstructorAccessorImpl.newInstance(Unknown Source)
        at java.base/java.lang.reflect.Constructor.newInstanceWithCaller(Unknown Source)
        at java.base/java.lang.reflect.Constructor.newInstance(Unknown Source)
        at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:247)
        at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:374)
        at py4j.Gateway.invoke(Gateway.java:238)
        at py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:80)
        at py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:69)
        at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:182)
        at py4j.ClientServerConnection.run(ClientServerConnection.java:106)
        at java.base/java.lang.Thread.run(Unknown Source)&lt;/LI-CODE&gt;</description>
      <pubDate>Sun, 26 Oct 2025 12:21:49 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Submit-job-spark-standalone-emit-error-kerberos-when/m-p/412750#M253669</guid>
      <dc:creator>yoonli</dc:creator>
      <dc:date>2025-10-26T12:21:49Z</dc:date>
    </item>
    <item>
      <title>Re: Submit job spark standalone emit error kerberos when enabling event log to s3.</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Submit-job-spark-standalone-emit-error-kerberos-when/m-p/412807#M253691</link>
      <description>&lt;P&gt;Hello&amp;nbsp;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/130962"&gt;@yoonli&lt;/a&gt;,&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks for contacting our Cloudera Community and sharing your question.&amp;nbsp;&lt;/P&gt;&lt;P&gt;Something I have to mention is that Cloudera does not support standalone Spark cluster, we only work with YARN clusters.&amp;nbsp;&lt;/P&gt;&lt;P&gt;Anyway, taking a quick look on this issue, I see that you're not mentioning any principal or keytab, when you use Kerberos, it will always try to run kinit, but if there are no principal and keytab, it will fail.&amp;nbsp;&lt;/P&gt;&lt;P&gt;Something you can try is using the Simple Auth method, you can add these two settings on your spark-defaults.conf:&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;spark.hadoop.fs.s3a.aws.credentials.provider=org.apache.hadoop.fs.s3a.SimpleAWSCredentialsProvider
spark.hadoop.security.authentication=simple
hadoop.security.authentication=simple&lt;/LI-CODE&gt;&lt;P&gt;&lt;A href="https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html" target="_blank"&gt;https://hadoop.apache.org/docs/stable/hadoop-aws/tools/hadoop-aws/index.html&lt;/A&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 03 Nov 2025 19:07:22 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Submit-job-spark-standalone-emit-error-kerberos-when/m-p/412807#M253691</guid>
      <dc:creator>vafs</dc:creator>
      <dc:date>2025-11-03T19:07:22Z</dc:date>
    </item>
    <item>
      <title>Re: Submit job spark standalone emit error kerberos when enabling event log to s3.</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Submit-job-spark-standalone-emit-error-kerberos-when/m-p/412953#M253798</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/130962"&gt;@yoonli&lt;/a&gt;&amp;nbsp;As per my knowledge Cloudera will not support Standalone setup and supported types are YARN and Kubernetes.&amp;nbsp;&lt;/P&gt;&lt;P&gt;For your error, you need to check kerberos setup for standalone cluster and try to run simple examples like SparkPi and see whether it is working or not.&lt;/P&gt;</description>
      <pubDate>Tue, 02 Dec 2025 02:23:57 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Submit-job-spark-standalone-emit-error-kerberos-when/m-p/412953#M253798</guid>
      <dc:creator>RangaReddy1988</dc:creator>
      <dc:date>2025-12-02T02:23:57Z</dc:date>
    </item>
  </channel>
</rss>

