<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Failed to initialize Atlas client using spark-atlas-connector in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Failed-to-initialize-Atlas-client-using-spark-atlas/m-p/241232#M203036</link>
    <description>&lt;P&gt;&lt;STRONG&gt;&lt;U&gt;COMMAND: &lt;/U&gt;&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;spark-shell --jars /home/atlas/tt/spark-atlas-connector-assembly_2.11-0.1.0-SNAPSHOT.jar --master local  --conf spark.extraListeners=com.hortonworks.spark.atlas.SparkAtlasEventTracker --conf spark.sql.queryExecutionListeners=com.hortonworks.spark.atlas.SparkAtlasEventTracker --conf spark.sql.streaming.streamingQueryListeners=com.hortonworks.spark.atlas.SparkAtlasStreamingQueryEventTracker&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;&lt;U&gt;&lt;STRONG&gt;ERROR:&lt;/STRONG&gt;&lt;/U&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;ERROR SparkAtlasEventTracker: Fail to initialize Atlas client, stop this listener&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;org.apache.atlas.AtlasServiceException: Metadata service API org.apache.atlas.AtlasClientV2$API_V2@398694a6 failed with status 401 (Unauthorized) Response Body ()&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;        at org.apache.atlas.AtlasBaseClient.callAPIWithResource(AtlasBaseClient.java:395)&lt;/P&gt;&lt;P&gt;        at org.apache.atlas.AtlasBaseClient.callAPIWithResource(AtlasBaseClient.java:323)&lt;/P&gt;&lt;P&gt;        at org.apache.atlas.AtlasBaseClient.callAPI(AtlasBaseClient.java:239)&lt;/P&gt;&lt;P&gt;        at org.apache.atlas.AtlasClientV2.getAllTypeDefs(AtlasClientV2.java:124)&lt;/P&gt;&lt;P&gt;        at com.hortonworks.spark.atlas.RestAtlasClient.getAtlasTypeDefs(RestAtlasClient.scala:58)&lt;/P&gt;&lt;P&gt;        at com.hortonworks.spark.atlas.types.SparkAtlasModel$$anonfun$checkAndGroupTypes$1.apply(SparkAtlasModel.scala:107)&lt;/P&gt;&lt;P&gt;        at com.hortonworks.spark.atlas.types.SparkAtlasModel$$anonfun$checkAndGroupTypes$1.apply(SparkAtlasModel.scala:104)&lt;/P&gt;&lt;P&gt;        at scala.collection.immutable.HashMap$HashMap1.foreach(HashMap.scala:221)&lt;/P&gt;&lt;P&gt;        at scala.collection.immutable.HashMap$HashTrieMap.foreach(HashMap.scala:428)&lt;/P&gt;&lt;P&gt;        at scala.collection.immutable.HashMap$HashTrieMap.foreach(HashMap.scala:428)&lt;/P&gt;&lt;P&gt;        at com.hortonworks.spark.atlas.types.SparkAtlasModel$.checkAndGroupTypes(SparkAtlasModel.scala:104)&lt;/P&gt;&lt;P&gt;        at com.hortonworks.spark.atlas.types.SparkAtlasModel$.checkAndCreateTypes(SparkAtlasModel.scala:71)&lt;/P&gt;&lt;P&gt;        at com.hortonworks.spark.atlas.SparkAtlasEventTracker.initializeSparkModel(SparkAtlasEventTracker.scala:108)&lt;/P&gt;&lt;P&gt;        at com.hortonworks.spark.atlas.SparkAtlasEventTracker.&amp;lt;init&amp;gt;(SparkAtlasEventTracker.scala:48)&lt;/P&gt;&lt;P&gt;        at com.hortonworks.spark.atlas.SparkAtlasEventTracker.&amp;lt;init&amp;gt;(SparkAtlasEventTracker.scala:39)&lt;/P&gt;&lt;P&gt;        at com.hortonworks.spark.atlas.SparkAtlasEventTracker.&amp;lt;init&amp;gt;(SparkAtlasEventTracker.scala:43)&lt;/P&gt;&lt;P&gt;        at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)&lt;/P&gt;&lt;P&gt;        at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)&lt;/P&gt;&lt;P&gt;        at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)&lt;/P&gt;&lt;P&gt;        at java.lang.reflect.Constructor.newInstance(Constructor.java:423)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.util.Utils$$anonfun$loadExtensions$1.apply(Utils.scala:2747)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.util.Utils$$anonfun$loadExtensions$1.apply(Utils.scala:2736)&lt;/P&gt;&lt;P&gt;        at scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241)&lt;/P&gt;&lt;P&gt;        at scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241)&lt;/P&gt;&lt;P&gt;        at scala.collection.mutable.ArraySeq.foreach(ArraySeq.scala:74)&lt;/P&gt;&lt;P&gt;        at scala.collection.TraversableLike$class.flatMap(TraversableLike.scala:241)&lt;/P&gt;&lt;P&gt;        at scala.collection.AbstractTraversable.flatMap(Traversable.scala:104)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.util.Utils$.loadExtensions(Utils.scala:2736)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.SparkContext$$anonfun$setupAndStartListenerBus$1.apply(SparkContext.scala:2360)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.SparkContext$$anonfun$setupAndStartListenerBus$1.apply(SparkContext.scala:2359)&lt;/P&gt;&lt;P&gt;        at scala.Option.foreach(Option.scala:257)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.SparkContext.setupAndStartListenerBus(SparkContext.scala:2359)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.SparkContext.&amp;lt;init&amp;gt;(SparkContext.scala:554)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2493)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:934)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:925)&lt;/P&gt;&lt;P&gt;        at scala.Option.getOrElse(Option.scala:121)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:925)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.repl.Main$.createSparkSession(Main.scala:103)&lt;/P&gt;&lt;P&gt;        at $line3.$read$$iw$$iw.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:15)&lt;/P&gt;&lt;P&gt;        at $line3.$read$$iw.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:43)&lt;/P&gt;&lt;P&gt;        at $line3.$read.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:45)&lt;/P&gt;&lt;P&gt;        at $line3.$read$.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:49)&lt;/P&gt;&lt;P&gt;        at $line3.$read$.&amp;lt;clinit&amp;gt;(&amp;lt;console&amp;gt;)&lt;/P&gt;&lt;P&gt;        at $line3.$eval$.$print$lzycompute(&amp;lt;console&amp;gt;:7)&lt;/P&gt;&lt;P&gt;        at $line3.$eval$.$print(&amp;lt;console&amp;gt;:6)&lt;/P&gt;&lt;P&gt;        at $line3.$eval.$print(&amp;lt;console&amp;gt;)&lt;/P&gt;&lt;P&gt;        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)&lt;/P&gt;&lt;P&gt;        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)&lt;/P&gt;&lt;P&gt;        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)&lt;/P&gt;&lt;P&gt;        at java.lang.reflect.Method.invoke(Method.java:498)&lt;/P&gt;&lt;P&gt;        at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.call(IMain.scala:786)&lt;/P&gt;&lt;P&gt;        at scala.tools.nsc.interpreter.IMain$Request.loadAndRun(IMain.scala:1047)&lt;/P&gt;&lt;P&gt;        at scala.tools.nsc.interpreter.IMain$WrappedRequest$$anonfun$loadAndRunReq$1.apply(IMain.scala:638)&lt;/P&gt;&lt;P&gt;        at scala.tools.nsc.interpreter.IMain$WrappedRequest$$anonfun$loadAndRunReq$1.apply(IMain.scala:637)&lt;/P&gt;&lt;P&gt;        at scala.reflect.internal.util.ScalaClassLoader$class.asContext(ScalaClassLoader.scala:31)&lt;/P&gt;&lt;P&gt;        at scala.reflect.internal.util.AbstractFileClassLoader.asContext(AbstractFileClassLoader.scala:19)&lt;/P&gt;&lt;P&gt;        at scala.tools.nsc.interpreter.IMain$WrappedRequest.loadAndRunReq(IMain.scala:637)&lt;/P&gt;&lt;P&gt;        at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:569)&lt;/P&gt;&lt;P&gt;        at scala.tools.nsc.interpreter.IMain.interpret(IMain.scala:565)&lt;/P&gt;&lt;P&gt;        at scala.tools.nsc.interpreter.ILoop.interpretStartingWith(ILoop.scala:807)&lt;/P&gt;&lt;P&gt;        at scala.tools.nsc.interpreter.ILoop.command(ILoop.scala:681)&lt;/P&gt;&lt;P&gt;        at scala.tools.nsc.interpreter.ILoop.processLine(ILoop.scala:395)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.repl.SparkILoop$$anonfun$initializeSpark$1$$anonfun$apply$mcV$sp$1$$anonfun$apply$mcV$sp$2.apply(SparkILoop.scala:79)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.repl.SparkILoop$$anonfun$initializeSpark$1$$anonfun$apply$mcV$sp$1$$anonfun$apply$mcV$sp$2.apply(SparkILoop.scala:79)&lt;/P&gt;&lt;P&gt;        at scala.collection.immutable.List.foreach(List.scala:381)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.repl.SparkILoop$$anonfun$initializeSpark$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(SparkILoop.scala:79)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.repl.SparkILoop$$anonfun$initializeSpark$1$$anonfun$apply$mcV$sp$1.apply(SparkILoop.scala:79)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.repl.SparkILoop$$anonfun$initializeSpark$1$$anonfun$apply$mcV$sp$1.apply(SparkILoop.scala:79)&lt;/P&gt;&lt;P&gt;        at scala.tools.nsc.interpreter.ILoop.savingReplayStack(ILoop.scala:91)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.repl.SparkILoop$$anonfun$initializeSpark$1.apply$mcV$sp(SparkILoop.scala:78)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.repl.SparkILoop$$anonfun$initializeSpark$1.apply(SparkILoop.scala:78)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.repl.SparkILoop$$anonfun$initializeSpark$1.apply(SparkILoop.scala:78)&lt;/P&gt;&lt;P&gt;        at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:214)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.repl.SparkILoop.initializeSpark(SparkILoop.scala:77)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.repl.SparkILoop.loadFiles(SparkILoop.scala:110)&lt;/P&gt;&lt;P&gt;        at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.apply$mcZ$sp(ILoop.scala:920)&lt;/P&gt;&lt;P&gt;        at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.apply(ILoop.scala:909)&lt;/P&gt;&lt;P&gt;        at scala.tools.nsc.interpreter.ILoop$$anonfun$process$1.apply(ILoop.scala:909)&lt;/P&gt;&lt;P&gt;        at scala.reflect.internal.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:97)&lt;/P&gt;&lt;P&gt;        at scala.tools.nsc.interpreter.ILoop.process(ILoop.scala:909)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.repl.Main$.doMain(Main.scala:76)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.repl.Main$.main(Main.scala:56)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.repl.Main.main(Main.scala)&lt;/P&gt;&lt;P&gt;        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)&lt;/P&gt;&lt;P&gt;        at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)&lt;/P&gt;&lt;P&gt;        at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)&lt;/P&gt;&lt;P&gt;        at java.lang.reflect.Method.invoke(Method.java:498)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:904)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:198)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:228)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:137)&lt;/P&gt;&lt;P&gt;        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;</description>
    <pubDate>Tue, 28 May 2019 22:39:12 GMT</pubDate>
    <dc:creator>srinucse9</dc:creator>
    <dc:date>2019-05-28T22:39:12Z</dc:date>
  </channel>
</rss>

