<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Spark-HBase connector issue (Caused by: java.lang.NoClassDefFoundError: org/apache/hbase/thirdparty/com/google/protobuf/RpcController) in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Spark-HBase-connector-issue-Caused-by-java-lang/m-p/309856#M223944</link>
    <description>&lt;P data-unlink="true"&gt;I am new to Spark and wanted to read/write data to/from HBase tables. I&amp;nbsp;encountered an error while reading from HBase tables.&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Versions:&lt;/STRONG&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;Spark: 2.4.7; HBase: 1.4.13; Scala: 2.11.12&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Command:&amp;nbsp;&lt;/STRONG&gt;&lt;/P&gt;&lt;PRE&gt;spark-shell --jars /usr/lib/hbase/shc/core/target/shc-core&lt;SPAN class="hljs-number"&gt;-1.1&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.3&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;-2.4&lt;/SPAN&gt;-s_2&lt;SPAN class="hljs-number"&gt;.11&lt;/SPAN&gt;.jar,/usr/lib/hbase/lib/htrace-core4&lt;SPAN class="hljs-number"&gt;-4.1&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.0&lt;/SPAN&gt;-incubating.jar,/usr/lib/hbase/hbase-client&lt;SPAN class="hljs-number"&gt;-2.4&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.0&lt;/SPAN&gt;.jar,/usr/lib/hbase/hbase-common&lt;SPAN class="hljs-number"&gt;-2.4&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.0&lt;/SPAN&gt;.jar,/usr/lib/hbase/hbase-server&lt;SPAN class="hljs-number"&gt;-2.4&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.0&lt;/SPAN&gt;.jar,/usr/lib/hbase/hbase-protocol&lt;SPAN class="hljs-number"&gt;-2.4&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.0&lt;/SPAN&gt;.jar,/usr/lib/hbase/lib/htrace-core4&lt;SPAN class="hljs-number"&gt;-4.1&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.0&lt;/SPAN&gt;-incubating.jar,/usr/lib/hbase/hbase-shaded-miscellaneous&lt;SPAN class="hljs-number"&gt;-2.2&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.1&lt;/SPAN&gt;.jar,/usr/lib/hbase/hbase-protocol-shaded&lt;SPAN class="hljs-number"&gt;-2.4&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.0&lt;/SPAN&gt;.jar&lt;/PRE&gt;&lt;P&gt;&lt;STRONG&gt;&amp;nbsp;Error:&lt;/STRONG&gt;&lt;/P&gt;&lt;PRE&gt;java.io.&lt;SPAN class="hljs-type"&gt;IOException&lt;/SPAN&gt;: java.lang.reflect.&lt;SPAN class="hljs-type"&gt;UndeclaredThrowableException&lt;/SPAN&gt;
  at org.apache.hadoop.hbase.client.&lt;SPAN class="hljs-type"&gt;ConnectionFactory&lt;/SPAN&gt;.createConnection(&lt;SPAN class="hljs-type"&gt;ConnectionFactory&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;232&lt;/SPAN&gt;)
  at org.apache.hadoop.hbase.client.&lt;SPAN class="hljs-type"&gt;ConnectionFactory&lt;/SPAN&gt;.createConnection(&lt;SPAN class="hljs-type"&gt;ConnectionFactory&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;128&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;$$anonfun$getConnection$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;144&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;$$anonfun$getConnection$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;144&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;$$anonfun$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;135&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;$$anonfun$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;133&lt;/SPAN&gt;)
  at scala.collection.mutable.&lt;SPAN class="hljs-type"&gt;HashMap&lt;/SPAN&gt;.getOrElseUpdate(&lt;SPAN class="hljs-type"&gt;HashMap&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;79&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;$.getConnection(&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;133&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;$.getConnection(&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;144&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;RegionResource&lt;/SPAN&gt;.init(&lt;SPAN class="hljs-type"&gt;HBaseResources&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;96&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;ReferencedResource&lt;/SPAN&gt;$&lt;SPAN class="hljs-keyword"&gt;class&lt;/SPAN&gt;.liftedTree1$&lt;SPAN class="hljs-number"&gt;1&lt;/SPAN&gt;(&lt;SPAN class="hljs-type"&gt;HBaseResources&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;60&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;ReferencedResource&lt;/SPAN&gt;$&lt;SPAN class="hljs-keyword"&gt;class&lt;/SPAN&gt;.acquire(&lt;SPAN class="hljs-type"&gt;HBaseResources&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;57&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;RegionResource&lt;/SPAN&gt;.acquire(&lt;SPAN class="hljs-type"&gt;HBaseResources&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;91&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;ReferencedResource&lt;/SPAN&gt;$&lt;SPAN class="hljs-keyword"&gt;class&lt;/SPAN&gt;.releaseOnException(&lt;SPAN class="hljs-type"&gt;HBaseResources&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;77&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;RegionResource&lt;/SPAN&gt;.releaseOnException(&lt;SPAN class="hljs-type"&gt;HBaseResources&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;91&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;RegionResource&lt;/SPAN&gt;.&amp;lt;init&amp;gt;(&lt;SPAN class="hljs-type"&gt;HBaseResources&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;111&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;HBaseTableScanRDD&lt;/SPAN&gt;.getPartitions(&lt;SPAN class="hljs-type"&gt;HBaseTableScan&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;66&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;273&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at scala.&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.getOrElse(&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;121&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.partitions(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;MapPartitionsRDD&lt;/SPAN&gt;.getPartitions(&lt;SPAN class="hljs-type"&gt;MapPartitionsRDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;49&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;273&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at scala.&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.getOrElse(&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;121&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.partitions(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;MapPartitionsRDD&lt;/SPAN&gt;.getPartitions(&lt;SPAN class="hljs-type"&gt;MapPartitionsRDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;49&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;273&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at scala.&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.getOrElse(&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;121&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.partitions(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;MapPartitionsRDD&lt;/SPAN&gt;.getPartitions(&lt;SPAN class="hljs-type"&gt;MapPartitionsRDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;49&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;273&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at scala.&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.getOrElse(&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;121&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.partitions(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;MapPartitionsRDD&lt;/SPAN&gt;.getPartitions(&lt;SPAN class="hljs-type"&gt;MapPartitionsRDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;49&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;273&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at scala.&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.getOrElse(&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;121&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.partitions(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;SparkPlan&lt;/SPAN&gt;.executeTake(&lt;SPAN class="hljs-type"&gt;SparkPlan&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;384&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;CollectLimitExec&lt;/SPAN&gt;.executeCollect(limit.scala:&lt;SPAN class="hljs-number"&gt;38&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.org$apache$spark$sql$&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;$$collectFromPlan(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;3416&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;$$anonfun$head$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;2553&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;$$anonfun$head$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;2553&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;$$anonfun$&lt;SPAN class="hljs-number"&gt;52.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;3391&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;$.org$apache$spark$sql$execution$&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;$$executeQuery$&lt;SPAN class="hljs-number"&gt;1&lt;/SPAN&gt;(&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;83&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;$$anonfun$withNewExecutionId$&lt;SPAN class="hljs-number"&gt;1&lt;/SPAN&gt;$$anonfun$apply$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;94&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;QueryExecutionMetrics&lt;/SPAN&gt;$.withMetrics(&lt;SPAN class="hljs-type"&gt;QueryExecutionMetrics&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;141&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;$.org$apache$spark$sql$execution$&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;$$withMetrics(&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;178&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;$$anonfun$withNewExecutionId$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;93&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;$.withSQLConfPropagated(&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;200&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;$.withNewExecutionId(&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;92&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.org$apache$spark$sql$&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;$$withAction(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;3390&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.head(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;2553&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.take(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;2767&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.getRows(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;256&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.showString(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;293&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.show(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;754&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.show(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;713&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.show(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;722&lt;/SPAN&gt;)
  ... &lt;SPAN class="hljs-number"&gt;55&lt;/SPAN&gt; elided
&lt;SPAN class="hljs-type"&gt;Caused&lt;/SPAN&gt; by: java.lang.reflect.&lt;SPAN class="hljs-type"&gt;UndeclaredThrowableException&lt;/SPAN&gt;: java.lang.reflect.&lt;SPAN class="hljs-type"&gt;InvocationTargetException&lt;/SPAN&gt;: java.lang.&lt;SPAN class="hljs-type"&gt;NoClassDefFoundError&lt;/SPAN&gt;: org/apache/hbase/thirdparty/com/google/protobuf/&lt;SPAN class="hljs-type"&gt;RpcController&lt;/SPAN&gt;
  at org.apache.hadoop.security.&lt;SPAN class="hljs-type"&gt;UserGroupInformation&lt;/SPAN&gt;.doAs(&lt;SPAN class="hljs-type"&gt;UserGroupInformation&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;1944&lt;/SPAN&gt;)
  at org.apache.hadoop.hbase.security.&lt;SPAN class="hljs-type"&gt;User&lt;/SPAN&gt;$&lt;SPAN class="hljs-type"&gt;SecureHadoopUser&lt;/SPAN&gt;.runAs(&lt;SPAN class="hljs-type"&gt;User&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;347&lt;/SPAN&gt;)
  at org.apache.hadoop.hbase.client.&lt;SPAN class="hljs-type"&gt;ConnectionFactory&lt;/SPAN&gt;.createConnection(&lt;SPAN class="hljs-type"&gt;ConnectionFactory&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;228&lt;/SPAN&gt;)
  ... &lt;SPAN class="hljs-number"&gt;116&lt;/SPAN&gt; more
&lt;SPAN class="hljs-type"&gt;Caused&lt;/SPAN&gt; by: java.lang.reflect.&lt;SPAN class="hljs-type"&gt;InvocationTargetException&lt;/SPAN&gt;: java.lang.&lt;SPAN class="hljs-type"&gt;NoClassDefFoundError&lt;/SPAN&gt;: org/apache/hbase/thirdparty/com/google/protobuf/&lt;SPAN class="hljs-type"&gt;RpcController&lt;/SPAN&gt;
  at sun.reflect.&lt;SPAN class="hljs-type"&gt;NativeConstructorAccessorImpl&lt;/SPAN&gt;.newInstance0(&lt;SPAN class="hljs-type"&gt;Native&lt;/SPAN&gt; &lt;SPAN class="hljs-type"&gt;Method&lt;/SPAN&gt;)
  at sun.reflect.&lt;SPAN class="hljs-type"&gt;NativeConstructorAccessorImpl&lt;/SPAN&gt;.newInstance(&lt;SPAN class="hljs-type"&gt;NativeConstructorAccessorImpl&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;62&lt;/SPAN&gt;)
  at sun.reflect.&lt;SPAN class="hljs-type"&gt;DelegatingConstructorAccessorImpl&lt;/SPAN&gt;.newInstance(&lt;SPAN class="hljs-type"&gt;DelegatingConstructorAccessorImpl&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;45&lt;/SPAN&gt;)
  at java.lang.reflect.&lt;SPAN class="hljs-type"&gt;Constructor&lt;/SPAN&gt;.newInstance(&lt;SPAN class="hljs-type"&gt;Constructor&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;423&lt;/SPAN&gt;)
  at org.apache.hadoop.hbase.client.&lt;SPAN class="hljs-type"&gt;ConnectionFactory&lt;/SPAN&gt;.lambda$createConnection$&lt;SPAN class="hljs-number"&gt;0&lt;/SPAN&gt;(&lt;SPAN class="hljs-type"&gt;ConnectionFactory&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;230&lt;/SPAN&gt;)
  at java.security.&lt;SPAN class="hljs-type"&gt;AccessController&lt;/SPAN&gt;.doPrivileged(&lt;SPAN class="hljs-type"&gt;Native&lt;/SPAN&gt; &lt;SPAN class="hljs-type"&gt;Method&lt;/SPAN&gt;)
  at javax.security.auth.&lt;SPAN class="hljs-type"&gt;Subject&lt;/SPAN&gt;.doAs(&lt;SPAN class="hljs-type"&gt;Subject&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;422&lt;/SPAN&gt;)
  at org.apache.hadoop.security.&lt;SPAN class="hljs-type"&gt;UserGroupInformation&lt;/SPAN&gt;.doAs(&lt;SPAN class="hljs-type"&gt;UserGroupInformation&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;1926&lt;/SPAN&gt;)
  ... &lt;SPAN class="hljs-number"&gt;118&lt;/SPAN&gt; more
&lt;SPAN class="hljs-type"&gt;Caused&lt;/SPAN&gt; by: java.lang.&lt;SPAN class="hljs-type"&gt;NoClassDefFoundError&lt;/SPAN&gt;: org/apache/hbase/thirdparty/com/google/protobuf/&lt;SPAN class="hljs-type"&gt;RpcController&lt;/SPAN&gt;
  at java.lang.&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.defineClass1(&lt;SPAN class="hljs-type"&gt;Native&lt;/SPAN&gt; &lt;SPAN class="hljs-type"&gt;Method&lt;/SPAN&gt;)
  at java.lang.&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.defineClass(&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;756&lt;/SPAN&gt;)
  at java.security.&lt;SPAN class="hljs-type"&gt;SecureClassLoader&lt;/SPAN&gt;.defineClass(&lt;SPAN class="hljs-type"&gt;SecureClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;142&lt;/SPAN&gt;)
  at java.net.&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.defineClass(&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;468&lt;/SPAN&gt;)
  at java.net.&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.access$&lt;SPAN class="hljs-number"&gt;100&lt;/SPAN&gt;(&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;74&lt;/SPAN&gt;)
  at java.net.&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;run(&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;369&lt;/SPAN&gt;)
  at java.net.&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;run(&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;363&lt;/SPAN&gt;)
  at java.security.&lt;SPAN class="hljs-type"&gt;AccessController&lt;/SPAN&gt;.doPrivileged(&lt;SPAN class="hljs-type"&gt;Native&lt;/SPAN&gt; &lt;SPAN class="hljs-type"&gt;Method&lt;/SPAN&gt;)
  at java.net.&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.findClass(&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;362&lt;/SPAN&gt;)
  at java.lang.&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.loadClass(&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;418&lt;/SPAN&gt;)
  at java.lang.&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.loadClass(&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;351&lt;/SPAN&gt;)
  at org.apache.hadoop.hbase.client.&lt;SPAN class="hljs-type"&gt;ConnectionImplementation&lt;/SPAN&gt;.&amp;lt;init&amp;gt;(&lt;SPAN class="hljs-type"&gt;ConnectionImplementation&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;286&lt;/SPAN&gt;)
  ... &lt;SPAN class="hljs-number"&gt;126&lt;/SPAN&gt; more
&lt;SPAN class="hljs-type"&gt;Caused&lt;/SPAN&gt; by: java.lang.&lt;SPAN class="hljs-type"&gt;ClassNotFoundException&lt;/SPAN&gt;: org.apache.hbase.thirdparty.com.google.protobuf.&lt;SPAN class="hljs-type"&gt;RpcController&lt;/SPAN&gt;
  at java.net.&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.findClass(&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;382&lt;/SPAN&gt;)
  at java.lang.&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.loadClass(&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;418&lt;/SPAN&gt;)
  at java.lang.&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.loadClass(&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;351&lt;/SPAN&gt;)
  ... &lt;SPAN class="hljs-number"&gt;138&lt;/SPAN&gt; more&lt;/PRE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Please guide on how to resolve this and be able to read/write data to/from HBase using Spark.&lt;/P&gt;</description>
    <pubDate>Sun, 17 Jan 2021 12:51:12 GMT</pubDate>
    <dc:creator>Paarth</dc:creator>
    <dc:date>2021-01-17T12:51:12Z</dc:date>
    <item>
      <title>Spark-HBase connector issue (Caused by: java.lang.NoClassDefFoundError: org/apache/hbase/thirdparty/com/google/protobuf/RpcController)</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-HBase-connector-issue-Caused-by-java-lang/m-p/309856#M223944</link>
      <description>&lt;P data-unlink="true"&gt;I am new to Spark and wanted to read/write data to/from HBase tables. I&amp;nbsp;encountered an error while reading from HBase tables.&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Versions:&lt;/STRONG&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;Spark: 2.4.7; HBase: 1.4.13; Scala: 2.11.12&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Command:&amp;nbsp;&lt;/STRONG&gt;&lt;/P&gt;&lt;PRE&gt;spark-shell --jars /usr/lib/hbase/shc/core/target/shc-core&lt;SPAN class="hljs-number"&gt;-1.1&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.3&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;-2.4&lt;/SPAN&gt;-s_2&lt;SPAN class="hljs-number"&gt;.11&lt;/SPAN&gt;.jar,/usr/lib/hbase/lib/htrace-core4&lt;SPAN class="hljs-number"&gt;-4.1&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.0&lt;/SPAN&gt;-incubating.jar,/usr/lib/hbase/hbase-client&lt;SPAN class="hljs-number"&gt;-2.4&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.0&lt;/SPAN&gt;.jar,/usr/lib/hbase/hbase-common&lt;SPAN class="hljs-number"&gt;-2.4&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.0&lt;/SPAN&gt;.jar,/usr/lib/hbase/hbase-server&lt;SPAN class="hljs-number"&gt;-2.4&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.0&lt;/SPAN&gt;.jar,/usr/lib/hbase/hbase-protocol&lt;SPAN class="hljs-number"&gt;-2.4&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.0&lt;/SPAN&gt;.jar,/usr/lib/hbase/lib/htrace-core4&lt;SPAN class="hljs-number"&gt;-4.1&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.0&lt;/SPAN&gt;-incubating.jar,/usr/lib/hbase/hbase-shaded-miscellaneous&lt;SPAN class="hljs-number"&gt;-2.2&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.1&lt;/SPAN&gt;.jar,/usr/lib/hbase/hbase-protocol-shaded&lt;SPAN class="hljs-number"&gt;-2.4&lt;/SPAN&gt;&lt;SPAN class="hljs-number"&gt;.0&lt;/SPAN&gt;.jar&lt;/PRE&gt;&lt;P&gt;&lt;STRONG&gt;&amp;nbsp;Error:&lt;/STRONG&gt;&lt;/P&gt;&lt;PRE&gt;java.io.&lt;SPAN class="hljs-type"&gt;IOException&lt;/SPAN&gt;: java.lang.reflect.&lt;SPAN class="hljs-type"&gt;UndeclaredThrowableException&lt;/SPAN&gt;
  at org.apache.hadoop.hbase.client.&lt;SPAN class="hljs-type"&gt;ConnectionFactory&lt;/SPAN&gt;.createConnection(&lt;SPAN class="hljs-type"&gt;ConnectionFactory&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;232&lt;/SPAN&gt;)
  at org.apache.hadoop.hbase.client.&lt;SPAN class="hljs-type"&gt;ConnectionFactory&lt;/SPAN&gt;.createConnection(&lt;SPAN class="hljs-type"&gt;ConnectionFactory&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;128&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;$$anonfun$getConnection$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;144&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;$$anonfun$getConnection$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;144&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;$$anonfun$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;135&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;$$anonfun$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;133&lt;/SPAN&gt;)
  at scala.collection.mutable.&lt;SPAN class="hljs-type"&gt;HashMap&lt;/SPAN&gt;.getOrElseUpdate(&lt;SPAN class="hljs-type"&gt;HashMap&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;79&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;$.getConnection(&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;133&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;$.getConnection(&lt;SPAN class="hljs-type"&gt;HBaseConnectionCache&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;144&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;RegionResource&lt;/SPAN&gt;.init(&lt;SPAN class="hljs-type"&gt;HBaseResources&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;96&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;ReferencedResource&lt;/SPAN&gt;$&lt;SPAN class="hljs-keyword"&gt;class&lt;/SPAN&gt;.liftedTree1$&lt;SPAN class="hljs-number"&gt;1&lt;/SPAN&gt;(&lt;SPAN class="hljs-type"&gt;HBaseResources&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;60&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;ReferencedResource&lt;/SPAN&gt;$&lt;SPAN class="hljs-keyword"&gt;class&lt;/SPAN&gt;.acquire(&lt;SPAN class="hljs-type"&gt;HBaseResources&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;57&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;RegionResource&lt;/SPAN&gt;.acquire(&lt;SPAN class="hljs-type"&gt;HBaseResources&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;91&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;ReferencedResource&lt;/SPAN&gt;$&lt;SPAN class="hljs-keyword"&gt;class&lt;/SPAN&gt;.releaseOnException(&lt;SPAN class="hljs-type"&gt;HBaseResources&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;77&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;RegionResource&lt;/SPAN&gt;.releaseOnException(&lt;SPAN class="hljs-type"&gt;HBaseResources&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;91&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;RegionResource&lt;/SPAN&gt;.&amp;lt;init&amp;gt;(&lt;SPAN class="hljs-type"&gt;HBaseResources&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;111&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.datasources.hbase.&lt;SPAN class="hljs-type"&gt;HBaseTableScanRDD&lt;/SPAN&gt;.getPartitions(&lt;SPAN class="hljs-type"&gt;HBaseTableScan&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;66&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;273&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at scala.&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.getOrElse(&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;121&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.partitions(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;MapPartitionsRDD&lt;/SPAN&gt;.getPartitions(&lt;SPAN class="hljs-type"&gt;MapPartitionsRDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;49&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;273&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at scala.&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.getOrElse(&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;121&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.partitions(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;MapPartitionsRDD&lt;/SPAN&gt;.getPartitions(&lt;SPAN class="hljs-type"&gt;MapPartitionsRDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;49&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;273&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at scala.&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.getOrElse(&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;121&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.partitions(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;MapPartitionsRDD&lt;/SPAN&gt;.getPartitions(&lt;SPAN class="hljs-type"&gt;MapPartitionsRDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;49&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;273&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at scala.&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.getOrElse(&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;121&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.partitions(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;MapPartitionsRDD&lt;/SPAN&gt;.getPartitions(&lt;SPAN class="hljs-type"&gt;MapPartitionsRDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;49&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;273&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;$$anonfun$partitions$&lt;SPAN class="hljs-number"&gt;2.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at scala.&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.getOrElse(&lt;SPAN class="hljs-type"&gt;Option&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;121&lt;/SPAN&gt;)
  at org.apache.spark.rdd.&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.partitions(&lt;SPAN class="hljs-type"&gt;RDD&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;269&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;SparkPlan&lt;/SPAN&gt;.executeTake(&lt;SPAN class="hljs-type"&gt;SparkPlan&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;384&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;CollectLimitExec&lt;/SPAN&gt;.executeCollect(limit.scala:&lt;SPAN class="hljs-number"&gt;38&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.org$apache$spark$sql$&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;$$collectFromPlan(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;3416&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;$$anonfun$head$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;2553&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;$$anonfun$head$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;2553&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;$$anonfun$&lt;SPAN class="hljs-number"&gt;52.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;3391&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;$.org$apache$spark$sql$execution$&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;$$executeQuery$&lt;SPAN class="hljs-number"&gt;1&lt;/SPAN&gt;(&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;83&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;$$anonfun$withNewExecutionId$&lt;SPAN class="hljs-number"&gt;1&lt;/SPAN&gt;$$anonfun$apply$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;94&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;QueryExecutionMetrics&lt;/SPAN&gt;$.withMetrics(&lt;SPAN class="hljs-type"&gt;QueryExecutionMetrics&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;141&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;$.org$apache$spark$sql$execution$&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;$$withMetrics(&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;178&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;$$anonfun$withNewExecutionId$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;apply(&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;93&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;$.withSQLConfPropagated(&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;200&lt;/SPAN&gt;)
  at org.apache.spark.sql.execution.&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;$.withNewExecutionId(&lt;SPAN class="hljs-type"&gt;SQLExecution&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;92&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.org$apache$spark$sql$&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;$$withAction(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;3390&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.head(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;2553&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.take(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;2767&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.getRows(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;256&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.showString(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;293&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.show(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;754&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.show(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;713&lt;/SPAN&gt;)
  at org.apache.spark.sql.&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.show(&lt;SPAN class="hljs-type"&gt;Dataset&lt;/SPAN&gt;.scala:&lt;SPAN class="hljs-number"&gt;722&lt;/SPAN&gt;)
  ... &lt;SPAN class="hljs-number"&gt;55&lt;/SPAN&gt; elided
&lt;SPAN class="hljs-type"&gt;Caused&lt;/SPAN&gt; by: java.lang.reflect.&lt;SPAN class="hljs-type"&gt;UndeclaredThrowableException&lt;/SPAN&gt;: java.lang.reflect.&lt;SPAN class="hljs-type"&gt;InvocationTargetException&lt;/SPAN&gt;: java.lang.&lt;SPAN class="hljs-type"&gt;NoClassDefFoundError&lt;/SPAN&gt;: org/apache/hbase/thirdparty/com/google/protobuf/&lt;SPAN class="hljs-type"&gt;RpcController&lt;/SPAN&gt;
  at org.apache.hadoop.security.&lt;SPAN class="hljs-type"&gt;UserGroupInformation&lt;/SPAN&gt;.doAs(&lt;SPAN class="hljs-type"&gt;UserGroupInformation&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;1944&lt;/SPAN&gt;)
  at org.apache.hadoop.hbase.security.&lt;SPAN class="hljs-type"&gt;User&lt;/SPAN&gt;$&lt;SPAN class="hljs-type"&gt;SecureHadoopUser&lt;/SPAN&gt;.runAs(&lt;SPAN class="hljs-type"&gt;User&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;347&lt;/SPAN&gt;)
  at org.apache.hadoop.hbase.client.&lt;SPAN class="hljs-type"&gt;ConnectionFactory&lt;/SPAN&gt;.createConnection(&lt;SPAN class="hljs-type"&gt;ConnectionFactory&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;228&lt;/SPAN&gt;)
  ... &lt;SPAN class="hljs-number"&gt;116&lt;/SPAN&gt; more
&lt;SPAN class="hljs-type"&gt;Caused&lt;/SPAN&gt; by: java.lang.reflect.&lt;SPAN class="hljs-type"&gt;InvocationTargetException&lt;/SPAN&gt;: java.lang.&lt;SPAN class="hljs-type"&gt;NoClassDefFoundError&lt;/SPAN&gt;: org/apache/hbase/thirdparty/com/google/protobuf/&lt;SPAN class="hljs-type"&gt;RpcController&lt;/SPAN&gt;
  at sun.reflect.&lt;SPAN class="hljs-type"&gt;NativeConstructorAccessorImpl&lt;/SPAN&gt;.newInstance0(&lt;SPAN class="hljs-type"&gt;Native&lt;/SPAN&gt; &lt;SPAN class="hljs-type"&gt;Method&lt;/SPAN&gt;)
  at sun.reflect.&lt;SPAN class="hljs-type"&gt;NativeConstructorAccessorImpl&lt;/SPAN&gt;.newInstance(&lt;SPAN class="hljs-type"&gt;NativeConstructorAccessorImpl&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;62&lt;/SPAN&gt;)
  at sun.reflect.&lt;SPAN class="hljs-type"&gt;DelegatingConstructorAccessorImpl&lt;/SPAN&gt;.newInstance(&lt;SPAN class="hljs-type"&gt;DelegatingConstructorAccessorImpl&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;45&lt;/SPAN&gt;)
  at java.lang.reflect.&lt;SPAN class="hljs-type"&gt;Constructor&lt;/SPAN&gt;.newInstance(&lt;SPAN class="hljs-type"&gt;Constructor&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;423&lt;/SPAN&gt;)
  at org.apache.hadoop.hbase.client.&lt;SPAN class="hljs-type"&gt;ConnectionFactory&lt;/SPAN&gt;.lambda$createConnection$&lt;SPAN class="hljs-number"&gt;0&lt;/SPAN&gt;(&lt;SPAN class="hljs-type"&gt;ConnectionFactory&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;230&lt;/SPAN&gt;)
  at java.security.&lt;SPAN class="hljs-type"&gt;AccessController&lt;/SPAN&gt;.doPrivileged(&lt;SPAN class="hljs-type"&gt;Native&lt;/SPAN&gt; &lt;SPAN class="hljs-type"&gt;Method&lt;/SPAN&gt;)
  at javax.security.auth.&lt;SPAN class="hljs-type"&gt;Subject&lt;/SPAN&gt;.doAs(&lt;SPAN class="hljs-type"&gt;Subject&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;422&lt;/SPAN&gt;)
  at org.apache.hadoop.security.&lt;SPAN class="hljs-type"&gt;UserGroupInformation&lt;/SPAN&gt;.doAs(&lt;SPAN class="hljs-type"&gt;UserGroupInformation&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;1926&lt;/SPAN&gt;)
  ... &lt;SPAN class="hljs-number"&gt;118&lt;/SPAN&gt; more
&lt;SPAN class="hljs-type"&gt;Caused&lt;/SPAN&gt; by: java.lang.&lt;SPAN class="hljs-type"&gt;NoClassDefFoundError&lt;/SPAN&gt;: org/apache/hbase/thirdparty/com/google/protobuf/&lt;SPAN class="hljs-type"&gt;RpcController&lt;/SPAN&gt;
  at java.lang.&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.defineClass1(&lt;SPAN class="hljs-type"&gt;Native&lt;/SPAN&gt; &lt;SPAN class="hljs-type"&gt;Method&lt;/SPAN&gt;)
  at java.lang.&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.defineClass(&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;756&lt;/SPAN&gt;)
  at java.security.&lt;SPAN class="hljs-type"&gt;SecureClassLoader&lt;/SPAN&gt;.defineClass(&lt;SPAN class="hljs-type"&gt;SecureClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;142&lt;/SPAN&gt;)
  at java.net.&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.defineClass(&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;468&lt;/SPAN&gt;)
  at java.net.&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.access$&lt;SPAN class="hljs-number"&gt;100&lt;/SPAN&gt;(&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;74&lt;/SPAN&gt;)
  at java.net.&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;run(&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;369&lt;/SPAN&gt;)
  at java.net.&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;$&lt;SPAN class="hljs-number"&gt;1.&lt;/SPAN&gt;run(&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;363&lt;/SPAN&gt;)
  at java.security.&lt;SPAN class="hljs-type"&gt;AccessController&lt;/SPAN&gt;.doPrivileged(&lt;SPAN class="hljs-type"&gt;Native&lt;/SPAN&gt; &lt;SPAN class="hljs-type"&gt;Method&lt;/SPAN&gt;)
  at java.net.&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.findClass(&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;362&lt;/SPAN&gt;)
  at java.lang.&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.loadClass(&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;418&lt;/SPAN&gt;)
  at java.lang.&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.loadClass(&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;351&lt;/SPAN&gt;)
  at org.apache.hadoop.hbase.client.&lt;SPAN class="hljs-type"&gt;ConnectionImplementation&lt;/SPAN&gt;.&amp;lt;init&amp;gt;(&lt;SPAN class="hljs-type"&gt;ConnectionImplementation&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;286&lt;/SPAN&gt;)
  ... &lt;SPAN class="hljs-number"&gt;126&lt;/SPAN&gt; more
&lt;SPAN class="hljs-type"&gt;Caused&lt;/SPAN&gt; by: java.lang.&lt;SPAN class="hljs-type"&gt;ClassNotFoundException&lt;/SPAN&gt;: org.apache.hbase.thirdparty.com.google.protobuf.&lt;SPAN class="hljs-type"&gt;RpcController&lt;/SPAN&gt;
  at java.net.&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.findClass(&lt;SPAN class="hljs-type"&gt;URLClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;382&lt;/SPAN&gt;)
  at java.lang.&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.loadClass(&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;418&lt;/SPAN&gt;)
  at java.lang.&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.loadClass(&lt;SPAN class="hljs-type"&gt;ClassLoader&lt;/SPAN&gt;.java:&lt;SPAN class="hljs-number"&gt;351&lt;/SPAN&gt;)
  ... &lt;SPAN class="hljs-number"&gt;138&lt;/SPAN&gt; more&lt;/PRE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Please guide on how to resolve this and be able to read/write data to/from HBase using Spark.&lt;/P&gt;</description>
      <pubDate>Sun, 17 Jan 2021 12:51:12 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-HBase-connector-issue-Caused-by-java-lang/m-p/309856#M223944</guid>
      <dc:creator>Paarth</dc:creator>
      <dc:date>2021-01-17T12:51:12Z</dc:date>
    </item>
    <item>
      <title>Re: Spark-HBase connector issue (Caused by: java.lang.NoClassDefFoundError: org/apache/hbase/thirdparty/com/google/protobuf/RpcController)</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-HBase-connector-issue-Caused-by-java-lang/m-p/310105#M224029</link>
      <description>&lt;P&gt;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/84524"&gt;@Paarth&lt;/a&gt;&amp;nbsp;You might have to adjust jars. Check one similar issue here:&amp;nbsp;&lt;A href="https://community.cloudera.com/t5/Support-Questions/java-lang-ClassNotFoundException-org-apache-hbase-thirdparty/m-p/290876/highlight/true#M215172" target="_blank"&gt;https://community.cloudera.com/t5/Support-Questions/java-lang-ClassNotFoundException-org-apache-hbase-thirdparty/m-p/290876/highlight/true#M215172&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 20 Jan 2021 12:59:42 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-HBase-connector-issue-Caused-by-java-lang/m-p/310105#M224029</guid>
      <dc:creator>GangWar</dc:creator>
      <dc:date>2021-01-20T12:59:42Z</dc:date>
    </item>
    <item>
      <title>Re: Spark-HBase connector issue (Caused by: java.lang.NoClassDefFoundError: org/apache/hbase/thirdparty/com/google/protobuf/RpcController)</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-HBase-connector-issue-Caused-by-java-lang/m-p/370855#M240832</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/84524"&gt;@Paarth&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Spark HBase Connector (SHC) is not supported in CDP. You need to use HBase Spark Connector to access the HBase data using Spark.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;You can find the sample reference:&lt;/P&gt;&lt;P&gt;&lt;A href="https://docs.cloudera.com/runtime/7.2.10/managing-hbase/topics/hbase-using-hbase-spark-connector.html" target="_blank"&gt;https://docs.cloudera.com/runtime/7.2.10/managing-hbase/topics/hbase-using-hbase-spark-connector.html&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 16 May 2023 09:08:34 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-HBase-connector-issue-Caused-by-java-lang/m-p/370855#M240832</guid>
      <dc:creator>RangaReddy</dc:creator>
      <dc:date>2023-05-16T09:08:34Z</dc:date>
    </item>
  </channel>
</rss>

