<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Spark with HIVE JDBC connection in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233232#M195060</link>
    <description>&lt;P&gt;It won't work on transactional table.&lt;/P&gt;</description>
    <pubDate>Wed, 17 Jan 2018 03:56:55 GMT</pubDate>
    <dc:creator>thanhtu3009</dc:creator>
    <dc:date>2018-01-17T03:56:55Z</dc:date>
    <item>
      <title>Spark with HIVE JDBC connection</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233221#M195049</link>
      <description>&lt;P&gt;Hi All, &lt;/P&gt;&lt;P&gt;I have a sample table(stuends1) in HIVE which I want to connect from 
Spark using JDBC (as Hive is not in same cluster). I was just 
trying with following code...&lt;/P&gt;&lt;PRE&gt;def main(args: Array[String]): Unit = {
//Class.forName("org.apache.hive.jdbc.HiveDriver").newInstance()
val conf = new SparkConf().setAppName("SOME APP NAME").setMaster("local[*]")

val sc = new SparkContext(conf)

val spark = SparkSession
  .builder()
  .appName("Spark Hive Example")
  .getOrCreate()

val jdbcDF = spark.read
  .format("jdbc")
  .option("url", "jdbc:hive2://34.223.237.55:10000")
  .option("dbtable", "students1")
  .option("user", "hduser")
  .option("password", "hadoop")
  //.option("driver", "org.apache.hadoop.hive.jdbc.HiveDriver")
  .load()

println("able to connect------------------")

jdbcDF.show

jdbcDF.printSchema()

jdbcDF.createOrReplaceTempView("std")

val sqlDF = spark.sql("select * from std")
println("Start println-----")
spark.sqlContext.sql("select * from std").collect().foreach(println)
println("end println-----")
sqlDF.show(false)
}&lt;/PRE&gt;&lt;P&gt;I tried in multiple ways but all the time its showing table structure with column name only. Like ...&lt;/P&gt;&lt;PRE&gt;+--------------+-------------+-------------+
|students1.name|students1.age|students1.gpa|
+--------------+-------------+-------------+
+--------------+-------------+-------------+&lt;/PRE&gt;&lt;P&gt;But not data, but able to get data when trying to with dbeaver from 
my local with SQL query. From spark,  jdbcDF.printSchema() also showing 
proper schema , so I guess no issue with connection. &lt;/P&gt;&lt;P&gt;I am using spark 2.1.1 with HIVE 1.2.1. My sbt.build file is like this .... &lt;/P&gt;&lt;PRE&gt;libraryDependencies ++= Seq(

"log4j"  %   "log4j"  % "1.2.17",
"org.apache.spark" % "spark-core_2.11"                        % "2.1.1" ,
"org.apache.spark" % "spark-streaming-kafka_2.10"             % "1.6.2",
"org.apache.spark" % "spark-hivecontext-compatibility_2.10"   % "2.0.0-preview",
"org.apache.spark" % "spark-sql_2.11"                         % "2.1.1" ,
"org.apache.spark" % "spark-hive_2.10"                        % "2.1.1",
"org.apache.hive"  % "hive-jdbc"                              % "1.2.1"

}&lt;/PRE&gt;&lt;P&gt;can any one suggest why I am not getting any output of show(). Thanks in advance...&lt;/P&gt;</description>
      <pubDate>Thu, 24 Aug 2017 15:17:41 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233221#M195049</guid>
      <dc:creator>Biswajit16</dc:creator>
      <dc:date>2017-08-24T15:17:41Z</dc:date>
    </item>
    <item>
      <title>Re: Spark with HIVE JDBC connection</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233222#M195050</link>
      <description>&lt;P&gt;Spark connects to the Hive metastore directly via a HiveContext. It does not (nor should, in my opinion) use JDBC. &lt;/P&gt;&lt;P&gt;First, you must compile Spark with Hive support, then you need to explicitly call enableHiveSupport() on the SparkSession bulider. &lt;/P&gt;&lt;P&gt;Additionally, Spark2 will need you to provide either &lt;/P&gt;&lt;P&gt;1. A hive-site.xml file in the classpath&lt;/P&gt;&lt;P&gt;2. Setting hive.metastore.uris . Refer: &lt;A href="https://stackoverflow.com/questions/31980584/how-to-connect-to-a-hive-metastore-programmatically-in-sparksql" target="_blank"&gt;https://stackoverflow.com/questions/31980584/how-to-connect-to-a-hive-metastore-programmatically-in-sparksql&lt;/A&gt;&lt;/P&gt;&lt;P&gt;Additional resources&lt;/P&gt;&lt;P&gt;- &lt;A href="https://spark.apache.org/docs/latest/sql-programming-guide.html#hive-tables" target="_blank"&gt;https://spark.apache.org/docs/latest/sql-programming-guide.html#hive-tables&lt;/A&gt;&lt;/P&gt;&lt;P&gt;- &lt;A href="https://jaceklaskowski.gitbooks.io/mastering-apache-spark/spark-sql-hive-integration.html" target="_blank"&gt;https://jaceklaskowski.gitbooks.io/mastering-apache-spark/spark-sql-hive-integration.html&lt;/A&gt; &lt;/P&gt;</description>
      <pubDate>Fri, 25 Aug 2017 01:29:47 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233222#M195050</guid>
      <dc:creator>JordanMoore</dc:creator>
      <dc:date>2017-08-25T01:29:47Z</dc:date>
    </item>
    <item>
      <title>Re: Spark with HIVE JDBC connection</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233223#M195051</link>
      <description>&lt;P&gt;Thanks Jodan... you saved my day ....thanks a lot .... &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt; &lt;/P&gt;</description>
      <pubDate>Fri, 25 Aug 2017 19:23:25 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233223#M195051</guid>
      <dc:creator>Biswajit16</dc:creator>
      <dc:date>2017-08-25T19:23:25Z</dc:date>
    </item>
    <item>
      <title>Re: Spark with HIVE JDBC connection</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233224#M195052</link>
      <description>&lt;P&gt;What about reading external Hive Data by JDBC from Spark SQL?&lt;/P&gt;</description>
      <pubDate>Tue, 10 Oct 2017 12:30:38 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233224#M195052</guid>
      <dc:creator>casel_chen</dc:creator>
      <dc:date>2017-10-10T12:30:38Z</dc:date>
    </item>
    <item>
      <title>Re: Spark with HIVE JDBC connection</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233225#M195053</link>
      <description>&lt;P&gt;@CaselChen Again, Spark connects directly to the HiveMetastore - using JDBC requires you to go through HiveServer2&lt;/P&gt;</description>
      <pubDate>Wed, 11 Oct 2017 03:52:03 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233225#M195053</guid>
      <dc:creator>JordanMoore</dc:creator>
      <dc:date>2017-10-11T03:52:03Z</dc:date>
    </item>
    <item>
      <title>Re: Spark with HIVE JDBC connection</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233226#M195054</link>
      <description>&lt;P&gt;Hi Jordan, is there anyway by using JDBC through HiveServer2 with spark, we can load Hive data  for non-transactional as well as transactional table ? I have tried it but my tables were empty after loading. All table structures returned ( column names )  are fine though. Thanks.&lt;/P&gt;</description>
      <pubDate>Tue, 16 Jan 2018 00:25:30 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233226#M195054</guid>
      <dc:creator>thanhtu3009</dc:creator>
      <dc:date>2018-01-16T00:25:30Z</dc:date>
    </item>
    <item>
      <title>Re: Spark with HIVE JDBC connection</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233227#M195055</link>
      <description>&lt;A rel="user" href="https://community.cloudera.com/users/57151/thanhtu3009.html" nodeid="57151"&gt;@Tu Nguyen&lt;/A&gt;&lt;P&gt; - I'm afraid I don't understand your question. Spark does not use JDBC to communicate with Hive, but it can load Hive with any type of data that's able to be represented as a Spark DataSet. &lt;/P&gt;&lt;P&gt;You may want to try a "MSCK REPAIR TABLE &amp;lt;tablename&amp;gt;;" in Hive, though&lt;/P&gt;</description>
      <pubDate>Tue, 16 Jan 2018 00:54:35 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233227#M195055</guid>
      <dc:creator>JordanMoore</dc:creator>
      <dc:date>2018-01-16T00:54:35Z</dc:date>
    </item>
    <item>
      <title>Re: Spark with HIVE JDBC connection</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233228#M195056</link>
      <description>&lt;P&gt;Sorry Jordan, I was not clear.  As spark context is not supporting Hive Transactional tables. I am trying to use SparkSession to load a hive transactional table through JDBC as below &lt;/P&gt;&lt;P&gt;spSession.read.format("jdbc").option("url", url).option("driver", "org.apache.hive.jdbc.HiveDriver").option("dbtable", "tnguy.table_transactional_test").load().show()&lt;/P&gt;&lt;P&gt;The result is an empty table but I was expecting 3 rows of data. &lt;/P&gt;&lt;P&gt;I also tried it on non-transactional table but it gave the same result. I am wondering  if we can use hive JDBC to load a hive data ?  &lt;/P&gt;&lt;P&gt;Thanks,&lt;/P&gt;</description>
      <pubDate>Tue, 16 Jan 2018 02:06:09 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233228#M195056</guid>
      <dc:creator>thanhtu3009</dc:creator>
      <dc:date>2018-01-16T02:06:09Z</dc:date>
    </item>
    <item>
      <title>Re: Spark with HIVE JDBC connection</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233229#M195057</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/57151/thanhtu3009.html" nodeid="57151"&gt;@Tu Nguyen&lt;/A&gt; Where are you reading that you need to use JDBC from Spark to communicate with Hive? It isn't in the SparkSQL documentation. &lt;/P&gt;&lt;P&gt;&lt;A href="https://spark.apache.org/docs/latest/sql-programming-guide.html#hive-tables" target="_blank"&gt;https://spark.apache.org/docs/latest/sql-programming-guide.html#hive-tables&lt;/A&gt;&lt;/P&gt;&lt;P&gt;1. Try using an alternative JDBC client, see if you get similar results. &lt;/P&gt;&lt;P&gt;2. What happens when you simply use the following?&lt;/P&gt;&lt;PRE&gt;val spark = SparkSession
  .builder()
  .appName("Spark Transactional Hive Example")
  .config("spark.sql.warehouse.dir", warehouseLocation)
  .enableHiveSupport()
  .getOrCreate()

spark.table("tnguy.table_transactional_test").count()&lt;/PRE&gt;</description>
      <pubDate>Tue, 16 Jan 2018 10:29:19 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233229#M195057</guid>
      <dc:creator>JordanMoore</dc:creator>
      <dc:date>2018-01-16T10:29:19Z</dc:date>
    </item>
    <item>
      <title>Re: Spark with HIVE JDBC connection</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233230#M195058</link>
      <description>&lt;P&gt;I guess Tu Nguyen want to load external Hive table into Spark, right? If so, think about the following code &lt;/P&gt;&lt;PRE&gt;object SparkHiveJdbc extends App {
  val spark = SparkSession.builder.master("local[2]").appName("SparkHiveJob").getOrCreate
  val sc = spark.sparkContext
  val sqlContext = spark.sqlContext
  val driverName = "org.apache.hive.jdbc.HiveDriver"
  Class.forName(driverName)
  val df = spark.read
    .format("jdbc")
    .option("url", "jdbc:hive2://localhost:10000/default")
    .option("dbtable", "clicks_json")
    .load()
  df.printSchema()
  println(df.count())
  df.show()
}&lt;/PRE&gt;&lt;P&gt;I run the above code and met error reported as &lt;/P&gt;&lt;PRE&gt;root&lt;BR /&gt;|-- clicks_json.ip: string (nullable = true) 
|-- clicks_json.timestamp: long (nullable = true) 
|-- clicks_json.url: string (nullable = true) 
|-- clicks_json.referrer: string (nullable = true) 
|-- clicks_json.useragent: string (nullable = true) 
|-- clicks_json.sessionid: integer (nullable = true) 
|-- clicks_json.tenantid: string (nullable = true) 
|-- clicks_json.datestamp: string (nullable = true) 
&lt;BR /&gt;998&lt;BR /&gt;
Caused by: java.lang.NumberFormatException: For input string: "clicks_json.timestamp" at java.lang.NumberFormatException.forInputString(NumberFormatException.java:65) at java.lang.Long.parseLong(Long.java:589) at java.lang.Long.parseLong(Long.java:631) at org.apache.hive.jdbc.HiveBaseResultSet.getLong(HiveBaseResultSet.java:368) ... 23 more&amp;lt;br&amp;gt;&lt;/PRE&gt;&lt;P&gt;The reason of error I think is spark load header title row as first row when convert ResultSet into internal Row object. &lt;BR /&gt;Anything wrong here?&lt;/P&gt;</description>
      <pubDate>Tue, 16 Jan 2018 12:12:30 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233230#M195058</guid>
      <dc:creator>casel_chen</dc:creator>
      <dc:date>2018-01-16T12:12:30Z</dc:date>
    </item>
    <item>
      <title>Re: Spark with HIVE JDBC connection</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233231#M195059</link>
      <description>&lt;P&gt;I have tried it and it returned me an empty table with all corrected column names.&lt;/P&gt;</description>
      <pubDate>Wed, 17 Jan 2018 03:55:14 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233231#M195059</guid>
      <dc:creator>thanhtu3009</dc:creator>
      <dc:date>2018-01-17T03:55:14Z</dc:date>
    </item>
    <item>
      <title>Re: Spark with HIVE JDBC connection</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233232#M195060</link>
      <description>&lt;P&gt;It won't work on transactional table.&lt;/P&gt;</description>
      <pubDate>Wed, 17 Jan 2018 03:56:55 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233232#M195060</guid>
      <dc:creator>thanhtu3009</dc:creator>
      <dc:date>2018-01-17T03:56:55Z</dc:date>
    </item>
    <item>
      <title>Re: Spark with HIVE JDBC connection</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233233#M195061</link>
      <description>&lt;P&gt;Hi &lt;A rel="user" href="https://community.cloudera.com/users/29170/moorej.html" nodeid="29170"&gt;@Jordan Moore&lt;/A&gt; &lt;/P&gt;&lt;P&gt;SparkSession only work on non-transactional tables. With transactional table, it gave me error below :&lt;/P&gt;&lt;P&gt;scala&amp;gt; spark.table("user_tnguy11.table_orc_transactional_test").count()
org.apache.spark.sql.catalyst.errors.package$TreeNodeException: execute, tree:
Exchange SinglePartition
+- *HashAggregate(keys=[], functions=[partial_count(1)], output=[count#33L])
   +- HiveTableScan HiveTableRelation `user_tnguy11`.`table_orc_transactional_test`, org.apache.hadoop.hive.ql.io.orc.OrcSerde, [year#17, month#18, s#19]
  at org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:56)
  at org.apache.spark.sql.execution.exchange.ShuffleExchange.doExecute(ShuffleExchange.scala:115)
  at org.apache.spark.sql.execution.SparkPlan$anonfun$execute$1.apply(SparkPlan.scala:117)
  at org.apache.spark.sql.execution.SparkPlan$anonfun$execute$1.apply(SparkPlan.scala:117)
  at org.apache.spark.sql.execution.SparkPlan$anonfun$executeQuery$1.apply(SparkPlan.scala:138)
  at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
  at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:135)
  at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:116)
  at org.apache.spark.sql.execution.InputAdapter.inputRDDs(WholeStageCodegenExec.scala:252)
  at org.apache.spark.sql.execution.aggregate.HashAggregateExec.inputRDDs(HashAggregateExec.scala:141)
  at org.apache.spark.sql.execution.WholeStageCodegenExec.doExecute(WholeStageCodegenExec.scala:386)
  at org.apache.spark.sql.execution.SparkPlan$anonfun$execute$1.apply(SparkPlan.scala:117)
  at org.apache.spark.sql.execution.SparkPlan$anonfun$execute$1.apply(SparkPlan.scala:117)
  at org.apache.spark.sql.execution.SparkPlan$anonfun$executeQuery$1.apply(SparkPlan.scala:138)
  at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
  at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:135)
  at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:116)
  at org.apache.spark.sql.execution.SparkPlan.getByteArrayRdd(SparkPlan.scala:228)
  at org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:275)
  at org.apache.spark.sql.Dataset$anonfun$count$1.apply(Dataset.scala:2431)
  at org.apache.spark.sql.Dataset$anonfun$count$1.apply(Dataset.scala:2430)
  at org.apache.spark.sql.Dataset$anonfun$55.apply(Dataset.scala:2838)
  at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:65)
  at org.apache.spark.sql.Dataset.withAction(Dataset.scala:2837)
  at org.apache.spark.sql.Dataset.count(Dataset.scala:2430)
  ... 50 elided
Caused by: java.lang.RuntimeException: serious problem
  at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.generateSplitsInfo(OrcInputFormat.java:1021)
  at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.getSplits(OrcInputFormat.java:1048)
  at org.apache.spark.rdd.HadoopRDD.getPartitions(HadoopRDD.scala:199)
  at org.apache.spark.rdd.RDD$anonfun$partitions$2.apply(RDD.scala:252)
  at org.apache.spark.rdd.RDD$anonfun$partitions$2.apply(RDD.scala:250)
  at scala.Option.getOrElse(Option.scala:121)
  at org.apache.spark.rdd.RDD.partitions(RDD.scala:250)
  at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)
  at org.apache.spark.rdd.RDD$anonfun$partitions$2.apply(RDD.scala:252)
  at org.apache.spark.rdd.RDD$anonfun$partitions$2.apply(RDD.scala:250)
  at scala.Option.getOrElse(Option.scala:121)
  at org.apache.spark.rdd.RDD.partitions(RDD.scala:250)
  at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)
  at org.apache.spark.rdd.RDD$anonfun$partitions$2.apply(RDD.scala:252)
  at org.apache.spark.rdd.RDD$anonfun$partitions$2.apply(RDD.scala:250)
  at scala.Option.getOrElse(Option.scala:121)
  at org.apache.spark.rdd.RDD.partitions(RDD.scala:250)
  at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)
  at org.apache.spark.rdd.RDD$anonfun$partitions$2.apply(RDD.scala:252)
  at org.apache.spark.rdd.RDD$anonfun$partitions$2.apply(RDD.scala:250)
  at scala.Option.getOrElse(Option.scala:121)
  at org.apache.spark.rdd.RDD.partitions(RDD.scala:250)
  at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)
  at org.apache.spark.rdd.RDD$anonfun$partitions$2.apply(RDD.scala:252)
  at org.apache.spark.rdd.RDD$anonfun$partitions$2.apply(RDD.scala:250)
  at scala.Option.getOrElse(Option.scala:121)
  at org.apache.spark.rdd.RDD.partitions(RDD.scala:250)
  at org.apache.spark.rdd.MapPartitionsRDD.getPartitions(MapPartitionsRDD.scala:35)
  at org.apache.spark.rdd.RDD$anonfun$partitions$2.apply(RDD.scala:252)
  at org.apache.spark.rdd.RDD$anonfun$partitions$2.apply(RDD.scala:250)
  at scala.Option.getOrElse(Option.scala:121)
  at org.apache.spark.rdd.RDD.partitions(RDD.scala:250)
  at org.apache.spark.ShuffleDependency.&amp;lt;init&amp;gt;(Dependency.scala:91)
  at org.apache.spark.sql.execution.exchange.ShuffleExchange$.prepareShuffleDependency(ShuffleExchange.scala:264)
  at org.apache.spark.sql.execution.exchange.ShuffleExchange.prepareShuffleDependency(ShuffleExchange.scala:87)
  at org.apache.spark.sql.execution.exchange.ShuffleExchange$anonfun$doExecute$1.apply(ShuffleExchange.scala:124)
  at org.apache.spark.sql.execution.exchange.ShuffleExchange$anonfun$doExecute$1.apply(ShuffleExchange.scala:115)
  at org.apache.spark.sql.catalyst.errors.package$.attachTree(package.scala:52)
  ... 74 more
Caused by: java.util.concurrent.ExecutionException: java.lang.NumberFormatException: For input string: "0248155_0000"
  at java.util.concurrent.FutureTask.report(FutureTask.java:122)
  at java.util.concurrent.FutureTask.get(FutureTask.java:192)
  at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.generateSplitsInfo(OrcInputFormat.java:998)
  ... 111 more
Caused by: java.lang.NumberFormatException: For input string: "0248155_0000"
  at java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
  at java.lang.Long.parseLong(Long.java:589)
  at java.lang.Long.parseLong(Long.java:631)
  at org.apache.hadoop.hive.ql.io.AcidUtils.parseDelta(AcidUtils.java:310)
  at org.apache.hadoop.hive.ql.io.AcidUtils.getAcidState(AcidUtils.java:379)
  at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$FileGenerator.call(OrcInputFormat.java:634)
  at org.apache.hadoop.hive.ql.io.orc.OrcInputFormat$FileGenerator.call(OrcInputFormat.java:620)
  at java.util.concurrent.FutureTask.run(FutureTask.java:266)
  at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
  at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
  at java.lang.Thread.run(Thread.java:748)&lt;/P&gt;</description>
      <pubDate>Wed, 17 Jan 2018 05:30:15 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233233#M195061</guid>
      <dc:creator>thanhtu3009</dc:creator>
      <dc:date>2018-01-17T05:30:15Z</dc:date>
    </item>
    <item>
      <title>Re: Spark with HIVE JDBC connection</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233234#M195062</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/57151/thanhtu3009.html" nodeid="57151"&gt;@Tu Nguyen&lt;/A&gt; I suggest you post a new question, rather than hijack this one. &lt;/P&gt;&lt;P&gt;Your error does not relate directly to transactional tables, but rather the OrcSplits generated by your table. &lt;/P&gt;&lt;P&gt;How about if you should try to use &lt;STRONG&gt;spark.read.format("orc") &lt;/STRONG&gt;from the filesystem?&lt;/P&gt;&lt;PRE&gt;org.apache.hadoop.hive.ql.io.orc.OrcInputFormat.generateSplitsInfo(OrcInputFormat.java:998) ... 111 more 
Caused by: java.lang.NumberFormatException: For input string: "0248155_0000" 
    at java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)&lt;/PRE&gt;</description>
      <pubDate>Wed, 17 Jan 2018 11:19:27 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233234#M195062</guid>
      <dc:creator>JordanMoore</dc:creator>
      <dc:date>2018-01-17T11:19:27Z</dc:date>
    </item>
    <item>
      <title>Re: Spark with HIVE JDBC connection</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233235#M195063</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/29170/moorej.html" nodeid="29170"&gt;@Jordan Moore&lt;/A&gt; &lt;/P&gt;&lt;P&gt;Hi Jordan ,&lt;/P&gt;&lt;P&gt;Have to develop Stand alone  spark submit in Yarn mode using Kerbores cluster .I need to run Hive SQL using spark session.&lt;/P&gt;&lt;P&gt;Could you please help me on this.&lt;/P&gt;&lt;P&gt;Thanks in Advance!!!&lt;/P&gt;&lt;P&gt;Regrads,&lt;/P&gt;&lt;P&gt;Bipin&lt;/P&gt;</description>
      <pubDate>Tue, 23 Jan 2018 11:53:50 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233235#M195063</guid>
      <dc:creator>bipin_pradhan</dc:creator>
      <dc:date>2018-01-23T11:53:50Z</dc:date>
    </item>
    <item>
      <title>Re: Spark with HIVE JDBC connection</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233236#M195064</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/63184/bipinpradhan.html" nodeid="63184"&gt;@Bipin Pradhan&lt;/A&gt;, please post your question as a brand new post. &lt;/P&gt;</description>
      <pubDate>Wed, 24 Jan 2018 02:47:25 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-with-HIVE-JDBC-connection/m-p/233236#M195064</guid>
      <dc:creator>JordanMoore</dc:creator>
      <dc:date>2018-01-24T02:47:25Z</dc:date>
    </item>
  </channel>
</rss>

