<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Spark SQL fails on empty ORC table, HDP 2.4.2 in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123739#M34434</link>
    <description>&lt;P&gt;Thanks &lt;A rel="user" href="https://community.cloudera.com/users/1404/mahan.html" nodeid="1404"&gt;@Mahan Hosseinzadeh&lt;/A&gt;!&lt;/P&gt;&lt;P&gt;I had a similar issue and your solution worked for me. I was trying to read from a hive table using the hive context.&lt;/P&gt;&lt;P&gt;This was the error I got, in case it's useful for someone else.&lt;/P&gt;&lt;P&gt;scala&amp;gt; val outputHiveTableSchema = hiveContext.read.table("table_name").schema java.util.NoSuchElementException: next on empty iterator at scala.collection.Iterator$anon$2.next(Iterator.scala:39) at scala.collection.Iterator$anon$2.next(Iterator.scala:37) at scala.collection.IndexedSeqLike$Elements.next(IndexedSeqLike.scala:64) at scala.collection.IterableLike$class.head(IterableLike.scala:91) at scala.collection.mutable.ArrayOps$ofRef.scala$collection$IndexedSeqOptimized$super$head(ArrayOps.scala:108) at scala.collection.IndexedSeqOptimized$class.head(IndexedSeqOptimized.scala:120) at scala.collection.mutable.ArrayOps$ofRef.head(ArrayOps.scala:108) at org.apache.spark.sql.hive.orc.OrcRelation$anonfun$2.apply(OrcRelation.scala:186) at org.apache.spark.sql.hive.orc.OrcRelation$anonfun$2.apply(OrcRelation.scala:185) at scala.Option.getOrElse(Option.scala:120) at org.apache.spark.sql.hive.orc.OrcRelation.&amp;lt;init&amp;gt;(OrcRelation.scala:184) at org.apache.spark.sql.hive.HiveMetastoreCatalog$anonfun$19.apply(HiveMetastoreCatalog.scala:567) at org.apache.spark.sql.hive.HiveMetastoreCatalog$anonfun$19.apply(HiveMetastoreCatalog.scala:565) at scala.Option.getOrElse(Option.scala:120) at org.apache.spark.sql.hive.HiveMetastoreCatalog.org$apache$spark$sql$hive$HiveMetastoreCatalog$convertToOrcRelation(HiveMetastoreCatalog.scala:565) at org.apache.spark.sql.hive.HiveMetastoreCatalog$OrcConversions$anonfun$apply$2.applyOrElse(HiveMetastoreCatalog.scala:647) at org.apache.spark.sql.hive.HiveMetastoreCatalog$OrcConversions$anonfun$apply$2.applyOrElse(HiveMetastoreCatalog.scala:643) at org.apache.spark.sql.catalyst.trees.TreeNode$anonfun$transformUp$1.apply(TreeNode.scala:335) at org.apache.spark.sql.catalyst.trees.TreeNode$anonfun$transformUp$1.apply(TreeNode.scala:335) at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:69) at org.apache.spark.sql.catalyst.trees.TreeNode.transformUp(TreeNode.scala:334) at org.apache.spark.sql.hive.HiveMetastoreCatalog$OrcConversions$.apply(HiveMetastoreCatalog.scala:643) at org.apache.spark.sql.hive.HiveMetastoreCatalog$OrcConversions$.apply(HiveMetastoreCatalog.scala:637) at org.apache.spark.sql.catalyst.rules.RuleExecutor$anonfun$execute$1$anonfun$apply$1.apply(RuleExecutor.scala:83) at org.apache.spark.sql.catalyst.rules.RuleExecutor$anonfun$execute$1$anonfun$apply$1.apply(RuleExecutor.scala:80) at scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:111) at scala.collection.immutable.List.foldLeft(List.scala:84) at org.apache.spark.sql.catalyst.rules.RuleExecutor$anonfun$execute$1.apply(RuleExecutor.scala:80) at org.apache.spark.sql.catalyst.rules.RuleExecutor$anonfun$execute$1.apply(RuleExecutor.scala:72) at scala.collection.immutable.List.foreach(List.scala:318) at org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:72) at org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:36) at org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:36) at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:34) at org.apache.spark.sql.DataFrame.&amp;lt;init&amp;gt;(DataFrame.scala:133) at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:52) at org.apache.spark.sql.DataFrameReader.table(DataFrameReader.scala:337) at $iwC$iwC$iwC$iwC$iwC$iwC$iwC$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:29) at $iwC$iwC$iwC$iwC$iwC$iwC$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:34) at $iwC$iwC$iwC$iwC$iwC$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:36) at $iwC$iwC$iwC$iwC$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:38) at $iwC$iwC$iwC$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:40) at $iwC$iwC$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:42) at $iwC$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:44) at $iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:46) at &amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:48) at .&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:52) at .&amp;lt;clinit&amp;gt;(&amp;lt;console&amp;gt;) at .&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:7) at .&amp;lt;clinit&amp;gt;(&amp;lt;console&amp;gt;) at $print(&amp;lt;console&amp;gt;) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065) at org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1346) at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:840) at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871) at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819) at org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:857) at org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:902) at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814) at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:657) at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:665) at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$loop(SparkILoop.scala:670) at org.apache.spark.repl.SparkILoop$anonfun$org$apache$spark$repl$SparkILoop$process$1.apply$mcZ$sp(SparkILoop.scala:997) at org.apache.spark.repl.SparkILoop$anonfun$org$apache$spark$repl$SparkILoop$process$1.apply(SparkILoop.scala:945) at org.apache.spark.repl.SparkILoop$anonfun$org$apache$spark$repl$SparkILoop$process$1.apply(SparkILoop.scala:945) at scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135) at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$process(SparkILoop.scala:945) at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1059) at org.apache.spark.repl.Main$.main(Main.scala:31) at org.apache.spark.repl.Main.main(Main.scala) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at&lt;/P&gt;</description>
    <pubDate>Thu, 05 Jan 2017 03:08:53 GMT</pubDate>
    <dc:creator>vinay_muttineni</dc:creator>
    <dc:date>2017-01-05T03:08:53Z</dc:date>
    <item>
      <title>Spark SQL fails on empty ORC table, HDP 2.4.2</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123732#M34427</link>
      <description>&lt;P&gt;A simple scenario, create an empty orc table using Hive, then try to query the table using Spark&lt;/P&gt;&lt;PRE&gt;Hive:
create table tbl(name string) stored as orc;
Spark:
sqlContext.sql("select * from tbl") // even collect is not needed to see the error!&lt;/PRE&gt;&lt;P&gt;Here is the error:&lt;/P&gt;&lt;PRE&gt;16/07/11 15:09:21 INFO ParseDriver: Parsing command: select * from tbl
16/07/11 15:09:22 INFO ParseDriver: Parse Completed
java.lang.IllegalArgumentException: orcFileOperator: path hdfs://dobbindata/apps/hive/warehouse/tbl does not have valid orc files matching the pattern
	at org.apache.spark.sql.hive.orc.OrcFileOperator$.listOrcFiles(OrcFileOperator.scala:104)
	at org.apache.spark.sql.hive.orc.OrcFileOperator$.getFileReader(OrcFileOperator.scala:69)
	at org.apache.spark.sql.hive.orc.OrcFileOperator$.readSchema(OrcFileOperator.scala:77)
	at org.apache.spark.sql.hive.orc.OrcRelation$$anonfun$2.apply(OrcRelation.scala:185)
	at org.apache.spark.sql.hive.orc.OrcRelation$$anonfun$2.apply(OrcRelation.scala:185)
	at scala.Option.getOrElse(Option.scala:120)
	at org.apache.spark.sql.hive.orc.OrcRelation.&amp;lt;init&amp;gt;(OrcRelation.scala:184)
	at org.apache.spark.sql.hive.HiveMetastoreCatalog$$anonfun$20.apply(HiveMetastoreCatalog.scala:580)
	at org.apache.spark.sql.hive.HiveMetastoreCatalog$$anonfun$20.apply(HiveMetastoreCatalog.scala:578)
	at scala.Option.getOrElse(Option.scala:120)
	at org.apache.spark.sql.hive.HiveMetastoreCatalog.org$apache$spark$sql$hive$HiveMetastoreCatalog$$convertToOrcRelation(HiveMetastoreCatalog.scala:578)
	at org.apache.spark.sql.hive.HiveMetastoreCatalog$OrcConversions$$anonfun$apply$2.applyOrElse(HiveMetastoreCatalog.scala:647)
	at org.apache.spark.sql.hive.HiveMetastoreCatalog$OrcConversions$$anonfun$apply$2.applyOrElse(HiveMetastoreCatalog.scala:643)
	at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformUp$1.apply(TreeNode.scala:335)
	at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$transformUp$1.apply(TreeNode.scala:335)
	at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:69)
	at org.apache.spark.sql.catalyst.trees.TreeNode.transformUp(TreeNode.scala:334)
	at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$5.apply(TreeNode.scala:332)
	at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$5.apply(TreeNode.scala:332)
	at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:281)
	at scala.collection.Iterator$$anon$11.next(Iterator.scala:328)
	at scala.collection.Iterator$class.foreach(Iterator.scala:727)
	at scala.collection.AbstractIterator.foreach(Iterator.scala:1157)
	at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48)
	at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103)
	at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47)
	at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:273)
	at scala.collection.AbstractIterator.to(Iterator.scala:1157)
	at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265)
	at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157)
	at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252)
	at scala.collection.AbstractIterator.toArray(Iterator.scala:1157)
	at org.apache.spark.sql.catalyst.trees.TreeNode.transformChildren(TreeNode.scala:321)
	at org.apache.spark.sql.catalyst.trees.TreeNode.transformUp(TreeNode.scala:332)
	at org.apache.spark.sql.hive.HiveMetastoreCatalog$OrcConversions$.apply(HiveMetastoreCatalog.scala:643)
	at org.apache.spark.sql.hive.HiveMetastoreCatalog$OrcConversions$.apply(HiveMetastoreCatalog.scala:637)
	at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:83)
	at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:80)
	at scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:111)
	at scala.collection.immutable.List.foldLeft(List.scala:84)
	at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:80)
	at org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:72)
	at scala.collection.immutable.List.foreach(List.scala:318)
	at org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:72)
	at org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:36)
	at org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:36)
	at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:34)
	at org.apache.spark.sql.DataFrame.&amp;lt;init&amp;gt;(DataFrame.scala:133)
	at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:52)
	at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:817)
	at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:26)
	at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:31)
	at $iwC$$iwC$$iwC$$iwC$$iwC$$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:33)
	at $iwC$$iwC$$iwC$$iwC$$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:35)
	at $iwC$$iwC$$iwC$$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:37)
	at $iwC$$iwC$$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:39)
	at $iwC$$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:41)
	at $iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:43)
	at &amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:45)
	at .&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:49)
	at .&amp;lt;clinit&amp;gt;(&amp;lt;console&amp;gt;)
	at .&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:7)
	at .&amp;lt;clinit&amp;gt;(&amp;lt;console&amp;gt;)
	at $print(&amp;lt;console&amp;gt;)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065)
	at org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1346)
	at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:840)
	at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871)
	at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819)
	at org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:857)
	at org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:902)
	at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814)
	at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:657)
	at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:665)
	at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$loop(SparkILoop.scala:670)
	at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply$mcZ$sp(SparkILoop.scala:997)
	at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
	at org.apache.spark.repl.SparkILoop$$anonfun$org$apache$spark$repl$SparkILoop$$process$1.apply(SparkILoop.scala:945)
	at scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135)
	at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$$process(SparkILoop.scala:945)
	at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1059)
	at org.apache.spark.repl.Main$.main(Main.scala:31)
	at org.apache.spark.repl.Main.main(Main.scala)
	at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
	at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:498)
	at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:731)
	at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
	at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
	at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
	at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
&lt;/PRE&gt;</description>
      <pubDate>Tue, 12 Jul 2016 02:23:32 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123732#M34427</guid>
      <dc:creator>mahan</dc:creator>
      <dc:date>2016-07-12T02:23:32Z</dc:date>
    </item>
    <item>
      <title>Re: Spark SQL fails on empty ORC table, HDP 2.4.2</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123733#M34428</link>
      <description>&lt;P&gt;There is also an open jira ticket since March&lt;/P&gt;&lt;P&gt;&lt;A href="https://issues.apache.org/jira/browse/SPARK-14286" target="_blank"&gt;https://issues.apache.org/jira/browse/SPARK-14286&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 12 Jul 2016 02:25:46 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123733#M34428</guid>
      <dc:creator>mahan</dc:creator>
      <dc:date>2016-07-12T02:25:46Z</dc:date>
    </item>
    <item>
      <title>Re: Spark SQL fails on empty ORC table, HDP 2.4.2</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123734#M34429</link>
      <description>&lt;P&gt;I think this bug fix is also related to the issue&lt;/P&gt;&lt;P&gt;&lt;A href="https://issues.apache.org/jira/browse/SPARK-8501" target="_blank"&gt;https://issues.apache.org/jira/browse/SPARK-8501&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 12 Jul 2016 03:02:24 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123734#M34429</guid>
      <dc:creator>mahan</dc:creator>
      <dc:date>2016-07-12T03:02:24Z</dc:date>
    </item>
    <item>
      <title>Re: Spark SQL fails on empty ORC table, HDP 2.4.2</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123735#M34430</link>
      <description>&lt;P&gt;Is there any workaround/Patch available  for the same in HDP 2.4.2 .. Becasue on HDP 2.3.6 said scenario is working fine on Spark Version 1.5.2.. below is sample:&lt;/P&gt;&lt;P&gt;[root@****-1316 ~]# spark-sql
SET hive.support.sql11.reserved.keywords=false
SET spark.sql.hive.version=1.2.1
SET spark.sql.hive.version=1.2.1
spark-sql&amp;gt; create table customer1(id int ,name string, email string) clustered by (id) into 2 buckets stored as orc ;
OK
Time taken: 5.225 seconds
spark-sql&amp;gt; select * from customer1;
Time taken: 0.402 seconds
spark-sql&amp;gt;&lt;/P&gt;</description>
      <pubDate>Wed, 13 Jul 2016 15:54:24 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123735#M34430</guid>
      <dc:creator>suniln_kumar</dc:creator>
      <dc:date>2016-07-13T15:54:24Z</dc:date>
    </item>
    <item>
      <title>Re: Spark SQL fails on empty ORC table, HDP 2.4.2</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123736#M34431</link>
      <description>&lt;P&gt;Now I know what is going on, there is a ticket &lt;STRONG&gt;in progress&lt;/STRONG&gt; at moment called "Enable OrcRelation even when connecting via spark thrift server", &lt;A href="https://issues.apache.org/jira/browse/SPARK-12998" target="_blank"&gt;https://issues.apache.org/jira/browse/SPARK-12998&lt;/A&gt;&lt;/P&gt;&lt;P&gt;But you can find it as a &lt;STRONG&gt;release improvement&lt;/STRONG&gt; on HDP 2.4.2, &lt;A href="https://github.com/hortonworks/spark-release/blob/HDP-2.4.2.0-tag/HDP-CHANGES.txt" target="_blank"&gt;https://github.com/hortonworks/spark-release/blob/HDP-2.4.2.0-tag/HDP-CHANGES.txt&lt;/A&gt;&lt;/P&gt;&lt;P&gt;To conclude, Spark in Hortonworks could be slightly different from main Spark!&lt;/P&gt;&lt;P&gt;Now this is a workaround to disable the improvement and directly read the schema from Hive metastore:&lt;/P&gt;&lt;PRE&gt;sqlContext.setConf("spark.sql.hive.convertMetastoreOrc", "false")&lt;/PRE&gt;</description>
      <pubDate>Wed, 13 Jul 2016 16:39:11 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123736#M34431</guid>
      <dc:creator>mahan</dc:creator>
      <dc:date>2016-07-13T16:39:11Z</dc:date>
    </item>
    <item>
      <title>Re: Spark SQL fails on empty ORC table, HDP 2.4.2</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123737#M34432</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/1404/mahan.html" nodeid="1404"&gt;@Mahan Hosseinzadeh&lt;/A&gt;   Which configuration file should I set &lt;/P&gt;&lt;P&gt;"spark.sql.hive.convertMetastoreOrc","false"  ?&lt;/P&gt;&lt;P&gt;Not work in hive-site.xml&lt;/P&gt;</description>
      <pubDate>Tue, 27 Dec 2016 18:07:03 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123737#M34432</guid>
      <dc:creator>diablo2</dc:creator>
      <dc:date>2016-12-27T18:07:03Z</dc:date>
    </item>
    <item>
      <title>Re: Spark SQL fails on empty ORC table, HDP 2.4.2</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123738#M34433</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/13286/huangdengke.html" nodeid="13286"&gt;@Huahua Wei&lt;/A&gt; this is a spark specific configuration is not in hive-site.xml! set it in your application or from Ambari&lt;/P&gt;</description>
      <pubDate>Sat, 31 Dec 2016 01:28:59 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123738#M34433</guid>
      <dc:creator>mahan</dc:creator>
      <dc:date>2016-12-31T01:28:59Z</dc:date>
    </item>
    <item>
      <title>Re: Spark SQL fails on empty ORC table, HDP 2.4.2</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123739#M34434</link>
      <description>&lt;P&gt;Thanks &lt;A rel="user" href="https://community.cloudera.com/users/1404/mahan.html" nodeid="1404"&gt;@Mahan Hosseinzadeh&lt;/A&gt;!&lt;/P&gt;&lt;P&gt;I had a similar issue and your solution worked for me. I was trying to read from a hive table using the hive context.&lt;/P&gt;&lt;P&gt;This was the error I got, in case it's useful for someone else.&lt;/P&gt;&lt;P&gt;scala&amp;gt; val outputHiveTableSchema = hiveContext.read.table("table_name").schema java.util.NoSuchElementException: next on empty iterator at scala.collection.Iterator$anon$2.next(Iterator.scala:39) at scala.collection.Iterator$anon$2.next(Iterator.scala:37) at scala.collection.IndexedSeqLike$Elements.next(IndexedSeqLike.scala:64) at scala.collection.IterableLike$class.head(IterableLike.scala:91) at scala.collection.mutable.ArrayOps$ofRef.scala$collection$IndexedSeqOptimized$super$head(ArrayOps.scala:108) at scala.collection.IndexedSeqOptimized$class.head(IndexedSeqOptimized.scala:120) at scala.collection.mutable.ArrayOps$ofRef.head(ArrayOps.scala:108) at org.apache.spark.sql.hive.orc.OrcRelation$anonfun$2.apply(OrcRelation.scala:186) at org.apache.spark.sql.hive.orc.OrcRelation$anonfun$2.apply(OrcRelation.scala:185) at scala.Option.getOrElse(Option.scala:120) at org.apache.spark.sql.hive.orc.OrcRelation.&amp;lt;init&amp;gt;(OrcRelation.scala:184) at org.apache.spark.sql.hive.HiveMetastoreCatalog$anonfun$19.apply(HiveMetastoreCatalog.scala:567) at org.apache.spark.sql.hive.HiveMetastoreCatalog$anonfun$19.apply(HiveMetastoreCatalog.scala:565) at scala.Option.getOrElse(Option.scala:120) at org.apache.spark.sql.hive.HiveMetastoreCatalog.org$apache$spark$sql$hive$HiveMetastoreCatalog$convertToOrcRelation(HiveMetastoreCatalog.scala:565) at org.apache.spark.sql.hive.HiveMetastoreCatalog$OrcConversions$anonfun$apply$2.applyOrElse(HiveMetastoreCatalog.scala:647) at org.apache.spark.sql.hive.HiveMetastoreCatalog$OrcConversions$anonfun$apply$2.applyOrElse(HiveMetastoreCatalog.scala:643) at org.apache.spark.sql.catalyst.trees.TreeNode$anonfun$transformUp$1.apply(TreeNode.scala:335) at org.apache.spark.sql.catalyst.trees.TreeNode$anonfun$transformUp$1.apply(TreeNode.scala:335) at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:69) at org.apache.spark.sql.catalyst.trees.TreeNode.transformUp(TreeNode.scala:334) at org.apache.spark.sql.hive.HiveMetastoreCatalog$OrcConversions$.apply(HiveMetastoreCatalog.scala:643) at org.apache.spark.sql.hive.HiveMetastoreCatalog$OrcConversions$.apply(HiveMetastoreCatalog.scala:637) at org.apache.spark.sql.catalyst.rules.RuleExecutor$anonfun$execute$1$anonfun$apply$1.apply(RuleExecutor.scala:83) at org.apache.spark.sql.catalyst.rules.RuleExecutor$anonfun$execute$1$anonfun$apply$1.apply(RuleExecutor.scala:80) at scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:111) at scala.collection.immutable.List.foldLeft(List.scala:84) at org.apache.spark.sql.catalyst.rules.RuleExecutor$anonfun$execute$1.apply(RuleExecutor.scala:80) at org.apache.spark.sql.catalyst.rules.RuleExecutor$anonfun$execute$1.apply(RuleExecutor.scala:72) at scala.collection.immutable.List.foreach(List.scala:318) at org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:72) at org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:36) at org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:36) at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:34) at org.apache.spark.sql.DataFrame.&amp;lt;init&amp;gt;(DataFrame.scala:133) at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:52) at org.apache.spark.sql.DataFrameReader.table(DataFrameReader.scala:337) at $iwC$iwC$iwC$iwC$iwC$iwC$iwC$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:29) at $iwC$iwC$iwC$iwC$iwC$iwC$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:34) at $iwC$iwC$iwC$iwC$iwC$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:36) at $iwC$iwC$iwC$iwC$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:38) at $iwC$iwC$iwC$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:40) at $iwC$iwC$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:42) at $iwC$iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:44) at $iwC.&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:46) at &amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:48) at .&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:52) at .&amp;lt;clinit&amp;gt;(&amp;lt;console&amp;gt;) at .&amp;lt;init&amp;gt;(&amp;lt;console&amp;gt;:7) at .&amp;lt;clinit&amp;gt;(&amp;lt;console&amp;gt;) at $print(&amp;lt;console&amp;gt;) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:1065) at org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1346) at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:840) at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:871) at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:819) at org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:857) at org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:902) at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:814) at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:657) at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:665) at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$loop(SparkILoop.scala:670) at org.apache.spark.repl.SparkILoop$anonfun$org$apache$spark$repl$SparkILoop$process$1.apply$mcZ$sp(SparkILoop.scala:997) at org.apache.spark.repl.SparkILoop$anonfun$org$apache$spark$repl$SparkILoop$process$1.apply(SparkILoop.scala:945) at org.apache.spark.repl.SparkILoop$anonfun$org$apache$spark$repl$SparkILoop$process$1.apply(SparkILoop.scala:945) at scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135) at org.apache.spark.repl.SparkILoop.org$apache$spark$repl$SparkILoop$process(SparkILoop.scala:945) at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:1059) at org.apache.spark.repl.Main$.main(Main.scala:31) at org.apache.spark.repl.Main.main(Main.scala) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at&lt;/P&gt;</description>
      <pubDate>Thu, 05 Jan 2017 03:08:53 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123739#M34434</guid>
      <dc:creator>vinay_muttineni</dc:creator>
      <dc:date>2017-01-05T03:08:53Z</dc:date>
    </item>
    <item>
      <title>Re: Spark SQL fails on empty ORC table, HDP 2.4.2</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123740#M34435</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/1404/mahan.html" nodeid="1404" target="_blank"&gt;@Mahan Hosseinzadeh&lt;/A&gt;   in Ambari, which section should I set &lt;/P&gt;&lt;P&gt;spark.sql.hive.convertMetastoreOrc ?&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="11173-9zzr4.png" style="width: 238px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/22743i42FE84B64858BE16/image-size/medium?v=v2&amp;amp;px=400" role="button" title="11173-9zzr4.png" alt="11173-9zzr4.png" /&gt;&lt;/span&gt;&lt;/P&gt; or  &lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="11174-o7w3o.png" style="width: 266px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/22744i1E7BFDB1683DDCB2/image-size/medium?v=v2&amp;amp;px=400" role="button" title="11174-o7w3o.png" alt="11174-o7w3o.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;or elsewhere? &lt;/P&gt;</description>
      <pubDate>Mon, 19 Aug 2019 10:17:21 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123740#M34435</guid>
      <dc:creator>diablo2</dc:creator>
      <dc:date>2019-08-19T10:17:21Z</dc:date>
    </item>
    <item>
      <title>Re: Spark SQL fails on empty ORC table, HDP 2.4.2</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123741#M34436</link>
      <description>&lt;P&gt;In ambari its Custom spark-defaults.&lt;/P&gt;&lt;P&gt;Value : 
spark.sql.hive.convertMetastoreOrc=false&lt;/P&gt;</description>
      <pubDate>Tue, 28 Feb 2017 16:21:11 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-SQL-fails-on-empty-ORC-table-HDP-2-4-2/m-p/123741#M34436</guid>
      <dc:creator>rakesh_mib_j</dc:creator>
      <dc:date>2017-02-28T16:21:11Z</dc:date>
    </item>
  </channel>
</rss>

