<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Job hang when Insert data into table in Spark Thrift Server in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Job-hang-when-Insert-data-into-table-in-Spark-Thrift-Server/m-p/395350#M248934</link>
    <description>&lt;P&gt;Hi everyone,&lt;/P&gt;&lt;P&gt;Thank you all for your responses. I am using Spark 3, and I’ve discovered that the issue is due to the improper configuration of the spark_shuffle settings in the yarn-site.xml file.&lt;/P&gt;&lt;P&gt;Thanks again!&lt;/P&gt;</description>
    <pubDate>Thu, 17 Oct 2024 01:28:17 GMT</pubDate>
    <dc:creator>allen_chu</dc:creator>
    <dc:date>2024-10-17T01:28:17Z</dc:date>
    <item>
      <title>Job hang when Insert data into table in Spark Thrift Server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Job-hang-when-Insert-data-into-table-in-Spark-Thrift-Server/m-p/395286#M248908</link>
      <description>&lt;P&gt;Hi Everyone,&lt;/P&gt;&lt;P&gt;I am facing a problem that I try to insert data into hiverserver2 by spark thrift server (actually I use beeline), the job of insert is stucked.&lt;/P&gt;&lt;P&gt;I have checked that spark MasterApplication UI page, and find that it shows as following figure.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="allen_chu_1-1729064115670.png" style="width: 400px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/42155iD14052526DA22654/image-size/medium?v=v2&amp;amp;px=400" role="button" title="allen_chu_1-1729064115670.png" alt="allen_chu_1-1729064115670.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;The log of spark thrift server is as following :&lt;/P&gt;&lt;P&gt;24/10/16 15:21:39 INFO SparkExecuteStatementOperation: Submitting query 'insert into test_database.test_table (a,b) values (2,33)' with a75190ac-d536-4ee1-a1ff-da42a195a40b&lt;BR /&gt;24/10/16 15:21:39 INFO SparkExecuteStatementOperation: Running query with a75190ac-d536-4ee1-a1ff-da42a195a40b&lt;BR /&gt;24/10/16 15:21:40 INFO FileUtils: Creating directory if it doesn't exist: hdfs://ha/warehouse/tablespace/managed/hive/test_database.db/test_table/.hive-staging_hive_2024-10-16_15-21-40_061_8849017887411502804-3&lt;BR /&gt;24/10/16 15:21:40 INFO FileOutputCommitter: File Output Committer Algorithm version is 1&lt;BR /&gt;24/10/16 15:21:40 INFO FileOutputCommitter: FileOutputCommitter skip cleanup _temporary folders under output directory:false, ignore cleanup failures: false&lt;BR /&gt;24/10/16 15:21:40 INFO SQLHadoopMapReduceCommitProtocol: Using output committer class org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter&lt;BR /&gt;24/10/16 15:21:40 INFO SparkContext: Starting job: run at AccessController.java:0&lt;BR /&gt;24/10/16 15:21:40 INFO DAGScheduler: Got job 2 (run at AccessController.java:0) with 1 output partitions&lt;BR /&gt;24/10/16 15:21:40 INFO DAGScheduler: Final stage: ResultStage 2 (run at AccessController.java:0)&lt;BR /&gt;24/10/16 15:21:40 INFO DAGScheduler: Parents of final stage: List()&lt;BR /&gt;24/10/16 15:21:40 INFO DAGScheduler: Missing parents: List()&lt;BR /&gt;24/10/16 15:21:40 INFO DAGScheduler: Submitting ResultStage 2 (MapPartitionsRDD[8] at run at AccessController.java:0), which has no missing parents&lt;BR /&gt;24/10/16 15:21:40 INFO MemoryStore: Block broadcast_2 stored as values in memory (estimated size 421.2 KiB, free 910.8 MiB)&lt;BR /&gt;24/10/16 15:21:40 INFO MemoryStore: Block broadcast_2_piece0 stored as bytes in memory (estimated size 147.0 KiB, free 910.6 MiB)&lt;BR /&gt;24/10/16 15:21:40 INFO BlockManagerInfo: Added broadcast_2_piece0 in memory on persp-6.persp.net:45131 (size: 147.0 KiB, free: 911.9 MiB)&lt;BR /&gt;24/10/16 15:21:40 INFO SparkContext: Created broadcast 2 from broadcast at DAGScheduler.scala:1535&lt;BR /&gt;24/10/16 15:21:40 INFO DAGScheduler: Submitting 1 missing tasks from ResultStage 2 (MapPartitionsRDD[8] at run at AccessController.java:0) (first 15 tasks are for partitions Vector(0))&lt;BR /&gt;24/10/16 15:21:40 INFO YarnScheduler: Adding task set 2.0 with 1 tasks resource profile 0&lt;BR /&gt;24/10/16 15:21:40 INFO FairSchedulableBuilder: Added task set TaskSet_2.0 tasks to pool default&lt;BR /&gt;24/10/16 15:21:50 WARN YarnScheduler: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient resources&lt;BR /&gt;24/10/16 15:22:05 WARN YarnScheduler: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient resources&lt;BR /&gt;24/10/16 15:22:20 WARN YarnScheduler: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient resources&lt;BR /&gt;24/10/16 15:22:35 WARN YarnScheduler: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient resources&lt;BR /&gt;24/10/16 15:22:50 WARN YarnScheduler: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient resources&lt;BR /&gt;24/10/16 15:23:05 WARN YarnScheduler: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient resources&lt;/P&gt;&lt;P&gt;Please help me to figure out what happens, thanks a lot.&lt;/P&gt;</description>
      <pubDate>Wed, 16 Oct 2024 07:42:43 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Job-hang-when-Insert-data-into-table-in-Spark-Thrift-Server/m-p/395286#M248908</guid>
      <dc:creator>allen_chu</dc:creator>
      <dc:date>2024-10-16T07:42:43Z</dc:date>
    </item>
    <item>
      <title>Re: Job hang when Insert data into table in Spark Thrift Server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Job-hang-when-Insert-data-into-table-in-Spark-Thrift-Server/m-p/395344#M248929</link>
      <description>&lt;P&gt;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/118899"&gt;@allen_chu&lt;/a&gt;&lt;/P&gt;&lt;P&gt;This looks like a Yarn Resource issue.&amp;nbsp; I would recommend opening a case in the Cloudera Support Portal under the Yarn Component to get further assistance with this.&lt;/P&gt;</description>
      <pubDate>Wed, 16 Oct 2024 20:25:19 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Job-hang-when-Insert-data-into-table-in-Spark-Thrift-Server/m-p/395344#M248929</guid>
      <dc:creator>MGreen</dc:creator>
      <dc:date>2024-10-16T20:25:19Z</dc:date>
    </item>
    <item>
      <title>Re: Job hang when Insert data into table in Spark Thrift Server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Job-hang-when-Insert-data-into-table-in-Spark-Thrift-Server/m-p/395345#M248930</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/118899"&gt;@allen_chu&lt;/a&gt;&amp;nbsp;, Let me know if my understanding is correct&amp;nbsp;&lt;BR /&gt;You are trying to insert data on Hive using&amp;nbsp;&lt;SPAN&gt;spark thrift server&amp;nbsp;and it is getting stuck.&lt;BR /&gt;However, when you insert the data using beeline you managed to insert the data.&lt;BR /&gt;Which CDP version you are using ?&amp;nbsp;&lt;BR /&gt;Do you see any Yarn application getting created ?&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 16 Oct 2024 20:29:34 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Job-hang-when-Insert-data-into-table-in-Spark-Thrift-Server/m-p/395345#M248930</guid>
      <dc:creator>caio_contente</dc:creator>
      <dc:date>2024-10-16T20:29:34Z</dc:date>
    </item>
    <item>
      <title>Re: Job hang when Insert data into table in Spark Thrift Server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Job-hang-when-Insert-data-into-table-in-Spark-Thrift-Server/m-p/395350#M248934</link>
      <description>&lt;P&gt;Hi everyone,&lt;/P&gt;&lt;P&gt;Thank you all for your responses. I am using Spark 3, and I’ve discovered that the issue is due to the improper configuration of the spark_shuffle settings in the yarn-site.xml file.&lt;/P&gt;&lt;P&gt;Thanks again!&lt;/P&gt;</description>
      <pubDate>Thu, 17 Oct 2024 01:28:17 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Job-hang-when-Insert-data-into-table-in-Spark-Thrift-Server/m-p/395350#M248934</guid>
      <dc:creator>allen_chu</dc:creator>
      <dc:date>2024-10-17T01:28:17Z</dc:date>
    </item>
  </channel>
</rss>

