<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Transfer files from hdfs to kudu in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Transfer-files-from-hdfs-to-kudu/m-p/293593#M216764</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/72966"&gt;@drgenious&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Are you getting a similar error which reported in&amp;nbsp;&lt;A href="https://issues.apache.org/jira/browse/KUDU-2633" target="_blank" rel="noopener"&gt;KUDU-2633&lt;/A&gt; It seems this is open JIRA reported in the community&lt;/P&gt;&lt;PRE&gt;ERROR core.JobRunShell: Job DEFAULT.EventKpisConsumer threw an unhandled Exception: 
org.apache.spark.SparkException: Job aborted due to stage failure: Aborting TaskSet 109.0 because task 3 (partition 3) cannot run anywhere due to node and executor blacklist.  Blacklisting behavior can be configured via spark.blacklist.*.&lt;/PRE&gt;&lt;P&gt;If you have the data in HDFS in (csv/avro/parquet) format, then you can use the below command to import the files to Kudu table.&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Prerequisites:&lt;/STRONG&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;Kudu jar with compatible version (1.6 or higher)&amp;nbsp;&lt;A href="https://stackoverflow.com/questions/47891020/loading-data-from-hdfs-to-kudu" target="_blank" rel="noopener"&gt;For more reference&lt;/A&gt;&lt;/P&gt;&lt;DIV class="post-text"&gt;&lt;PRE&gt;spark2-submit  --master yarn/local  --class org.apache.kudu.spark.tools.ImportExportFiles &amp;lt;path of kudu jar&amp;gt;/kudu-spark2-tools_2.11-1.6.0.jar --operation=import --format=&amp;lt;parquet/avro/csv&amp;gt; --master-addrs=&amp;lt;kudu master host&amp;gt;:&amp;lt;port number&amp;gt;  --path=&amp;lt;hdfs path for data&amp;gt; --table-name=impala::&amp;lt;table name&amp;gt;&lt;/PRE&gt;&lt;/DIV&gt;&lt;DIV class="grid mb0 fw-wrap ai-start jc-end gs8 gsy"&gt;&lt;DIV class="grid--cell mr16"&gt;Hope this helps. Please accept the answer and vote up if it did.&lt;/DIV&gt;&lt;/DIV&gt;</description>
    <pubDate>Thu, 09 Apr 2020 08:56:59 GMT</pubDate>
    <dc:creator>jagadeesan</dc:creator>
    <dc:date>2020-04-09T08:56:59Z</dc:date>
    <item>
      <title>Transfer files from hdfs to kudu</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Transfer-files-from-hdfs-to-kudu/m-p/293515#M216731</link>
      <description>&lt;P&gt;Hello,&lt;/P&gt;
&lt;P&gt;I want to transfer files from hdfs to kudu.&amp;nbsp; I tried through talend fabric and its components but I have an error : Cannot run anywhere due to node and executor blacklist.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Can you help me please? Thanks a lot.&lt;/P&gt;</description>
      <pubDate>Wed, 08 Apr 2020 16:51:05 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Transfer-files-from-hdfs-to-kudu/m-p/293515#M216731</guid>
      <dc:creator>drgenious</dc:creator>
      <dc:date>2020-04-08T16:51:05Z</dc:date>
    </item>
    <item>
      <title>Re: Transfer files from hdfs to kudu</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Transfer-files-from-hdfs-to-kudu/m-p/293593#M216764</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/72966"&gt;@drgenious&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Are you getting a similar error which reported in&amp;nbsp;&lt;A href="https://issues.apache.org/jira/browse/KUDU-2633" target="_blank" rel="noopener"&gt;KUDU-2633&lt;/A&gt; It seems this is open JIRA reported in the community&lt;/P&gt;&lt;PRE&gt;ERROR core.JobRunShell: Job DEFAULT.EventKpisConsumer threw an unhandled Exception: 
org.apache.spark.SparkException: Job aborted due to stage failure: Aborting TaskSet 109.0 because task 3 (partition 3) cannot run anywhere due to node and executor blacklist.  Blacklisting behavior can be configured via spark.blacklist.*.&lt;/PRE&gt;&lt;P&gt;If you have the data in HDFS in (csv/avro/parquet) format, then you can use the below command to import the files to Kudu table.&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Prerequisites:&lt;/STRONG&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;Kudu jar with compatible version (1.6 or higher)&amp;nbsp;&lt;A href="https://stackoverflow.com/questions/47891020/loading-data-from-hdfs-to-kudu" target="_blank" rel="noopener"&gt;For more reference&lt;/A&gt;&lt;/P&gt;&lt;DIV class="post-text"&gt;&lt;PRE&gt;spark2-submit  --master yarn/local  --class org.apache.kudu.spark.tools.ImportExportFiles &amp;lt;path of kudu jar&amp;gt;/kudu-spark2-tools_2.11-1.6.0.jar --operation=import --format=&amp;lt;parquet/avro/csv&amp;gt; --master-addrs=&amp;lt;kudu master host&amp;gt;:&amp;lt;port number&amp;gt;  --path=&amp;lt;hdfs path for data&amp;gt; --table-name=impala::&amp;lt;table name&amp;gt;&lt;/PRE&gt;&lt;/DIV&gt;&lt;DIV class="grid mb0 fw-wrap ai-start jc-end gs8 gsy"&gt;&lt;DIV class="grid--cell mr16"&gt;Hope this helps. Please accept the answer and vote up if it did.&lt;/DIV&gt;&lt;/DIV&gt;</description>
      <pubDate>Thu, 09 Apr 2020 08:56:59 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Transfer-files-from-hdfs-to-kudu/m-p/293593#M216764</guid>
      <dc:creator>jagadeesan</dc:creator>
      <dc:date>2020-04-09T08:56:59Z</dc:date>
    </item>
  </channel>
</rss>

