<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: HBase increase num of reducers for bulk loading with ImportTSV in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/HBase-increase-num-of-reducers-for-bulk-loading-with/m-p/39915#M25581</link>
    <description>The reduce phase of a bulk load preparation job is used to align the output&lt;BR /&gt;files against the # of regions under the targeted table. You will always&lt;BR /&gt;see the number of reducers equal the number of regions in the targeted&lt;BR /&gt;table during the time of launching the job.&lt;BR /&gt;&lt;BR /&gt;If you desire more reducers, you will need to pre-split your table&lt;BR /&gt;appropriately. Read up more on pre-splitting at&lt;BR /&gt;&lt;A href="http://hbase.apache.org/book.html#manual_region_splitting_decisions" target="_blank"&gt;http://hbase.apache.org/book.html#manual_region_splitting_decisions&lt;/A&gt;&lt;BR /&gt;</description>
    <pubDate>Wed, 20 Apr 2016 09:07:09 GMT</pubDate>
    <dc:creator>Harsh J</dc:creator>
    <dc:date>2016-04-20T09:07:09Z</dc:date>
    <item>
      <title>HBase increase num of reducers for bulk loading with ImportTSV</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/HBase-increase-num-of-reducers-for-bulk-loading-with/m-p/39911#M25580</link>
      <description>&lt;P&gt;Hi dear experts!&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;i'm trying to load data with &lt;A href="http://hbase.apache.org/0.94/book/ops_mgt.html#importtsv" target="_self"&gt;ImportTSV tool&amp;nbsp;&lt;/A&gt;, like this:&lt;/P&gt;&lt;P&gt;hbase org.apache.hadoop.hbase.mapreduce.ImportTsv -Dmapreduce.job.reduces=1000 -Dimporttsv.columns="data:SS_SOLD_DATE_SK, HBASE_ROW_KEY" -Dimporttsv.separator="|" -Dimporttsv.bulk.output=/tmp/store_sales_hbase store_sales /user/root/benchmarks/bigbench/data/store_sales/*,&lt;/P&gt;&lt;P&gt;but have only one reducer (despite on &lt;SPAN&gt;-Dmapreduce.job.reduces=1000 setting&lt;/SPAN&gt;).&lt;/P&gt;&lt;P&gt;i even set&amp;nbsp;&lt;SPAN&gt;mapreduce.job.reduces=1000 on the Cluster wide, but still have only one reducer.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Could anybody hint how to resolve this?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;thank you in advance for any input!&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 10:14:41 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/HBase-increase-num-of-reducers-for-bulk-loading-with/m-p/39911#M25580</guid>
      <dc:creator>fil</dc:creator>
      <dc:date>2022-09-16T10:14:41Z</dc:date>
    </item>
    <item>
      <title>Re: HBase increase num of reducers for bulk loading with ImportTSV</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/HBase-increase-num-of-reducers-for-bulk-loading-with/m-p/39915#M25581</link>
      <description>The reduce phase of a bulk load preparation job is used to align the output&lt;BR /&gt;files against the # of regions under the targeted table. You will always&lt;BR /&gt;see the number of reducers equal the number of regions in the targeted&lt;BR /&gt;table during the time of launching the job.&lt;BR /&gt;&lt;BR /&gt;If you desire more reducers, you will need to pre-split your table&lt;BR /&gt;appropriately. Read up more on pre-splitting at&lt;BR /&gt;&lt;A href="http://hbase.apache.org/book.html#manual_region_splitting_decisions" target="_blank"&gt;http://hbase.apache.org/book.html#manual_region_splitting_decisions&lt;/A&gt;&lt;BR /&gt;</description>
      <pubDate>Wed, 20 Apr 2016 09:07:09 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/HBase-increase-num-of-reducers-for-bulk-loading-with/m-p/39915#M25581</guid>
      <dc:creator>Harsh J</dc:creator>
      <dc:date>2016-04-20T09:07:09Z</dc:date>
    </item>
    <item>
      <title>Re: HBase increase num of reducers for bulk loading with ImportTSV</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/HBase-increase-num-of-reducers-for-bulk-loading-with/m-p/39967#M25582</link>
      <description>&lt;P&gt;As Harsh suggested, for a new (empty) table without any 'split' defined, the number of reducers will always be 1. If you pre-split the table before import, you'd get that many number of reducers (Of course pre-splitting needs a good idea of how your row keys are designed and is broad topic in itself. See HBase The Definitive Guide &amp;gt; Chapter 11 &amp;gt; Optimizing Splits and Compactions &amp;gt;Presplitting Regions )&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Example: When the table is pre-splitted with 6 regions:&lt;/P&gt;&lt;P&gt;hbase(main):002:0&amp;gt;create 'hly_temp2', {NAME =&amp;gt; 't', VERSIONS =&amp;gt; 1}, {SPLITS =&amp;gt; ['USW000138290206', 'USW000149290623', 'USW000231870807', 'USW000242331116', 'USW000937411119']} # hadoop jar /usr/lib/hbase/hbase-server.jar importtsv -Dimporttsv.bulk.output=/user/hac/output/2-4 -Dimporttsv.columns=HBASE_ROW_KEY,t:v01 hly_temp2 /user/hac/input/2-1&lt;/P&gt;&lt;P&gt;... ....&lt;/P&gt;&lt;P&gt;Job Counters&lt;/P&gt;&lt;P&gt;Launched map tasks=1&lt;/P&gt;&lt;P&gt;Launched reduce tasks=6 &amp;lt;&amp;lt;&amp;lt;&lt;/P&gt;</description>
      <pubDate>Thu, 21 Apr 2016 07:07:54 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/HBase-increase-num-of-reducers-for-bulk-loading-with/m-p/39967#M25582</guid>
      <dc:creator>AutoIN</dc:creator>
      <dc:date>2016-04-21T07:07:54Z</dc:date>
    </item>
  </channel>
</rss>

