<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Hive Partitioning - maximum for cluster in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hive-Partitioning-maximum-for-cluster/m-p/56576#M63696</link>
    <description>I don't know of any hard limits. There are limitations as a table with 10k+ partitions will likely fail on operations against all partitions like 'drop table'. That is generally the soft cap on partitions per table.&lt;BR /&gt;&lt;BR /&gt;For the full cluster, the backend RDBMS hosting the metastore will dictate this somewhat. Again there is no hard limit. I have seen some near 10 million partitions across all tables. Granted HMS, HS2, and CatalogD were not stable due to the large partitions count. A single or set of large queries or full table scans would bring them down each time. Your HMS heap will also be large. Hive does have settings now to prevent full partitions grabs or limiting the partition count per query.&lt;BR /&gt;&lt;BR /&gt;The Hive community is moving HMS to be backed by HBase to address the scalability of partitions, tables, and databases.</description>
    <pubDate>Tue, 27 Jun 2017 14:19:05 GMT</pubDate>
    <dc:creator>mbigelow</dc:creator>
    <dc:date>2017-06-27T14:19:05Z</dc:date>
    <item>
      <title>Hive Partitioning - maximum for cluster</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hive-Partitioning-maximum-for-cluster/m-p/56542#M63695</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Our cluster is in CDH 5.7. We want to know the maximum partitions we can create for a hive/table and maximum allowed for a cluster ?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Best regards,&lt;/P&gt;&lt;P&gt;Olivier&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 11:50:01 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hive-Partitioning-maximum-for-cluster/m-p/56542#M63695</guid>
      <dc:creator>odanjean</dc:creator>
      <dc:date>2022-09-16T11:50:01Z</dc:date>
    </item>
    <item>
      <title>Re: Hive Partitioning - maximum for cluster</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hive-Partitioning-maximum-for-cluster/m-p/56576#M63696</link>
      <description>I don't know of any hard limits. There are limitations as a table with 10k+ partitions will likely fail on operations against all partitions like 'drop table'. That is generally the soft cap on partitions per table.&lt;BR /&gt;&lt;BR /&gt;For the full cluster, the backend RDBMS hosting the metastore will dictate this somewhat. Again there is no hard limit. I have seen some near 10 million partitions across all tables. Granted HMS, HS2, and CatalogD were not stable due to the large partitions count. A single or set of large queries or full table scans would bring them down each time. Your HMS heap will also be large. Hive does have settings now to prevent full partitions grabs or limiting the partition count per query.&lt;BR /&gt;&lt;BR /&gt;The Hive community is moving HMS to be backed by HBase to address the scalability of partitions, tables, and databases.</description>
      <pubDate>Tue, 27 Jun 2017 14:19:05 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hive-Partitioning-maximum-for-cluster/m-p/56576#M63696</guid>
      <dc:creator>mbigelow</dc:creator>
      <dc:date>2017-06-27T14:19:05Z</dc:date>
    </item>
    <item>
      <title>Re: Hive Partitioning - maximum for cluster</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hive-Partitioning-maximum-for-cluster/m-p/56628#M63697</link>
      <description>&lt;P&gt;Thank for your answer mbigelow&lt;/P&gt;</description>
      <pubDate>Wed, 28 Jun 2017 12:04:09 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hive-Partitioning-maximum-for-cluster/m-p/56628#M63697</guid>
      <dc:creator>odanjean</dc:creator>
      <dc:date>2017-06-28T12:04:09Z</dc:date>
    </item>
    <item>
      <title>Re: Hive Partitioning - maximum for cluster</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hive-Partitioning-maximum-for-cluster/m-p/56630#M63698</link>
      <description>&lt;P&gt;Would you consider using Bucketing instead of partitioning because it&lt;/P&gt;&lt;P&gt;decomposes data sets into much more manageable parts but still depends on the use case.&amp;nbsp;&lt;/P&gt;&lt;P&gt;insertion may take some time but intended for fast read .&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 28 Jun 2017 12:45:37 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hive-Partitioning-maximum-for-cluster/m-p/56630#M63698</guid>
      <dc:creator>csguna</dc:creator>
      <dc:date>2017-06-28T12:45:37Z</dc:date>
    </item>
    <item>
      <title>Re: Hive Partitioning - maximum for cluster</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hive-Partitioning-maximum-for-cluster/m-p/56953#M63699</link>
      <description>&lt;P&gt;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/18127"&gt;@mbigelow&lt;/a&gt;&amp;nbsp;i belive you can enforce partition count per query with this parameter&amp;nbsp;&lt;/P&gt;&lt;PRE&gt;hive.metastore.limit.partition.request&lt;/PRE&gt;&lt;P&gt;correct me if I am wrong.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 05 Jul 2017 09:55:33 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hive-Partitioning-maximum-for-cluster/m-p/56953#M63699</guid>
      <dc:creator>csguna</dc:creator>
      <dc:date>2017-07-05T09:55:33Z</dc:date>
    </item>
    <item>
      <title>Re: Hive Partitioning - maximum for cluster</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hive-Partitioning-maximum-for-cluster/m-p/65232#M63700</link>
      <description>&lt;P&gt;For other people reading this in 2018 and beyond NB&amp;nbsp;&lt;A href="https://issues.apache.org/jira/browse/HIVE-9452" target="_blank"&gt;https://issues.apache.org/jira/browse/HIVE-9452&lt;/A&gt; and&amp;nbsp;&lt;A href="https://issues.apache.org/jira/browse/HIVE-17234.&amp;nbsp;" target="_blank"&gt;https://issues.apache.org/jira/browse/HIVE-17234.&amp;nbsp;&lt;/A&gt; Essentially AFAIK development for an HBase backed metastore has stalled.&lt;/P&gt;</description>
      <pubDate>Fri, 09 Mar 2018 12:57:31 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hive-Partitioning-maximum-for-cluster/m-p/65232#M63700</guid>
      <dc:creator>russell_pierce</dc:creator>
      <dc:date>2018-03-09T12:57:31Z</dc:date>
    </item>
  </channel>
</rss>

