<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Cannot modify hive.exec.max.dynamic.partitions.pernode at runtime in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Cannot-modify-hive-exec-max-dynamic-partitions-pernode-at/m-p/173820#M136083</link>
    <description>&lt;P&gt;I figured out the issue. The problem was in the format of the hive.security.authorization.sqlstd.confwhitelist in Ambari. It say's it's just a comma delimited list, but if you inspect the value in beeline e.g. &lt;/P&gt;&lt;PRE&gt;set hive.security.authorization.sqlstd.confwhitelist;&lt;/PRE&gt;&lt;P&gt;you see it's actually pipe delimited, and all of the periods are escaped with a backslash. If I put in exactly that format. e.g.:&lt;/P&gt;&lt;PRE&gt;hive\.auto\..*|hive\.cbo\..*|hive\.convert\..*|hive\.exec\.dynamic\.partition.*|hive\.exec\..*\.dynamic\.partitions\..*|hive\.exec\.compress\..*|hive\.exec\.infer\..*|hive\.exec\.mode.local\..*|hive\.exec\.orc\..*|hive\.exec\.parallel.*|hive\.explain\..*|hive\.fetch.task\..*|hive\.groupby\..*|hive\.hbase\..*|hive\.index\..*|hive\.index\..*|hive\.intermediate\..*|hive\.join\..*|hive\.limit\..*|hive\.log\..*|hive\.mapjoin\..*|hive\.merge\..*|hive\.optimize\..*|hive\.orc\..*|hive\.outerjoin\..*|hive\.parquet\..*|hive\.ppd\..*|hive\.prewarm\..*|hive\.server2\.proxy\.user|hive\.skewjoin\..*|hive\.smbjoin\..*|hive\.stats\..*|hive\.tez\..*|hive\.vectorized\..*|mapred\.map\..*|mapred\.reduce\..*|mapred\.output\.compression\.codec|mapred\.job\.queuename|mapred\.output\.compression\.type|mapred\.min\.split\.size|mapreduce\.job\.reduce\.slowstart\.completedmaps|mapreduce\.job\.queuename|mapreduce\.job\.tags|mapreduce\.input\.fileinputformat\.split\.minsize|mapreduce\.map\..*|mapreduce\.reduce\..*|mapreduce\.output\.fileoutputformat\.compress\.codec|mapreduce\.output\.fileoutputformat\.compress\.type|tez\.am\..*|tez\.task\..*|tez\.runtime\..*|tez.queue.name|hive\.exec\.reducers\.bytes\.per\.reducer|hive\.client\.stats\.counters|hive\.exec\.default\.partition\.name|hive\.exec\.drop\.ignorenonexistent|hive\.counters\.group\.name|hive\.default\.fileformat\.managed|hive\.enforce\.bucketing|hive\.enforce\.bucketmapjoin|hive\.enforce\.sorting|hive\.enforce\.sortmergebucketmapjoin|hive\.cache\.expr\.evaluation|hive\.hashtable\.loadfactor|hive\.hashtable\.initialCapacity|hive\.ignore\.mapjoin\.hint|hive\.limit\.row\.max\.size|hive\.mapred\.mode|hive\.map\.aggr|hive\.compute\.query\.using\.stats|hive\.exec\.rowoffset|hive\.variable\.substitute|hive\.variable\.substitute\.depth|hive\.autogen\.columnalias\.prefix\.includefuncname|hive\.autogen\.columnalias\.prefix\.label|hive\.exec\.check\.crossproducts|hive\.compat|hive\.exec\.concatenate\.check\.index|hive\.display\.partition\.cols\.separately|hive\.error\.on\.empty\.partition|hive\.execution\.engine|hive\.exim\.uri\.scheme\.whitelist|hive\.file\.max\.footer|hive\.mapred\.supports\.subdirectories|hive\.insert\.into\.multilevel\.dirs|hive\.localize\.resource\.num\.wait\.attempts|hive\.multi\.insert\.move\.tasks\.share\.dependencies|hive\.support\.quoted\.identifiers|hive\.resultset\.use\.unique\.column\.names|hive\.analyze\.stmt\.collect\.partlevel\.stats|hive\.server2\.logging\.operation\.level|hive\.support\.sql11\.reserved\.keywords|hive\.exec\.job\.debug\.capture\.stacktraces|hive\.exec\.job\.debug\.timeout|hive\.exec\.max\.created\.files|hive\.exec\.reducers\.max|hive\.reorder\.nway\.joins|hive\.output\.file\.extension|hive\.exec\.show\.job\.failure\.debug\.info|hive\.exec\.tasklog\.debug\.timeout|parquet\.compression|hive\.exec\.max\.dynamic\.partitions|hive\.exec\.max\.dynamic\.partitions\.pernode&lt;/PRE&gt;&lt;P&gt;Then I can set these values at runtime:&lt;/P&gt;&lt;PRE&gt;0: jdbc:hive2://whqlhdpm004:10000/lab&amp;gt; set hive.exec.max.dynamic.partitions.pernode;
+------------------------------------------------+--+
|                      set                       |
+------------------------------------------------+--+
| hive.exec.max.dynamic.partitions.pernode=5000  |
+------------------------------------------------+--+
1 row selected (0.117 seconds)
0: jdbc:hive2://whqlhdpm004:10000/lab&amp;gt; set hive.exec.max.dynamic.partitions.pernode=10000;
No rows affected (0.003 seconds)
0: jdbc:hive2://whqlhdpm004:10000/lab&amp;gt; set hive.exec.max.dynamic.partitions.pernode;
+-------------------------------------------------+--+
|                       set                       |
+-------------------------------------------------+--+
| hive.exec.max.dynamic.partitions.pernode=10000  |
+-------------------------------------------------+--+
1 row selected (0.005 seconds)
0: jdbc:hive2://whqlhdpm004:10000/lab&amp;gt;&lt;/PRE&gt;</description>
    <pubDate>Wed, 04 Jan 2017 06:28:47 GMT</pubDate>
    <dc:creator>barret_miller</dc:creator>
    <dc:date>2017-01-04T06:28:47Z</dc:date>
  </channel>
</rss>

