<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Not able to run hive benchmark test in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-run-hive-benchmark-test/m-p/137782#M23442</link>
    <description>&lt;P&gt;Same here! Deployed Hortonworks Data Platform on Google Cloud using bdutil. Ambari is ok. When I try to gen data using tpcds setup it fails with same error as you though.&lt;/P&gt;&lt;P&gt;Did you find any solution? &lt;/P&gt;</description>
    <pubDate>Fri, 08 Apr 2016 11:46:07 GMT</pubDate>
    <dc:creator>tuliobragam</dc:creator>
    <dc:date>2016-04-08T11:46:07Z</dc:date>
    <item>
      <title>Not able to run hive benchmark test</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-run-hive-benchmark-test/m-p/137781#M23441</link>
      <description>&lt;P&gt;
	I am trying to do hive bench marking(https://github.com/hortonworks/hive-testbench)&lt;/P&gt;&lt;P&gt; but when I run setup script it loads data is some table but fails after sometime fails with the following error: &lt;/P&gt;&lt;P&gt;OK
Time taken: 0.264 seconds
+ '[' X = X ']'
+ FORMAT=orc
+ i=1
+ total=24
+ DATABASE=tpcds_bin_partitioned_orc_2
+ for t in '${FACTS}'
+ echo 'Optimizing table store_sales (1/24).'
Optimizing table store_sales (1/24).
+ COMMAND='hive -i settings/load-partitioned.sql -f ddl-tpcds/bin_partitioned/store_sales.sql     -d DB=tpcds_bin_partitioned_orc_2  -d SCALE=2     -d SOURCE=tpcds_text_2 -d BUCKETS=1     -d RETURN_BUCKETS=1 -d FILE=orc'
+ runcommand 'hive -i settings/load-partitioned.sql -f ddl-tpcds/bin_partitioned/store_sales.sql     -d DB=tpcds_bin_partitioned_orc_2  -d SCALE=2     -d SOURCE=tpcds_text_2 -d BUCKETS=1     -d RETURN_BUCKETS=1 -d FILE=orc'
+ '[' XON '!=' X ']'
+ hive -i settings/load-partitioned.sql -f ddl-tpcds/bin_partitioned/store_sales.sql -d DB=tpcds_bin_partitioned_orc_2 -d SCALE=2 -d SOURCE=tpcds_text_2 -d BUCKETS=1 -d RETURN_BUCKETS=1 -d FILE=orc
WARNING: Use "yarn jar" to launch YARN applications.

Logging initialized using configuration in file:/etc/hive/2.4.0.0-169/0/hive-log4j.properties
...&lt;/P&gt;&lt;P&gt;OK
Time taken: 0.948 seconds
OK
Time taken: 0.238 seconds
OK
Time taken: 0.629 seconds
OK
Time taken: 0.248 seconds
Query ID = hdfs_20160322014240_60c3f689-816d-409e-b8c7-c6ea636fa12a
Total jobs = 1
Launching Job 1 out of 1
Dag submit failed due to Invalid TaskLaunchCmdOpts defined for Vertex Map 1 : Invalid/conflicting GC options found, cmdOpts="-server -Djava.net.preferIPv4Stack=true -Dhdp.version=2.4.0.0-169 -XX:+PrintGCDetails -verbose:gc -XX:+PrintGCTimeStamps -XX:+UseNUMA -XX:+UseParallelGC -XX:+PrintGCDetails -verbose:gc -XX:+PrintGCTimeStamps -XX:+UseNUMA -XX:+UseG1GC -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=/tmp/ -Dlog4j.configuratorClass=org.apache.tez.common.TezLog4jConfigurator -Dlog4j.configuration=tez-container-log4j.properties -Dyarn.app.container.log.dir=&amp;lt;LOG_DIR&amp;gt; -Dtez.root.logger=INFO,CLA " stack trace: [org.apache.tez.dag.api.DAG.createDag(DAG.java:859), org.apache.tez.client.TezClientUtils.prepareAndCreateDAGPlan(TezClientUtils.java:694), org.apache.tez.client.TezClient.submitDAGSession(TezClient.java:487), org.apache.tez.client.TezClient.submitDAG(TezClient.java:434), org.apache.hadoop.hive.ql.exec.tez.TezTask.submit(TezTask.java:439), org.apache.hadoop.hive.ql.exec.tez.TezTask.execute(TezTask.java:180), org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160), org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:89), org.apache.hadoop.hive.ql.exec.TaskRunner.run(TaskRunner.java:75)] retrying...
FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.tez.TezTask
+ '[' 1 -ne 0 ']'
+ echo 'Command failed, try '\''export DEBUG_SCRIPT=ON'\'' and re-running'
Command failed, try 'export DEBUG_SCRIPT=ON' and re-running
+ exit 1&lt;/P&gt;&lt;P&gt;Not sure what is wrong. &lt;/P&gt;&lt;P&gt;Anyhelp is appreciated.&lt;/P&gt;</description>
      <pubDate>Tue, 22 Mar 2016 08:13:48 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-run-hive-benchmark-test/m-p/137781#M23441</guid>
      <dc:creator>amey_jain</dc:creator>
      <dc:date>2016-03-22T08:13:48Z</dc:date>
    </item>
    <item>
      <title>Re: Not able to run hive benchmark test</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-run-hive-benchmark-test/m-p/137782#M23442</link>
      <description>&lt;P&gt;Same here! Deployed Hortonworks Data Platform on Google Cloud using bdutil. Ambari is ok. When I try to gen data using tpcds setup it fails with same error as you though.&lt;/P&gt;&lt;P&gt;Did you find any solution? &lt;/P&gt;</description>
      <pubDate>Fri, 08 Apr 2016 11:46:07 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-run-hive-benchmark-test/m-p/137782#M23442</guid>
      <dc:creator>tuliobragam</dc:creator>
      <dc:date>2016-04-08T11:46:07Z</dc:date>
    </item>
    <item>
      <title>Re: Not able to run hive benchmark test</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-run-hive-benchmark-test/m-p/137783#M23443</link>
      <description>&lt;P&gt;So, I found out that the problem was caused by hive-testbench/settings/load-partitioned.sql. This file is used as init file for hive on generating TPC-DS data. It has some configs for hive, including &lt;STRONG&gt;hive.tez.java.opts.&lt;/STRONG&gt;&lt;/P&gt;&lt;PRE&gt;set hive.tez.java.opts=-XX:+PrintGCDetails -verbose:gc -XX:+PrintGCTimeStamps -XX:+UseNUMA -XX:+UseG1GC -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=/tmp/;&lt;/PRE&gt;&lt;P&gt;This config conflicts with default HDP Hive config.&lt;/P&gt;&lt;P&gt;Two ways to solve it: &lt;/P&gt;&lt;P&gt;1. Change &lt;STRONG&gt;hive.tez.java.opts&lt;/STRONG&gt; in &lt;STRONG&gt;hive-testbench/settings/load-partitioned.sql&lt;/STRONG&gt; to use UseParallelGC (recommended).&lt;/P&gt;&lt;P&gt;set hive.tez.java.opts=-XX:+PrintGCDetails -verbose:gc -XX:+PrintGCTimeStamps -XX:+UseNUMA -XX:+UseParallelGC -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=/tmp/;&lt;/P&gt;&lt;P&gt;or&lt;/P&gt;&lt;P&gt;2. Set hive config on Ambari to use UseG1GC java garbage collecto&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;tez.am.launch.cmd-opts:&lt;/STRONG&gt; -XX:+PrintGCDetails -verbose:gc -XX:+PrintGCTimeStamps -XX:+UseNUMA --XX:+UseG1GC&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;tez.task.launch.cmd-opts:&lt;/STRONG&gt; -XX:+PrintGCDetails -verbose:gc -XX:+PrintGCTimeStamps -XX:+UseNUMA -XX:+UseG1GC&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;hive.tez.java.&lt;/STRONG&gt;&lt;STRONG&gt;opt:&lt;/STRONG&gt; -server -Djava.net.preferIPv4Stack=true -XX:NewRatio=8 -XX:+UseNUMA -XX:+UseG1GC -XX:+PrintGCDetails -verbose:gc -XX:+PrintGCTimeStamps&lt;/P&gt;</description>
      <pubDate>Fri, 29 Apr 2016 23:55:27 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-run-hive-benchmark-test/m-p/137783#M23443</guid>
      <dc:creator>tuliobragam</dc:creator>
      <dc:date>2016-04-29T23:55:27Z</dc:date>
    </item>
    <item>
      <title>Re: Not able to run hive benchmark test</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-run-hive-benchmark-test/m-p/137784#M23444</link>
      <description>&lt;P&gt;+1 Another solution is to comment out hive.tez.java.opts in that sql file and manage the GC from Ambari.&lt;/P&gt;</description>
      <pubDate>Sat, 30 Apr 2016 16:45:27 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Not-able-to-run-hive-benchmark-test/m-p/137784#M23444</guid>
      <dc:creator>pminovic</dc:creator>
      <dc:date>2016-04-30T16:45:27Z</dc:date>
    </item>
  </channel>
</rss>

