Member since
12-10-2015
24
Posts
13
Kudos Received
0
Solutions
08-17-2016
09:24 PM
for my questions : Yes the alert_definition can be updated using a PUT api call.
... View more
08-12-2016
01:47 AM
@Neeraj if I create my custom alert script and register it with the json. Will it be possible in future ti change the threshold ? Can I edit the alert definition? or I have to make the script not to take threshold values from the alert definition?
... View more
08-11-2016
07:32 PM
@Jay SenSharmaHi, I almost developed my script and then found this. Why can't we simply execute df -h from python and parse the result instead of doing all the calculations? Any problem with this approach?
... View more
06-24-2016
09:26 PM
I restarted the agents but looks like the errors are still present. I see the old cluster name in a execution_commands table. I have identified the task_id which are having the command haivng my old clustername. Should I delete them from the database?
... View more
06-23-2016
11:56 PM
1 Kudo
i am getting following logs in my ambari-server logs . I check the database and there is no entry of this cluster name "oldclustername" in any tables(i grepped on my dump file and also checked manually in all tables). I named my cluster as "oldclustername" but then renamed it to "newclustername" but logs are still showing these messages. Only table I see suspicious entries is ambari.request which has cluster id=-1 and 2, where cluster_id correctly maps to my newclustername but -1 looks like a wrong entry. Any suggestions where should I look? 23 Jun 2016 23:38:32,233 ERROR [alert-event-bus-2] AlertReceivedListener:329 - Unable to process alert yarn_nodemanager_health for an invalid cluster named oldclustername
23 Jun 2016 23:38:32,361 ERROR [alert-event-bus-2] AlertReceivedListener:329 - Unable to process alert ams_metrics_monitor_process for an invalid cluster named oldclustername
23 Jun 2016 23:38:32,362 WARN [alert-event-bus-2] AlertReceivedListener:248 - Cluster lookup failed for clusterName=oldclustername
23 Jun 2016 23:38:32,362 WARN [alert-event-bus-2] AlertReceivedListener:134 - Received an alert for ambari_agent_disk_usage which is a definition that does not exist anymore
23 Jun 2016 23:38:32,362 ERROR [alert-event-bus-2] AlertReceivedListener:329 - Unable to process alert kafka_broker_process for an invalid cluster named oldclustername
23 Jun 2016 23:38:32,942 ERROR [alert-event-bus-2] AlertReceivedListener:329 - Unable to process alert datanode_webui for an invalid cluster named oldclustername
23 Jun 2016 23:38:32,943 ERROR [alert-event-bus-2] AlertReceivedListener:329 - Unable to process alert datanode_process for an invalid cluster named oldclustername
23 Jun 2016 23:38:32,943 WARN [alert-event-bus-2] AlertReceivedListener:248 - Cluster lookup failed for clusterName=oldclustername
23 Jun 2016 23:38:32,943 WARN [alert-event-bus-2] AlertReceivedListener:134 - Received an alert for ambari_agent_disk_usage which is a definition that does not exist anymore
23 Jun 2016 23:38:32,943 ERROR [alert-event-bus-2] AlertReceivedListener:329 - Unable to process alert yarn_nodemanager_webui for an invalid cluster named oldclustername
23 Jun 2016 23:38:32,943 ERROR [alert-event-bus-2] AlertReceivedListener:329 - Unable to process alert ams_metrics_monitor_process for an invalid cluster named oldclustername
23 Jun 2016 23:38:32,943 ERROR [alert-event-bus-2] AlertReceivedListener:329 - Unable to process alert yarn_nodemanager_health for an invalid cluster named oldclustername
... View more
Labels:
- Labels:
-
Apache Ambari
06-22-2016
02:11 AM
Got it. Tried it on a different cluster and this is working
... View more
06-21-2016
11:08 PM
I added ambari.display.url = http://100.123.123.123:8080 in properties file but I still dont see ambari url in the email. I manually added $ambari.getUrl() but it gets in the email as it is. Any other thing that I can do?
... View more
06-21-2016
09:31 PM
I am trying to modify the template of the alert notification. I see in the template code it tries to display ambari url but it never displays that.
The lower end of the template reads like <div class="ambari-footer">
This notification was sent to $dispatch.getTargetName()
<br/>
Apache Ambari $ambari.getServerVersion()
#if( $ambari.hasUrl() )
<br/>
Ambari Server link: <a href=$ambari.getUrl()>$ambari.getUrl()</a>
#end
</div>
</html>
]]>
</body>
But I think method ambari.hasUrl() is always false. I tried printing the $ambari.getUrl() but it displays as it is. In the emails I see "$ambari.getUrl()". I checked the code https://github.com/apache/ambari/blob/71a1f7e0e5985b1a77bf09b976ebda3ab3fdbbf5/ambari-server/src/main/java/org/apache/ambari/server/state/services/AlertNoticeDispatchService.java @Inject
private Configuration m_configuration; I am not sure from where it is getting injected from. Any suggestions why ambari url is not coming in the configs.
... View more
Labels:
- Labels:
-
Apache Ambari
06-17-2016
12:35 AM
@Neeraj Sabharwal I deleted my service name from all these tables but I am still seeing
the same issue in the logs. I see this name is present in lot of tables clusterconfig execution_command upgrade_item stage serviceconfig requestresourcefilter requestoperationlevel request and there are many others. Should I delete from everywhere?
... View more
03-22-2016
01:13 AM
3 Kudos
I am trying to do hive bench marking(https://github.com/hortonworks/hive-testbench) but when I run setup script it loads data is some table but fails after sometime fails with the following error: OK
Time taken: 0.264 seconds
+ '[' X = X ']'
+ FORMAT=orc
+ i=1
+ total=24
+ DATABASE=tpcds_bin_partitioned_orc_2
+ for t in '${FACTS}'
+ echo 'Optimizing table store_sales (1/24).'
Optimizing table store_sales (1/24).
+ COMMAND='hive -i settings/load-partitioned.sql -f ddl-tpcds/bin_partitioned/store_sales.sql -d DB=tpcds_bin_partitioned_orc_2 -d SCALE=2 -d SOURCE=tpcds_text_2 -d BUCKETS=1 -d RETURN_BUCKETS=1 -d FILE=orc'
+ runcommand 'hive -i settings/load-partitioned.sql -f ddl-tpcds/bin_partitioned/store_sales.sql -d DB=tpcds_bin_partitioned_orc_2 -d SCALE=2 -d SOURCE=tpcds_text_2 -d BUCKETS=1 -d RETURN_BUCKETS=1 -d FILE=orc'
+ '[' XON '!=' X ']'
+ hive -i settings/load-partitioned.sql -f ddl-tpcds/bin_partitioned/store_sales.sql -d DB=tpcds_bin_partitioned_orc_2 -d SCALE=2 -d SOURCE=tpcds_text_2 -d BUCKETS=1 -d RETURN_BUCKETS=1 -d FILE=orc
WARNING: Use "yarn jar" to launch YARN applications.
Logging initialized using configuration in file:/etc/hive/2.4.0.0-169/0/hive-log4j.properties
... OK
Time taken: 0.948 seconds
OK
Time taken: 0.238 seconds
OK
Time taken: 0.629 seconds
OK
Time taken: 0.248 seconds
Query ID = hdfs_20160322014240_60c3f689-816d-409e-b8c7-c6ea636fa12a
Total jobs = 1
Launching Job 1 out of 1
Dag submit failed due to Invalid TaskLaunchCmdOpts defined for Vertex Map 1 : Invalid/conflicting GC options found, cmdOpts="-server -Djava.net.preferIPv4Stack=true -Dhdp.version=2.4.0.0-169 -XX:+PrintGCDetails -verbose:gc -XX:+PrintGCTimeStamps -XX:+UseNUMA -XX:+UseParallelGC -XX:+PrintGCDetails -verbose:gc -XX:+PrintGCTimeStamps -XX:+UseNUMA -XX:+UseG1GC -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=/tmp/ -Dlog4j.configuratorClass=org.apache.tez.common.TezLog4jConfigurator -Dlog4j.configuration=tez-container-log4j.properties -Dyarn.app.container.log.dir=<LOG_DIR> -Dtez.root.logger=INFO,CLA " stack trace: [org.apache.tez.dag.api.DAG.createDag(DAG.java:859), org.apache.tez.client.TezClientUtils.prepareAndCreateDAGPlan(TezClientUtils.java:694), org.apache.tez.client.TezClient.submitDAGSession(TezClient.java:487), org.apache.tez.client.TezClient.submitDAG(TezClient.java:434), org.apache.hadoop.hive.ql.exec.tez.TezTask.submit(TezTask.java:439), org.apache.hadoop.hive.ql.exec.tez.TezTask.execute(TezTask.java:180), org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160), org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:89), org.apache.hadoop.hive.ql.exec.TaskRunner.run(TaskRunner.java:75)] retrying...
FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.tez.TezTask
+ '[' 1 -ne 0 ']'
+ echo 'Command failed, try '\''export DEBUG_SCRIPT=ON'\'' and re-running'
Command failed, try 'export DEBUG_SCRIPT=ON' and re-running
+ exit 1 Not sure what is wrong. Anyhelp is appreciated.
... View more
Labels:
- Labels:
-
Apache Hive
-
Apache Tez