<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Job submitted on edge node runs in local host and cannot be tracked under resource manager web u in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Job-submitted-on-edge-node-runs-in-local-host-and-cannot-be/m-p/62025#M71460</link>
    <description>&lt;P&gt;Thanks a lot.. This resolved the issue : )&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I have one more doubt, If I get java heap size issue like,&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Caused by: java.lang.OutOfMemoryError: Java heap space when running any mapreduce job, how to increase the java heap size runtime? Does&amp;nbsp;“-Dmapreduce.map.java.opts=-Xmx2048m” this really do something there? I dint find any changes. Could you please advice the best way to increase java heap size? Thanks in advance&lt;/SPAN&gt;&lt;/P&gt;</description>
    <pubDate>Tue, 21 Nov 2017 06:16:50 GMT</pubDate>
    <dc:creator>bvk</dc:creator>
    <dc:date>2017-11-21T06:16:50Z</dc:date>
    <item>
      <title>Job submitted on edge node runs in local host and cannot be tracked under resource manager web ui</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Job-submitted-on-edge-node-runs-in-local-host-and-cannot-be/m-p/62001#M71458</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I have 8 node cluster, when i submit job in edge node (Pi program), it creates job in local and executes&lt;/P&gt;&lt;P&gt;hadoop jar /opt/cloudera/parcels/CDH-5.7.1-1.cdh5.7.1.p0.11/jars/hadoop-examples.jar pi 10 10&lt;BR /&gt;Number of Maps = 10&lt;BR /&gt;Samples per Map = 10&lt;BR /&gt;Wrote input for Map #0&lt;BR /&gt;Wrote input for Map #1&lt;BR /&gt;Wrote input for Map #2&lt;BR /&gt;Wrote input for Map #3&lt;BR /&gt;Wrote input for Map #4&lt;BR /&gt;Wrote input for Map #5&lt;BR /&gt;Wrote input for Map #6&lt;BR /&gt;Wrote input for Map #7&lt;BR /&gt;Wrote input for Map #8&lt;BR /&gt;Wrote input for Map #9&lt;BR /&gt;Starting Job&lt;BR /&gt;17/11/20 08:47:57 INFO Configuration.deprecation: session.id is deprecated. Instead, use dfs.metrics.session-id&lt;BR /&gt;17/11/20 08:47:57 INFO jvm.JvmMetrics: Initializing JVM Metrics with processName=JobTracker, sessionId=&lt;BR /&gt;17/11/20 08:47:58 INFO input.FileInputFormat: Total input paths to process : 10&lt;BR /&gt;17/11/20 08:47:58 INFO mapreduce.JobSubmitter: number of splits:10&lt;BR /&gt;17/11/20 08:47:58 INFO mapreduce.JobSubmitter: Submitting tokens for job:&lt;STRONG&gt; job_local635221628_0001&lt;/STRONG&gt;&lt;BR /&gt;17/11/20 08:47:58 INFO mapreduce.Job: The url to track the job: &lt;A href="http://localhost:8080/" target="_blank"&gt;http://localhost:8080/&lt;/A&gt;&lt;BR /&gt;17/11/20 08:47:58 INFO mapreduce.Job: Running job: job_local635221628_0001&lt;BR /&gt;17/11/20 08:47:58 INFO mapred.LocalJobRunner: OutputCommitter set in config null&lt;BR /&gt;17/11/20 08:47:58 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1&lt;BR /&gt;17/11/20 08:47:58 INFO mapred.LocalJobRunner: OutputCommitter is org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter&lt;BR /&gt;17/11/20 08:47:58 INFO mapred.LocalJobRunner: Waiting for map tasks&lt;BR /&gt;17/11/20 08:47:58 INFO mapred.LocalJobRunner: Starting task: attempt_local635221628_0001_m_000000_0&lt;BR /&gt;17/11/20 08:47:58 INFO output.FileOutputCommitter: File Output Committer Algorithm version is 1&lt;BR /&gt;17/11/20 08:47:58 INFO mapred.Task: Using ResourceCalculatorProcessTree : [ ]&lt;BR /&gt;17/11/20 08:47:58 INFO mapred.MapTask: Processing split: hdfs://nameservice-ha/user/hduser/QuasiMonteCarlo_1511185676373_1845096796/in/part0:0+118&lt;/P&gt;&lt;P&gt;but it executes succssfully.. But the job id&amp;nbsp;&lt;STRONG&gt;job_localxxx cannot be tracked under Resource manager web ui.&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;When I run the same job on any other node (Name node or worker node, proper job_id is getting created which will be available in resoure manager web ui)&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;Also I noticed, when I run mapred job -list in &lt;STRONG&gt;edge node&lt;/STRONG&gt;, throws me below error&lt;/P&gt;&lt;P&gt;mapred job -list&lt;BR /&gt;17/11/20 08:52:30 INFO Configuration.deprecation: session.id is deprecated. Instead, use dfs.metrics.session-id&lt;BR /&gt;17/11/20 08:52:30 INFO jvm.JvmMetrics: Initializing JVM Metrics with processName=JobTracker, sessionId=&lt;BR /&gt;Exception in thread "main" java.lang.NullPointerException&lt;BR /&gt;at org.apache.hadoop.mapreduce.tools.CLI.listJobs(CLI.java:604)&lt;BR /&gt;at org.apache.hadoop.mapreduce.tools.CLI.run(CLI.java:382)&lt;BR /&gt;at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)&lt;BR /&gt;at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)&lt;BR /&gt;at org.apache.hadoop.mapred.JobClient.main(JobClient.java:1269)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;And when I run&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;yarn application -list&lt;/STRONG&gt;&lt;BR /&gt;17/11/20 08:52:59 INFO client.RMProxy: Connecting to ResourceManager at /0.0.0.0:8032&lt;BR /&gt;17/11/20 08:53:00 INFO ipc.Client: Retrying connect to server: 0.0.0.0/0.0.0.0:8032. Already tried 0 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)&lt;BR /&gt;17/11/20 08:53:01 INFO ipc.Client: Retrying connect to server: 0.0.0.0/0.0.0.0:8032. Already tried 1 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)&lt;BR /&gt;17/11/20 08:53:02 INFO ipc.Client: Retrying connect to server: 0.0.0.0/0.0.0.0:8032. Already tried 2 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)&lt;BR /&gt;17/11/20 08:53:03 INFO ipc.Client: Retrying connect to server: 0.0.0.0/0.0.0.0:8032. Already tried 3 time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Where as these commands works fine in other nodes, I have oozie service installed and&amp;nbsp;&lt;SPAN&gt;ResourceManager Address is set to 8032.&amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Can some one tell me what went wrong? How can I fix this issue?&amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 12:32:21 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Job-submitted-on-edge-node-runs-in-local-host-and-cannot-be/m-p/62001#M71458</guid>
      <dc:creator>bvk</dc:creator>
      <dc:date>2022-09-16T12:32:21Z</dc:date>
    </item>
    <item>
      <title>Re: Job submitted on edge node runs in local host and cannot be tracked under resource manager web u</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Job-submitted-on-edge-node-runs-in-local-host-and-cannot-be/m-p/62019#M71459</link>
      <description>The default behaviour of Hadoop is to run things locally, in face of no&lt;BR /&gt;found YARN cluster configuration. In CM managed clusters, cluster&lt;BR /&gt;configuration for client programs are deployed by means of a Gateway role.&lt;BR /&gt;Your edge host is missing a gateway role and the subsequent config files&lt;BR /&gt;required to discover and use the cluster daemons.&lt;BR /&gt;&lt;BR /&gt;Do these two steps:&lt;BR /&gt;&lt;BR /&gt;1. Visit YARN -&amp;gt; Instances page in CM, then click 'Add Role Instances' and&lt;BR /&gt;under the Gateway type in the modal dialog, find and add your edge hostname&lt;BR /&gt;to it (this edge host should already be running a CM agent for it to show&lt;BR /&gt;up here).&lt;BR /&gt;&lt;BR /&gt;2. Deploy cluster-wide client configs, following this:&lt;BR /&gt;&lt;A href="https://www.youtube.com/watch?v=4S9H3wftM_0" target="_blank"&gt;https://www.youtube.com/watch?v=4S9H3wftM_0&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;Retry your commands after this completes. Also verify that your edge host&lt;BR /&gt;now has a proper /etc/hadoop/conf symlink, with the directory contents&lt;BR /&gt;carrying info about the cluster.&lt;BR /&gt;&lt;BR /&gt;P.s. Having HDFS Gateways is insufficient to connect to YARN, you will need&lt;BR /&gt;a YARN Gateway to connect to YARN.&lt;BR /&gt;</description>
      <pubDate>Tue, 21 Nov 2017 01:18:35 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Job-submitted-on-edge-node-runs-in-local-host-and-cannot-be/m-p/62019#M71459</guid>
      <dc:creator>Harsh J</dc:creator>
      <dc:date>2017-11-21T01:18:35Z</dc:date>
    </item>
    <item>
      <title>Re: Job submitted on edge node runs in local host and cannot be tracked under resource manager web u</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Job-submitted-on-edge-node-runs-in-local-host-and-cannot-be/m-p/62025#M71460</link>
      <description>&lt;P&gt;Thanks a lot.. This resolved the issue : )&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I have one more doubt, If I get java heap size issue like,&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Caused by: java.lang.OutOfMemoryError: Java heap space when running any mapreduce job, how to increase the java heap size runtime? Does&amp;nbsp;“-Dmapreduce.map.java.opts=-Xmx2048m” this really do something there? I dint find any changes. Could you please advice the best way to increase java heap size? Thanks in advance&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 21 Nov 2017 06:16:50 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Job-submitted-on-edge-node-runs-in-local-host-and-cannot-be/m-p/62025#M71460</guid>
      <dc:creator>bvk</dc:creator>
      <dc:date>2017-11-21T06:16:50Z</dc:date>
    </item>
  </channel>
</rss>

