<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Problem when Distcp between two HA Cluster. in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216467#M84166</link>
    <description>&lt;P&gt;Hi, &lt;A rel="user" href="https://community.cloudera.com/users/84313/jsneep.html" nodeid="84313"&gt;@Jonathan Sneep&lt;/A&gt; Thank you for your reply. I tried the method you suggest but got the same error. Below is the command what I ran:&lt;/P&gt;&lt;PRE&gt;hdfs --config /configurations/hadoop distcp -Dmapreduce.job.hdfs-servers.token-renewal.exclude=d001.server.edu.tk,d002.server.edu.tk -update -p hdfs://cluster_1:8020/tmp/ hdfs:/cluster_2:8020/tmp/&lt;/PRE&gt;</description>
    <pubDate>Tue, 09 Oct 2018 10:52:02 GMT</pubDate>
    <dc:creator>mrdashanzi</dc:creator>
    <dc:date>2018-10-09T10:52:02Z</dc:date>
    <item>
      <title>Problem when Distcp between two HA Cluster.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216463#M84162</link>
      <description>&lt;P&gt;
	Hi, ALL;&lt;/P&gt;&lt;P&gt;
	I have two hadoop cluster named as cluster_1 and cluster_2 with different zookeepers. Now I want to Distcp hdfs files from cluster_1 to cluster_2.&lt;/P&gt;&lt;P&gt;
	following as cluster_1's and cluster_2's information.&lt;/P&gt;
&lt;PRE&gt;### cluster_1
1, active namenode: g001.server.edu.tk standby namenode: g002.server.edu.tk
2, zookeeper hosts: g003.server.edu.tk g004.server.edu.tk g005.server.edu.tk
### cluster_2
1, active namenode: d001.server.edu.tk standby namenode: d002.server.edu.tk
2, zookeeper hosts: d003.server.edu.tk d004.server.edu.tk d005.server.edu.tk&amp;lt;br&amp;gt;
&lt;/PRE&gt;&lt;P&gt;1, In order to distcp data from cluster_1 to cluster_2, I copied whole hadoop configuration files from $HADOOP_HOME/etc/hadoop to /configurations/hadoop in cluster_1 and added following properties in hdfs-site.xml in g001.server.org.tk:&lt;/P&gt;&lt;PRE&gt;&amp;lt;property&amp;gt;
    &amp;lt;name&amp;gt;dfs.nameservices&amp;lt;/name&amp;gt;  
    &amp;lt;value&amp;gt;cluster_1,cluster_2&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt;

&amp;lt;property&amp;gt;   
    &amp;lt;name&amp;gt;dfs.client.failover.proxy.provider.cluster_2&amp;lt;/name&amp;gt; 
    &amp;lt;value&amp;gt;org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt; 
&amp;lt;property&amp;gt;
     &amp;lt;name&amp;gt;dfs.ha.namenodes.cluster_2&amp;lt;/name&amp;gt;    
    &amp;lt;value&amp;gt;d001.server.edu.tk,d002.server.edu.tk&amp;lt;/value&amp;gt; &amp;lt;/property&amp;gt;
&amp;lt;property&amp;gt;
    &amp;lt;name&amp;gt;dfs.namenode.rpc-address.cluster_2.d001.server.edu.tk&amp;lt;/name&amp;gt; 
    &amp;lt;value&amp;gt;d001.server.edu.tk:8020&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt; 

&amp;lt;property&amp;gt; 
    &amp;lt;name&amp;gt;dfs.namenode.servicerpc-address.cluster_2.d001.server.edu.tk&amp;lt;/name&amp;gt; 
    &amp;lt;value&amp;gt;d001.server.edu.tk:54321&amp;lt;/value&amp;gt; 
&amp;lt;/property&amp;gt;

&amp;lt;property&amp;gt;
    &amp;lt;name&amp;gt;dfs.namenode.http-address.cluster_2.d001.server.edu.tk&amp;lt;/name&amp;gt; 
    &amp;lt;value&amp;gt;d001.server.edu.tk:50070&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt;

&amp;lt;property&amp;gt; 
    &amp;lt;name&amp;gt;dfs.namenode.https-address.cluster_2.d001.server.edu.tk&amp;lt;/name&amp;gt; 
    &amp;lt;value&amp;gt;d001.server.edu.tk:50470&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt;

&amp;lt;property&amp;gt;
    &amp;lt;name&amp;gt;dfs.namenode.rpc-address.cluster_2.d002.server.edu.tk&amp;lt;/name&amp;gt; 
    &amp;lt;value&amp;gt;d002.server.edu.tk:8020&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt;

&amp;lt;property&amp;gt;
    &amp;lt;name&amp;gt;dfs.namenode.servicerpc-address.cluster_2.d002.server.edu.tk&amp;lt;/name&amp;gt; 
    &amp;lt;value&amp;gt;d002.server.edu.tk:54321&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt;

&amp;lt;property&amp;gt;
    &amp;lt;name&amp;gt;dfs.namenode.http-address.cluster_2.d002.server.edu.tk&amp;lt;/name&amp;gt;
    &amp;lt;value&amp;gt;d002.server.edu.tk:50070&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt;

&amp;lt;property&amp;gt; 
    &amp;lt;name&amp;gt;dfs.namenode.https-address.cluster_2.d002.server.edu.tk&amp;lt;/name&amp;gt;  
    &amp;lt;value&amp;gt;d002.server.edu.tk:50470&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt;

&amp;lt;property&amp;gt;
    &amp;lt;name&amp;gt;dfs.client.failover.proxy.provider.cluster_1&amp;lt;/name&amp;gt;     
    &amp;lt;value&amp;gt;org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider&amp;lt;/value&amp;gt; 
&amp;lt;/property&amp;gt; 

&amp;lt;property&amp;gt;
     &amp;lt;name&amp;gt;dfs.ha.namenodes.cluster_1&amp;lt;/name&amp;gt; 
    &amp;lt;value&amp;gt;g001.server.edu.tk,g002.server.edu.tk&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt;
&amp;lt;property&amp;gt;   
    &amp;lt;name&amp;gt;dfs.namenode.rpc-address.cluster_1.g001.server.edu.tk&amp;lt;/name&amp;gt; 
    &amp;lt;value&amp;gt;g001.server.edu.tk:8020&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt;

&amp;lt;property&amp;gt;
    &amp;lt;name&amp;gt;dfs.namenode.servicerpc-address.cluster_1.g001.server.edu.tk&amp;lt;/name&amp;gt;   
    &amp;lt;value&amp;gt;g001.server.edu.tk:54321&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt;

&amp;lt;property&amp;gt;
    &amp;lt;name&amp;gt;dfs.namenode.http-address.cluster_1.g001.server.edu.tk&amp;lt;/name&amp;gt; 
    &amp;lt;value&amp;gt;g001.server.edu.tk:50070&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt;
&amp;lt;property&amp;gt; 
    &amp;lt;name&amp;gt;dfs.namenode.https-address.cluster_1.g001.server.edu.tk&amp;lt;/name&amp;gt; 
    &amp;lt;value&amp;gt;g001.server.edu.tk:50470&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt;

&amp;lt;property&amp;gt;
    &amp;lt;name&amp;gt;dfs.namenode.rpc-address.cluster_1.g002.server.edu.tk&amp;lt;/name&amp;gt;
    &amp;lt;value&amp;gt;g002.server.edu.tk:8020&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt;

&amp;lt;property&amp;gt;
    &amp;lt;name&amp;gt;dfs.namenode.servicerpc-address.cluster_1.g002.server.edu.tk&amp;lt;/name&amp;gt; 
    &amp;lt;value&amp;gt;g002.server.edu.tk:54321&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt;

&amp;lt;property&amp;gt; 
    &amp;lt;name&amp;gt;dfs.namenode.http-address.cluster_1.g002.server.edu.tk&amp;lt;/name&amp;gt; 
     &amp;lt;value&amp;gt;g002.server.edu.tk:50070&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt;

&amp;lt;property&amp;gt;
    &amp;lt;name&amp;gt;dfs.namenode.https-address.cluster_1.g002.server.edu.tk&amp;lt;/name&amp;gt; 
    &amp;lt;value&amp;gt;g002.server.edu.tk:50470&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt;
&lt;/PRE&gt;&lt;P&gt;2, Then I ran the Distcp command in g001.server.org.tk:&lt;/P&gt;&lt;PRE&gt;hdfs --config /configurations/hadoop distcp -update -p hdfs://cluster_1:8020/tmp/ hdfs:/cluster_2:8020/tmp/&lt;/PRE&gt;&lt;P&gt;3, But got errors like:&lt;/P&gt;&lt;PRE&gt;18/10/08 07:55:00 ERROR tools.Distcp: Exception encountered
java.io.IOException: org.apache.hadoop.yarn.exception.YarnException: Failed to submit application_xxx to YARN: Failed to renew token Kind: HDFS_DELEGATION_YOKEN, Service: hdfs:cluster_2, Ident: (HDFS_DELEGATION_TOKEN token 50168 for hdfs)
   at org.apache.hadoop.mapred.YARNRunner.submitJob(YARNRunner.java:306)
   at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(Jobsubmitter.java:240)
   at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
   at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject;.java:422)
   at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698)
   at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
   at org.apache.hadoop.tools.DistCp.createAndSubmitJob(DistCp.java:183)
   at org.apache.hadoop.tools.DistCp.execute(DistCp.java:153)
   at org.apache.hadoop.tools.DistCp.run(DistCp.java:126)
   at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
   at org.apache.hadoop.tools.DistCp.main(DistCp.java:430)
Caused by: org.apache.hadoop.yarn.exceptions.YarnException: Failed to submit application_xxxx to YARN: Failed to RENEW token: Kind: HDFS_DELEGATION_TOKEN, Service: hfs:cluster_2, Ident: (HDFS_DELEGATION_TOKEN token 50168 for hdfs)
   at org.apache.hadoop.yarn.client.api.impl.YarnClientImpl.submitApplication(YarnClientImpl.java:271)
   at org.apache.hadoop.mapred.ResourceMgrDelegate.submitApplication(ResourceMgrDelegate.java:290)
   at org.apache.hadoop.mapred.YARNRunner.submitJob(YARNRunner.java:291)
   ... 12 more&lt;/PRE&gt;&lt;P&gt;4, Hadoop version is 2.7.3&lt;/P&gt;</description>
      <pubDate>Tue, 09 Oct 2018 10:52:00 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216463#M84162</guid>
      <dc:creator>mrdashanzi</dc:creator>
      <dc:date>2018-10-09T10:52:00Z</dc:date>
    </item>
    <item>
      <title>Re: Problem when Distcp between two HA Cluster.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216464#M84163</link>
      <description>&lt;P&gt;Use this, and server name and port if you are doing distcp directly to the active NN on remote cluster:&lt;/P&gt;&lt;PRE&gt;-Dmapreduce.job.hdfs-servers.token-renewal.exclude=cluster_2&lt;/PRE&gt;</description>
      <pubDate>Tue, 09 Oct 2018 10:52:01 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216464#M84163</guid>
      <dc:creator>pminovic</dc:creator>
      <dc:date>2018-10-09T10:52:01Z</dc:date>
    </item>
    <item>
      <title>Re: Problem when Distcp between two HA Cluster.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216465#M84164</link>
      <description>&lt;P&gt;Hi &lt;A rel="user" href="https://community.cloudera.com/users/98237/mrdashanzi.html" nodeid="98237"&gt;@Shen Sean&lt;/A&gt;&lt;/P&gt;&lt;P&gt;It looks like you may be hitting YARN-3021 - &lt;A href="https://issues.apache.org/jira/browse/YARN-3021" target="_blank"&gt;https://issues.apache.org/jira/browse/YARN-3021&lt;/A&gt;&lt;/P&gt;&lt;P&gt;Try the same distcp operation after adding in following parameters to distcp command&lt;/P&gt;&lt;PRE&gt;-Dmapreduce.job.hdfs-servers.token-renewal.exclude=&amp;lt;destinationNN1&amp;gt;,&amp;lt;destinationNN2&amp;gt;&lt;/PRE&gt;</description>
      <pubDate>Tue, 09 Oct 2018 10:52:01 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216465#M84164</guid>
      <dc:creator>JonathanSneep</dc:creator>
      <dc:date>2018-10-09T10:52:01Z</dc:date>
    </item>
    <item>
      <title>Re: Problem when Distcp between two HA Cluster.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216466#M84165</link>
      <description>&lt;P&gt;4, Hadoop version is 2.7.3&lt;/P&gt;</description>
      <pubDate>Tue, 09 Oct 2018 10:52:01 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216466#M84165</guid>
      <dc:creator>mrdashanzi</dc:creator>
      <dc:date>2018-10-09T10:52:01Z</dc:date>
    </item>
    <item>
      <title>Re: Problem when Distcp between two HA Cluster.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216467#M84166</link>
      <description>&lt;P&gt;Hi, &lt;A rel="user" href="https://community.cloudera.com/users/84313/jsneep.html" nodeid="84313"&gt;@Jonathan Sneep&lt;/A&gt; Thank you for your reply. I tried the method you suggest but got the same error. Below is the command what I ran:&lt;/P&gt;&lt;PRE&gt;hdfs --config /configurations/hadoop distcp -Dmapreduce.job.hdfs-servers.token-renewal.exclude=d001.server.edu.tk,d002.server.edu.tk -update -p hdfs://cluster_1:8020/tmp/ hdfs:/cluster_2:8020/tmp/&lt;/PRE&gt;</description>
      <pubDate>Tue, 09 Oct 2018 10:52:02 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216467#M84166</guid>
      <dc:creator>mrdashanzi</dc:creator>
      <dc:date>2018-10-09T10:52:02Z</dc:date>
    </item>
    <item>
      <title>Re: Problem when Distcp between two HA Cluster.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216468#M84167</link>
      <description>&lt;P&gt;Hi, &lt;A rel="user" href="https://community.cloudera.com/users/177/pminovic.html" nodeid="177"&gt;@Predrag Minovic&lt;/A&gt; Thank you for your reply. I tried the method you suggest but got the same error. Below is the command what I ran:&lt;/P&gt;&lt;PRE&gt;hdfs --config /configurations/hadoop distcp -Dmapreduce.job.hdfs-servers.token-renewal.exclude=d001.server.edu.tk,d002.server.edu.tk -update -p hdfs://cluster_1:8020/tmp/ hdfs:/cluster_2:8020/tmp/

hdfs --config /configurations/hadoop distcp -Dmapreduce.job.hdfs-servers.token-renewal.exclude=d001.server.edu.tk:8020,d002.server.edu.tk:8020 -update -p hdfs://cluster_1:8020/tmp/ hdfs:/cluster_2:8020/tmp/


hdfs --config /configurations/hadoop distcp -Dmapreduce.job.hdfs-servers.token-renewal.exclude=d001.server.edu.tk,d002.server.edu.tk:8020 -update -p hdfs://cluster_1:8020/tmp/ hdfs:/cluster_2:8020/tmp/


hdfs --config /configurations/hadoop distcp -Dmapreduce.job.hdfs-servers.token-renewal.exclude=cluster_2 -update -p hdfs://cluster_1:8020/tmp/ hdfs:/cluster_2:8020/tmp/


hdfs --config /configurations/hadoop distcp -Dmapreduce.job.hdfs-servers.token-renewal.exclude=cluster_2:8020 -update -p hdfs://cluster_1:8020/tmp/ hdfs:/cluster_2:8020/tmp/
&lt;BR /&gt;&lt;/PRE&gt;</description>
      <pubDate>Tue, 09 Oct 2018 10:52:10 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216468#M84167</guid>
      <dc:creator>mrdashanzi</dc:creator>
      <dc:date>2018-10-09T10:52:10Z</dc:date>
    </item>
    <item>
      <title>Re: Problem when Distcp between two HA Cluster.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216469#M84168</link>
      <description>&lt;P&gt;Sorry for hard to understand message, try this:&lt;/P&gt;&lt;PRE&gt;hdfs --config /configurations/hadoop distcp -Dmapreduce.job.hdfs-servers.token-renewal.exclude=cluster_2 -update -p hdfs://cluster_1/tmp/ hdfs:/cluster_2/tmp/&lt;BR /&gt;&lt;/PRE&gt;&lt;P&gt;Note that you don't need port when using NN service name. Also I suggest to copy first a small file or directory in /tmp, like /tmp/mydir1, just create that dir and put a few files inside. Also remove '-update -p' during initial tests. Once it starts working you can try all that.&lt;/P&gt;</description>
      <pubDate>Tue, 09 Oct 2018 16:06:42 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216469#M84168</guid>
      <dc:creator>pminovic</dc:creator>
      <dc:date>2018-10-09T16:06:42Z</dc:date>
    </item>
    <item>
      <title>Re: Problem when Distcp between two HA Cluster.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216470#M84169</link>
      <description>&lt;P&gt;Hi, &lt;A rel="user" href="https://community.cloudera.com/users/177/pminovic.html" nodeid="177"&gt;@Predrag Minovic&lt;/A&gt;; The hadoop version is 2.7.3. It seems property 'mapreduce.job.hdfs-servers.token-renewal.exclude' is not available for hadoop version under 2.8.0. I've tried the method you provided, but got same error.&lt;/P&gt;</description>
      <pubDate>Wed, 10 Oct 2018 09:07:40 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Problem-when-Distcp-between-two-HA-Cluster/m-p/216470#M84169</guid>
      <dc:creator>mrdashanzi</dc:creator>
      <dc:date>2018-10-10T09:07:40Z</dc:date>
    </item>
  </channel>
</rss>

