Member since
03-14-2018
8
Posts
0
Kudos Received
0
Solutions
09-10-2018
09:36 PM
@Harsh J yeap with <property>
<name>dfs.ha.fencing.methods</name>
<value>shell(/bin/true)</value>
</property> It working perfect now Thanks you very much
... View more
09-10-2018
02:29 AM
Dear @Harsh J , Does 'Automatic Failover Configuration' need config 'Fencing Configuration' , It is 2 dependent section or I need both to config Automatic Failover . Because I met this error ou must configure a fencing method before using automatic failover.
org.apache.hadoop.ha.BadFencingConfigurationException: No fencer configured for NameNode at node1/x.x.x.x:8020
at org.apache.hadoop.hdfs.tools.NNHAServiceTarget.checkFencingConfigured(NNHAServiceTarget.java:132)
at org.apache.hadoop.ha.ZKFailoverController.doRun(ZKFailoverController.java:225)
at org.apache.hadoop.ha.ZKFailoverController.access$000(ZKFailoverController.java:60)
at org.apache.hadoop.ha.ZKFailoverController$1.run(ZKFailoverController.java:171)
at org.apache.hadoop.ha.ZKFailoverController$1.run(ZKFailoverController.java:167)
at org.apache.hadoop.security.SecurityUtil.doAsLoginUserOrFatal(SecurityUtil.java:444)
at org.apache.hadoop.ha.ZKFailoverController.run(ZKFailoverController.java:167)
at org.apache.hadoop.hdfs.tools.DFSZKFailoverController.main(DFSZKFailoverController.java:192)
2018-09-10 15:56:53,262 INFO org.apache.zookeeper.ZooKeeper: Session: 0x365c2b22a1e0000 closed
2018-09-10 15:56:53,262 INFO org.apache.zookeeper.ClientCnxn: EventThread shut down If I need both of them , so <property>
<name>dfs.ha.fencing.methods</name>
<value>shell(/path/to/my/script.sh --nameservice=$target_nameserviceid $target_host:$target_port)</value>
</property> What is the /path/to/my/script.sh The contain of this script I am not clear about this , pls explain and may be give me an example . Thanks you
... View more
09-08-2018
02:44 AM
Thanks @saranvisa but I am not using CM , just install by command line.
... View more
09-07-2018
04:10 AM
Yeap as @Harsh J said I am using a CDH package based (non-CM) installation. I will show more about my config I have 3 nodes: node 1 , node 2 , node 3 zookeeper in 3 nodes: 3 nodes with the same config maxClientCnxns=50
tickTime=2000
initLimit=10
syncLimit=5
dataDir=/var/lib/zookeeper
clientPort=2181
dataLogDir=/var/lib/zookeeper hdfs-site.xml in 3 nodes: <property>
<name>dfs.nameservices</name>
<value>mycluster</value>
</property>
<property>
<name>dfs.ha.namenodes.mycluster</name>
<value>node1,node2</value>
</property>
<property>
<name>dfs.namenode.rpc-address.mycluster.node1</name>
<value>node1:8020</value>
</property>
<property>
<name>dfs.namenode.rpc-address.mycluster.node2</name>
<value>node2:8020</value>
</property>
<property>
<name>dfs.namenode.http-address.mycluster.node1</name>
<value>node1:50070</value>
</property>
<property>
<name>dfs.namenode.http-address.mycluster.node2</name>
<value>node2:50070</value>
</property>
<property>
<name>dfs.namenode.shared.edits.dir</name>
<value>qjournal://node1:8485;node2:8485;node3:8485/mycluster</value>
</property>
<property>
<name>dfs.client.failover.proxy.provider.mycluster</name>
<value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
</property>
<property>
<name>dfs.ha.automatic-failover.enabled</name>
<value>true</value>
</property>
<property>
<name>dfs.journalnode.edits.dir</name>
<value>/namenode/dfs/jn</value>
</property> core-site.xml in 3 nodes: <property>
<name>fs.defaultFS</name>
<value>hdfs://mycluster</value>
</property>
<property>
<name>ha.zookeeper.quorum</name>
<value>node1:2181,node2:2181,node3:2181</value>
</property>
Above is config relate to cluster namnode. I doupt about zookeeper config , does is it enough ? Service install on each node Node 1:
hadoop-hdfs-journalnode hadoop-hdfs-namenode hadoop-hdfs-zkfc zookeeper-server
Node 2:
hadoop-hdfs-journalnode hadoop-hdfs-namenode hadoop-hdfs-zkfc zookeeper-server
Node 3:
hadoop-hdfs-journalnode zookeeper-server The firt time initial is ok: Node 1 active , Node 2 standby Stop namenode service on Node 1 => node 2 active => OK But when start service NameNode on Node 1 again node 1 active And node 2 active too => fail
... View more
09-05-2018
09:01 PM
Hi everyone, My system have 2 datanode, 2 namenode, 3 journalnode, 3 zookeeper service I had config cluster namenode ok , when browsing the admin page namenode:50070 , I had see 1 name node status (active) and one namenode status (standby). => OK When I stop active namenode the other with standby become active . => OK But the problem is how start the namenode which I had stop again ? I do the following : sudo -u hdfs hdfs namenode -bootstrapStandby -force
/etc/init.d/hadoop-hdfs-namenode start With above process sometime namenode start ok with standby mode , but sometime it start with active mode and then I have 2 active node (split brain !!) So what I have wrong , what is the right process to start a namenode had stop again Thanks you
... View more
Labels:
- Labels:
-
HDFS
-
Manual Installation
03-15-2018
08:10 PM
After check again hdfs-site.xml on 2 DataNode where Yarn Node Manager stand on , I see that the hdfs-site file missing this line when compare with the hdfs-site on Name Node <property>
<name>dfs.client.failover.proxy.provider.radcluster</name>
<value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
</property> restart the Yarn Node Manager Service , It work now , Thanks you
... View more
03-15-2018
02:22 AM
My system have
2 NameNode as Cluster (nn1 active, nn2 standby)
2 DataNode
1 Yarn Resource Manager service is on nn1
2 Yarn Node Manager is on 2 DataNode
Before config cluster NameNode everything work well
After config Namde Node as cluster , Yarn start ok but when submit script from Hue I have some error
2018-03-15 15:46:49,203 INFO org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ResourceLocalizationService: Failed to download resource { { hdfs://mycluster/user/oozie/share/lib/lib_20171215093741/pig/jsp-2.1-6.1.14.jar, 1513305464001, FILE, null },pending,[(container_1521103549281_0001_02_000001)],8742037577451777,DOWNLOADING}
java.lang.IllegalArgumentException: java.net.UnknownHostException: mycluster
at org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.java:406)
at org.apache.hadoop.hdfs.NameNodeProxies.createNonHAProxy(NameNodeProxies.java:310)
at org.apache.hadoop.hdfs.NameNodeProxies.createProxy(NameNodeProxies.java:176)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:728)
at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:671)
at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:155)
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2815)
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:98)
at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2852)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2834)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:387)
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:296)
at org.apache.hadoop.yarn.util.FSDownload.copy(FSDownload.java:249)
at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:356)
at org.apache.hadoop.yarn.util.FSDownload.call(FSDownload.java:60)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.net.UnknownHostException: mycluster
View log on both Yarn Node manager service I see above error .
How Can I fix this
Thanks
... View more
Labels:
- Labels:
-
Apache YARN
-
Cloudera Hue
-
HDFS