<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Setup a Hadoop Multinode Cluster (2 Nodes) in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Setup-a-Hadoop-Multinode-Cluster-2-Nodes/m-p/221831#M74826</link>
    <description>&lt;P&gt;Was able to get an answer. Removing the "file" in hdfs-site.xml did the trick.&lt;/P&gt;</description>
    <pubDate>Mon, 26 Feb 2018 23:25:52 GMT</pubDate>
    <dc:creator>RhettDB</dc:creator>
    <dc:date>2018-02-26T23:25:52Z</dc:date>
    <item>
      <title>Setup a Hadoop Multinode Cluster (2 Nodes)</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Setup-a-Hadoop-Multinode-Cluster-2-Nodes/m-p/221829#M74824</link>
      <description>&lt;P&gt;I have two computers, the one I work on (CENTOS installed) and a second computer (also CENTOS (server), to act as the datanode), both not in a VM environment. I want to create a multi-node cluster with these computers. I have directly connected the computers together to test for possible network issues (ports etc.) and found that not to be the issue. I also used the guide by &lt;/P&gt;&lt;P&gt;&lt;A href="https://tecadmin.net/set-up-hadoop-multi-node-cluster-on-centos-redhat/#" target="_blank"&gt;https://tecadmin.net/set-up-hadoop-multi-node-cluster-on-centos-redhat/#&lt;/A&gt; and &lt;/P&gt;&lt;P&gt;&lt;A href="https://dwbi.org/etl/bigdata/183-setup-hadoop-cluster" target="_blank"&gt;https://dwbi.org/etl/bigdata/183-setup-hadoop-cluster&lt;/A&gt;. I have created a 'hadoop' user on both machines, with permissions, and established a password-less SSH access between both.&lt;/P&gt;&lt;P&gt;The hostname for the computers are:&lt;/P&gt;&lt;P&gt;1. NameNode (main computer): master&lt;/P&gt;&lt;P&gt;2. DataNode (the server): datanode1&lt;/P&gt;&lt;P&gt;The /etc/hosts file I have as(showing  'computerIP' in place of the actual IP's):&lt;/P&gt;&lt;P&gt;computerIP master&lt;/P&gt;&lt;P&gt;computerIP datanode1&lt;/P&gt;&lt;P&gt;My .xml file configurations on the NameNode  are:&lt;/P&gt;&lt;P&gt;1. core-site.xml:&lt;/P&gt;&lt;P&gt;&amp;lt;configuration&amp;gt;&lt;/P&gt;&lt;P&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;fs.defaultFS&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;hdfs://master:8020/&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;io.file.buffer.size&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;131072&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;/P&gt;&lt;P&gt;&amp;lt;/configuration&amp;gt;&lt;/P&gt;&lt;P&gt;2. hdfs-site.xml:&lt;/P&gt;&lt;P&gt;&amp;lt;configuration&amp;gt;&lt;/P&gt;&lt;P&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;dfs.namenode.name.dir&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;file:/opt/volume/namenode&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;dfs.datanode.data.dir&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;file:/opt/volume/datanode&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;dfs.namenode.checkpoint.dir&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;file:/opt/volume/namesecondary&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;dfs.replication&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;fs.defaultFS&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;hdfs://master:8020/&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;io.file.buffer.size&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;131072&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;/P&gt;&lt;P&gt; &amp;lt;/configuration&amp;gt;&lt;/P&gt;&lt;P&gt;3. mapred-site.xml:&lt;/P&gt;&lt;P&gt;&amp;lt;configuration&amp;gt;&lt;/P&gt;&lt;P&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;mapreduce.framework.name&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;yarn&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;mapreduce.jobhistory.address&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;master:10020&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;mapreduce.jobhistory.webapp.address&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;master:19888&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;yarn.app.mapreduce.am.staging-dir&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;/user/app&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;mapred.child.java.opts&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;-Djava.security.egd=file:/dev/../dev/urandom&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;/P&gt;&lt;P&gt;&amp;lt;/configuration&amp;gt;&lt;/P&gt;&lt;P&gt;4. yarn-site.xml:&lt;/P&gt;&lt;P&gt;&amp;lt;configuration&amp;gt;&lt;/P&gt;&lt;P&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;yarn.resourcemanager.hostname&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;master&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;yarn.resourcemanager.bind-host&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;0.0.0.0&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;yarn.nodemanager.bind-host&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;0.0.0.0&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;yarn.nodemanager.aux-services&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;mapreduce_shuffle&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;yarn.nodemanager.aux-services.mapreduce_shuffle.class&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;org.apache.hadoop.mapred.ShuffleHandler&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;yarn.log-aggregation-enable&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;true&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;yarn.nodemanager.local-dirs&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;file:/opt/volume/local&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;yarn.nodemanager.log-dirs&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;file:/opt/volume/yarn/log&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;BR /&gt;&amp;lt;property&amp;gt;&lt;BR /&gt;&amp;lt;name&amp;gt;yarn.nodemanager.remote-app-log-dir&amp;lt;/name&amp;gt;&lt;BR /&gt;&amp;lt;value&amp;gt;hdfs://master:8020/var/log/hadoop-yarn/apps&amp;lt;/value&amp;gt;&lt;BR /&gt;&amp;lt;/property&amp;gt;&lt;/P&gt;&lt;P&gt;&amp;lt;/configuration&amp;gt;&lt;/P&gt;&lt;P&gt;5.  JAVA_HOME (Where java is located):&lt;/P&gt;&lt;P&gt;# The java implementation to use.&lt;BR /&gt;export JAVA_HOME=/usr/lib/jvm/jre-1.8.0-openjdk&lt;/P&gt;&lt;P&gt;6. Slaves file:&lt;/P&gt;&lt;P&gt;datanode1&lt;/P&gt;&lt;P&gt;7. Masters file:&lt;/P&gt;&lt;P&gt;master&lt;/P&gt;&lt;P&gt;My .bashrc file is as follows:&lt;/P&gt;&lt;P&gt;export JAVA_HOME=/usr/lib/java-1.8.0&lt;BR /&gt;export PATH=$PATH:$JAVA_HOME/bin&lt;BR /&gt;export HADOOP_HOME=/opt/hadoop/hadoop-2.8.3&lt;BR /&gt;export PATH=$PATH:$HADOOP_HOME/bin&lt;BR /&gt;export PATH=$PATH:$HADOOP_HOME/sbin&lt;BR /&gt;export HADOOP_MAPRED_HOME=$HADOOP_HOME&lt;BR /&gt;export HADOOP_COMMON_HOME=$HADOOP_HOME&lt;BR /&gt;export HADOOP_HDFS_HOME=$HADOOP_HOME&lt;BR /&gt;export YARN_HOME=$HADOOP_HOME&lt;BR /&gt;export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native&lt;BR /&gt;export HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib/native"&lt;BR /&gt;export CLASSPATH=$CLASSPATH:/usr/local/hadoop/lib/*:.&lt;BR /&gt;&lt;BR /&gt;export HADOOP_OPTS="$HADOOP_OPTS -Djava.security.egd=file:/dev/../dev/urandom"&lt;/P&gt;&lt;P&gt;The permissions are as follows on both machines (from terminal):&lt;/P&gt;&lt;P&gt;[hadoop@master hadoop]$ ls -al /opt&lt;BR /&gt;total 0&lt;BR /&gt;drwxr-xr-x.  5 hadoop hadoop  44 Feb 15 16:05 .&lt;BR /&gt;dr-xr-xr-x. 17 root  root  242 Feb 21 11:38 ..&lt;BR /&gt;drwxr-xr-x.  3 hadoop hadoop  53 Feb 15 16:00 hadoop&lt;BR /&gt;drwxr-xr-x.  2 hadoop hadoop  6 Sep  7 01:11 rh&lt;BR /&gt;drwxr-xr-x.  7 hadoop hadoop  84 Feb 20 11:27 volume&lt;/P&gt;&lt;P&gt;For the DataNode:&lt;/P&gt;&lt;P&gt;[hadoop@datanode1 ~]$ ls -al /opt&lt;BR /&gt;total 0&lt;BR /&gt;drwxrwxrwx.  4 hadoop hadoop  34 Feb 20 11:06 .&lt;BR /&gt;dr-xr-xr-x. 17 root  root  242 Feb 19 16:13 ..&lt;BR /&gt;drwxr-xr-x.  3 hadoop hadoop  53 Feb 20 11:07 hadoop&lt;BR /&gt;drwxrwxrwx.  5 hadoop hadoop  59 Feb 21 09:53 volume&lt;/P&gt;&lt;P&gt;So when I go to format the namenode: hdfs namenode -format, I get that the NameNode is formatted on the 'master'.&lt;/P&gt;&lt;P&gt;And then I go start the system, $HADOOP_HOME/sbin/start-dfs.sh and get the following output:&lt;/P&gt;&lt;P&gt;[hadoop@master hadoop]$ $HADOOP_HOME/sbin/start-dfs.sh&lt;BR /&gt;Starting namenodes on [master]&lt;BR /&gt;master: starting namenode, logging to /opt/hadoop/hadoop-2.8.3/logs/hadoop-hadoop-namenode-master.out&lt;BR /&gt;datanode1: starting datanode, logging to /opt/hadoop/hadoop-2.8.3/logs/hadoop-hadoop-datanode-datanode1.out&lt;BR /&gt;Starting secondary namenodes [0.0.0.0]&lt;BR /&gt;0.0.0.0: starting secondarynamenode, logging to /opt/hadoop/hadoop-2.8.3/logs/hadoop-hadoop-secondarynamenode-master.out&lt;/P&gt;&lt;P&gt;Showing that the datanode is started, yet i go to the 50070 terminal to find that the datanode storage is not configured. I then stop the whole process, $HADOOP_HOME/sbin/stop-dfs.sh only to find that indeed the datanode didnt even start in the first place.&lt;/P&gt;&lt;P&gt;[hadoop@master hadoop]$ $HADOOP_HOME/sbin/stop-dfs.sh&lt;BR /&gt;Stopping namenodes on [master]&lt;BR /&gt;master: stopping namenode&lt;BR /&gt;datanode1: no datanode to stop&lt;BR /&gt;Stopping secondary namenodes [0.0.0.0]&lt;BR /&gt;0.0.0.0: stopping secondarynamenode&lt;/P&gt;&lt;P&gt;This is even when the computers are directly connected together. I have no idea why the datanode is not starting, and I hope someone could help. Need this for my masters thesis. &lt;/P&gt;&lt;P&gt;Thanks!&lt;/P&gt;</description>
      <pubDate>Wed, 21 Feb 2018 18:48:40 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Setup-a-Hadoop-Multinode-Cluster-2-Nodes/m-p/221829#M74824</guid>
      <dc:creator>RhettDB</dc:creator>
      <dc:date>2018-02-21T18:48:40Z</dc:date>
    </item>
    <item>
      <title>Re: Setup a Hadoop Multinode Cluster (2 Nodes)</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Setup-a-Hadoop-Multinode-Cluster-2-Nodes/m-p/221830#M74825</link>
      <description>&lt;P&gt;Apologies for the line spacing, if it might be difficult to follow.&lt;/P&gt;</description>
      <pubDate>Wed, 21 Feb 2018 18:50:29 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Setup-a-Hadoop-Multinode-Cluster-2-Nodes/m-p/221830#M74825</guid>
      <dc:creator>RhettDB</dc:creator>
      <dc:date>2018-02-21T18:50:29Z</dc:date>
    </item>
    <item>
      <title>Re: Setup a Hadoop Multinode Cluster (2 Nodes)</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Setup-a-Hadoop-Multinode-Cluster-2-Nodes/m-p/221831#M74826</link>
      <description>&lt;P&gt;Was able to get an answer. Removing the "file" in hdfs-site.xml did the trick.&lt;/P&gt;</description>
      <pubDate>Mon, 26 Feb 2018 23:25:52 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Setup-a-Hadoop-Multinode-Cluster-2-Nodes/m-p/221831#M74826</guid>
      <dc:creator>RhettDB</dc:creator>
      <dc:date>2018-02-26T23:25:52Z</dc:date>
    </item>
  </channel>
</rss>

