Created 09-12-2017 11:38 PM
After Install Kerberos and enable to system, Restart my namenode1 and namenode2 got this error.
17/09/13 11:39:39 INFO namenode.FSNamesystem: fsOwner = hdfs/namenode1.xx@MOPH.COM (auth:KERBEROS) 17/09/13 11:39:39 INFO namenode.FSNamesystem: supergroup = root 17/09/13 11:39:39 INFO namenode.FSNamesystem: isPermissionEnabled = true 17/09/13 11:39:39 INFO namenode.FSNamesystem: Determined nameservice ID: nameservice1 17/09/13 11:39:39 INFO namenode.FSNamesystem: HA Enabled: true 17/09/13 11:39:39 INFO namenode.FSNamesystem: Append Enabled: true 17/09/13 11:39:39 INFO util.GSet: Computing capacity for map INodeMap 17/09/13 11:39:39 INFO util.GSet: VM type = 64-bit 17/09/13 11:39:39 INFO util.GSet: 1.0% max memory 3.9 GB = 39.6 MB 17/09/13 11:39:39 INFO util.GSet: capacity = 2^22 = 4194304 entries 17/09/13 11:39:39 INFO namenode.FSDirectory: POSIX ACL inheritance enabled? false 17/09/13 11:39:39 INFO namenode.NameNode: Caching file names occuring more than 10 times 17/09/13 11:39:39 INFO util.GSet: Computing capacity for map cachedBlocks 17/09/13 11:39:39 INFO util.GSet: VM type = 64-bit 17/09/13 11:39:39 INFO util.GSet: 0.25% max memory 3.9 GB = 9.9 MB 17/09/13 11:39:39 INFO util.GSet: capacity = 2^20 = 1048576 entries 17/09/13 11:39:39 INFO namenode.FSNamesystem: dfs.namenode.safemode.threshold-pct = 0.9990000128746033 17/09/13 11:39:39 INFO namenode.FSNamesystem: dfs.namenode.safemode.min.datanodes = 1 17/09/13 11:39:39 INFO namenode.FSNamesystem: dfs.namenode.safemode.extension = 30000 17/09/13 11:39:39 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.window.num.buckets = 10 17/09/13 11:39:39 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.num.users = 10 17/09/13 11:39:39 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25 17/09/13 11:39:39 INFO namenode.FSNamesystem: Retry cache on namenode is enabled 17/09/13 11:39:39 INFO namenode.FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis 17/09/13 11:39:39 INFO util.GSet: Computing capacity for map NameNodeRetryCache 17/09/13 11:39:39 INFO util.GSet: VM type = 64-bit 17/09/13 11:39:39 INFO util.GSet: 0.029999999329447746% max memory 3.9 GB = 1.2 MB 17/09/13 11:39:39 INFO util.GSet: capacity = 2^17 = 131072 entries 17/09/13 11:39:39 INFO namenode.FSNamesystem: ACLs enabled? true 17/09/13 11:39:39 INFO namenode.FSNamesystem: XAttrs enabled? true 17/09/13 11:39:39 INFO namenode.FSNamesystem: Maximum size of an xattr: 16384 Running in non-interactive mode, and data appears to exist in QJM to [172.16.120.31:8485, 172.16.120.32:8485, 172.16.120.46:8485]. Not formatting. 17/09/13 11:39:39 INFO util.ExitUtil: Exiting with status 1 17/09/13 11:39:39 INFO namenode.NameNode: SHUTDOWN_MSG:
When i'm trying to run it manually with command
exec /usr/lib64/cmf/service/hdfs/hdfs.sh format-namenode cluster14
It error as
Wed Sep 13 13:20:59 ICT 2017
Wed Sep 13 13:20:59 ICT 2017
+ source_parcel_environment
+ '[' '!' -z '' ']'
+ locate_cdh_java_home
+ '[' -z /usr/java/jdk1.7.0_67-cloudera ']'
+ verify_java_home
+ '[' -z /usr/java/jdk1.7.0_67-cloudera ']'
+ echo JAVA_HOME=/usr/java/jdk1.7.0_67-cloudera
JAVA_HOME=/usr/java/jdk1.7.0_67-cloudera
+ . /usr/lib64/cmf/service/common/cdh-default-hadoop
++ [[ -z 5 ]]
++ '[' 5 = 3 ']'
++ '[' 5 = -3 ']'
++ '[' 5 -ge 4 ']'
++ export HADOOP_HOME_WARN_SUPPRESS=true
++ HADOOP_HOME_WARN_SUPPRESS=true
++ export HADOOP_PREFIX=
++ HADOOP_PREFIX=
++ export HADOOP_LIBEXEC_DIR=/libexec
++ HADOOP_LIBEXEC_DIR=/libexec
++ export HADOOP_CONF_DIR=/var/run/cloudera-scm-agent/process/5020-hdfs-NAMENODE-format
++ HADOOP_CONF_DIR=/var/run/cloudera-scm-agent/process/5020-hdfs-NAMENODE-format
++ export HADOOP_COMMON_HOME=
++ HADOOP_COMMON_HOME=
++ export HADOOP_HDFS_HOME=
++ HADOOP_HDFS_HOME=
++ export HADOOP_MAPRED_HOME=
++ HADOOP_MAPRED_HOME=
++ '[' 5 = 4 ']'
++ '[' 5 = 5 ']'
++ export HADOOP_YARN_HOME=
++ HADOOP_YARN_HOME=
++ replace_pid
++ echo
++ sed 's#{{PID}}#35248#g'
+ export HADOOP_NAMENODE_OPTS=
+ HADOOP_NAMENODE_OPTS=
++ replace_pid
++ echo
++ sed 's#{{PID}}#35248#g'
+ export HADOOP_DATANODE_OPTS=
+ HADOOP_DATANODE_OPTS=
++ replace_pid
++ echo
++ sed 's#{{PID}}#35248#g'
+ export HADOOP_SECONDARYNAMENODE_OPTS=
+ HADOOP_SECONDARYNAMENODE_OPTS=
++ replace_pid
++ echo
++ sed 's#{{PID}}#35248#g'
+ export HADOOP_NFS3_OPTS=
+ HADOOP_NFS3_OPTS=
++ replace_pid
++ echo
++ sed 's#{{PID}}#35248#g'
+ export HADOOP_JOURNALNODE_OPTS=
+ HADOOP_JOURNALNODE_OPTS=
+ '[' 5 -ge 4 ']'
+ HDFS_BIN=/bin/hdfs
+ export 'HADOOP_OPTS=-Djava.net.preferIPv4Stack=true '
+ HADOOP_OPTS='-Djava.net.preferIPv4Stack=true '
+ echo 'using /usr/java/jdk1.7.0_67-cloudera as JAVA_HOME'
using /usr/java/jdk1.7.0_67-cloudera as JAVA_HOME
+ echo 'using 5 as CDH_VERSION'
using 5 as CDH_VERSION
+ echo 'using /var/run/cloudera-scm-agent/process/5020-hdfs-NAMENODE-format as CONF_DIR'
using /var/run/cloudera-scm-agent/process/5020-hdfs-NAMENODE-format as CONF_DIR
+ echo 'using hdfs as SECURE_USER'
using hdfs as SECURE_USER
+ echo 'using hadoop as SECURE_GROUP'
using hadoop as SECURE_GROUP
+ set_hadoop_classpath
+ set_classpath_in_var HADOOP_CLASSPATH
+ '[' -z HADOOP_CLASSPATH ']'
+ [[ -n /usr/share/cmf ]]
++ find /usr/share/cmf/lib/plugins -maxdepth 1 -name '*.jar'
++ tr '\n' :
+ ADD_TO_CP=/usr/share/cmf/lib/plugins/event-publish-5.10.0-shaded.jar:/usr/share/cmf/lib/plugins/tt-instrumentation-5.10.0.jar:
+ [[ -n '' ]]
+ eval 'OLD_VALUE=$HADOOP_CLASSPATH'
++ OLD_VALUE=
+ NEW_VALUE=/usr/share/cmf/lib/plugins/event-publish-5.10.0-shaded.jar:/usr/share/cmf/lib/plugins/tt-instrumentation-5.10.0.jar:
+ export HADOOP_CLASSPATH=/usr/share/cmf/lib/plugins/event-publish-5.10.0-shaded.jar:/usr/share/cmf/lib/plugins/tt-instrumentation-5.10.0.jar
+ HADOOP_CLASSPATH=/usr/share/cmf/lib/plugins/event-publish-5.10.0-shaded.jar:/usr/share/cmf/lib/plugins/tt-instrumentation-5.10.0.jar
+ set -x
+ replace_conf_dir
+ echo CONF_DIR=/var/run/cloudera-scm-agent/process/5020-hdfs-NAMENODE-format
CONF_DIR=/var/run/cloudera-scm-agent/process/5020-hdfs-NAMENODE-format
+ echo CMF_CONF_DIR=/etc/cloudera-scm-agent
CMF_CONF_DIR=/etc/cloudera-scm-agent
+ EXCLUDE_CMF_FILES=('cloudera-config.sh' 'httpfs.sh' 'hue.sh' 'impala.sh' 'sqoop.sh' 'supervisor.conf' '*.log' '*.keytab' '*jceks')
++ printf '! -name %s ' cloudera-config.sh httpfs.sh hue.sh impala.sh sqoop.sh supervisor.conf '*.log' hdfs.keytab '*jceks'
+ find /var/run/cloudera-scm-agent/process/5020-hdfs-NAMENODE-format -type f '!' -path '/var/run/cloudera-scm-agent/process/5020-hdfs-NAMENODE-format/logs/*' '!' -name cloudera-config.sh '!' -name httpfs.sh '!' -name hue.sh '!' -name impala.sh '!' -name sqoop.sh '!' -name supervisor.conf '!' -name '*.log' '!' -name hdfs.keytab '!' -name '*jceks' -exec perl -pi -e 's#{{CMF_CONF_DIR}}#/var/run/cloudera-scm-agent/process/5020-hdfs-NAMENODE-format#g' '{}' ';'
+ make_scripts_executable
+ find /var/run/cloudera-scm-agent/process/5020-hdfs-NAMENODE-format -regex '.*\.\(py\|sh\)$' -exec chmod u+x '{}' ';'
+ '[' DATANODE_MAX_LOCKED_MEMORY '!=' '' ']'
+ ulimit -l
64
+ export HADOOP_IDENT_STRING=hdfs
+ HADOOP_IDENT_STRING=hdfs
+ '[' -n '' ']'
+ '[' mkdir '!=' format-namenode ']'
+ acquire_kerberos_tgt hdfs.keytab
+ '[' -z hdfs.keytab ']'
+ '[' -n '' ']'
+ '[' validate-writable-empty-dirs = format-namenode ']'
+ '[' file-operation = format-namenode ']'
+ '[' bootstrap = format-namenode ']'
+ '[' failover = format-namenode ']'
+ '[' transition-to-active = format-namenode ']'
+ '[' initializeSharedEdits = format-namenode ']'
+ '[' initialize-znode = format-namenode ']'
+ '[' format-namenode = format-namenode ']'
+ '[' -z '' ']'
+ echo 'No storage dirs specified.'
No storage dirs specified.The configuration hdfs-site.xml looks like this
<property>
<name>dfs.ha.namenodes.nameservice1</name>
<value>namenode108,namenode123</value>
</property>
<property>
<name>dfs.namenode.name.dir.nameservice1.namenode108</name>
<value>file:///mnt/disk1/dfs/nn,file:///mnt/disk2/dfs/nn</value>
</property>
<property>
<name>dfs.namenode.shared.edits.dir.nameservice1.namenode108</name>
<value>qjournal://namenode1.xx:8485;namenode2.xx:8485;service.moph.com:8485/nameservice1</value>
</property>
<property>
<name>dfs.namenode.rpc-address.nameservice1.namenode108</name>
<value>namenode1.xx:8020</value>
</property>
<property>
<name>dfs.namenode.servicerpc-address.nameservice1.namenode108</name>
<value>namenode1.xx:8022</value>
</property>
<property>
<name>dfs.namenode.http-address.nameservice1.namenode108</name>
<value>namenode1.xx:50070</value>
</property>
<property>
<name>dfs.namenode.https-address.nameservice1.namenode108</name>
<value>namenode1.xx:50470</value>
</property>
<property>
<name>dfs.namenode.name.dir.nameservice1.namenode123</name>
<value>file:///mnt/disk1/dfs/nn,file:///mnt/disk2/dfs/nn</value>
</property>
Created 09-13-2017 10:05 PM
Created 09-13-2017 10:05 PM
Created 09-20-2017 03:12 AM