Member since
03-14-2016
4721
Posts
1111
Kudos Received
874
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2460 | 04-27-2020 03:48 AM | |
4908 | 04-26-2020 06:18 PM | |
3985 | 04-26-2020 06:05 PM | |
3238 | 04-13-2020 08:53 PM | |
4950 | 03-31-2020 02:10 AM |
10-26-2018
01:07 AM
@Ayoub Bounfour Posting my response from other thread. here : https://community.hortonworks.com/comments/226829/view.html As we see the message: Will retry 6 time(s), caught exception: ZooKeeper node /hiveserver2 is not ready yet . So it looks like the following command is returning Empty JSON. Please replace
the xxxxxxx with your Zookeeper hostname. Also it is better to have
more than one zookeeper to maintain a zookeeper quorum so that if one ZK
does not work still we have other Zookeepers working. # /usr/hdp/current/zookeeper-client/bin/zkCli.sh -server xxxxxxxxx.internal:2181 ls /hiveserver2 | grep 'serverUri='' Can
you try running that command manually to validate if your Zookeeper is
accessible and running fine and is returning the znode properly? It might happen if your Zookeepers are not healthy (not running) or showing some errors in their logs.
... View more
10-26-2018
01:03 AM
@Ayoub Bounfour Looks like the following call is returning Empty JSON. Please replace the xxxxxxx with your Zookeeper hostname. Also it is better to have more than one zookeeper to maintain a zookeeper quorum so that if one ZK does not work still we have other Zookeepers working. # /usr/hdp/current/zookeeper-client/bin/zkCli.sh -server xxxxxxxxx.internal:2181 ls /hiveserver2 | grep 'serverUri='' Can you try running that command manually to validate if your Zookeeper is accessible and running fine and is returning the znode properly? It might happen if your Zookeepers are not healthy (not running) or showing some errors in their logs. .
... View more
10-25-2018
07:28 AM
@Tarun Parimi Thank you for sharing the working solution. I am marking this thread as resolved.
... View more
10-25-2018
05:44 AM
@Zholaman Kubaliyev Ambair also allows to disable alerts .. so if it is your test cluster and if you just want to suppress that alert then just click on that alert link in ambari UI and "Disable" button in the alerts page. https://docs.hortonworks.com/HDPDocuments/Ambari-2.7.1.0/managing-and-monitoring-ambari/content/amb_disabling_alerts.html
... View more
10-25-2018
05:39 AM
@Zholaman Kubaliyev Can you please check if the following "status" command is giving you proper output? # su - yarn-ats -c "/usr/hdp/current/hadoop-yarn-client/bin/yarn app -status ats-hbase" If "ats-hbase" service is down the you might see the error. Because the failing alert will also run the same query and then will attempt to parse the JSON output so if the status is not returning fine they you should see the alert. Please check if your "ats-hbase" service is running? Reference: https://docs.hortonworks.com/HDPDocuments/HDP3/HDP-3.0.1/data-operating-system/content/options_to_restart_ats-hbase.html https://docs.hortonworks.com/HDPDocuments/HDP3/HDP-3.0.1/data-operating-system/content/enable_system_service_mode.html
... View more
10-24-2018
03:00 AM
@Mudit Kumar Your configuration looks good however the Curl command seems incorrect. Your command as following: # curl -k -i -vvvv --negotiate -u : "http://<knox1>:50070/gateway/<cluster_name>/webhdfs/v1/user?=op=LISTSTATUS" . In the above command we see that you are using "http://<knox1>:50070". (which is a direct port 50070 for HDFS namenode) ideally you should be hitting the Knox port. Also as knox is a trusted proxy hence you do not need to use "--negotiate" in the curl command (as long as you have not configured knox for SPNEGO authentication) Hence you should try with the following kind of commands: As knox by default runs on "https" (instead of "http" and it';s default port is 8443 (instead of 50070) hence polease try this. Also please pass knox LDAP credentials. # curl -k -i -vvvv -u guest:guest-password "https://<knox1>:8443/gateway/<topology_name>/webhdfs/v1/user?=op=LISTSTATUS"
Example:
-----------
# curl -k -i -vvvv -u guest:guest-password "https://<knox1>:8443/gateway/defaqult/webhdfs/v1/user?=op=LISTSTATUS" . Here the <topolofy_name> should be your knox topology name like "default"
... View more
10-23-2018
10:50 PM
@Alex
Goron
Do you see any discrepancy between the output of the following commands? Can you please share the output? # sudo du -sh /disks/disk1/
# sudo df -h Are you able to make file on your own manually? (this is just to isolate the issue) # echo "ABCDEFGH" > /disks/disk1/log/ambari-metrics-collector/hbase-ams-master-node4.hdp.com.out
# echo "ABCDEFGH" > /disks/disk1/log/ambari-metrics-collector/hbase-ams-master-node4.hdp.com.log
# echo "ABCDEFGH" > /disks/disk1/log/ambari-metrics-collector/hbase-ams-master-node4.hdp.com.log There might be possibility that the OS still might have some Deleted File Reserved by Processes that we can find out by running the following command. # sudo lsof / | grep deleted If you find some oprocesses aer still holding up "Deleted" files then restart those processes. # sudo systemctl restart $SERVICE_NAME There is a set of metadata on filesystems called “inodes.” Inodes track information about files. # sudo df -i / Also please check if there are any Bad Blocks # sudo fsck -vcck /disks/disk1/ If None of the above helps the Please check if you can reboot the host where AMS is installed.
... View more
10-23-2018
06:32 AM
5 Kudos
@Arif Ishak As we see the error is basically happening because the Port 53 Already Bind 2018-10-23 13:51:37,962 ERROR dns.PrivilegedRegistryDNSStarter (PrivilegedRegistryDNSStarter.java:init(61)) - Error initializing Registry DNS
java.net.BindException: Problem binding to [node.b2be.com:53] java.net.BindException: Address already in use; For more details see: http://wiki.apache.org/hadoop/BindException . So please check if your host where you are planning to run the "YARN Registry DNS" has the port 53 available Or some other process is already using that port? # netstat -tnlpa | grep 53 . OR try verifying if you can open the port 53 using simple program like "nc" (netcat) onn that host? Or do you see Already Used port error there as well? # nc -l `hostname -f` 53
Ncat: bind to 10.10.20.30:53: Address already in use. QUITTING. . If yes then Try changing the port to see if it works Or stop the other process which is using that port? Check the value of the following properyy in your YARN configs "hadoop.registry.dns.bind-port". (registry.dns.bind-port) default value = 53 .
... View more
10-23-2018
04:23 AM
@Ronnie 10 Your JAR name is different. "hadoop-hdfs-client-tests.jar" (wrong jar) and I am using "hadoop-mapreduce-client-jobclient-tests.jar" (correct JAR) Your command: # sudo -u hdfs hadoop jar hadoop-hdfs-client-tests.jar TestDFSIO -write -nrFiles 10 -fileSize 100000 My Command: # cd /usr/hdp/3.0.0.0-1634/hadoop-mapreduce
# sudo -u hdfs hadoop jar hadoop-mapreduce-client-jobclient-tests.jar TestDFSIO -write -nrFiles 10 -fileSize 100000 . You can chekc if your JAR contains that class or not ass following: # /usr/jdk64/jdk1.8.0_112/bin/jar -tvf /usr/hdp/3.0.0.0-1634/hadoop-hdfs/hadoop-hdfs-client-tests.jar | grep TestDFSIO
Not Found . But if you use the correct JAR then you can see it. # /usr/jdk64/jdk1.8.0_112/bin/jar -tvf /usr/hdp/3.0.0.0-1634/hadoop-mapreduce/hadoop-mapreduce-client-jobclient-tests.jar | grep TestDFSIO
2323 Thu Jul 12 20:36:52 UTC 2018 org/apache/hadoop/fs/TestDFSIO$ByteMultiple.class
2974 Thu Jul 12 20:36:52 UTC 2018 org/apache/hadoop/fs/TestDFSIO$ReadMapper.class
4215 Thu Jul 12 20:36:52 UTC 2018 org/apache/hadoop/fs/TestDFSIO$RandomReadMapper.class
3148 Thu Jul 12 20:36:52 UTC 2018 org/apache/hadoop/fs/TestDFSIO$TruncateMapper.class
3233 Thu Jul 12 20:36:52 UTC 2018 org/apache/hadoop/fs/TestDFSIO$WriteMapper.class
4014 Thu Jul 12 20:36:52 UTC 2018 org/apache/hadoop/fs/TestDFSIO$IOStatMapper.class
1086 Thu Jul 12 20:36:52 UTC 2018 org/apache/hadoop/fs/TestDFSIO$1.class
24732 Thu Jul 12 20:36:52 UTC 2018 org/apache/hadoop/fs/TestDFSIO.class
3078 Thu Jul 12 20:36:52 UTC 2018 org/apache/hadoop/fs/TestDFSIO$AppendMapper.class
1895 Thu Jul 12 20:36:52 UTC 2018 org/apache/hadoop/fs/TestDFSIO$TestType.class .
... View more
10-23-2018
03:24 AM
@Ronnie 10 How are you running it ? Can you share the exact command? Example: # su - hdfs
# hadoop jar /usr/hdp/3.0.0.0-1634/hadoop-mapreduce/hadoop-mapreduce-client-jobclient-tests.jar TestDFSIO -write -nrFiles 10 -fileSize 100000
.
... View more