Member since
03-14-2016
4721
Posts
1111
Kudos Received
874
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 2724 | 04-27-2020 03:48 AM | |
| 5283 | 04-26-2020 06:18 PM | |
| 4448 | 04-26-2020 06:05 PM | |
| 3575 | 04-13-2020 08:53 PM | |
| 5377 | 03-31-2020 02:10 AM |
06-28-2018
10:55 AM
@kanna k 2018-06-28 09:45:44,144 WARN [main] AbstractLifeCycle:204 - FAILED SelectChannelConnector@0.0.0.0:9000: java.net.BindException: Address already in use
2018-06-28 09:45:44,148 ERROR [main] SupportToolServer:624 - Failed to run the SmartSense Tool Server
java.net.BindException: Address already in use
On the host where you see the above error please check if there is any other process occupied that port some how? If yet then kill that process and then try starting the HST server again. Find Which process is using that port 9000 # netstat -tnlpa | grep 9000 The above command will show the PID which is using the port 9000 please check what that process is and then kill it if needed. # ps -ef | grep $PID_FROM_ABOVE_COMMAND
# kill -9 $PID_FROM_ABOVE_COMMAND . Alternate approach: Or try changing the SmartSense port Ambari UI --> SmartSense --> Configs --> Operations --> Web Ui Port (default is 9000) But better to find the process which is using the port 9000 and then stop it to ensure that SmartSense can use that port.
... View more
06-28-2018
06:57 AM
1 Kudo
@Takefumi Oide According to the ambari code base both "ntp" and "chrony" are fine any of these services are fine. https://github.com/apache/ambari/blob/release-2.6.2/ambari-agent/src/main/python/ambari_agent/HostInfo.py#L128-L134 . I see that there are few code changes made recently to also include the : https://issues.apache.org/jira/browse/AMBARI-18761 Doc can be enhanced.
... View more
06-27-2018
10:33 AM
1 Kudo
@Jyotsna Ganji You might want to see this JIRA: https://issues.apache.org/jira/browse/HADOOP-14217 and https://issues.apache.org/jira/browse/HDFS-13 ( the use of : is not yet resolved)
... View more
06-26-2018
05:19 AM
@Adi Jabkowsky Grafana pulls all historical metrics related data from AMS collector, So your AMS might be having the old host information. Please try accessing the following URL of AMS collector to see how many hosts are there and to verify the AMS database still has Old host info: http://<ams-host>:6188/ws/v1/timeline/metrics/hosts
... View more
06-26-2018
05:11 AM
1 Kudo
@Adi Jabkowsky Grafana collects information from Ambari Metrics collector. Ambari metrics collecotr stores hisgtorical metrics data. So your Old Host data will still be there. If you want to verify your host removal changes then click at the top right corner of the Grafana UI "Last 6 Hours" data to change the time after you deleted the Host. Other options may be to set the TTL value for the host level metrics (timeline.metrics.host.aggregator.daily.ttl) to a small value so that older data from the specified date is not maintained inside the AMS database (However this approach will cause deletion of other host metrics from the specified date) Please see this link for more information on TTL: https://docs.hortonworks.com/HDPDocuments/Ambari-2.6.0.0/bk_ambari-operations/content/ams_aggregated_metrics_ttl_settings.html
... View more
06-26-2018
01:47 AM
1 Kudo
@Gerg Git Ambari agent finds out all the host specific informations like diskspace/ memory(RAM)/ CPU and sends it to ambari server as part of it's registration request. If the cluster is already created and components/services are already installed then ambari can simply show the recommendations while making any configuration changes while ambari UI. You can refer to the Ambari Stack Advisory script: https://github.com/apache/ambari/blob/trunk/ambari-server/src/main/resources/stacks/stack_advisor.py . If you want to know more about it then there are few options available to determine the config requirements like: SmartSense - The memory-related rules are updated very frequently, at least once a quarter and have the most context as it has all of the SmartSense diagnostic information at its disposal and can take into account actual use versus configured use of services (cores, spindles, memory, other services running on that machine, other 3rd party utilities being run on that machine) Stack Advisor - Updated frequently, but tied to Ambari releases so depends on if the customer is using Ambari and if they are which specific version of Ambari and how up to date it is 1.7 vs 2.0 vs 2.1, etc.. HDP Configuration Utility - Most basic and least frequently updated, but if the customer does not have Ambari or SmartSense and is manually deploying HDP is better than nothing. . - https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.6.5/bk_command-line-installation/content/determine-hdp-memory-config.html https://cwiki.apache.org/confluence/display/AMBARI/How-To+Define+Stacks+and+Services#How-ToDefineStacksandServices-StackAdvisor https://community.hortonworks.com/questions/141855/stack-advisor-how-to-use-it.html
... View more
06-22-2018
11:31 PM
1 Kudo
@Utkarsh Jadhav Regarding your query: "The problem is Ambari itself is not doing the checkpoint (Assuming Ambari to do it)." >>>> Ambari is not responsible for doing the HDFS check pointing (rather it can simply alert if checkpoint did not happen). The Alert that you are getting is simply checking the HDFS Checkpoint time and reporting the alert. The "NameNode Last Checkpoint" can be triggered if Too much time elapsed since last NameNode checkpoint. We can see this alert if the last time that the NameNode performed a checkpoint was too long ago or if the number of uncommitted transactions is beyond a certain threshold. Checkpoointing is controlled by the following properties of HDFS configs so if it is not happening in regular interval then we will have to look the NN logs / gc logs / settings. dfs.namenode.checkpoint.period, set to 1 hour by default, specifies the maximum delay between two consecutive checkpoints dfs.namenode.checkpoint.txns, set to 1 million by default, defines the number of uncheckpointed transactions on the NameNode which will force an urgent checkpoint, even if the checkpoint period has not been reached. You could go to the Namenode current folder and check when was the last fsimage created. .
... View more
06-22-2018
07:14 AM
@Saurav D Good to know that it resolved your issue. If it helped then please mark this HCC thread as answered by clicking on "Accept" link on the correct answer, That way it will help other HCC users to quickly find the answers.
... View more
06-22-2018
06:35 AM
1 Kudo
@Saurav D Do you see any error/Warning on your Yarn logs? As we see and looks like resource manager port ? Retrying connect to server: sandbox-hdp.hortonworks.com/<masked IP>:8032 . Can we check if the port 8032 is opened and accessible? # netstat -tnlpa | grep 8032 . In Sandbox environment sometimes when we keep running the unwanted services then it creates a lots of load on the VM hence some services might not work properly sometimes. So you might want to stop some unwanted services in your sandbox and then try again.
... View more