Member since
03-14-2016
4721
Posts
1111
Kudos Received
874
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 2497 | 04-27-2020 03:48 AM | |
| 4964 | 04-26-2020 06:18 PM | |
| 4047 | 04-26-2020 06:05 PM | |
| 3279 | 04-13-2020 08:53 PM | |
| 4999 | 03-31-2020 02:10 AM |
02-12-2019
03:14 PM
@Michael Bronson Just create the home directory as follows # su - hdfs
$ hdfs dfs -mkdir /user/slider
$ hdfs dfs -chown slider:hdfs /user/slider That should be enough .. good luck
... View more
02-06-2019
03:06 AM
abandoning this effort. was not getting anywhere as I think the ldaps in corp.example.com and machein hostname setup was causing all my issues., it is not a preferred setup anyhow I needed to correct so i renamed all cluster membre with corp....
... View more
02-06-2019
08:51 PM
Hi @Jay Kumar SenSharma, I perform a fresh installation of the cluster hdp 2.6.5. I follow steps mentioned and it works now 🙂 Many Thanks for your help. Regards,
... View more
02-03-2019
07:33 AM
@Michael Bronson As we see that it is basically a "500 Error" which basically indicates an Internal Server Error hence you must see a very detailed Stack Trace inside your ambari-server.log. Can you please share the complete ambari-server.log so that we can check what might be failing.
... View more
02-12-2019
01:07 PM
Logs from Infra Solr now constantly show errors as below: 2019-02-12T10:20:33,548 [zkCallback-7-thread-4] WARN [c:hadoop_logs s:shard5 r:core_node19 x:hadoop_logs_shard5_replica_n16] org.apache.solr.update.PeerSync (PeerSync.java:489) - PeerSync: core=hadoop_logs_shard5_replica_n16 url=http://ip-10-241-10-96:8886/solr exception talking to http://ip-10-241-10-72:8886/solr/hadoop_logs_shard5_replica_n18/, failed
org.apache.solr.client.solrj.impl.HttpSolrClient$RemoteSolrException: Error from server at http://ip-10-241-10-72:8886/solr/hadoop_logs_shard5_replica_n18: Expected mime type application/octet-stream but got text/html. <html>
<head>
<meta http-equiv="Content-Type" content="text/html;charset=utf-8"/>
<title>Error 403 GSSException: Failure unspecified at GSS-API level (Mechanism level: Invalid argument (400) - Cannot find key of appropriate type to decrypt AP REP - RC4 with HMAC)</title>
</head>
<body><h2>HTTP ERROR 403</h2>
<p>Problem accessing /solr/hadoop_logs_shard5_replica_n18/get. Reason:
<pre> GSSException: Failure unspecified at GSS-API level (Mechanism level: Invalid argument (400) - Cannot find key of appropriate type to decrypt AP REP - RC4 with HMAC)</pre></p>
</body>
</html>
at org.apache.solr.client.solrj.impl.HttpSolrClient.executeMethod(HttpSolrClient.java:607) ~[solr-solrj-7.4.0.jar:7.4.0 9060ac689c270b02143f375de0348b7f626adebc - jpountz - 2018-06-18 16:55:14]
at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:255) ~[solr-solrj-7.4.0.jar:7.4.0 9060ac689c270b02143f375de0348b7f626adebc - jpountz - 2018-06-18 16:55:14]
at org.apache.solr.client.solrj.impl.HttpSolrClient.request(HttpSolrClient.java:244) ~[solr-solrj-7.4.0.jar:7.4.0 9060ac689c270b02143f375de0348b7f626adebc - jpountz - 2018-06-18 16:55:14]
at org.apache.solr.client.solrj.SolrClient.request(SolrClient.java:1219) ~[solr-solrj-7.4.0.jar:7.4.0 9060ac689c270b02143f375de0348b7f626adebc - jpountz - 2018-06-18 16:55:14]
at org.apache.solr.handler.component.HttpShardHandler.lambda$submit$0(HttpShardHandler.java:172) ~[solr-core-7.4.0.jar:7.4.0 9060ac689c270b02143f375de0348b7f626adebc - jpountz - 2018-06-18 16:55:13]
at java.util.concurrent.FutureTask.run(FutureTask.java:266) ~[?:1.8.0_112]
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) ~[?:1.8.0_112]
at java.util.concurrent.FutureTask.run(FutureTask.java:266) ~[?:1.8.0_112]
at com.codahale.metrics.InstrumentedExecutorService$InstrumentedRunnable.run(InstrumentedExecutorService.java:176) ~[metrics-core-3.2.6.jar:3.2.6]
at org.apache.solr.common.util.ExecutorUtil$MDCAwareThreadPoolExecutor.lambda$execute$0(ExecutorUtil.java:209) ~[solr-solrj-7.4.0.jar:7.4.0 9060ac689c270b02143f375de0348b7f626adebc - jpountz - 2018-06-18 16:55:14]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) [?:1.8.0_112]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) [?:1.8.0_112]
at java.lang.Thread.run(Thread.java:745) [?:1.8.0_112]
... View more
01-30-2019
12:08 PM
1 Kudo
@Yee Zee Please refer to the previously shared detailed steps to test this scenario. This works for any Abrupt Termination of the component (either due to host reboot or if somehow the process was killed abruptly). As mentioned earlier that Auto start of a component is based on its current state and "desired state". Which means if you stop a component using Ambari API Or using Ambari UI then you wont see the recovery because in that case the ambari DB will have the correct info.
... View more
01-21-2019
09:27 AM
thank you so much @Jay Kumar SenSharma for good answer. I wanted to make changes the file cluster-env.xml that auto start works properly. The way you said it changed the blueprint and it's perfectly correct, but my services are still starting after rebooting the system after about 40 minutes. can you tell me what i do step by step for active service auto start?
... View more
01-21-2019
12:37 PM
@Jay , after we do the decommission on some datanode , do we need also to stop the components on that datanode ? and then replaced the disk , or maybe it is inoufgh to decomission without to stop the componet and then replaced the disk ?
... View more
01-15-2019
04:51 AM
@Magical Jelly Good to know that the issue is resolved. It will be wonderful if you can mark this HCC thread as Answered by clicking on the "Accept" link then other HCC users can quickly find the answered threads.
... View more
01-06-2019
09:43 AM
Also there are too many alerts automatically after starting,,.. So is it normally started or abnormally ? @Jay Kumar SenSharma
... View more