Member since
08-05-2017
30
Posts
2
Kudos Received
1
Solution
My Accepted Solutions
Title | Views | Posted |
---|---|---|
4671 | 08-11-2017 11:29 AM |
12-22-2023
02:17 AM
@rlevas what if the password has changed? what steps do we need to take as a fix?
... View more
06-22-2018
02:03 AM
@Geoffrey Shelton Okot thanks for you reply. I have used the same database for reinstalling the Ranger. I was assuming that Ranger will delete all of its tables during the uninstallation
... View more
06-24-2018
07:56 PM
@Dassi Jean Fongang Unfortunately there is no FORCE command for decommissioning in Hadoop. Once you have the host in the excludes file and you run the yarn rmadmin -refreshNodes command that should trigger the decommissioning. It isn't recommended and good architecture to have a NameNode and DataNode on the same host (Master and Slave/worker respectively) with over 24 nodes you should have planned 3 to 5 master nodes and strictly have DataNode,NodeManager and eg Zk client on the slave (workernodes). Moving the NameNode to a new node and running the decommissioning will make your work easier and isolate your Master processes from the Slave this is the ONLY solution I see left for you. HTH
... View more
02-01-2018
11:13 AM
Hi @Krishnaswami Rajagopalan I don't know exactly the detail of this Sandbox in the Azure Cloud. Are you connecting to the Sandbox or to the docker container inside? The docker container is where zeppelin and other services are located. To connect to the docker container use the port 2222 in your SSH command. Example: ssh root@127.0.0.1 -p 2222 It doesn't matter I guess if your cluster or sandbox is running on the cloud. You should be able to find zeppelin under /usr/hdp/current/zeppelin-server Hope this helps. BR. Paul
... View more
10-18-2017
03:18 AM
@Jay SenSharma Thanks for helping out. Restarting the shellinabox service in the sandbox helped me to access the webssh.
... View more
08-11-2017
11:29 AM
2 Kudos
I am able to find the solution, I was running Ranger admin on host1 and Hiveserver on host2 and i created the admin OS account in host1 but not in host2(where hive server is running). Creating the admin account and its group (hdpmasters) in host2 resolved this issue. I guess Ambari views might need the OS account/ Group to be present on the server where the service(being accessed by view) is installed.
... View more
08-06-2017
11:29 AM
1 Kudo
@Chiranjeevi Nimmala Good to hear that your issue is resolved. It will be also wonderful if you can mark this thread as "Answered" (Accepted) so that it will be useful for other HCC users to quickly browse the correct answer for their issues.
... View more
07-15-2019
01:20 PM
Hi everyone! I installed the ambari server and ambari-agent to my laptop (SuSE Tumbleweed 2019) to set up a local single-node Hadoop cluster. (Server and agent versions 2.1.2.1-418) Server starts without issues, and I can open the console at http://localhost:8080 Agent says it started, but it can't connect the server, here's what the log says : ERROR 2019-07-15 11:06:06,862 NetUtil.py:77 - [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:727)ERROR 2019-07-15 11:06:06,862 NetUtil.py:78 - SSLError: Failed to connect. Please check openssl library versions.Refer to: https://bugzilla.redhat.com/show_bug.cgi?id=1022468 for more details. I have python (2.7.16-2.1) installed. Also python3 is installed, but the executable defaults to version 2.7 : ls -l /usr/bin/python lrwxrwxrwx 1 root root 9 Jun 24 22:12 /usr/bin/python -> python2.7 I would implement this "verify=disable" solution mentioned above, but there is no file named cert-verification.cfg in my system. There is no python dir under etc at all ! So I think maybe python under SuSE might have a different configuration than other distributions. I google'd a lot, but couldn't find a solution for this yet. Can someone guide me thru this issue please ?
... View more