Member since
08-19-2013
392
Posts
29
Kudos Received
9
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2545 | 09-12-2019 01:04 PM | |
2419 | 08-21-2019 04:56 PM | |
8211 | 07-03-2018 07:59 AM | |
5865 | 10-09-2015 08:02 AM | |
2522 | 04-29-2015 12:14 PM |
07-12-2021
11:57 AM
Hi Tony, I will reach out via private message
... View more
11-06-2019
08:25 AM
1 Kudo
@sundar_gampa Try a different browser or a bigger monitor. Apparently, the page design makes the page adaptive to the viewing device.
... View more
09-04-2019
01:28 PM
Is there a reason you can share, or a link to something that explains why disabling SELinux is necessary, and why we can't have nice things?
... View more
08-21-2019
05:24 PM
When your first post titled "Ambari-server start error" was caught by the spam filter, you created a second post with the same content. I approved one of them so we didn't have duplicate questions on the board.
I am unsure why your posts were flagged as spam. Our Community Admins are still tuning the spam filter rules after merging into the Unified Cloudera Community.
Our moderators will get posts out of spam quarantine as quickly as we are able. I apologize again for the inconvenience.
... View more
05-22-2019
06:35 AM
Since you were able to access the Cloudera repo, you should be able to install the Oracle JDK on the agent with:
sudo apt-get install oracle-j2sdk1.8
If that does not work, you can install it with the instructions under Manually Installing OpenJDK
When starting, the agent looks through common paths for JDKs and will select the Oracle 1.8 JDK first if it finds it.
If you would prefer to use OpenJDK on all the systems, in Cloudera Manager navigate to Hosts > Hosts Configuration and set the Java Home Directory to your preferred JAVA_HOME.
... View more
03-15-2019
02:27 PM
@Somanath, Based on my code review and testing, the original logging that was provided in this thread is caused by a minor bug in CM 5.12 and higher only when single-user mode is configured or the agent is not running as root. I opened a new internal Cloudera Jira for this issue: OPSAPS-49735. In my case, though, even though I reproduced the errors, this did not prevent the Zookeeper server from starting. I think it would be advised that you still review the logs to make certain of the cause of the server failing to start. On any host showing the "UnboundLocalError: local variable 'mdata' referenced before assignment" error: (1) Back up your os_ops.py file so you can role back if required Assuming you have Python 2.7 like was posted in the error in this thread, you can find the os_ops.py file here: /usr/lib64/cmf/agent/build/env/lib/python2.7/site-packages/cmf-5.16.1-py2.7.egg/cmf/util/os_ops.py prompt> cd /usr/lib64/cmf/agent/build/env/lib/python2.7/site-packages/cmf-5.16.1-py2.7.egg/cmf/util/ prompt> cp ./os_ops.py ./os_ops.py.original (2) Edit os_ops.py to move "mdata = self.get_path_metadata(path)" before the "if" conditional: prompt> vim os_ops.py Locate the following block of code in mkabsdir(): if os.path.isdir(path):
# Log warnings if user/group/mode are different than what's expected
if self.honor_users_and_groups:
mdata = self.get_path_metadata(path) Move this line above "if self.honor_users_and_groups:": mdata = self.get_path_metadata(path) The result should look like this: if os.path.isdir(path):
# Log warnings if user/group/mode are different than what's expected mdata = self.get_path_metadata(path)
if self.honor_users_and_groups:
if user is not None and user != mdata.user:
LOG.warning('Expected user %s for %s but was %s', user, path, mdata.user)
if group is not None and group != mdata.group:
LOG.warning('Expected group %s for %s but was %s', group, path, mdata.group)
if mode is not None and oct(mode) != mdata.mode:
LOG.warning('Expected mode %s for %s but was %s', oct(mode), path, mdata.mode)
return False Save your edits This change will make sure that "mdata" is assigned a value before it is referenced. (3) Restart the agent on the host where you updated os_ops.py: prompt> systemctl restart cloudera-scm-agent or on el6 oses: prompt> service cloudera-scm-agent restart (4) If the agent does not restart and it cites some python problem, you can revert by copying the "os_ops.py.original" file to overwrite the "os_ops.py" file you edited. Restart after that.
... View more
03-07-2019
02:34 AM
Hello Experts, I finally found and resolved the problem. I destroyed all the VM and built new ones. I kept only the database VM on which I droped the database scm and re-created it (to empty it). In our company, we use proxy to go on Internet, direct access is forbidden. So I re-wrote the repo files in /etc/yum.repos.d to put proxy=http://<proxy_user>:<proxy_password>@<proxy_url>:8080/ after each repo definition. I canceled in /etc/yum.conf all proxy definition. The repos for cloudera-cm and cloudera-cdh were written with the local repo I created and filled with the good packages. I ran yum update to get rpg-key (for epel-release). The only disadvantage with this method is : no parcels can be used (the proxy definition in the wizard crashed the download (timeout connexion)). Thanks for your help. Alain
... View more
07-27-2018
06:05 AM
Hi, We have Kerberos configured in our Hadoop cluster. We did a Wizard installation (https://www.cloudera.com/documentation/enterprise/5-6-x/topics/cm_sg_intro_kerb.html), it works well. We try to have a high level of availability, we have configured a secondary kdc-server (we followed the kerberos documentation). We have a replication of the credentials from the first Kerberos server to the second (like in the topic : https://community.hortonworks.com/articles/92333/configure-two-kerberos-kdcs-as-a-masterslave.html) We set Kerberos configuration on Cloudera Manager to add the secondary kdc server. The configuration generate by Cloudera in /etc/krb5.conf contains : [realms]
XXXXXX.COM = {
kdc = master1.com
admin_server = master1.com
kdc = worker1.com
} We have the following configuration: master1 : Kerberos server + Namenode (active) HDFS worker1 : Kerberos server + Namenode HDFS worker2 : Kerberos client + Datanode HDFS We are testing the replication of Kerberos. Case 1 : stop Kerberos server (kdc + kadmin) on master1 and init user ticket on worker2 with kinit It works well. Case 2 : stop Kerberos server (kdc + kadmin) and Namenode HDFS on master1 (to simulate the crash of the server master1) Normaly, the Namenode on worker1 should be activate. But, there is an error : "This role's process exited. This role is supposed to be started." on worker1. Message in log: PriviledgedActionException as:hdfs/worker1.com@XXXXXX.COM (auth:KERBEROS) cause:java.io.IOException: org.apache.hadoop.security.authentication.client.AuthenticationException: GSSException: No valid credentials provided (Mechanism level: Connection refused (Connection refused)) Conclusion/Question So my conclusion is that the Namenode on worker1 doesn't use the secondary kdc (there is nothing in the kadmin.log on the worker1). But if I do a kinit manually, that works. So, is not a problem of Kerberos. If the server with the main Kerberos kdc crash, the hadoop services crash too.. This is a big problem. Do you have a solution? Or any suggestion? Thank you, Martin.
... View more
07-04-2018
04:23 PM
Thank you. it would be good if it comes as integrated with QuickStart VM for the future release.
... View more