Member since
01-05-2016
60
Posts
42
Kudos Received
7
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
150 | 10-18-2024 02:15 PM | |
1154 | 10-21-2019 05:16 AM | |
4392 | 01-29-2018 07:05 AM | |
3203 | 06-27-2017 06:42 AM | |
39124 | 05-26-2016 04:05 AM |
10-18-2024
02:15 PM
1 Kudo
Solved by: - Deleting the Namenode role entirely from "HDFS --> Instances --> NameNode", clicking the checkbox near the Namenode Instance and selecting "delete" from the "Actions" dropdown menu - Redeploying a new "Namenode" role on the same host where the primary Namenode was previously running - Reenabling High Availability on the Namenode - Reconstructing the Namenode Metadata in the "Hive" service, under "Hive --> actions --> rebuild Namenode metadata"
... View more
10-18-2024
09:58 AM
1 Kudo
Hi, here is a summary of our blocking situation on our CDH Express 6.3.1 installation: - A crash happened on the storage where all the virtual nodes composing our Cluster insist - After solving the issue on the storage, our Cloudera cluster started again working, but with a critical error on HDFS Service Nodename - In fact, we have High Availability configured on 2 nodes, and exploring the situation we saw that Namenode 2 was active and Namenode 1 was DOWN - But the cluster worked - In an attempt to fix the issue, we tried several times to manually start the Namenode 1 from "HDFS --> Instances --> federation --> Namenode1 --> Start" but it was impossible to start manually - So, we decided to disable High Availability but the new configuration has not completed and actually is in pending status - It's in fact impossible to start HDFS service at all because an error pops up saying that "Nameservice nameservice1 has no SecondaryNameNode or High-Availability partner" It is interesting to note that (see attached picture) apparently we don't have High Availability configured anymore but cluster does not start anyhow As you can see from the second screenshot, if we try to manually start HDFS service a configuration issue is shown and it still believes to be running in High Availability mode and still looking for a "partner node" QUESTION: - Is there a way to force a NON-High-availability state and start the HDFS Service? Thank you for any advice and/or insights
... View more
Labels:
- Labels:
-
HDFS
07-29-2024
03:05 AM
1 Kudo
Thank you again @Rajat_710 . I Asked the Customer to revert the changes the OS Team performed at OS level. At this point, the Stale Configurations were still showing as pending but following a reboot of the "spurious" machine they cleared up. The situation is back to normal. I'm grateful for your valuable insight!
... View more
07-29-2024
12:49 AM
1 Kudo
Thank you so much, I'm going to dig more in deep in this direction and update you on the findings!
... View more
07-28-2024
06:55 AM
1 Kudo
Hello, a customer has a CDH Express 6.3.1 setup and: - It appears that suddenly one (and just one!) of the hosts listed in Hosts --> All Hosts is now showing as having a complete FQDN. It never used to have it (I have screenshots of previous configurations where ALL hosts in the list are shown with just the hostname and no domain) - Also, there are Stale Configurations waiting to be applied. If I click on them, they are related to a number of Client Configuration files waiting to be deployed, and sure enough ALL the lines in ALL the files are showing a single difference for that particular host! Now, I don't know what happened here. Nobody is reporting a CM configuration change made by them. I don't have History/Rollback capabilities on this cluster, as it's an Express Edition. Also, I have no Audits as well to check what (if any?) parameters have been recently changed I just know that Sysadmins recently performed a change in FQDN names on those hosts at OS level, because a monitoring tool (Zabbix) required it. But according to them nothing has been changed at CM Gui level My questions: - How can I manually check what (if any?) parameters have been recently changed in CM Gui? If I'd be able to identify a change/changes where the FQDN has been put in for that host, I could manually overwrite it - Or maybe the problem could be that CM "detected" a change at OS level (that config made by the sysadmins) and is now proposing a configuration change automatically? I really doubt it, but... Is it even a possibility? Thank you for any advice/help/support. Unfortunately I'm not directly managing that cluster so any requests for screenshots/checks on the system would receive slow replies... I have to ask the Customer. Knowing that, I eventually encourage to simply point out what configurations in CM can lead to change in single host FQDN changes being distributed in the cluster as client configurations
... View more
Labels:
- Labels:
-
Cloudera Manager
05-26-2020
05:01 AM
Hi, as we all know CDH 6.3.3 and subsequent versions are not available anymore under "Express licensing" model. Yet, I was in the process of setting up a 6.3.1 installation and apparently this is not possible, because a valid authentication (enterprise) is required. What am I doing wrong? Below, a screenshot from the Installation Guide: But when I try to import the 6.3.1 Repository (WITHOUT Username and Password as stated on the documentation itself for version < 6.3.3 I get the following error: # rpm --import https://archive.cloudera.com/p/cm6/6.3.1/redhat7/yum/RPM-GPG-KEY-cloudera
curl: (22) The requested URL returned error: 401 Authentication required
errore: https://archive.cloudera.com/p/cm6/6.3.1/redhat7/yum/RPM-GPG-KEY-cloudera: lettura importazione fallita(2). What am I doing wrong? @Cloudera1 Is this a mistake I'm committing or is it a problem with Cloudera implementing their policies in a wrong way? Thanks for any insights
... View more
Labels:
10-21-2019
05:16 AM
1 Kudo
You can query the API exposed by Cloudera Manager and simplify your life. For example, you can run the following: curl -u <CM_USER>:<CM_PASSWD> http://<CM_IP_ADDRESS>:7180/api/v19/clusters/<CLUSTER_NAME>/services/hive2 You'll get a Json answer in reply to your Query, with all the details related to the desired service's status. You can finally parse your Json answer (e.g. using "jq" or directly inside your bash script) and take the desired actions HTH
... View more
01-29-2018
07:26 AM
1) Apparently, yes 2) The name of the user you're trying to use to log in to the remote system, I suppose. Pls note that the user you specify here would be the user "oozie" will run as, so you'd eventually get other problems of unpredictable nature when using Oozie 3) I don't really know, sorry about that... The fact is that even if I'm pretty sure to have understood the cause of your issue, I never had to deal with it directly myself. Maybe the easier way could be to follow the additional suggestions I wrote in my previous answer (give permissions to OS User "yarn" to "ssh" and/or "su"). Or, maybe, another possibility would be for you to create a "yarn" user on the remote system and grant this user with the correct permissions to get to the final working directory I hope you'll manage to get through the problems and make it 🙂
... View more
01-29-2018
07:05 AM
This is probably related to the fact that the shell action, when run from Oozie, runs as user "yarn" and not as the desired user you're specifying in the ssh command You can refer to this thread for more information about the issue: https://community.cloudera.com/t5/Batch-Processing-and-Workflow/How-to-run-Oozie-workfllow-or-action-as-another-user/td-p/26794 It should all boil down (in case your cluster is not secured with kerberos) to try and set up your environment, specifically the "linux-container-executor" configuration parameter (you go in Cloudera Admin UI --> Yarn --> Configuration). It's all explained in the linked document. Another alternative could be to grant OS user "yarn" with permissions to execute "ssh" and/or "su" so you can eventually switch user in your script before executing the remote ssh command HTH
... View more