Member since
01-19-2017
3681
Posts
633
Kudos Received
372
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1626 | 06-04-2025 11:36 PM | |
| 2085 | 03-23-2025 05:23 AM | |
| 988 | 03-17-2025 10:18 AM | |
| 3764 | 03-05-2025 01:34 PM | |
| 2591 | 03-03-2025 01:09 PM |
06-05-2024
12:51 AM
1 Kudo
@rizalt There are a couple of things to validate. Step 1 Pre-requisites Kerberos Server: Ensure you have a Kerberos Key Distribution Center (KDC) and an administrative server set up. DNS: Proper DNS setup is required for both forward and reverse lookups. NTP: Time synchronization across all nodes using Network Time Protocol (NTP). HDP Cluster: A running Hortonworks Data Platform (HDP) cluster. Step 2: Check your /etc/host file ensure your KDC host is assigned the domain HADOOP.COM to match your KDC credentials # hostname -f Step 3: Once that matches then edit the Kerberos configuration file (/etc/krb5.conf) on all nodes to point to your KDC you can scramble the sensitive info and share [libdefaults] default_realm = HADOOP.COM dns_lookup_realm = false dns_lookup_kdc = false [realms] HADOOP.COM = { kdc = kdc.hadoop.com admin_server = admin.hadoop.com } [domain_realm] .hadoop.com = HADOOP.COM hadoop.com = HADOOP.COM Step 4: Locate your kadm5.acl file and ensure it looks like this */admin@HADOOP.COM * Step 5: Restart the KDC and admin servers as root or with sudo # systemctl restart krb5kdc # systemctl restart kadmin Step 6: Check Kerberos Ticket: Ensure that the Kerberos ticket is obtained correctly. kinit -kt /etc/security/keytabs/hdfs.keytab hdfs/hostname@HADOOP.COM klist If your setup is correct you will see an output like below Ticket cache: FILE:/tmp/krb5cc_1000 Default principal: hdfs/hostname@HADOOP.COM Valid starting Expires Service principal 06/05/2024 09:50:21 06/06/2024 09:50:21 krbtgt/HADOOP.COM@HADOOP.COM renew until 06/05/2024 09:50:21 06/05/2024 09:50:22 06/06/2024 09:50:21 HTTP/hostname@HADOOP.COM renew until 06/05/2024 09:50:21 Hope that helps
... View more
01-09-2024
10:05 AM
@achemeleu Welcome Acemeleu, @DianaTorres for pinging me on this one. I provide 2 solutions see threads about a similar case Ambari stuck1 Ambari.stuck2 Can you check the above solution and see if that works out for you too? In case it doesn't please can you share your HDP version, database type/version, ambari-server logs and OS type/version, and the brief background of the steps you executed before getting stuck Please let us know whether that resolved your issue. Geoffrey
... View more
04-09-2023
09:06 AM
@YasBHK File /user/hdfs/data/file.xlsx could only be written to 0 of the 1 minReplication nodes. There are 1 datanode(s) running and 1 node(s) are excluded in this operation. So that means your data node is down can you restart the HDFS service and retry?
... View more
04-09-2023
03:26 AM
@YasBHK Please ensure both data nodes (2) are running. You definitely have an issue with one of the data nodes and because of your replication factor which I guess is 2 from the output the file /user/hdfs/data/file.xlsx can't be persisted if it can't meet the min replication of 2. Firstly understand why the second data node has been excluded by YARN either its space related issue or it just isn't started. Please check the dfs.hosts.exclude location usually in HDP /etc/hadoop/conf/dfs.exclude remove the host in the file and run the below hdfs dfsadmin -refreshNodes or from Ambari Ui just run the refresh nodes That should resolve the issue. Restart the faulty datanode and your HDFS put command will succeed
... View more
04-07-2023
12:47 PM
@Sanchari I suspect the java.io.IOException: Mkdirs failed to create is due to permissions on the edge-node Assuming you are the HDFS copy is being run as hdfs and your edge node directory belongs to a different user/group that. Just for test purposes can you do the following on the edgenode # mkdir -p /some_specific/path/in_edge_server Then run chmod on the destination path # chmod 777 /some_specific/path/in_edge_server Finally, rerun you spark-submit and let me know
... View more
04-07-2023
01:14 AM
@Sanchari It could be good to share a snippet of your code. logically I think you copy FROM -->TO Below is the function being used: fs.copyFromLocalFile(new Path(src_HDFSPath), new Path(dest_edgePath)) Disclaimer I am not a Spark/Python developer
... View more
04-06-2023
02:15 PM
@BrianChan You will need to manually perform the checkpoint on the faulty node. If the standby NameNode is faulty for a long time, generated edit log will accumulate. In this case, this will cause the HDFS or active NN to take a long time to restart and could even fail to restart because if the HDFS or active NameNode is restarted, the active NameNode reads a large amount of unmerged editlog. Is your NN setup active/standby? Fr the below steps you could as well use CM UI to perfom the tasks Quickest solution 1 I have had occasions when a simple rolling restart of the Zk's would resolve that biut I see the checkpoint lag goes to > 2 days Solution 2 Check the most up to date on both NN by comparing the dates of files in the directory. $ ls -lrt /dfs/nn/current/ On the Active NN with the latest editlogs as hdfs user $ hdfs dfsadmin -safemode enter $ hdfs dfsadmin -saveNamespace Check whether the latest generated fsimage timestamp is the current time. If yes, the combination is executed correctly and is complete. $ hdfs dfsadmin -safemode leave Before restarting the HDFS or active NameNode, perform a checkpoint manually to merge the metadata of the active NameNode. The restart the standby the newly generated files should now automatically be shipped and synced this could take a while < 5 minutes and your NN should all be green
... View more
04-06-2023
12:49 PM
@SSandhu The first question is I am just wondering if you have an HDP subscription. If not sure you could independently run the below to see whether the repo URL is valid and reachable Display enabled HDP software repositories # yum repolist Clean out all HDP packages and metadata from cache # yum clean all Refresh HDP packages on your system # yum update Reinstall AMC yum reinstall ambari-metrics-collector The above steps should help you resolve AMC does not exist in the stack-select package
... View more
03-29-2023
04:47 PM
2 Kudos
@vciampa I tried to recreate your scenarion using HDP and ODBC ODBC screens I created a hive database and table and used a non ssl connection as shown below as my HDP is not Secured, though the ODBS delivers a certificate in C:\Program Files\Cloudera ODBC Driver for Apache Hive\lib\cacerts.pem Ping from Windows DSN Config ODBC config SSL Config Table creation Connect test So it does work for non-TLS but when I enable TLS I get the below because TLS is not enabled on my HDP cluster , I will try in install sa self signed certificate and revert Geoffrey
... View more
03-24-2023
06:53 AM
@ambari275 These are the steps to follow assuming you are logged in as root # su - hdfs $ klist -kt /etc/security/keytabs/hdfs-headless.keytab Then the output should give you the principal to use $ kinit -kt /etc/security/keytabs/hdfs-headless.keytab
... View more