Member since
01-19-2017
3652
Posts
623
Kudos Received
364
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
176 | 12-22-2024 07:33 AM | |
113 | 12-18-2024 12:21 PM | |
442 | 12-17-2024 07:48 AM | |
298 | 08-02-2024 08:15 AM | |
3584 | 04-06-2023 12:49 PM |
05-29-2024
08:11 PM
what are you setting in /etc/hostname ? the same issue in my setup Kerberos
... View more
01-12-2024
01:40 AM
Hello, sorry but due to the urgency I abandoned the migration of the components and finally reinstalled everything. I now have a hbase table export problem and submitted a new topic. If anyone here would take the time to help me find a solution, that would be very kind.
... View more
11-08-2023
04:02 PM
@KundanRND Welcome to the Cloudera Community! As this is an older post, you would have a better chance of receiving a resolution by starting a new thread. This will also be an opportunity to provide details specific to your environment that could aid others in assisting you with a more accurate answer to your question. You can link this thread as a reference in your new post. Thanks.
... View more
10-01-2023
06:41 AM
@Shelton I could not start namenode with ambari ui when setting safemode to off mode. Please could you helpme out! And also while setting safemode to 'ON' does not allow me to do task.
... View more
09-16-2023
08:51 AM
Hi, You need to set in hive-site.xml, these three tags to get this working with hive: <property>
<name>google.cloud.auth.service.account.json.keyfile</name>
<value>/home/hadoop/keyfile.json</value>
</property>
<property>
<name>fs.gs.reported.permissions</name>
<value>777</value>
</property>
<property>
<name>fs.gs.path.encoding</name>
<value>/home/hadoop/</value>
</property> Same xml tags we can have it on hadoop in core-site.xml to have it working with hdfs, On beeline, just execute this and it shall work: INSERT OVERWRITE DIRECTORY 'gs://bucket/table' ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' SELECT * FROM table; Please upvote if you found helpful!
... View more
09-13-2023
07:00 AM
Hello @hebamahmoud If the issue is has been from any of the above responses, Could you accept it as a solution? Regards, Chethan YM
... View more
06-02-2023
04:22 PM
1 Kudo
This response is NOT to fix "files with corrupt replicas" but to find and fix files that are completely corrupt, that is that there are not good replicas to recover the files. The warning of files with corrupt replicas is when the file has at least one replica which is corrupt, but the file can still be recovered from the remaining replicas. In this case hdfs fsck /path ... will not show these files because it considere these healty. These files and the corrupted replicas are only reported by the command hdfs dfsadmin -report and as far as I known there is no direct command to fix this. Only way I have found I to wait for the Hadoop cluster to health itself by reallocating the bad replicas from the good ones.
... View more
04-22-2023
07:48 PM
1 Kudo
I got it. you can do like this. link:https://cwiki.apache.org/confluence/display/ZOOKEEPER/Server-Server+mutual+authentication
... View more
04-21-2023
02:58 PM
I'm also getting same error when my spark application is trying to connect hbase. Found no valid authentication method from options. @tarekabouzeid91 @Ninads are you able to find the solution to fix this issue?
... View more
04-21-2023
07:27 AM
@Shelton, wow apologies, now I´m reading, excellent test. Could you do testing with TLS/SSL? I still haven`t solved this.. 😞 Thank you for your support.
... View more