Member since
03-11-2020
197
Posts
30
Kudos Received
40
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
2133 | 11-07-2024 08:47 AM | |
1504 | 11-07-2024 08:36 AM | |
1052 | 06-18-2024 01:34 AM | |
727 | 06-18-2024 01:25 AM | |
887 | 06-18-2024 01:16 AM |
04-12-2023
02:07 AM
@aleezeh Letus know if you also getting this error unable to find valid certification path to requested target. If yes there could be chances that kafka truststore does not have ranger admin certificate you can import the ranger admin certificate to kafka truststore. keytool -importcert -file /tmp/ranger.cer -keystore kafka_plugin_truststore.jks If you found that the provided solution(s) assisted you with your query, please take a moment to login and click Accept as Solution below each response that helped.
... View more
04-12-2023
01:59 AM
@aleezeh Can you please attach the log file to further investigate on this? What version of HDP or CDP you are using?
... View more
04-12-2023
01:48 AM
@rajilion It seems that you are using the -update flag with distcp command, which is causing the command to skip files that exist in the destination and have a modification time equal to or newer than the source file. This is the expected behavior of distcp when the -update flag is used. In your case, even though the content of the file has changed, the size and modification time are still the same, which is causing distcp to skip the file during the copy process. To copy the updated file to S3, you can try removing the -update flag from the distcp command. This will force distcp to copy all files from the source directory to the destination, regardless of whether they exist in the destination or not. Your updated command would look like this: hadoop distcp -pu -delete hdfs_path s3a://bucket The -pu flag is used to preserve the user and group ownership of the files during the copy process. Please note that removing the -update flag can cause distcp to copy all files from the source directory to the destination, even if they haven't been modified. This can be time-consuming and may result in unnecessary data transfer costs if you have a large number of files to copy. If you only want to copy specific files that have been modified, you can use a different tool such as s3-dist-cp or aws s3 sync that supports checksum-based incremental copies. These tools use checksums to determine which files have been modified and need to be copied, rather than relying on modification times or file sizes. If you found that the provided solution(s) assisted you with your query, please take a moment to login and click Accept as Solution below each response that helped.
... View more
04-12-2023
01:44 AM
You should be able to access cluster using below link. https://www.cloudera.com/campaign/try-cdp-public-cloud.html#:~:text=Try%20CDP%20Public%20Cloud%20for,hybrid%20and%20multi%2Dcloud%20data If you found that the provided solution(s) assisted you with your query, please take a moment to login and click Accept as Solution below each response that helped.
... View more
03-30-2023
04:11 AM
@jjjjanine Thanks for providing your valuable inputs.
... View more
03-27-2023
09:27 PM
I have checked this internally currently it's not supported for CDP as of now but we have an internal Jira to test and certify this in the future.
... View more
01-17-2023
12:01 AM
@fernando_lopez I'm checking on this internally and will keep you posted.
... View more
01-16-2023
11:50 PM
1 Kudo
The spark/yarn/hive jobs uses local directories for localization purpose. The required data which is used by the application is stored on local directories when the application is in execution. Please manually delete the older data from local /tmp directory. Also, Please follow the steps mentioned in following article to clear the cache memory from local directory : https://community.cloudera.com/t5/Community-Articles/How-to-clear-local-file-cache-and-user-cache-for-yarn/ta-p/245160 Please manually delete the unwanted data from local /tmp directory and also follow the above article If you found that the provided solution(s) assisted you with your query, please take a moment to login and click Accept as Solution below each response that helped.
... View more
01-16-2023
11:44 PM
@techfriend this can be resolved after modifiying the principle. WARNING: no policy specified for mapred/ip-172-31-46-169.us-west-2.compute.internal@HADM.RU; defaulting to no policy
add_principal: Principal or policy already exists while creating "mapred/ip-172-31-46-169.us-west-2.compute.internal@HADM.RU".
+ '[' 604800 -gt 0 ']'
++ kadmin -k -t /var/run/cloudera-scm-server/cmf5922922234613877041.keytab -p cloudera-scm/admin@HADM.RU -r HADM.RU -q 'getprinc -terse mapred/ip-172-31-46-169.us-west-2.compute.internal@HADM.RU'
++ tail -1
++ cut -f 12
+ RENEW_LIFETIME=0
+ '[' 0 -eq 0 ']'
+ echo 'Unable to set maxrenewlife'
+ exit 1 modprinc -maxrenewlife 90day +allow_renewable mapred/ip-172-31-46-169.us-west-2.compute.internal@HADM.RU
... View more
01-10-2023
05:47 AM
To resolve this issue, open /opt/cloudera/cm/bin/gen_credentials.sh for editing and add the following to the very end of the script: exit ${PIPESTATUS[0]} And then try generating Kerberos credentials once more.
... View more