Member since
11-01-2019
146
Posts
5
Kudos Received
4
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1249 | 09-26-2025 08:06 AM | |
| 2183 | 03-04-2025 08:16 AM | |
| 3191 | 03-23-2023 03:30 PM | |
| 3799 | 02-01-2023 12:44 PM |
01-09-2026
10:46 PM
@Hadoop16 FYI ➤ This error occurs because of a token delegation gap between Hive and the HDFS Router. In a Kerberized cluster, when Hive (running on a DataNode/Compute node via Tez or MapReduce) attempts to write to HDFS, it needs a Delegation Token. When you use an HDFS Router address, Hive must be explicitly told to obtain a token specifically for the Router's service principal, which may be different from the backend NameNodes. ➤ The Root Cause The error Client cannot authenticate via:[TOKEN, KERBEROS] at the FileSinkOperator stage indicates that the tasks running on your worker nodes do not have a valid token to "speak" to the Router at router_host:8888. When Hive plans the job, it usually fetches tokens for the default filesystem. If your fs.defaultFS is set to a regular NameNode but your table location is an RBF address, Hive might not be fetching the secondary token required for the Router. ➤ The Fix: Configure Token Requirements You need to ensure Hive and the underlying MapReduce/Tez framework know to fetch tokens for the Router's URI. 1. Add the Router URI to Hive's Token List In your Hive session (or globally in hive-site.xml), you must define the Router as a "known" filesystem that requires tokens. SET hive.metastore.token.signature=hdfs://router_host:8888; SET mapreduce.job.hdfs-servers=hdfs://router_host:8888,hdfs://nameservice-backend; 2. Configure HDFS Client to "Trust" the Router for Tokens In core-site.xml or hdfs-site.xml, you need to enable the Router to act as a proxy for the backend NameNodes so it can pass the tokens correctly. <property> <name>dfs.federation.router.delegation.token.enable</name> <value>true</value> </property> ➤ Critical Kerberos Configuration Because the Router is an intermediary, it must be allowed to impersonate the user (Hive) when talking to the backend. Ensure your ProxyUser settings in core-site.xml include the Router's service principal. Assuming your Router runs as the hdfs or router user: <property> <name>hadoop.proxyuser.router.groups</name> <value>*</value> </property> <property> <name>hadoop.proxyuser.router.hosts</name> <value>*</value> </property> ➤ Diagnostic Verification To prove if the token is missing, run this command from the datanode_host mentioned in your error logs using the same user running the Hive job: # Check if you can manually get a token for the router hdfs fetchdt --renewer hdfs hdfs://router_host:8888 router.token # Check the contents of your current credentials cache klist -f If fetchdt fails, the issue is with the Router's ability to issue tokens. If it succeeds but Hive fails, the issue is with Hive's Job Submission not including the Router URI in the mapreduce.job.hdfs-servers list.
... View more
10-13-2025
10:41 AM
Hi @ishashrestha , Yes, the user needs to have permission to write to that directory. You can test to execute with another user that already have that permissions. Let me know if works. Best Regards
... View more
10-03-2025
09:49 AM
@champa Has the reply helped resolve your issue? If so, please mark the appropriate reply as the solution, as it will make it easier for others to find the answer in the future. Thanks.
... View more
09-23-2025
09:40 AM
Hi @Jaguar , This issue seems to be related with Knox. Service Mismatch: The URL in the error (.../gateway/dt/knoxtoken/api/v1/token) suggests a conflict between the Hive Metastore and the Knox Token Service. This often happens after a cluster upgrade where a new token service is implemented, but the clients (in this case, the Metastore) are still configured for the old one. Solution: Check the Knox configuration in your cluster management tool (e.g., Cloudera Manager). Verify that the HiveServer2 and Hive Metastore services are using the correct Knox topology and that the token service settings are correctly configured to match the Knox server. Let me know if this helps.
... View more
07-07-2025
08:38 AM
Adding a LoadBalancer to Hue. Solved the problem. Thanks
... View more
07-04-2025
11:19 AM
@Jackallboy Has the reply helped resolve your issue? If so, please mark the appropriate reply as the solution, as it will make it easier for others to find the answer in the future. If you are still experiencing the issue, can you provide the information requested? Thanks.
... View more
04-10-2025
10:44 PM
To provide the exact HQL query, Please share the following : DDL for both the tables Sample records from each table The expected output based on the sample data. The above information will help to understand the problem statement better and validate the solution.
... View more
03-12-2025
09:12 AM
@rosejo Has the reply helped resolve your issue? If so, please mark the appropriate reply as the solution, as it will make it easier for others to find the answer in the future. Thanks.
... View more
11-22-2024
10:27 AM
1 Kudo
Hi @LucasM , could you describe more of your use case? Your server in GCP have access to Cloudera CML? It's in the same network? Best Regards
... View more
08-29-2023
07:40 PM
create connection between Dbeaver and Hive on a CDP 7.1.8 centos 7 cloudera 7.1.8 i am also facing same issue in cluster Kerberos is enabled
... View more