Member since
10-28-2020
622
Posts
47
Kudos Received
40
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1999 | 02-17-2025 06:54 AM | |
6713 | 07-23-2024 11:49 PM | |
1337 | 05-28-2024 11:06 AM | |
1891 | 05-05-2024 01:27 PM | |
1268 | 05-05-2024 01:09 PM |
06-26-2023
06:17 AM
@aafc could you please share the CDP (HIve) version you are using? Also, have you tried with the latest Clouldera JDBC driver for hive?
... View more
06-24-2023
11:53 AM
@rahuledavalath In a 3-node ZK cluster, we can only afford to lose 1 instance/node at once. However, having this cluster across two geo-locations could put us in a difficult position. So, it will be a good idea to build the cluster across 3 regions, with one instance in each.
... View more
06-21-2023
06:39 AM
@Choolake Try : entries=$((count2-count1)) This should work provided we have valid values on both variables.
... View more
06-14-2023
10:47 PM
@xiamu this error could appear if the data nodes are not healthy. Does the job fail repeatedly, or it succeeds at times? Have you tried running it with a different user? This is where it is failing: private void setupPipelineForAppendOrRecovery() throws IOException {
// Check number of datanodes. Note that if there is no healthy datanode,
// this must be internal error because we mark external error in striped
// outputstream only when all the streamers are in the DATA_STREAMING stage
if (nodes == null || nodes.length == 0) {
String msg = "Could not get block locations. " + "Source file \""
+ src + "\" - Aborting..." + this;
LOG.warn(msg);
lastException.set(new IOException(msg));
streamerClosed = true;
return;
}
setupPipelineInternal(nodes, storageTypes, storageIDs);
}
... View more
06-14-2023
10:21 PM
@haihua Do you mean it works from beeline but not from Hive CLI? If it works with beeline, why don't we run it run it with that instead? beeline ... -f query.hql Also, could you try providing required privileges to the role with GRANT OPTION ? Refer to https://docs.cloudera.com/documentation/enterprise/latest/topics/sg_hive_sql.html#grant_privilege_with_grant
... View more
06-09-2023
08:47 AM
@snm1523 From beeline use sys;
select cd_id, count(cd_id) as column_count from columns_v2 group by cd_id order by cd_id asc; -- this will return column_count for each table Every individual table will have a unique cd_id. To map the table names with cd_id, try the following. select t.tbl_name, s.cd_id from tbls t join sds s where t.sd_id=s.sd_id order by s.cd_id asc; You could also merge the two queries to get the o/p together.
... View more
06-09-2023
05:55 AM
@BrianChan Does the AD user have permissions to view the Hive table? This privilege need to be set under Hadoop SQL policies in Ranger. If this is already done, you also need to set required atlas policies. Refer to https://docs.cloudera.com/cdp-private-cloud-base/7.1.6/atlas-securing/topics/atlas-configure-ranger-authorization.html Make sure that the entiity-type is set correctly.
... View more
06-09-2023
05:36 AM
@snm1523 It should be SET FILEFORMAT INPUTFORMAT... Please try it as follows. ALTER TABLE alter_file_format_test SET FILEFORMAT INPUTFORMAT 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat' OUTPUTFORMAT 'org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat' SERDE 'org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe';
... View more
05-20-2023
11:37 PM
@mwblee I am not sure if you are using any Cloudera Hive distribution, if you are, consider upgrading to latest CDP version where we have fixed many issues around compactor initiator/worker/cleaner. e.g. For initiator: upstream Jiras - HIVE-21917, HIVE-22568, HIVE-22081 For this specific issue, you may check take a look at multiple factors, such as Hive metastore being overloaded, slow/large(certain txn related tables) metastore database. You may enable DEBUG logging in Hive metastore, and this will provide more information on why/where the compactor is stuck. If you are using opensource Hive, upgrade to Hive 4.x; you will have much better experience w.r.t. compaction.
... View more
05-12-2023
02:10 AM
@quangbilly79 cloudera-manager-server, daemon and agent package needs to be installed in one node that will host your Cloudera Manager UI. Please refer to the Installation guide and follow the step by step procedure. You also need to setup the manager DB as you have mentioned above. Start the Manager and access UI. Once the Manager is accessible, you may add the other nodes by specifying the IP addresses as described here : https://docs.cloudera.com/documentation/enterprise/6/6.3/topics/install_software_cm_wizard.html
... View more