Member since
10-28-2020
622
Posts
47
Kudos Received
40
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1993 | 02-17-2025 06:54 AM | |
6708 | 07-23-2024 11:49 PM | |
1335 | 05-28-2024 11:06 AM | |
1885 | 05-05-2024 01:27 PM | |
1267 | 05-05-2024 01:09 PM |
08-29-2023
03:56 AM
1 Kudo
@AndreaCavenago For that you will have to check if the connection is getting interrupted/closed between the client and hiveserver2. Without thorough log analysis, it will be difficult to answer that. Could you open a support case for the same?
... View more
08-22-2023
03:03 AM
@Kaher Could you find out what's the value you have set for "tez.staging-dir"; if it's not set, the default path is /tmp/${user.name}/staging. Do verify if there is any issue with the /tmp filesystem. Also, there is a dash missing in the following value: <name>tez.am.java.opts</name>
<value>-Xmx2024m</value>
... View more
08-21-2023
03:08 AM
@itdm_bmi We are still seeing the error "java.lang.ClassNotFoundException: Class org.apache.hadoop.hive.contrib.serde2.MultiDelimitSerDe" This means, you have not uploaded the "hive-contrib" jar to Hive classpath as mentioned here. If you are on CDP(>7.1.5 onward) then, we will not have to upload hive-contrib jar to the class path. This serde class is already added to Hive native serde list. You just need to alter the original table, and change the serde(Note: only applies to CDP version > 7.1.5) ALTER TABLE <table name> SET SERDE 'org.apache.hadoop.hive.serde2.MultiDelimitSerDe' If you are on an older version, do consider uploading to hive-contrib jar to hive classpath by uploading this to aux jar path location.
... View more
08-16-2023
01:06 PM
@AndreaCavenago Does this error appear every time we run this spark-submit command? As this is a warning message, and it does not have any real impact, we can avoid it by changing the log level. In the script.py file, add the following two lines: from pyspark import SparkContext
SparkContext.setLogLevel("ERROR") This will avoid the WARN message. But it will still be good to address the actual issue.
... View more
08-02-2023
01:37 AM
@itdm_bmi This error message is very generic. We could not say what caused the job to fail. Is it possible for you to share the YARN application log of application_1690451589703_12741 ? You may collect it using the command : yarn logs -applicationId application_1690451589703_12741 > /tmp/application_1690451589703_12741.log Attach this application_1690451589703_12741.log file. If you are not comfortable sharing the log files here, you may file a support case. We'll be happy to assist.
... View more
07-25-2023
07:29 AM
@novice_tester when you create a database, it takes the external warehouse directory path from hive.metastore.warehouse.external.dir, and sets that as the database LOCATION for external tables only(when created without setting location clause). Please note, there is another MANAGEDLOCATION field that is blank. e.g. describe database extended question;
+-----------+----------+----------------------------------------------------+------------------+-------------+-------------+-------------+
| db_name | comment | location | managedlocation | owner_name | owner_type | parameters |
+-----------+----------+----------------------------------------------------+------------------+-------------+-------------+-------------+
| question | | hdfs://node2.cloudera.com:8020/warehouse/tablespace/external/hive/question.db | | hive | USER | |
+-----------+----------+----------------------------------------------------+------------------+-------------+-------------+-------------+ A Managedlocation can be set on a database if you want to use a different location for managed tables other than hive.metastore.warehouse.dir. Otherwise it is kept blank. I hope this helps. To summarize, you do not need to worry about the LOCATION of a database. It can host both managed and external tables. When you create a managed table under the same DB, it will pick the path set in hive.metastore.warehouse.dir, irrespective of the database Location, and when you create an external table without the location clause, it gets stored in the DB LOCATION path.
... View more
07-24-2023
10:22 AM
@novice_tester Could you please make it a bit clearer for us? What are the DDLs(create table command) you used to create the managed table, and the external table? Creating a managed table in any location outside of 'hive.metastore.warehouse.dir' path should prompt the following error: A managed table's location should be located within managed warehouse root directory or within its database's managedLocationUri. /warehouse/tablespace/managed/hive/ seems like the warehouse directory for the external tables. So, I doubt creating the managed table picked this location on its own. Could you also share the outputs of the following commands from beeline: beeline> set hive.metastore.warehouse.dir;
beeline> hive.metastore.warehouse.external.dir; An external table can be created with the LOCATION clause, and we can set any path w/ it. Refer to this Cloudera Doc.
... View more
07-04-2023
11:55 AM
@Mannoj RDBMS HA for metastore is not officially supported yet. We have a knowledge article on the same. Cloudera's statement on RDBMS HA can be found here.
... View more
07-03-2023
01:35 PM
@vaibhavgokhale You could try: --conf spark.sql.hive.conf.list="tez.queue.name=queue1"
... View more
06-28-2023
11:31 AM
@Choolake Try this: count1=$(beeline -u "jdbc:hive2://dev-lisa.realm.com:10000/default;principal=hive/dev-lisa.intranet.slt.com.lk@REALM.COM;ssl=true;sslTrustStore=/var/lib/cloudera-scm-agent/agent-cert/cm-auto-global_truststore.jks" --showHeader=false --silent=true --outputformat=tsv2 -e 'SELECT count(*) from table_name;')
These beeline flags will remove all the unnecessary texts from the stdout. Compute count2 the same way.
... View more