Member since
10-28-2020
396
Posts
18
Kudos Received
24
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
22 | 09-29-2023 08:59 PM | |
138 | 08-31-2023 03:16 AM | |
219 | 08-21-2023 03:08 AM | |
369 | 08-16-2023 01:06 PM | |
355 | 07-25-2023 07:29 AM |
09-29-2023
08:59 PM
@Srinivas-M You may set these properties in a safety valve for core-site.xml. CM UI > HDFS > Configuration > Cluster-wide Advanced Configuration Snippet (Safety Valve) for core-site.xml
... View more
09-20-2023
08:18 AM
@PetiaLeshiy Adding to @asish 's comment, as its a struct column, we could write the query something like this: SELECT * FROM TABLE_NAME LATERAL VIEW explode(struct_col_name.list_name) exploded_column AS xyz WHERE xyz IS NOT NULL; You may make changes where required.
... View more
08-31-2023
03:16 AM
we tried replicating the issue with the data shared by @Shivakuk . Left/Right Single/Double Quotation Mark(smart quotes) in the text did not show up correctly and got converted to ? . I was able to fix this issue by changing the LC_CTYPE from "UTF-8" to "en_US.UTF-8". Check "locale" command output: # locale
LANG=en_US.UTF-8
LC_CTYPE=en_US.UTF-8
LC_NUMERIC="en_US.UTF-8"
LC_TIME="en_US.UTF-8"
LC_COLLATE="en_US.UTF-8"
LC_MONETARY="en_US.UTF-8"
LC_MESSAGES="en_US.UTF-8"
LC_PAPER="en_US.UTF-8"
LC_NAME="en_US.UTF-8"
LC_ADDRESS="en_US.UTF-8"
LC_TELEPHONE="en_US.UTF-8"
LC_MEASUREMENT="en_US.UTF-8"
LC_IDENTIFICATION="en_US.UTF-8"
LC_ALL=
See what your LC_CTYPE read.
If it does not read "en_US.UTF-8", do the following:
vi ~/.bash_profile
Add the following two lines at the bottom:
+++
LC_CTYPE=en_US.UTF-8
export LC_CTYPE
+++
Save the file, and source it for it to take effect:
#source ~/.bash_profile
Now connect to beeline, and see if the data show up correctly.
... View more
08-29-2023
03:56 AM
1 Kudo
@AndreaCavenago For that you will have to check if the connection is getting interrupted/closed between the client and hiveserver2. Without thorough log analysis, it will be difficult to answer that. Could you open a support case for the same?
... View more
08-22-2023
03:03 AM
@Kaher Could you find out what's the value you have set for " tez.staging-dir"; if it's not set, the default path is /tmp/${user.name}/staging. Do verify if there is any issue with the /tmp filesystem. Also, there is a dash missing in the following value: <name>tez.am.java.opts</name>
<value>-Xmx2024m</value>
... View more
08-21-2023
03:08 AM
@itdm_bmi We are still seeing the error " java.lang.ClassNotFoundException: Class org.apache.hadoop.hive.contrib.serde2.MultiDelimitSerDe" This means, you have not uploaded the "hive-contrib" jar to Hive classpath as mentioned here. If you are on CDP(>7.1.5 onward) then, we will not have to upload hive-contrib jar to the class path. This serde class is already added to Hive native serde list. You just need to alter the original table, and change the serde(Note: only applies to CDP version > 7.1.5) ALTER TABLE <table name> SET SERDE 'org.apache.hadoop.hive.serde2.MultiDelimitSerDe' If you are on an older version, do consider uploading to hive-contrib jar to hive classpath by uploading this to aux jar path location.
... View more
08-16-2023
01:06 PM
@AndreaCavenago Does this error appear every time we run this spark-submit command? As this is a warning message, and it does not have any real impact, we can avoid it by changing the log level. In the script.py file, add the following two lines: from pyspark import SparkContext
SparkContext.setLogLevel("ERROR") This will avoid the WARN message. But it will still be good to address the actual issue.
... View more
08-02-2023
01:37 AM
@itdm_bmi This error message is very generic. We could not say what caused the job to fail. Is it possible for you to share the YARN application log of application_1690451589703_12741 ? You may collect it using the command : yarn logs -applicationId application_1690451589703_12741 > /tmp/application_1690451589703_12741.log Attach this application_1690451589703_12741.log file. If you are not comfortable sharing the log files here, you may file a support case. We'll be happy to assist.
... View more
07-25-2023
07:29 AM
@novice_tester when you create a database, it takes the external warehouse directory path from hive.metastore.warehouse.external.dir, and sets that as the database LOCATION for external tables only(when created without setting location clause). Please note, there is another MANAGEDLOCATION field that is blank. e.g. describe database extended question;
+-----------+----------+----------------------------------------------------+------------------+-------------+-------------+-------------+
| db_name | comment | location | managedlocation | owner_name | owner_type | parameters |
+-----------+----------+----------------------------------------------------+------------------+-------------+-------------+-------------+
| question | | hdfs://node2.cloudera.com:8020/warehouse/tablespace/external/hive/question.db | | hive | USER | |
+-----------+----------+----------------------------------------------------+------------------+-------------+-------------+-------------+ A Managedlocation can be set on a database if you want to use a different location for managed tables other than hive.metastore.warehouse.dir. Otherwise it is kept blank. I hope this helps. To summarize, you do not need to worry about the LOCATION of a database. It can host both managed and external tables. When you create a managed table under the same DB, it will pick the path set in hive.metastore.warehouse.dir, irrespective of the database Location, and when you create an external table without the location clause, it gets stored in the DB LOCATION path.
... View more
07-24-2023
10:22 AM
@novice_tester Could you please make it a bit clearer for us? What are the DDLs(create table command) you used to create the managed table, and the external table? Creating a managed table in any location outside of 'hive.metastore.warehouse.dir' path should prompt the following error: A managed table's location should be located within managed warehouse root directory or within its database's managedLocationUri. /warehouse/tablespace/managed/hive/ seems like the warehouse directory for the external tables. So, I doubt creating the managed table picked this location on its own. Could you also share the outputs of the following commands from beeline: beeline> set hive.metastore.warehouse.dir;
beeline> hive.metastore.warehouse.external.dir; An external table can be created with the LOCATION clause, and we can set any path w/ it. Refer to this Cloudera Doc.
... View more