Member since
09-16-2021
420
Posts
54
Kudos Received
38
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
175 | 09-05-2025 07:19 AM | |
598 | 07-15-2025 02:22 AM | |
1172 | 06-02-2025 06:55 AM | |
1455 | 05-22-2025 03:00 AM | |
810 | 05-19-2025 03:02 AM |
09-07-2025
11:04 PM
@lbourgeois, Did the response assist in resolving your query? If it did, please mark the relevant reply as the solution, as it will help others locate the answer more easily in the future.
... View more
08-19-2025
10:33 PM
Is the source table a JdbcStorageHandler table? Please provide the DDL of the source table, the query used, and any sample data if possible. This information will help us understand the problem better. Also, validate the set -v command, especially configurations like hive.tez.container.size.
... View more
07-24-2025
05:01 AM
SELECT * FROM your_table WHERE data_dt = '__HIVE_DEFAULT_PARTITION__'; ok thank's for your reply ,i understand this example But my question is, in my Third SQL statement, using length() function in the select statement is effective and the result is 26, while in the first and second SQL , why can't this record be filtered by length in the where statement when I use the length function on the partition key? Thank You
... View more
07-17-2025
09:50 AM
I will check with the development team to see where they are failing with the Hive 4 support and let you know. Thanks
... View more
07-15-2025
08:50 AM
@LSIMS PageWriter.java#L99 is where your operation is failing We need simple steps to reproduce the same error on a brand new environment to investigate further
... View more
07-15-2025
02:22 AM
https://mvnrepository.com/artifact/org.apache.hadoop/hadoop-common/3.1.1.3.1.5.0-152
... View more
07-04-2025
11:19 AM
@Jackallboy Has the reply helped resolve your issue? If so, please mark the appropriate reply as the solution, as it will make it easier for others to find the answer in the future. If you are still experiencing the issue, can you provide the information requested? Thanks.
... View more
06-02-2025
06:55 AM
It does looks like query failed with ClassCastException. It indicates that ( org.apache.hadoop.hive.serde2.io.HiveDecimalWritable cannot be cast to org.apache.hadoop.io.LongWritable ) a mismatch between the data type Hive expects and the data type it's actually encountering while processing the query. From the Error trace , Hive treats a value as a DECIMAL(HiveDecimalWritable) but the metadata seems to be Long(LongWritable). One possible Reason might be Schema Mismatch: Hive table schema defines a column but the underlying data file (e.g., Parquet, ORC, ...) for that column actually contains DECIMAL Values. To validate , Run DESCRIBE FORMATTED <your_table_name>; for the table involved in the failing query. Pay close attention to the data types of all columns, especially those that might be involved in the conversion. Compare these Hive schema data types with the actual data types in your source data files. For example, if you're using Parquet, use tools like parquet-tools to inspect the schema of Parquet files. if you're using ORC , use hive --orcfiledump to inspect the schema of orc files. Also make sure that Serde's pointing to valid underlying file formats.
... View more
05-26-2025
07:02 AM
Perfect, thanks four your help and quick responses.
... View more
05-26-2025
06:35 AM
Too simple 😋 I can't use clause IN as already written in my first post. Then my credential are "read only," I can't open a temporary tables.
... View more