Member since
03-06-2020
406
Posts
56
Kudos Received
36
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
909 | 11-21-2024 10:40 PM | |
878 | 11-21-2024 10:12 PM | |
2681 | 07-23-2024 10:52 PM | |
2003 | 05-16-2024 12:27 AM | |
6693 | 05-01-2024 04:50 AM |
08-29-2025
12:27 AM
@AEAT The log message "ORC read request to already read range. Falling back to readRandom" is a sign of a suboptimal read pattern. While not a fatal error, it means Impala is not reading the ORC file as efficiently as it could. Impala's ORC scanner is designed to read data in a sequential, read-ahead fashion to optimize I/O from HDFS. It attempts to predict what data a query will need next and reads it in large, efficient chunks. -> Random reads are slower than sequential reads on both spinning disks and SSDs. -> The process of seeking to a different location in the file and reading a small chunk of data consumes more CPU resources. -> The cumulative effect of these inefficient reads can add significant time to a query's execution, especially for large datasets. The most common cause of this issue is a large number of small files. Impala has to make many I/O requests to process each file, which can disrupt the efficient read pattern. Please check if you have such pattren of files and compress it as per the hdfs block size. Manually monitor the resources usage while running the query. Regards, Chethan YM
... View more
08-28-2025
09:19 PM
@Pratibha123 ORA-12705 - This is the key error. The Oracle JDBC driver needs to set up a language and character set session with the Oracle database. To do this, it attempts to read NLS configuration data from files on the local filesystem of the machine where Sqoop is running. This error occurs because it cannot find or access those NLS data files, or the environment variable that points to them is invalid. ORA-00604: This is a cascading error. It means an internal, recursive SQL statement that Oracle runs during the connection/session setup failed, because the session setup itself was incomplete. References: https://docs.oracle.com/en/error-help/db/ora-12705/?r=23ai https://stackoverflow.com/questions/7700330/ora-12705-cannot-access-nls-data-files-or-invalid-environment https://docs.oracle.com/en/error-help/db/ora-00604/?r=23ai https://stackoverflow.com/questions/30478070/how-to-solve-sql-error-ora-00604-error-occurred-at-recursive-sql-level-1 "oracle.jdbc.NLS_LANG" Is not seems to be valid property. Can you export it before running the job and check? > export NLS_LANG="AMERICAN_AMERICA.AL32UTF8" Ensure libjars has the correct ojdbc8.jar and it's compatible with your Oracle DB version. Regards, Chethan YM
... View more
08-05-2025
03:45 AM
1 Kudo
@Pratibha123 ORA-12705 - This is the key error. The Oracle JDBC driver needs to set up a language and character set session with the Oracle database. To do this, it attempts to read NLS configuration data from files on the local filesystem of the machine where Sqoop is running. This error occurs because it cannot find or access those NLS data files, or the environment variable that points to them is invalid. ORA-00604: This is a cascading error. It means an internal, recursive SQL statement that Oracle runs during the connection/session setup failed, because the session setup itself was incomplete. References: https://docs.oracle.com/en/error-help/db/ora-12705/?r=23ai https://stackoverflow.com/questions/7700330/ora-12705-cannot-access-nls-data-files-or-invalid-environment https://docs.oracle.com/en/error-help/db/ora-00604/?r=23ai https://stackoverflow.com/questions/30478070/how-to-solve-sql-error-ora-00604-error-occurred-at-recursive-sql-level-1 "oracle.jdbc.NLS_LANG" Is not seems to be valid property. Can you export it before running the job and check? > export NLS_LANG="AMERICAN_AMERICA.AL32UTF8" Ensure libjars has the correct ojdbc8.jar and it's compatible with your Oracle DB version. Regards, Chethan YM
... View more
05-07-2025
05:48 AM
1 Kudo
@Yigal It is not supported in Impala, Below is the Jira for your reference it is still in open state and not Resolved. https://issues.apache.org/jira/browse/IMPALA-5226 Regards, Chethan YM
... View more
05-05-2025
06:35 AM
Hi @Rich_Learner can you try this: "SELECT get_json_object(product_json, '$.ProductCOde') AS product_code, get_json_object(product_json, '$.Type') AS product_type FROM customer_table LATERAL VIEW json_tuple(json_column, 'Customer') c AS customer_json LATERAL VIEW json_tuple(customer_json, 'products') p AS products_json LATERAL VIEW explode(from_json(products_json, 'array<map<string,string>>')) product_table AS product_json ;"" OR WITH cleaned_json AS ( SELECT regexp_replace( regexp_replace( get_json_object(json_column, '$.Customer.products'), '\\}\\s*,\\s*\\{', '}~{' ), '\\[|\\]', '' ) AS flat_products FROM customer_table ), split_json AS ( SELECT split(flat_products, '~') AS product_array FROM cleaned_json ) SELECT get_json_object(item, '$.ProductCOde') AS product_code, get_json_object(item, '$.Type') AS product_type FROM split_json LATERAL VIEW explode(product_array) exploded_table AS item; Ensure your JSON keys match case-sensitively and Use consistent JSON structure. If offer is both a number and an array in different objects, consider preprocessing or cleaning up such inconsistencies. Regards, Chethan YM
... View more
05-05-2025
06:31 AM
Hi @rdhau You can go through the below Cloudera documentations completely to understand to work with HWC. https://docs.cloudera.com/cdp-private-cloud-base/7.1.8/integrating-hive-and-bi/topics/hive_hivewarehouseconnector_for_handling_apache_spark_data.html https://docs.cloudera.com/cdp-private-cloud-base/7.3.1/integrating-hive-and-bi/topics/hive-hwc-reader-mode.html https://docs.cloudera.com/cdw-runtime/1.5.4/hive-metastore/topics/hive_apache_spark_hive_connection_configuration.html Regards, Chethan YM
... View more
02-13-2025
05:42 AM
Hi, I have not tested it may be you can try something like this and see if that works. curl -X POST "https://cdp.company.com/gateway/cdp-proxy-api/Impala/api/v1/query" \ -H "Authorization: Bearer YOUR_ACCESS_TOKEN" \ -H "Content-Type: application/json" \ -d '{ "query": "SELECT * FROM sales_data WHERE date >= current_date - interval 7 days;", "database": "analytics_db", "async": false }' Cloudera API overview: https://docs.cloudera.com/cdp-public-cloud/cloud/api/topics/mc-api-overview.html?utm_source=chatgpt.com Regards, Chethan YM
... View more
02-13-2025
05:33 AM
@ysong26 The "out of sequence response" error suggests that the Impala JDBC driver expected a specific response sequence but received an unexpected one. This can happen due to network latency, dropped packets, or an issue in the client-server communication. See if you have set high value for "defaultRowBatchSize" property, If yes you can lower it and check which JDBC driver version you are using try to use the latest version available and see if that helps. Also verify if any network issues in the system. Regards, Chethan YM
... View more
12-11-2024
04:09 AM
1 Kudo
Hi @sayebogbon Could you restart the CM agent on the hosts where Impala daemon is in bad health and also restart service monitor from CM and check it out? Regards, Chethan YM
... View more
11-21-2024
10:40 PM
1 Kudo
Hi @mrblack To avoid full table scan you follow these tips: 1. Ensure proper partition pruning: https://impala.apache.org/docs/build/html/topics/impala_partitioning.html#:~:text=the%20impalad%20daemon.-,Partition%20Pruning%20for%20Queries,-Partition%20pruning%20refers 2. Re write the query with sub queries. 3. Add explicit hints for join behaviour. Impala supports join hints like brodcast and shuffle that can influence query planning. After optimising check the explain plan. Regards, Chethan YM
... View more