Member since
03-23-2015
1288
Posts
114
Kudos Received
98
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3299 | 06-11-2020 02:45 PM | |
5016 | 05-01-2020 12:23 AM | |
2816 | 04-21-2020 03:38 PM | |
2622 | 04-14-2020 12:26 AM | |
2323 | 02-27-2020 05:51 PM |
09-18-2017
04:03 AM
1 Kudo
Nah, nothing is stupid, they are all questions that lots of people will face one day. Glad that I am helpful here :).
... View more
09-18-2017
02:35 AM
I am not aware of Impala can do that, you have need to write custom code to convert the result into JSON.
... View more
09-18-2017
02:34 AM
What version of Hive are you using? If you are using CDH, which version of CDH are you using? Have you tried to set SPARK_HOME environment variable?
... View more
09-18-2017
02:32 AM
1 Kudo
This looks like that the table in Impala is not field delimited by tab. I suggest you re-create the table with following statement: CREATE TABLE my_table (a int, b int) ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t'; Basically you need to have "ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t'" in the table definition, so that impala/hive knows what the delimiter is, otherwise, the default Ctrl-A (hex 01) character will be used.
... View more
09-17-2017
11:52 PM
A hive table consists the following: 1. metadata info (all table and column definitions and HDFS location) 2. actual HDFS data stored in HDFS If you delete a managed table, both 1 and 2 will be deleted. However, if you delete an external table, then only 1 will be deleted, meaning, the table reference will be removed in Hive's backend database (show tables will not return the table and you can't query the table any more). The underlining HDFS file will remain on HDFS path untouched. To confirm this, you can check where the backend database is stored. If it is mysql, simply login and check the table under TBLS and check if you can query the table (mysql table, not hive table): SELECT * FROM TBLS WHERE TBL_NAME = "{your_table_name}"; Hope above helps.
... View more
09-17-2017
11:47 PM
2 Kudos
Impala does not support DATE data type, please refer to Cloudera doc: https://www.cloudera.com/documentation/enterprise/latest/topics/impala_langref_unsupported.html You mentioned you changed to timestamp, it also failed, can you please let me know how you made the change?
... View more
09-17-2017
11:46 PM
Please enable trace logging on ODBC and examine the log for hints. See user manual: http://www.cloudera.com/documentation/other/connectors/hive-odbc/latest/Cloudera-ODBC-Driver-for-Apache-Hive-Install-Guide.pdf on how to enable trace logging for ODBC driver.
... View more
09-17-2017
11:35 PM
Hi, I think it should be doable, though not through Hue. You can try to put the environment specific variables into job.properties file and pass them into the workflow, so that the workflow is generic and can be used in other environments. Hue will hard code things like credentials into the workflow if you are in secured environment, so it won't work. You might have to try to manually create those workflows and submit jobs from command line.
... View more
08-20-2017
10:41 PM
Hi Francesco, Why do you need to read from an empty parquet file? Can you simply remove them? Have you also tried to use parquet-tools command line utility to confirm if the parquet file is valid?
... View more
08-20-2017
10:33 PM
Hi, I believe that you are hitting https://issues.apache.org/jira/browse/HIVE-14037. HIVE-14037 has been fixed from CDH5.9.0 onwards. Did you create function with "USING JAR" keywords? If not, please give it a try to see if it will help. Otherwise, upgrade to latest CDH should help.
... View more