Member since
07-24-2014
44
Posts
0
Kudos Received
0
Solutions
12-13-2014
03:04 AM
Hello Harsh, Even if I am using the Hive CLI, which i presume uses ODBC driver, still I get the same error if my path is /root/filename.csv. hive> load data local inpath "/root/veraxinternal.csv" overwrite into table veraxexternal; FAILED: SemanticException Line 1:23 Invalid path '"/root/veraxinternal.csv"': No files matching path file:/root/veraxinternal.csv If I copy the same file to a different a location and do loadoperation it goes well. What's the logic here? does not hive accept files from /Root? By the way how can I trace this if I need to debug hive> load data local inpath "/home/cloudera/veraxinternal.csv" overwrite into table veraxexternal; Copying data from file:/home/cloudera/veraxinternal.csv Copying file: file:/home/cloudera/veraxinternal.csv Loading data to table default.veraxexternal chgrp: changing ownership of 'hdfs://quickstart.cloudera:8020/user/hive/warehouse/veraxexternal': User does not belong to hive Table default.veraxexternal stats: [numFiles=1, numRows=0, totalSize=113, rawDataSize=0] OK
... View more
10-21-2014
06:31 AM
Hello Friends, 1) Created one table in Hive with column X as timestamp and fileformat=Text 2) Changed the file format to Parquet using alter command in Hive 3) Now inserted the data from other source table successfully into this changed parquet table 4) Now tried to change the data type for column X as timestamp, Hive thrown error saying it is not supported which is valid. Hive does not support parquet-timestamp format 5) Now I did the same alter command through Impala successfully 6) But if I read that column X then I get the following error. The same query works fine with Text file format and timestamp column through impala and also even Parquet- String as data type for column X. Any ideas why the below errors 'hdfs://nameservice1/work/oca/wrkoca01/wrkoca01_files/parquet_format_ceh_busevents_extract/000044_0' has an incompatible type with the table schema for column 'X'. Expected type: INT96. Actual type: BYTE_ARRAY The X column data format is as 2014-10-01 21:51:09. Not sure what's wrong with the format here. I have found one issue IMPALA-779 but it does not seem to be applicable to my case Regards Sree
... View more
Labels:
- Labels:
-
Apache Hive
-
Apache Impala
-
HDFS
10-17-2014
03:53 AM
Hello Alex, Thanks for your time. Finally I got the point where I was doing the mistake, solved but still lot of questions in the mind. Objective: Convert the existing text table to parquet table for performance gains The actions performed are as below hive> create table t2 like t1; # t1 is text formatted table hive> insert into table t2 select * from t1; hive>alter table t2 set fileformat=parquet Impala> Invalidate metadata t2; Impala> desc formatted t2; Result: error, failed to load metadata, invalidate may help message. After posting this query and did some mor work, found that even Hive was not able to describe the table. NOte: Here I see a problem, we are in Hive version 12, the command returned successful, but it is not able to describe the table. I really see this as major issue, because when we allow the command and made successful changes to metadata , but unable to load metadata then we are lost, we will not be able to do any thing for this table, neither alter to original format nor drop the table. Don't you see this as issue? Now I found for Hive 12 version I need to follow the below alter table <tablename>set serde '<parquet serde>'; alter table <tablename> set fileformat inputformat "<parquet inputformat class>" alter table <tablename> set fileformat outputformat "<parquet outputformat class>" Now the first problem is solved, of course not on the same table as t2, i did exerice on another new tables, so now I could describe the table both on Impala and Hive. but not able to read the data because as you mentioned the underlying data is csv not parquet. Here I forgot to reload the data so that the data can be converted to parquet while copying it to new table with new altered format. finally now the issue is solved and more over it was not impala issue. but still I have doubt, may be fundamental: alter table set fileformat will just change table metadata information, so do we see really such cases where user will just need to alter table properties to parquet or ORC. Not able think of such scenario. when user has to take care of underlying data formats, then a warning mentioning that, user should alter data file to the respective <fileformat> to read the data would really help, I think
... View more
10-16-2014
01:12 AM
Impala 1.4.0. Also additional input is it seems hive also not able to see the table after using alert command, so may not be a problem. But I don't understand why we allow alter command to change file format and after doing that then it says Table not found. Really interesting to know what's happening here.
... View more
10-15-2014
04:54 AM
Hi, I see today that Impala some how not able to read the metadata. I created a table in Hive with text format, data loaded into Hive, Alter the file format to PARQUET using alter table <table name>SET FILEFORMAT PARQUET. Now I did invalidate metadata <tablename> in Impala, I can very well see the table in Impala by doing Show tables, but if I do describe it say's it can't load the metadata. E1015 06:20:11.609848 34004 Analyzer.java:1445] Failed to load metadata for table: XXXXXX Failed to load metadata for table: XXXXX. Running 'invalidate metadata XXXXXX may resolve this problem. CAUSED BY: NullPointerException: null I1015 06:20:11.610817 34004 jni-util.cc:177] com.cloudera.impala.common.AnalysisException: Failed to load metadata for table: xxxx at com.cloudera.impala.analysis.Analyzer.getTable(Analyzer.java:1446) at com.cloudera.impala.analysis.Analyzer.getTable(Analyzer.java:1462) at com.cloudera.impala.analysis.DescribeStmt.analyze(DescribeStmt.java:61) at com.cloudera.impala.analysis.AnalysisContext.analyze(AnalysisContext.java:252) at com.cloudera.impala.service.Frontend.analyzeStmt(Frontend.java:597) at com.cloudera.impala.service.Frontend.createExecRequest(Frontend.java:621) at com.cloudera.impala.service.JniFrontend.createExecRequest(JniFrontend.java:135) Caused by: com.cloudera.impala.catalog.TableLoadingException: Failed to load metadata for table: xxxxxx CAUSED BY: NullPointerException: null at com.cloudera.impala.catalog.IncompleteTable.loadFromThrift(IncompleteTable.java:113) at com.cloudera.impala.catalog.Table.fromThrift(Table.java:206) at com.cloudera.impala.catalog.ImpaladCatalog.addTable(ImpaladCatalog.java:606) at com.cloudera.impala.catalog.ImpaladCatalog.addCatalogObject(ImpaladCatalog.java:495) at com.cloudera.impala.catalog.ImpaladCatalog.updateCatalog(ImpaladCatalog.java:233) at com.cloudera.impala.service.Frontend.updateCatalogCache(Frontend.java:150) at com.cloudera.impala.service.JniFrontend.updateCatalogCache(JniFrontend.java:152) Any suggestions/inputs to share
... View more
Labels:
- Labels:
-
Apache Hive
-
Apache Impala
10-10-2014
02:03 AM
That's wonderful. Thanks a lot Lenny.
... View more
10-09-2014
01:59 AM
Hello Friends, It seems Impala does not support column names being a reserved word, is it correct? we have table with column name DATE if we select date or select distinct date query it says syntax error.Encountered: DATE Expected: ALL, CASE, CAST, DISTINCT, FALSE, IF, INTERVAL, NOT, NULL, STRAIGHT_JOIN, TRUE, IDENTIFIER. Hive does not show up any problem. Is there any plan to fix this issue in Impala or it is built like that intentionally if yes why is it so? Is not it a problem if people created a table in Hive and then retrieving data from Impala. I am cautious that having column names same as reserved word is not good practice but this is not completly avoidable as well Regards Sree
... View more
Labels:
- Labels:
-
Apache Hive
-
Apache Impala
08-05-2014
11:02 PM
Hi I am facing the same problem but unable to understand your solution. Is it that you first set Advanced Configuration snippet field to True and then passed on ldap_domain details in command line? Regards Sreeman
... View more
08-05-2014
02:56 AM
Thank you Henry. In case if we don't set SSL between AD and Impala/Hive, does it encrypt the password automatically? I see if we set ldap_passwords_in_clear_ok, false then password will not be in clear form? But I am not sure about user id? If both user id and password encrypted by default what kind of encryption logic is used? Regards Sreeman
... View more
08-03-2014
11:34 PM
Thanks Henry for giving me the github link. Let's in this case I want to establish SSL connection between Tableau server and Impala server and also between Impala server and Active Directory.Does impala support it? I mean while enabling SSL, we can provide only one SSL certificate for the ssl server certifcate path parameter? Am I right? Regards Sreeman
... View more