Created on 07-10-2018 01:46 PM - edited 09-16-2022 06:26 AM
I have a kudu table with more than a million records, i have been asked to do some query performance test through both impala-shell and also java. Through impala-shell i am able to perform all the queries and it gives me the timestamp through which i get to know the time taken for running the query,but whne it comes to perfrom queries using java api kudu,is it possible if i can perform join query operation using java api kudu and benchmark the logs
Created 07-10-2018 02:54 PM
Hi HJ,
It is not possible to do a join using the native Kudu NoSQL API. You will need to use SQL with Impala or Spark SQL, or using the Spark data frame APIs to do the join.
Mike
Created 07-13-2018 03:17 PM
The only way I know of to do complex queries through Java is to use the Impala JDBC connector, which you can find here: https://www.cloudera.com/downloads/connectors/impala/jdbc/2-6-3.html
Created 07-16-2018 04:12 PM
If you have the data in Oracle I would suggest writing it to Parquet on HDFS using Sqoop first. After that, you will be able to transfer the data to Kudu using Impala with a command like CREATE TABLE kudu_table STORED AS KUDU AS SELECT * FROM parquet_table;
Created 07-16-2018 04:13 PM
However I see you wrote a separate forum post which is good, we try to stick with one topic per thread in the forums.
Created 07-17-2018 05:33 AM
thank you again for the solution.I found a java code which loads data into kudu table from csv. is it efficient.?