Created on 01-07-2020 06:43 AM - last edited on 01-07-2020 08:49 AM by VidyaSargur
Hi,
I have a database and i want to find the best way to import it into impala. Please help me.
Created 01-09-2020 03:22 PM
Impala shares metadata and data with Hive. You can use sqoop to import the tables from your database into hive. Don't forget to run 'invalidate metadata' in Impala after the ingestion is done. Otherwise, you can't see the imported tables.
Created 02-07-2020 04:13 AM
If a sqoop job failed/crashed in the middle of importing a table, the table is imported. When you run this job again, it will start from zero so you need to clear the partially imported data first.
Alternatively, if you know which rows are not imported yet, you can use the WHERE clause when you restart the job to import the rest rows.
Created 01-09-2020 03:22 PM
Impala shares metadata and data with Hive. You can use sqoop to import the tables from your database into hive. Don't forget to run 'invalidate metadata' in Impala after the ingestion is done. Otherwise, you can't see the imported tables.
Created 02-06-2020 01:15 AM
Thanks for your answer, but I want also to ask something.
If I have a big amount of data which are to be parsed and something goes wrong in the process of parsing, the parsing will start from the zero or I can do something in order to start again from the part which crashed the process?
Created 02-07-2020 04:13 AM
If a sqoop job failed/crashed in the middle of importing a table, the table is imported. When you run this job again, it will start from zero so you need to clear the partially imported data first.
Alternatively, if you know which rows are not imported yet, you can use the WHERE clause when you restart the job to import the rest rows.