- Subscribe to RSS Feed
- Mark Question as New
- Mark Question as Read
- Float this Question for Current User
- Bookmark
- Subscribe
- Mute
- Printer Friendly Page
Exporting a phoenix table into a csvfile
- Labels:
-
Apache HBase
-
Apache Phoenix
Created ‎11-29-2016 08:53 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Is there a way to export a phoenix table into a csv file using the phoenix pherf ulitly. If so can anyone provide the details. I couldnt get the proper documetnation for this. we are supposed to export around 26 million rows into csv file.
Created ‎11-29-2016 08:58 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Phoenix pherf utility exports details like how much time the query took etc..not the data. There is no way to export data into csv from Phoenix. Can't you use Export utility provided by HBase.
Created ‎11-29-2016 09:06 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Better to go with @Ankit Singhal suggestion of using pig/hive or spark to export.
Created ‎11-29-2016 09:16 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
hi @Rajeshbabu Chintaguntla, the export utility provided by hbase generates sequence files i guess. how to convert them into csv files.
I tried the pig export, but the datatypes are not mapped.
i used a query similar to this
A = load 'hbase://query/select * from TRANSACTION' using org.apache.phoenix.pig.PhoenixHBaseLoader('localhost');
Should i need to explicitly specify the schema and if so how to do that.
Created ‎11-29-2016 09:40 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Yes Export generate sequence files. If you want to import back the exported data to other HBase table in different cluster then you can use it other wise it won't help.
Pig should map the data types properly. Can you try specifying the columns list than * and check. For eg:
A = load 'hbase://query/select col1,col2.. from TRANSACTION' using org.apache.phoenix.pig.PhoenixHBaseLoader('localhost');
Created ‎11-29-2016 09:51 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Or else can you try something similar to type convertion.
A = load 'hbase://query/select col1,col2.. from TRANSACTION' using org.apache.phoenix.pig.PhoenixHBaseLoader('localhost') as (rowKey:chararray,col_a:int, col_b:double, col_c:chararray);
Created ‎11-29-2016 08:59 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
You can use hive storage handler or pig or spark to do that.
https://phoenix.apache.org/hive_storage_handler.html
Created ‎11-29-2016 04:59 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Depending on the size of the data you want to export, you can just run a normal query.
SELECT col1 || ',' || col2 || ',' || col3 from my_table;
Created ‎02-26-2017 08:34 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
@ARUN did you get it, how to export a phoenix table into a csv file?
Created ‎02-26-2017 08:36 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
@ARUN did you get it, how to export a phoenix table into a csv file?
