Member since
08-20-2018
7
Posts
0
Kudos Received
0
Solutions
10-03-2016
11:38 PM
Bump. Any recommendations to this above question? We are also looking for such a solution.
... View more
10-03-2016
06:54 PM
@Rajeshbabu Chintaguntla Thanks for that detailed post, there seems to be two really good approaches there. Which approach would likely provide better performance? It seems like the CsvBulkLoadTool might better than ImportTsv but wanted to verify.
... View more
10-03-2016
06:51 PM
@Constantin Stanca Thanks for the insight. Based on your comment, does Phoenix chunk the data automically if we ingest it through it?
... View more
09-29-2016
07:25 PM
We have a 250GB CSV file that contains 60 Million records and roughly 600 columns. The file lives within HDFS currently and we are trying to ingest it into HBase and have a phoenix table on top of it. The approach we tried so far was to create a Hive table backed by HBase and then execute an overwrite command in Hive which ingests the data in HBase. The biggest problem we have is that the job currently takes about 3-4 days to run!! this is running on a 10 node cluster with medium spec cluster (30GB of RAM each node, and 2TB on each). Any advice on how to speed this up or different methods that can be more efficient?
... View more
Labels:
- Labels:
-
Apache HBase
-
Apache Hive
-
Apache Phoenix