Member since
05-02-2017
360
Posts
65
Kudos Received
22
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
13343 | 02-20-2018 12:33 PM | |
1499 | 02-19-2018 05:12 AM | |
1858 | 12-28-2017 06:13 AM | |
7135 | 09-28-2017 09:25 AM | |
12162 | 09-25-2017 11:19 AM |
02-13-2017
08:56 PM
Is that so? In relational databases I agree that there be a significant difference. But in hadoop i always thought that it will read the entire record if it is stored as TEXTFILE. Doesnt mapper reads entire record to parse it to reducer?
... View more
02-13-2017
07:03 PM
1 Kudo
I have a hive managed table stored as TEXTFILE. Will there be any change in the performance between select col1,col2 from hive_tabl than select * from hive_tabl. Consider the table has 300 columns with 20 billion of data. Will there be any performance which impacts based on the select clause. If so can ORC storage overcome it or what will be the best way to store the data which will not affect based on the select clause.
... View more
Labels:
- Labels:
-
Apache Hadoop
02-09-2017
02:20 PM
Thanks Frank. I have tried both ways.. But the compression ratio is still the same which has provided in the question.
... View more
02-07-2017
08:08 PM
I have text file with 2.6Gb. I have loaded it into hive table with text as storage type. From the text hive table i have loaded into avro based hive table by insert into table avro_hive table which is a snappy compression table. Please feel free if you need more details
... View more
02-07-2017
08:22 AM
I have created a hive avro based table with snappy compression. The size of avro file is 2628MB. The data in the hive avro based table without snappy compression is 2296MB. I have created one more avro hive table with snappy compression and loaded the same data. But there is no big change in the compression size. Also if I describe the table properties it shows that the compression as 'No'. Please find below the table property. Table Parameters:
COLUMN_STATS_ACCURATE True
avro.compress SNAPPY
transient_lastDdlTime 1486455066
# Storage Information
SerDe Library: org.apache.hadoop.hive.serde2.avro.AvroSerDe
InputFormat: org.apache.hadoop.hive.ql.io.avro.AvroContainerInputFormat
OutputFormat: org.apache.hadoop.hive.ql.io.avro.AvroContainerOutputFormat
Compressed: No
Num Buckets: -1
Bucket Columns: []
Sort Columns: []
Storage Desc Params:
serialization.format 1
... View more
Labels:
- Labels:
-
Apache Hive
01-04-2017
06:45 PM
@Sergey SoldatovFixed width files where there will not be any delimiters available. Each column data will be avilable in the specific length but with no delimiters.
... View more
01-04-2017
02:34 PM
Can we create external hive table on top of Fixed width file? If yes then how it can be done.
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache Hive
09-25-2016
12:40 PM
3 Kudos
Im trying to load hive table. I have two different source which has to loaded into same target. Is it ok If run that job in parallel?
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache Hive
09-07-2016
08:20 AM
1 Kudo
I need to know which performs better in hive. EXISTS or IN?
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache Hive
08-11-2016
01:24 PM
I have pdf file. I have copied the file from local system to hdfs. But i need to convert the pdf file into a hive table. Is there anyway to do in hive? I know we can handle the same in pig.
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache Hive
-
Apache Pig
- « Previous
- Next »