Member since
01-25-2016
4
Posts
2
Kudos Received
0
Solutions
04-20-2016
09:41 AM
@Neeraj Sabharwal..looking forward for your inputs
... View more
04-20-2016
07:28 AM
1 Kudo
hi, i have a text file on HDFS(generated by sqoop job) which i am compressing using basic GzipCodec(streaming), i created a external hive table pointing to the compressed file, when i am do a select query on the hive table i am getting NULL's in the first column. - tried compressing the file in the sqoop job using --compress option it works fine -tried copying the file to local , zipped it and copy back to HDFS it also works fine -since the first column is an INT, tried changed it to STRING then some junk characters appear in the first column Below is the command m using for compression: hadoop jar /usr/local/hadoop/share/hadoop/tools/lib/hadoop-streaming-2.4.0.jar
-Dmapred.output.compress=true -Dmapred.compress.map.output=true
-Dmapred.output.compression.codec=org.apache.hadoop.io.compress.GzipCodec
-Dmapred.reduce.task=0
-input <input file>
-output <output file folder> Not sure if i am missing something here .Any help on this will be appreciated. Thanks.
... View more
Labels:
01-29-2016
06:29 AM
hi experts, I am trying to import data from Oracle DB. I just wanted to check if we can use a string field delimiter instead of comma(default delimiter for columns) Thanks.
... View more
Labels:
01-25-2016
12:42 PM
1 Kudo
Hi, In Hive DB I got a table named "employee" with employee id as one field, can I set HiveContext and delete data from Hive table() like below (if not what is the best way to do) val sqlContext = new HiveContext(sc) sqlContext.sql("DELETE FROM employee WHERE employee_id=6") Thanks
... View more