Member since
05-02-2017
360
Posts
65
Kudos Received
22
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
13349 | 02-20-2018 12:33 PM | |
1506 | 02-19-2018 05:12 AM | |
1862 | 12-28-2017 06:13 AM | |
7140 | 09-28-2017 09:25 AM | |
12180 | 09-25-2017 11:19 AM |
04-18-2017
06:07 AM
@Pardhu T You might have to check this Link. Its related to the ticket here that you might want to look at.
... View more
04-17-2017
07:21 PM
@Saikrishna Tarapareddy you mean to say that 172.16.1.4 should be generated?. The other way of doing it is split the data and then use regexp on the second string you got from split. Example: SPLIT('hive:hadoop',':') returns ["hive","hadoop"]
... View more
04-16-2017
11:51 AM
Hi @Constantin Stanca Thanks. At present im using the combination of substr and instr only. Just wanted to know if there are any other possibilities. My current solution is Substr('28 May 2016[35]',1,instr('28 May 2016[35]','[' - 1 ))
... View more
04-15-2017
02:36 PM
Hi.. Is there a way to find '[' from a column. I have a field which has a value of '28 May 2016[3]' and I need the output as '28 May 2016' I tried with regexp and split but while using '[' im facing an error. Also please dont suggest substr because my value will change and it will contain like '7 September 2015[456]' , '2 Sep 2014[34]'. Is there any way out in hive?
... View more
Labels:
- Labels:
-
Apache Hadoop
-
Apache Hive
04-13-2017
01:57 PM
@Kelvin Tong Copy your file from windows to the host 127.0.0.1. From local directory copy to HDFS using hadoop fs -ls <local file path> <hdfs path>
... View more
04-13-2017
01:07 PM
@Kelvin Tong Are you trying to copy the file from windows to hadoop? I think the file which you are trying to be copied should be present in the server on which hadoop is installed..
... View more
04-12-2017
11:06 AM
@Jonathan Samelson Glad you have solved. Be careful in choosing block size if you are going to deal with large chuck of data then its better to choose a higher value. But then if your just getting know few stuffs in HDFS for small amount of data then smaller block size wont affect the process.
... View more
04-11-2017
03:10 PM
@Christopher Daumer Check the delimiter which you used in creating external table is correct. Also could you share the sample data and DDL which you have used to create the hive table? I dont think the compression property or size will be the reason for the issue.
... View more
04-10-2017
07:17 AM
@Anand Pawar Its kind of tricky here. You can have the header when storing in HDFS. While processing the data for analysis you should remember that file contains header and it should be skipped orelse it will cause errors. As mentioned above if you use skip header properties it will be skipped by default in hive. However the base data lying underneath the hive table will contain header which can be used for any further processing. In simple when storing it you can have header but when processing the data you should not have header. If you feel it satisfies your question then accept the answer.
... View more
04-09-2017
06:44 PM
Hi @kerra Bucketing is supported for hive 2.x and above. set hive.enforce.bucketing = true ; The main reason is that it allows the correct number of reducers and the cluster by column to be automatically selected based on the table. Otherwise, you would need to set the number of reducers to be the same as the number of buckets as in set mapred.reduce.tasks = 256; and have a CLUSTER BY ... clause in the select.
... View more