Member since
12-28-2017
7
Posts
1
Kudos Received
0
Solutions
01-11-2018
05:56 PM
Hello @Bala Vignesh N V I am sorry for the late reply although I was too busy lately and not following the forum. So finally we solved this by making the table internal, keeping the TextFile format and storing data under default Hive directory. The definition of table look like this at the moment: create table test1(c1 int, c2 int) CLUSTERED BY(c1) SORTED BY(c1) INTO 4 BUCKETS ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED AS TEXTFILE;
So every time I was specifying the LOCATION Hive would overwrite the data but when changing to internal everything worked fine under the /apps/hive/warehouse/ directory
... View more
01-11-2018
05:52 PM
In case that someone will face the same problem we solved this by making the table internal, keeping the TextFile format and storing data under default Hive directory. The definition of table look like this at the moment: create table test1(c1 int, c2 int) CLUSTERED BY(c1) SORTED BY(c1) INTO 4 BUCKETS ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED AS TEXTFILE;
... View more
01-09-2018
08:01 PM
Hello guys, any update on this? I remind you that in my case it occurs on HDFS not S3. Thank you in advance
... View more
12-29-2017
02:30 PM
1 Kudo
Hello guys, I have a similar issue but with external table onto HDFS. Is there any solution on this so far? we are using HDP-2.6.3.0 and here is how my table looks like: create external table test1(c1 int, c2 int) CLUSTERED BY(c1) SORTED BY(c1) INTO 4 BUCKETS ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED AS TEXTFILE LOCATION '/user/companies/'; Currently the data is always overwritten Thank you
... View more
12-29-2017
10:43 AM
Hello @Bala Vignesh N V thank you for your reply and my excuses for the insufficient information but I missed out to mention that in our case we have this behavior on HDFS not S3. Also I really want to avoid such a union because our inserts are already invoking huge select with join so this work around would exhaust our cluster resources and would make the ETL process really slow. Are you aware of such a bug? or it could be a configuration issue as well?
... View more
12-28-2017
10:01 PM
Hello guys,
I am trying to execute multiple insert queries into a table which looks like the next one:
create external table test1(c1 int, c2 int) CLUSTERED BY(c1) SORTED BY(c1) INTO 4 BUCKETS ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' STORED AS TEXTFILE LOCATION '/user/companies/';
Although I noticed that Hive is overriding the existing data on each insert(). I am running Hive v.1.2.1000 with HDP-2.6.3.0 I already found this similar issue and I was wondering if someone has any suggestions or solution for this one Thank you
... View more
Labels:
- Labels:
-
Apache Hive
12-28-2017
11:24 AM
Hello @Huahua Wei yes, you can connect through beeline e.g: jdbc:hive2://yournode.yourdomain.office:10500/default
... View more