Member since
02-08-2016
39
Posts
29
Kudos Received
5
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1486 | 06-22-2017 05:05 PM | |
2050 | 03-26-2017 11:55 PM | |
2472 | 07-18-2016 03:15 PM | |
17893 | 06-29-2016 07:43 PM | |
1450 | 06-20-2016 06:11 PM |
03-27-2017
02:25 AM
One option is to delete existing external table and create new table that includes new column. Since this is Hive metadata operation, your data files wont be touched. Downside is that you will have to execute alter table command to redefine partitions on new table.
... View more
08-09-2016
07:35 AM
HI Pierre, We would need to look at the code. Can you a do a persist just before stage 63 and before stage 65 check the spark UI storage tab and executor tab for data skew. If there is data skew, you will need to add a salt key to your key. You could also look at creating a dataframe from the RDD rdd.toDF() and apply UDF on it. DF manage memory more efficiently. Best, Amit
... View more
07-18-2016
03:15 PM
2 Kudos
We explicitly listed out FQDN's of all hosts in both the clusters under [domain_realm] section of krb5.conf file. We have to update this file everytime we add node to our clusters and our clusters are currently less than <100 nodes and this solution is manageable but for large clusters this may be challenge.
... View more
05-25-2018
02:17 AM
this seems like a better solution to me https://community.hortonworks.com/questions/8010/hives-alter-table-partition-concatenate-not-workin.html
... View more