Member since
01-08-2018
133
Posts
31
Kudos Received
21
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
17324 | 07-18-2018 01:29 AM | |
3106 | 06-26-2018 06:21 AM | |
5269 | 06-26-2018 04:33 AM | |
2709 | 06-21-2018 07:48 AM | |
2241 | 05-04-2018 04:04 AM |
04-18-2018
02:42 AM
Hi Gekas, Thank you very much for your answer. I will test it on UAT & will give you & community the update. Thanks
... View more
04-17-2018
12:57 PM
@GeKas you are the Hero of the day. You saved me from having to call a production maintenance window. Thanks for your help.
... View more
04-17-2018
08:27 AM
1 Kudo
I use the same command and have no issues. According to logs: Caused by: org.apache.hadoop.util.DiskChecker$DiskErrorException: Could not find any valid local directory for output/attempt_1523546159827_0013_r_000000_0/map_0.out So, I would guess that you csv is too big and when the reducer tries to load it, there is no sufficient space in local dirs of YARN nodemanager. Can you try set more reducers by using : --reducers 4 or more (based on your partitions and the csv size). You can also set more mappers, but based on log the reducer is suffering. More details: https://www.cloudera.com/documentation/enterprise/5-13-x/topics/search_mapreduceindexertool.html#concept_pjs_3sd_3v
... View more
04-17-2018
01:47 AM
Hi.First of all sorry for late reply, I was out for some time. According to this "yyyy-mm-dd hh:mm:ss[.f...]", yes you have to store it in UTC. In order to be able to store date in other timezones, the format should include the "Z" which is hours from UTC.
... View more
04-17-2018
01:19 AM
2 Kudos
This is a mis-leading of the "free" output. The first line (starting with "Mem") displays that you have 62G of memory and 56G are used. This memory is used but not from procesess. At the end of the line, you will see a number of 39G cached. In few words, Linux uses a part of free RAM to store data from files used often, in order to save some interactions with the hard disk. Once an application request memory and there is no "free", Linux automatically drops these caches. You cannot turn this feature off. The only thing you can do is just drop the current cached data, but Linux will store something the very next second. In any case, when the output of "free" is similar to the one you provided, you should always refer to the second line "-/+ buffers/cache: 16G 49G" This is the real status, which show "16G" used and "49G" free. Finally, CM displays the disk and memory usage of the host (in Hosts view) regardless of what process is using it. It is the same output as "free".
... View more
03-26-2018
03:31 AM
Hi Tricky part is that some users (with same permission) can see tables and some cannot like me. We are using sssd, i have restarted it and did "id userid" command. Groups appear correct but still no hive tables.
... View more
03-24-2018
08:06 PM
to add a point to @GeKas you can see the permission denied in your log trace ( future reference ) . AccessControlException: Permission denied: user=hive, access=EXECUTE, inode="/tmp":hdfs:supergroup:d-wx
... View more
03-20-2018
08:11 AM
1 Kudo
The error complains about the value of "hadoop.security.authentication". You have set it to "Kerberos" while the accepted values are "simple" and "kerberos" (all letters in lowercase).
... View more
03-16-2018
08:22 AM
1 Kudo
According to ther error you posted, your NameNode is in Safe Mode. There should be an Active Name Node in order to perform any HDFS action.
... View more
03-15-2018
10:23 AM
Thank you, @Lars Volker. Definitely learned something new there!
... View more