Member since
07-31-2013
1924
Posts
462
Kudos Received
311
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1952 | 07-09-2019 12:53 AM | |
| 11789 | 06-23-2019 08:37 PM | |
| 9077 | 06-18-2019 11:28 PM | |
| 10031 | 05-23-2019 08:46 PM | |
| 4442 | 05-20-2019 01:14 AM |
09-21-2015
11:56 PM
To add onto Wilfred's response, what is your CDH version? HDFS does cache all positive entries for 5 minutes, but negative caching wasn't supported until CDH 5.2.0 onwards (via HADOOP-10755). See also http://archive.cloudera.com/cdh5/cdh/5/hadoop/hadoop-project-dist/hadoop-common/core-default.xml#hadoop.security.groups.negative-cache.secs (which lists negative caching's TTL default being 30s, vs. positive's 300s). NSCD does also do negative caching by default, which could explain why the problem is gone, depending on how many negative, WARN group-lookup failure entries you observe in the log.
... View more
09-20-2015
06:55 AM
Hi Harish, thanks for your reply. I have another doubt to ask you, how can we determine the no of mappers in the above mentioned wordcount programme. Can we determine that only using those 2 input files a.txt & b.txt ??. Is it mandatory that we should know the file size & block size? Please help...
... View more
09-18-2015
10:13 AM
Hrsh, I was able to find the property and I modified queueMaxAppsDefault property and now I get more than 8 apps running concurently. Thanks for your help Nitin
... View more
09-18-2015
04:59 AM
The value on the doc page is picked as about 20% of the RAM for overhead reservation, but you could set it lower. Our past overcommit testing does show that the values can reach close to extra 20% in use for some tested workloads, but that would not be an always-as-such case - and this may have changed overall lately also. We're reworking the docs for these recommendations soon in future, as developments happen. For now, please rely on the XLSX file for a more closer guideline on the recommended calculated values.
... View more
09-18-2015
04:23 AM
Glad to hear you were able to figure it out. In spirit of https://xkcd.com/979/, please mark the thread solved with the solution post selected, so others with a similar problem can find their solution quicker on the web.
... View more
09-09-2015
10:36 PM
Start here, and drill further down into the DFSClient and DFSInputStream, etc. classes: https://github.com/cloudera/hadoop-common/blob/cdh5.4.5-release/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DistributedFileSystem.java#L294-L303
... View more
09-03-2015
05:51 PM
In spirit of https://xkcd.com/979/, feel free to mark the thread as resolved if it does help your cause, so others may find a solution quicker.
... View more
09-03-2015
05:49 PM
1 Kudo
Currently, the CM BDR feature does not carry any HBase replication abilities (we do support schedulable snapshot policies, but no replication/copies yet). You will need to utilise standard HBase techniques to copy over the data between your two clusters: http://blog.cloudera.com/blog/2013/11/approaches-to-backup-and-disaster-recovery-in-hbase/, and I'd recommend the ExportSnapshot method (if not live replication).
... View more
09-03-2015
05:16 PM
You will need the gateway copy, which exists under /etc/hive/conf/ on a Hive Gateway designated node (check Hive -> Instances in CM to find which hosts have a gateway role).
... View more