- Subscribe to RSS Feed
- Mark Question as New
- Mark Question as Read
- Float this Question for Current User
- Bookmark
- Subscribe
- Mute
- Printer Friendly Page
CM enforces max locked memory on datanode process
- Labels:
-
Cloudera Manager
-
HDFS
Created on ‎09-11-2014 07:09 AM - edited ‎09-16-2022 02:07 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hello,
When setting the parameter "dfs.datanode.max.locked.memory" to a value other than 0, it is supposed to be the user's responsibility to set the MEMLOCK limits accordingly. CM does not leave that for the user and enforces it. (as seen on /proc/<pid>/limits)
This parameter is used by HDFS only to limt cache resources, so there is no real reason for CM to enforce this limit on an OS level. By doing so, CM does not allow external libraries (non-cache related) on the datanode to allocate memory past this limit.
To make a long story short, I'd like to cancel CM's enforcement of this limit and set the limits myself. Is there any way I could do so?
Thanks,
Elad
Created ‎09-11-2014 07:24 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
You can set probably dfs.datanode.max.locked.memory to the value you want to enforced by the OS, then use the Advanced Configuration Snippet for hdfs-site.xml to change the value that HDFS actually sees for this parameter in hdfs-site.
If CM didn't set memlock, then HDFS caching would be at the mercy of the default memlock setting and could fail unexpectedly, with essentially no workaround. So it's pretty important that CM sets this limit to something that will make caching work.
Thanks,
Darren
Created ‎09-11-2014 07:24 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
You can set probably dfs.datanode.max.locked.memory to the value you want to enforced by the OS, then use the Advanced Configuration Snippet for hdfs-site.xml to change the value that HDFS actually sees for this parameter in hdfs-site.
If CM didn't set memlock, then HDFS caching would be at the mercy of the default memlock setting and could fail unexpectedly, with essentially no workaround. So it's pretty important that CM sets this limit to something that will make caching work.
Thanks,
Darren
Created ‎09-11-2014 07:47 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Thanks a lot Darren, a very simple solution 🙂
