- Subscribe to RSS Feed
- Mark Question as New
- Mark Question as Read
- Float this Question for Current User
- Bookmark
- Subscribe
- Mute
- Printer Friendly Page
Best practices for ulimits? (number of open file decriptors and processes)
- Labels:
-
Apache Ambari
Created on ‎10-23-2015 06:00 PM - edited ‎09-16-2022 02:45 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
What are the recommended starting values for ulimits for each component?
The Ambari 2.1 doc says 10000, but should some services be started higher, say 32k?
Is there a good way to estimate these values based on cluster size, memory/cpu, number of blocks, etc? How can we proactively adjust ulimit, to avoid waiting until a service fails because the limit is hit?
Created on ‎10-23-2015 06:16 PM - edited ‎08-19-2019 05:56 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Ulimit does not depend on the size of the cluster but more on the individual node, workloads and user concurrency. I set these values as best practice, which BTW is a large enough value and is probably never reached.
* - nofile 32768 * - nproc 65536
Ambari actually changes the limit of the core files created as part of the commands executed like in this case -
Created on ‎10-23-2015 06:16 PM - edited ‎08-19-2019 05:56 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Ulimit does not depend on the size of the cluster but more on the individual node, workloads and user concurrency. I set these values as best practice, which BTW is a large enough value and is probably never reached.
* - nofile 32768 * - nproc 65536
Ambari actually changes the limit of the core files created as part of the commands executed like in this case -
