<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Too many open files in region server logs in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Too-many-open-files-in-region-server-logs/m-p/124182#M26740</link>
    <description>&lt;P&gt;hello Sumit&lt;/P&gt;&lt;P&gt;If your ulimit is already set to unlimited or a very high number, you could actually getting insight on the number of open files with lsof | wc -l. You may need to increase the max number of filed handles in the os. check fs.file-max to see if this helps. this is to try to solve the cause.&lt;/P&gt;&lt;P&gt;An offlineMetaRepair, fix meta should help with the consequence.&lt;/P&gt;</description>
    <pubDate>Sat, 30 Apr 2016 16:24:33 GMT</pubDate>
    <dc:creator>nmaillard1</dc:creator>
    <dc:date>2016-04-30T16:24:33Z</dc:date>
    <item>
      <title>Too many open files in region server logs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Too-many-open-files-in-region-server-logs/m-p/124181#M26739</link>
      <description>&lt;P&gt;I have 3 region servers and their total size on HDFS is ~50G only. I have ulimit set to unlimited and for the hbase user also the value is very high (32K +). I am noticing following in my logs very often after which I start getting HFile corruption exceptions:&lt;/P&gt;&lt;P&gt;2016-04-27 16:44:46,845 WARN 
[StoreFileOpenerThread-g-1] hdfs.DFSClient: Failed to connect to
/10.45.0.51:50010 for block, add to deadNodes and continue. java.net.SocketException:
&lt;STRONG&gt;Too many open files&lt;/STRONG&gt;
java.net.SocketException: Too many open files
       at
sun.nio.ch.Net.socket0(Native Method)&lt;/P&gt;&lt;P&gt;After many of these open files issues, I get a barrage of HFile corrupt issues too and hbase fails to come up:&lt;/P&gt;&lt;P&gt;2016-04-27 16:44:46,313 ERROR
[RS_OPEN_REGION-secas01aplpd:44461-1] handler.OpenRegionHandler: Failed open of
region=lm:DS_326_A_stage,\x7F\xFF\xFF\xF8,1460147940285.1a764b8679b8565c5d6d63e349212cbf.,
starting to roll back the global memstore size.&lt;/P&gt;&lt;P&gt;java.io.IOException: java.io.IOException:
org.apache.hadoop.hbase.io.hfile.&lt;STRONG&gt;CorruptHFileException&lt;/STRONG&gt;: &lt;STRONG&gt;Problem reading HFile
Trailer from file&lt;/STRONG&gt; hdfs://mycluster/MA/hbase/data/lm/DS_326_A_stage/1a764b8679b8565c5d6d63e349212cbf/e/63083720d739491eb97544e16969ffc7&lt;/P&gt;&lt;P&gt;     at
org.apache.hadoop.hbase.regionserver.HRegion.initializeRegionStores(HRegion.java:836)
     at
org.apache.hadoop.hbase.regionserver.HRegion.initializeRegionInternals(HRegion.java:747)
     at
org.apache.hadoop.hbase.regionserver.HRegion.initialize(HRegion.java:718)&lt;/P&gt;&lt;P&gt;My questions are two:&lt;/P&gt;&lt;P&gt;1. No other process on this node shows up too many open files issue. Even data node seems to not show this error in logs. Not sure, why then this error should be reported.&lt;/P&gt;&lt;P&gt;2. Would an OfflineMetaRepair following by hbck -fixMeta and hbck -fixAssignments solve the issue?&lt;/P&gt;</description>
      <pubDate>Sat, 30 Apr 2016 14:13:16 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Too-many-open-files-in-region-server-logs/m-p/124181#M26739</guid>
      <dc:creator>sumit_nigam</dc:creator>
      <dc:date>2016-04-30T14:13:16Z</dc:date>
    </item>
    <item>
      <title>Re: Too many open files in region server logs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Too-many-open-files-in-region-server-logs/m-p/124182#M26740</link>
      <description>&lt;P&gt;hello Sumit&lt;/P&gt;&lt;P&gt;If your ulimit is already set to unlimited or a very high number, you could actually getting insight on the number of open files with lsof | wc -l. You may need to increase the max number of filed handles in the os. check fs.file-max to see if this helps. this is to try to solve the cause.&lt;/P&gt;&lt;P&gt;An offlineMetaRepair, fix meta should help with the consequence.&lt;/P&gt;</description>
      <pubDate>Sat, 30 Apr 2016 16:24:33 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Too-many-open-files-in-region-server-logs/m-p/124182#M26740</guid>
      <dc:creator>nmaillard1</dc:creator>
      <dc:date>2016-04-30T16:24:33Z</dc:date>
    </item>
    <item>
      <title>Re: Too many open files in region server logs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Too-many-open-files-in-region-server-logs/m-p/124183#M26741</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/131/nmaillard.html" nodeid="131"&gt;@nmaillard&lt;/A&gt; - Thanks. Yes, I am aware of lsof and was planning to use it. Also could there be a setting in hbase which restricts number of open file handles in hbase itself and throws this error?&lt;/P&gt;&lt;P&gt;Also, you meant /proc/sys/fs/file-max?&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;</description>
      <pubDate>Sat, 30 Apr 2016 19:51:33 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Too-many-open-files-in-region-server-logs/m-p/124183#M26741</guid>
      <dc:creator>sumit_nigam</dc:creator>
      <dc:date>2016-04-30T19:51:33Z</dc:date>
    </item>
    <item>
      <title>Re: Too many open files in region server logs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Too-many-open-files-in-region-server-logs/m-p/124184#M26742</link>
      <description>&lt;P&gt;Hi Sumit,&lt;/P&gt;&lt;P&gt;You may also want to verify that the ulimit that is set, is actually applied to the process : &lt;/P&gt;&lt;P&gt;# cat /proc/&amp;lt;Region Server PID&amp;gt;/limits&lt;/P&gt;&lt;P&gt;It is possible that somehow the user limits are overridden when the process starts up.&lt;/P&gt;</description>
      <pubDate>Mon, 02 May 2016 16:21:22 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Too-many-open-files-in-region-server-logs/m-p/124184#M26742</guid>
      <dc:creator>rmaruthiyodan</dc:creator>
      <dc:date>2016-05-02T16:21:22Z</dc:date>
    </item>
    <item>
      <title>Re: Too many open files in region server logs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Too-many-open-files-in-region-server-logs/m-p/124185#M26743</link>
      <description>&lt;P&gt;Depending on your OS the setting might be different then you expect. To check the actual value become root and switch to the user hbase and print the actual limits.&lt;/P&gt;&lt;PRE&gt;# on Hbase Region Server:
sudo -i
su hbase

# print limits for the user hbase:
ulimit -a&lt;/PRE&gt;&lt;P&gt;On our RedHat 6 system, there was a file 90-nproc.conf in /etc/security/limits.d/ deployed. This limits the nr of processes for users to 1024. The user ambari received these limits and when starting hbase from ambari the limits are passed over somehow.&lt;/P&gt;&lt;P&gt;As &lt;A rel="user" href="https://community.cloudera.com/users/329/rmaruthiyodan.html" nodeid="329"&gt;@rmaruthiyodan&lt;/A&gt; mentions you can check the running process limits.&lt;/P&gt;&lt;PRE&gt;grep 'open files' /proc/&amp;lt;Ambari Agent PID&amp;gt;/limits
grep 'open files' /proc/&amp;lt;Region Server PID&amp;gt;/limits&lt;/PRE&gt;&lt;P&gt;&lt;A href="http://hbase.apache.org/0.94/book/configuration.html"&gt;Hbase book config&lt;/A&gt; suggests: &lt;EM&gt;'Set it to north of 10k'&lt;/EM&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 02 May 2016 17:12:00 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Too-many-open-files-in-region-server-logs/m-p/124185#M26743</guid>
      <dc:creator>MrBee</dc:creator>
      <dc:date>2016-05-02T17:12:00Z</dc:date>
    </item>
    <item>
      <title>Re: Too many open files in region server logs</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Too-many-open-files-in-region-server-logs/m-p/124186#M26744</link>
      <description>&lt;P&gt;Hey &lt;A rel="user" href="https://community.cloudera.com/users/329/rmaruthiyodan.html" nodeid="329"&gt;@rmaruthiyodan&lt;/A&gt; - Thanks. Yes, I had to use /proc to find region server PID specific limits. Basically, ambari restricts this number to 32K by default and this can be overridden in blueprint being submitted.&lt;/P&gt;</description>
      <pubDate>Mon, 02 May 2016 21:44:53 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Too-many-open-files-in-region-server-logs/m-p/124186#M26744</guid>
      <dc:creator>sumit_nigam</dc:creator>
      <dc:date>2016-05-02T21:44:53Z</dc:date>
    </item>
  </channel>
</rss>

