<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN">
<html>
<head>
<meta content="text/html; charset=UTF-8" http-equiv="Content-Type">
</head>
<body bgcolor="#ffffff" text="#000000">
That's the number of files open on the system. So this looks like<br>
an app problem. Some app has many files open.<br>
<br>
On 06/01/2011 10:37 PM, Vasyl S. Kostroma wrote:
<blockquote cite="mid:4B669AA1916A4A56A3D63CE75CE69142@VasylD"
type="cite">
<div dir="ltr">
<div style="font-family: 'Calibri'; color: rgb(0, 0, 0);
font-size: 12pt;">
<div>Hi guys!</div>
<div>I can’t find an answer in google, so my last hope is this
mailing list.</div>
<div> </div>
<div>Story.</div>
<div>I have two servers with same arrays. Servers connected by
DRBD.</div>
<div>I used ocfs2 as file system, also I used NFS4 to access
to the </div>
<div>ocfs2 drive. I do not have any idea, but the allocated
descriptors in</div>
<div>/proc/sys/fs/file-nr increasing every time while drive
accessed.</div>
<div>So after some time allocated descriptions over max value
and all</div>
<div>processes make error “To much opened files” (something
like</div>
<div>this). I do not see any error messages in log files...
Any idea?</div>
<div>I haven’t sleep two days <img style="border-style:
none;" class="wlEmoticon wlEmoticon-sadsmile"
alt="Печальная рожица"
src="cid:part1.09050400.05010109@oracle.com"></div>
<div> </div>
<div>Thank you all in advance.</div>
<div> </div>
<div>Configs:</div>
<div><em>cat /etc/drbd.conf</em></div>
<div><em># You can find an example in
/usr/share/doc/drbd.../drbd.conf.example</em></div>
<div><em></em> </div>
<div><em>include "drbd.d/global_common.conf";</em></div>
<div><em>include "drbd.d/*.res";</em></div>
<div><em></em> </div>
<div><em></em> </div>
<div><em>resource nfs {</em></div>
<div><em></em> </div>
<div><em> protocol C;</em></div>
<div><em></em> </div>
<div><em> handlers {</em></div>
<div><em> split-brain
"/usr/lib/drbd/notify-split-brain.sh root";</em></div>
<div><em> pri-on-incon-degr "echo o >
/proc/sysrq-trigger ; halt -f";</em></div>
<div><em> pri-lost-after-sb "echo o >
/proc/sysrq-trigger ; halt -f";</em></div>
<div><em> local-io-error "echo o >
/proc/sysrq-trigger ; halt -f";</em></div>
<div><em> }</em></div>
<div><em></em> </div>
<div><em> startup {</em></div>
<div><em> become-primary-on both;</em></div>
<div><em> degr-wfc-timeout 120;</em></div>
<div><em> }</em></div>
<div><em></em> </div>
<div><em> disk {</em></div>
<div><em> on-io-error detach;</em></div>
<div><em> }</em></div>
<div><em></em> </div>
<div><em> net {</em></div>
<div><em> cram-hmac-alg sha1;</em></div>
<div><em> shared-secret "password";</em></div>
<div>
<div><em></em></div>
<em> allow-two-primaries;</em></div>
<div><em> after-sb-0pri discard-zero-changes;</em></div>
<div><em> after-sb-1pri discard-secondary;</em></div>
<div><em> after-sb-2pri disconnect;</em></div>
<div><em> rr-conflict disconnect;</em></div>
<div><em> }</em></div>
<div><em></em> </div>
<div><em> syncer {</em></div>
<div><em> rate 500M;</em></div>
<div><em> verify-alg sha1;</em></div>
<div><em> al-extents 257;</em></div>
<div><em> }</em></div>
<div><em></em> </div>
<div><em> on st01 {</em></div>
<div><em> device /dev/drbd0;</em></div>
<div><em> disk /dev/sdb;</em></div>
<div><em> address 192.168.3.151:7788;</em></div>
<div><em> meta-disk internal;</em></div>
<div><em> }</em></div>
<div><em></em> </div>
<div><em> on st02 {</em></div>
<div><em> device /dev/drbd0;</em></div>
<div><em> disk /dev/sdb;</em></div>
<div><em> address 192.168.3.152:7788;</em></div>
<div><em> meta-disk internal;</em></div>
<div><em> }</em></div>
<div><em>}</em></div>
<div> </div>
<div>---</div>
<div><em>cat /etc/ocfs2/cluster.conf</em></div>
<div><em>#/etc/ocfs2/cluster.conf</em></div>
<div><em>node:</em></div>
<div><em> ip_port = 7777</em></div>
<div><em> ip_address = 192.168.1.151</em></div>
<div><em> number = 1</em></div>
<div><em> name = st01</em></div>
<div><em> cluster = ocfs2</em></div>
<div><em></em> </div>
<div><em>node:</em></div>
<div><em> ip_port = 7777</em></div>
<div><em> ip_address = 192.168.1.152</em></div>
<div><em> number = 2</em></div>
<div><em> name = st02</em></div>
<div><em> cluster = ocfs2</em></div>
<div><em></em> </div>
<div><em>cluster:</em></div>
<div><em> node_count = 2</em></div>
<div><em> name = ocfs2</em></div>
<div><em>---</em></div>
<div><em>cat /etc/exports</em></div>
<div><em># /etc/exports: the access control list for
filesystems which may be exported</em></div>
<div><em># to NFS clients. See exports(5).</em></div>
<div><em>/snfs
192.168.1.0/24(rw,sync,no_root_squash,no_subtree_check,fsid=0)</em></div>
<div><em>/snfs/projects
192.168.1.0/24(rw,sync,no_root_squash,no_subtree_check)</em></div>
<div><em>/snfs/configs
192.168.1.0/24(rw,sync,no_root_squash,no_subtree_check)</em></div>
<div><em>/snfs/variables
192.168.1.0/24(rw,sync,no_root_squash,no_subtree_check)</em></div>
<div><em>/snfs/backups
192.168.1.0/24(rw,sync,no_root_squash,no_subtree_check)</em></div>
<div>---</div>
</div>
</div>
<pre wrap="">
<fieldset class="mimeAttachmentHeader"></fieldset>
_______________________________________________
Ocfs2-users mailing list
<a class="moz-txt-link-abbreviated" href="mailto:Ocfs2-users@oss.oracle.com">Ocfs2-users@oss.oracle.com</a>
<a class="moz-txt-link-freetext" href="http://oss.oracle.com/mailman/listinfo/ocfs2-users">http://oss.oracle.com/mailman/listinfo/ocfs2-users</a></pre>
</blockquote>
<br>
</body>
</html>