[Ocfs2-users] Sometime Crash

Sebastien WACHTER sebastien.wachter at netasq.com
Fri Feb 5 06:48:53 PST 2010


Hi all 

this my configuration 

2 servers on debian 2.6.26-2-amd64 #1 SMP Thu Nov 5 02:23:12 UTC 2009 x86_64 GNU/Linux

each server run nfsd 

a pool a server can be access on drdb/ocfs disk via nfs 


We're using OCFS2 Node Manager and DLM 1.5.0.
drbd: initialised. Version: 8.0.14 (api:86/proto:86)
drbd: GIT-hash: bb447522fc9a87d0069b7e14f0234911ebdab0f7 build by phil at fat-tyre, 2008-11-12 16:40:33

sometime i have crash like this 
[26969.636164] CPU 1 
[26969.636189] Modules linked in: ocfs2 drbd cn ocfs2_dlmfs ocfs2_stack_o2cb ocfs2_dlm ocfs2_nodeman
ager ocfs2_stackglue configfs nfsd auth_rpcgss exportfs nfs lockd nfs_acl sunrpc ipv6 loop parport_p
c parport snd_pcm snd_timer snd soundcore serio_raw i2c_i801 rng_core snd_page_alloc pcspkr i2c_core
 psmouse iTCO_wdt container button shpchp i5000_edac edac_core pci_hotplug evdev ext3 jbd mbcache id
e_cd_mod cdrom ata_generic libata dock sd_mod piix floppy ide_pci_generic ide_core ehci_hcd 3w_9xxx 
scsi_mod uhci_hcd e1000e thermal processor fan thermal_sys
[26969.636526] Pid: 2579, comm: nfsd Not tainted 2.6.26-2-amd64 #1
[26969.636558] RIP: 0010:[<ffffffffa03df427>]  [<ffffffffa03df427>] :ocfs2:ocfs2_inode_l
ock_full+0x875/0xd08
[26969.636631] RSP: 0018:ffff8104174c1cf0  EFLAGS: 00010286
[26969.636662] RAX: 0000000000000091 RBX: ffff810355220000 RCX: 000000000000d8c7
[26969.636697] RDX: 000000000000d8c7 RSI: 0000000000000046 RDI: 0000000000000286
[26969.636732] RBP: ffff8103c5c394e8 R08: ffff8103c5c39100 R09: ffff8104174c1800
[26969.636766] R10: 0000000000000000 R11: 000001ff804fe0f0 R12: 0000000100080000
[26969.636801] R13: ffff8103c5c39100 R14: 0000000000000000 R15: ffff8103c5c39218
[26969.636836] FS:  00007f230680d6e0(0000) GS:ffff81042e48a8c0(0000) knlGS:0000000000000000
[26969.636889] CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
[26969.636921] CR2: 00007f93fd951000 CR3: 000000042b96f000 CR4: 00000000000006e0
[26969.636955] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
[26969.636990] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
[26969.637025] Process nfsd (pid: 2579, threadinfo ffff8104174c0000, task ffff81042e6e2100)
[26969.637078] Stack:  ffff81032bfe9812 ffff81042bfe9736 00000000b5168528 0000000000000000
[26969.637138]  0000000100000000 ffff8103d1073080 ffff8103c5cee8e8 ffff8103c5c394e8
[26969.637196]  ffff8103c5c394e8 ffff8104174c1db0 ffff81042cd0a800 ffff81042b3a8000
[26969.637235] Call Trace:
[26969.637295]  [<ffffffffa03eda7e>] ? :ocfs2:ocfs2_inode_revalidate+0x13a/0x214
[26969.637343]  [<ffffffffa03e98e9>] ? :ocfs2:ocfs2_getattr+0x78/0x167
[26969.637390]  [<ffffffffa02df513>] ? :nfsd:encode_post_op_attr+0x3e/0x84
[26969.637435]  [<ffffffffa02d8141>] ? :nfsd:nfsd_lookup+0xa1/0xae
[26969.637480]  [<ffffffffa02e091b>] ? :nfsd:nfs3svc_encode_diropres+0x32/0x62
[26969.637524]  [<ffffffffa02d32d2>] ? :nfsd:nfsd_dispatch+0x168/0x1b6
[26969.637577]  [<ffffffffa024859b>] ? :sunrpc:svc_process+0x408/0x6e9
[26969.637615]  [<ffffffff804295a4>] ? __down_read+0x12/0xa1
[26969.637655]  [<ffffffffa02d3671>] ? :nfsd:nfsd+0x0/0x2a3
[26969.637695]  [<ffffffffa02d3805>] ? :nfsd:nfsd+0x194/0x2a3
[26969.637729]  [<ffffffff802301ec>] ? schedule_tail+0x27/0x5c
[26969.637763]  [<ffffffff8020cf28>] ? child_rip+0xa/0x12
[26969.637802]  [<ffffffffa02d3671>] ? :nfsd:nfsd+0x0/0x2a3
[26969.637841]  [<ffffffffa02d3671>] ? :nfsd:nfsd+0x0/0x2a3
[26969.637882]  [<ffffffffa02d3671>] ? :nfsd:nfsd+0x0/0x2a3
[26969.637913]  [<ffffffff8020cf1e>] ? child_rip+0x0/0x12
[26969.637944] 
[26969.637966] 
[26969.638274]  RSP <ffff8104174c1cf0>
[26969.638655] ---[ end trace de22b7902a3499d8 ]


i Hope you can help me to find a way to resolv this problem 


sébastien.



More information about the Ocfs2-users mailing list