[Ocfs2-users] system crash

Sunil Mushran sunil.mushran at oracle.com
Thu Feb 11 12:54:53 PST 2010


No. Very hard to hit it.

Brad Plant wrote:
> Hi Sunil,
>
> Did this also go into the 2.6.32.y stable branch?
>
> Cheers,
>
> Brad
>
>
> On Thu, 11 Feb 2010 10:50:24 -0800
> Sunil Mushran <sunil.mushran at oracle.com> wrote:
>
>   
>> It looks similar to bz#1202.
>> http://oss.oracle.com/bugzilla/show_bug.cgi?id=1202
>>
>> File a bugzilla with novell. They should be able to look at the
>> objdump and confirm.
>>
>> If so, the fix is already in mainline.
>> http://git.kernel.org/?p=linux/kernel/git/torvalds/linux-2.6.git;a=commitdiff;h=71656fa6ec10473eb9b646c10a2173fdea2f83c9
>>
>> Sunil
>>
>> Charlie Sharkey wrote:
>>     
>>>  
>>>
>>> We had a system crash while unmounting some volumes. It is a 6 node
>>>
>>> cluster, hostnames are n1 -> n6. Info below,  Any ideas ?
>>>
>>>  
>>>
>>> Thanks in advance,
>>>
>>>  
>>>
>>> charlie
>>>
>>>  
>>>
>>>  
>>>
>>>  
>>>
>>> ocfs2-tools-1.4.0-0.5
>>>
>>> ocfs2console-1.4.0-0.5
>>>
>>> Linux sr2600-1 2.6.16.60-0.34-smp #1 SMP Fri Jan 16 14:59:01 UTC 2009 
>>> x86_64 x86_64 x86_64 GNU/Linux
>>>
>>>  
>>>
>>> OCFS2 Node Manager 1.4.1-1-SLES Wed Jul 23 18:33:42 UTC 2008 (build 
>>> f922955d99ef972235bd0c1fc236c5ddbb368611)
>>>
>>> o2cb heartbeat: registered disk 
>>> mode                                                                          
>>>
>>>
>>> OCFS2 DLM 1.4.1-1-SLES Wed Jul 23 18:33:42 UTC 2008 (build 
>>> f922955d99ef972235bd0c1fc236c5ddbb368611)         
>>>
>>> OCFS2 DLMFS 1.4.1-1-SLES Wed Jul 23 18:33:42 UTC 2008 (build 
>>> f922955d99ef972235bd0c1fc236c5ddbb368611)       
>>>                                                                      
>>>
>>>  
>>>
>>> from /var/log/messages
>>>
>>> Feb  2 15:04:05 n2 kernel: ocfs2_dlm: Node 3 leaves domain 
>>> 27169AFBA10C425B91D386F55FC37AF5
>>>
>>> Feb  2 15:04:05 n2 kernel: ocfs2_dlm: Nodes in domain 
>>> ("27169AFBA10C425B91D386F55FC37AF5"): 1 2 4
>>>
>>> Feb  2 15:04:05 n2 kernel: ocfs2_dlm: Node 4 leaves domain 
>>> 27169AFBA10C425B91D386F55FC37AF5
>>>
>>> Feb  2 15:04:05 n2 kernel: ocfs2_dlm: Nodes in domain 
>>> ("27169AFBA10C425B91D386F55FC37AF5"): 1 2
>>>
>>> Feb  2 15:04:05 n2 kernel: ocfs2: Unmounting device (253,8) on (node 1)
>>>
>>> Feb  2 15:04:05 n2 kernel: ocfs2_dlm: Node 5 leaves domain 
>>> 65EAD2E63E7549B69335BCAD58C5A476
>>>
>>> Feb  2 15:04:05 n2 kernel: ocfs2_dlm: Nodes in domain 
>>> ("65EAD2E63E7549B69335BCAD58C5A476"): 0 1 2 3 4
>>>
>>> Feb  2 15:04:06 n2 kernel: o2net: no longer connected to node n6 (num 
>>> 5) at 192.168.100.60:7777
>>>
>>> Feb  2 15:04:08 n2 kernel: ocfs2_dlm: Node 0 leaves domain 
>>> 65EAD2E63E7549B69335BCAD58C5A476
>>>
>>> Feb  2 15:04:08 n2 kernel: ocfs2_dlm: Nodes in domain 
>>> ("65EAD2E63E7549B69335BCAD58C5A476"): 1 2 3 4
>>>
>>> Feb  2 15:04:09 n2 kernel: o2net: no longer connected to node n1 (num 
>>> 0) at 192.168.100.10:7777
>>>
>>> Feb  2 15:04:09 n2 kernel: ocfs2: Unmounting device (253,9) on (node 1)
>>>
>>> Feb  2 15:04:13 n2 kernel: ocfs2: Unmounting device (253,6) on (node 1)
>>>
>>> Feb  2 15:04:18 n2 kernel: ocfs2: Unmounting device (253,7) on (node 1)
>>>
>>>  Feb  2 15:07:22 n2 syslog-ng[5679]: syslog-ng version 1.6.8 
>>> starting               ß------------ system restart
>>>
>>>  
>>>
>>>  
>>>
>>> stack trace from core dump file:
>>>
>>> PID: 14669  TASK: ffff8101289cb7b0  CPU: 1   COMMAND: "umount"
>>>
>>>  #0 [ffff8101040ef810] machine_kexec at ffffffff8011c036
>>>
>>>  #1 [ffff8101040ef8e0] crash_kexec   at ffffffff80153ea9
>>>
>>>  #2 [ffff8101040ef9a0] __die         at ffffffff802ec510
>>>
>>>  #3 [ffff8101040ef9e0] die           at ffffffff8010c786
>>>
>>>  #4 [ffff8101040efa10] do_invalid_op at ffffffff8010cd2d
>>>
>>>  #5 [ffff8101040efad0] error_exit    at ffffffff8010bc91
>>>
>>>     [exception RIP: dlm_add_lock_to_array+249]
>>>
>>>     RIP: ffffffff8852750f  RSP: ffff8101040efb88  RFLAGS: 00010286
>>>
>>>     RAX: ffff810103897ac8  RBX: ffff810103e02ec0  RCX: 0000000000016a72
>>>
>>>     RDX: 0000000000000000  RSI: 0000000000000292  RDI: ffffffff8035a99c
>>>
>>>     RBP: ffff810104056000   R8: ffffffff8045a260   R9: 0000000000000001
>>>
>>>     R10: ffff810103897a80  R11: 0000000000000000  R12: ffff810104056080
>>>
>>>     R13: 0000000000000000  R14: 0000000000000001  R15: ffff810104056000
>>>
>>>     ORIG_RAX: ffffffffffffffff  CS: 0010  SS: 0018
>>>
>>>  #6 [ffff8101040efb80] dlm_add_lock_to_array  at ffffffff8852750f
>>>
>>>  #7 [ffff8101040efba0] dlm_send_one_lockres   at ffffffff88527a55
>>>
>>>  #8 [ffff8101040efc80] dlm_empty_lockres      at ffffffff88520ae2
>>>
>>>  #9 [ffff8101040efdb0] ocfs2_dlm_shutdown     at ffffffff8856fc8a
>>>
>>> #10 [ffff8101040efdd0] ocfs2_dismount_volume  at ffffffff8859416c
>>>
>>> #11 [ffff8101040efe40] ocfs2_put_super        at ffffffff88594668
>>>
>>> #12 [ffff8101040efe50] generic_shutdown_super at ffffffff8018c305
>>>
>>> #13 [ffff8101040efe70] kill_block_super       at ffffffff8018c3d6
>>>
>>> #14 [ffff8101040efe90] deactivate_super       at ffffffff8018c4ac
>>>
>>> #15 [ffff8101040efeb0] sys_umount             at ffffffff801a1376
>>>
>>> #16 [ffff8101040eff80] system_call            at ffffffff8010adba
>>>
>>>     RIP: 00002b4ba5ba0987  RSP: 00007fff050e3990  RFLAGS: 00010246
>>>
>>>     RAX: 00000000000000a6  RBX: ffffffff8010adba  RCX: 00007fff050e4140
>>>
>>>     RDX: 000000000000006e  RSI: 0000000000000001  RDI: 00000000005101e0
>>>
>>>     RBP: 00000000005101c0   R8: 000000000000006f   R9: fefeff344d544b2e
>>>
>>>     R10: 0000000000000000  R11: 0000000000000206  R12: 0000000000510250
>>>
>>>     R13: 00000000005101e0  R14: 0000000000510200  R15: 00007fff050e42e0
>>>
>>>     ORIG_RAX: 00000000000000a6  CS: 0033  SS: 002b
>>>
>>>  
>>>
>>>  
>>>
>>>  
>>>
>>>  
>>>
>>>  
>>>
>>>  
>>>
>>>  
>>>
>>>  
>>>
>>>  
>>>
>>> ------------------------------------------------------------------------
>>>
>>> _______________________________________________
>>> Ocfs2-users mailing list
>>> Ocfs2-users at oss.oracle.com
>>> http://oss.oracle.com/mailman/listinfo/ocfs2-users
>>>       
>> _______________________________________________
>> Ocfs2-users mailing list
>> Ocfs2-users at oss.oracle.com
>> http://oss.oracle.com/mailman/listinfo/ocfs2-users
>>     
>> ------------------------------------------------------------------------
>>
>> _______________________________________________
>> Ocfs2-users mailing list
>> Ocfs2-users at oss.oracle.com
>> http://oss.oracle.com/mailman/listinfo/ocfs2-users




More information about the Ocfs2-users mailing list