[Ocfs2-users] ocfs2 - Kernel panic on many write/read from both

Marek Królikowski admin at wset.edu.pl
Fri Dec 23 13:19:40 PST 2011


Hello
I get oops on TEST-MAIL2:

INFO: task ocfs2dc:15430 blocked for more than 120 seconds.
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
ocfs2dc         D ffff88107f232c40     0 15430      2 0x00000000
ffff881014889080 0000000000000046 ffff881000000000 ffff88102060c080
0000000000012c40 ffff88101eefbfd8 0000000000012c40 ffff88101eefa010
ffff88101eefbfd8 0000000000012c40 0000000000000001 00000001130a4380
Call Trace:
[<ffffffff8148db41>] ? __mutex_lock_slowpath+0xd1/0x140
[<ffffffff8148da53>] ? mutex_lock+0x23/0x40
[<ffffffff81181eb6>] ? dqget+0x246/0x3a0
[<ffffffff81182281>] ? __dquot_initialize+0x121/0x210
[<ffffffff8114c90d>] ? d_kill+0x9d/0x100
[<ffffffffa0a601c3>] ? ocfs2_find_local_alias+0x23/0x100 [ocfs2]
[<ffffffffa0a7fca8>] ? ocfs2_delete_inode+0x98/0x3e0 [ocfs2]
[<ffffffffa0a7106c>] ? ocfs2_unblock_lock+0x10c/0x770 [ocfs2]
[<ffffffffa0a80969>] ? ocfs2_evict_inode+0x19/0x40 [ocfs2]
[<ffffffff8114e9cc>] ? evict+0x8c/0x170
[<ffffffffa0a5fccd>] ? ocfs2_dentry_lock_put+0x5d/0x90 [ocfs2]
[<ffffffffa0a7177a>] ? ocfs2_process_blocked_lock+0xaa/0x280 [ocfs2]
[<ffffffff8107beb2>] ? prepare_to_wait+0x82/0x90
[<ffffffff8107bceb>] ? finish_wait+0x4b/0xa0
[<ffffffffa0a71aa0>] ? ocfs2_downconvert_thread+0x150/0x270 [ocfs2]
[<ffffffff8107bb60>] ? wake_up_bit+0x40/0x40
[<ffffffffa0a71950>] ? ocfs2_process_blocked_lock+0x280/0x280 [ocfs2]
[<ffffffffa0a71950>] ? ocfs2_process_blocked_lock+0x280/0x280 [ocfs2]
[<ffffffff8107b686>] ? kthread+0x96/0xa0
[<ffffffff81498a74>] ? kernel_thread_helper+0x4/0x10
[<ffffffff8107b5f0>] ? kthread_worker_fn+0x190/0x190
[<ffffffff81498a70>] ? gs_change+0x13/0x13
INFO: task kworker/0:1:30806 blocked for more than 120 seconds.
"echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.
kworker/0:1     D ffff88107f212c40     0 30806      2 0x00000000
ffff8810152f4080 0000000000000046 0000000000000000 ffffffff81a0d020
0000000000012c40 ffff880c28a57fd8 0000000000012c40 ffff880c28a56010
ffff880c28a57fd8 0000000000012c40 ffff880c28a57a08 00000001152f4080
Call Trace:
[<ffffffff8148d45d>] ? schedule_timeout+0x1ed/0x2d0
[<ffffffffa05244c0>] ? __jbd2_journal_file_buffer+0xd0/0x230 [jbd2]
[<ffffffff8148ce5c>] ? wait_for_common+0x12c/0x1a0
[<ffffffff81052230>] ? try_to_wake_up+0x280/0x280
[<ffffffff81085e21>] ? ktime_get+0x61/0xf0
[<ffffffffa0a6e850>] ? __ocfs2_cluster_lock+0x1f0/0x780 [ocfs2]
[<ffffffff81046fa7>] ? find_busiest_group+0x1f7/0xb00
[<ffffffffa0a73a56>] ? ocfs2_inode_lock_full_nested+0x126/0x540 [ocfs2]
[<ffffffffa0ad4da9>] ? ocfs2_lock_global_qf+0x29/0xd0 [ocfs2]
[<ffffffffa0ad4da9>] ? ocfs2_lock_global_qf+0x29/0xd0 [ocfs2]
[<ffffffffa0ad71df>] ? ocfs2_sync_dquot_helper+0xbf/0x330 [ocfs2]
[<ffffffffa0ad7120>] ? ocfs2_acquire_dquot+0x390/0x390 [ocfs2]
[<ffffffff81181c3a>] ? dquot_scan_active+0xda/0x110
[<ffffffffa0ad4ca0>] ? ocfs2_global_is_id+0x60/0x60 [ocfs2]
[<ffffffffa0ad4cc1>] ? qsync_work_fn+0x21/0x40 [ocfs2]
[<ffffffff810753f3>] ? process_one_work+0x123/0x450
[<ffffffff8107690b>] ? worker_thread+0x15b/0x370
[<ffffffff810767b0>] ? manage_workers+0x110/0x110
[<ffffffff810767b0>] ? manage_workers+0x110/0x110
[<ffffffff8107b686>] ? kthread+0x96/0xa0
[<ffffffff81498a74>] ? kernel_thread_helper+0x4/0x10
[<ffffffff8107b5f0>] ? kthread_worker_fn+0x190/0x190
[<ffffffff81498a70>] ? gs_change+0x13/0x13

And i can`t login to TEST-MAIL1 after give login and password console say 
when i lastlog but i don`t get bash - console don`t answer... but there is 
no OOPS or something like this on screen.
I don`t restart both server tell me what to do now.
Thanks


-----Oryginalna wiadomość----- 
From: srinivas eeda
Sent: Thursday, December 22, 2011 9:12 PM
To: Marek Królikowski
Cc: ocfs2-users at oss.oracle.com
Subject: Re: [Ocfs2-users] ocfs2 - Kernel panic on many write/read from both

We need to know what happened to node 2. Was the node rebooted because
of a network timeout or kernel panic? can you please configure
netconsole, serial console and rerun the test?

On 12/22/2011 8:08 AM, Marek Królikowski wrote:
> Hello
> After 24 hours i see TEST-MAIL2 reboot ( possible kernel panic) but 
> TEST-MAIL1 got in dmesg:
> TEST-MAIL1 ~ #dmesg
> [cut]
> o2net: accepted connection from node TEST-MAIL2 (num 1) at 
> 172.17.1.252:7777
> o2dlm: Node 1 joins domain B24C4493BBC74FEAA3371E2534BB3611
> o2dlm: Nodes in domain B24C4493BBC74FEAA3371E2534BB3611: 0 1
> o2net: connection to node TEST-MAIL2 (num 1) at 172.17.1.252:7777 has been 
> idle for 60.0 seconds, shutting it down.
> (swapper,0,0):o2net_idle_timer:1562 Here are some times that might help 
> debug the situation: (Timer: 33127732045, Now 33187808090, DataReady 
> 33127732039, Advance 33127732051-33127732051, Key 0xebb9cd47, Func 506, 
> FuncTime 33127732045-33127732048)
> o2net: no longer connected to node TEST-MAIL2 (num 1) at 172.17.1.252:7777
> (du,5099,12):dlm_do_master_request:1324 ERROR: link to 1 went down!
> (du,5099,12):dlm_get_lock_resource:907 ERROR: status = -112
> (dlm_thread,14321,1):dlm_send_proxy_ast_msg:484 ERROR: 
> B24C4493BBC74FEAA3371E2534BB3611: res M000000000000000000000cf023ef70, 
> error -112 send AST to node 1
> (dlm_thread,14321,1):dlm_flush_asts:605 ERROR: status = -112
> (dlm_thread,14321,1):dlm_send_proxy_ast_msg:484 ERROR: 
> B24C4493BBC74FEAA3371E2534BB3611: res P000000000000000000000000000000, 
> error -107 send AST to node 1
> (dlm_thread,14321,1):dlm_flush_asts:605 ERROR: status = -107
> (kworker/u:3,5071,0):o2net_connect_expired:1724 ERROR: no connection 
> established with node 1 after 60.0 seconds, giving up and returning 
> errors.
> (o2hb-B24C4493BB,14310,0):o2dlm_eviction_cb:267 o2dlm has evicted node 1 
> from group B24C4493BBC74FEAA3371E2534BB3611
> (ocfs2rec,5504,6):dlm_get_lock_resource:834 
> B24C4493BBC74FEAA3371E2534BB3611:M0000000000000000000015f023ef70: at least 
> one node (1) to recover before lock mastery can begin
> (ocfs2rec,5504,6):dlm_get_lock_resource:888 
> B24C4493BBC74FEAA3371E2534BB3611:M0000000000000000000015f023ef70: at least 
> one node (1) to recover before lock mastery can begin
> (du,5099,12):dlm_restart_lock_mastery:1213 ERROR: node down! 1
> (du,5099,12):dlm_wait_for_lock_mastery:1030 ERROR: status = -11
> (du,5099,12):dlm_get_lock_resource:888 
> B24C4493BBC74FEAA3371E2534BB3611:N000000000020924f: at least one node (1) 
> to recover before lock mastery can begin
> (dlm_reco_thread,14322,0):dlm_get_lock_resource:834 
> B24C4493BBC74FEAA3371E2534BB3611:$RECOVERY: at least one node (1) to 
> recover before lock mastery can begin
> (dlm_reco_thread,14322,0):dlm_get_lock_resource:868 
> B24C4493BBC74FEAA3371E2534BB3611: recovery map is not empty, but must 
> master $RECOVERY lock now
> (dlm_reco_thread,14322,0):dlm_do_recovery:523 (14322) Node 0 is the 
> Recovery Master for the Dead Node 1 for Domain 
> B24C4493BBC74FEAA3371E2534BB3611
> (ocfs2rec,5504,6):ocfs2_replay_journal:1549 Recovering node 1 from slot 1 
> on device (253,0)
> (ocfs2rec,5504,6):ocfs2_begin_quota_recovery:407 Beginning quota recovery 
> in slot 1
> (kworker/u:0,2909,0):ocfs2_finish_quota_recovery:599 Finishing quota 
> recovery in slot 1
>
> And i try give this command:
> debugfs.ocfs2 -l ENTRY EXIT DLM_GLUE QUOTA INODE DISK_ALLOC EXTENT_MAP 
> allow
> debugfs.ocfs2: Unable to write log mask "ENTRY": No such file or directory
> debugfs.ocfs2 -l ENTRY EXIT DLM_GLUE QUOTA INODE DISK_ALLOC EXTENT_MAP off
> debugfs.ocfs2: Unable to write log mask "ENTRY": No such file or directory
>
> But not working....
>
>
> -----Oryginalna wiadomość----- From: Srinivas Eeda
> Sent: Wednesday, December 21, 2011 8:43 PM
> To: Marek Królikowski
> Cc: ocfs2-users at oss.oracle.com
> Subject: Re: [Ocfs2-users] ocfs2 - Kernel panic on many write/read from 
> both
>
> Those numbers look good. Basically with the fixes backed out and another
> fix I gave, you are not seeing that many orphans hanging around and
> hence not seeing the process stuck kernel stacks. You can run the test
> longer or if you are satisfied, please enable quotas and re-run the test
> with the modified kernel. You might see a dead lock which needs to be
> fixed(I was not able to reproduce this yet). If the system hangs, please
> capture the following and provide me the output
>
> 1. echo t > /proc/sysrq-trigger
> 2. debugfs.ocfs2 -l ENTRY EXIT DLM_GLUE QUOTA INODE DISK_ALLOC EXTENT_MAP 
> allow
> 3. wait for 10 minutes
> 4. debugfs.ocfs2 -l ENTRY EXIT DLM_GLUE QUOTA INODE DISK_ALLOC EXTENT_MAP 
> off
> 5. echo t > /proc/sysrq-trigger
> 




More information about the Ocfs2-users mailing list