[Ocfs2-users] ocfs2 - Kernel panic on many write/read from both servers

Marek Królikowski admin at wset.edu.pl
Wed Dec 7 12:00:38 PST 2011


Hey
This is log from fsck:

fsck.ocfs2 /dev/dm-0
fsck.ocfs2 1.6.4
Checking OCFS2 filesystem in /dev/dm-0:
  Label:              NEW-PLK
  UUID:               9FBF7DCC2D5842E8B3C7A30B058DC255
  Number of blocks:   422137056
  Block size:         4096
  Number of clusters: 422137056
  Cluster size:       4096
  Number of slots:    2

/dev/dm-0 wasn't cleanly unmounted by all nodes.  Attempting to replay the 
journals for nodes that didn't unmount cleanly
Checking each slot's journal.
Replaying slot 0's journal.
Slot 0's journal replayed successfully.
Replaying slot 1's journal.
Slot 1's journal replayed successfully.
Slot 0's local alloc replayed successfully
Slot 1's local alloc replayed successfully
Slot 1's truncate log replayed successfully
Pass 1: Checking inodes and blocks.
[CLUSTER_ALLOC_BIT] Cluster 792097 is marked in the global cluster bitmap 
but it isn't in use.  Clear its bit in the bitmap? <y> y
[CLUSTER_ALLOC_BIT] Cluster 792482 is marked in the global cluster bitmap 
but it isn't in use.  Clear its bit in the bitmap? <y> y
[CLUSTER_ALLOC_BIT] Cluster 795026 is marked in the global cluster bitmap 
but it isn't in use.  Clear its bit in the bitmap? <y> y
[cut many many here this same infor with other Cluster  number]
[CLUSTER_ALLOC_BIT] Cluster 16568868 is marked in the global cluster bitmap 
but it isn't in use.  Clear its bit in the bitmap? <y> y
Pass 2: Checking directory entries. 
Pass 3: Checking directory connectivity.
Pass 4a: checking for orphaned inodes
Pass 4b: Checking inodes link counts.
Pass 5: Checking quota information.
All passes succeeded.

Thanks


-----Oryginalna wiadomość----- 
From: Eduardo Diaz - Gmail
Sent: Wednesday, December 07, 2011 11:42 AM
To: Marek Krolikowski
Cc: ocfs2-users at oss.oracle.com
Subject: Re: [Ocfs2-users] ocfs2 - Kernel panic on many write/read from both 
servers

You can make a fsck.ocfs2 to see if the filesystem are broken?

On Wed, Dec 7, 2011 at 11:30 AM, Marek Krolikowski <admin at wset.edu.pl> 
wrote:
> Hello
> I use: sys-fs/ocfs2-tools-1.6.4 and create file system with all features:
> mkfs.ocfs2 -N 2 -L MAIL --fs-feature-level=max-features /dev/dm-0
> and after this got kernel panic :(
>
>
>
> -----Oryginalna wiadomość----- From: Eduardo Diaz - Gmail
> Sent: Wednesday, December 07, 2011 11:08 AM
> To: Marek Krolikowski
>
> Cc: ocfs2-users at oss.oracle.com
> Subject: Re: [Ocfs2-users] ocfs2 - Kernel panic on many write/read from 
> both
> servers
>
> Try to use other filesystem for example xfs and make a full test.
>
> Try create a one node cluster and filesystem and make tests..
> (fileindex can be the problem)
>
> make a fsck to the filesystem
>
> Try upgrade ocfs2 to last version and use the max features, only has
> two nodes?..
>
> I will do. make a backup, create a new filesystem with all features
> that you need and make mkfs. the cluster only with the number of the
> nodes that you will use.
>
> Restore de data.
>
> Make extensive test for a week before put in production :-)..
>
> On Tue, Dec 6, 2011 at 2:04 PM, Marek Krolikowski <admin at wset.edu.pl> 
> wrote:
>>
>> hey m8
>> Like i say i am not expert too but when i use ext3 and write/read working
>> with np.
>>
>>
>> -----Oryginalna wiadomość----- From: Eduardo Diaz - Gmail
>> Sent: Tuesday, December 06, 2011 3:06 AM
>> To: Marek Królikowski
>> Cc: ocfs2-users at oss.oracle.com
>> Subject: Re: [Ocfs2-users] ocfs2 - Kernel panic on many write/read from
>> both
>> servers
>>
>>
>> I am not a expert, but you have a problem in you EMC system (multipath
>> system) or drivers.
>>
>> Did you test this before put in production? or test this NAS with
>> other filesystem, xfs for example?.
>>
>> As I read I see" that hung_task_timeout_secs" some task wait more that
>> 120 seg may be a problem of you EMC/Fiber/cable problem...
>>
>> 2011/12/4 Marek Królikowski <admin at wset.edu.pl>:
>>>
>>>
>>> I do for all night tests with write/read files from ocfs2 on both 
>>> servers
>>> something like this:
>>> On MAIL1 server:
>>> #!/bin/bash
>>> while true
>>> do
>>> rm -rf /mnt/EMC/MAIL1
>>> mkdir /mnt/EMC/MAIL1
>>> cp -r /usr /mnt/EMC/MAIL1
>>> rm -rf /mnt/EMC/MAIL1
>>> done;
>>> On MAIL2 server:
>>> #!/bin/bash
>>> while true
>>> do
>>> rm -rf /mnt/EMC/MAIL2
>>> mkdir /mnt/EMC/MAIL2
>>> cp -r /usr /mnt/EMC/MAIL2
>>> rm -rf /mnt/EMC/MAIL2
>>> done;
>>>
>>> Today i check logs and see:
>>> o2dlm: Node 1 joins domain EAC7942B71964050AE2046D3F0CDD7B2
>>> o2dlm: Nodes in domain EAC7942B71964050AE2046D3F0CDD7B2: 0 1
>>> (rm,26136,0):ocfs2_unlink:953 ERROR: status = -2
>>> (touch,26137,0):ocfs2_check_dir_for_entry:2120 ERROR: status = -17
>>> (touch,26137,0):ocfs2_mknod:461 ERROR: status = -17
>>> (touch,26137,0):ocfs2_create:631 ERROR: status = -17
>>> (rm,26142,0):ocfs2_unlink:953 ERROR: status = -2
>>> INFO: task kworker/u:2:20246 blocked for more than 120 seconds.
>>> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
>>> message.
>>> kworker/u:2     D ffff88107f4525c0     0 20246      2 0x00000000
>>> ffff880b730b57d0 0000000000000046 ffff8810201297d0 00000000000125c0
>>> ffff880f5a399fd8 00000000000125c0 00000000000125c0 00000000000125c0
>>> ffff880f5a398000 00000000000125c0 ffff880f5a399fd8 00000000000125c0
>>> Call Trace:
>>> [<ffffffff81481b71>] ? __mutex_lock_slowpath+0xd1/0x140
>>> [<ffffffff814818d3>] ? mutex_lock+0x23/0x40
>>> [<ffffffffa0937d95>] ? ocfs2_wipe_inode+0x105/0x690 [ocfs2]
>>> [<ffffffffa0935cfb>] ? ocfs2_query_inode_wipe.clone.9+0xcb/0x370 [ocfs2]
>>> [<ffffffffa09385a4>] ? ocfs2_delete_inode+0x284/0x3f0 [ocfs2]
>>> [<ffffffffa0919a10>] ? ocfs2_dentry_attach_lock+0x5a0/0x5a0 [ocfs2]
>>> [<ffffffffa093872e>] ? ocfs2_evict_inode+0x1e/0x50 [ocfs2]
>>> [<ffffffff81145900>] ? evict+0x70/0x140
>>> [<ffffffffa0919322>] ? __ocfs2_drop_dl_inodes.clone.2+0x32/0x60 [ocfs2]
>>> [<ffffffffa0919a39>] ? ocfs2_drop_dl_inodes+0x29/0x90 [ocfs2]
>>> [<ffffffff8106e56f>] ? process_one_work+0x11f/0x440
>>> [<ffffffff8106f279>] ? worker_thread+0x159/0x330
>>> [<ffffffff8106f120>] ? manage_workers.clone.21+0x120/0x120
>>> [<ffffffff8106f120>] ? manage_workers.clone.21+0x120/0x120
>>> [<ffffffff81073fa6>] ? kthread+0x96/0xa0
>>> [<ffffffff8148bb24>] ? kernel_thread_helper+0x4/0x10
>>> [<ffffffff81073f10>] ? kthread_worker_fn+0x1a0/0x1a0
>>> [<ffffffff8148bb20>] ? gs_change+0x13/0x13
>>> INFO: task rm:5192 blocked for more than 120 seconds.
>>> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
>>> message.
>>> rm              D ffff88107f2725c0     0  5192  16338 0x00000000
>>> ffff881014ccb040 0000000000000082 ffff8810206b8040 00000000000125c0
>>> ffff8804d7697fd8 00000000000125c0 00000000000125c0 00000000000125c0
>>> ffff8804d7696000 00000000000125c0 ffff8804d7697fd8 00000000000125c0
>>> Call Trace:
>>> [<ffffffff8148148d>] ? schedule_timeout+0x1ed/0x2e0
>>> [<ffffffffa0886162>] ? dlmconvert_master+0xe2/0x190 [ocfs2_dlm]
>>> [<ffffffffa08878bf>] ? dlmlock+0x7f/0xb70 [ocfs2_dlm]
>>> [<ffffffff81480e0a>] ? wait_for_common+0x13a/0x190
>>> [<ffffffff8104bc50>] ? try_to_wake_up+0x280/0x280
>>> [<ffffffffa0928a38>] ? __ocfs2_cluster_lock.clone.21+0x1d8/0x6b0 [ocfs2]
>>> [<ffffffffa0928fcc>] ? ocfs2_inode_lock_full_nested+0xbc/0x490 [ocfs2]
>>> [<ffffffffa0943c1b>] ? ocfs2_lookup_lock_orphan_dir+0x6b/0x1b0 [ocfs2]
>>> [<ffffffffa09454ba>] ? ocfs2_prepare_orphan_dir+0x4a/0x280 [ocfs2]
>>> [<ffffffffa094616f>] ? ocfs2_unlink+0x6ef/0xb90 [ocfs2]
>>> [<ffffffff811b35a9>] ? may_link.clone.22+0xd9/0x170
>>> [<ffffffff8113aa58>] ? vfs_unlink+0x98/0x100
>>> [<ffffffff8113ac41>] ? do_unlinkat+0x181/0x1b0
>>> [<ffffffff8113e7cd>] ? vfs_readdir+0x9d/0xe0
>>> [<ffffffff811653d8>] ? fsnotify_find_inode_mark+0x28/0x40
>>> [<ffffffff81166324>] ? dnotify_flush+0x54/0x110
>>> [<ffffffff8112b07f>] ? filp_close+0x5f/0x90
>>> [<ffffffff8148aa12>] ? system_call_fastpath+0x16/0x1b
>>> INFO: task kworker/u:2:20246 blocked for more than 120 seconds.
>>> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
>>> message.
>>> kworker/u:2     D ffff88107f4525c0     0 20246      2 0x00000000
>>> ffff880b730b57d0 0000000000000046 ffff8810201297d0 00000000000125c0
>>> ffff880f5a399fd8 00000000000125c0 00000000000125c0 00000000000125c0
>>> ffff880f5a398000 00000000000125c0 ffff880f5a399fd8 00000000000125c0
>>> Call Trace:
>>> [<ffffffff81481b71>] ? __mutex_lock_slowpath+0xd1/0x140
>>> [<ffffffff814818d3>] ? mutex_lock+0x23/0x40
>>> [<ffffffffa0937d95>] ? ocfs2_wipe_inode+0x105/0x690 [ocfs2]
>>> [<ffffffffa0935cfb>] ? ocfs2_query_inode_wipe.clone.9+0xcb/0x370 [ocfs2]
>>> [<ffffffffa09385a4>] ? ocfs2_delete_inode+0x284/0x3f0 [ocfs2]
>>> [<ffffffffa0919a10>] ? ocfs2_dentry_attach_lock+0x5a0/0x5a0 [ocfs2]
>>> [<ffffffffa093872e>] ? ocfs2_evict_inode+0x1e/0x50 [ocfs2]
>>> [<ffffffff81145900>] ? evict+0x70/0x140
>>> [<ffffffffa0919322>] ? __ocfs2_drop_dl_inodes.clone.2+0x32/0x60 [ocfs2]
>>> [<ffffffffa0919a39>] ? ocfs2_drop_dl_inodes+0x29/0x90 [ocfs2]
>>> [<ffffffff8106e56f>] ? process_one_work+0x11f/0x440
>>> [<ffffffff8106f279>] ? worker_thread+0x159/0x330
>>> [<ffffffff8106f120>] ? manage_workers.clone.21+0x120/0x120
>>> [<ffffffff8106f120>] ? manage_workers.clone.21+0x120/0x120
>>> [<ffffffff81073fa6>] ? kthread+0x96/0xa0
>>> [<ffffffff8148bb24>] ? kernel_thread_helper+0x4/0x10
>>> [<ffffffff81073f10>] ? kthread_worker_fn+0x1a0/0x1a0
>>> [<ffffffff8148bb20>] ? gs_change+0x13/0x13
>>> INFO: task rm:5192 blocked for more than 120 seconds.
>>> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
>>> message.
>>> rm              D ffff88107f2725c0     0  5192  16338 0x00000000
>>> ffff881014ccb040 0000000000000082 ffff8810206b8040 00000000000125c0
>>> ffff8804d7697fd8 00000000000125c0 00000000000125c0 00000000000125c0
>>> ffff8804d7696000 00000000000125c0 ffff8804d7697fd8 00000000000125c0
>>> Call Trace:
>>> [<ffffffff8148148d>] ? schedule_timeout+0x1ed/0x2e0
>>> [<ffffffffa0886162>] ? dlmconvert_master+0xe2/0x190 [ocfs2_dlm]
>>> [<ffffffffa08878bf>] ? dlmlock+0x7f/0xb70 [ocfs2_dlm]
>>> [<ffffffff81480e0a>] ? wait_for_common+0x13a/0x190
>>> [<ffffffff8104bc50>] ? try_to_wake_up+0x280/0x280
>>> [<ffffffffa0928a38>] ? __ocfs2_cluster_lock.clone.21+0x1d8/0x6b0 [ocfs2]
>>> [<ffffffffa0928fcc>] ? ocfs2_inode_lock_full_nested+0xbc/0x490 [ocfs2]
>>> [<ffffffffa0943c1b>] ? ocfs2_lookup_lock_orphan_dir+0x6b/0x1b0 [ocfs2]
>>> [<ffffffffa09454ba>] ? ocfs2_prepare_orphan_dir+0x4a/0x280 [ocfs2]
>>> [<ffffffffa094616f>] ? ocfs2_unlink+0x6ef/0xb90 [ocfs2]
>>> [<ffffffff811b35a9>] ? may_link.clone.22+0xd9/0x170
>>> [<ffffffff8113aa58>] ? vfs_unlink+0x98/0x100
>>> [<ffffffff8113ac41>] ? do_unlinkat+0x181/0x1b0
>>> [<ffffffff8113e7cd>] ? vfs_readdir+0x9d/0xe0
>>> [<ffffffff811653d8>] ? fsnotify_find_inode_mark+0x28/0x40
>>> [<ffffffff81166324>] ? dnotify_flush+0x54/0x110
>>> [<ffffffff8112b07f>] ? filp_close+0x5f/0x90
>>> [<ffffffff8148aa12>] ? system_call_fastpath+0x16/0x1b
>>> INFO: task kworker/u:2:20246 blocked for more than 120 seconds.
>>> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
>>> message.
>>> kworker/u:2     D ffff88107f4525c0     0 20246      2 0x00000000
>>> ffff880b730b57d0 0000000000000046 ffff8810201297d0 00000000000125c0
>>> ffff880f5a399fd8 00000000000125c0 00000000000125c0 00000000000125c0
>>> ffff880f5a398000 00000000000125c0 ffff880f5a399fd8 00000000000125c0
>>> Call Trace:
>>> [<ffffffff81481b71>] ? __mutex_lock_slowpath+0xd1/0x140
>>> [<ffffffff814818d3>] ? mutex_lock+0x23/0x40
>>> [<ffffffffa0937d95>] ? ocfs2_wipe_inode+0x105/0x690 [ocfs2]
>>> [<ffffffffa0935cfb>] ? ocfs2_query_inode_wipe.clone.9+0xcb/0x370 [ocfs2]
>>> [<ffffffffa09385a4>] ? ocfs2_delete_inode+0x284/0x3f0 [ocfs2]
>>> [<ffffffffa0919a10>] ? ocfs2_dentry_attach_lock+0x5a0/0x5a0 [ocfs2]
>>> [<ffffffffa093872e>] ? ocfs2_evict_inode+0x1e/0x50 [ocfs2]
>>> [<ffffffff81145900>] ? evict+0x70/0x140
>>> [<ffffffffa0919322>] ? __ocfs2_drop_dl_inodes.clone.2+0x32/0x60 [ocfs2]
>>> [<ffffffffa0919a39>] ? ocfs2_drop_dl_inodes+0x29/0x90 [ocfs2]
>>> [<ffffffff8106e56f>] ? process_one_work+0x11f/0x440
>>> [<ffffffff8106f279>] ? worker_thread+0x159/0x330
>>> [<ffffffff8106f120>] ? manage_workers.clone.21+0x120/0x120
>>> [<ffffffff8106f120>] ? manage_workers.clone.21+0x120/0x120
>>> [<ffffffff81073fa6>] ? kthread+0x96/0xa0
>>> [<ffffffff8148bb24>] ? kernel_thread_helper+0x4/0x10
>>> [<ffffffff81073f10>] ? kthread_worker_fn+0x1a0/0x1a0
>>> [<ffffffff8148bb20>] ? gs_change+0x13/0x13
>>> INFO: task rm:5192 blocked for more than 120 seconds.
>>> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
>>> message.
>>> rm              D ffff88107f2725c0     0  5192  16338 0x00000000
>>> ffff881014ccb040 0000000000000082 ffff8810206b8040 00000000000125c0
>>> ffff8804d7697fd8 00000000000125c0 00000000000125c0 00000000000125c0
>>> ffff8804d7696000 00000000000125c0 ffff8804d7697fd8 00000000000125c0
>>> Call Trace:
>>> [<ffffffff8148148d>] ? schedule_timeout+0x1ed/0x2e0
>>> [<ffffffffa0886162>] ? dlmconvert_master+0xe2/0x190 [ocfs2_dlm]
>>> [<ffffffffa08878bf>] ? dlmlock+0x7f/0xb70 [ocfs2_dlm]
>>> [<ffffffff81480e0a>] ? wait_for_common+0x13a/0x190
>>> [<ffffffff8104bc50>] ? try_to_wake_up+0x280/0x280
>>> [<ffffffffa0928a38>] ? __ocfs2_cluster_lock.clone.21+0x1d8/0x6b0 [ocfs2]
>>> [<ffffffffa0928fcc>] ? ocfs2_inode_lock_full_nested+0xbc/0x490 [ocfs2]
>>> [<ffffffffa0943c1b>] ? ocfs2_lookup_lock_orphan_dir+0x6b/0x1b0 [ocfs2]
>>> [<ffffffffa09454ba>] ? ocfs2_prepare_orphan_dir+0x4a/0x280 [ocfs2]
>>> [<ffffffffa094616f>] ? ocfs2_unlink+0x6ef/0xb90 [ocfs2]
>>> [<ffffffff811b35a9>] ? may_link.clone.22+0xd9/0x170
>>> [<ffffffff8113aa58>] ? vfs_unlink+0x98/0x100
>>> [<ffffffff8113ac41>] ? do_unlinkat+0x181/0x1b0
>>> [<ffffffff8113e7cd>] ? vfs_readdir+0x9d/0xe0
>>> [<ffffffff811653d8>] ? fsnotify_find_inode_mark+0x28/0x40
>>> [<ffffffff81166324>] ? dnotify_flush+0x54/0x110
>>> [<ffffffff8112b07f>] ? filp_close+0x5f/0x90
>>> [<ffffffff8148aa12>] ? system_call_fastpath+0x16/0x1b
>>> INFO: task kworker/u:2:20246 blocked for more than 120 seconds.
>>> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
>>> message.
>>> kworker/u:2     D ffff88107f4525c0     0 20246      2 0x00000000
>>> ffff880b730b57d0 0000000000000046 ffff8810201297d0 00000000000125c0
>>> ffff880f5a399fd8 00000000000125c0 00000000000125c0 00000000000125c0
>>> ffff880f5a398000 00000000000125c0 ffff880f5a399fd8 00000000000125c0
>>> Call Trace:
>>> [<ffffffff81481b71>] ? __mutex_lock_slowpath+0xd1/0x140
>>> [<ffffffff814818d3>] ? mutex_lock+0x23/0x40
>>> [<ffffffffa0937d95>] ? ocfs2_wipe_inode+0x105/0x690 [ocfs2]
>>> [<ffffffffa0935cfb>] ? ocfs2_query_inode_wipe.clone.9+0xcb/0x370 [ocfs2]
>>> [<ffffffffa09385a4>] ? ocfs2_delete_inode+0x284/0x3f0 [ocfs2]
>>> [<ffffffffa0919a10>] ? ocfs2_dentry_attach_lock+0x5a0/0x5a0 [ocfs2]
>>> [<ffffffffa093872e>] ? ocfs2_evict_inode+0x1e/0x50 [ocfs2]
>>> [<ffffffff81145900>] ? evict+0x70/0x140
>>> [<ffffffffa0919322>] ? __ocfs2_drop_dl_inodes.clone.2+0x32/0x60 [ocfs2]
>>> [<ffffffffa0919a39>] ? ocfs2_drop_dl_inodes+0x29/0x90 [ocfs2]
>>> [<ffffffff8106e56f>] ? process_one_work+0x11f/0x440
>>> [<ffffffff8106f279>] ? worker_thread+0x159/0x330
>>> [<ffffffff8106f120>] ? manage_workers.clone.21+0x120/0x120
>>> [<ffffffff8106f120>] ? manage_workers.clone.21+0x120/0x120
>>> [<ffffffff81073fa6>] ? kthread+0x96/0xa0
>>> [<ffffffff8148bb24>] ? kernel_thread_helper+0x4/0x10
>>> [<ffffffff81073f10>] ? kthread_worker_fn+0x1a0/0x1a0
>>> [<ffffffff8148bb20>] ? gs_change+0x13/0x13
>>> INFO: task rm:5192 blocked for more than 120 seconds.
>>> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
>>> message.
>>> rm              D ffff88107f2725c0     0  5192  16338 0x00000000
>>> ffff881014ccb040 0000000000000082 ffff8810206b8040 00000000000125c0
>>> ffff8804d7697fd8 00000000000125c0 00000000000125c0 00000000000125c0
>>> ffff8804d7696000 00000000000125c0 ffff8804d7697fd8 00000000000125c0
>>> Call Trace:
>>> [<ffffffff8148148d>] ? schedule_timeout+0x1ed/0x2e0
>>> [<ffffffffa0886162>] ? dlmconvert_master+0xe2/0x190 [ocfs2_dlm]
>>> [<ffffffffa08878bf>] ? dlmlock+0x7f/0xb70 [ocfs2_dlm]
>>> [<ffffffff81480e0a>] ? wait_for_common+0x13a/0x190
>>> [<ffffffff8104bc50>] ? try_to_wake_up+0x280/0x280
>>> [<ffffffffa0928a38>] ? __ocfs2_cluster_lock.clone.21+0x1d8/0x6b0 [ocfs2]
>>> [<ffffffffa0928fcc>] ? ocfs2_inode_lock_full_nested+0xbc/0x490 [ocfs2]
>>> [<ffffffffa0943c1b>] ? ocfs2_lookup_lock_orphan_dir+0x6b/0x1b0 [ocfs2]
>>> [<ffffffffa09454ba>] ? ocfs2_prepare_orphan_dir+0x4a/0x280 [ocfs2]
>>> [<ffffffffa094616f>] ? ocfs2_unlink+0x6ef/0xb90 [ocfs2]
>>> [<ffffffff811b35a9>] ? may_link.clone.22+0xd9/0x170
>>> [<ffffffff8113aa58>] ? vfs_unlink+0x98/0x100
>>> [<ffffffff8113ac41>] ? do_unlinkat+0x181/0x1b0
>>> [<ffffffff8113e7cd>] ? vfs_readdir+0x9d/0xe0
>>> [<ffffffff811653d8>] ? fsnotify_find_inode_mark+0x28/0x40
>>> [<ffffffff81166324>] ? dnotify_flush+0x54/0x110
>>> [<ffffffff8112b07f>] ? filp_close+0x5f/0x90
>>> [<ffffffff8148aa12>] ? system_call_fastpath+0x16/0x1b
>>> INFO: task kworker/u:2:20246 blocked for more than 120 seconds.
>>> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
>>> message.
>>> kworker/u:2     D ffff88107f4525c0     0 20246      2 0x00000000
>>> ffff880b730b57d0 0000000000000046 ffff8810201297d0 00000000000125c0
>>> ffff880f5a399fd8 00000000000125c0 00000000000125c0 00000000000125c0
>>> ffff880f5a398000 00000000000125c0 ffff880f5a399fd8 00000000000125c0
>>> Call Trace:
>>> [<ffffffff81481b71>] ? __mutex_lock_slowpath+0xd1/0x140
>>> [<ffffffff814818d3>] ? mutex_lock+0x23/0x40
>>> [<ffffffffa0937d95>] ? ocfs2_wipe_inode+0x105/0x690 [ocfs2]
>>> [<ffffffffa0935cfb>] ? ocfs2_query_inode_wipe.clone.9+0xcb/0x370 [ocfs2]
>>> [<ffffffffa09385a4>] ? ocfs2_delete_inode+0x284/0x3f0 [ocfs2]
>>> [<ffffffffa0919a10>] ? ocfs2_dentry_attach_lock+0x5a0/0x5a0 [ocfs2]
>>> [<ffffffffa093872e>] ? ocfs2_evict_inode+0x1e/0x50 [ocfs2]
>>> [<ffffffff81145900>] ? evict+0x70/0x140
>>> [<ffffffffa0919322>] ? __ocfs2_drop_dl_inodes.clone.2+0x32/0x60 [ocfs2]
>>> [<ffffffffa0919a39>] ? ocfs2_drop_dl_inodes+0x29/0x90 [ocfs2]
>>> [<ffffffff8106e56f>] ? process_one_work+0x11f/0x440
>>> [<ffffffff8106f279>] ? worker_thread+0x159/0x330
>>> [<ffffffff8106f120>] ? manage_workers.clone.21+0x120/0x120
>>> [<ffffffff8106f120>] ? manage_workers.clone.21+0x120/0x120
>>> [<ffffffff81073fa6>] ? kthread+0x96/0xa0
>>> [<ffffffff8148bb24>] ? kernel_thread_helper+0x4/0x10
>>> [<ffffffff81073f10>] ? kthread_worker_fn+0x1a0/0x1a0
>>> [<ffffffff8148bb20>] ? gs_change+0x13/0x13
>>> INFO: task rm:5192 blocked for more than 120 seconds.
>>> "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this 
>>> message.
>>> rm              D ffff88107f2725c0     0  5192  16338 0x00000000
>>> ffff881014ccb040 0000000000000082 ffff8810206b8040 00000000000125c0
>>> ffff8804d7697fd8 00000000000125c0 00000000000125c0 00000000000125c0
>>> ffff8804d7696000 00000000000125c0 ffff8804d7697fd8 00000000000125c0
>>> Call Trace:
>>> [<ffffffff8148148d>] ? schedule_timeout+0x1ed/0x2e0
>>> [<ffffffffa0886162>] ? dlmconvert_master+0xe2/0x190 [ocfs2_dlm]
>>> [<ffffffffa08878bf>] ? dlmlock+0x7f/0xb70 [ocfs2_dlm]
>>> [<ffffffff81480e0a>] ? wait_for_common+0x13a/0x190
>>> [<ffffffff8104bc50>] ? try_to_wake_up+0x280/0x280
>>> [<ffffffffa0928a38>] ? __ocfs2_cluster_lock.clone.21+0x1d8/0x6b0 [ocfs2]
>>> [<ffffffffa0928fcc>] ? ocfs2_inode_lock_full_nested+0xbc/0x490 [ocfs2]
>>> [<ffffffffa0943c1b>] ? ocfs2_lookup_lock_orphan_dir+0x6b/0x1b0 [ocfs2]
>>> [<ffffffffa09454ba>] ? ocfs2_prepare_orphan_dir+0x4a/0x280 [ocfs2]
>>> [<ffffffffa094616f>] ? ocfs2_unlink+0x6ef/0xb90 [ocfs2]
>>> [<ffffffff811b35a9>] ? may_link.clone.22+0xd9/0x170
>>> [<ffffffff8113aa58>] ? vfs_unlink+0x98/0x100
>>> [<ffffffff8113ac41>] ? do_unlinkat+0x181/0x1b0
>>> [<ffffffff8113e7cd>] ? vfs_readdir+0x9d/0xe0
>>> [<ffffffff811653d8>] ? fsnotify_find_inode_mark+0x28/0x40
>>> [<ffffffff81166324>] ? dnotify_flush+0x54/0x110
>>> [<ffffffff8112b07f>] ? filp_close+0x5f/0x90
>>> [<ffffffff8148aa12>] ? system_call_fastpath+0x16/0x1b
>>>
>>>
>>> _______________________________________________
>>> Ocfs2-users mailing list
>>> Ocfs2-users at oss.oracle.com
>>> http://oss.oracle.com/mailman/listinfo/ocfs2-users
>>
>>
>>
> 




More information about the Ocfs2-users mailing list