[Ocfs2-users] No space left on device
Alexander Barton
alex at barton.de
Wed Sep 29 02:40:22 PDT 2010
Am 29.09.2010 um 11:25 schrieb Tao Ma:
> On 09/29/2010 05:13 PM, Alexander Barton wrote:
>
>> cl1-n2 kernel: [ 4006.829327] general protection fault: 0000 [#21] SMP
>> cl1-n2 kernel: [ 4006.829487] last sysfs file: /sys/devices/platform/coretemp.7/temp1_label
>> cl1-n2 kernel: [ 4006.829558] CPU 1
>> cl1-n2 kernel: [ 4006.829611] Modules linked in: ocfs2 jbd2 quota_tree tun xt_tcpudp iptable_filter hmac sha1_generic ocfs2_dlmfs ocfs2_stack_o2cb ocfs2_dlm ocfs2_nodemanager ocfs2_stackglue iptable_nat nf_nat configfs nf_conntrack_ipv4 nf_conntrack nf_defrag_ipv4 ip_tables x_tables nfsd exportfs nfs lockd fscache nfs_acl auth_rpcgss sunrpc ext2 coretemp drbd lru_cache cn loop hed tpm_tis snd_pcm snd_timer snd soundcore psmouse snd_page_alloc processor tpm pcspkr evdev joydev tpm_bios dcdbas serio_raw i5k_amb button rng_core shpchp pci_hotplug i5000_edac edac_core ext3 jbd mbcache dm_mirror dm_region_hash dm_log dm_snapshot dm_mod usbhid hid sg sr_mod cdrom ata_generic sd_mod ses crc_t10dif enclosure ata_piix ehci_hcd uhci_hcd usbcore bnx2 libata nls_base megaraid_sas scsi_mod e1000e thermal fan thermal_sys [last unloaded: scsi_wait_scan]
>> cl1-n2 kernel: [ 4006.833215]
>> cl1-n2 kernel: [ 4006.833215] Pid: 7699, comm: apache2 Tainted: G D 2.6.35-trunk-amd64 #1 0H603H/PowerEdge 2950
>> cl1-n2 kernel: [ 4006.833215] RIP: 0010:[<ffffffff810e1886>] [<ffffffff810e1886>] __kmalloc+0xd3/0x136
>> cl1-n2 kernel: [ 4006.833215] RSP: 0018:ffff88012e277cd8 EFLAGS: 00010006
>> cl1-n2 kernel: [ 4006.833215] RAX: 0000000000000000 RBX: 6f635f6465727265 RCX: ffffffffa0686032
>> cl1-n2 kernel: [ 4006.833215] RDX: 0000000000000000 RSI: ffff88012e277da8 RDI: 0000000000000004
>> cl1-n2 kernel: [ 4006.833215] RBP: ffffffff81625520 R08: ffff880001a52510 R09: 0000000000000003
>> cl1-n2 kernel: [ 4006.833215] R10: ffff88009a561b40 R11: ffff88022d62f400 R12: 000000000000000b
>> cl1-n2 kernel: [ 4006.833215] R13: 0000000000008050 R14: 0000000000008050 R15: 0000000000000246
>> cl1-n2 kernel: [ 4006.833215] FS: 00007f9199715740(0000) GS:ffff880001a40000(0000) knlGS:0000000000000000
>> cl1-n2 kernel: [ 4006.833215] CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b
>> cl1-n2 kernel: [ 4006.833215] CR2: 00000000402de9d0 CR3: 00000001372a1000 CR4: 00000000000406e0
>> cl1-n2 kernel: [ 4006.833215] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
>> cl1-n2 kernel: [ 4006.833215] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
>> cl1-n2 kernel: [ 4006.833215] Process apache2 (pid: 7699, threadinfo ffff88012e276000, task ffff88009a561b40)
>> cl1-n2 kernel: [ 4006.833215] Stack:
>> cl1-n2 kernel: [ 4006.833215] ffff8801b1af9c20 ffffffffa0686032 ffff8801b1a2da20 ffff88018f5f30c0
>> cl1-n2 kernel: [ 4006.833215]<0> ffff88012e277e88 000000000000000a ffff88018d105300 ffff88009a561b40
>> cl1-n2 kernel: [ 4006.833215]<0> ffff88012e277da8 ffffffffa0686032 ffff88012e277e88 ffff88012e277da8
>> cl1-n2 kernel: [ 4006.833215] Call Trace:
>> cl1-n2 kernel: [ 4006.833215] [<ffffffffa0686032>] ? ocfs2_fast_follow_link+0x166/0x284 [ocfs2]
>> cl1-n2 kernel: [ 4006.833215] [<ffffffff810f29fa>] ? do_follow_link+0xdb/0x24c
>> cl1-n2 kernel: [ 4006.833215] [<ffffffff810f2d55>] ? link_path_walk+0x1ea/0x482
>> cl1-n2 kernel: [ 4006.833215] [<ffffffff810f311f>] ? path_walk+0x63/0xd6
>> cl1-n2 kernel: [ 4006.833215] [<ffffffff810f27ba>] ? path_init+0x46/0x1ab
>> cl1-n2 kernel: [ 4006.833215] [<ffffffff810f3288>] ? do_path_lookup+0x20/0x85
>> cl1-n2 kernel: [ 4006.833215] [<ffffffff810f3cd9>] ? user_path_at+0x46/0x78
>> cl1-n2 kernel: [ 4006.833215] [<ffffffff81038bac>] ? pick_next_task_fair+0xe6/0xf6
>> cl1-n2 kernel: [ 4006.833215] [<ffffffff81305101>] ? schedule+0x4d4/0x530
>> cl1-n2 kernel: [ 4006.833215] [<ffffffff81060526>] ? prepare_creds+0x87/0x9c
>> cl1-n2 kernel: [ 4006.833215] [<ffffffff810e8649>] ? sys_faccessat+0x96/0x15b
>> cl1-n2 kernel: [ 4006.833215] [<ffffffff810089c2>] ? system_call_fastpath+0x16/0x1b
>> cl1-n2 kernel: [ 4006.833215] Code: 0f 1f 44 00 00 49 89 c7 fa 66 0f 1f 44 00 00 65 4c 8b 04 25 b0 ea 00 00 48 8b 45 00 49 01 c0 49 8b 18 48 85 db 74 0d 48 63 45 18<48> 8b 04 03 49 89 00 eb 11 83 ca ff 44 89 f6 48 89 ef e8 a1 f1
>> cl1-n2 kernel: [ 4006.833215] RIP [<ffffffff810e1886>] __kmalloc+0xd3/0x136
>> cl1-n2 kernel: [ 4006.833215] RSP<ffff88012e277cd8>
>> cl1-n2 kernel: [ 4006.833215] ---[ end trace b1eead7c8752b710 ]—
>
> Interesting, this seems to unrelated to discontig bg. Please file a bug for it.
Done. Bug #1292: <http://oss.oracle.com/bugzilla/show_bug.cgi?id=1292>
>> This fault then repeats over and over again …
>> Sometime the node seems to lock up completely; then Node 1 logs:
>>
>> cl1-n1 kernel: [ 7314.523473] block drbd0: peer( Primary -> Unknown ) conn( Connected -> NetworkFailure ) pdsk( UpToDate -> DUnknown )
>> cl1-n1 kernel: [ 7314.523485] block drbd0: asender terminated
>> cl1-n1 kernel: [ 7314.523488] block drbd0: Terminating drbd0_asender
>> cl1-n1 kernel: [ 7314.523495] block drbd0: short read receiving data: read 2960 expected 4096
>> cl1-n1 kernel: [ 7314.523502] block drbd0: Creating new current UUID
>> cl1-n1 kernel: [ 7314.523509] block drbd0: error receiving Data, l: 4120!
>> cl1-n1 kernel: [ 7314.523705] block drbd0: Connection closed
>> cl1-n1 kernel: [ 7314.523710] block drbd0: conn( NetworkFailure -> Unconnected )
>> cl1-n1 kernel: [ 7314.523718] block drbd0: receiver terminated
>> cl1-n1 kernel: [ 7314.523721] block drbd0: Restarting drbd0_receiver
>> cl1-n1 kernel: [ 7314.523723] block drbd0: receiver (re)started
>> cl1-n1 kernel: [ 7314.523730] block drbd0: conn( Unconnected -> WFConnection )
>>
>> … which looks „fine“ because of the 1st node not responding any more.
>> But then we get:
>>
>> cl1-n1 kernel: [ 7319.136065] o2net: connection to node cl1-n2 (num 1) at 10.0.1.2:6999 has been idle for 10.0 seconds, shutting it down.
>> cl1-n1 kernel: [ 7319.136079] (swapper,0,0):o2net_idle_timer1O:ee n ee n ee n ee n ee n ee n ee n ee ee ee ee ee ee ee ee ee ee ee ee ee sending message 506 (key 0xf5dfae8c) to node 1
>>
>> And endless messages like the following …
>>
>> cl1-n1 kernel: [ 7319.196135] (TunCtl,3278,6):dlm_send_remote_unlock_request:358 ERROR: Error -107 when sending message 506 (key 0xf5dfae8c) to node 1
>> cl1-n1 kernel: [ 7319.196192] (IPaddr2,3416,4):dlm_send_remote_unlock_request:358 ERROR: Error -107 when sending message 506 (key 0xf5dfae8c) to node 1
>> cl1-n1 kernel: [ 7319.196256] (IPaddr2,3282,1):dlm_send_remote_unlock_request:358 ERROR: Error -107 when sending message 506 (key 0xf5dfae8c) to node 1
>> cl1-n1 kernel: [ 7319.196373] (linux,3271,3):dlm_send_remote_unlock_request:358 ERROR: Error -107 when sending message 506 (key 0xf5dfae8c) to node 1
>> cl1-n1 kernel: [ 7319.196622] (apache2,5474,0):dlm_send_remote_unlock_request:358 ERROR: Error -107 when sending message 506 (key 0xf5dfae8c) to node 1
>> …
>> …
>>
>> … until node 1 hangs forever, too.
>>
>> After rebooting both of the nodes, the cluster runs for some time … and errors out like above again. And again. And … :-/
>>
>> Two other error messages observed while trying to mount the OCFS2 filesystem on one node (but we can’t reproduce any more):
>>
>> cl1-n2 kernel: [ 361.012172] INFO: task mount.ocfs2:4969 blocked for more than 120 seconds.
>> cl1-n2 kernel: [ 361.012252] "echo 0> /proc/sys/kernel/hung_task_timeout_secs" disables this message.
>> cl1-n2 kernel: [ 361.012343] mount.ocfs2 D 00000000ffff3314 0 4969 1 0x00000000
>> cl1-n2 kernel: [ 361.012347] ffff88022c8c1b40 0000000000000082 ffff880200000000 ffff88022c133e90
>> cl1-n2 kernel: [ 361.012351] ffff88022ec751c0 00000000000154c0 00000000000154c0 00000000000154c0
>> cl1-n2 kernel: [ 361.012354] ffff8802184b3fd8 00000000000154c0 ffff88022c8c1b40 ffff8802184b3fd8
>> cl1-n2 kernel: [ 361.012356] Call Trace:
>> cl1-n2 kernel: [ 361.012365] [<ffffffff813054df>] ? schedule_timeout+0x2d/0xd7
>> cl1-n2 kernel: [ 361.012369] [<ffffffff81305354>] ? wait_for_common+0xd1/0x14e
>> cl1-n2 kernel: [ 361.012374] [<ffffffff8103f630>] ? default_wake_function+0x0/0xf
>> cl1-n2 kernel: [ 361.012394] [<ffffffffa05bc21f>] ? __ocfs2_cluster_lock+0x6e0/0x890 [ocfs2]
>> cl1-n2 kernel: [ 361.012403] [<ffffffffa0551bb4>] ? dlm_register_domain+0x9e4/0xaf0 [ocfs2_dlm]
>> cl1-n2 kernel: [ 361.012408] [<ffffffff81190a3e>] ? hweight_long+0x5/0x6
>> cl1-n2 kernel: [ 361.012420] [<ffffffffa05bd242>] ? T.775+0x18/0x1d [ocfs2]
>> cl1-n2 kernel: [ 361.012432] [<ffffffffa05bd2f3>] ? ocfs2_super_lock+0xac/0x2bd [ocfs2]
>> cl1-n2 kernel: [ 361.012443] [<ffffffffa05b4d3b>] ? ocfs2_is_hard_readonly+0x10/0x23 [ocfs2]
>> cl1-n2 kernel: [ 361.012455] [<ffffffffa05bd2f3>] ? ocfs2_super_lock+0xac/0x2bd [ocfs2]
>> cl1-n2 kernel: [ 361.012470] [<ffffffffa05f5d66>] ? ocfs2_fill_super+0x1227/0x2101 [ocfs2]
>> cl1-n2 kernel: [ 361.012475] [<ffffffff8118e2f3>] ? snprintf+0x36/0x3b
>> cl1-n2 kernel: [ 361.012478] [<ffffffff810eb86e>] ? get_sb_bdev+0x137/0x19a
>> cl1-n2 kernel: [ 361.012492] [<ffffffffa05f4b3f>] ? ocfs2_fill_super+0x0/0x2101 [ocfs2]
>> cl1-n2 kernel: [ 361.012495] [<ffffffff810eaf45>] ? vfs_kern_mount+0xa6/0x196
>> cl1-n2 kernel: [ 361.012498] [<ffffffff810eb094>] ? do_kern_mount+0x49/0xe7
>> cl1-n2 kernel: [ 361.012502] [<ffffffff810ff38b>] ? do_mount+0x75c/0x7d6
>> cl1-n2 kernel: [ 361.012506] [<ffffffff810d90ba>] ? alloc_pages_current+0x9f/0xc2
>> cl1-n2 kernel: [ 361.012508] [<ffffffff810ff48d>] ? sys_mount+0x88/0xc3
>> cl1-n2 kernel: [ 361.012513] [<ffffffff810089c2>] ? system_call_fastpath+0x16/0x1b
>>
>> cl1-n2 kernel: [ 361.012518] INFO: task mount.ocfs2:5481 blocked for more than 120 seconds.
>> cl1-n2 kernel: [ 361.012588] "echo 0> /proc/sys/kernel/hung_task_timeout_secs" disables this message.
>> cl1-n2 kernel: [ 361.012674] mount.ocfs2 D 00000000ffff63dd 0 5481 1 0x00000000
>> cl1-n2 kernel: [ 361.012677] ffff88022c7a2210 0000000000000082 ffff88022c5ae2c8 ffffffff00000000
>> cl1-n2 kernel: [ 361.012680] ffff88022ec76d00 00000000000154c0 00000000000154c0 00000000000154c0
>> cl1-n2 kernel: [ 361.012683] ffff880218485fd8 00000000000154c0 ffff88022c7a2210 ffff880218485fd8
>> cl1-n2 kernel: [ 361.012685] Call Trace:
>> cl1-n2 kernel: [ 361.012691] [<ffffffff813063b5>] ? rwsem_down_failed_common+0x97/0xcb
>> cl1-n2 kernel: [ 361.012694] [<ffffffff810ea58f>] ? test_bdev_super+0x0/0xd
>> cl1-n2 kernel: [ 361.012697] [<ffffffff81306405>] ? rwsem_down_write_failed+0x1c/0x25
>> cl1-n2 kernel: [ 361.012700] [<ffffffff8118f2c3>] ? call_rwsem_down_write_failed+0x13/0x20
>> cl1-n2 kernel: [ 361.012703] [<ffffffff81305d2e>] ? down_write+0x25/0x27
>> cl1-n2 kernel: [ 361.012705] [<ffffffff810eb361>] ? sget+0x99/0x34d
>> cl1-n2 kernel: [ 361.012708] [<ffffffff810ea565>] ? set_bdev_super+0x0/0x2a
>> cl1-n2 kernel: [ 361.012710] [<ffffffff810eb7d5>] ? get_sb_bdev+0x9e/0x19a
>> cl1-n2 kernel: [ 361.012724] [<ffffffffa05f4b3f>] ? ocfs2_fill_super+0x0/0x2101 [ocfs2]
>> cl1-n2 kernel: [ 361.012727] [<ffffffff810eaf45>] ? vfs_kern_mount+0xa6/0x196
>> cl1-n2 kernel: [ 361.012730] [<ffffffff810eb094>] ? do_kern_mount+0x49/0xe7
>> cl1-n2 kernel: [ 361.012733] [<ffffffff810ff38b>] ? do_mount+0x75c/0x7d6
>> cl1-n2 kernel: [ 361.012735] [<ffffffff810d90ba>] ? alloc_pages_current+0x9f/0xc2
>> cl1-n2 kernel: [ 361.012738] [<ffffffff810ff48d>] ? sys_mount+0x88/0xc3
>>
Do you have any comments to these errors and traces? Or anybody else?
>> Is it possible for us to downgrade the kernel and ocfs2-tools after „tunefs.ocfs2 --fs-features=discontig-bg“?
>
> just use tunefs.ocfs2 --fs-features=nodiscontig-bg to disable it and you can go back to the old kernel and use it.
Ok, will try this most probably to get a stable system again.
But then will will encounter the „no space left“ problem again. Hmpf :-(
Thanks for your support!
Alex
More information about the Ocfs2-users
mailing list