[Ocfs2-users] OCFS2 Node silent death.

Eric Ren zren at suse.com
Thu Dec 22 17:59:39 PST 2016


Hi,

Looks like this patch is for this issue:

commit 3d46a44a0c01b15d385ccaae24b56f619613c256
Author: Tariq Saeed <tariq.x.saeed at oracle.com>
Date:   Fri Sep 4 15:44:31 2015 -0700

     ocfs2: fix BUG_ON() in ocfs2_ci_checkpointed()

     PID: 614    TASK: ffff882a739da580  CPU: 3   COMMAND: "ocfs2dc"
       #0 [ffff882ecc3759b0] machine_kexec at ffffffff8103b35d
       #1 [ffff882ecc375a20] crash_kexec at ffffffff810b95b5
       #2 [ffff882ecc375af0] oops_end at ffffffff815091d8
       #3 [ffff882ecc375b20] die at ffffffff8101868b
       #4 [ffff882ecc375b50] do_trap at ffffffff81508bb0
       #5 [ffff882ecc375ba0] do_invalid_op at ffffffff810165e5
       #6 [ffff882ecc375c40] invalid_op at ffffffff815116fb
          [exception RIP: ocfs2_ci_checkpointed+208]
          RIP: ffffffffa0a7e940  RSP: ffff882ecc375cf0  RFLAGS: 00010002
          RAX: 0000000000000001  RBX: 000000000000654b  RCX: ffff8812dc83f1f8
          RDX: 00000000000017d9  RSI: ffff8812dc83f1f8  RDI: ffffffffa0b2c318
          RBP: ffff882ecc375d20   R8: ffff882ef6ecfa60   R9: ffff88301f272200
          R10: 0000000000000000  R11: 0000000000000000  R12: ffffffffffffffff
          R13: ffff8812dc83f4f0  R14: 0000000000000000  R15: ffff8812dc83f1f8
          ORIG_RAX: ffffffffffffffff  CS: 0010  SS: 0018
       #7 [ffff882ecc375d28] ocfs2_check_meta_downconvert at ffffffffa0a7edbd [ocfs2]
       #8 [ffff882ecc375d38] ocfs2_unblock_lock at ffffffffa0a84af8 [ocfs2]
       #9 [ffff882ecc375dc8] ocfs2_process_blocked_lock at ffffffffa0a85285 [ocfs2]
     #10 [ffff882ecc375e18] ocfs2_downconvert_thread_do_work at ffffffffa0a85445 [ocfs2]
     #11 [ffff882ecc375e68] ocfs2_downconvert_thread at ffffffffa0a854de [ocfs2]
     #12 [ffff882ecc375ee8] kthread at ffffffff81090da7
     #13 [ffff882ecc375f48] kernel_thread_helper at ffffffff81511884
     assert is tripped because the tran is not checkpointed and the lock level is PR.

     Some time ago, chmod command had been executed. As result, the following call
     chain left the inode cluster lock in PR state, latter on causing the assert.
     system_call_fastpath
       -> my_chmod
        -> sys_chmod
         -> sys_fchmodat
          -> notify_change
           -> ocfs2_setattr
            -> posix_acl_chmod
             -> ocfs2_iop_set_acl
              -> ocfs2_set_acl
               -> ocfs2_acl_set_mode
     Here is how.
     1119 int ocfs2_setattr(struct dentry *dentry, struct iattr *attr)
     1120 {
     1247         ocfs2_inode_unlock(inode, 1); <<< WRONG thing to do.
     ..
     1258         if (!status && attr->ia_valid & ATTR_MODE) {
     1259                 status =  posix_acl_chmod(inode, inode->i_mode);

     519 posix_acl_chmod(struct inode *inode, umode_t mode)
     520 {
     ..
     539         ret = inode->i_op->set_acl(inode, acl, ACL_TYPE_ACCESS);

     287 int ocfs2_iop_set_acl(struct inode *inode, struct posix_acl *acl, ...
     288 {
     289         return ocfs2_set_acl(NULL, inode, NULL, type, acl, NULL, NULL);

     224 int ocfs2_set_acl(handle_t *handle,
     225                          struct inode *inode, ...
     231 {
     ..
     252                                 ret = ocfs2_acl_set_mode(inode, di_bh,
     253 handle, mode);

     168 static int ocfs2_acl_set_mode(struct inode *inode, struct buffer_head ...
     170 {
     183         if (handle == NULL) {
                         >>> BUG: inode lock not held in ex at this point <<<
     184                 handle = ocfs2_start_trans(OCFS2_SB(inode->i_sb),
     185 OCFS2_INODE_UPDATE_CREDITS);

     ocfs2_setattr.#1247 we unlock and at #1259 call posix_acl_chmod. When we reach
     ocfs2_acl_set_mode.#181 and do trans, the inode cluster lock is not held in EX
     mode (it should be). How this could have happended?

     We are the lock master, were holding lock EX and have released it in
     ocfs2_setattr.#1247.  Note that there are no holders of this lock at
     this point.  Another node needs the lock in PR, and we downconvert from
     EX to PR.  So the inode lock is PR when do the trans in
     ocfs2_acl_set_mode.#184.  The trans stays in core (not flushed to disc).
     Now another node want the lock in EX, downconvert thread gets kicked
     (the one that tripped assert abovt), finds an unflushed trans but the
     lock is not EX (it is PR).  If the lock was at EX, it would have flushed
     the trans ocfs2_ci_checkpointed -> ocfs2_start_checkpoint before
     downconverting (to NULL) for the request.

     ocfs2_setattr must not drop inode lock ex in this code path. If it
     does, takes it again before the trans, say in ocfs2_set_acl, another
     cluster node can get in between, execute another setattr, overwriting
     the one in progress on this node, resulting in a mode acl size combo
     that is a mix of the two.

     Orabug: 20189959
     Signed-off-by: Tariq Saeed <tariq.x.saeed at oracle.com>
     Reviewed-by: Mark Fasheh <mfasheh at suse.de>
     Cc: Joel Becker <jlbec at evilplan.org>
     Cc: Joseph Qi <joseph.qi at huawei.com>
     Signed-off-by: Andrew Morton <akpm at linux-foundation.org>

Thanks,
Eric

On 12/19/2016 07:55 PM, Joseph Cluney wrote:
> Hi,
>
> I have a 3 node cluster running Oracle VM server release 3.4.1.
> We are having an occasional instances of a node disappearing, the only
> indication in the logs is the eviction and recovery messages.
>
> typical dmesg output.
>
> [3312552.140992] o2net: Connection to node cs-apps-node3 (num 2) at
> 192.168.200.3:7777 has been idle for 60.114 secs.
> [3312612.171263] o2net: Connection to node cs-apps-node3 (num 2) at
> 192.168.200.3:7777 has been idle for 60.32 secs.
> [3312613.015723] o2net: No longer connected to node cs-apps-node3 (num 2)
> at 192.168.200.3:7777
> [3312613.015783] (dlm_thread,5404,8):dlm_send_proxy_ast_msg:486 ERROR:
> 0004FB0000050000BC74A5DD09C33779: res M0000000000000000000207c84b1f72,
> error -112 send AST to node 2
> [3312613.015877] o2cb: o2dlm has evicted node 2 from domain
> 0004FB00000500007A4523B76F1051D5
> [3312613.016354] o2cb: o2dlm has evicted node 2 from domain
> 0004FB0000050000BC74A5DD09C33779
> [3312613.016652] (dlm_thread,5404,8):dlm_flush_asts:609 ERROR: status = -112
> [3312613.018071] o2cb: o2dlm has evicted node 2 from domain ovm
> [3312613.106251] o2dlm: Begin recovery on domain ovm for node 2
> [3312613.107043] o2dlm: Node 0 (me) is the Recovery Master for the dead
> node 2 in domain ovm
> [3312614.039237] o2dlm: Waiting on the recovery of node 2 in domain
> 0004FB0000050000BC74A5DD09C33779
> [3312614.050222] o2dlm: Waiting on the recovery of node 2 in domain
> 0004FB00000500007A4523B76F1051D5
> [3312614.982240] o2cb: o2dlm has evicted node 2 from domain
> 0004FB00000500007A4523B76F1051D5
> [3312615.913169] o2dlm: Begin recovery on domain
> 0004FB0000050000BC74A5DD09C33779 for node 2
> [3312615.913901] o2dlm: Node 0 (me) is the Recovery Master for the dead
> node 2 in domain 0004FB0000050000BC74A5DD09C33779
> [3312615.976172] o2dlm: Begin recovery on domain
> 0004FB00000500007A4523B76F1051D5 for node 2
> [3312615.976179] o2dlm: Node 1 (he) is the Recovery Master for the dead
> node 2 in domain 0004FB00000500007A4523B76F1051D5
> [3312615.976182] o2dlm: End recovery on domain
> 0004FB00000500007A4523B76F1051D5
> [3312618.106659] o2dlm: End recovery on domain ovm
> [3312620.914435] o2dlm: End recovery on domain
> 0004FB0000050000BC74A5DD09C33779
> [3312620.983993] ocfs2: Begin replay journal (node 2, slot 1) on device
> (7,0)
> [3312621.077920] ocfs2: End replay journal (node 2, slot 1) on device (7,0)
> [3312621.113673] ocfs2: Beginning quota recovery on device (7,0) for slot 1
> [3312625.040978] ocfs2: Begin replay journal (node 2, slot 0) on device
> (251,18)
> [3312627.147646] ocfs2: End replay journal (node 2, slot 0) on device
> (251,18)
> [3312627.187473] ocfs2: Beginning quota recovery on device (251,18) for
> slot 0
> [3312643.137839] ocfs2: Finishing quota recovery on device (251,18) for
> slot 0
> [3312643.139285] ocfs2: Finishing quota recovery on device (7,0) for slot 1
> [3312891.479712] o2net: Accepted connection from node cs-apps-node3 (num 2)
> at 192.168.200.3:7777
> [3312899.822790] o2dlm: Node 2 joins domain
> 0004FB00000500007A4523B76F1051D5 ( 0 1 2 ) 3 nodes
> [3312900.046000] o2dlm: Node 2 joins domain
> 0004FB0000050000BC74A5DD09C33779 ( 0 1 2 ) 3 nodes
> [3312901.134648] o2dlm: Node 2 joins domain ovm ( 0 1 2 ) 3 nodes
>
> Have done a net console to capture the output from the failing node.
>
> The node that fails in the cluster is performing a reflink operation on VM
> files for snapshots, the crash happens during this snapshot window.
> To note this does not happen every time the snapshots are taken.
>
> Two nodes are 10GB capable and have multipath links to the storage device,
> when these links are activated the crashes are more frequent.
>
> Only the node that is doing the snapshot (reflink) is affected.
>
> netconsole output.
> [1382139.668117] ------------[ cut here ]------------
> [1382139.668172] kernel BUG at fs/ocfs2/dlmglue.c:3647!
> [1382139.668687] invalid opcode: 0000 [#1] SMP
> [1382139.669185] Modules linked in: netconsole tun ocfs2 jbd2 nfsv3 nfs_acl
> rpcsec_gss_krb5 auth_rpcgss nfsv4 nfs fscache lockd sunrpc grace
> xen_pciback xen_netback xen_blkback xen_gntalloc xen_gntdev xen_evtchn
> xenfs xen_privcmd ocfs2_d
> lmfs ocfs2_stack_o2cb ocfs2_dlm ocfs2_nodemanager ocfs2_stackglue configfs
> bnx2fc fcoe libfcoe libfc scsi_transport_fc 8021q mrp garp dm_round_robin
> bridge stp llc bonding ib_iser rdma_cm ib_cm iw_cm ib_sa ib_mad ib_core
> ib_addr iscsi_tc
> p dm_multipath ipmi_devintf ipmi_si ipmi_msghandler iTCO_wdt
> iTCO_vendor_support dcdbas pcspkr serio_raw lpc_ich mfd_core ioatdma dca
> i7core_edac edac_core sg ext3 jbd mbcache sr_mod cdrom sd_mod wmi pata_acpi
> ata_generic ata_piix megara
> id_sas crc32c_intel bnx2 be2iscsi bnx2i cnic uio cxgb4i cxgb4 cxgb3i
> libcxgbi ipv6 cxgb3 mdio libiscsi_tcp qla4xxx iscsi_boot_sysfs libiscsi
> scsi_transport_iscsi usb_storage mgag200 ttm drm_kms_helper drm
> i2c_algo_bit sysimgblt sysfillre
> ct i2c_core syscopyarea dm_mirror dm_region_hash dm_log dm_mod [last
> unloaded: netconsole]
> [1382139.675600] CPU: 15 PID: 4240 Comm: ocfs2dc Tainted: G        W
> 4.1.12-32.1.3.el6uek.x86_64 #2
> [1382139.676458] Hardware name: Dell Inc. PowerEdge R710/00W9X3, BIOS 6.4.0
> 07/23/2013
> [1382139.677348] task: ffff8800b6648e00 ti: ffff8800b6640000 task.ti:
> ffff8800b6640000
> [1382139.678224] RIP: e030:[<ffffffffa0abf42f>]  [<ffffffffa0abf42f>]
> ocfs2_ci_checkpointed+0xbf/0xd0 [ocfs2]
> [1382139.679202] RSP: e02b:ffff8800b6643ca8  EFLAGS: 00010093
> [1382139.680185] RAX: 0000000000008652 RBX: ffff8800023156c8 RCX:
> 0000000000000005
> [1382139.681180] RDX: 0000000000008652 RSI: ffff8800023153b8 RDI:
> ffffffffa0b659d8
> [1382139.682203] RBP: ffff8800b6643cd8 R08: ffff8800023153c8 R09:
> 0000000000000000
> [1382139.683324] R10: 0000000000007ff0 R11: 0000000000000002 R12:
> ffff8800023153b8
> [1382139.684381] R13: 0000000000000000 R14: 0000000000001f2a R15:
> 0000000000001f2a
> [1382139.685455] FS:  00007fe6eb8667c0(0000) GS:ffff8801307c0000(0000)
> knlGS:0000000000000000
> [1382139.686585] CS:  e033 DS: 0000 ES: 0000 CR0: 000000008005003b
> [1382139.687701] CR2: 00007f510cad0000 CR3: 00000000bdaf0000 CR4:
> 0000000000002660
> [1382139.688977] Stack:
> [1382139.690147]  ffff8800b6643cd8 ffff8800023153b8 00000000000002d0
> ffff880002315424
> [1382139.691562]  0000000000000000 ffff8800b6643da8 ffff8800b6643ce8
> ffffffffa0abf489
> [1382139.692889]  ffff8800b6643d88 ffffffffa0ac2dca ffff8801307ca210
> ffff8801307caa10
> [1382139.694167] Call Trace:
> [1382139.695462]  [<ffffffffa0abf489>]
> ocfs2_check_meta_downconvert+0x29/0x40 [ocfs2]
> [1382139.696808]  [<ffffffffa0ac2dca>] ocfs2_unblock_lock+0xca/0x750 [ocfs2]
> [1382139.698166]  [<ffffffff81012982>] ? __switch_to+0x212/0x5b0
> [1382139.699548]  [<ffffffffa0ac35fa>]
> ocfs2_process_blocked_lock+0x1aa/0x270 [ocfs2]
> [1382139.701015]  [<ffffffffa0ac3b92>]
> ocfs2_downconvert_thread_do_work+0xb2/0xe0 [ocfs2]
> [1382139.702491]  [<ffffffffa0ac3c36>] ocfs2_downconvert_thread+0x76/0x180
> [ocfs2]
> [1382139.704059]  [<ffffffff810c3b10>] ? wait_woken+0x90/0x90
> [1382139.705537]  [<ffffffffa0ac3bc0>] ?
> ocfs2_downconvert_thread_do_work+0xe0/0xe0 [ocfs2]
> [1382139.707051]  [<ffffffffa0ac3bc0>] ?
> ocfs2_downconvert_thread_do_work+0xe0/0xe0 [ocfs2]
> [1382139.708558]  [<ffffffff810a012e>] kthread+0xce/0xf0
> [1382139.710075]  [<ffffffff810a0060>] ?
> kthread_freezable_should_stop+0x70/0x70
> [1382139.711622]  [<ffffffff816b7e22>] ret_from_fork+0x42/0x70
> [1382139.713182]  [<ffffffff810a0060>] ?
> kthread_freezable_should_stop+0x70/0x70
> [1382139.714769] Code: 48 89 df e8 34 a6 05 00 48 8b b8 40 03 00 00 31 c9
> ba 01 00 00 00 be 03 00 00 00 48 81 c7 68 01 00 00 e8 a5 49 60 e0 31 c0 eb
> ae <0f> 0b eb fe 66 66 66 66 2e 0f 1f 84 00 00 00 00 00 55 48 89 e5
> [1382139.718379] RIP  [<ffffffffa0abf42f>] ocfs2_ci_checkpointed+0xbf/0xd0
> [ocfs2]
> [1382139.720221]  RSP <ffff8800b6643ca8>
> [1382139.728533] ------------[ cut here ]------------
> [1382139.730240] kernel BUG at arch/x86/mm/pageattr.c:214!
> [1382139.731902] invalid opcode: 0000 [#2] SMP
> [1382139.733542] Modules linked in: netconsole tun ocfs2 jbd2 nfsv3 nfs_acl
> rpcsec_gss_krb5 auth_rpcgss nfsv4 nfs fscache lockd sunrpc grace
> xen_pciback xen_netback xen_blkback xen_gntalloc xen_gntdev xen_evtchn
> xenfs xen_privcmd ocfs2_d
> lmfs ocfs2_stack_o2cb ocfs2_dlm ocfs2_nodemanager ocfs2_stackglue configfs
> bnx2fc fcoe libfcoe libfc scsi_transport_fc 8021q mrp garp dm_round_robin
> bridge stp llc bonding ib_iser rdma_cm ib_cm iw_cm ib_sa ib_mad ib_core
> ib_addr iscsi_tc
> p dm_multipath ipmi_devintf ipmi_si ipmi_msghandler iTCO_wdt
> iTCO_vendor_support dcdbas pcspkr serio_raw lpc_ich mfd_core ioatdma dca
> i7core_edac edac_core sg ext3 jbd mbcache sr_mod cdrom sd_mod wmi pata_acpi
> ata_generic ata_piix megara
> id_sas crc32c_intel bnx2 be2iscsi bnx2i cnic uio cxgb4i cxgb4 cxgb3i
> libcxgbi ipv6 cxgb3 mdio libiscsi_tcp qla4xxx iscsi_boot_sysfs libiscsi
> scsi_transport_iscsi usb_storage mgag200 ttm drm_kms_helper drm
> i2c_algo_bit sysimgblt sysfillre
> ct i2c_core syscopyarea dm_mirror dm_region_hash dm_log dm_mod [last
> unloaded: netconsole]
> [1382139.749332] CPU: 15 PID: 4240 Comm: ocfs2dc Tainted: G        W
> 4.1.12-32.1.3.el6uek.x86_64 #2
> [1382139.751099] Hardware name: Dell Inc. PowerEdge R710/00W9X3, BIOS 6.4.0
> 07/23/2013
> [1382139.752908] task: ffff8800b6648e00 ti: ffff8800b6640000 task.ti:
> ffff8800b6640000
> [1382139.754653] RIP: e030:[<ffffffff8106cf02>]  [<ffffffff8106cf02>]
> change_page_attr_set_clr+0x4e2/0x520
> [1382139.756420] RSP: e02b:ffff8800b6642da8  EFLAGS: 00010046
> [1382139.758187] RAX: 201008001fc900f5 RBX: 0000000000000200 RCX:
> 0000000000000000
> [1382139.759913] RDX: 0000000000000000 RSI: 0000000080000000 RDI:
> 0000000080000000
> [1382139.761592] RBP: ffff8800b6642e58 R08: 801000006ff01067 R09:
> 000000000006ff01
> [1382139.763221] R10: 0000000000007ff0 R11: 0000000000000000 R12:
> 0000000000000005
> [1382139.764843] R13: 0000000000000001 R14: 0000000000000000 R15:
> 0000000000000000
> [1382139.766467] FS:  00007fe6eb8667c0(0000) GS:ffff8801307c0000(0000)
> knlGS:0000000000000000
> [1382139.768089] CS:  e033 DS: 0000 ES: 0000 CR0: 000000008005003b
> [1382139.769697] CR2: 00007f510cad0000 CR3: 0000000001a62000 CR4:
> 0000000000002660
> [1382139.771320] Stack:
> [1382139.772907]  ffff880000000000 0000000000000000 0000000000000010
> 0000000000000001
> [1382139.774646]  0000000000000000 0000000000000000 0000000000000000
> 0000000000000000
> [1382139.776372]  0000000000000010 0000000000000000 0000000500000001
> 000000000006ff01
> [1382139.777968] Call Trace:
> [1382139.779530]  [<ffffffff8106d21c>] _set_pages_array+0xec/0x140
> [1382139.781092]  [<ffffffff8106d283>] set_pages_array_wc+0x13/0x20
> [1382139.782669]  [<ffffffffa010f54e>] ttm_set_pages_caching+0x4e/0x80 [ttm]
> [1382139.784258]  [<ffffffffa010fb76>] ttm_alloc_new_pages+0xb6/0x180 [ttm]
> [1382139.785824]  [<ffffffffa010fcdc>]
> ttm_page_pool_fill_locked.clone.1+0x9c/0x140 [ttm]
> [1382139.787394]  [<ffffffffa010fdc4>]
> ttm_page_pool_get_pages.clone.2+0x44/0xf0 [ttm]
> [1382139.788970]  [<ffffffffa010ff13>] ttm_get_pages.clone.0+0xa3/0x200
> [ttm]
> [1382139.790551]  [<ffffffffa01100fc>] ttm_pool_populate+0x8c/0xf0 [ttm]
> [1382139.792114]  [<ffffffffa010c2b4>] ? ttm_mem_reg_ioremap+0x64/0x100
> [ttm]
> [1382139.793795]  [<ffffffffa012deee>] mgag200_ttm_tt_populate+0xe/0x10
> [mgag200]
> [1382139.795373]  [<ffffffffa010c8c2>] ttm_bo_move_memcpy+0x4f2/0x540 [ttm]
> [1382139.796935]  [<ffffffffa01081ac>] ? ttm_tt_init+0x8c/0xb0 [ttm]
> [1382139.798498]  [<ffffffff811c321e>] ? __vmalloc_node+0x3e/0x40
> [1382139.800052]  [<ffffffffa012de18>] mgag200_bo_move+0x18/0x20 [mgag200]
> [1382139.801607]  [<ffffffffa0108ce9>] ttm_bo_handle_move_mem+0x299/0x650
> [ttm]
> [1382139.803099]  [<ffffffff811c2600>] ? vunmap_page_range+0x100/0x180
> [1382139.804590]  [<ffffffffa010af6e>] ttm_bo_validate+0x1de/0x1f0 [ttm]
> [1382139.806058]  [<ffffffff8106a3aa>] ? iounmap+0x8a/0xd0
> [1382139.807541]  [<ffffffffa012dfe3>] mgag200_bo_push_sysram+0x83/0xd0
> [mgag200]
> [1382139.809092]  [<ffffffffa012af3f>]
> mga_crtc_do_set_base.clone.0+0x7f/0x1e0 [mgag200]
> [1382139.810654]  [<ffffffff816b72e0>] ?
> _raw_spin_unlock_irqrestore+0x20/0x50
> [1382139.812146]  [<ffffffffa012bbe1>] mga_crtc_mode_set+0xb41/0xe60
> [mgag200]
> [1382139.813585]  [<ffffffffa00da9a9>] drm_crtc_helper_set_mode+0x389/0x5d0
> [drm_kms_helper]
> [1382139.815096]  [<ffffffffa00dba59>]
> drm_crtc_helper_set_config+0x879/0xae0 [drm_kms_helper]
> [1382139.816468]  [<ffffffff816b72e0>] ?
> _raw_spin_unlock_irqrestore+0x20/0x50
> [1382139.817845]  [<ffffffffa0078576>]
> drm_mode_set_config_internal+0x66/0x110 [drm]
> [1382139.819177]  [<ffffffffa00e6a21>]
> drm_fb_helper_pan_display+0x111/0x160 [drm_kms_helper]
> [1382139.820487]  [<ffffffff8138abd6>] fb_pan_display+0xe6/0x140
> [1382139.821761]  [<ffffffff81384d4a>] bit_update_start+0x2a/0x60
> [1382139.822981]  [<ffffffff813810d6>] fbcon_switch+0x386/0x530
> [1382139.824198]  [<ffffffff8140f5a9>] redraw_screen+0x189/0x230
> [1382139.825370]  [<ffffffff81380c9a>] fbcon_blank+0x21a/0x2d0
> [1382139.826642]  [<ffffffff816b72e0>] ?
> _raw_spin_unlock_irqrestore+0x20/0x50
> [1382139.827813]  [<ffffffff810d736c>] ? vprintk_emit+0x2ac/0x520
> [1382139.828933]  [<ffffffff810e8091>] ? internal_add_timer+0x91/0xc0
> [1382139.830070]  [<ffffffff816b72e0>] ?
> _raw_spin_unlock_irqrestore+0x20/0x50
> [1382139.831192]  [<ffffffff810ea06e>] ? mod_timer+0xfe/0x1d0
> [1382139.832306]  [<ffffffff81410676>] do_unblank_screen+0xb6/0x1f0
> [1382139.833438]  [<ffffffff814107c0>] unblank_screen+0x10/0x20
> [1382139.834562]  [<ffffffff8131f7bd>] bust_spinlocks+0x1d/0x40
> [1382139.835731]  [<ffffffff81017643>] oops_end+0x43/0x120
> [1382139.836953]  [<ffffffff81017c4b>] die+0x5b/0x90
> [1382139.838020]  [<ffffffff810140f9>] do_trap+0x169/0x170
> [1382139.839090]  [<ffffffff810a0ef2>] ?
> __atomic_notifier_call_chain+0x12/0x20
> [1382139.840144]  [<ffffffff81014af9>] do_error_trap+0xd9/0x180
> [1382139.841221]  [<ffffffffa0abf42f>] ? ocfs2_ci_checkpointed+0xbf/0xd0
> [ocfs2]
> [1382139.842344]  [<ffffffff816b72e0>] ?
> _raw_spin_unlock_irqrestore+0x20/0x50
> [1382139.843520]  [<ffffffff810c3e23>] ? __wake_up+0x53/0x70
> [1382139.844581]  [<ffffffff81014cb0>] do_invalid_op+0x20/0x30
> [1382139.845662]  [<ffffffff816b93be>] invalid_op+0x1e/0x30
> [1382139.846734]  [<ffffffffa0abf42f>] ? ocfs2_ci_checkpointed+0xbf/0xd0
> [ocfs2]
> [1382139.847826]  [<ffffffffa0abf3cd>] ? ocfs2_ci_checkpointed+0x5d/0xd0
> [ocfs2]
> [1382139.848992]  [<ffffffffa0abf489>]
> ocfs2_check_meta_downconvert+0x29/0x40 [ocfs2]
> [1382139.850223]  [<ffffffffa0ac2dca>] ocfs2_unblock_lock+0xca/0x750 [ocfs2]
> [1382139.851302]  [<ffffffff81012982>] ? __switch_to+0x212/0x5b0
> [1382139.852384]  [<ffffffffa0ac35fa>]
> ocfs2_process_blocked_lock+0x1aa/0x270 [ocfs2]
> [1382139.853480]  [<ffffffffa0ac3b92>]
> ocfs2_downconvert_thread_do_work+0xb2/0xe0 [ocfs2]
> [1382139.854580]  [<ffffffffa0ac3c36>] ocfs2_downconvert_thread+0x76/0x180
> [ocfs2]
> [1382139.855649]  [<ffffffff810c3b10>] ? wait_woken+0x90/0x90
> [1382139.856730]  [<ffffffffa0ac3bc0>] ?
> ocfs2_downconvert_thread_do_work+0xe0/0xe0 [ocfs2]
> [1382139.857835]  [<ffffffffa0ac3bc0>] ?
> ocfs2_downconvert_thread_do_work+0xe0/0xe0 [ocfs2]
> [1382139.858899]  [<ffffffff810a012e>] kthread+0xce/0xf0
> [1382139.860106]  [<ffffffff810a0060>] ?
> kthread_freezable_should_stop+0x70/0x70
> [1382139.861175]  [<ffffffff816b7e22>] ret_from_fork+0x42/0x70
> [1382139.862236]  [<ffffffff810a0060>] ?
> kthread_freezable_should_stop+0x70/0x70
> [1382139.863303] Code: 58 ff ff ff e9 e2 fb ff ff be b8 00 00 00 48 c7 c7
> f8 56 95 81 e8 bf 35 01 00 e9 03 ff ff ff 0f 0b eb fe 0f 0b 0f 1f 40 00 eb
> fa <0f> 0b eb fe 48 89 ca 48 c1 e9 05 45 31 ed 48 c1 ea 03 48 89 d0
> [1382139.865840] RIP  [<ffffffff8106cf02>]
> change_page_attr_set_clr+0x4e2/0x520
> [1382139.866978]  RSP <ffff8800b6642da8>
> [1382139.868102] ---[ end trace c97984eb9f0c5481 ]---
> [1382139.877139] Kernel panic - not syncing: Fatal exception
> [1382139.878299] Kernel Offset: disabled
> [1382139.879386] drm_kms_helper: panic occurred, switching back to text
> console
>
>
> If anyone has experienced this issue and could give any pointers on what to
> try next it would be much appreciated.
>
> Kind Regards
>
> Joe Cluney
>
>
>
> _______________________________________________
> Ocfs2-users mailing list
> Ocfs2-users at oss.oracle.com
> https://oss.oracle.com/mailman/listinfo/ocfs2-users





More information about the Ocfs2-users mailing list