[Ocfs2-users] sanity check - Xen+iSCSI+LVM+OCFS2 at dom0/domU

Sunil Mushran Sunil.Mushran at oracle.com
Thu Feb 7 09:56:43 PST 2008


Are you building ocfs2 with this kernel or are using the ones we
provide for RHEL5?

I am assuming you have built it yourself as we did not release
packages for the latest 2.6.18-53.1.6 kernel till last night.

If you are using your own, then use the one from oss.

If you are using the one from oss, then file a bugzilla with the
full oops trace.

Thanks
Sunil

Alok K. Dhir wrote:
> Hello all - we're evaluating OCFS2 in our development environment to 
> see if it meets our needs.
>
> We're testing it with an iSCSI storage array (Dell MD3000i) and 5 
> servers running Centos 5.1 (2.6.18-53.1.6.el5xen).
>
> 1) Each of the 5 servers is running the Centos 5.1 open-iscsi 
> initiator, and sees the volumes exposed by the array just fine.  So 
> far so good.
>
> 2) Created a volume group using the exposed iscsi volumes and created 
> a few LVM2 logical volumes.
>
> 3) vgscan; vgchange -a y; on all the cluster members.  all see the 
> "md3000vg" volume group.  looking good. (we have no intention of 
> changing the LVM2 configurations much if at all, and can make sure all 
> such changes are done when the volumes are off-line on all cluster 
> members, so theoretically this should not be a problem).
>
> 4) mkfs.ocfs2 /dev/md3000vg/testvol0 -- works great
>
> 5) mount on all Xen dom0 boxes in the cluster, works great.
>
> 6) create a VM on one of the cluster members, set up iscsi, vgscan, 
> md3000vg shows up -- looking good.
>
> 7) install ocfs2, 'service o2cb enable', starts up fine.  mount 
> /dev/md3000vg/testvol0, works fine.
>
> ** Thanks for making it this far -- this is where is gets interesting
>
> 8) run 'iozone' in domU against ocfs2 share - BANG - immediate kernel 
> panic, repeatable all day long.
>
>     "kernel BUG at fs/inode.c"
>
> So my questions:
>
> 1) should this work?
>
> 2) if not, what should we do differently?
>
> 3) currently we're tracking the latest RHEL/Centos 5.1 kernels -- 
> would we have better luck using the latest mainline kernel?
>
> Thanks for any assistance.
>
> Alok Dhir
>
>
> _______________________________________________
> Ocfs2-users mailing list
> Ocfs2-users at oss.oracle.com
> http://oss.oracle.com/mailman/listinfo/ocfs2-users




More information about the Ocfs2-users mailing list