[Ocfs2-users] sanity check - Xen+iSCSI+LVM+OCFS2 at dom0/domU

Sunil Mushran Sunil.Mushran at oracle.com
Thu Feb 7 10:15:44 PST 2008


http://oss.oracle.com/projects/ocfs2/dist/documentation/ocfs2-new-features.html

Alok Dhir wrote:
> We were indeed using a self-built module due to the lack of an OSS one 
> for the latest kernel.  Thanks for your response, I will test with the 
> new version.
>
> What are we leaving on the table by not using the latest mainline kernel?
>
> On Feb 7, 2008, at 12:56 PM, Sunil Mushran wrote:
>
>> Are you building ocfs2 with this kernel or are using the ones we
>> provide for RHEL5?
>>
>> I am assuming you have built it yourself as we did not release
>> packages for the latest 2.6.18-53.1.6 kernel till last night.
>>
>> If you are using your own, then use the one from oss.
>>
>> If you are using the one from oss, then file a bugzilla with the
>> full oops trace.
>>
>> Thanks
>> Sunil
>>
>> Alok K. Dhir wrote:
>>> Hello all - we're evaluating OCFS2 in our development environment to 
>>> see if it meets our needs.
>>>
>>> We're testing it with an iSCSI storage array (Dell MD3000i) and 5 
>>> servers running Centos 5.1 (2.6.18-53.1.6.el5xen).
>>>
>>> 1) Each of the 5 servers is running the Centos 5.1 open-iscsi 
>>> initiator, and sees the volumes exposed by the array just fine.  So 
>>> far so good.
>>>
>>> 2) Created a volume group using the exposed iscsi volumes and 
>>> created a few LVM2 logical volumes.
>>>
>>> 3) vgscan; vgchange -a y; on all the cluster members.  all see the 
>>> "md3000vg" volume group.  looking good. (we have no intention of 
>>> changing the LVM2 configurations much if at all, and can make sure 
>>> all such changes are done when the volumes are off-line on all 
>>> cluster members, so theoretically this should not be a problem).
>>>
>>> 4) mkfs.ocfs2 /dev/md3000vg/testvol0 -- works great
>>>
>>> 5) mount on all Xen dom0 boxes in the cluster, works great.
>>>
>>> 6) create a VM on one of the cluster members, set up iscsi, vgscan, 
>>> md3000vg shows up -- looking good.
>>>
>>> 7) install ocfs2, 'service o2cb enable', starts up fine.  mount 
>>> /dev/md3000vg/testvol0, works fine.
>>>
>>> ** Thanks for making it this far -- this is where is gets interesting
>>>
>>> 8) run 'iozone' in domU against ocfs2 share - BANG - immediate 
>>> kernel panic, repeatable all day long.
>>>
>>>    "kernel BUG at fs/inode.c"
>>>
>>> So my questions:
>>>
>>> 1) should this work?
>>>
>>> 2) if not, what should we do differently?
>>>
>>> 3) currently we're tracking the latest RHEL/Centos 5.1 kernels -- 
>>> would we have better luck using the latest mainline kernel?
>>>
>>> Thanks for any assistance.
>>>
>>> Alok Dhir
>>>
>>>
>>> _______________________________________________
>>> Ocfs2-users mailing list
>>> Ocfs2-users at oss.oracle.com
>>> http://oss.oracle.com/mailman/listinfo/ocfs2-users
>>
>
>
> _______________________________________________
> Ocfs2-users mailing list
> Ocfs2-users at oss.oracle.com
> http://oss.oracle.com/mailman/listinfo/ocfs2-users




More information about the Ocfs2-users mailing list