[Ocfs2-users] sanity check - Xen+iSCSI+LVM+OCFS2 at dom0/domU

Alok Dhir adhir at symplicity.com
Thu Feb 7 10:19:58 PST 2008


I've seen that -- I was under the impression that some of those were  
being backported into the release kernels.

Thanks,

Alok

On Feb 7, 2008, at 1:15 PM, Sunil Mushran wrote:

> http://oss.oracle.com/projects/ocfs2/dist/documentation/ocfs2-new-features.html
>
> Alok Dhir wrote:
>> We were indeed using a self-built module due to the lack of an OSS  
>> one for the latest kernel.  Thanks for your response, I will test  
>> with the new version.
>>
>> What are we leaving on the table by not using the latest mainline  
>> kernel?
>>
>> On Feb 7, 2008, at 12:56 PM, Sunil Mushran wrote:
>>
>>> Are you building ocfs2 with this kernel or are using the ones we
>>> provide for RHEL5?
>>>
>>> I am assuming you have built it yourself as we did not release
>>> packages for the latest 2.6.18-53.1.6 kernel till last night.
>>>
>>> If you are using your own, then use the one from oss.
>>>
>>> If you are using the one from oss, then file a bugzilla with the
>>> full oops trace.
>>>
>>> Thanks
>>> Sunil
>>>
>>> Alok K. Dhir wrote:
>>>> Hello all - we're evaluating OCFS2 in our development environment  
>>>> to see if it meets our needs.
>>>>
>>>> We're testing it with an iSCSI storage array (Dell MD3000i) and 5  
>>>> servers running Centos 5.1 (2.6.18-53.1.6.el5xen).
>>>>
>>>> 1) Each of the 5 servers is running the Centos 5.1 open-iscsi  
>>>> initiator, and sees the volumes exposed by the array just fine.   
>>>> So far so good.
>>>>
>>>> 2) Created a volume group using the exposed iscsi volumes and  
>>>> created a few LVM2 logical volumes.
>>>>
>>>> 3) vgscan; vgchange -a y; on all the cluster members.  all see  
>>>> the "md3000vg" volume group.  looking good. (we have no intention  
>>>> of changing the LVM2 configurations much if at all, and can make  
>>>> sure all such changes are done when the volumes are off-line on  
>>>> all cluster members, so theoretically this should not be a  
>>>> problem).
>>>>
>>>> 4) mkfs.ocfs2 /dev/md3000vg/testvol0 -- works great
>>>>
>>>> 5) mount on all Xen dom0 boxes in the cluster, works great.
>>>>
>>>> 6) create a VM on one of the cluster members, set up iscsi,  
>>>> vgscan, md3000vg shows up -- looking good.
>>>>
>>>> 7) install ocfs2, 'service o2cb enable', starts up fine.  mount / 
>>>> dev/md3000vg/testvol0, works fine.
>>>>
>>>> ** Thanks for making it this far -- this is where is gets  
>>>> interesting
>>>>
>>>> 8) run 'iozone' in domU against ocfs2 share - BANG - immediate  
>>>> kernel panic, repeatable all day long.
>>>>
>>>>   "kernel BUG at fs/inode.c"
>>>>
>>>> So my questions:
>>>>
>>>> 1) should this work?
>>>>
>>>> 2) if not, what should we do differently?
>>>>
>>>> 3) currently we're tracking the latest RHEL/Centos 5.1 kernels --  
>>>> would we have better luck using the latest mainline kernel?
>>>>
>>>> Thanks for any assistance.
>>>>
>>>> Alok Dhir
>>>>
>>>>
>>>> _______________________________________________
>>>> Ocfs2-users mailing list
>>>> Ocfs2-users at oss.oracle.com
>>>> http://oss.oracle.com/mailman/listinfo/ocfs2-users
>>>
>>
>>
>> _______________________________________________
>> Ocfs2-users mailing list
>> Ocfs2-users at oss.oracle.com
>> http://oss.oracle.com/mailman/listinfo/ocfs2-users
>




More information about the Ocfs2-users mailing list