[Ocfs2-users] RHEL 5.8, ocfs2 v1.4.4, stability issues

Nathan Patwardhan npatwardhan at llbean.com
Fri Apr 27 08:32:16 PDT 2012



From: Kushnir, Michael (NIH/NLM/LHC) [C] [mailto:michael.kushnir at nih.gov]
Sent: Wednesday, April 25, 2012 12:52 PM
To: Nathan Patwardhan; ocfs2-users at oss.oracle.com
Subject: RE: RHEL 5.8, ocfs2 v1.4.4, stability issues


Ø  Hi Nathan,


Ø  I've no particular insight on splunk...


Ø  We use OCFS 1.4  with RHEL 5.8 VMs on ESX 4.1 and ESXi 5 as storage for a Hadoop cluster as well as an index and document store for application servers running a search engine application (java, tomcat, apache). We also export the same volumes from a few cluster nodes over NFS for developer access. We've had no stability issues or corruption.

I think I've identified the issue.

http://communities.vmware.com/thread/164967

In short, we had been allocated a shared vmdk through shared bus rather than as an RDM.  As a result, and after some digging into the logs, I found a bunch of SCSI reservation conflicts.  I'll get setup with RDM within the next week or so and will update with results as I have them.


From: Nathan Patwardhan [mailto:npatwardhan at llbean.com]<mailto:[mailto:npatwardhan at llbean.com]>
Sent: Tuesday, April 24, 2012 9:44 AM
To: ocfs2-users at oss.oracle.com<mailto:ocfs2-users at oss.oracle.com>
Subject: [Ocfs2-users] RHEL 5.8, ocfs2 v1.4.4, stability issues

Hi everyone,

We're running a two-node cluster w/ocfs2 v.1.4.4 under RHEL 5.8 and we're having major stability issues:

[root at splwww02 ~]# lsb_release -a ; uname -a ; rpm -qa |grep -i ocfs
LSB Version:    :core-4.0-amd64:core-4.0-ia32:core-4.0-noarch:graphics-4.0-amd64:graphics-4.0-ia32:graphics-4.0-noarch:printing-4.0-amd64:printing-4.0-ia32:printing-4.0-noarch
Distributor ID: RedHatEnterpriseServer
Description:    Red Hat Enterprise Linux Server release 5.8 (Tikanga)
Release:        5.8
Codename:       Tikanga

Linux splwww02.llbean.com 2.6.18-308.1.1.el5 #1 SMP Fri Feb 17 16:51:01 EST 2012 x86_64 x86_64 x86_64 GNU/Linux

ocfs2-2.6.18-308.1.1.el5-1.4.7-1.el5
ocfs2-tools-1.4.4-1.el5
ocfs2console-1.4.4-1.el5

These rpms were downloaded and installed from oracle.com.

The node are ESX VMs and the storage is shared to both VMs through vSphere.

We're using ocfs2 because we needed shared back end storage for a splunk implementation (splunk web search heads can used pooled storage and we didn't want to just use NFS for the obvious reasons - where shared storage seemed to be a better solution all the way around).

/etc/ocfs2/cluster.conf contains the following on each node:

cluster:
        node_count = 2
        name = splwww

node:
        ip_port = 7777
        ip_address = 10.130.245.40
        number = 0
        name = splwww01
        cluster = splwww

node:
        ip_port = 7777
        ip_address = 10.130.245.41
        number = 1
        name = splwww02
        cluster = splwww

Note that each node uses eth0 for clustering.  We do NOT currently have a private network setup in ESX for this purpose but CAN if this is recommended by people here.

o2cb status on both nodes:

Driver for "configfs": Loaded
Filesystem "configfs": Mounted
Driver for "ocfs2_dlmfs": Loaded
Filesystem "ocfs2_dlmfs": Mounted
Checking O2CB cluster splwww: Online
Heartbeat dead threshold = 3
  Network idle timeout: 90000
  Network keepalive delay: 5000
  Network reconnect delay: 5000
Checking O2CB heartbeat: Active

/etc/sysconfig/o2cb:

O2CB_ENABLED=true
O2CB_STACK=o2cb
O2CB_BOOTCLUSTER=splwww
O2CB_HEARTBEAT_THRESHOLD=3
O2CB_IDLE_TIMEOUT_MS=90000
O2CB_KEEPALIVE_DELAY_MS=5000
O2CB_RECONNECT_DELAY_MS=5000

Each host mounts the (ocfs2) file system then turns around and exports the  file system where it mounts it on localhost as shown below (yes I know this is awful but splunk does not support ocfs2 and as such is unable to lock files properly no matter what options I give to mount an ocfs2 file system):

# From /etc/fstab
/dev/mapper/splwwwshared-splwwwshared--lv /opt/splunk/pooled ocfs2 _netdev,datavolume,errors=remount-ro,rw,noatime 0 0

# From /etc/exports
/opt/splunk/pooled 127.0.0.1(rw,no_root_squash)

# From /bin/mount
/dev/mapper/splwwwshared-splwwwshared--lv on /opt/splunk/pooled type ocfs2 (rw,_netdev,noatime,datavolume,errors=remount-ro,heartbeat=local)
localhost:/opt/splunk/pooled on /opt/splunk/mnt type nfs (rw,nordirplus,addr=127.0.0.1)

The above should supposedly allow us to enjoy the benefits of an ocfs2 shared/clustered file system while functionally supporting what splunk needs.  Unfortunately we have experienced a number of problems with stability, summarized here:


1.       Cluster timeouts, node(s) leaving cluster.



Apr 21 21:03:50 splwww01 kernel: (o2hb-DA9199FC9F,3090,0):o2hb_do_disk_heartbeat:768 ERROR: status = -52

Apr 21 21:03:52 splwww01 kernel: (o2hb-DA9199FC9F,3090,0):o2hb_do_disk_heartbeat:777 ERROR: Device "dm-5": another node is heartbeating in our slot!



2.       Ocfs2 crashes, nfs mount becoming unavailable, /opt/splunk/pooled becoming inaccessible, splunk crashing.

Apr 24 06:53:43 splwww01 kernel: (o2hb-DA9199FC9F,3089,0):ocfs2_dlm_eviction_cb:98 device (253,5): dlm has evicted node 1


3.       inode irregularities on either/both node(s).



Apr 24 06:40:47 splwww01 kernel: (nfsd,3330,1):ocfs2_read_locked_inode:472 ERROR: status = -5



Apr 21 10:09:31 splwww01 kernel: (ocfs2rec,7676,0):ocfs2_clear_journal_error:690 ERROR: File system error -5 recorded in journal 0.

Apr 21 10:09:31 splwww01 kernel: (ocfs2rec,7676,0):ocfs2_clear_journal_error:692 ERROR: File system on device dm-5 needs checking.



4.       Full system crashes on both node(s).

Ideally, I'd love to continue using ocfs2 because it solves an age-old problem for us, but since we can't keep either system stable for more than an hour or so, I'm looking for some more insight as to what we're seeing and how we might resolve some of these issues.

Note also that I'm using monit to check a .state file I created under /opt/splunk/pool and re-mount all the fs, restart splunk, etc, but this only takes us so far once the system(s) crash and reboot.

--
Nathan Patwardhan, Sr. System Engineer
npatwardhan at llbean.com<mailto:npatwardhan at llbean.com>
x26662

-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://oss.oracle.com/pipermail/ocfs2-users/attachments/20120427/11067063/attachment-0001.html 


More information about the Ocfs2-users mailing list