[Ocfs2-users] Slow concurrent actions on the same LVM logicalvolume

Arnold Maderthaner arnold.maderthaner at j4care.com
Fri Aug 24 10:44:58 PDT 2007


We have the newest QLogic driver installed and I can write on full speed
(40mb/sec) with ocfs2 (also with ext3) when I'm the only one writing to the
device (even if it is mounted on both servers). The things I noticed today
is that it is not a write problem it is more a read problem. When I'm
reading files every 3-5 seconds it would halt for lets say 2-3 seconds.
Between it is fast. Even when I read the same data again then it is fast.
Do you have any comments on the server load ?

yours

Arnold


 I h
On 8/24/07, Sunil Mushran <Sunil.Mushran at oracle.com> wrote:
>
> mounted.ocfs2 is still not devmapper savvy. We'll address that
> in the next full release.
>
> Arnold Maderthaner wrote:
> > FYI the vg is setuped on /dev/sdc1.
> > Also I noticed some strange thing. The mounted command doesn't show
> > anything:
> > [root at impaxdb ~]# mounted.ocfs2 -d
> > Device                FS     UUID                                  Label
> > [root at impaxdb ~]# mounted.ocfs2 -f
> > Device                FS     Nodes
> > [root at impaxdb ~]#
> >
> > [root at impax ~]# mounted.ocfs2 -d
> > Device                FS     UUID                                  Label
> > [root at impax ~]# mounted.ocfs2 -f
> > Device                FS     Nodes
> > [root at impax ~]#
> >
> >
> > but devices are mounted:
> >
> > [root at impax ~]# mount -t ocfs2
> > /dev/mapper/raidVG-sts001LV on /srv/dcm4chee/sts/001 type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-sts002LV on /srv/dcm4chee/sts/002 type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-sts003LV on /srv/dcm4chee/sts/003 type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-cacheLV on /srv/dcm4chee/sts/cache type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-u01LV on /srv/oracle/u01 type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-u02LV on /srv/oracle/u02 type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-u03LV on /srv/oracle/u03 type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-u04LV on /srv/oracle/u04 type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-webstartLV on /var/www/html/webstart type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-configLV on /opt/dcm4chee/server/default/conf type
> > ocfs2 (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-xmbeanattrsLV on
> > /opt/dcm4chee/server/default/data/xmbean-attrs type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-installLV on /install type ocfs2
> > (rw,_netdev,heartbeat=local)
> >
> > [root at impaxdb ~]# mount -t ocfs2
> > /dev/mapper/raidVG-sts001LV on /srv/dcm4chee/sts/001 type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-sts002LV on /srv/dcm4chee/sts/002 type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-sts003LV on /srv/dcm4chee/sts/003 type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-cacheLV on /srv/dcm4chee/sts/cache type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-u01LV on /srv/oracle/u01 type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-u02LV on /srv/oracle/u02 type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-u03LV on /srv/oracle/u03 type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-u04LV on /srv/oracle/u04 type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-webstartLV on /var/www/html/webstart type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-configLV on /opt/dcm4chee/server/default/conf type
> > ocfs2 (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-xmbeanattrsLV on
> > /opt/dcm4chee/server/default/data/xmbean-attrs type ocfs2
> > (rw,_netdev,heartbeat=local)
> > /dev/mapper/raidVG-installLV on /install type ocfs2
> > (rw,_netdev,heartbeat=local)
> >
> >
> > yours
> >
> > arnold
> >
> >
> >
> >
> > On 8/24/07, *Arnold Maderthaner* <arnold.maderthaner at j4care.com
> > <mailto:arnold.maderthaner at j4care.com>> wrote:
> >
> >     vmstat server1:
> >     [root at impax ~]# vmstat 5
> >     procs -----------memory---------- ---swap-- -----io---- --system--
> >     -----cpu------
> >      r  b   swpd   free   buff  cache   si   so    bi    bo   in   cs
> >     us sy id wa st
> >      0  0      0 2760468 119924 872240    0    0     4     4  107
> >     63  0  0 99  0  0
> >      0  2      0 2760468 119928 872240    0    0     4    30 1095
> >     277  1  0 99  0  0
> >      0  0      0 2760468 119940 872248    0    0     7    11 1076
> >     232  0  0 100  0  0
> >      0  0      0 2760468 119948 872272    0    0     4    18 1084
> >     244  0  0 100  0  0
> >      0  0      0 2760468 119948 872272    0    0     7    34 1063
> >     220  0  0 100  0  0
> >      0  0      0 2760468 119948 872272    0    0     0     1 1086
> >     243  0  0 100  0  0
> >      0  0      0 2760716 119956 872272    0    0     7    10 1065
> >     234  2  0 98  0  0
> >     vmstart server2:
> >     [root at impaxdb ~]# vmstat 5
> >     procs -----------memory---------- ---swap-- -----io---- --system--
> >     -----cpu------
> >      r  b   swpd   free   buff  cache   si   so    bi    bo   in   cs
> >     us sy id wa st
> >      0  0  24676 143784  46428 3652848    0    0    11    20  107
> >     122  0  0 84 16  0
> >      0  1  24676 143784  46452 3652832    0    0     4    14 1080
> >     237  0  0 89 11  0
> >      0  1  24672 143660  46460 3652812    6    0    14    34 1074
> >     246  0  0 90 10  0
> >      0  0  24672 143660  46460 3652852    0    0     7    22 1076
> >     246  0  0 80 20  0
> >      0  1  24672 143660  46476 3652844    0    0     7    10 1068
> >     227  0  0 85 15  0
> >
> >     iostat server1:
> >
> >
> >     [root at impax ~]# iostat -x 5
> >     Linux 2.6.18-8.el5PAE (impax)   08/24/2007
> >
> >     avg-cpu:  %user   %nice %system %iowait  %steal   %idle
> >                0.07    0.00    0.04    0.45    0.00   99.44
> >
> >     Device:         rrqm/s   wrqm/s   r/s   w/s   rsec/s   wsec/s
> >     avgrq-sz avgqu-sz   await  svctm  %util
> >     sda               0.24     0.90  0.23  0.76    15.72    13.33
> >     29.22     0.01   12.68   4.77   0.47
> >     sdb               0.03     1.23  0.02  0.58     1.34     6.81
> >     13.59     0.00    0.79   0.30   0.02
> >     sdc               0.58     0.87  3.88  3.78    16.69    10.87
> >     3.60     7.98 1042.08  89.69  68.68
> >     dm-0              0.00     0.00   0.31  1.14     0.96     6.99
> >     5.46     1.55 1061.28 459.00  66.92
> >     dm-1              0.00     0.00  0.31  0.31     0.95     0.31
> >     2.01     0.66 1054.16 1054.17  66.15
> >     dm-2              0.00     0.00  0.31  0.31      0.95     0.31
> >     2.01     0.66 1057.72 1057.84  66.24
> >     dm-3              0.00     0.00  0.96  0.34     6.15     0.55
> >     5.12     0.77  592.01 504.72  66.04
> >     dm-4              0.00     0.00  0.32  0.33     0.96      0.41
> >     2.14     0.66 1031.13 1028.75  66.00
> >     dm-5              0.00     0.00  0.31  0.32     0.95     0.40
> >     2.12     0.66 1037.27 1036.55  66.09
> >     dm-6              0.00     0.00  0.31  0.31     0.94     0.31
> >     2.01     0.66 1064.34 1064.45  66.37
> >     dm-7              0.00     0.00  0.32  0.31     0.95     0.32
> >     2.01     0.66 1043.05 1042.99  65.77
> >     dm-8              0.00     0.00  0.31  0.31     0.95     0.32
> >     2.03      0.66 1050.73 1048.23  65.97
> >     dm-9              0.00     0.00  0.34  0.31     0.99     0.31
> >     2.01     0.67 1027.41 1006.68  65.71
> >     dm-10             0.00     0.00  0.32  0.32     0.95     0.32
> >     2.01     0.68 1078.71 1050.22  66.41
> >     dm-12             0.00     0.00  0.31  0.31     0.96     0.31
> >     2.03     0.66 1065.33 1065.10  66.43
> >     dm-13             0.00     0.00  0.04  1.27     1.27     2.54
> >     2.91     0.00    2.43   0.10   0.01
> >     dm-18             0.00     0.00  0.01  0.35     0.04     2.80
> >     8.00     0.01   25.42   0.03   0.00
> >     dm-19             0.00     0.00  0.00  0.18     0.01     1.47
> >     8.00     0.00    0.42   0.21   0.00
> >     dm-21             0.00     0.00  0.02  0.01     2.20     0.11
> >     65.49     0.00    9.84   3.01   0.01
> >     dm-26             0.00     0.00  0.01  0.05     0.12     0.43
> >     9.21     0.00   11.66   7.05   0.04
> >     dm-27             0.00     0.00   0.00  0.21     0.02     1.70
> >     8.02     0.00   12.09   4.57   0.10
> >     dm-28             0.00     0.00  0.26  0.11     8.90     0.91
> >     26.40     0.01   14.94   1.85   0.07
> >     dm-29             0.00     0.00   0.05  1.03      1.04
> >     8.22     8.56     0.01   13.76   2.48   0.27
> >
> >     avg-cpu:  %user   %nice %system %iowait  %steal   %idle
> >                0.00    0.00    0.00    0.40    0.00   99.60
> >
> >     Device:         rrqm/s   wrqm/s   r/s   w/s   rsec/s   wsec/s
> >     avgrq-sz avgqu-sz   await  svctm  %util
> >     sda               0.00     2.80  0.00  2.20     0.00    40.00
> >     18.18     0.03   12.27   4.73   1.04
> >     sdb               0.00     0.40  0.00  0.40     0.00     1.60
> >     4.00     0.00    5.50   5.50   0.22
> >     sdc               0.00     0.00  4.80  4.80    14.40     4.80
> >     2.00     7.21 1560.12  63.67  61.12
> >     dm-0              0.00     0.00   0.40  0.40     1.20     0.40
> >     2.00     0.60 1559.50 750.25  60.02
> >     dm-1              0.00     0.00  0.40  0.40     1.20     0.40
> >     2.00     0.60 1558.50 750.50  60.04
> >     dm-2              0.00     0.00  0.40  0.40      1.20     0.40
> >     2.00     0.60 1559.50 752.75  60.22
> >     dm-3              0.00     0.00  0.40  0.40     1.20     0.40
> >     2.00     0.60 1562.75 752.25  60.18
> >     dm-4              0.00     0.00  0.40  0.40     1.20     0.40
> >     2.00     0.60 1559.50 752.50  60.20
> >     dm-5              0.00     0.00  0.40  0.40     1.20     0.40
> >     2.00     0.60 1563.00 750.25  60.02
> >     dm-6              0.00     0.00  0.40  0.40     1.20     0.40
> >     2.00      0.60 1558.50 750.25  60.02
> >     dm-7              0.00     0.00  0.40  0.40     1.20     0.40
> >     2.00     0.60 1559.50 752.50  60.20
> >     dm-8              0.00     0.00  0.40  0.40     1.20     0.40
> >     2.00     0.60 1559.75 750.25  60.02
> >     dm-9              0.00     0.00  0.40  0.40     1.20     0.40
> >     2.00     0.60 1554.00 750.25  60.02
> >     dm-10             0.00     0.00  0.40  0.40     1.20     0.40
> >     2.00     0.60 1555.75 752.25   60.18
> >     dm-12             0.00     0.00  0.40  0.40     1.20     0.40
> >     2.00     0.60 1571.25 750.25  60.02
> >     dm-13             0.00     0.00  0.00  0.80     0.00     1.60
> >     2.00     0.01    8.25   2.75   0.22
> >     dm-18             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-19             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-21             0.00     0.00   0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-26             0.00     0.00  0.00  0.60     0.00     4.80
> >     8.00     0.01   12.67   8.33   0.50
> >     dm-27             0.00     0.00   0.00  0.00      0.00
> >     0.00     0.00     0.00    0.00   0.00   0.00
> >     dm-28             0.00     0.00  0.00  1.80     0.00    14.40
> >     8.00     0.02   10.22   2.44   0.44
> >     dm-29             0.00     0.00  0.00  0.80      0.00     6.40
> >     8.00     0.01   10.25   3.75   0.30
> >
> >     avg-cpu:  %user   %nice %system %iowait  %steal   %idle
> >                0.05    0.00    0.00    0.00    0.00   99.95
> >
> >     Device:         rrqm/s   wrqm/s   r/s   w/s   rsec/s   wsec/s
> >     avgrq-sz avgqu-sz   await  svctm  %util
> >     sda               0.00     0.20  0.00  0.60     0.00     6.40
> >     10.67     0.01   14.67   6.00   0.36
> >     sdb               0.00     0.40  0.00  0.40     0.00     1.60
> >     4.00     0.00    0.00   0.00   0.00
> >     sdc               0.00     0.00  4.80  4.80    14.40     4.80
> >     2.00     8.81 1292.92  77.06  73.98
> >     dm-0              0.00     0.00   0.20  0.40     0.60     0.40
> >     1.67     0.74 1723.33 1233.00  73.98
> >     dm-1              0.00     0.00  0.20  0.40     0.60     0.40
> >     1.67     0.73 1722.00 1223.33  73.40
> >     dm-2              0.00     0.00  0.20  0.40      0.60     0.40
> >     1.67     0.73 1726.33 1222.00  73.32
> >     dm-3              0.00     0.00  0.20  0.40     0.60     0.40
> >     1.67     0.73 1725.67 1222.00  73.32
> >     dm-4              0.00     0.00  0.20  0.40     0.60      0.40
> >     1.67     0.73 1726.00 1222.00  73.32
> >     dm-5              0.00     0.00  0.20  0.40     0.60     0.40
> >     1.67     0.73 1722.33 1222.00  73.32
> >     dm-6              0.00     0.00  0.20  0.40     0.60     0.40
> >     1.67     0.73 1722.00 1223.67  73.42
> >     dm-7              0.00     0.00  0.20  0.40     0.60     0.40
> >     1.67     0.73 1726.00 1222.00  73.32
> >     dm-8              0.00     0.00  0.20  0.40     0.60     0.40
> >     1.67      0.73 1722.33 1222.00  73.32
> >     dm-9              0.00     0.00  0.20  0.40     0.60     0.40
> >     1.67     0.73 1723.00 1214.33  72.86
> >     dm-10             0.00     0.00  0.20  0.40     0.60     0.40
> >     1.67     0.73 1725.67 1222.00  73.32
> >     dm-12             0.00     0.00  0.20  0.40     0.60     0.40
> >     1.67     0.74 1722.00 1231.67  73.90
> >     dm-13             0.00     0.00  0.00  0.80     0.00     1.60
> >     2.00     0.00    0.00    0.00   0.00
> >     dm-18             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-19             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-21             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-26             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-27             0.00     0.00  0.00  0.60     0.00     4.80
> >     8.00     0.01   13.00   4.33   0.26
> >     dm-28             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-29             0.00     0.00   0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >
> >
> >     iostat server2:
> >
> >     [root at impaxdb ~]# iostat -x 5
> >     Linux 2.6.18-8.el5PAE (impaxdb)         08/24/2007
> >
> >     avg-cpu:  %user   %nice %system %iowait  %steal   %idle
> >                0.48    0.00    0.10   15.70    0.00   83.72
> >
> >     Device:         rrqm/s   wrqm/s   r/s   w/s   rsec/s   wsec/s
> >     avgrq-sz avgqu-sz   await  svctm  %util
> >     sda               0.28     1.38  0.32  0.82    18.52    17.59
> >     31.55     0.00    1.98   0.83   0.10
> >     sdb               0.01     1.73  0.02  0.04     3.20    14.15
> >     297.21     0.00    9.18   0.87   0.01
> >     sdc               6.44    14.59  3.94  4.26    65.36   126.12
> >     23.36     9.52 1160.63  91.33  74.88
> >     dm-0              0.00     0.00   0.29  0.29     0.89     0.29
> >     2.01     0.71 1214.34 1214.01  71.18
> >     dm-1              0.00     0.00  0.29  0.29     0.89     0.29
> >     2.01     0.71 1206.47 1206.38  70.95
> >     dm-2              0.00     0.00  0.29  0.29      0.89     0.29
> >     2.01     0.71 1207.35 1207.27  70.96
> >     dm-3              0.00     0.00  7.10  9.84    55.33    76.63
> >     7.79    25.51 1506.27  43.35  73.42
> >     dm-4              0.00     0.00  0.32  5.48     1.10    41.82
> >     7.39     9.21 1586.36 122.41  71.06
> >     dm-5              0.00     0.00  0.30  0.88     0.90     5.02
> >     5.02     0.72  606.80 600.72  70.89
> >     dm-6              0.00     0.00  0.29  0.29     0.89     0.29
> >     2.01      0.71 1206.90 1206.94  70.91
> >     dm-7              0.00     0.00  0.29  0.29     0.89     0.29
> >     2.01     0.71 1202.65 1202.69  70.82
> >     dm-8              0.00     0.00  0.29  0.29     0.89     0.29
> >     2.01      0.71 1204.47 1204.43  70.84
> >     dm-9              0.00     0.00  0.29  0.29     0.89     0.29
> >     2.01     0.71 1202.38 1202.42  70.77
> >     dm-10             0.00     0.00  0.29  0.29     0.88     0.29
> >     2.01     0.71 1208.53 1208.48  70.93
> >     dm-12             0.00     0.00  0.30  0.29     0.91     0.29
> >     2.04     0.71 1202.82 1202.49  70.92
> >     dm-13             0.00     0.00  0.00  0.00     0.01     0.00
> >     6.99     0.00    0.47   0.06   0.00
> >     dm-18             0.00     0.00  0.02  1.77     3.16    14.15
> >     9.67     0.03   16.69   0.03   0.01
> >     dm-19             0.00     0.00  0.00  0.00     0.01     0.00
> >     8.02     0.00    0.69   0.34   0.00
> >     dm-20             0.00     0.00  0.05  0.28     3.48     2.26
> >     17.18     0.00    3.37   0.47   0.02
> >     dm-25             0.00     0.00  0.01  0.05     0.16     0.44
> >     9.68     0.00    1.20   0.65   0.00
> >     dm-26             0.00     0.00   0.00  0.14     0.02     1.11
> >     8.03     0.00    3.28   0.05   0.00
> >     dm-27             0.00     0.00  0.31  0.15     9.69     1.24
> >     23.61     0.00    9.39   0.88   0.04
> >     dm-28             0.00     0.00   0.06  1.08      1.24
> >     8.61     8.65     0.00    1.04   0.21   0.02
> >
> >     avg-cpu:  %user   %nice %system %iowait  %steal   %idle
> >                0.05    0.00    0.05   23.62    0.00   76.27
> >
> >     Device:         rrqm/s   wrqm/s   r/s   w/s   rsec/s   wsec/s
> >     avgrq-sz avgqu-sz   await  svctm  %util
> >     sda               0.00     4.85  0.00  2.22     0.00    56.57
> >     25.45     0.00    0.91   0.36   0.08
> >     sdb               0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     sdc               0.00     1.21  4.85  5.45    14.55    24.24
> >     3.76    12.51 1556.98  94.24  97.09
> >     dm-0              0.00     0.00   0.20  0.40     0.61     0.40
> >     1.67     0.97 2055.67 1602.00  97.09
> >     dm-1              0.00     0.00  0.20  0.40     0.61     0.40
> >     1.67     0.96 2064.67 1589.00  96.30
> >     dm-2              0.00     0.00  0.20  0.40      0.61     0.40
> >     1.67     0.96 2063.67 1582.67  95.92
> >     dm-3              0.00     0.00  0.20  0.40     0.61     0.40
> >     1.67     4.79 8988.33 1582.00  95.88
> >     dm-4              0.00     0.00  0.20  1.21     0.61      6.87
> >     5.29     0.96  886.57 679.14  96.04
> >     dm-5              0.00     0.00  0.20  1.21     0.61     6.87
> >     5.29     0.96  878.00 680.00  96.16
> >     dm-6              0.00     0.00  0.20  0.40     0.61     0.40
> >     1.67     0.97 2055.67 1593.00  96.55
> >     dm-7              0.00     0.00  0.20  0.40     0.61     0.40
> >     1.67     0.96 2064.67 1584.33  96.02
> >     dm-8              0.00     0.00  0.20  0.40     0.61     0.40
> >     1.67      0.96 2064.67 1586.33  96.14
> >     dm-9              0.00     0.00  0.20  0.40     0.61     0.40
> >     1.67     0.96 2064.67 1583.33  95.96
> >     dm-10             0.00     0.00  0.20  0.40     0.61     0.40
> >     1.67     0.97 2055.67 1597.67  96.83
> >     dm-12             0.00     0.00  0.20  0.40     0.61     0.40
> >     1.67     0.97 2071.67 1595.67  96.71
> >     dm-13             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-18             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-19             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-20             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-25             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-26             0.00     0.00   0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-27             0.00     0.00  0.00  0.81     0.00     6.46
> >     8.00     0.00    0.00   0.00   0.00
> >     dm-28             0.00     0.00   0.00  4.85      0.00
> >     38.79     8.00     0.01    2.17   0.17   0.08
> >
> >     avg-cpu:  %user   %nice %system %iowait  %steal   %idle
> >                0.00    0.00    0.00   19.85    0.00   80.15
> >
> >     Device:         rrqm/s   wrqm/s   r/s   w/s   rsec/s   wsec/s
> >     avgrq-sz avgqu-sz   await  svctm  %util
> >     sda               0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     sdb               0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     sdc               0.00     0.61  0.00  0.00     0.00     0.00
> >     0.00     9.12    0.00   0.00  80.26
> >     dm-0              0.00     0.00   0.20  0.00     0.61     0.00
> >     3.00     0.65    0.00 3196.00  64.70
> >     dm-1              0.00     0.00  0.20  0.00     0.61     0.00
> >     3.00     0.65    0.00 3235.00  65.49
> >     dm-2              0.00     0.00  0.20  0.00      0.61     0.00
> >     3.00     0.66    0.00 3254.00  65.87
> >     dm-3              0.00     0.00  0.40  0.81     2.23     6.48
> >     7.17     3.29    0.00 660.83  80.26
> >     dm-4              0.00     0.00  0.20  0.00     0.61     0.00
> >     3.00     0.66    0.00 3248.00  65.75
> >     dm-5              0.00     0.00  0.20  0.00     0.61     0.00
> >     3.00     0.66    0.00 3242.00  65.63
> >     dm-6              0.00     0.00  0.20  0.00     0.61     0.00
> >     3.00     0.65    0.00 3223.00  65.24
> >     dm-7              0.00     0.00  0.20  0.00     0.61     0.00
> >     3.00     0.66    0.00 3249.00  65.77
> >     dm-8              0.00     0.00  0.20  0.00     0.61     0.00
> >     3.00     0.66    0.00 3243.00  65.65
> >     dm-9              0.00     0.00  0.20  0.00     0.61     0.00
> >     3.00     0.66    0.00 3252.00  65.83
> >     dm-10             0.00     0.00  0.20  0.00     0.61     0.00
> >     3.00     0.65    0.00 3209.00  64.96
> >     dm-12             0.00     0.00  0.20  0.00     0.61     0.00
> >     3.00     0.65    0.00 3215.00  65.08
> >     dm-13             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00    0.00   0.00
> >     dm-18             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-19             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-20             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-25             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-26             0.00     0.00   0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-27             0.00     0.00  0.00  0.00     0.00     0.00
> >     0.00     0.00    0.00   0.00   0.00
> >     dm-28             0.00     0.00   0.00  0.00      0.00
> >     0.00     0.00     0.00    0.00   0.00   0.00
> >
> >     top server1:
> >
> >
> >
> >     Tasks: 217 total,   1 running, 214 sleeping,   0 stopped,   2 zombie
> >     Cpu(s):  0.1%us,  0.0%sy,  0.0%ni, 99.4%id,  0.4%wa,  0.0%hi,
> >     0.0%si,  0.0%st
> >     Mem:   4084444k total,  1323728k used,  2760716k free,   120072k
> >     buffers
> >     Swap:  8193020k total,        0k used,  8193020k free,   872284k
> >     cached
> >
> >       PID USER      PR  NI  VIRT  RES  SHR S %CPU %MEM    TIME+  COMMAND
> >     14245 root      16   0  2288  988  704 R    2  0.0   0:00.01 top
> >         1 root      15   0  2036  652  564 S    0  0.0   0:01.13 init
> >         2 root      RT   0     0    0    0 S    0  0.0   0:00.00
> >     migration/0
> >         3 root      34  19     0    0    0 S    0  0.0   0:00.46
> >     ksoftirqd/0
> >         4 root      RT   0     0    0    0 S    0  0.0   0:00.00
> >     watchdog/0
> >         5 root      RT   0     0    0    0 S    0  0.0   0:00.00
> >     migration/1
> >         6 root      34  19     0    0    0 S    0  0.0   0:00.07
> >     ksoftirqd/1
> >         7 root      RT   0     0    0    0 S    0  0.0   0:00.00
> >     watchdog/1
> >         8 root      RT   0     0    0    0 S    0  0.0   0:00.00
> >     migration/2
> >         9 root      34  19     0    0    0 S    0  0.0   0:00.19
> >     ksoftirqd/2
> >        10 root      RT   0     0    0    0 S    0  0.0   0:00.00
> >     watchdog/2
> >        11 root      RT   0     0    0    0 S    0  0.0   0:00.00
> >     migration/3
> >        12 root      34  19     0    0    0 S    0  0.0   0:00.23
> >     ksoftirqd/3
> >        13 root      RT   0     0    0    0 S    0  0.0   0:00.00
> >     watchdog/3
> >        14 root      10  -5     0    0    0 S    0  0.0   0:00.00events/0
> >        15 root      10  -5     0    0    0 S    0  0.0   0:00.00events/1
> >        16 root      10  -5     0    0    0 S    0  0.0   0:00.00events/2
> >        17 root      10  -5     0    0    0 S    0  0.0    0:00.00events/3
> >        18 root      19  -5     0    0    0 S    0  0.0   0:00.00 khelper
> >        19 root      10  -5     0    0    0 S    0  0.0   0:00.00 kthread
> >        25 root      10  -5     0    0    0 S    0  0.0   0: 00.00
> >     kblockd/0
> >        26 root      10  -5     0    0    0 S    0  0.0   0:00.04kblockd/1
> >        27 root      10  -5     0    0    0 S    0  0.0   0:00.00kblockd/2
> >        28 root      10  -5     0    0    0 S    0  0.0   0: 00.06
> >     kblockd/3
> >        29 root      14  -5     0    0    0 S    0  0.0   0:00.00 kacpid
> >       147 root      14  -5     0    0    0 S    0  0.0   0:00.00cqueue/0
> >       148 root      14  -5     0    0    0 S    0  0.0   0:00.00cqueue/1
> >       149 root      14  -5     0    0    0 S    0  0.0   0:00.00cqueue/2
> >       150 root      14  -5     0    0    0 S    0  0.0   0:00.00cqueue/3
> >       153 root      10  -5     0    0    0 S    0  0.0   0:00.00 khubd
> >       155 root      10  -5     0    0    0 S    0  0.0   0:00.00 kseriod
> >       231 root      19   0     0    0    0 S    0  0.0   0:00.00 pdflush
> >       232 root      15   0     0    0    0 S    0  0.0   0:00.15 pdflush
> >       233 root      14  -5     0    0    0 S    0  0.0   0:00.00 kswapd0
> >       234 root      14  -5     0    0    0 S    0  0.0   0:00.00 aio/0
> >       235 root      14  -5     0    0    0 S    0  0.0   0:00.00 aio/1
> >       236 root      14  -5     0    0    0 S    0  0.0   0:00.00 aio/2
> >
> >
> >     top server2:
> >
> >     Tasks: 264 total,   1 running, 260 sleeping,   0 stopped,   3 zombie
> >     Cpu(s):  0.5%us,  0.1%sy,  0.0%ni, 83.7%id, 15.7%wa,  0.0%hi,
> >     0.0%si,  0.0%st
> >     Mem:   4084444k total,  3941216k used,   143228k free,    46608k
> >     buffers
> >     Swap:  8193020k total,    24660k used,  8168360k free,  3652864k
> >     cached
> >
> >       PID USER      PR  NI  VIRT  RES  SHR S %CPU %MEM    TIME+  COMMAND
> >      9507 root      15   0  2292 1020  704 R    2  0.0   0:00.01 top
> >         1 root      15   0  2032  652  564 S    0  0.0   0:01.20 init
> >         2 root      RT   0     0    0    0 S    0  0.0   0:00.00
> >     migration/0
> >         3 root      34  19     0    0    0 S    0  0.0   0:00.25
> >     ksoftirqd/0
> >         4 root      RT   0     0    0    0 S    0  0.0   0:00.00
> >     watchdog/0
> >         5 root      RT   0     0    0    0 S    0  0.0   0:00.00
> >     migration/1
> >         6 root      34  19     0    0    0 S    0  0.0   0:00.42
> >     ksoftirqd/1
> >         7 root      RT   0     0    0    0 S    0  0.0   0:00.00
> >     watchdog/1
> >         8 root      RT   0     0    0    0 S    0  0.0   0:00.00
> >     migration/2
> >         9 root      34  19     0    0    0 S    0  0.0   0:00.56
> >     ksoftirqd/2
> >        10 root      RT   0     0    0    0 S    0  0.0   0:00.00
> >     watchdog/2
> >        11 root      RT   0     0    0    0 S    0  0.0   0:00.00
> >     migration/3
> >        12 root      39  19     0    0    0 S    0  0.0   0:00.08
> >     ksoftirqd/3
> >        13 root      RT   0     0    0    0 S    0  0.0   0:00.00
> >     watchdog/3
> >        14 root      10  -5     0    0    0 S    0  0.0   0:00.03events/0
> >        15 root      10  -5     0    0    0 S    0  0.0   0:00.00events/1
> >        16 root      10  -5     0    0    0 S    0  0.0   0:00.02events/2
> >        17 root      10  -5     0    0    0 S    0  0.0    0:00.02events/3
> >        18 root      10  -5     0    0    0 S    0  0.0   0:00.00 khelper
> >        19 root      11  -5     0    0    0 S    0  0.0   0:00.00 kthread
> >        25 root      10  -5     0    0    0 S    0  0.0   0: 00.08
> >     kblockd/0
> >        26 root      10  -5     0    0    0 S    0  0.0   0:00.00kblockd/1
> >        27 root      10  -5     0    0    0 S    0  0.0   0:00.00kblockd/2
> >        28 root      10  -5     0    0    0 S    0  0.0   0: 00.08
> >     kblockd/3
> >        29 root      18  -5     0    0    0 S    0  0.0   0:00.00 kacpid
> >       147 root      18  -5     0    0    0 S    0  0.0   0:00.00cqueue/0
> >       148 root      20  -5     0    0    0 S    0  0.0   0:00.00cqueue/1
> >       149 root      10  -5     0    0    0 S    0  0.0   0:00.00cqueue/2
> >       150 root      10  -5     0    0    0 S    0  0.0   0:00.00cqueue/3
> >       153 root      10  -5     0    0    0 S    0  0.0   0:00.00 khubd
> >       155 root      10  -5     0    0    0 S    0  0.0   0:00.00 kseriod
> >       233 root      15  -5     0    0    0 S    0  0.0   0:00.84 kswapd0
> >       234 root      20  -5     0    0    0 S    0  0.0   0:00.00 aio/0
> >       235 root      20  -5     0    0    0 S    0  0.0   0:00.00 aio/1
> >       236 root      20  -5     0    0    0 S    0  0.0   0:00.00 aio/2
> >       237 root      20  -5     0    0    0 S    0  0.0   0:00.00 aio/3
> >       409 root      11  -5     0    0    0 S    0  0.0   0:00.00kpsmoused
> >
> >     slabtop server1:
> >
> >      Active / Total Objects (% used)    : 301060 / 312027 (96.5%)
> >      Active / Total Slabs (% used)      : 12055 / 12055 (100.0%)
> >      Active / Total Caches (% used)     : 106 / 146 (72.6%)
> >      Active / Total Size (% used)       : 45714.46K / 46814.27K (97.7%)
> >      Minimum / Average / Maximum Object : 0.01K / 0.15K / 128.00K
> >
> >       OBJS ACTIVE  USE OBJ SIZE  SLABS OBJ/SLAB CACHE SIZE NAME
> >     148176 148095  99%    0.05K   2058       72      8232K buffer_head
> >      44573  44497  99%    0.13K   1537       29      6148K dentry_cache
> >      32888  32878  99%    0.48K   4111        8     16444K
> >     ext3_inode_cache
> >      10878  10878 100%    0.27K    777       14      3108K
> radix_tree_node
> >       7943   7766  97%    0.02K     47      169       188K dm_io
> >       7917   7766  98%    0.02K     39      203       156K dm_tio
> >       6424   5414  84%    0.09K    146       44       584K
> vm_area_struct
> >       6384   6223  97%    0.04K     76       84       304K
> sysfs_dir_cache
> >       5989   5823  97%    0.03K     53      113       212K size-32
> >       5074   4833  95%    0.06K     86       59       344K size-64
> >       4488   4488 100%    0.33K    408       11      1632K inode_cache
> >       2610   2561  98%    0.12K     87       30       348K size-128
> >       2540   1213  47%    0.01K     10      254        40K anon_vma
> >       2380   1475  61%    0.19K    119       20       476K filp
> >       1430   1334  93%    0.35K    130       11       520K
> >     proc_inode_cache
> >       1326   1031  77%    0.05K     17       78        68K
> >     selinux_inode_security
> >       1320   1304  98%    0.25K     88       15       352K size-256
> >       1015    629  61%    0.02K      5      203        20K biovec-1
> >       1008    125  12%    0.05K     14       72        56K journal_head
> >        930    841  90%    0.12K     31       30       124K bio
> >        920    666  72%    0.19K     46       20       184K
> >     skbuff_head_cache
> >        798    757  94%    2.00K    399        2      1596K size-2048
> >        791    690  87%    0.03K      7      113        28K ocfs2_em_ent
> >        784    748  95%    0.50K     98        8       392K size-512
> >        590    503  85%    0.06K     10       59        40K biovec-4
> >        564    562  99%    0.88K    141        4       564K
> >     ocfs2_inode_cache
> >        546    324  59%    0.05K      7       78        28K
> delayacct_cache
> >        531    497  93%    0.43K     59        9       236K
> >     shmem_inode_cache
> >        520    487  93%    0.19K     26       20       104K biovec-16
> >        505    327  64%    0.04K       5      101        20K pid
> >        500    487  97%    0.75K    100        5       400K biovec-64
> >        460    376  81%    0.04K      5       92        20K Acpi-Operand
> >        452    104  23%    0.03K      4      113        16K pgd
> >        347    347 100%    4.00K    347        1      1388K size-4096
> >        338    303  89%    0.02K      2      169         8K
> Acpi-Namespace
> >        336    175  52%    0.04K      4       84        16K crq_pool
> >
> >
> >     slabtop server2:
> >
> >
> >      Active / Total Objects (% used)    : 886188 / 907066 (97.7%)
> >      Active / Total Slabs (% used)      : 20722 / 20723 (100.0%)
> >      Active / Total Caches (% used)     : 104 / 147 (70.7%)
> >      Active / Total Size (% used)       : 76455.01K / 78620.30K (97.2%)
> >      Minimum / Average / Maximum Object : 0.01K / 0.09K / 128.00K
> >
> >       OBJS ACTIVE  USE OBJ SIZE  SLABS OBJ/SLAB CACHE SIZE NAME
> >     736344 729536  99%    0.05K  10227       72     40908K buffer_head
> >      37671  34929  92%    0.13K   1299       29      5196K dentry_cache
> >      32872  32666  99%    0.48K   4109        8     16436K
> >     ext3_inode_cache
> >      20118  20097  99%    0.27K   1437       14      5748K
> radix_tree_node
> >      10912   9437  86%    0.09K    248       44       992K
> vm_area_struct
> >       7714   7452  96%    0.02K     38      203       152K dm_tio
> >       7605   7452  97%    0.02K     45      169       180K dm_io
> >       6552   6475  98%    0.04K     78       84       312K
> sysfs_dir_cache
> >       5763   5461  94%    0.03K     51      113       204K size-32
> >       4661   4276  91%    0.06K     79       59       316K size-64
> >       4572   3569  78%    0.01K     18      254        72K anon_vma
> >       4440   3770  84%    0.19K    222       20       888K filp
> >       2288   1791  78%    0.35K    208       11       832K
> >     proc_inode_cache
> >       1980   1884  95%    0.12K     66       30       264K size-128
> >       1804   1471  81%    0.33K    164       11       656K inode_cache
> >       1404   1046  74%    0.05K     18       78        72K
> >     selinux_inode_security
> >       1060    670  63%    0.19K     53       20       212K
> >     skbuff_head_cache
> >        904    767  84%    0.03K      8      113        32K ocfs2_em_ent
> >        870    870 100%    0.12K     29       30       116K bio
> >        840    816  97%    0.50K    105        8       420K size-512
> >        812    578  71%    0.02K      4      203        16K biovec-1
> >        810    800  98%    0.25K     54       15       216K size-256
> >        800    761  95%    2.00K    400        2      1600K size-2048
> >        549    514  93%    0.43K     61        9       244K
> >     shmem_inode_cache
> >        546    289  52%    0.05K      7       78        28K
> delayacct_cache
> >        531    478  90%    0.06K      9       59        36K biovec-4
> >        520    471  90%    0.19K     26       20       104K biovec-16
> >        505    292  57%    0.04K      5      101        20K pid
> >        490    479  97%    0.75K     98        5       392K biovec-64
> >        460    376  81%    0.04K       5       92        20K Acpi-Operand
> >        452    175  38%    0.03K      4      113        16K pgd
> >        440    357  81%    0.38K     44       10       176K
> >     sock_inode_cache
> >        369    303  82%    0.44K     41        9       164K UNIX
> >        360    320  88%    1.00K     90        4       360K size-1024
> >        354    124  35%    0.06K      6       59        24K fs_cache
> >        351    351 100%    4.00K    351        1      1404K pmd
> >
> >     uptime server1:
> >
> >     [root at impax ~]# uptime
> >      10:21:09 up 18:16,  1 user,  load average: 7.22, 7.72, 7.83
> >
> >     uptime server2:
> >
> >     [root at impaxdb ~]# uptime
> >      10:21:17 up 18:16,  1 user,  load average: 8.79, 9.02, 8.98
> >
> >     yours
> >
> >     arnold
> >
> >     On 8/24/07, *Alexei_Roudnev* <Alexei_Roudnev at exigengroup.com
> >     <mailto:Alexei_Roudnev at exigengroup.com>> wrote:
> >
> >         Run and send here, please:
> >
> >         vmstat 5
> >         (5 lines)
> >
> >         iostat -x 5
> >         (2 - 3 outputs)
> >
> >         top
> >         (1 screen)
> >
> >         slabtop (or equivalent, I dont remember)
> >
> >
> >         ----- Original Message -----
> >
> >             *From:* Arnold Maderthaner
> >             <mailto:arnold.maderthaner at j4care.com>
> >             *To:* Sunil Mushran <mailto:Sunil.Mushran at oracle.com> ;
> >             ocfs2-users at oss.oracle.com
> >             <mailto:ocfs2-users at oss.oracle.com>
> >             *Sent:* Thursday, August 23, 2007 4:12 AM
> >             *Subject:* Re: [Ocfs2-users] Slow concurrent actions on
> >             the same LVM logicalvolume
> >
> >
> >             Hi !
> >
> >             I watched the servers today and noticed that the server
> >             load is very high but not much is working on the server:
> >             server1:
> >              16:37:11 up 32 min,  1 user,  load average: 6.45, 7.90,
> 7.45
> >             server2:
> >              16:37:19 up 32 min,  4 users,  load average: 9.79, 9.76,
> 8.26
> >             i also tried to do some selects on the oracle database
> today.
> >             to do a "select count(*) from xyz" where xyz has 120 000
> >             rows it took me about 10 sec.
> >             always the process who does the filesystem access is the
> >             first one.
> >             and every command (ls,du,vgdisplay,lvdisplay,dd,cp,....)
> >             which needs access to a ocfs2 filesystem seams like
> >             blocking or it seams to me that there is a lock.
> >             Can anyone please help me ?
> >
> >             yours
> >
> >             Arnold
> >
> >             On 8/22/07, *Arnold Maderthaner*
> >             <arnold.maderthaner at j4care.com
> >             <mailto:arnold.maderthaner at j4care.com>> wrote:
> >
> >                 There is not really a high server load when writing to
> >                 the disks but will try it asap. what do you mean with
> >                 AST ?
> >                 Can I enable some debuging on ocfs2 ? We have 2 Dell
> >                 servers with about 4gb of ram each and 2 Dual Core
> >                 CPUs with 2GHZ each.
> >                 Can you please describe which tests todo again ?
> >
> >                 Thx for your help
> >
> >                 Arnold
> >
> >
> >                 On 8/22/07, *Sunil Mushran *< Sunil.Mushran at oracle.com
> >                 <mailto:Sunil.Mushran at oracle.com>> wrote:
> >
> >                     Repeat the first test. This time run top on the
> >                     first server. Which
> >                     process is eating the cpu? From the time it
> >                     appears that the second
> >                     server is waiting for the AST that the first
> >                     server is slow to send. And
> >                     it could be slow because it may be busy flushing
> >                     the data to disk.
> >                     How much memory/cpu do these servers have?
> >
> >                     As far as the second test goes, what is the
> >                     concurrent write performance
> >                     directly to the LV (minus fs).
> >
> >                     On Wed, Aug 22, 2007 at 01:14:02PM +0530, Arnold
> >                     Maderthaner wrote:
> >                     > Hi 2 all !
> >                     >
> >                     > I have problems with concurrent filesystem
> >                     actions on a ocfs2
> >                     > filesystem which is mounted by 2 nodes. OS=RH5ES
> >                     and OCFS2= 1.2.6
> >                     > F.e.: If I have a LV called testlv which is
> >                     mounted on /mnt on both
> >                     > servers and I do a "dd if=/dev/zero
> >                     of=/mnt/test.a bs=1024
> >                     > count=1000000" on server 1 and do at the same
> >                     time a du -hs
> >                     > /mnt/test.a it takes about 5 seconds for du -hs
> >                     to execute:
> >                     > 270M     test.a
> >                     >
> >                     > real    0m3.627s
> >                     > user    0m0.000s
> >                     > sys     0m0.002s
> >                     >
> >                     > or even longer. If i do a du -hs ond a file which
> >                     is not written by
> >                     > another server its fast:
> >                     > 977M     test.b
> >                     >
> >                     > real    0m0.001s
> >                     > user    0m0.001s
> >                     > sys     0m0.000s
> >                     >
> >                     > If I write 2 different files from both servers to
> >                     the same LV I get
> >                     > only 2.8mb/sec.
> >                     > If I write 2 different files from both servers to
> >                     a different LV
> >                     > (testlv,test2lv) then I get about 30mb/sec on
> >                     each write (so 60mb/sec
> >                     > combined which is very good).
> >                     > does anyone has an idea what could be wrong. Are
> >                     there any mount
> >                     > options for ocfs2 which could help me? is this a
> >                     lvm2 or ocfs2 problem
> >                     > (locking,blocking,.... issue) ?
> >                     >
> >                     > Any help will be appriciated.
> >                     >
> >                     > yours
> >                     >
> >                     > Arnold
> >                     > --
> >                     > Arnold Maderthaner
> >                     > J4Care Inc.
> >                     >
> >                     > _______________________________________________
> >                     > Ocfs2-users mailing list
> >                     > Ocfs2-users at oss.oracle.com
> >                     <mailto:Ocfs2-users at oss.oracle.com>
> >                     > http://oss.oracle.com/mailman/listinfo/ocfs2-users
> >                     <http://oss.oracle.com/mailman/listinfo/ocfs2-users>
> >
> >
> >
> >
> >                 --
> >                 Arnold Maderthaner
> >                 J4Care Inc.
> >
> >
> >
> >
> >             --
> >             Arnold Maderthaner
> >             J4Care Inc.
> >
> >
> ------------------------------------------------------------------------
> >             _______________________________________________
> >             Ocfs2-users mailing list
> >             Ocfs2-users at oss.oracle.com <mailto:
> Ocfs2-users at oss.oracle.com>
> >             http://oss.oracle.com/mailman/listinfo/ocfs2-users
> >
> >
> >
> >
> >     --
> >     Arnold Maderthaner
> >     J4Care Inc.
> >
> >
> >
> >
> > --
> > Arnold Maderthaner
> > J4Care Inc.
> > ------------------------------------------------------------------------
> >
> > _______________________________________________
> > Ocfs2-users mailing list
> > Ocfs2-users at oss.oracle.com
> > http://oss.oracle.com/mailman/listinfo/ocfs2-users
>
>


-- 
Arnold Maderthaner
J4Care Inc.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://oss.oracle.com/pipermail/ocfs2-users/attachments/20070824/af55e046/attachment-0001.html


More information about the Ocfs2-users mailing list