[Ocfs2-users] 10 Node OCFS2 Cluster - Performance

Sunil Mushran sunil.mushran at oracle.com
Tue Sep 15 10:58:13 PDT 2009


Is the o2cb interconnect and iscsi sharing the same network channel?

Laurence Mayer wrote:
> *1 x Node*:
> root at n1 <mailto:root at n1>:~# dd if=/dev/sdc1 of=/dev/null bs=1M 
> count=1000 skip=2000
> 1000+0 records in
> 1000+0 records out
> 1048576000 bytes (1.0 GB) copied, 10.9246 s, 96.0 MB/s
> *2 x Nodes*
> root at n1:/cfs1/laurence <mailto:root at n1:/cfs1/laurence># cat run.sh.e7470.1
> 1000+0 records in
> 1000+0 records out
> 1048576000 bytes (1.0 GB) copied, 18.6313 s, 56.3 MB/s
> root at n1:/cfs1/laurence <mailto:root at n1:/cfs1/laurence># cat run.sh.e7470.2
> 1000+0 records in
> 1000+0 records out
> 1048576000 bytes (1.0 GB) copied, 19.0982 s, 54.9 MB/s
> real 0m21.557s
> user 0m0.010s
> sys 0m0.000s
> *5 x Nodes*
> run.sh.e7471.1:1048576000 bytes (1.0 GB) copied, 45.7561 s, 22.9 MB/s
> run.sh.e7471.2:1048576000 bytes (1.0 GB) copied, 43.3075 s, 24.2 MB/s
> run.sh.e7471.3:1048576000 bytes (1.0 GB) copied, 38.9945 s, 26.9 MB/s
> run.sh.e7471.4:1048576000 bytes (1.0 GB) copied, 43.535 s, 24.1 MB/s
> run.sh.e7471.5:1048576000 bytes (1.0 GB) copied, 41.4462 s, 25.3 MB/s
> real 0m49.552s
> user 0m0.000s
> sys 0m0.010s
> *8 x Nodes:*
> run.sh.e7472.1:1048576000 bytes (1.0 GB) copied, 60.7164 s, 17.3 MB/s
> run.sh.e7472.2:1048576000 bytes (1.0 GB) copied, 50.3527 s, 20.8 MB/s
> run.sh.e7472.3:1048576000 bytes (1.0 GB) copied, 57.4285 s, 18.3 MB/s
> run.sh.e7472.4:1048576000 bytes (1.0 GB) copied, 47.4362 s, 22.1 MB/s
> run.sh.e7472.5:1048576000 bytes (1.0 GB) copied, 61.4835 s, 17.1 MB/s
> run.sh.e7472.6:1048576000 bytes (1.0 GB) copied, 48.5347 s, 21.6 MB/s
> run.sh.e7472.7:1048576000 bytes (1.0 GB) copied, 63.9391 s, 16.4 MB/s
> run.sh.e7472.8:1048576000 bytes (1.0 GB) copied, 60.6223 s, 17.3 MB/s
> real 1m7.497s
> user 0m0.010s
> sys 0m0.010s
> *10 x Nodes:*
> run.sh.e7473.1:1048576000 bytes (1.0 GB) copied, 58.4126 s, 18.0 MB/s
> run.sh.e7473.10:1048576000 bytes (1.0 GB) copied, 50.982 s, 20.6 MB/s
> run.sh.e7473.2:1048576000 bytes (1.0 GB) copied, 53.1949 s, 19.7 MB/s
> run.sh.e7473.3:1048576000 bytes (1.0 GB) copied, 48.3755 s, 21.7 MB/s
> run.sh.e7473.4:1048576000 bytes (1.0 GB) copied, 60.8544 s, 17.2 MB/s
> run.sh.e7473.5:1048576000 bytes (1.0 GB) copied, 59.9801 s, 17.5 MB/s
> run.sh.e7473.6:1048576000 bytes (1.0 GB) copied, 61.6221 s, 17.0 MB/s
> run.sh.e7473.7:1048576000 bytes (1.0 GB) copied, 59.2011 s, 17.7 MB/s
> run.sh.e7473.8:1048576000 bytes (1.0 GB) copied, 56.3118 s, 18.6 MB/s
> run.sh.e7473.9:1048576000 bytes (1.0 GB) copied, 54.2202 s, 19.3 MB/s
> real 1m6.979s
> user 0m0.010s
> sys 0m0.010s
> Do you think the hardware cannot handle the load?
>
>  
> On Tue, Sep 15, 2009 at 7:53 PM, Sunil Mushran 
> <sunil.mushran at oracle.com <mailto:sunil.mushran at oracle.com>> wrote:
>
>     All clusters are running release tests. So not at the moment.
>
>     But you can see if your hardware is limiting you.
>
>     # time dd if=/dev/sdX1 of=/dev/null bs=1M count=1000 skip=2000
>
>     Run this on one node, then two nodes concurrently, 5 nodes, 10 nodes.
>     The idea is to see whether you see any drop off in read performance
>     when multiple nodes are hitting the iscsi io stack.
>
>     # echo 3 > /proc/sys/vm/drop_caches
>     Do remember to clear the caches between runs.
>
>     Sunil
>
>
>     Laurence Mayer wrote:
>
>         Hi Sunil
>         I am running iostat on only one of the nodes, so the results
>         you see is only from a single node.
>         However I am running this concurrently on the 10 nodes,
>         resulting in a total of 2Gig being written, so yes on this node
>         it took 8 secs to write 205Megs.
>
>         My latest results (using sync after the dd) show that when
>         running on the 10 nodes concurrently it take 37secs
>         to write the 10 x 205Meg files (2Gig),
>         Here are the results from ALL the nodes:
>         run.sh.e7212.1:204800000 bytes (205 MB) copied, 17.9657 s,
>         11.4 MB/s
>         run.sh.e7212.10:204800000 bytes (205 MB) copied, 30.1489 s,
>         6.8 MB/s
>         run.sh.e7212.2:204800000 bytes (205 MB) copied, 16.4605 s,
>         12.4 MB/s
>         run.sh.e7212.3:204800000 bytes (205 MB) copied, 18.1461 s,
>         11.3 MB/s
>         run.sh.e7212.4:204800000 bytes (205 MB) copied, 20.9716 s, 9.8
>         MB/s
>         run.sh.e7212.5:204800000 bytes (205 MB) copied, 22.6265 s, 9.1
>         MB/s
>         run.sh.e7212.6:204800000 bytes (205 MB) copied, 12.9318 s,
>         15.8 MB/s
>         run.sh.e7212.7:204800000 bytes (205 MB) copied, 15.1739 s,
>         13.5 MB/s
>         run.sh.e7212.8:204800000 bytes (205 MB) copied, 13.8953 s,
>         14.7 MB/s
>         run.sh.e7212.9:204800000 bytes (205 MB) copied, 29.5445 s, 6.9
>         MB/s
>
>         real    0m37.920s
>         user    0m0.000s
>         sys     0m0.030s
>
>         (This averages 11.17MB/sec per node, which seems very low.)
>
>         compared to 23.5secs when writing 2Gig from a single node.
>
>         root at n2:# time (dd if=/dev/zero of=txt bs=2048000 count=1000;
>         sync)
>         1000+0 records in
>         1000+0 records out
>         2048000000 bytes (2.0 GB) copied, 16.1369 s, 127 MB/s
>
>         real    0m23.495s
>         user    0m0.000s
>         sys     0m15.180s
>
>
>         Sunil, do you have any way to run the same test (10 x 200Megs)
>         concurrently on 10 or more nodes to compare results?
>
>         Thanks again
>
>         Laurence
>
>
>         Sunil Mushran wrote:
>
>             Always cc ocfs2-users.
>
>             Strange. The ocfs2 numbers look incomplete. It shows only
>             200MB written.
>             You said it was taking 16 secs. Yet the iostat numbers are
>             for 8 secs only.
>
>             The xfs numbers look complete. Shows 90+ MB/s.
>
>             On my iscsi setup (netapp backend, gige, node with single
>             cpu box and
>             512M RAM), I get 85MB/s.
>
>             # time (dd if=/dev/zero of=/mnt/boq7 count=2000 bs=1M ;
>             sync ;)
>             sync
>             2000+0 records in
>             2000+0 records out
>             2097152000 bytes (2.1 GB) copied, 24.4168 seconds, 85.9 MB/s
>
>             real   0m24.515s
>             user   0m0.035s
>             sys    0m14.967s
>
>             This is with data=writeback.
>
>             The 2.2 secs is probably because of delayed allocation.
>             Since your box has
>             enough memory, xfs can cache all the writes and return to
>             the user. Its
>             writeback then flushes the data in the background. The
>             iostat/vmstat
>             numbers should show similar writeback numbers.
>
>             Sunil
>
>             Laurence Mayer wrote:
>
>                  iostat from cfs volume
>                  avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>                           0.00    0.00    1.77    2.28    0.00   95.95
>
>                 Device:         rrqm/s   wrqm/s     r/s     w/s  
>                 rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
>                 sdc               0.00     4.00    2.00    4.00  
>                  16.00    64.00    13.33     0.12   15.00  15.00   9.00
>
>                 avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>                           0.00    0.00    6.90    7.14    0.00   85.96
>
>                 Device:         rrqm/s   wrqm/s     r/s     w/s  
>                 rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
>                 sdc               0.00    16.00    9.00   40.00  
>                  75.00   441.00    10.53     0.43    9.39   6.73  33.00
>
>                 avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>                           0.00    0.00    7.67    7.18    0.00   85.15
>
>                 Device:         rrqm/s   wrqm/s     r/s     w/s  
>                 rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
>                 sdc               0.00    20.00   11.00   47.00  
>                  88.00   536.00    10.76     0.36    6.21   4.48  26.00
>
>                 avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>                           0.00    0.00    5.65   10.07    0.00   84.28
>
>                 Device:         rrqm/s   wrqm/s     r/s     w/s  
>                 rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
>                 sdc               0.00    16.00    9.00   37.00  
>                  75.00   417.00    10.70     0.55   11.96   8.48  39.00
>
>                 avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>                           0.25    0.00   12.69   31.22    0.00   55.84
>
>                 Device:         rrqm/s   wrqm/s     r/s     w/s  
>                 rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
>                 sdc               0.00 40324.00    2.00  181.00  
>                  16.00 174648.00   954.45    94.58  364.86   4.81  88.00
>
>                 avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>                           0.00    0.00   13.35   14.14    0.00   72.51
>
>                 Device:         rrqm/s   wrqm/s     r/s     w/s  
>                 rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
>                 sdc               0.00  9281.00    1.00  228.00  
>                  11.00 224441.00   980.14   100.93  559.17   4.37 100.00
>
>                 avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>                           0.00    0.00    0.25    0.50    0.00   99.25
>
>                 Device:         rrqm/s   wrqm/s     r/s     w/s  
>                 rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
>                 sdc               0.00     0.00    0.00    3.00    
>                 0.00  1040.00   346.67     0.03  240.00   6.67   2.00
>
>                 avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>                           0.00    0.00    0.00    0.00    0.00  100.00
>
>                 Device:         rrqm/s   wrqm/s     r/s     w/s  
>                 rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
>                 sdc               0.00     0.00    1.00    1.00  
>                  11.00     1.00     6.00     0.04   20.00  20.00   4.00
>
>                 vmstat from cfs volume:
>                 procs -----------memory---------- ---swap--
>                 -----io---- -system-- ----cpu----
>                  r  b   swpd   free   buff  cache   si   so    bi  
>                  bo   in   cs us sy id wa
>                  0  0      0 447656 279416 15254408    0    0     0  
>                   0   39  350  0  0 100  0
>                  0  0      0 447656 279416 15254408    0    0     5  
>                  21   61  358  0  0 100  0
>                  0  0      0 447656 279416 15254408    0    0     0  
>                   0   49  369  0  0 100  0
>                  0  0      0 447656 279416 15254408    0    0     6  
>                   0   28  318  0  0 100  0
>                  0  0      0 447656 279416 15254408    0    0     0  
>                   0   26  321  0  0 100  0
>                  0  0      0 447656 279416 15254408    0    0     5  
>                   1   45  339  0  0 100  0
>                  0  0      0 447656 279416 15254412    0    0     0  
>                   0    8  283  0  0 100  0
>                  0  1      0 439472 279424 15262604    0    0    14  
>                  80   93  379  0  1 90  9
>                  0  0      0 439472 279424 15262604    0    0     0  
>                   4   43  338  0  0 97  2
>                  0  0      0 382312 279456 15319964    0    0    37  
>                 209  208  562  0  7 85  8
>                  0  0      0 324524 279500 15377292    0    0    44  
>                 264  250  647  0  7 86  7
>                  0  0      0 266864 279532 15434636    0    0    38  
>                 208  213  548  0  7 83 10
>                  0  3      0 250072 279544 15450584    0    0    44
>                 124832 13558 2038  0 11 62 27
>                  0  1      0 250948 279564 15450584    0    0     5
>                 75341 19596 2735  0 13 71 16
>                  0  0      0 252808 279564 15450548    0    0     0  
>                  52 2777  849  0  2 95  3
>                  0  0      0 252808 279564 15450548    0    0     6  
>                   0   21  310  0  0 100  0
>                  0  0      0 252808 279564 15450548    0    0     0  
>                   0   15  298  0  0 100  0
>                  0  0      0 253012 279564 15450548    0    0     5  
>                   1   29  310  0  0 100  0
>                  0  0      0 253048 279564 15450552    0    0     0  
>                   0   19  290  0  0 100  0
>                  0  0      0 253048 279564 15450552    0    0     6  
>                   0   26  305  0  0 100  0
>                  1  0      0 253172 279564 15450552    0    0     0  
>                  60   28  326  0  0 100  0
>                   xfs volume:
>                 iostat
>                 Device:         rrqm/s   wrqm/s     r/s     w/s  
>                 rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
>                 sdd               0.00     0.00    4.00    0.00  
>                  40.00     0.00    10.00     0.05   12.00  12.00   4.80
>                 avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>                           0.00    0.00   14.98    0.25    0.00   84.77
>                 Device:         rrqm/s   wrqm/s     r/s     w/s  
>                 rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
>                 sdd               0.00     0.00    3.00    5.00  
>                  24.00  3088.00   389.00     6.54   44.00  17.00  13.60
>                 avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>                           0.00    0.00   10.67   21.86    0.00   67.47
>                 Device:         rrqm/s   wrqm/s     r/s     w/s  
>                 rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
>                 sdd               0.00     1.00    0.00  221.00    
>                 0.00 202936.00   918.26   110.51  398.39   4.52 100.00
>                 avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>                           0.00    0.00    4.92   21.84    0.00   73.23
>                 Device:         rrqm/s   wrqm/s     r/s     w/s  
>                 rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
>                 sdd               0.00     2.00    0.00  232.00    
>                 0.00 209152.00   901.52   110.67  493.50   4.31 100.00
>                 avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>                           0.00    0.00    3.67   22.78    0.00   73.54
>                 Device:         rrqm/s   wrqm/s     r/s     w/s  
>                 rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
>                 sdd               0.00     1.00    0.00  215.00    
>                 0.00 185717.00   863.80   111.37  501.67   4.65 100.00
>                 avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>                           0.12    0.00    6.24   12.61    0.00   81.02
>                 Device:         rrqm/s   wrqm/s     r/s     w/s  
>                 rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
>                 sdd               0.00     1.00    0.00  200.00    
>                 0.00 178456.00   892.28    80.01  541.82   4.88  97.60
>                 avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>                           0.12    0.00    4.61    8.34    0.00   86.92
>                 Device:         rrqm/s   wrqm/s     r/s     w/s  
>                 rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
>                 sdd               0.00     0.00    0.00  179.00    
>                 0.00 183296.00  1024.00   134.56  470.61   5.21  93.20
>                 avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>                           0.00    0.00    4.25    9.96    0.00   85.79
>                 Device:         rrqm/s   wrqm/s     r/s     w/s  
>                 rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
>                 sdd               0.00     0.00    0.00  201.00    
>                 0.00 205824.00  1024.00   142.86  703.92   4.98 100.00
>                    vmstat
>                 procs -----------memory---------- ---swap--
>                 -----io---- -system-- ----cpu----
>                  r  b   swpd   free   buff  cache   si   so    bi  
>                  bo   in   cs us sy id wa
>                  1  0  45396 214592   6332 31771312    0    0   668  
>                 908    3    6  3  2 92  3
>                  0  0  45396 214460   6332 31771336    0    0     0  
>                   0   14 4874  0  0 100  0
>                  2  0  45396 161032   6324 31822524    0    0    20  
>                   0   42 6074  0 13 87  0
>                  5  1  45396 166380   6324 31820072    0    0    12
>                 77948 8166 6416  0 16 77  7
>                  1  2  45396 163176   6324 31824580    0    0    28
>                 102920 24190 6660  0  6 73 21
>                  0  2  45396 163096   6332 31824580    0    0     0
>                 102743 22576 6700  0  5 72 23
>                  0  2  45396 163076   6332 31824580    0    0     0
>                 90400 21831 6500  0  4 76 21
>                  0  1  45396 163012   6332 31824580    0    0     0
>                 114732 19686 5894  0  7 83 10
>                  0  1  45396 162972   6332 31824580    0    0     0
>                 98304 24882 6314  0  4 87  8
>                  0  1  45396 163064   6332 31824580    0    0     0
>                 98304 24118 6285  0  4 84 12
>                  0  1  45396 163096   6340 31824576    0    0     0
>                 114720 24800 6166  0  4 87  9
>                  0  1  45396 162964   6340 31824584    0    0     0
>                 98304 24829 6105  0  3 85 12
>                  0  1  45396 162856   6340 31824584    0    0     0
>                 98304 23506 6402  0  5 83 12
>                  0  1  45396 162888   6340 31824584    0    0     0
>                 114688 24685 7057  0  4 87  9
>                  0  1  45396 162600   6340 31824584    0    0     0
>                 98304 24902 7107  0  4 86 10
>                  0  1  45396 162740   6340 31824584    0    0     0
>                 98304 24906 7019  0  4 91  6
>                  0  1  45396 162616   6348 31824584    0    0     0
>                 114728 24997 7169  0  4 86  9
>                  0  1  45396 162896   6348 31824584    0    0     0
>                 98304 23700 6857  0  4 85 11
>                  0  1  45396 162732   6348 31824584    0    0     0
>                 94512 24468 6995  0  3 89  8
>                  0  1  45396 162836   6348 31824584    0    0     0
>                 81920 19764 6604  0  7 81 11
>                  0  3  45396 162996   6348 31824584    0    0     0
>                 114691 24303 7270  0  4 81 14
>                 procs -----------memory---------- ---swap--
>                 -----io---- -system-- ----cpu----
>                  r  b   swpd   free   buff  cache   si   so    bi  
>                  bo   in   cs us sy id wa
>                  0  1  45396 163160   6356 31824584    0    0     0
>                 98332 22695 7174  0  4 78 18
>                  0  1  45396 162848   6356 31824584    0    0     0
>                 90549 24836 7347  0  4 82 15
>                  1  0  45396 163092   6364 31824580    0    0     0  
>                  37 13990 6216  0  6 83 11
>                  0  0  45396 163272   6364 31824588    0    0     0  
>                 320   65 3817  0  0 100  0
>                  0  0  45396 163272   6364 31824588    0    0     0  
>                   0    8 3694  0  0 100  0
>                  0  0  45396 163272   6364 31824588    0    0     0  
>                   0   25 3833  0  0 100  0
>                  0  0  45396 163272   6364 31824588    0    0     0  
>                   1   13 3690  0  0 100  0
>                    
>                  On Mon, Sep 14, 2009 at 10:15 PM, Sunil Mushran
>                 <sunil.mushran at oracle.com
>                 <mailto:sunil.mushran at oracle.com>
>                 <mailto:sunil.mushran at oracle.com
>                 <mailto:sunil.mushran at oracle.com>>> wrote:
>
>                    Add a sync. Both utils are showing very little io.
>                 And do the same
>                    for runs on both ocfs2 and xfs.
>
>                    # dd if...  ; sync;
>
>                    Laurence Mayer wrote:
>
>                        Here is the output of iostat while running the
>                 test on all the
>                        OCFS volume.
>                         avg-cpu:  %user   %nice %system %iowait
>                  %steal   %idle
>                                  0.23    0.00   15.80    0.45    0.00
>                   83.52
>
>                        Device:         rrqm/s   wrqm/s     r/s     w/s
>                   rsec/s          wsec/s avgrq-sz avgqu-sz   await
>                  svctm  %util
>                        sdc               0.00     4.00    5.00    4.00
>                    43.00           57.00    11.11     0.08    8.89  
>                 8.89   8.00
>
>                        avg-cpu:  %user   %nice %system %iowait  %steal
>                   %idle
>                                  0.28    0.00    4.46    0.00    0.00
>                   95.26
>
>                        Device:         rrqm/s   wrqm/s     r/s     w/s
>                   rsec/s          wsec/s avgrq-sz avgqu-sz   await
>                  svctm  %util
>                        sdc               0.00     0.00    0.00    0.00
>                     0.00            0.00     0.00     0.00    0.00  
>                 0.00   0.00
>
>                        avg-cpu:  %user   %nice %system %iowait  %steal
>                   %idle
>                                  0.25    0.00    0.25    3.23    0.00
>                   96.28
>
>                        Device:         rrqm/s   wrqm/s     r/s     w/s
>                   rsec/s          wsec/s avgrq-sz avgqu-sz   await
>                  svctm  %util
>                        sdc               0.00     7.00    1.00   13.00
>                    11.00          153.00    11.71     0.24   17.14
>                  11.43  16.00
>
>                        avg-cpu:  %user   %nice %system %iowait  %steal
>                   %idle
>                                  0.00    0.00    0.00    0.00    0.00
>                  100.00
>
>                        Device:         rrqm/s   wrqm/s     r/s     w/s
>                   rsec/s          wsec/s avgrq-sz avgqu-sz   await
>                  svctm  %util
>                        sdc               0.00     0.00    0.00    0.00
>                     0.00            0.00     0.00     0.00    0.00  
>                 0.00   0.00
>
>                        avg-cpu:  %user   %nice %system %iowait  %steal
>                   %idle
>                                  0.00    0.00    0.00    0.00    0.00
>                  100.00
>
>                        Device:         rrqm/s   wrqm/s     r/s     w/s
>                   rsec/s          wsec/s avgrq-sz avgqu-sz   await
>                  svctm  %util
>                        sdc               0.00     0.00    1.00    1.00
>                    11.00            1.00     6.00     0.03   15.00
>                  15.00   3.00
>
>                        vmstat:
>                         procs -----------memory---------- ---swap--
>                 -----io----
>                        -system-- ----cpu----
>                         r  b   swpd   free   buff  cache   si   so  
>                  bi    bo   in          cs us sy id wa
>                         0  0      0  54400 279320 15651312    0    0  
>                   9     8    2
>                           4 30  1 69  0
>                         0  0      0  54384 279320 15651316    0    0  
>                   6     0   24
>                         299  0  0 100  0
>                         0  0      0  54384 279320 15651316    0    0  
>                   0     0   92
>                         409  0  0 100  0
>                         2  0      0  54384 279320 15651316    0    0  
>                   5     1   81
>                         386  0  0 100  0
>                         0  0      0  53756 279320 15651352    0    0  
>                   8     0  730
>                        1664  0  1 99  0
>                         0  0      0  53232 279320 15651352    0    0  
>                   6    88  586
>                        1480  0  0 99  0
>                         0  0      0 242848 279320 15458608    0    0  
>                   8     0  348
>                        1149  0  3 97  0
>                         0  0      0 242868 279320 15458608    0    0  
>                   5     1  220
>                         721  0  0 100  0
>                         0  0      0 242868 279320 15458608    0    0  
>                   0     0  201
>                         709  0  0 100  0
>                         0  0      0 243116 279320 15458608    0    0  
>                   6     0  239
>                         775  0  0 100  0
>                         0  0      0 243116 279320 15458608    0    0  
>                   0     0  184
>                         676  0  0 100  0
>                         0  0      0 243116 279336 15458608    0    0  
>                   5    65  236
>                         756  0  0 99  0
>                         0  0      0 243488 279336 15458608    0    0  
>                   0     0  231
>                         791  0  0 100  0
>                         1  0      0 243488 279336 15458608    0    0  
>                   6     0  193
>                         697  0  1 100  0
>                         0  0      0 243488 279336 15458608    0    0  
>                   0     0  221
>                         762  0  0 100  0
>                         0  0      0 243860 279336 15458608    0    0  
>                   9     1  240
>                         793  0  0 100  0
>                         0  0      0 243860 279336 15458608    0    0  
>                   0     0  197
>                         708  0  0 100  0
>                         1  0      0 117384 279348 15585384    0    0  
>                  26    16  124
>                         524  0 15 84  1
>                         0  0      0  53204 279356 15651364    0    0  
>                   0   112  141
>                         432  0  8 91  1
>                         0  0      0  53212 279356 15651320    0    0  
>                   5     1   79
>                         388  0  0 100  0
>                         0  0      0  53212 279356 15651320    0    0  
>                   0    20   30
>                         301  0  0 100  0
>                         Does this give you any clue to the bottle neck?
>                                   On Mon, Sep 14, 2009 at 9:42 PM,
>                 Sunil Mushran
>                        <sunil.mushran at oracle.com
>                 <mailto:sunil.mushran at oracle.com>
>                 <mailto:sunil.mushran at oracle.com
>                 <mailto:sunil.mushran at oracle.com>>
>                        <mailto:sunil.mushran at oracle.com
>                 <mailto:sunil.mushran at oracle.com>
>                        <mailto:sunil.mushran at oracle.com
>                 <mailto:sunil.mushran at oracle.com>>>> wrote:
>
>                           Get some iostat/vmstat numbers.
>                           # iostat -x /dev/sdX 1
>                           # vmstat 1
>
>                           How much memory do the nodes have? If more
>                 than 2G, XFS
>                           is probably leveraging its delayed
>                 allocation feature to
>                        heavily
>                           cache the writes. iostat/vmstat should show
>                 that.
>
>                           Is the timing for the 10 node test cumulative?
>
>                           Laurence Mayer wrote:
>
>                               Hi,
>
>                               I am currently running a 10 Node OCFS2
>                  Cluster (version
>                               1.3.9-0ubuntu1) on Ubuntu Server 8.04
>                 x86_64.
>                               Linux n1 2.6.24-24-server #1 SMP Tue Jul
>                 7 19:39:36 UTC
>                        2009
>                               x86_64 GNU/Linux
>
>                               The Cluster is connected to a 1Tera
>                 iSCSI Device
>                        presented by
>                               an IBM 3300 Storage System, running over
>                 a 1Gig Network.
>                               Mounted on all nodes:  /dev/sdc1 on
>                 /cfs1 type ocfs2
>                               (rw,_netdev,data=writeback,heartbeat=local)
>                               Maximum Nodes: 32
>                               Block Size=4k
>                               Cluster Size=4k
>
>                               My testing shows that to write
>                 simultaneously from the 10
>                               nodes, 10 x 200Meg files (1 file per
>                 node,  total of 2Gig)
>                               takes ~23.54secs.
>                               Reading the files back can take just as
>                 long.
>
>                               Do these numbers sound correct?
>
>                               Doing dd if=/dev/zero of=/cfs1/xxxxx/txt
>                 count=1000
>                        bs=2048000
>                               (2Gig) from a single node takes 16secs.
>
>                               (running the same dd command on an XFS
>                 filesystem
>                        connected to
>                               the same iSCSI Storage takes 2.2secs)
>
>                               Is there any tips & tricks to improve
>                 performance on OCFS2?
>
>                               Thanks in advance
>                               Laurence
>
>                              
>                 _______________________________________________
>                               Ocfs2-users mailing list
>                               Ocfs2-users at oss.oracle.com
>                 <mailto:Ocfs2-users at oss.oracle.com>
>                        <mailto:Ocfs2-users at oss.oracle.com
>                 <mailto:Ocfs2-users at oss.oracle.com>>
>                        <mailto:Ocfs2-users at oss.oracle.com
>                 <mailto:Ocfs2-users at oss.oracle.com>
>                        <mailto:Ocfs2-users at oss.oracle.com
>                 <mailto:Ocfs2-users at oss.oracle.com>>>
>
>                              
>                 http://oss.oracle.com/mailman/listinfo/ocfs2-users
>                            
>
>
>
>




More information about the Ocfs2-users mailing list