[Ocfs2-users] The trouble with the ocfs2 partition continues ...

Juan Pablo Marco Cardona jmarco.cardona at cast-info.es
Thu Oct 29 14:48:04 PDT 2009


Hi Srinivas,
the trouble happens creating files like this:

dd if=/dev/zero of=./file bs=100M count=60

When the file arrives to 5GB the filesystem outputs a error about no 
more free space!

Theorically there is about 63 GB of free space:

df -kh | grep mail
300G  238G   63G  80% /home/mail

This trouble happens creating small files, too

The inodes seems not to be exhausted:

df -i | grep mail
78642183 62365063 16277120   80% /home/mail

I think, there are so few orphaned inodes and maybe it doesn't affect:

echo "ls //orphan_dir:0000" | debugfs.ocfs2 /dev/sda1

debugfs.ocfs2 1.4.3
debugfs:        16              16   1    2  .
        10              16   2    2  ..
        79813023        28   16   1  0000000004c1d99f
        28472617        28   16   1  0000000001b27529
        8438318         28   16   1  000000000080c22e
        80973610        28   16   1  0000000004d38f2a
        213406992       28   16   1  000000000cb85510
        13234609        28   16   1  0000000000c9f1b1
        228704523       28   16   1  000000000da1c10b
        225968869       28   16   1  000000000d7802e5
        256692752       28   16   1  000000000f4cd210
        103224605       28   16   1  000000000627151d
        83675914        28   16   1  0000000004fccb0a
        225968588       28   16   1  000000000d7801cc
        278103558       28   16   1  0000000010938606
        256692760       28   16   1  000000000f4cd218
        13235439        28   16   1  0000000000c9f4ef
        8861985         28   16   1  0000000000873921
        228997031       28   16   1  000000000da637a7
        111786205       28   16   1  0000000006a9b8dd
        24850261        28   16   1  00000000017b2f55
        29889095        28   16   1  0000000001c81247
        311022924       28   16   1  000000001289d54c
        13235309        28   16   1  0000000000c9f46d
        129665645       28   16   1  0000000007ba8a6d
        79605831        28   16   1  0000000004beb047
        28104589        28   16   1  0000000001acd78d
        294769884       28   16   1  000000001191d4dc
        253519012       28   16   1  000000000f1c64a4
        80973232        28   16   1  0000000004d38db0
        13234376        28   16   1  0000000000c9f0c8
        312527073       28   16   1  0000000012a0c8e1
        25863407        28   16   1  00000000018aa4ef
        305612210       28   16   1  00000000123745b2
        226494741       28   16   1  000000000d800915
        228705439       28   16   1  000000000da1c49f
        79604433        40   16   1  0000000004beaad1
        80973605        28   16   1  0000000004d38f25
        226494890       28   16   1  000000000d8009aa
        80973659        28   16   1  0000000004d38f5b
        13236194        56   16   1  0000000000c9f7e2
        80973320        56   16   1  0000000004d38e08
        46226861        28   16   1  0000000002c15dad
        228995934       28   16   1  000000000da6335e
        294770302       28   16   1  000000001191d67e
        256692906       28   16   1  000000000f4cd2aa
        255321636       56   16   1  000000000f37e624
        80974273        28   16   1  0000000004d391c1
        209062548       56   16   1  000000000c760a94
        46227042        168  16   1  0000000002c15e62
        108197615       28   16   1  000000000672f6ef
        13236185        56   16   1  0000000000c9f7d9
        294768980       112  16   1  000000001191d154
        294768983       212  16   1  000000001191d157


I will show you some info about the partition:

Block size:

tunefs.ocfs2 -q -Q "BS=%5B\n" /dev/sda1
BS= 1024

Cluster size:

tunefs.ocfs2 -q -Q "CS=%5T\n" /dev/sda1
CS= 4096

Number Cluster Nodes:

tunefs.ocfs2 -q -Q "CN=%5N\n" /dev/sda1
CN=    5

I think is better to give you more information about this partition.

In this 300 GB ocfs2 partition are the maildir's of about 3K users with 
a lot of small files ( ~ 62 M used inodes).
In the near future we are planning to mount a 2-mail-cluster system, 
because this volume is in a SAN.

Perhaps the cluster and node size are to much big? Maybe the trouble is 
fragmentation?
What are the best ocfs2 options to format this mail volume?

Thanks in advanced.

Regards,
Pablo




Srinivas Eeda escribió:
> Do you run into ENOSPC creating new files or extending existing files? 
> What is the cluster size? Don't think this may be the issue, but any 
> files under orphan directory? run (echo "ls //orphan_dir:000X" | 
> debugfs.ocfs2 <device>) to check if there are any.
>
>
> Juan Pablo Marco Cardona wrote:
>> Hi Sunil,
>> we still have problems with the ocfs2 mail partition. The old problem 
>> with the free space still continues.
>> Theorically we hace 62 GB free space, but we only can use about 5 GB!
>>
>> df -kh | grep mail
>> 300G  238G   63G  80% /home/mail
>>
>> Also, yesterday we have a power outage that force a fsck.ocfs2 of the 
>> partition.
>>
>> We have upgraded to ocfs2 1.4.4 and ocfs2 1.4.3 tools and the kernel 
>> version is the same:
>>
>> rpm -qa | grep ocfs2
>> ocfs2-2.6.18-128.2.1.el5-1.4.4-1.el5
>> ocfs2-tools-1.4.3-1.el5
>> ocfs2console-1.4.3-1.el5
>>
>> modinfo ocfs2 | grep ^version
>> version:        1.4.4
>>
>> uname -r
>> 2.6.18-128.2.1.el5
>>
>> What we can make? Maybe make a fsck.ocfs2 or mkfs.ocfs2 with the new 
>> tools version (1.4.3) ??
>>
>> Thanks in advance.
>>
>> Regards,
>> Pablo
>>
>>
>>
>>
>>
>>  
>>
>>
>>
>> -- 
>>
>> ------------------------------------------------------------------------
>>
>> _______________________________________________
>> Ocfs2-users mailing list
>> Ocfs2-users at oss.oracle.com
>> http://oss.oracle.com/mailman/listinfo/ocfs2-users

-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://oss.oracle.com/pipermail/ocfs2-users/attachments/20091029/0e455ddb/attachment.html 


More information about the Ocfs2-users mailing list