[Ocfs2-devel] 60% full and writes fail..
Sunil Mushran
Sunil.Mushran at oracle.com
Thu Sep 6 10:33:03 PDT 2007
http://oss.oracle.com/~smushran/.debug/scripts/stat_sysdir.sh
File a bugzilla and attach the output of the above script. It will
dump the superblock and the system directory.
paul vogel wrote:
>
> I have a setup with lot's of small files (Maildir), in 4 different
> volumes and for some
> reason the volumes are full when they reach 60% usage (as reported by
> df ).
>
> This was ofcourse a bit of a supprise for me .. lots of failed writes,
> bounced messages
> and very angry customers.
>
> Has anybody on this list seen this before (not the angry customers ;-) ?
>
> Regards,
>
> =paulv
>
>
>
>
> # echo "ls -l //" | debugfs.ocfs2 /dev/drbd6
> debugfs.ocfs2 1.2.1
> debugfs: 34 drwxr-xr-x 6 0 0
> 4096 16-Aug-2007 22:01 .
> 34 drwxr-xr-x 6 0 0 4096
> 16-Aug-2007 22:01 ..
> 35 -rw-r--r-- 1 0 0 0
> 16-Aug-2007 22:01 bad_blocks
> 36 -rw-r--r-- 1 0 0 851968
> 16-Aug-2007 22:01 global_inode_alloc
> 37 -rw-r--r-- 1 0 0 65536
> 16-Aug-2007 22:01 slot_map
> 38 -rw-r--r-- 1 0 0 1048576
> 16-Aug-2007 22:01 heartbeat
> 39 -rw-r--r-- 1 0 0 249999654912
> 16-Aug-2007 22:01 global_bitmap
> 40 drwxr-xr-x 2 0 0 4096
> 6-Sep-2007 19:13 orphan_dir:0000
> 41 drwxr-xr-x 2 0 0 16384
> 6-Sep-2007 09:34 orphan_dir:0001
> 42 drwxr-xr-x 2 0 0 4096
> 16-Aug-2007 22:01 orphan_dir:0002
> 43 drwxr-xr-x 2 0 0 4096
> 16-Aug-2007 22:01 orphan_dir:0003
> 44 -rw-r--r-- 1 0 0 4194304
> 16-Aug-2007 22:01 extent_alloc:0000
> 45 -rw-r--r-- 1 0 0 0
> 16-Aug-2007 22:01 extent_alloc:0001
> 46 -rw-r--r-- 1 0 0 0
> 16-Aug-2007 22:01 extent_alloc:0002
> 47 -rw-r--r-- 1 0 0 0
> 16-Aug-2007 22:01 extent_alloc:0003
> 48 -rw-r--r-- 1 0 0 142606336
> 16-Aug-2007 22:01 inode_alloc:0000
> 49 -rw-r--r-- 1 0 0 6966738944
> 16-Aug-2007 22:01 inode_alloc:0001
> 50 -rw-r--r-- 1 0 0 0
> 16-Aug-2007 22:01 inode_alloc:0002
> 51 -rw-r--r-- 1 0 0 0
> 16-Aug-2007 22:01 inode_alloc:0003
> 52 -rw-r--r-- 1 0 0 268435456
> 16-Aug-2007 22:01 journal:0000
> 53 -rw-r--r-- 1 0 0 268435456
> 16-Aug-2007 22:02 journal:0001
> 54 -rw-r--r-- 1 0 0 268435456
> 16-Aug-2007 22:02 journal:0002
> 55 -rw-r--r-- 1 0 0 268435456
> 16-Aug-2007 22:03 journal:0003
> 56 -rw-r--r-- 1 0 0 0
> 16-Aug-2007 22:01 local_alloc:0000
> 57 -rw-r--r-- 1 0 0 0
> 16-Aug-2007 22:01 local_alloc:0001
> 58 -rw-r--r-- 1 0 0 0
> 16-Aug-2007 22:01 local_alloc:0002
> 59 -rw-r--r-- 1 0 0 0
> 16-Aug-2007 22:01 local_alloc:0003
> 60 -rw-r--r-- 1 0 0 0
> 16-Aug-2007 22:01 truncate_log:0000
> 61 -rw-r--r-- 1 0 0 0
> 16-Aug-2007 22:01 truncate_log:0001
> 62 -rw-r--r-- 1 0 0 0
> 16-Aug-2007 22:01 truncate_log:0002
> 63 -rw-r--r-- 1 0 0 0
> 16-Aug-2007 22:01 truncate_log:0003
> debugfs:
>
>
>
> My custer.conf :
>
>
> cluster:
> node_count = 2
> name = pool_5
> node:
> ip_port = 1015
> ip_address = 10.17.178.132
> number = 1
> name = jaguar
> cluster = pool_5
> node:
> ip_port = 1015
> ip_address = 10.17.178.133
> number = 2
> name = joon
> cluster = pool_5
>
>
> cluster:
> node_count = 2
> name = pool_6
> node:
> ip_port = 1016
> ip_address = 10.17.178.132
> number = 3
> name = jaguar
> cluster = pool_6
> node:
> ip_port = 1016
> ip_address = 10.17.178.133
> number = 4
> name = joon
> cluster = pool_6
>
>
> cluster:
> node_count = 2
> name = pool_7
> node:
> ip_port = 1017
> ip_address = 10.17.178.132
> number = 5
> name = jaguar
> cluster = pool_7
> node:
> ip_port = 1017
> ip_address = 10.17.178.133
> number = 6
> name = joon
> cluster = pool_7
>
>
> cluster:
> node_count = 2
> name = pool_8
> node:
> ip_port = 1018
> ip_address = 10.17.178.132
> number = 7
> name = jaguar
> cluster = pool_8
> node:
> ip_port = 1018
> ip_address = 10.17.178.133
> number = 8
> name = joon
> cluster = pool_8
>
>
>
>
> _______________________________________________
> Ocfs2-devel mailing list
> Ocfs2-devel at oss.oracle.com
> http://oss.oracle.com/mailman/listinfo/ocfs2-devel
More information about the Ocfs2-devel
mailing list