Dear all,
We envolve in a same situation as problem discussed here:
http://lists.lustre.org/pipermail/lustre-discuss/2009-January/009512.html
One OST is set as read only the first time after it is remounted after a server
crash.
Apr 16 17:40:31 boss27 kernel: LDISKFS-fs error (device sdd1):
ldiskfs_mb_check_ondisk_bitmap: on-disk bitmap for group 32486corrupted: 8664
blocks free in bitmap, 13310 - in gd
Apr 16 17:40:31 boss27 kernel:
Apr 16 17:40:31 boss27 kernel: Remounting filesystem read-only
Apr 16 17:40:31 boss27 kernel: LDISKFS-fs error (device sdd1):
ldiskfs_mb_check_ondisk_bitmap: on-disk bitmap for group 32486corrupted: 8664
blocks free in bitmap, 13310 - in gd
Apr 16 17:40:31 boss27 kernel:
Apr 16 17:40:31 boss27 kernel: LDISKFS-fs error (device sdd1):
ldiskfs_mb_check_ondisk_bitmap: on-disk bitmap for group 32486corrupted: 8664
blocks free in bitmap, 13310 - in gd
Apr 16 17:40:31 boss27 kernel:
Apr 16 17:40:31 boss27 kernel: LDISKFS-fs error (device sdd1):
ldiskfs_mb_check_ondisk_bitmap: on-disk bitmap for group 32486corrupted: 8664
blocks free in bitmap, 13310 - in gd
Apr 16 17:40:31 boss27 kernel:
Apr 16 17:40:31 boss27 kernel: LDISKFS-fs error (device sdd1):
ldiskfs_mb_check_ondisk_bitmap: on-disk bitmap for group 32486corrupted: 8664
blocks free in bitmap, 13310 - in gd
Apr 16 17:40:31 boss27 kernel:
Apr 16 17:40:31 boss27 kernel: LustreError:
6158:0:(fsfilt-ldiskfs.c:1288:fsfilt_ldiskfs_write_record()) can''t
start transaction for 37 blocks (128 bytes)
Apr 16 17:40:31 boss27 kernel: LustreError:
6240:0:(fsfilt-ldiskfs.c:1288:fsfilt_ldiskfs_write_record()) can''t
start transaction for 37 blocks (128 bytes)
Apr 16 17:40:31 boss27 kernel: LustreError:
6166:0:(fsfilt-ldiskfs.c:470:fsfilt_ldiskfs_brw_start()) can''t get
handle for 555 credits: rc = -30
@
This OST is unwritable since last Friday, and its disk usage is quite different
from neigbour OSTs.
[root at boss27 ~]# df -h
Filesystem Size Used Avail Use% Mounted on
/dev/sda1 4.1T 3.8T 105G 98% /lustre/ost1
/dev/sda2 4.1T 3.7T 158G 96% /lustre/ost2
/dev/sdb1 4.1T 3.8T 124G 97% /lustre/ost3
/dev/sdb2 4.1T 3.8T 84G 98% /lustre/ost4
/dev/sdc1 4.1T 3.7T 128G 97% /lustre/ost5
/dev/sdc2 4.1T 3.7T 131G 97% /lustre/ost6
/dev/sdd1 4.1T 3.3T 591G 85% /lustre/ost7
/dev/sdd2 4.1T 3.8T 52G 99% /lustre/ost8
Is is prossible to fix this problem without lfsck the whole file system? Our
system is about 500TB(93% full). We are running Lustre 1.8.1.1. File stripe=1
Best Regards
Lu Wang
--------------------------------------------------------------
Computing Center
IHEP Office: Computing Center,123
19B Yuquan Road Tel: (+86) 10 88236012-607
P.O. Box 918-7 Fax: (+86) 10 8823 6839
Beijing 100049,China Email: Lu.Wang at ihep.ac.cn
--------------------------------------------------------------