Hi Guys,
I have a FreeBSD9.3-stable system with 3x 68GB IBM Disks (Cheeta)
configured wirh gpt and zfs as pool zroot. In the last days one of the
disks was dying and I've disconnected the bad going drive to connect
an other one. The "new" disk is physically identical but replacing the
bad
disk with the new one failed, since the new disk reported only 65 insted of
68GByte free space (other firmware, IBMAS400 in the inquiry string).
So I looked at the bad disks and found one with defective electronics
and one with defective platters and exchanged the controllers (all Disks
are in "real life ST373453LC disks, original with 534 Byte sector size
or so).
I've rebooted the system, set up a camcontrol cmd for changing the sector
size to 512Byte and begun to format the disk with camcontrol format.
While formating for a while the system apniced with an adaptec register
dump and would'nt boot anymore with the infamous "ZFS: i/o error - all
block
copies unavailabe", kernel booted but mountroot failed.
I've found out that I could boot the system when I manually load the zfs.ko
module at the loader prompt.
The filesystem is somewhat broken now, /boot/loader.conf isn't readable and
so are other files, /etc/rc.conf for example, /usr/home is empty.
I've tried to reformat the disk a 2nd time, but it never finished. Tried
the next disk and this worked, tried a replace in the with zpool but I have
an unclear state here:
# zpool status -v
pool: zroot
state: DEGRADED
status: One or more devices has experienced an error resulting in data
corruption. Applications may be affected.
action: Restore the file in question if possible. Otherwise restore the
entire pool from backup.
see: http://illumos.org/msg/ZFS-8000-8A
scan: resilvered 42.4G in 1h25m with 2 errors on Sat Jul 11 12:28:53 2015
config:
NAME STATE READ WRITE CKSUM
zroot DEGRADED 0 0 86
raidz1-0 DEGRADED 0 0 344
gpt/disk0 ONLINE 0 0 0
gpt/disk1 ONLINE 0 0 0
replacing-2 DEGRADED 0 0 0
6329731321389372496 OFFLINE 0 0 0 was
/dev/gpt/disk
2/old
gpt/disk2 ONLINE 0 0 0
errors: Permanent errors have been detected in the following files:
zroot:<0x0>
zroot/var/db:<0x0>
#
I'm unable to remove thedisk with that id 6329731321389372496 and don't
know what replacing-2 could mean at all.
I do have an tar backup from my /home from almost a week ago, but in the
meantime I've done some serious work with Kicad (PCB's) tat I really
don't
want to loose. The Backups from the system itselves are really old...
How can I repair this trouble? Can I get my data back at all?
zpool scrub ran several times, resilvering too. I can't remove that
loader.conf file or others. (cat /etc/rc.conf .. unknown error 122)
I think I need ZFS Guru here...please help..
Regards,
Holm
--
Technik Service u. Handel Tiffe, www.tsht.de, Holm Tiffe,
Freiberger Stra?e 42, 09600 Obersch?na, USt-Id: DE253710583
www.tsht.de, info at tsht.de, Fax +49 3731 74200, Mobil: 0172 8790 741