So... issues with reslivering yet again. This is ~3TB pool. I have one raid-z of 5 500GB disks, and a second pool of 3 300GB disks. One of the 300GB disks failed, so I have replaced the drive. After doing the resliver, it takes approximately 5 minutes for it to complete 68.05% of the reslivering... then it appears to just hang. It''s been this way for 30 hours now. If I do a zpool status, the command does not finish, it just hangs after presenting the scrub: resliver in progress, 68.05% done If I do a zpool iostat, it shows zero disk activity. If I do a zpool iostat -v, the command hangs as well. There''s 0 activity to this pool as I suspended all shares while doing the reslivering in hopes it would speed things up. I''ve seen previous threads about how slow this can be, but this is a bit ridiculous. At this point I''m afraid to stick anymore data into a zpool if one disk failing will take weeks to rebuild. This message posted from opensolaris.org
After messing around... who knows what''s going on with it now. Finally rebooted because I was sick of it hanging. After that, this is what it came back with: root:=> zpool status pool: fserv state: DEGRADED status: One or more devices is currently being resilvered. The pool will continue to function, possibly in a degraded state. action: Wait for the resilver to complete. scrub: resilver in progress, 0.00% done, 87h56m to go config: NAME STATE READ WRITE CKSUM fserv DEGRADED 0 0 0 raidz1 ONLINE 0 0 0 c4t0d0 ONLINE 0 0 0 c4t1d0 ONLINE 0 0 0 c4t2d0 ONLINE 0 0 0 c4t3d0 ONLINE 0 0 0 c4t4d0 ONLINE 0 0 0 raidz1 DEGRADED 0 0 0 c4t6d0 ONLINE 0 0 0 c4t7d0 ONLINE 0 0 0 replacing DEGRADED 0 0 0 12544952246745011915 FAULTED 0 0 0 was /dev/dsk/c4t5d0s0/old c4t5d0 ONLINE 0 0 0 root:=> zpool iostat -v capacity operations bandwidth pool used avail read write read write -------------------------- ----- ----- ----- ----- ----- ----- fserv 990G 2.11T 397 25 25.3M 101K raidz1 866G 1.42T 201 1 804K 5.29K c4t0d0 - - 133 1 533K 9.86K c4t1d0 - - 133 1 544K 9.89K c4t2d0 - - 133 1 541K 9.88K c4t3d0 - - 133 1 535K 9.82K c4t4d0 - - 132 1 525K 9.86K raidz1 124G 708G 196 23 24.5M 95.6K c4t6d0 - - 102 31 12.3M 84.1K c4t7d0 - - 102 31 12.3M 83.8K replacing - - 0 42 0 1.48M 12544952246745011915 - - 0 0 1.67K 0 c4t5d0 - - 0 25 2.04K 1.50M -------------------------- ----- ----- ----- ----- ----- ----- This message posted from opensolaris.org
That locked up pretty quickly as well, one more reboot and this is what I''m seeing now: root:=> zpool status pool: fserv state: DEGRADED status: One or more devices is currently being resilvered. The pool will continue to function, possibly in a degraded state. action: Wait for the resilver to complete. scrub: resilver in progress, 1.81% done, 0h19m to go config: NAME STATE READ WRITE CKSUM fserv DEGRADED 0 0 0 raidz1 ONLINE 0 0 0 c4t0d0 ONLINE 0 0 0 c4t1d0 ONLINE 0 0 0 c4t2d0 ONLINE 0 0 0 c4t3d0 ONLINE 0 0 0 c4t4d0 ONLINE 0 0 0 raidz1 DEGRADED 0 0 0 c4t6d0 ONLINE 0 0 0 c4t7d0 ONLINE 0 0 0 replacing DEGRADED 0 0 0 c4t5d0s0/o FAULTED 0 0 0 corrupted data c4t5d0 ONLINE 0 0 0 errors: No known data errors The "corrupted data" seems a bit scary to me... This message posted from opensolaris.org