I physically removed a disk (c3t8d0 used by ZFS ''pool01'') from a 3310 JBOD connected to a V210 running s10u3 (11/06) and ''zpool status'' reported this: # zpool status pool: pool01 state: DEGRADED status: One or more devices could not be opened. Sufficient replicas exist for the pool to continue functioning in a degraded state. action: Attach the missing device and online it using ''zpool online''. see: http://www.sun.com/msg/ZFS-8000-D3 scrub: resilver completed with 0 errors on Mon Jan 8 15:56:20 2007 config: NAME STATE READ WRITE CKSUM pool01 DEGRADED 0 0 0 mirror DEGRADED 0 0 0 c2t4d0 ONLINE 0 0 0 c3t8d0 UNAVAIL 0 0 0 cannot open mirror ONLINE 0 0 0 c2t5d0 ONLINE 0 0 0 c3t9d0 ONLINE 0 0 0 spares c2t8d0 AVAIL c3t10d0 AVAIL Why doesn''t ZFS automatically use one of the hot spares? Is this expected behavior or a bug? Rob This message posted from opensolaris.org
Rob, It (hot-spare) should have kicked in. How long did you wait for it ? Was there any IO happening on the pool ? Try doing some IO to the disk and see if it kicks in. Also, another point to note is the size of the the hotspares. Please ensure that the hot-spares are of the same size as the mirrors. I think the hot-spares don''t kickin if there is a size mismatch. If none of the above works then we will have a take a closer look at the details :-) Regards, Sanjeev. Rob wrote:>I physically removed a disk (c3t8d0 used by ZFS ''pool01'') from a 3310 JBOD connected to a V210 running s10u3 (11/06) and ''zpool status'' reported this: > ># zpool status > pool: pool01 > state: DEGRADED >status: One or more devices could not be opened. Sufficient replicas exist for > the pool to continue functioning in a degraded state. >action: Attach the missing device and online it using ''zpool online''. > see: http://www.sun.com/msg/ZFS-8000-D3 > scrub: resilver completed with 0 errors on Mon Jan 8 15:56:20 2007 >config: > > NAME STATE READ WRITE CKSUM > pool01 DEGRADED 0 0 0 > mirror DEGRADED 0 0 0 > c2t4d0 ONLINE 0 0 0 > c3t8d0 UNAVAIL 0 0 0 cannot open > mirror ONLINE 0 0 0 > c2t5d0 ONLINE 0 0 0 > c3t9d0 ONLINE 0 0 0 > spares > c2t8d0 AVAIL > c3t10d0 AVAIL > >Why doesn''t ZFS automatically use one of the hot spares? Is this expected behavior or a bug? > >Rob > > >This message posted from opensolaris.org >_______________________________________________ >zfs-discuss mailing list >zfs-discuss at opensolaris.org >http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > >-- Solaris Revenue Products Engineering, India Engineering Center, Sun Microsystems India Pvt Ltd. Tel: x27521 +91 80 669 27521
Thanks for the response, please see my comments below . . .> It (hot-spare) should have kicked in. How long did you wait for it ?10 to 15 minutes.> Was there any IO happening on the pool ?Not when I noticed the problem. I''m evaluating ZFS+NFS as a replacement for our NetApp filers and didn''t want to continue testing until the disk issue was resolved.> Try doing some IO to the disk and see if it kicks in.Before noticing the drive was unavailable, I had performed several bringovers to the FS. This did nothing to trigger the hot spare (assuming the drive was already unavailable). The box has been reinstalled with s10u3 since posting my original message, so I''ll try to reproduce the problem tomorrow (Wednesday).> Also, another point to note is the size of the the hotspares. Please > ensure that the hot-spares are of the same size as the mirrors. I > think the hot-spares don''t kickin if there is a size mismatch.Since I''ve reinstalled, the controller numbers are different, but the disks remain the same: AVAILABLE DISK SELECTIONS: 0. c0t0d0 <SUN72G cyl 14087 alt 2 hd 24 sec 424> /pci at 1c,600000/scsi at 2/sd at 0,0 1. c1t4d0 <SEAGATE-ST373307LSUN72G-0207-68.37GB> /pci at 1d,700000/scsi at 1/sd at 4,0 2. c1t5d0 <SEAGATE-ST373307LSUN72G-0207-68.37GB> /pci at 1d,700000/scsi at 1/sd at 5,0 3. c1t8d0 <SEAGATE-ST373307LSUN72G-0207-68.37GB> /pci at 1d,700000/scsi at 1/sd at 8,0 4. c1t9d0 <SUN72G cyl 14087 alt 2 hd 24 sec 424> /pci at 1d,700000/scsi at 1/sd at 9,0 5. c1t10d0 <SUN72G cyl 14087 alt 2 hd 24 sec 424> /pci at 1d,700000/scsi at 1/sd at a,0 6. c1t11d0 <SUN72G cyl 14087 alt 2 hd 24 sec 424> /pci at 1d,700000/scsi at 1/sd at b,0 7. c2t8d0 <SEAGATE-ST373307LSUN72G-0207-68.37GB> /pci at 1d,700000/scsi at 1,1/sd at 8,0 8. c2t9d0 <SEAGATE-ST373307LSUN72G-0307-68.37GB> /pci at 1d,700000/scsi at 1,1/sd at 9,0 9. c2t10d0 <SEAGATE-ST373307LSUN72G-0307-68.37GB> /pci at 1d,700000/scsi at 1,1/sd at a,0 10. c2t11d0 <SUN72G cyl 14087 alt 2 hd 24 sec 424> /pci at 1d,700000/scsi at 1,1/sd at b,0 11. c2t12d0 <SUN72G cyl 14087 alt 2 hd 24 sec 424> /pci at 1d,700000/scsi at 1,1/sd at c,0 12. c2t13d0 <SUN72G cyl 14087 alt 2 hd 24 sec 424> /pci at 1d,700000/scsi at 1,1/sd at d,0 The spares are the same size.> If none of the above works then we will have a take a closer > look at the details :-)Ok, let me try to reproduce the problem and get you more info. Rob This message posted from opensolaris.org