Hi list, I''ve a strange behaviour with autoreplace property. It is set to off by default, ok. I want to manually manage disk replacement so default "off" matches my need. # zpool get autoreplace mypool NAME PROPERTY VALUE SOURCE mypool autoreplace off default Then I added 2 spare disks. spares c1t18d0 AVAIL c1t19d0 AVAIL Ok, fine. Then I had failures with 1 disk of the pool and can see in logs the following : ------------------------ DESC: The number of I/O errors associated with a ZFS device exceeded acceptable levels. Refer to http://sun.com/msg/ZFS-8000-FD for more information. AUTO-RESPONSE: The device has been offlined and marked as faulted. An attempt will be made to activate a hot spare if available. ----------------------- This is where my problem occurs , zfs automatically replaced faulted disk by a spare ! even with autoreplace=off # zpool status pool: mypool state: DEGRADED status: One or more devices are faulted in response to persistent errors. Sufficient replicas exist for the pool to continue functioning in a degraded state. action: Replace the faulted device, or use ''zpool clear'' to mark the device repaired. scrub: resilver completed after 0h0m with 0 errors on Thu Feb 4 00:10:25 2010 config: NAME STATE READ WRITE CKSUM mypool DEGRADED 0 0 0 mirror ONLINE 0 0 0 c0t2d0 ONLINE 0 0 0 c0t3d0 ONLINE 0 0 0 c0t4d0 ONLINE 0 0 0 c0t5d0 ONLINE 0 0 0 mirror DEGRADED 0 0 0 c0t6d0 ONLINE 0 0 0 c0t7d0 ONLINE 0 0 0 spare DEGRADED 4 0 0 c1t8d0 FAULTED 3 26 0 too many errors c1t18d0 ONLINE 0 0 4 56K resilvered c1t9d0 ONLINE 0 0 0 mirror ONLINE 0 0 0 c1t10d0 ONLINE 0 0 0 c1t11d0 ONLINE 0 0 0 c1t12d0 ONLINE 0 0 0 c1t13d0 ONLINE 0 0 0 mirror ONLINE 0 0 0 c1t14d0 ONLINE 0 0 0 c1t15d0 ONLINE 0 0 0 c1t16d0 ONLINE 0 0 0 c1t17d0 ONLINE 0 0 0 cache c2d0 ONLINE 0 0 0 c3d0 ONLINE 0 0 0 spares c1t18d0 INUSE currently in use c1t19d0 AVAIL errors: No known data errors Any idea why it has been done automatically ? solaris 10U8 Generic_141445-09 - zpool version 15 - zfs version 4 Thx for your answers. -- Francois
Hi Francois, The autoreplace property works independently of the spare feature. Spares are activated automatically when a device in the main pool fails. Thanks, Cindy On 02/05/10 09:43, Francois wrote:> Hi list, > > I''ve a strange behaviour with autoreplace property. It is set to off by > default, ok. I want to manually manage disk replacement so default "off" > matches my need. > > # zpool get autoreplace mypool > NAME PROPERTY VALUE SOURCE > mypool autoreplace off default > > Then I added 2 spare disks. > > spares > c1t18d0 AVAIL > c1t19d0 AVAIL > > Ok, fine. > > Then I had failures with 1 disk of the pool and can see in logs the > following : > > ------------------------ > DESC: The number of I/O errors associated with a ZFS device exceeded > acceptable levels. Refer to http://sun.com/msg/ZFS-8000-FD for more > information. > AUTO-RESPONSE: The device has been offlined and marked as faulted. An > attempt will be made to activate a hot spare if available. > ----------------------- > > This is where my problem occurs , zfs automatically replaced faulted > disk by a spare ! even with autoreplace=off > > # zpool status > pool: mypool > state: DEGRADED > status: One or more devices are faulted in response to persistent errors. > Sufficient replicas exist for the pool to continue functioning in a > degraded state. > action: Replace the faulted device, or use ''zpool clear'' to mark the device > repaired. > scrub: resilver completed after 0h0m with 0 errors on Thu Feb 4 > 00:10:25 2010 > config: > > NAME STATE READ WRITE CKSUM > mypool DEGRADED 0 0 0 > mirror ONLINE 0 0 0 > c0t2d0 ONLINE 0 0 0 > c0t3d0 ONLINE 0 0 0 > c0t4d0 ONLINE 0 0 0 > c0t5d0 ONLINE 0 0 0 > mirror DEGRADED 0 0 0 > c0t6d0 ONLINE 0 0 0 > c0t7d0 ONLINE 0 0 0 > spare DEGRADED 4 0 0 > c1t8d0 FAULTED 3 26 0 too many errors > c1t18d0 ONLINE 0 0 4 56K resilvered > c1t9d0 ONLINE 0 0 0 > mirror ONLINE 0 0 0 > c1t10d0 ONLINE 0 0 0 > c1t11d0 ONLINE 0 0 0 > c1t12d0 ONLINE 0 0 0 > c1t13d0 ONLINE 0 0 0 > mirror ONLINE 0 0 0 > c1t14d0 ONLINE 0 0 0 > c1t15d0 ONLINE 0 0 0 > c1t16d0 ONLINE 0 0 0 > c1t17d0 ONLINE 0 0 0 > cache > c2d0 ONLINE 0 0 0 > c3d0 ONLINE 0 0 0 > spares > c1t18d0 INUSE currently in use > c1t19d0 AVAIL > > errors: No known data errors > > > Any idea why it has been done automatically ? > > solaris 10U8 Generic_141445-09 - zpool version 15 - zfs version 4 > > > Thx for your answers. > > -- > Francois > _______________________________________________ > zfs-discuss mailing list > zfs-discuss at opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
On Fri, Feb 5, 2010 at 12:11 PM, Cindy Swearingen <Cindy.Swearingen at sun.com>wrote:> Hi Francois, > > The autoreplace property works independently of the spare > feature. > > Spares are activated automatically when a device in the main > pool fails. > > Thanks, > > Cindy > > > On 02/05/10 09:43, Francois wrote: > >> Hi list, >> >> I''ve a strange behaviour with autoreplace property. It is set to off by >> default, ok. I want to manually manage disk replacement so default "off" >> matches my need. >> >> # zpool get autoreplace mypool >> NAME PROPERTY VALUE SOURCE >> mypool autoreplace off default >> >> Then I added 2 spare disks. >> >> spares >> c1t18d0 AVAIL >> c1t19d0 AVAIL >> >> Ok, fine. >> >> Then I had failures with 1 disk of the pool and can see in logs the >> following : >> >> ------------------------ >> DESC: The number of I/O errors associated with a ZFS device exceeded >> acceptable levels. Refer to http://sun.com/msg/ZFS-8000-FD for more >> information. >> AUTO-RESPONSE: The device has been offlined and marked as faulted. An >> attempt will be made to activate a hot spare if available. >> ----------------------- >> >> This is where my problem occurs , zfs automatically replaced faulted disk >> by a spare ! even with autoreplace=off >> >> # zpool status >> pool: mypool >> state: DEGRADED >> status: One or more devices are faulted in response to persistent errors. >> Sufficient replicas exist for the pool to continue functioning in a >> degraded state. >> action: Replace the faulted device, or use ''zpool clear'' to mark the >> device >> repaired. >> scrub: resilver completed after 0h0m with 0 errors on Thu Feb 4 00:10:25 >> 2010 >> config: >> >> NAME STATE READ WRITE CKSUM >> mypool DEGRADED 0 0 0 >> mirror ONLINE 0 0 0 >> c0t2d0 ONLINE 0 0 0 >> c0t3d0 ONLINE 0 0 0 >> c0t4d0 ONLINE 0 0 0 >> c0t5d0 ONLINE 0 0 0 >> mirror DEGRADED 0 0 0 >> c0t6d0 ONLINE 0 0 0 >> c0t7d0 ONLINE 0 0 0 >> spare DEGRADED 4 0 0 >> c1t8d0 FAULTED 3 26 0 too many errors >> c1t18d0 ONLINE 0 0 4 56K resilvered >> c1t9d0 ONLINE 0 0 0 >> mirror ONLINE 0 0 0 >> c1t10d0 ONLINE 0 0 0 >> c1t11d0 ONLINE 0 0 0 >> c1t12d0 ONLINE 0 0 0 >> c1t13d0 ONLINE 0 0 0 >> mirror ONLINE 0 0 0 >> c1t14d0 ONLINE 0 0 0 >> c1t15d0 ONLINE 0 0 0 >> c1t16d0 ONLINE 0 0 0 >> c1t17d0 ONLINE 0 0 0 >> cache >> c2d0 ONLINE 0 0 0 >> c3d0 ONLINE 0 0 0 >> spares >> c1t18d0 INUSE currently in use >> c1t19d0 AVAIL >> >> errors: No known data errors >> >> >> Any idea why it has been done automatically ? >> >> solaris 10U8 Generic_141445-09 - zpool version 15 - zfs version 4 >> >> >> Thx for your answers. >> >> -- >> Francois >> _______________________________________________ >> >I think it might be helpful to explain exactly what that means. I''ll give it a shot, feel free to correct my mistake(s). Francois: when you have autoreplace on, what that means is if you remove the bad drive, and stick in a new one to replace it, it will automatically be added to the pool. To do what you''re trying to do, you shouldn''t have drives added as hot spares at all. If you want it to be a "cold" spare, put it in the system, and just leave it unassigned. --Tim -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://mail.opensolaris.org/pipermail/zfs-discuss/attachments/20100205/b3f57f5d/attachment.html>