Charles Emery
2008-Jul-28 18:30 UTC
[zfs-discuss] zpool status my_pool , shows a pulled disk c1t6d0 as ONLINE ???
New server build with Solaris-10 u5/08, on a SunFire t5220, and this is our first rollout of ZFS and Zpools. Have 8 disks, boot disk is hardware mirrored (c1t0d0 + c1t1d0) Created Zpool my_pool as RaidZ using 5 disks + 1 spare: c1t2d0, c1t3d0, c1t4d0, c1t5d0, c1t6d0, and spare c1t7d0 I am working on alerting & recovery plans for disks failures in the zpool. As a test, I have pulled disk c1t6d0, to see what a disk failure will look like. "zpool status -v mypool" Still reports disk c1t6d0 as ONLINE. "iostat -En" also does not yet realize that the disk is pulled. By contrast, format realizes the disk is missing, and the disk pull did generate errors in /var/adm/messages. Do I need to hit the device bus with some a command to get a more accurate status, or something like that? Would appreciate any recommendations for zpool disk failure monitoring? See the attachment for output from iostat -En, format, and the tail of /var/adm/messages: Here is the output from "zpool status -v": ===========================================newserver:/# zpool status -v pool: my_pool state: ONLINE scrub: none requested config: NAME STATE READ WRITE CKSUM my_pool ONLINE 0 0 0 raidz1 ONLINE 0 0 0 c1t2d0 ONLINE 0 0 0 c1t3d0 ONLINE 0 0 0 c1t4d0 ONLINE 0 0 0 c1t5d0 ONLINE 0 0 0 c1t6d0 ONLINE 0 0 0 spares c1t7d0 AVAIL errors: No known data errors ============================================ Message was edited by: cemery This message posted from opensolaris.org -------------- next part -------------- zpool status my_pool , shows a pulled disk c1t6d0 as ONLINE ??? New server build, on a SunFire t5220, and this is our first rollout of ZFS and Zpools. Have 8 disks, boot disk is hardware mirrored (c1t0d0 + c1t1d0) Created a Raidz (raid5) zpool (called my_pool), including 1 spare, Created my_pool as RaidZ pool using 5 disks + 1 spare disk: c1t2d0, c1t3d0, c1t4d0, c1t5d0, c1t6d0, and spare c1t7d0 I am working on alerting & recovery plans for disks failures in the zpool. As a test, I have pulled disk c1t6d0, to see what a disk failure will look like. "zpool status -v mypool" Still reports disk c1t6d0 as ONLINE. "iostat -En" also does not yet realize that the disk is pulled. By contrast, format realizes the disk is missing, and the disk pull did generate errors in /var/adm/messages. Do I need to hit the device bus with some a command to get a more accurate status? Would appreciate any recommendations for zpool disk failure monitoring? Below are the output from "zpool status -v", "iostat -En", "format", and the tail of /var/adm/messages: ===========================================newserver:/# zpool status -v pool: my_pool state: ONLINE scrub: none requested config: NAME STATE READ WRITE CKSUM my_pool ONLINE 0 0 0 raidz1 ONLINE 0 0 0 c1t2d0 ONLINE 0 0 0 c1t3d0 ONLINE 0 0 0 c1t4d0 ONLINE 0 0 0 c1t5d0 ONLINE 0 0 0 c1t6d0 ONLINE 0 0 0 spares c1t7d0 AVAIL errors: No known data errors ============================================ newserver:/# iostat -En c1t0d0 Soft Errors: 2 Hard Errors: 0 Transport Errors: 0 Vendor: LSILOGIC Product: Logical Volume Revision: 3000 Serial No: Size: 146.56GB <146561286144 bytes> Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0 Illegal Request: 2 Predictive Failure Analysis: 0 c1t2d0 Soft Errors: 0 Hard Errors: 0 Transport Errors: 0 Vendor: SEAGATE Product: ST914602SSUN146G Revision: 0603 Serial No: 0811953XZG Size: 146.81GB <146810536448 bytes> Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0 Illegal Request: 0 Predictive Failure Analysis: 0 c0t0d0 Soft Errors: 4 Hard Errors: 3 Transport Errors: 0 Vendor: TSSTcorp Product: CD/DVDW TS-T632A Revision: SR03 Serial No: Size: 0.00GB <0 bytes> Media Error: 0 Device Not Ready: 3 No Device: 0 Recoverable: 0 Illegal Request: 4 Predictive Failure Analysis: 0 c1t3d0 Soft Errors: 0 Hard Errors: 0 Transport Errors: 0 Vendor: SEAGATE Product: ST914602SSUN146G Revision: 0603 Serial No: 08139591NN Size: 146.81GB <146810536448 bytes> Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0 Illegal Request: 0 Predictive Failure Analysis: 0 c1t4d0 Soft Errors: 0 Hard Errors: 0 Transport Errors: 0 Vendor: SEAGATE Product: ST914602SSUN146G Revision: 0603 Serial No: 0813957V3R Size: 146.81GB <146810536448 bytes> Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0 Illegal Request: 0 Predictive Failure Analysis: 0 c1t5d0 Soft Errors: 0 Hard Errors: 0 Transport Errors: 0 Vendor: SEAGATE Product: ST914602SSUN146G Revision: 0603 Serial No: 0813957V2J Size: 146.81GB <146810536448 bytes> Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0 Illegal Request: 0 Predictive Failure Analysis: 0 c1t6d0 Soft Errors: 0 Hard Errors: 0 Transport Errors: 0 Vendor: SEAGATE Product: ST914602SSUN146G Revision: 0603 Serial No: 0812951XFK Size: 146.81GB <146810536448 bytes> Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0 Illegal Request: 0 Predictive Failure Analysis: 0 c1t7d0 Soft Errors: 0 Hard Errors: 0 Transport Errors: 0 Vendor: SEAGATE Product: ST914602SSUN146G Revision: 0603 Serial No: 0813957V4B Size: 146.81GB <146810536448 bytes> Media Error: 0 Device Not Ready: 0 No Device: 0 Recoverable: 0 Illegal Request: 0 Predictive Failure Analysis: 0 ============================================ newserver:/# format Searching for disks...done AVAILABLE DISK SELECTIONS: ... 5. c1t6d0 <drive not available> /pci at 0/pci at 0/pci at 2/scsi at 0/sd at 6,0 ... Specify disk (enter its number): 5 AVAILABLE DRIVE TYPES: 0. Auto configure 1. Quantum ProDrive 80S 2. Quantum ProDrive 105S ... 20. other Specify disk type (enter its number): ^C ============================================ newserver:/# tail -20 /var/adm/messages Jul 9 14:49:26 sbknwsxapd3 scsi: [ID 107833 kern.notice] Device is gone Jul 9 14:49:26 sbknwsxapd3 last message repeated 1 time Jul 9 14:49:26 sbknwsxapd3 scsi: [ID 107833 kern.warning] WARNING: /pci at 0/pci at 0/pci at 2/scsi at 0/sd at 6,0 (sd6): Jul 9 14:49:26 sbknwsxapd3 offline or reservation conflict Jul 9 14:51:01 sbknwsxapd3 scsi: [ID 107833 kern.notice] Device is gone Jul 9 14:51:01 sbknwsxapd3 last message repeated 1 time Jul 9 14:51:01 sbknwsxapd3 scsi: [ID 107833 kern.warning] WARNING: /pci at 0/pci at 0/pci at 2/scsi at 0/sd at 6,0 (sd6): Jul 9 14:51:01 sbknwsxapd3 offline or reservation conflict Jul 9 14:51:01 sbknwsxapd3 scsi: [ID 107833 kern.notice] Device is gone Jul 9 14:51:01 sbknwsxapd3 last message repeated 1 time Jul 9 14:51:01 sbknwsxapd3 scsi: [ID 107833 kern.warning] WARNING: /pci at 0/pci at 0/pci at 2/scsi at 0/sd at 6,0 (sd6): Jul 9 14:51:01 sbknwsxapd3 offline or reservation conflict Jul 9 14:51:12 sbknwsxapd3 scsi: [ID 107833 kern.notice] Device is gone Jul 9 14:51:12 sbknwsxapd3 last message repeated 2 times Jul 9 14:51:12 sbknwsxapd3 scsi: [ID 107833 kern.warning] WARNING: /pci at 0/pci at 0/pci at 2/scsi at 0/sd at 6,0 (sd6): Jul 9 14:51:12 sbknwsxapd3 offline or reservation conflict Jul 9 14:51:14 sbknwsxapd3 scsi: [ID 107833 kern.notice] Device is gone Jul 9 14:51:14 sbknwsxapd3 last message repeated 1 time Jul 9 14:51:14 sbknwsxapd3 scsi: [ID 107833 kern.warning] WARNING: /pci at 0/pci at 0/pci at 2/scsi at 0/sd at 6,0 (sd6): Jul 9 14:51:14 sbknwsxapd3 offline or reservation conflict
Richard Elling
2008-Jul-28 19:22 UTC
[zfs-discuss] zpool status my_pool , shows a pulled disk c1t6d0 as ONLINE ???
Charles Emery wrote:> New server build with Solaris-10 u5/08, >Can you try on a later release? The enhanced FMA for disks did not make the Solaris 10 5/08 release. http://www.opensolaris.org/os/community/on/flag-days/pages/2007080901/ -- richard> on a SunFire t5220, and this is our first rollout of ZFS and Zpools. > > Have 8 disks, boot disk is hardware mirrored (c1t0d0 + c1t1d0) > > Created Zpool my_pool as RaidZ using 5 disks + 1 spare: > c1t2d0, c1t3d0, c1t4d0, c1t5d0, c1t6d0, and spare c1t7d0 > > I am working on alerting & recovery plans for disks failures in the zpool. > As a test, I have pulled disk c1t6d0, to see what a disk failure will look like. > "zpool status -v mypool" Still reports disk c1t6d0 as ONLINE. > "iostat -En" also does not yet realize that the disk is pulled. > > By contrast, format realizes the disk is missing, > and the disk pull did generate errors in /var/adm/messages. > > Do I need to hit the device bus with some a command to get a more accurate status, or something like that? > Would appreciate any recommendations for zpool disk failure monitoring? > > See the attachment for output from iostat -En, format, and the tail of /var/adm/messages: > > Here is the output from "zpool status -v": > ===========================================> newserver:/# zpool status -v > pool: my_pool > state: ONLINE > scrub: none requested > config: > NAME STATE READ WRITE CKSUM > my_pool ONLINE 0 0 0 > raidz1 ONLINE 0 0 0 > c1t2d0 ONLINE 0 0 0 > c1t3d0 ONLINE 0 0 0 > c1t4d0 ONLINE 0 0 0 > c1t5d0 ONLINE 0 0 0 > c1t6d0 ONLINE 0 0 0 > spares > c1t7d0 AVAIL > errors: No known data errors > ============================================> > Message was edited by: > cemery > > > This message posted from opensolaris.org > > ------------------------------------------------------------------------ > > _______________________________________________ > zfs-discuss mailing list > zfs-discuss at opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss