While not strictly a ZFS issue as such I thought I''d post here as this and the storage forums are my best bet in terms of getting some help. I have a machine that I recently set up with b130, b131 and b132. With each build I have been playing around with ZFS raidz2 and mirroring to do a little performance testing. This is a 6 SATA port ICH10 Intel motherboard, running in AHCI mode. OS is on a USB flash. Suffice to say I have noticed that 1 particular drive out of 6 seems to have very high asvc_t practically all the time. This is an excerpt from ''iostat -xnM c6t2d0 2''; [i] r/s w/s Mr/s Mw/s wait actv wsvc_t asvc_t %w %b device 70.5 502.0 0.0 4.1 0.0 1.3 0.0 2.2 0 54 c6t2d0 50.5 137.5 0.0 3.0 0.0 0.7 0.0 3.9 0 47 c6t2d0 71.0 163.5 0.0 4.8 0.0 0.8 0.0 3.4 0 61 c6t2d0 13.5 29.5 0.0 1.0 0.0 2.6 0.0 61.4 0 88 c6t2d0 1.0 0.5 0.0 0.0 0.0 3.6 0.0 2406.2 0 100 c6t2d0 1.0 1.0 0.0 0.0 0.0 4.0 0.0 1993.4 0 100 c6t2d0 1.0 1.5 0.0 0.0 0.0 4.0 0.0 1593.8 0 100 c6t2d0 2.0 3.0 0.0 0.1 0.0 4.0 0.0 791.6 0 100 c6t2d0 1.0 2.0 0.0 0.1 0.0 4.0 0.0 1320.3 0 100 c6t2d0 1.0 5.0 0.0 0.3 0.0 3.6 0.0 595.1 0 100 c6t2d0 [/i] and here is the drive shown with the other in the raidz2 pool extended device statistics r/s w/s Mr/s Mw/s wait actv wsvc_t asvc_t %w %b device 0.0 1.5 0.0 0.0 0.0 0.0 0.0 0.5 0 0 c6t0d0 0.0 1.5 0.0 0.0 0.0 0.0 0.0 0.3 0 0 c6t1d0 1.0 1.0 0.0 0.0 0.0 4.0 0.0 1994.8 0 100 c6t2d0 1.0 1.5 0.0 0.0 0.0 0.0 0.0 5.2 0 1 c6t3d0 1.0 1.5 0.0 0.0 0.0 0.0 0.0 6.9 0 1 c6t4d0 1.0 1.5 0.0 0.0 0.0 0.0 0.0 10.1 0 2 c6t5d0 extended device statistics r/s w/s Mr/s Mw/s wait actv wsvc_t asvc_t %w %b device 1.0 5.5 0.0 0.2 0.0 0.0 0.0 1.6 0 1 c6t0d0 1.0 5.5 0.0 0.2 0.0 0.0 0.0 1.5 0 1 c6t1d0 2.0 3.5 0.0 0.1 0.0 4.0 0.0 721.8 0 100 c6t2d0 1.0 5.5 0.0 0.2 0.0 0.0 0.0 1.9 0 1 c6t3d0 1.0 5.5 0.0 0.2 0.0 0.0 0.0 1.6 0 1 c6t4d0 2.0 5.5 0.0 0.2 0.0 0.0 0.0 3.1 0 2 c6t5d0 extended device statistics r/s w/s Mr/s Mw/s wait actv wsvc_t asvc_t %w %b device 0.0 3.5 0.0 0.1 0.0 0.0 0.0 0.4 0 0 c6t0d0 0.0 3.5 0.0 0.1 0.0 0.0 0.0 1.8 0 0 c6t1d0 1.0 2.0 0.0 0.1 0.0 4.0 0.0 1327.1 0 100 c6t2d0 1.0 3.5 0.0 0.1 0.0 0.0 0.0 4.9 0 1 c6t3d0 1.0 3.5 0.0 0.1 0.0 0.0 0.0 3.9 0 1 c6t4d0 1.0 3.5 0.0 0.1 0.0 0.0 0.0 2.0 0 1 c6t5d0 I have seen asvc_t as high as 20000. There do not appear to be any errors hardware wise as ''iostat -e'' shows [i] ---- errors --- device s/w h/w trn tot sd0 0 0 0 0 sd2 0 0 0 0 sd3 0 0 0 0 sd4 0 0 0 0 sd5 0 0 0 0 sd6 0 0 0 0 sd7 0 0 0 0 [/i] ''zpool iostat -v 2'' pauses for anywhere between 3 and 10 seconds before it prints the stats for that particular drive in the pool; capacity operations bandwidth pool alloc free read write read write ------------------ ----- ----- ----- ----- ----- ----- data 185G 5.26T 3 115 8.96K 2.48M raidz2 185G 5.26T 3 115 8.96K 2.48M c6t0d0 - - 2 26 2.70K 643K c6t1d0 - - 2 26 2.49K 643K * INSERT ~10 SECOND PAUSE* c6t2d0 - - 2 24 2.81K 643K c6t3d0 - - 2 26 2.75K 643K c6t4d0 - - 2 26 2.45K 643K c6t5d0 - - 2 26 2.71K 643K ------------------ ----- ----- ----- ----- ----- ----- rpool 3.50G 3.94G 0 0 9.99K 1010 c5t0d0s0 3.50G 3.94G 0 0 9.99K 1010 ------------------ ----- ----- ----- ----- ----- ----- swpool 102K 3.69G 0 0 19 0 /dev/rdsk/c7t0d0 102K 3.69G 0 0 19 0 ------------------ ----- ----- ----- ----- ----- ----- I have booted up a linux rescue CD that has S.M.A.R.T support (system rescue CD) and performed the ''long'' test on each drive. All drives pass the ''test''. There also appears to be no system errors with the drives under linux either. Can anyone shed any light on this issue, or suggest what I could try next ? I am sort of discounting hardware problems given that I do not see errors from the live linux CD. Maybe I should install linux and see if the problem persists ? Cheers. -- This message posted from opensolaris.org
On Sunday, February 7, 2010, Brian McKerr <brian at datamatters.com.au> wrote:> While not strictly a ZFS issue as such I thought I''d post here as this and the storage forums are my best bet in terms of getting some help. > > I have a machine that I recently set up with b130, b131 and b132. With each build I have been playing around with ZFS raidz2 and mirroring to do a little performance testing. This is a 6 SATA port ICH10 Intel motherboard, running in AHCI mode. OS is on a USB flash. Suffice to say I have noticed that 1 particular drive out of 6 seems to have very high asvc_t practically all the time. This is an excerpt from ''iostat -xnM c6t2d0 2''; > > ?[i] ? r/s ? ?w/s ? Mr/s ? Mw/s wait actv wsvc_t asvc_t ?%w ?%b device > ? 70.5 ?502.0 ? ?0.0 ? ?4.1 ?0.0 ?1.3 ? ?0.0 ? ?2.2 ? 0 ?54 c6t2d0 > ? 50.5 ?137.5 ? ?0.0 ? ?3.0 ?0.0 ?0.7 ? ?0.0 ? ?3.9 ? 0 ?47 c6t2d0 > ? 71.0 ?163.5 ? ?0.0 ? ?4.8 ?0.0 ?0.8 ? ?0.0 ? ?3.4 ? 0 ?61 c6t2d0 > ? 13.5 ? 29.5 ? ?0.0 ? ?1.0 ?0.0 ?2.6 ? ?0.0 ? 61.4 ? 0 ?88 c6t2d0 > ? ?1.0 ? ?0.5 ? ?0.0 ? ?0.0 ?0.0 ?3.6 ? ?0.0 2406.2 ? 0 100 c6t2d0 > ? ?1.0 ? ?1.0 ? ?0.0 ? ?0.0 ?0.0 ?4.0 ? ?0.0 1993.4 ? 0 100 c6t2d0 > ? ?1.0 ? ?1.5 ? ?0.0 ? ?0.0 ?0.0 ?4.0 ? ?0.0 1593.8 ? 0 100 c6t2d0 > ? ?2.0 ? ?3.0 ? ?0.0 ? ?0.1 ?0.0 ?4.0 ? ?0.0 ?791.6 ? 0 100 c6t2d0 > ? ?1.0 ? ?2.0 ? ?0.0 ? ?0.1 ?0.0 ?4.0 ? ?0.0 1320.3 ? 0 100 c6t2d0 > ? ?1.0 ? ?5.0 ? ?0.0 ? ?0.3 ?0.0 ?3.6 ? ?0.0 ?595.1 ? 0 100 c6t2d0 > [/i] > > and here is the drive shown with the other in the raidz2 pool > > ? ? ? ? ? ? ? ? ? ?extended device statistics > ? ?r/s ? ?w/s ? Mr/s ? Mw/s wait actv wsvc_t asvc_t ?%w ?%b device > ? ?0.0 ? ?1.5 ? ?0.0 ? ?0.0 ?0.0 ?0.0 ? ?0.0 ? ?0.5 ? 0 ? 0 c6t0d0 > ? ?0.0 ? ?1.5 ? ?0.0 ? ?0.0 ?0.0 ?0.0 ? ?0.0 ? ?0.3 ? 0 ? 0 c6t1d0 > ? ?1.0 ? ?1.0 ? ?0.0 ? ?0.0 ?0.0 ?4.0 ? ?0.0 1994.8 ? 0 100 c6t2d0 > ? ?1.0 ? ?1.5 ? ?0.0 ? ?0.0 ?0.0 ?0.0 ? ?0.0 ? ?5.2 ? 0 ? 1 c6t3d0 > ? ?1.0 ? ?1.5 ? ?0.0 ? ?0.0 ?0.0 ?0.0 ? ?0.0 ? ?6.9 ? 0 ? 1 c6t4d0 > ? ?1.0 ? ?1.5 ? ?0.0 ? ?0.0 ?0.0 ?0.0 ? ?0.0 ? 10.1 ? 0 ? 2 c6t5d0 > ? ? ? ? ? ? ? ? ? ?extended device statistics > ? ?r/s ? ?w/s ? Mr/s ? Mw/s wait actv wsvc_t asvc_t ?%w ?%b device > ? ?1.0 ? ?5.5 ? ?0.0 ? ?0.2 ?0.0 ?0.0 ? ?0.0 ? ?1.6 ? 0 ? 1 c6t0d0 > ? ?1.0 ? ?5.5 ? ?0.0 ? ?0.2 ?0.0 ?0.0 ? ?0.0 ? ?1.5 ? 0 ? 1 c6t1d0 > ? ?2.0 ? ?3.5 ? ?0.0 ? ?0.1 ?0.0 ?4.0 ? ?0.0 ?721.8 ? 0 100 c6t2d0 > ? ?1.0 ? ?5.5 ? ?0.0 ? ?0.2 ?0.0 ?0.0 ? ?0.0 ? ?1.9 ? 0 ? 1 c6t3d0 > ? ?1.0 ? ?5.5 ? ?0.0 ? ?0.2 ?0.0 ?0.0 ? ?0.0 ? ?1.6 ? 0 ? 1 c6t4d0 > ? ?2.0 ? ?5.5 ? ?0.0 ? ?0.2 ?0.0 ?0.0 ? ?0.0 ? ?3.1 ? 0 ? 2 c6t5d0 > ? ? ? ? ? ? ? ? ? ?extended device statistics > ? ?r/s ? ?w/s ? Mr/s ? Mw/s wait actv wsvc_t asvc_t ?%w ?%b device > ? ?0.0 ? ?3.5 ? ?0.0 ? ?0.1 ?0.0 ?0.0 ? ?0.0 ? ?0.4 ? 0 ? 0 c6t0d0 > ? ?0.0 ? ?3.5 ? ?0.0 ? ?0.1 ?0.0 ?0.0 ? ?0.0 ? ?1.8 ? 0 ? 0 c6t1d0 > ? ?1.0 ? ?2.0 ? ?0.0 ? ?0.1 ?0.0 ?4.0 ? ?0.0 1327.1 ? 0 100 c6t2d0 > ? ?1.0 ? ?3.5 ? ?0.0 ? ?0.1 ?0.0 ?0.0 ? ?0.0 ? ?4.9 ? 0 ? 1 c6t3d0 > ? ?1.0 ? ?3.5 ? ?0.0 ? ?0.1 ?0.0 ?0.0 ? ?0.0 ? ?3.9 ? 0 ? 1 c6t4d0 > ? ?1.0 ? ?3.5 ? ?0.0 ? ?0.1 ?0.0 ?0.0 ? ?0.0 ? ?2.0 ? 0 ? 1 c6t5d0 > > I have seen asvc_t as high as 20000. > > There do not appear to be any errors hardware wise as ''iostat -e'' shows > > ?[i] ? ? ? ? ?---- errors --- > device ?s/w h/w trn tot > sd0 ? ? ? 0 ? 0 ? 0 ? 0 > sd2 ? ? ? 0 ? 0 ? 0 ? 0 > sd3 ? ? ? 0 ? 0 ? 0 ? 0 > sd4 ? ? ? 0 ? 0 ? 0 ? 0 > sd5 ? ? ? 0 ? 0 ? 0 ? 0 > sd6 ? ? ? 0 ? 0 ? 0 ? 0 > sd7 ? ? ? 0 ? 0 ? 0 ? 0 > [/i] > > > > ''zpool iostat -v 2'' pauses for anywhere between 3 and 10 seconds before it prints the stats for that particular drive in the pool; > > ? ? ? ? ? ? ? ? ? ? ? capacity ? ? operations ? ?bandwidth > pool ? ? ? ? ? ? ? ?alloc ? free ? read ?write ? read ?write > ------------------ ?----- ?----- ?----- ?----- ?----- ?----- > data ? ? ? ? ? ? ? ? 185G ?5.26T ? ? ?3 ? ?115 ?8.96K ?2.48M > ?raidz2 ? ? ? ? ? ? 185G ?5.26T ? ? ?3 ? ?115 ?8.96K ?2.48M > ? ?c6t0d0 ? ? ? ? ? ? ?- ? ? ?- ? ? ?2 ? ? 26 ?2.70K ? 643K > ? ?c6t1d0 ? ? ? ? ? ? ?- ? ? ?- ? ? ?2 ? ? 26 ?2.49K ? 643K > > * INSERT ~10 SECOND PAUSE* > > ? ?c6t2d0 ? ? ? ? ? ? ?- ? ? ?- ? ? ?2 ? ? 24 ?2.81K ? 643K > ? ?c6t3d0 ? ? ? ? ? ? ?- ? ? ?- ? ? ?2 ? ? 26 ?2.75K ? 643K > ? ?c6t4d0 ? ? ? ? ? ? ?- ? ? ?- ? ? ?2 ? ? 26 ?2.45K ? 643K > ? ?c6t5d0 ? ? ? ? ? ? ?- ? ? ?- ? ? ?2 ? ? 26 ?2.71K ? 643K > ------------------ ?----- ?----- ?----- ?----- ?----- ?----- > rpool ? ? ? ? ? ? ? 3.50G ?3.94G ? ? ?0 ? ? ?0 ?9.99K ? 1010 > ?c5t0d0s0 ? ? ? ? ?3.50G ?3.94G ? ? ?0 ? ? ?0 ?9.99K ? 1010 > ------------------ ?----- ?----- ?----- ?----- ?----- ?----- > swpool ? ? ? ? ? ? ? 102K ?3.69G ? ? ?0 ? ? ?0 ? ? 19 ? ? ?0 > ?/dev/rdsk/c7t0d0 ? 102K ?3.69G ? ? ?0 ? ? ?0 ? ? 19 ? ? ?0 > ------------------ ?----- ?----- ?----- ?----- ?----- ?----- > > I have booted up a linux rescue CD that has S.M.A.R.T support (system rescue CD) and performed the ''long'' test on each drive. All drives pass the ''test''. There also appears to be no system errors with the drives under linux either. > > Can anyone shed any light on this issue, or suggest what I could try next ? I am sort of discounting hardware problems given that I do not see errors from the live linux CD. Maybe I should install linux and see if the problem persists ? > > Cheers. > -- > This message posted from opensolaris.org > _______________________________________________ > zfs-discuss mailing list > zfs-discuss at opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discussI''d say your easiest two options are swap ports and see if the problem follows the drive. If it does, swap the drive out. --Tim
> > > I''d say your easiest two options are swap ports and > see if the problem > follows the drive. If it does, swap the drive out. > > > --Tim > _______________________________________________Yep, that sounds like a plan. Thanks for your suggestion. -- This message posted from opensolaris.org
Ok, I changed the cable and also tried swapping the port on the motherboard. The drive continued to have huge asvc_t and also started to have huge wsvc_t. I unplugged it and the ''pool'' is now operating as per expected performance wise. See the ''storage'' forum for any further updates as I am now convinced this has nothing to do with ZFS or my attempt to disable the ZIL. 8-) -- This message posted from opensolaris.org
Hi, If I may - you mentioned that you use ICH10 over ahci. As far as I know ICH10 is not officially supported by the ahci module. I have also tried myself on various ICH10 systems without success. OSOL wouldn''t even install on pre-130 builds, and I haven''t tried since. Regards, Tonmaus -- This message posted from opensolaris.org
I have booted up an osol-dev-131 live CD on a Dell Precision T7500, and the AHCI driver successfully loaded, to give access to the two sata DVD drives in the machine. (Unfortunately, I did not have the opportunity to attach any hard drives, but I would expect that also to work.) ''scanpci'' identified the southbridge as an Intel 82801JI (ICH10 family) Vendor 0x8086, device 0x3a22 AFAIK, as long as the SATA interface report a PCI ID class-code of 010601, then the AHCI device driver should load. The mode of the SATA interface will need to be selected in the BIOS. There are normally three modes: Native IDE, RAID or AHCI. ''scanpci'' should report different class-codes depending on the mode selected in the BIOS. RAID mode should report a class-code of 010400 IDE mode should report a class-code of 0101xx With OpenSolaris, you can see the class-code in the output from ''prtconf -pv''. If Native IDE is selected the ICH10 SATA interface should appear as two controllers, the first for ports 0-3, and the second for ports 4 & 5. Regards Nigel Smith -- This message posted from opensolaris.org