Jan-Aage Frydenbø-Bruvoll
2011-Dec-18 14:53 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
Dear List, I have a storage server running OpenIndiana with a number of storage pools on it. All the pools'' disks come off the same controller, and all pools are backed by SSD-based l2arc and ZIL. Performance is excellent on all pools but one, and I am struggling greatly to figure out what is wrong. A very basic test shows the following - pretty much typical performance at the moment: root at stor:/# for a in pool1 pool2 pool3; do dd if=/dev/zero of=$a/file bs=1M count=10; done 10+0 records in 10+0 records out 10485760 bytes (10 MB) copied, 0.00772965 s, 1.4 GB/s 10+0 records in 10+0 records out 10485760 bytes (10 MB) copied, 0.00996472 s, 1.1 GB/s 10+0 records in 10+0 records out 10485760 bytes (10 MB) copied, 71.8995 s, 146 kB/s The zpool status of the affected pool is: root at stor:/# zpool status pool3 pool: pool3 state: ONLINE scan: resilvered 222G in 24h2m with 0 errors on Wed Dec 14 15:20:11 2011 config: NAME STATE READ WRITE CKSUM pool3 ONLINE 0 0 0 c1t0d0 ONLINE 0 0 0 c1t1d0 ONLINE 0 0 0 c1t2d0 ONLINE 0 0 0 c1t3d0 ONLINE 0 0 0 c1t4d0 ONLINE 0 0 0 c1t5d0 ONLINE 0 0 0 c1t6d0 ONLINE 0 0 0 c1t7d0 ONLINE 0 0 0 c1t8d0 ONLINE 0 0 0 c1t9d0 ONLINE 0 0 0 c1t10d0 ONLINE 0 0 0 mirror-12 ONLINE 0 0 0 c1t26d0 ONLINE 0 0 0 c1t27d0 ONLINE 0 0 0 mirror-13 ONLINE 0 0 0 c1t28d0 ONLINE 0 0 0 c1t29d0 ONLINE 0 0 0 mirror-14 ONLINE 0 0 0 c1t34d0 ONLINE 0 0 0 c1t35d0 ONLINE 0 0 0 logs mirror-11 ONLINE 0 0 0 c2t2d0p8 ONLINE 0 0 0 c2t3d0p8 ONLINE 0 0 0 cache c2t2d0p12 ONLINE 0 0 0 c2t3d0p12 ONLINE 0 0 0 errors: No known data errors Ditto for the disk controller - MegaCli reports zero errors, be that on the controller itself, on this pool''s disks or on any of the other attached disks. I am pretty sure I am dealing with a disk-based problem here, i.e. a flaky disk that is "just" slow without exhibiting any actual data errors, holding the rest of the pool back, but I am at a miss as how to pinpoint what is going on. Would anybody on the list be able to give me any pointers as how to dig up more detailed information about the pool''s/hardware''s performance? Thank you in advance for your kind assistance. Best regards Jan
"Hung-Sheng Tsao (Lao Tsao 老曹) Ph.D."
2011-Dec-18 15:13 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
what are the output of zpool status pool1 and pool2 it seems that you have mix configuration of pool3 with disk and mirror On 12/18/2011 9:53 AM, Jan-Aage Frydenb?-Bruvoll wrote:> Dear List, > > I have a storage server running OpenIndiana with a number of storage > pools on it. All the pools'' disks come off the same controller, and > all pools are backed by SSD-based l2arc and ZIL. Performance is > excellent on all pools but one, and I am struggling greatly to figure > out what is wrong. > > A very basic test shows the following - pretty much typical > performance at the moment: > > root at stor:/# for a in pool1 pool2 pool3; do dd if=/dev/zero of=$a/file > bs=1M count=10; done > 10+0 records in > 10+0 records out > 10485760 bytes (10 MB) copied, 0.00772965 s, 1.4 GB/s > 10+0 records in > 10+0 records out > 10485760 bytes (10 MB) copied, 0.00996472 s, 1.1 GB/s > 10+0 records in > 10+0 records out > 10485760 bytes (10 MB) copied, 71.8995 s, 146 kB/s > > The zpool status of the affected pool is: > > root at stor:/# zpool status pool3 > pool: pool3 > state: ONLINE > scan: resilvered 222G in 24h2m with 0 errors on Wed Dec 14 15:20:11 2011 > config: > > NAME STATE READ WRITE CKSUM > pool3 ONLINE 0 0 0 > c1t0d0 ONLINE 0 0 0 > c1t1d0 ONLINE 0 0 0 > c1t2d0 ONLINE 0 0 0 > c1t3d0 ONLINE 0 0 0 > c1t4d0 ONLINE 0 0 0 > c1t5d0 ONLINE 0 0 0 > c1t6d0 ONLINE 0 0 0 > c1t7d0 ONLINE 0 0 0 > c1t8d0 ONLINE 0 0 0 > c1t9d0 ONLINE 0 0 0 > c1t10d0 ONLINE 0 0 0 > mirror-12 ONLINE 0 0 0 > c1t26d0 ONLINE 0 0 0 > c1t27d0 ONLINE 0 0 0 > mirror-13 ONLINE 0 0 0 > c1t28d0 ONLINE 0 0 0 > c1t29d0 ONLINE 0 0 0 > mirror-14 ONLINE 0 0 0 > c1t34d0 ONLINE 0 0 0 > c1t35d0 ONLINE 0 0 0 > logs > mirror-11 ONLINE 0 0 0 > c2t2d0p8 ONLINE 0 0 0 > c2t3d0p8 ONLINE 0 0 0 > cache > c2t2d0p12 ONLINE 0 0 0 > c2t3d0p12 ONLINE 0 0 0 > > errors: No known data errors > > Ditto for the disk controller - MegaCli reports zero errors, be that > on the controller itself, on this pool''s disks or on any of the other > attached disks. > > I am pretty sure I am dealing with a disk-based problem here, i.e. a > flaky disk that is "just" slow without exhibiting any actual data > errors, holding the rest of the pool back, but I am at a miss as how > to pinpoint what is going on. > > Would anybody on the list be able to give me any pointers as how to > dig up more detailed information about the pool''s/hardware''s > performance? > > Thank you in advance for your kind assistance. > > Best regards > Jan > _______________________________________________ > zfs-discuss mailing list > zfs-discuss at opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss-- Hung-Sheng Tsao Ph D. Founder& Principal HopBit GridComputing LLC cell: 9734950840 http://laotsao.wordpress.com/ http://laotsao.blogspot.com/ -------------- next part -------------- A non-text attachment was scrubbed... Name: laotsao.vcf Type: text/x-vcard Size: 153 bytes Desc: not available URL: <http://mail.opensolaris.org/pipermail/zfs-discuss/attachments/20111218/15c7dfda/attachment.vcf>
Jan-Aage Frydenbø-Bruvoll
2011-Dec-18 15:46 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
Hi, On Sun, Dec 18, 2011 at 15:13, "Hung-Sheng Tsao (Lao Tsao ??) Ph.D." <laotsao at gmail.com> wrote:> what are the output of zpool status pool1 and pool2 > it seems that you have mix configuration of pool3 with disk and mirrorThe other two pools show very similar outputs: root at stor:~# zpool status pool1 pool: pool1 state: ONLINE scan: resilvered 1.41M in 0h0m with 0 errors on Sun Dec 4 17:42:35 2011 config: NAME STATE READ WRITE CKSUM pool1 ONLINE 0 0 0 mirror-0 ONLINE 0 0 0 c1t12d0 ONLINE 0 0 0 c1t13d0 ONLINE 0 0 0 mirror-1 ONLINE 0 0 0 c1t24d0 ONLINE 0 0 0 c1t25d0 ONLINE 0 0 0 mirror-2 ONLINE 0 0 0 c1t30d0 ONLINE 0 0 0 c1t31d0 ONLINE 0 0 0 mirror-3 ONLINE 0 0 0 c1t32d0 ONLINE 0 0 0 c1t33d0 ONLINE 0 0 0 logs mirror-4 ONLINE 0 0 0 c2t2d0p6 ONLINE 0 0 0 c2t3d0p6 ONLINE 0 0 0 cache c2t2d0p10 ONLINE 0 0 0 c2t3d0p10 ONLINE 0 0 0 errors: No known data errors root at stor:~# zpool status pool2 pool: pool2 state: ONLINE scan: scrub canceled on Wed Dec 14 07:51:50 2011 config: NAME STATE READ WRITE CKSUM pool2 ONLINE 0 0 0 mirror-0 ONLINE 0 0 0 c1t14d0 ONLINE 0 0 0 c1t15d0 ONLINE 0 0 0 mirror-1 ONLINE 0 0 0 c1t18d0 ONLINE 0 0 0 c1t19d0 ONLINE 0 0 0 mirror-2 ONLINE 0 0 0 c1t20d0 ONLINE 0 0 0 c1t21d0 ONLINE 0 0 0 mirror-3 ONLINE 0 0 0 c1t22d0 ONLINE 0 0 0 c1t23d0 ONLINE 0 0 0 logs mirror-4 ONLINE 0 0 0 c2t2d0p7 ONLINE 0 0 0 c2t3d0p7 ONLINE 0 0 0 cache c2t2d0p11 ONLINE 0 0 0 c2t3d0p11 ONLINE 0 0 0 The affected pool does indeed have a mix of straight disks and mirrored disks (due to running out of vdevs on the controller), however it has to be added that the performance of the affected pool was excellent until around 3 weeks ago, and there have been no structural changes nor to the pools neither to anything else on this server in the last half year or so. -jan
Fajar A. Nugraha
2011-Dec-18 16:41 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
On Sun, Dec 18, 2011 at 10:46 PM, Jan-Aage Frydenb?-Bruvoll <jan at architechs.eu> wrote:> The affected pool does indeed have a mix of straight disks and > mirrored disks (due to running out of vdevs on the controller), > however it has to be added that the performance of the affected pool > was excellent until around 3 weeks ago, and there have been no > structural changes nor to the pools neither to anything else on this > server in the last half year or so.Is the pool over 80% full? Do you have dedup enabled (even if it was turned off later, see "zpool history")? -- Fajar
Jan-Aage Frydenbø-Bruvoll
2011-Dec-18 17:40 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
Hi, On Sun, Dec 18, 2011 at 16:41, Fajar A. Nugraha <work at fajar.net> wrote:> Is the pool over 80% full? Do you have dedup enabled (even if it was > turned off later, see "zpool history")?The pool stands at 86%, but that has not changed in any way that corresponds chronologically with the sudden drop in performance on the pool. With regards to dedup - I have already got my fingers burned thoroughly on dedup, so that setting is, has been and will remain firmly off. -jan
Fajar A. Nugraha
2011-Dec-18 22:00 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
On Mon, Dec 19, 2011 at 12:40 AM, Jan-Aage Frydenb?-Bruvoll <jan at architechs.eu> wrote:> Hi, > > On Sun, Dec 18, 2011 at 16:41, Fajar A. Nugraha <work at fajar.net> wrote: >> Is the pool over 80% full? Do you have dedup enabled (even if it was >> turned off later, see "zpool history")? > > The pool stands at 86%, but that has not changed in any way that > corresponds chronologically with the sudden drop in performance on the > pool.>From http://www.solarisinternals.com/wiki/index.php/ZFS_Best_Practices_Guide(or at least Google''s cache of it, since it seems to be inaccessible now: " Keep pool space under 80% utilization to maintain pool performance. Currently, pool performance can degrade when a pool is very full and file systems are updated frequently, such as on a busy mail server. Full pools might cause a performance penalty, but no other issues. If the primary workload is immutable files (write once, never remove), then you can keep a pool in the 95-96% utilization range. Keep in mind that even with mostly static content in the 95-96% range, write, read, and resilvering performance might suffer. " I''m guessing that your nearly-full disk, combined with your usage performance, is the cause of slow down. Try freeing up some space (e.g. make it about 75% full), just tot be sure. -- Fajar
Jan-Aage Frydenbø-Bruvoll
2011-Dec-18 22:05 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
Hi, On Sun, Dec 18, 2011 at 22:00, Fajar A. Nugraha <work at fajar.net> wrote:> From http://www.solarisinternals.com/wiki/index.php/ZFS_Best_Practices_Guide > (or at least Google''s cache of it, since it seems to be inaccessible > now: > > " > Keep pool space under 80% utilization to maintain pool performance. > Currently, pool performance can degrade when a pool is very full and > file systems are updated frequently, such as on a busy mail server. > Full pools might cause a performance penalty, but no other issues. If > the primary workload is immutable files (write once, never remove), > then you can keep a pool in the 95-96% utilization range. Keep in mind > that even with mostly static content in the 95-96% range, write, read, > and resilvering performance might suffer. > " > > I''m guessing that your nearly-full disk, combined with your usage > performance, is the cause of slow down. Try freeing up some space > (e.g. make it about 75% full), just tot be sure.I''m aware of the guidelines you refer to, and I have had slowdowns before due to the pool being too full, but that was in the 9x% range and the slowdown was in the order of a few percent. At the moment I am slightly above the recommended limit, and the performance is currently between 1/10000 and 1/2000 of what the other pools achieve - i.e. a few hundred kB/s versus 2GB/s on the other pools - surely allocation above 80% cannot carry such extreme penalties?! For the record - the read/write load on the pool is almost exclusively WORM. Best regards Jan
Nathan Kroenert
2011-Dec-18 22:14 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
I know some others may already have pointed this out - but I can''t see it and not say something... Do you realise that losing a single disk in that pool could pretty much render the whole thing busted? At least for me - the rate at which _I_ seem to lose disks, it would be worth considering something different ;) Cheers! Nathan. On 12/19/11 09:05 AM, Jan-Aage Frydenb?-Bruvoll wrote:> Hi, > > On Sun, Dec 18, 2011 at 22:00, Fajar A. Nugraha<work at fajar.net> wrote: >> From http://www.solarisinternals.com/wiki/index.php/ZFS_Best_Practices_Guide >> (or at least Google''s cache of it, since it seems to be inaccessible >> now: >> >> " >> Keep pool space under 80% utilization to maintain pool performance. >> Currently, pool performance can degrade when a pool is very full and >> file systems are updated frequently, such as on a busy mail server. >> Full pools might cause a performance penalty, but no other issues. If >> the primary workload is immutable files (write once, never remove), >> then you can keep a pool in the 95-96% utilization range. Keep in mind >> that even with mostly static content in the 95-96% range, write, read, >> and resilvering performance might suffer. >> " >> >> I''m guessing that your nearly-full disk, combined with your usage >> performance, is the cause of slow down. Try freeing up some space >> (e.g. make it about 75% full), just tot be sure. > I''m aware of the guidelines you refer to, and I have had slowdowns > before due to the pool being too full, but that was in the 9x% range > and the slowdown was in the order of a few percent. > > At the moment I am slightly above the recommended limit, and the > performance is currently between 1/10000 and 1/2000 of what the other > pools achieve - i.e. a few hundred kB/s versus 2GB/s on the other > pools - surely allocation above 80% cannot carry such extreme > penalties?! > > For the record - the read/write load on the pool is almost exclusively WORM. > > Best regards > Jan > _______________________________________________ > zfs-discuss mailing list > zfs-discuss at opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Jan-Aage Frydenbø-Bruvoll
2011-Dec-18 22:23 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
Hi, On Sun, Dec 18, 2011 at 22:14, Nathan Kroenert <nathan at tuneunix.com> wrote:> ?I know some others may already have pointed this out - but I can''t see it > and not say something... > > Do you realise that losing a single disk in that pool could pretty much > render the whole thing busted? > > At least for me - the rate at which _I_ seem to lose disks, it would be > worth considering something different ;)Yeah, I have thought that thought myself. I am pretty sure I have a broken disk, however I cannot for the life of me find out which one. zpool status gives me nothing to work on, MegaCli reports that all virtual and physical drives are fine, and iostat gives me nothing either. What other tools are there out there that could help me pinpoint what''s going on? Best regards Jan
Craig Morgan
2011-Dec-18 22:33 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
Try fmdump -e and then fmdump -eV, it could be a pathological disk just this side of failure doing heavy retries that id dragging the pool down. Craig -- Craig Morgan On 18 Dec 2011, at 16:23, Jan-Aage Frydenb?-Bruvoll <jan at architechs.eu> wrote:> Hi, > > On Sun, Dec 18, 2011 at 22:14, Nathan Kroenert <nathan at tuneunix.com> wrote: >> I know some others may already have pointed this out - but I can''t see it >> and not say something... >> >> Do you realise that losing a single disk in that pool could pretty much >> render the whole thing busted? >> >> At least for me - the rate at which _I_ seem to lose disks, it would be >> worth considering something different ;) > > Yeah, I have thought that thought myself. I am pretty sure I have a > broken disk, however I cannot for the life of me find out which one. > zpool status gives me nothing to work on, MegaCli reports that all > virtual and physical drives are fine, and iostat gives me nothing > either. > > What other tools are there out there that could help me pinpoint > what''s going on? > > Best regards > Jan > _______________________________________________ > zfs-discuss mailing list > zfs-discuss at opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Matt Breitbach
2011-Dec-18 22:38 UTC
[zfs-discuss] Very poor pool performance - no zfs/controllererrors?!
I''d look at iostat -En. It will give you a good breakdown of disks that have seen errors. I''ve also spotted failing disks just by watching an iostat -nxz and looking for the one who''s spending more %busy than the rest of them, or exhibiting longer than normal service times. -Matt -----Original Message----- From: zfs-discuss-bounces at opensolaris.org [mailto:zfs-discuss-bounces at opensolaris.org] On Behalf Of Jan-Aage Frydenb?-Bruvoll Sent: Sunday, December 18, 2011 4:24 PM To: Nathan Kroenert Cc: zfs-discuss at opensolaris.org Subject: Re: [zfs-discuss] Very poor pool performance - no zfs/controllererrors?! Hi, On Sun, Dec 18, 2011 at 22:14, Nathan Kroenert <nathan at tuneunix.com> wrote:> ?I know some others may already have pointed this out - but I can''t see it > and not say something... > > Do you realise that losing a single disk in that pool could pretty much > render the whole thing busted? > > At least for me - the rate at which _I_ seem to lose disks, it would be > worth considering something different ;)Yeah, I have thought that thought myself. I am pretty sure I have a broken disk, however I cannot for the life of me find out which one. zpool status gives me nothing to work on, MegaCli reports that all virtual and physical drives are fine, and iostat gives me nothing either. What other tools are there out there that could help me pinpoint what''s going on? Best regards Jan _______________________________________________ zfs-discuss mailing list zfs-discuss at opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
Jan-Aage Frydenbø-Bruvoll
2011-Dec-18 22:47 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
Hi Craig, On Sun, Dec 18, 2011 at 22:33, Craig Morgan <crgmrgn at gmail.com> wrote:> Try fmdump -e and then fmdump -eV, it could be a pathological disk just this side of failure doing heavy retries that id dragging the pool down.Thanks for the hint - didn''t know about fmdump. Nothing in the log since 13 Dec, though - when the previous failed disk was replaced. Best regards Jan
Jan-Aage Frydenbø-Bruvoll
2011-Dec-18 22:51 UTC
[zfs-discuss] Very poor pool performance - no zfs/controllererrors?!
Hi, On Sun, Dec 18, 2011 at 22:38, Matt Breitbach <matthewb at flash.shanje.com> wrote:> I''d look at iostat -En. ?It will give you a good breakdown of disks that > have seen errors. ?I''ve also spotted failing disks just by watching an > iostat -nxz and looking for the one who''s spending more %busy than the rest > of them, or exhibiting longer than normal service times.Thanks for that - I''ve been looking at iostat output for a while without being able to make proper sense of it, i.e. it doesn''t really look that weird. However, on a side note - would you happen to know whether it is possible to reset the error counter for a particular device? In the output I get this: c1t29d0 Soft Errors: 0 Hard Errors: 5395 Transport Errors: 5394 And c1t29d0 is actually a striped pair of disks where one disk failed recently. (c1t29d0 has a mirror on the zpool level - the reason for this weird config was running out of vdevs on the controller). The device should be just fine now - the counter has stopped incrementing - however having that number there is confusing for debugging. Best regards Jan
Jan-Aage Frydenbø-Bruvoll
2011-Dec-18 22:53 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
On Sun, Dec 18, 2011 at 22:14, Nathan Kroenert <nathan at tuneunix.com> wrote:> Do you realise that losing a single disk in that pool could pretty much > render the whole thing busted?Ah - didn''t pick up on that one until someone here pointed it out - all my disks are mirrored, however some of them are mirrored on the controller level. Best regards Jan
Hung-Sheng Tsao (laoTsao)
2011-Dec-19 11:51 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
what is the ram size? are there many snap? create then delete? did you run a scrub? Sent from my iPad On Dec 18, 2011, at 10:46, Jan-Aage Frydenb?-Bruvoll <jan at architechs.eu> wrote:> Hi, > > On Sun, Dec 18, 2011 at 15:13, "Hung-Sheng Tsao (Lao Tsao ??) Ph.D." > <laotsao at gmail.com> wrote: >> what are the output of zpool status pool1 and pool2 >> it seems that you have mix configuration of pool3 with disk and mirror > > The other two pools show very similar outputs: > > root at stor:~# zpool status pool1 > pool: pool1 > state: ONLINE > scan: resilvered 1.41M in 0h0m with 0 errors on Sun Dec 4 17:42:35 2011 > config: > > NAME STATE READ WRITE CKSUM > pool1 ONLINE 0 0 0 > mirror-0 ONLINE 0 0 0 > c1t12d0 ONLINE 0 0 0 > c1t13d0 ONLINE 0 0 0 > mirror-1 ONLINE 0 0 0 > c1t24d0 ONLINE 0 0 0 > c1t25d0 ONLINE 0 0 0 > mirror-2 ONLINE 0 0 0 > c1t30d0 ONLINE 0 0 0 > c1t31d0 ONLINE 0 0 0 > mirror-3 ONLINE 0 0 0 > c1t32d0 ONLINE 0 0 0 > c1t33d0 ONLINE 0 0 0 > logs > mirror-4 ONLINE 0 0 0 > c2t2d0p6 ONLINE 0 0 0 > c2t3d0p6 ONLINE 0 0 0 > cache > c2t2d0p10 ONLINE 0 0 0 > c2t3d0p10 ONLINE 0 0 0 > > errors: No known data errors > root at stor:~# zpool status pool2 > pool: pool2 > state: ONLINE > scan: scrub canceled on Wed Dec 14 07:51:50 2011 > config: > > NAME STATE READ WRITE CKSUM > pool2 ONLINE 0 0 0 > mirror-0 ONLINE 0 0 0 > c1t14d0 ONLINE 0 0 0 > c1t15d0 ONLINE 0 0 0 > mirror-1 ONLINE 0 0 0 > c1t18d0 ONLINE 0 0 0 > c1t19d0 ONLINE 0 0 0 > mirror-2 ONLINE 0 0 0 > c1t20d0 ONLINE 0 0 0 > c1t21d0 ONLINE 0 0 0 > mirror-3 ONLINE 0 0 0 > c1t22d0 ONLINE 0 0 0 > c1t23d0 ONLINE 0 0 0 > logs > mirror-4 ONLINE 0 0 0 > c2t2d0p7 ONLINE 0 0 0 > c2t3d0p7 ONLINE 0 0 0 > cache > c2t2d0p11 ONLINE 0 0 0 > c2t3d0p11 ONLINE 0 0 0 > > The affected pool does indeed have a mix of straight disks and > mirrored disks (due to running out of vdevs on the controller), > however it has to be added that the performance of the affected pool > was excellent until around 3 weeks ago, and there have been no > structural changes nor to the pools neither to anything else on this > server in the last half year or so. > > -jan
Jan-Aage Frydenbø-Bruvoll
2011-Dec-19 11:58 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
Hi, 2011/12/19 Hung-Sheng Tsao (laoTsao) <laotsao at gmail.com>:> what is the ram size?32 GB> are there many snap? create then delete?Currently, there are 36 snapshots on the pool - it is part of a fairly normal backup regime of snapshots every 5 min, hour, day, week and month.> did you run a scrub?Yes, as part of the previous drive failure. Nothing reported there. Now, interestingly - I deleted two of the oldest snapshots yesterday, and guess what - the performance went back to normal for a while. Now it is severely dropping again - after a good while on 1.5-2GB/s I am again seeing write performance in the 1-10MB/s range. Is there an upper limit on the number of snapshots on a ZFS pool? Best regards Jan
Jim Klimov
2011-Dec-19 13:10 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
2011-12-19 2:00, Fajar A. Nugraha wrote:> From http://www.solarisinternals.com/wiki/index.php/ZFS_Best_Practices_Guide > (or at least Google''s cache of it, since it seems to be inaccessible > now: > > " > Keep pool space under 80% utilization to maintain pool performance. > Currently, pool performance can degrade when a pool is very full and > file systems are updated frequently, such as on a busy mail server. > Full pools might cause a performance penalty, but no other issues. If > the primary workload is immutable files (write once, never remove), > then you can keep a pool in the 95-96% utilization range. Keep in mind > that even with mostly static content in the 95-96% range, write, read, > and resilvering performance might suffer. > " >This reminds me that I had a question :) If I were to "reserve" space on a pool by creating a dataset with a reservation totalling, say, 20% of all pool size - but otherwise keep this dataset empty - would it help the pool to maintain performance until the rest of the pool is 100% full (or the said 80% of total pool size)? Technically the pool would always have large empty slabs, but would be forbidden to write more than 80% of pool size... Basically this should be equivalent for "root-reserved 5%" on traditional FSes like UFS, EXT3, etc. Would it be indeed? Thanks, //Jim
Gary Mills
2011-Dec-19 13:15 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
On Mon, Dec 19, 2011 at 11:58:57AM +0000, Jan-Aage Frydenb?-Bruvoll wrote:> > 2011/12/19 Hung-Sheng Tsao (laoTsao) <laotsao at gmail.com>: > > did you run a scrub? > > Yes, as part of the previous drive failure. Nothing reported there. > > Now, interestingly - I deleted two of the oldest snapshots yesterday, > and guess what - the performance went back to normal for a while. Now > it is severely dropping again - after a good while on 1.5-2GB/s I am > again seeing write performance in the 1-10MB/s range.That behavior is a symptom of fragmentation. Writes slow down dramatically when there are no contiguous blocks available. Deleting a snapshot provides some of these, but only temporarily. -- -Gary Mills- -refurb- -Winnipeg, Manitoba, Canada-
Jim Klimov
2011-Dec-19 13:20 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
2011-12-19 2:53, Jan-Aage Frydenb?-Bruvoll ?????:> On Sun, Dec 18, 2011 at 22:14, Nathan Kroenert<nathan at tuneunix.com> wrote: >> Do you realise that losing a single disk in that pool could pretty much >> render the whole thing busted? > > Ah - didn''t pick up on that one until someone here pointed it out - > all my disks are mirrored, however some of them are mirrored on the > controller level.The problem somewhat remains: it is unknown (to us and to ZFS) *how* the disks are mirrored by hardware. For example, if a single-sector error exists, would the controller detect it quickly? Would it choose the good copy correctly or use the "first disk" blindly, for the lack of other clues? Many RAID controllers are relatively dumb in what they do, and if an error does get detected, the whole problematic disk is overwritten. This is long, error-prone (if the other disk in the pair is also imperfect), and has a tendency to ignore small errors - such as those detected by ZFS with its per-block checksums. So, in case of one HW disk having an error, you might be having random data presented by the HW mirror. Since in your case ZFS is only used to stripe over HW mirrors, it has no redundancy to intelligently detect and fix such "small errors". And depending on the error''s location in the block tree, the problem might range from ignorable to fatal. //Jim
Hung-Sheng Tsao (laoTsao)
2011-Dec-19 13:29 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
not sure oi support shadow migration or you may be to send zpool to another server then send back to do defrag regards Sent from my iPad On Dec 19, 2011, at 8:15, Gary Mills <gary_mills at fastmail.fm> wrote:> On Mon, Dec 19, 2011 at 11:58:57AM +0000, Jan-Aage Frydenb?-Bruvoll wrote: >> >> 2011/12/19 Hung-Sheng Tsao (laoTsao) <laotsao at gmail.com>: >>> did you run a scrub? >> >> Yes, as part of the previous drive failure. Nothing reported there. >> >> Now, interestingly - I deleted two of the oldest snapshots yesterday, >> and guess what - the performance went back to normal for a while. Now >> it is severely dropping again - after a good while on 1.5-2GB/s I am >> again seeing write performance in the 1-10MB/s range. > > That behavior is a symptom of fragmentation. Writes slow down > dramatically when there are no contiguous blocks available. Deleting > a snapshot provides some of these, but only temporarily. > > -- > -Gary Mills- -refurb- -Winnipeg, Manitoba, Canada-
Gregg Wonderly
2011-Dec-20 00:03 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
On 12/18/2011 4:23 PM, Jan-Aage Frydenb?-Bruvoll wrote:> Hi, > > On Sun, Dec 18, 2011 at 22:14, Nathan Kroenert<nathan at tuneunix.com> wrote: >> I know some others may already have pointed this out - but I can''t see it >> and not say something... >> >> Do you realise that losing a single disk in that pool could pretty much >> render the whole thing busted? >> >> At least for me - the rate at which _I_ seem to lose disks, it would be >> worth considering something different ;) > Yeah, I have thought that thought myself. I am pretty sure I have a > broken disk, however I cannot for the life of me find out which one. > zpool status gives me nothing to work on, MegaCli reports that all > virtual and physical drives are fine, and iostat gives me nothing > either. > > What other tools are there out there that could help me pinpoint > what''s going on? >One choice would be to take a single drive that you believe is in good working condition, and add it as a mirror to each single disk in turn. If there is a bad disk, you will find out if the mirror fails because of a read error. Scrub, though, should really be telling you everything that you need to know about disk failings, once the surface becomes corrupted enough that it can''t be corrected by re-reading enough times. It looks like you''ve started mirroring some of the drives. That''s really what you should be doing for the other non-mirror drives. Gregg Wonderly
Richard Elling
2011-Dec-20 02:00 UTC
[zfs-discuss] Very poor pool performance - no zfs/controller errors?!
comments below? On Dec 18, 2011, at 6:53 AM, Jan-Aage Frydenb?-Bruvoll wrote:> Dear List, > > I have a storage server running OpenIndiana with a number of storage > pools on it. All the pools'' disks come off the same controller, and > all pools are backed by SSD-based l2arc and ZIL. Performance is > excellent on all pools but one, and I am struggling greatly to figure > out what is wrong. > > A very basic test shows the following - pretty much typical > performance at the moment: > > root at stor:/# for a in pool1 pool2 pool3; do dd if=/dev/zero of=$a/file > bs=1M count=10; done > 10+0 records in > 10+0 records out > 10485760 bytes (10 MB) copied, 0.00772965 s, 1.4 GB/s > 10+0 records in > 10+0 records out > 10485760 bytes (10 MB) copied, 0.00996472 s, 1.1 GB/s > 10+0 records in > 10+0 records out > 10485760 bytes (10 MB) copied, 71.8995 s, 146 kB/sEnable compression and they should all go fast :-) But seriously, you could be getting tripped up by the allocator. There are several different allocator algorithms and they all begin to thrash at high utilization. Some are better than others for various cases. For OpenIndiana, you might be getting bit by the allocator. One troubleshooting tip would be to observe the utilization of the metaslabs: zdb -m pool3 If there are metaslabs that are > 96% full, then look more closely at the allocator algorithms.> > The zpool status of the affected pool is: > > root at stor:/# zpool status pool3 > pool: pool3 > state: ONLINE > scan: resilvered 222G in 24h2m with 0 errors on Wed Dec 14 15:20:11 2011 > config: > > NAME STATE READ WRITE CKSUM > pool3 ONLINE 0 0 0 > c1t0d0 ONLINE 0 0 0 > c1t1d0 ONLINE 0 0 0 > c1t2d0 ONLINE 0 0 0 > c1t3d0 ONLINE 0 0 0 > c1t4d0 ONLINE 0 0 0 > c1t5d0 ONLINE 0 0 0 > c1t6d0 ONLINE 0 0 0 > c1t7d0 ONLINE 0 0 0 > c1t8d0 ONLINE 0 0 0 > c1t9d0 ONLINE 0 0 0 > c1t10d0 ONLINE 0 0 0 > mirror-12 ONLINE 0 0 0 > c1t26d0 ONLINE 0 0 0 > c1t27d0 ONLINE 0 0 0 > mirror-13 ONLINE 0 0 0 > c1t28d0 ONLINE 0 0 0 > c1t29d0 ONLINE 0 0 0 > mirror-14 ONLINE 0 0 0 > c1t34d0 ONLINE 0 0 0 > c1t35d0 ONLINE 0 0 0 > logs > mirror-11 ONLINE 0 0 0 > c2t2d0p8 ONLINE 0 0 0 > c2t3d0p8 ONLINE 0 0 0 > cache > c2t2d0p12 ONLINE 0 0 0 > c2t3d0p12 ONLINE 0 0 0 > > errors: No known data errors > > Ditto for the disk controller - MegaCli reports zero errors, be that > on the controller itself, on this pool''s disks or on any of the other > attached disks. > > I am pretty sure I am dealing with a disk-based problem here, i.e. a > flaky disk that is "just" slow without exhibiting any actual data > errors, holding the rest of the pool back, but I am at a miss as how > to pinpoint what is going on."iostat -x" shows the average service time of each disk. If one disk or set of disks is a lot slower, when also busy, then it should be clearly visible in the iostat output. Personally, I often use something like "iostat -zxCn 10" for 10-second samples. -- richard> > Would anybody on the list be able to give me any pointers as how to > dig up more detailed information about the pool''s/hardware''s > performance? > > Thank you in advance for your kind assistance. > > Best regards > Jan > _______________________________________________ > zfs-discuss mailing list > zfs-discuss at opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss-- ZFS and performance consulting http://www.RichardElling.com LISA ''11, Boston, MA, December 4-9