Hello, first time posting. I''ve been working with zfs on and off with
limited *nix experience for a year or so now, and have read a lot of things by a
lot of you I''m sure. Still tons I don''t understand/know
I''m sure.
We''ve been having awful IO latencies on our 7210 running about 40
VM''s spread over 3 hosts, no SSD''s / Intent Logs. I am trying
to get some, but the price... so I had to work up some sort of PoC to show it
would help. It so happens I just purchased 3, X25-M''s for my own use,
and could spare one for a few weeks (though I hate to think how many cells I
burned testing), and we also happen to have a couple home built small ZFS
servers around to test with.
Pretty limited resources, we have a home built box with 5, 250gb 7200rpm sata
disks, each connected to the Intel "Server board''s" built in
sata ports. I reduced the RAM to 2GB for the tests. The OS is on a sata disk
in its own single disk pool. The X25-M was used as a ZIL Log for the SSD Tests.
I created 5 VM''s on a single ESX Host (Dell 1950) with a Data Store
connected to the "mini-thumper" running 2009.06 snv_111b via NFS over
a single GB link.
Each VM runs Windows 2003 R2 on a 4GB C:\ vmdk. Dynamo runs 1 worker on the
local C: vmdk on each guest and reports to my workstation, so the numbers below
are totals of the dynamo on all 5 guests.
Each test consisted of an 8k transfer with a 67% read, and 70% random pattern.
The tests were run for 5 minutes each.
Queue Depth, IOPS, Avg Latency (ms)
RAID0 - 5 Disk
1, 326, 15.3
2, 453, 22
4, 516, 38.7
8, 503, 72.3
16, 526, 152
32, 494, 323
RAID0-4 Disk +SSD
1, 494, 10.1
2, 579, 17.2
4, 580, 34.4
8, 603, 66.3
16, 598, 133.6
32, 600, 266
RAIDz - 5 Disk
1, 144, 34
2, 162, 60
4, 184, 108
8, 183, 218
16, 175, 455
32, 185, 864
RAIDz - 4 Disk +SSD
1, 222, 22
2, 201, 50
4, 221, 90
8, 219, 181
16, 228, 348
32, 228, 700
RAID10 - 4 Disk
1, 159, 31
2, 206, 48
4, 236, 84
8, 194, 205
16, 243, 328
32, 219, 728
RAID10 - 4 Disk +SSD
1, 270, 18
2, 332, 30
4, 363, 54
8, 320, 124
16, 325, 245
32, 333, 479
(wonders how the formatting will turn out)
Its interesting that going from a 5 disk RAIDz to a 4 disk Mirror (both with no
log device) has a bigger increase then using X25-M Log with a 4 disk RAIDz.
The increase in IO''s adding the X25-M to the Mirror setup is nice, but
smaller then I had expected, but the halving of the latencies is even nicer.
I am curious how this would scale with a lot more disks, the SSD didn''t
increase performance as much as I had hoped, but its still nice to see...
I''m thinking that''s mostly due to my limit of 4-5 disks.
I''m not sure how much difference there is between the X25-M and the SUN
SSD''s for the 7000 series.
>From what I''ve read so far the X25-E needs to have its write-cache
forced off to function proper, where the X25-M seems to obey the flush commands?
I was also curious if I would have seen a bigger increase with an SLC drive
instead of the MLC... searching turns up so much old info.
Comments welcome!
--
This message posted from opensolaris.org