Hi Gang, I recently rented a server at a datacenter with Centos 5.7 X64, Q9550 Processor, 8GB Ram, and dual 250GB SATA HDs (with 16mb cache). They had loaded it with KVM, and installed a 30-day trial of Virtualizor as the front-end for KVM. I was so impressed with how fasts the guests ran that I want to build a few of these machines for myself. I just installed one: same Q9550 processor, 4GB ram, and dual 250GB SATA HDs (with 32mb cache). I installed Centos 6.2 X64, and installed Webmin's Cloudmin as the front-end. Immediately when I was installing stuff, I could tell this new system I just built was not nearly as fast as the first one. I ran some CPU and disk benchmarking programs, and saw that while the CPU stuff tested similarly, the disk thruput was much different... Down-right poor in one of the guests! On both systems, /dev/md2 is a LVM reserved exclusively for KVM guests. So each guest is running in its own logical volume, in software raid. Thinking there may be something wrong with the HDs, I ran Bonnie ( http://www.coker.com.au/bonnie++/ ) and compared both host machines. They tested fairly similar (within 10%). Yet comparing their guests is like night and day. Example: On good machine's Centos 5.7 x32 guest install: # hdparm -tT /dev/hda /dev/hda: Timing cached reads: 26760 MB in 1.99 seconds = 13417.10 MB/sec Timing buffered disk reads: 388 MB in 3.01 seconds = 128.86 MB/sec On my machine's Centos 5.7 x32 guest install: # hdparm -tT /dev/hda /dev/hda: Timing cached reads: 1864 MB in 2.16 seconds = 863.87 MB/sec Timing buffered disk reads: 358 MB in 3.08 seconds = 116.17 MB/sec On one of my machine's Mandrake 8.2 x32 guest install: # hdparm -tT /dev/hda /dev/hda: Timing buffer-cache reads: 27000 MB in 2.00 seconds = 13500.00 MB/sec Timing buffered disk reads: 12 MB in 3.66 seconds = 3.28 MB/sec On that system, the hdparm's -i output shows: # hdparm -i /dev/hda /dev/hda: Model=QEMU HARDDISK, FwRev=0.12.1, SerialNo=QM00001 Config={ Fixed } RawCHS=16383/16/63, TrkSize=32256, SectSize=512, ECCbytes=4 BuffType=DualPortCache, BuffSize=256kB, MaxMultSect=16, MultSect=16 CurCHS=16383/16/63, CurSects=16514064, LBA=yes, LBAsects=73400320 IORDY=yes, tPIO={min:120,w/IORDY:120}, tDMA={min:120,rec:120} PIO modes: pio0 pio1 pio2 DMA modes: sdma0 sdma1 sdma2 mdma0 mdma1 mdma2 UDMA modes: udma0 udma1 udma2 udma3 udma4 *udma5 AdvancedPM=no Drive conforms to: ATA/ATAPI-5 published, ANSI NCITS 340-2000: * signifies the current active mode The bonnie numbers show for sequential output: Good Machine Host: 76,857K/Sec My Machine Host: 72,561K/Sec Good Machine Centos 5.7 Guest: 66,266K/sec My Machine Centos 5.7 Guest: 20,623K/sec My machine Mandrake Guest: 1,365K/sec Where should I look? I realize I do have two different front-ends to KVM, and perhaps they are passing different parameters to it. I am also running the KVM from Centos 6.2 on my machine, vs the other server is running on 5.7, but I would have thought that "newer is better". Also note that my hard drives have a larger cache. On a side note, I'm not thrilled with the Virtualizor's tech support, but the product seems easy to use, once it actually works. Cloudmin seems to be buggy, and not let you do things like change cd images on the fly, access the console before the machine fully boots (!)... Any suggestions on other, preferably open-source options? I'm a definite newbie to this virtualization stuff. Bob
How much did you paid for this? On 20/02/2012 07:26 AM, Bob Puff wrote:> Hi Gang, > > I recently rented a server at a datacenter with Centos 5.7 X64, Q9550 > Processor, 8GB Ram, and dual 250GB SATA HDs (with 16mb cache). They had > loaded it with KVM, and installed a 30-day trial of Virtualizor as the > front-end for KVM. > > I was so impressed with how fasts the guests ran that I want to build a few of > these machines for myself. I just installed one: same Q9550 processor, 4GB > ram, and dual 250GB SATA HDs (with 32mb cache). I installed Centos 6.2 X64, > and installed Webmin's Cloudmin as the front-end. > > Immediately when I was installing stuff, I could tell this new system I just > built was not nearly as fast as the first one. I ran some CPU and disk > benchmarking programs, and saw that while the CPU stuff tested similarly, the > disk thruput was much different... Down-right poor in one of the guests! > > On both systems, /dev/md2 is a LVM reserved exclusively for KVM guests. So > each guest is running in its own logical volume, in software raid. > > Thinking there may be something wrong with the HDs, I ran Bonnie ( > http://www.coker.com.au/bonnie++/ ) and compared both host machines. They > tested fairly similar (within 10%). Yet comparing their guests is like night > and day. Example: > > On good machine's Centos 5.7 x32 guest install: > # hdparm -tT /dev/hda > > /dev/hda: > Timing cached reads: 26760 MB in 1.99 seconds = 13417.10 MB/sec > Timing buffered disk reads: 388 MB in 3.01 seconds = 128.86 MB/sec > > On my machine's Centos 5.7 x32 guest install: > # hdparm -tT /dev/hda > > /dev/hda: > Timing cached reads: 1864 MB in 2.16 seconds = 863.87 MB/sec > Timing buffered disk reads: 358 MB in 3.08 seconds = 116.17 MB/sec > > On one of my machine's Mandrake 8.2 x32 guest install: > # hdparm -tT /dev/hda > > /dev/hda: > Timing buffer-cache reads: 27000 MB in 2.00 seconds = 13500.00 MB/sec > Timing buffered disk reads: 12 MB in 3.66 seconds = 3.28 MB/sec > > On that system, the hdparm's -i output shows: > # hdparm -i /dev/hda > > /dev/hda: > > Model=QEMU HARDDISK, FwRev=0.12.1, SerialNo=QM00001 > Config={ Fixed } > RawCHS=16383/16/63, TrkSize=32256, SectSize=512, ECCbytes=4 > BuffType=DualPortCache, BuffSize=256kB, MaxMultSect=16, MultSect=16 > CurCHS=16383/16/63, CurSects=16514064, LBA=yes, LBAsects=73400320 > IORDY=yes, tPIO={min:120,w/IORDY:120}, tDMA={min:120,rec:120} > PIO modes: pio0 pio1 pio2 > DMA modes: sdma0 sdma1 sdma2 mdma0 mdma1 mdma2 > UDMA modes: udma0 udma1 udma2 udma3 udma4 *udma5 > AdvancedPM=no > Drive conforms to: ATA/ATAPI-5 published, ANSI NCITS 340-2000: > > * signifies the current active mode > > The bonnie numbers show for sequential output: > Good Machine Host: 76,857K/Sec > My Machine Host: 72,561K/Sec > > Good Machine Centos 5.7 Guest: 66,266K/sec > My Machine Centos 5.7 Guest: 20,623K/sec > My machine Mandrake Guest: 1,365K/sec > > Where should I look? I realize I do have two different front-ends to KVM, and > perhaps they are passing different parameters to it. I am also running the > KVM from Centos 6.2 on my machine, vs the other server is running on 5.7, but > I would have thought that "newer is better". Also note that my hard drives > have a larger cache. > > > On a side note, I'm not thrilled with the Virtualizor's tech support, but the > product seems easy to use, once it actually works. Cloudmin seems to be > buggy, and not let you do things like change cd images on the fly, access the > console before the machine fully boots (!)... Any suggestions on other, > preferably open-source options? I'm a definite newbie to this virtualization > stuff. > > Bob > > > _______________________________________________ > CentOS mailing list > CentOS at centos.org > http://lists.centos.org/mailman/listinfo/centos
Bob Puff writes:> /dev/hda: >It should be /dev/vda.. it means you're not taking advantage of KVM's paravirtualisation capabilities. That probably explains the crappy performance. Run your disk and network with virtio and it should be much faster. In terms of management tools, check http://www.linux-kvm.org/page/Management_Tools hth -- Nux! www.nux.ro
On 02/19/2012 09:26 PM, Bob Puff wrote:> > Immediately when I was installing stuff, I could tell this new system I just > built was not nearly as fast as the first one. I ran some CPU and disk > benchmarking programs, and saw that while the CPU stuff tested similarly, the > disk thruput was much different... Down-right poor in one of the guests!If you use Red Hat's "virt-install" to set up a guest, you can specify the --os-type and --os-variant. If you specify the a variant that supports virtualized IO (such as type: linux and variant: rhel5.4 or rhel 6), the guest's disk and network IO will have much better throughput. You'll want to look for those options in whatever front-end you're using, or use virt-install.
On Mon, Feb 20, 2012 at 02:26, Bob Puff <bob at nleaudio.com> wrote:> > Hi Gang,[...]> On my machine's Centos 5.7 x32 guest install: > # hdparm -tT /dev/hda > > /dev/hda: > ?Timing cached reads: ? 1864 MB in ?2.16 seconds = 863.87 MB/sec > ?Timing buffered disk reads: ?358 MB in ?3.08 seconds = 116.17 MB/sec"cached reads" is a measure of linux buffer bandwitdth, not disk performance!!. It should be several thousands... In a real machine, it could be a motherboard problem... -- Marcelo "?No ser? acaso que ?sta vida moderna est? teniendo m?s de moderna que de vida?" (Mafalda)