Hi
We have successfully deployed glusterfs. We have deployed two servers with one
volume as a replica. We use Citrix XenServer for our virtual environment. In
XenServer we have created a storage repository via NFS using the glusterfs
volume. This has been successful and XenServer can now use that storage
repository for storing virtual machine images.
We are finding that the storage performance for the virtual machine images being
stored across the network is too slow. I am comparing this to using a straight
Linux NFS server in the same scenario. glusterfs is very appealing as a
replacement for NFS due to its support for NFS clients and its redundancy.
Virtual machine images range from 10GB to 100GB, most are 10GB.
The two glusterfs servers are on servers with RAID 6 hardware and we get great
storage performance out of them.
Current volume configuration:
gluster volume info
Volume Name: storage-pool-01
Type: Replicate
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: scluster01.corp.assureprograms.com.au:/gluster-export/storage-pool-01
Brick2: scluster02.corp.assureprograms.com.au:/gluster-export/storage-pool-01
Options Reconfigured:
nfs.volume-access: read-write
auth.allow: 192.168.10.*
nfs.ports-insecure: on
nfs.addr-namelookup: on
server.allow-insecure: on
nfs.export-volumes: on
As you can see above we have applied some volume configuration changes like
"nfs.volume-access: read-write". Does anyone have any suggested
configurations changes that can be made to glusterfs to improve performance of
the I/O in this scenario?
Kind regards
Stewart Campbell