Dear all,
I am setting up a big scale GlusterFS with tow servers and about 400
clients.
The problems taht I encountered is every time I log in client, it takes 10
to 30 secs as well as ls command(even not in share dirs).
Could any help to increase the performance? Thanks.
Environment:
Servers:
Two RadHat 5 x86_64 machines.
Each shares 2TB storage making RAID 1 for HA.
Clients:
400 machines with Two RadHat 5 x86_64OS and install fuse-2.6.3-3.slc5.
I have tried this guide, it doesn't work. (
http://www.mail-archive.com/gluster-devel at nongnu.org/msg01985.html )
Files:
In gluster01 already have 1T files and gluster02 not really sync yet.
I have tried close gluster02 to prevent sync latency, but it still takes
about 5 to 10 secs.
Most of files are less than 10M but been read really frequently.
Configs:
gluster01-export.vol
## file auto generated by /usr/bin/glusterfs-volgen (export.vol)
# Cmd line:
# $ /usr/bin/glusterfs-volgen --name gluster --raid 1
gluster01:/opt/exp_soft/ gluster02:/opt/exp_soft/
volume posix1
  type storage/posix
  option directory /opt/exp_soft/
end-volume
volume locks1
    type features/locks
    subvolumes posix1
end-volume
volume brick1
    type performance/io-threads
    option thread-count 16
    subvolumes locks1
end-volume
volume server-tcp
    type protocol/server
    option transport-type tcp
    option auth.addr.brick1.allow *
    option transport.socket.bind-address MY-IP-1
    option transport.socket.listen-port 6996
    option transport.socket.nodelay on
    subvolumes brick1
end-volume
===================================================================================gluster02-export.vol
gluster01-export.vol
## file auto generated by /usr/bin/glusterfs-volgen (export.vol)
# Cmd line:
# $ /usr/bin/glusterfs-volgen --name gluster --raid 1
gluster01:/opt/exp_soft/ gluster02:/opt/exp_soft/
volume posix1
  type storage/posix
  option directory /opt/exp_soft/
end-volume
volume locks1
    type features/locks
    subvolumes posix1
end-volume
volume brick1
    type performance/io-threads
    option thread-count 16
    subvolumes locks1
end-volume
volume server-tcp
    type protocol/server
    option transport-type tcp
    option auth.addr.brick1.allow *
    option transport.socket.bind-address MY-IP-2
    option transport.socket.listen-port 6996
    option transport.socket.nodelay on
    subvolumes brick1
end-volume
===================================================================================Clients:
glusterfs.vol
## file auto generated by /usr/bin/glusterfs-volgen (mount.vol)
# Cmd line:
# $ /usr/bin/glusterfs-volgen --name gluster --raid 1
gluster01:/opt/exp_soft/ gluster02:/opt/exp_soft/
# RAID 1
# TRANSPORT-TYPE tcp
volume gluster02-1
    type protocol/client
    option transport-type tcp
    option remote-host gluster02
    option transport.socket.nodelay on
    option transport.remote-port 6996
    option remote-subvolume brick1
end-volume
volume gluster01-1
    type protocol/client
    option transport-type tcp
    option remote-host gluster01
    option transport.socket.nodelay on
    option transport.remote-port 6996
    option remote-subvolume brick1
end-volume
volume mirror-0
    type cluster/replicate
    subvolumes gluster01-1 gluster02-1
end-volume
volume readahead
    type performance/read-ahead
    option page-count 4
    subvolumes mirror-0
end-volume
volume iocache
    type performance/io-cache
    option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed
's/[^0-9]//g') / 5120 ))`MB
    option cache-timeout 1
    subvolumes readahead
end-volume
volume quickread
    type performance/quick-read
    option cache-timeout 1
    option max-file-size 64kB
    subvolumes iocache
end-volume
volume writebehind
    type performance/write-behind
    option cache-size 4MB
    subvolumes quickread
end-volume
volume statprefetch
    type performance/stat-prefetch
    subvolumes writebehind
end-volume
BR,
Todd