Hi i?m using GlusterFs V3.0.2 in Fedora 12. I have configure AFR with 3 nodes and mount the volume in the client, its work fine, but when a node fail, the file system in the client locks and I can?t execute any operations for about 40 to 50 Seconds. After 40 to 50 Seconds the File system on the client start to work again. How I can resolve this problem, because the file system can?t bee inaccessible for so long Time. Bellow is my Server Configuration and Client Configuration: -------------------------------------------------------------------------- Server Configuration: -------------------------------------------------------------------------- volume s3ap2posix type storage/posix option directory /gluster/s3ap2 end-volume volume s3ap2locks type features/locks subvolumes s3ap2posix end-volume volume s3ap2brick type performance/io-threads option thread-count 8 subvolumes s3ap2locks end-volume volume server-tcp type protocol/server option transport-type tcp option auth.addr.s3ap2brick.allow 192.168.30.* option transport.socket.listen-port 6996 option transport.socket.nodelay on subvolumes s3ap2brick end-volume -------------------------------------------------------------------------- Cliente Configuration -------------------------------------------------------------------------- volume 192.168.30.1-s3ap2 type protocol/client option transport-type tcp option remote-host 192.168.30.1 option transport.socket.nodelay on option transport.remote-port 6996 option remote-subvolume s3ap2brick end-volume volume 192.168.30.2-s3ap2 type protocol/client option transport-type tcp option remote-host 192.168.30.2 option transport.socket.nodelay on option transport.remote-port 6996 option remote-subvolume s3ap2brick end-volume volume 192.168.30.3-s3ap2 type protocol/client option transport-type tcp option remote-host 192.168.30.3 option transport.socket.nodelay on option transport.remote-port 6996 option remote-subvolume s3ap2brick end-volume volume s3ap2afr type cluster/afr subvolumes 192.168.30.1-s3ap2 192.168.30.2-s3ap2 192.168.30.3-s3ap2 end-volume volume s3ap2writebehind type performance/write-behind option cache-size 4MB subvolumes s3ap2afr end-volume volume s3ap2readahead type performance/read-ahead option page-count 4 subvolumes s3ap2writebehind end-volume volume s3ap2iocache type performance/io-cache option cache-size 1GB option cache-timeout 1 subvolumes s3ap2readahead end-volume volume s3ap2quickread type performance/quick-read option cache-timeout 1 option max-file-size 64kB subvolumes s3ap2iocache end-volume volume s3ap2 type performance/stat-prefetch subvolumes s3ap2quickread end-volume Thank?s Eduardo
Hello?Eduardo, We had the same problem over here, with two nodes that are both server and client. You can try to lower the ping-timeout in the client volume file: option ping-timeout 5 5 seconds is the sadly the lowest possible ping-timeout, our applications on the main node can hang for about 5 seconds in case the standby node fails (although we have a dedicated interconnect). Maybe the gluster developers have a better solution to this. Regards, Richard> Hi i?m using GlusterFs V3.0.2 in Fedora 12. > > I have configure AFR with 3 nodes and mount the volume in the client, its > work fine, but when a node fail, the file system in the client locks and I > can?t execute any operations for about 40 to 50 Seconds.> After 40 to 50 Seconds the File system on the client start to work again. > How I can resolve this problem, because the file system can?t bee > inaccessible for so long Time.