Łukasz Zygmański
2015-Jun-16 13:56 UTC
[Gluster-users] GlusterFS replica synchronization problem
Hello, Could you please tell me what should I do to enable/fix synchronous replication between two GlusterFS nodes, because at that time my files are synching about every 7 mins. Here is my configuration: Gluster01 server = 10.75.3.43 (and also 10.75.2.41 for clients) Gluster02 server = 10.75.3.44 (and also 10.75.2.42 for clients) [root at gluster01 vins]# gluster volume status Status of volume: testapi_vol Gluster process TCP Port RDMA Port Online Pid ------------------------------------------------------------------------------ Brick gluster01.int:/glusterfs/testapi/bric k 49153 0 Y 2693 Brick gluster02.int:/glusterfs/testapi/bric k 49153 0 Y 5214 NFS Server on localhost 2049 0 Y 3388 Self-heal Daemon on localhost N/A N/A Y 3396 NFS Server on gluster02.int 2049 0 Y 6468 Self-heal Daemon on gluster02.int N/A N/A Y 6476 Task Status of Volume testapi_vol ------------------------------------------------------------------------------ There are no active volume tasks [root at gluster02 vins]# gluster volume status Status of volume: testapi_vol Gluster process TCP Port RDMA Port Online Pid ------------------------------------------------------------------------------ Brick gluster01.int:/glusterfs/testapi/bric k 49153 0 Y 2693 Brick gluster02.int:/glusterfs/testapi/bric k 49153 0 Y 5214 NFS Server on localhost 2049 0 Y 6468 Self-heal Daemon on localhost N/A N/A Y 6476 NFS Server on 10.75.3.43 2049 0 Y 3388 Self-heal Daemon on 10.75.3.43 N/A N/A Y 3396 Task Status of Volume testapi_vol ------------------------------------------------------------------------------ There are no active volume tasks [root at gluster01 vins]# gluster volume info Volume Name: testapi_vol Type: Replicate Volume ID: 7800a682-1f07-4464-bf7c-e1aba11f5190 Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: gluster01.int:/glusterfs/testapi/brick Brick2: gluster02.int:/glusterfs/testapi/brick [root at gluster02 vins]# gluster volume info Volume Name: testapi_vol Type: Replicate Volume ID: 7800a682-1f07-4464-bf7c-e1aba11f5190 Status: Started Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: gluster01.int:/glusterfs/testapi/brick Brick2: gluster02.int:/glusterfs/testapi/brick [root at gluster01 vins]# ip a 2: eno16780032: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP qlen 1000 link/ether 00:50:56:b2:11:15 brd ff:ff:ff:ff:ff:ff inet 10.75.2.41/24 brd 10.75.2.255 scope global eno16780032 valid_lft forever preferred_lft forever 3: eno33559296: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP qlen 1000 link/ether 00:50:56:b2:7f:d4 brd ff:ff:ff:ff:ff:ff inet 10.75.3.43/24 brd 10.75.3.255 scope global eno33559296 valid_lft forever preferred_lft forever [root at gluster02 vins]# ip a 2: eno16780032: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP qlen 1000 link/ether 00:50:56:b2:4b:5c brd ff:ff:ff:ff:ff:ff inet 10.75.2.42/24 brd 10.75.2.255 scope global eno16780032 valid_lft forever preferred_lft forever 3: eno33559296: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP qlen 1000 link/ether 00:50:56:b2:6c:8e brd ff:ff:ff:ff:ff:ff inet 10.75.3.44/24 brd 10.75.3.255 scope global eno33559296 valid_lft forever preferred_lft forever [root at gluster01 vins]# iptables -nxvL | grep 3.44 10 600 ACCEPT all -- * * 10.75.3.44 0.0.0.0/0 [root at gluster02 vins]# iptables -nxvL | grep 3.43 101 7398 ACCEPT all -- * * 10.75.3.43 0.0.0.0/0 [root at gluster01 vins]# cat /etc/hosts | grep 10.75.3 10.75.3.43 gluster01.int gluster01.int.uci.umk.pl 10.75.3.44 gluster02.int gluster02.int.uci.umk.pl [root at gluster02 vins]# cat /etc/hosts | grep 10.75.3 10.75.3.43 gluster01.int gluster01.int.uci.umk.pl 10.75.3.44 gluster02.int gluster02.int.uci.umk.pl Mount on the client: gluster01.int:/testapi_vol on /mnt/glusterfs type fuse.glusterfs (rw,default_permissions,allow_other,max_read=131072) # cat /etc/fstab | grep gluster gluster01.int:/testapi_vol /mnt/glusterfs glusterfs defaults,_netdev 0 0 My test (from client): # dd if=/dev/zero of=test bs=1024 count=10240 and I see the file on gluster01, but I have to wait the aforementioned about 7 minutes for the files to appear on gluster02. Could you please tell me what can I do to fix that, where should I look, is it possible? If there is more info needed please do let me know. Best regards Lukasz -- ?ukasz Zygma?ski Uczelniane Centrum Information & Communication Informatyczne Technology Centre Uniwersytet Mikolaja Kopernika Nicolaus Copernicus University Coll. Maximum, pl. Rapackiego 1, 87-100 Torun, Poland tel.: +48 56 611 27 36 fax: +48 56-622-18-50 email: Lukasz.Zygmanski at umk.pl -------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 5279 bytes Desc: Kryptograficzna sygnatura S/MIME URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150616/127b35ae/attachment.p7s>
Ravishankar N
2015-Jun-17 16:16 UTC
[Gluster-users] GlusterFS replica synchronization problem
On 06/16/2015 07:26 PM, ?ukasz Zygma?ski wrote:> > My test (from client): > # dd if=/dev/zero of=test bs=1024 count=10240 > and I see the file on gluster01, but I have to wait the aforementioned > about 7 minutes for the files to appear on gluster02.Looks like your fuse client has lost connection to gluster02 and writes only to gluster01. The self-heal daemon might later on be healing the file. Do the client logs indicate anything about being disconnected from the 2nd node? -Ravi