chaica at ohmytux.com
2011-Jul-01 09:53 UTC
[Gluster-users] Issues with adding a volume with glusterfs 3.2.1
Hi, I can't succeed adding two new replicated bricks in my volume. I do have a replicated volume on two servers. I created it with the following commands : root at glusterfs1:/var/log# gluster peer probe 192.168.1.31 Probe successful root at glusterfs1:~# gluster volume create test-volume replica 2 transport tcp 192.168.1.30:/sharedspace 192.168.1.31:/sharedspace Creation of volume test-volume has been successful. Please start the volume to access data. root at glusterfs1:~# gluster volume start test-volume root at glusterfs1:~# gluster volume info Volume Name: test-volume Type: Replicate Status: Started Number of Bricks: 2 Transport-type: tcp Bricks: Brick1: 192.168.1.30:/sharedspace Brick2: 192.168.1.31:/sharedspace Ok now from a client : client~# mount -t glusterfs 192.168.1.30:test-volume /distributed-volume client~# echo "hello" > /distributed-volume/foo Foo is well replicated across the two bricks : root at glusterfs1:~# ll /sharedspace/ total 28 -rw-r--r-- 1 root root 26 1 juil. 10:29 foo drwx------ 2 root root 16384 11 janv. 12:18 lost+found root at glusterfs2:~# ll /sharedspace/ total 28 -rw-r--r-- 1 root root 26 1 juil. 10:29 foo drwx------ 2 root root 16384 11 janv. 12:18 lost+found That's perfect... until I try to add two new bricks : root at glusterfs1:~# gluster peer probe 192.168.1.32 Probe successful root at glusterfs1:~# gluster peer probe 192.168.1.33 Probe successful root at glusterfs1:~# gluster volume add-brick test-volume 192.168.1.32:/sharedspace 192.168.1.33:/sharedspace Add Brick successful root at glusterfs1:~# gluster volume info Volume Name: test-volume Type: Distributed-Replicate Status: Started Number of Bricks: 2 x 2 = 4 Transport-type: tcp Bricks: Brick1: 192.168.1.30:/sharedspace Brick2: 192.168.1.31:/sharedspace Brick3: 192.168.1.32:/sharedspace Brick4: 192.168.1.33:/sharedspace root at glusterfs1:~# gluster volume rebalance test-volume start starting rebalance on volume test-volume has been successful root at glusterfs1:~# gluster volume rebalance test-volume status rebalance completed Ok, so If I'm correct, data on Brick1 and Brick2 should be now also available from Brick3 and Brick4. But that's not the case : root at glusterfs3:~# ll /sharedspace/ total 20 drwx------ 2 root root 16384 11 janv. 12:18 lost+found root at glusterfs4:~# ll /sharedspace/ total 20 drwx------ 2 root root 16384 11 janv. 12:18 lost+found I'm using Debian Sid with Glusterfs 3.2.1 from official Debian repository. Am I wrong somewhere ? Regards, Carl Chenet
Burnash, James
2011-Jul-01 13:01 UTC
[Gluster-users] Issues with adding a volume with glusterfs 3.2.1
Hi Carl. I was similarly confused at first when I did the exact same thing as you - adding two additional server/bricks to an existing Distributed-Replicate volume. Here's the thing - you specified "replica 2" when you created the volume - that means that two replicas will be kept, and only two. Going forward after adding bricks 3 and 4, those replicas will on either the first pair of bricks (Brick1 and Brick2) or the second pair of bricks (Brick3 and Brick4). What happens when you add bricks to a Distributed-Replicate volume such as yours is that the Distribute part kicks in, and "distributes" the files amongst the mirror (replica) pairs given above. Clear as mud? :-) James Burnash Unix Engineer Knight Capital Group -----Original Message----- From: gluster-users-bounces at gluster.org [mailto:gluster-users-bounces at gluster.org] On Behalf Of chaica at ohmytux.com Sent: Friday, July 01, 2011 5:54 AM To: gluster-users at gluster.org Subject: [Gluster-users] Issues with adding a volume with glusterfs 3.2.1 Hi, I can't succeed adding two new replicated bricks in my volume. I do have a replicated volume on two servers. I created it with the following commands : root at glusterfs1:/var/log# gluster peer probe 192.168.1.31 Probe successful root at glusterfs1:~# gluster volume create test-volume replica 2 transport tcp 192.168.1.30:/sharedspace 192.168.1.31:/sharedspace Creation of volume test-volume has been successful. Please start the volume to access data. root at glusterfs1:~# gluster volume start test-volume root at glusterfs1:~# gluster volume info Volume Name: test-volume Type: Replicate Status: Started Number of Bricks: 2 Transport-type: tcp Bricks: Brick1: 192.168.1.30:/sharedspace Brick2: 192.168.1.31:/sharedspace Ok now from a client : client~# mount -t glusterfs 192.168.1.30:test-volume /distributed-volume client~# echo "hello" > /distributed-volume/foo Foo is well replicated across the two bricks : root at glusterfs1:~# ll /sharedspace/ total 28 -rw-r--r-- 1 root root 26 1 juil. 10:29 foo drwx------ 2 root root 16384 11 janv. 12:18 lost+found root at glusterfs2:~# ll /sharedspace/ total 28 -rw-r--r-- 1 root root 26 1 juil. 10:29 foo drwx------ 2 root root 16384 11 janv. 12:18 lost+found That's perfect... until I try to add two new bricks : root at glusterfs1:~# gluster peer probe 192.168.1.32 Probe successful root at glusterfs1:~# gluster peer probe 192.168.1.33 Probe successful root at glusterfs1:~# gluster volume add-brick test-volume 192.168.1.32:/sharedspace 192.168.1.33:/sharedspace Add Brick successful root at glusterfs1:~# gluster volume info Volume Name: test-volume Type: Distributed-Replicate Status: Started Number of Bricks: 2 x 2 = 4 Transport-type: tcp Bricks: Brick1: 192.168.1.30:/sharedspace Brick2: 192.168.1.31:/sharedspace Brick3: 192.168.1.32:/sharedspace Brick4: 192.168.1.33:/sharedspace root at glusterfs1:~# gluster volume rebalance test-volume start starting rebalance on volume test-volume has been successful root at glusterfs1:~# gluster volume rebalance test-volume status rebalance completed Ok, so If I'm correct, data on Brick1 and Brick2 should be now also available from Brick3 and Brick4. But that's not the case : root at glusterfs3:~# ll /sharedspace/ total 20 drwx------ 2 root root 16384 11 janv. 12:18 lost+found root at glusterfs4:~# ll /sharedspace/ total 20 drwx------ 2 root root 16384 11 janv. 12:18 lost+found I'm using Debian Sid with Glusterfs 3.2.1 from official Debian repository. Am I wrong somewhere ? Regards, Carl Chenet _______________________________________________ Gluster-users mailing list Gluster-users at gluster.org http://gluster.org/cgi-bin/mailman/listinfo/gluster-users DISCLAIMER: This e-mail, and any attachments thereto, is intended only for use by the addressee(s) named herein and may contain legally privileged and/or confidential information. If you are not the intended recipient of this e-mail, you are hereby notified that any dissemination, distribution or copying of this e-mail, and any attachments thereto, is strictly prohibited. If you have received this in error, please immediately notify me and permanently delete the original and any copy of any e-mail and any printout thereof. E-mail transmission cannot be guaranteed to be secure or error-free. The sender therefore does not accept liability for any errors or omissions in the contents of this message which arise as a result of e-mail transmission. NOTICE REGARDING PRIVACY AND CONFIDENTIALITY Knight Capital Group may, at its discretion, monitor and review the content of all e-mail communications. http://www.knight.com