Strahil
2019-Oct-26 09:00 UTC
[Gluster-users] distributed replicated volume - added bricks
Hi Herb, First check your data usage, as you need free space on the remaining bricks. You have to remove bricks multiple per replica set. For replica 2 distributed volume you must specify 2 bricks at least , or any number multiple by 2 (2,4,6,8 bricks). The removal consists of several actions . 1. Starting removal via : gluster volume remove-brick <VOLNAME> <BRICKNAME> start This actually only migrates the data to the rest of the bricks. 2. Check status via: gluster volume remove-brick <VOLNAME> <BRICKNAME> status Once completed, check the logs for any rebalance issuues. You can verify data/inode count on the brick via: gluster volume status <VOLNAME> detail If everything is fine proceed with the actual removal: # gluster volume remove-brick <VOLNAME> <BRICKNAME> commit 5. Last verify the volume status. Source: https://docs.gluster.org/en/latest/Administrator%20Guide/Managing%20Volumes/#shrinking-volumes Best Regards, Strahil Nikolov On Oct 25, 2019 20:26, Herb Burnswell <herbert.burnswell at gmail.com> wrote:> > Thank you for the reply Strahil. > > Unfortunately we did do the rebalance already so the data should?be written across all brinks currently.? I'm fine with pulling these newly added bricks out of the volume.? However, is it as simple as pulling them out and the data will rebalance to the disks that are left? > > Thanks, > > HB? > > On Sat, Oct 19, 2019 at 4:13 PM Strahil Nikolov <hunter86_bg at yahoo.com> wrote: >> >> Most probably this means that data on? >> Brick server1:/gluster_bricks/data3 ? ? ? 49164 ? ? 0 ? ? ? ? ?Y ? ? ? 4625? >> Brick server1:/gluster_bricks/data4 ? ? ? 49165 ? ? 0 ? ? ? ? ?Y ? ? ? 4644 >> >> is the same and when server1 goes down , you will have no access to the data on this set. >> Same should be valid for : >> Brick server1:/gluster_bricks/data5 ? ? ? 49166 ? ? 0 ? ? ? ? ?Y ? ? ? 5088? >> Brick server1:/gluster_bricks/data6 ? ? ? 49167 ? ? 0 ? ? ? ? ?Y ? ? ? 5128? >> Brick server2:/gluster_bricks/data3 ? ? ? 49168 ? ? 0 ? ? ? ? ?Y ? ? ? 22314 >> Brick server2:/gluster_bricks/data4 ? ? ? 49169 ? ? 0 ? ? ? ? ?Y ? ? ? 22345 >> Brick server2:/gluster_bricks/data5 ? ? ? 49170 ? ? 0 ? ? ? ? ?Y ? ? ? 22889 >> Brick server2:/gluster_bricks/data6 ? ? ? 49171 ? ? 0 ? ? ? ? ?Y ? ? ? 22932 >> >> I would remove those bricks and add them again this type always specifying one bri-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.gluster.org/pipermail/gluster-users/attachments/20191026/4e2a304e/attachment.html>