Lilley, John F.
2015-Mar-30 23:42 UTC
[Gluster-users] Shrinking gluster filesystem in 3.6.2
Hi, I'd like to shrink my aws/ebs-based *distribute-only* gluster file system by migrating the data to other already existing, active and partially utilized bricks but found the 'replace-brick start' mentioned in the documentation is now deprecated. I see that there has been some back and forth on the mailing list regarding migrating data using self-heal on a replicated system but not so much on a distribute-only file system. Can anyone tell me the blessed way of doing this in 3.6.2? Is there one? To be clear, all of the ebs-based bricks are partially utilized at this point so I'd need a method to migrate the data first. Thank You, John -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150330/939ebd06/attachment.html>
On 03/31/2015 05:12 AM, Lilley, John F. wrote:> Hi, > > I'd like to shrink my aws/ebs-based *distribute-only* gluster file > system by migrating the data to other already existing, active and > partially utilized bricks but found the 'replace-brick start' > mentioned in the documentation is now deprecated. I see that there > has been some back and forth on the mailing list regarding migrating > data using self-heal on a replicated system but not so much on a > distribute-only file system. Can anyone tell me the blessed way of > doing this in 3.6.2? Is there one? > > To be clear, all of the ebs-based bricks are partially utilized at > this point so I'd need a method to migrate the data first. >If I understand you correctly, you want to replace a brick in a distribute volume with one of lesser capacity. You could first add a new brick and then remove the existing brick with remove-brick start/status/commit sequence. Something like this: ------------------------------------------------ [root at tuxpad ~]# gluster volume info testvol Volume Name: testvol Type: Distribute Volume ID: a89aa154-885c-4e14-8d3a-b555733b11f1 Status: Started Number of Bricks: 3 Transport-type: tcp Bricks: Brick1: 127.0.0.2:/home/ravi/bricks/brick1 Brick2: 127.0.0.2:/home/ravi/bricks/brick2 Brick3: 127.0.0.2:/home/ravi/bricks/brick3 [root at tuxpad ~]# [root at tuxpad ~]# [root at tuxpad ~]# gluster volume add-brick testvol 127.0.0.2:/home/ravi/bricks/brick{4..6} volume add-brick: success [root at tuxpad ~]# [root at tuxpad ~]# gluster volume info testvol Volume Name: testvol Type: Distribute Volume ID: a89aa154-885c-4e14-8d3a-b555733b11f1 Status: Started Number of Bricks: 6 Transport-type: tcp Bricks: Brick1: 127.0.0.2:/home/ravi/bricks/brick1 Brick2: 127.0.0.2:/home/ravi/bricks/brick2 Brick3: 127.0.0.2:/home/ravi/bricks/brick3 Brick4: 127.0.0.2:/home/ravi/bricks/brick4 Brick5: 127.0.0.2:/home/ravi/bricks/brick5 Brick6: 127.0.0.2:/home/ravi/bricks/brick6 [root at tuxpad ~]# [root at tuxpad ~]# [root at tuxpad ~]# [root at tuxpad ~]# gluster v remove-brick testvol 127.0.0.2:/home/ravi/bricks/brick{1..3} start volume remove-brick start: success ID: d535675e-8362-4a44-a291-1e567a77531e [root at tuxpad ~]# gluster v remove-brick testvol 127.0.0.2:/home/ravi/bricks/brick{1..3} status Node Rebalanced-files size scanned failures skipped status run time in secs --------- ----------- ----------- ----------- ----------- ----------- ------------ -------------- localhost 10 0Bytes 20 0 0 *completed * 0.00 [root at tuxpad ~]# [root at tuxpad ~]# gluster v remove-brick testvol 127.0.0.2:/home/ravi/bricks/brick{1..3} commit Removing brick(s) can result in data loss. Do you want to Continue? (y/n) y volume remove-brick commit: success *Check the removed bricks to ensure all files are migrated.** **If files with data are found on the brick path, copy them via a gluster mount point before re-purposing the removed brick. * [root at tuxpad ~]# [root at tuxpad ~]# gluster volume info testvol Volume Name: testvol Type: Distribute Volume ID: a89aa154-885c-4e14-8d3a-b555733b11f1 Status: Started Number of Bricks: 3 Transport-type: tcp Bricks: Brick1: 127.0.0.2:/home/ravi/bricks/brick4 Brick2: 127.0.0.2:/home/ravi/bricks/brick5 Brick3: 127.0.0.2:/home/ravi/bricks/brick6 [root at tuxpad ~]# ------------------------------------------------ Hope this helps. Ravi> Thank You, > John > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://www.gluster.org/mailman/listinfo/gluster-users-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20150401/02fdb403/attachment.html>