Just started trying gluster, to decide if we want to put it into production.
Running version 3.7.11-1
Replicated, distributed volume, two servers, 20 bricks per server:
[root at storinator1 ~]# gluster volume status gv0
Status of volume: gv0
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick storinator1:/export/brick1/gv0        49153     0          Y
2554
Brick storinator2:/export/brick1/gv0        49153     0          Y
9686
Brick storinator1:/export/brick2/gv0        49154     0          Y
2562
Brick storinator2:/export/brick2/gv0        49154     0          Y
9708
Brick storinator1:/export/brick3/gv0        49155     0          Y
2568
Brick storinator2:/export/brick3/gv0        49155     0          Y
9692
Brick storinator1:/export/brick4/gv0        49156     0          Y
2574
Brick storinator2:/export/brick4/gv0        49156     0          Y
9765
Brick storinator1:/export/brick5/gv0        49173     0          Y
16901
Brick storinator2:/export/brick5/gv0        49173     0          Y
9727
Brick storinator1:/export/brick6/gv0        49174     0          Y
16920
Brick storinator2:/export/brick6/gv0        49174     0          Y
9733
Brick storinator1:/export/brick7/gv0        49175     0          Y
16939
Brick storinator2:/export/brick7/gv0        49175     0          Y
9739
Brick storinator1:/export/brick8/gv0        49176     0          Y
16958
Brick storinator2:/export/brick8/gv0        49176     0          Y
9703
Brick storinator1:/export/brick9/gv0        49177     0          Y
16977
Brick storinator2:/export/brick9/gv0        49177     0          Y
9713
Brick storinator1:/export/brick10/gv0       49178     0          Y
16996
Brick storinator2:/export/brick10/gv0       49178     0          Y
9718
Brick storinator1:/export/brick11/gv0       49179     0          Y
17015
Brick storinator2:/export/brick11/gv0       49179     0          Y
9746
Brick storinator1:/export/brick12/gv0       49180     0          Y
17034
Brick storinator2:/export/brick12/gv0       49180     0          Y
9792
Brick storinator1:/export/brick13/gv0       49181     0          Y
17053
Brick storinator2:/export/brick13/gv0       49181     0          Y
9755
Brick storinator1:/export/brick14/gv0       49182     0          Y
17072
Brick storinator2:/export/brick14/gv0       49182     0          Y
9767
Brick storinator1:/export/brick15/gv0       49183     0          Y
17091
Brick storinator2:/export/brick15/gv0       N/A       N/A        N
N/A
Brick storinator1:/export/brick16/gv0       49184     0          Y
17110
Brick storinator2:/export/brick16/gv0       49184     0          Y
9791
Brick storinator1:/export/brick17/gv0       49185     0          Y
17129
Brick storinator2:/export/brick17/gv0       49185     0          Y
9756
Brick storinator1:/export/brick18/gv0       49186     0          Y
17148
Brick storinator2:/export/brick18/gv0       49186     0          Y
9766
Brick storinator1:/export/brick19/gv0       49187     0          Y
17167
Brick storinator2:/export/brick19/gv0       49187     0          Y
9745
Brick storinator1:/export/brick20/gv0       49188     0          Y
17186
Brick storinator2:/export/brick20/gv0       49188     0          Y
9783
NFS Server on localhost                     2049      0          Y
17206
Self-heal Daemon on localhost               N/A       N/A        Y
17214
NFS Server on storinator2                   2049      0          Y
9657
Self-heal Daemon on storinator2             N/A       N/A        Y
9677
Task Status of Volume gv0
------------------------------------------------------------------------------
Task                 : Rebalance
ID                   : 28c733e9-d618-44fc-873f-405d3b29a609
Status               : completed
Wouldn't you know it, within a week or two of pulling the hardware together
and getting gluster installed and configured, a disk dies. Note the dead
process for brick15 on server storinator2.
I would like to remove (not replace) the failed brick (and its replica).
 (I don't have a spare disk handy, and there's plenty of room on the
other
bricks.)  But gluster doesn't seem to want to remove a brick if the brick
is dead:
[root at storinator1 ~]# gluster volume remove-brick gv0
storinator{1..2}:/export/brick15/gv0 start
volume remove-brick start: failed: Staging failed on storinator2. Error:
Found stopped brick storinator2:/export/brick15/gv0
So what do I do?  I can't remove the brick while the brick is bad, but I
want to remove the brick *because* the brick is bad.  Bit of a Catch-22.
Thanks in advance for any help you can give.
phil
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://www.gluster.org/pipermail/gluster-users/attachments/20160609/3a24e46a/attachment.html>
Nithya Balachandran
2016-Jun-15  04:49 UTC
[Gluster-users] Disk failed, how do I remove brick?
On Fri, Jun 10, 2016 at 1:25 AM, Phil Dumont <phil at solidstatescientific.com> wrote:> Just started trying gluster, to decide if we want to put it into > production. > > Running version 3.7.11-1 > > Replicated, distributed volume, two servers, 20 bricks per server: > > [root at storinator1 ~]# gluster volume status gv0 > Status of volume: gv0 > Gluster process TCP Port RDMA Port Online > Pid > > ------------------------------------------------------------------------------ > Brick storinator1:/export/brick1/gv0 49153 0 Y > 2554 > Brick storinator2:/export/brick1/gv0 49153 0 Y > 9686 > Brick storinator1:/export/brick2/gv0 49154 0 Y > 2562 > Brick storinator2:/export/brick2/gv0 49154 0 Y > 9708 > Brick storinator1:/export/brick3/gv0 49155 0 Y > 2568 > Brick storinator2:/export/brick3/gv0 49155 0 Y > 9692 > Brick storinator1:/export/brick4/gv0 49156 0 Y > 2574 > Brick storinator2:/export/brick4/gv0 49156 0 Y > 9765 > Brick storinator1:/export/brick5/gv0 49173 0 Y > 16901 > Brick storinator2:/export/brick5/gv0 49173 0 Y > 9727 > Brick storinator1:/export/brick6/gv0 49174 0 Y > 16920 > Brick storinator2:/export/brick6/gv0 49174 0 Y > 9733 > Brick storinator1:/export/brick7/gv0 49175 0 Y > 16939 > Brick storinator2:/export/brick7/gv0 49175 0 Y > 9739 > Brick storinator1:/export/brick8/gv0 49176 0 Y > 16958 > Brick storinator2:/export/brick8/gv0 49176 0 Y > 9703 > Brick storinator1:/export/brick9/gv0 49177 0 Y > 16977 > Brick storinator2:/export/brick9/gv0 49177 0 Y > 9713 > Brick storinator1:/export/brick10/gv0 49178 0 Y > 16996 > Brick storinator2:/export/brick10/gv0 49178 0 Y > 9718 > Brick storinator1:/export/brick11/gv0 49179 0 Y > 17015 > Brick storinator2:/export/brick11/gv0 49179 0 Y > 9746 > Brick storinator1:/export/brick12/gv0 49180 0 Y > 17034 > Brick storinator2:/export/brick12/gv0 49180 0 Y > 9792 > Brick storinator1:/export/brick13/gv0 49181 0 Y > 17053 > Brick storinator2:/export/brick13/gv0 49181 0 Y > 9755 > Brick storinator1:/export/brick14/gv0 49182 0 Y > 17072 > Brick storinator2:/export/brick14/gv0 49182 0 Y > 9767 > Brick storinator1:/export/brick15/gv0 49183 0 Y > 17091 > Brick storinator2:/export/brick15/gv0 N/A N/A N > N/A > Brick storinator1:/export/brick16/gv0 49184 0 Y > 17110 > Brick storinator2:/export/brick16/gv0 49184 0 Y > 9791 > Brick storinator1:/export/brick17/gv0 49185 0 Y > 17129 > Brick storinator2:/export/brick17/gv0 49185 0 Y > 9756 > Brick storinator1:/export/brick18/gv0 49186 0 Y > 17148 > Brick storinator2:/export/brick18/gv0 49186 0 Y > 9766 > Brick storinator1:/export/brick19/gv0 49187 0 Y > 17167 > Brick storinator2:/export/brick19/gv0 49187 0 Y > 9745 > Brick storinator1:/export/brick20/gv0 49188 0 Y > 17186 > Brick storinator2:/export/brick20/gv0 49188 0 Y > 9783 > NFS Server on localhost 2049 0 Y > 17206 > Self-heal Daemon on localhost N/A N/A Y > 17214 > NFS Server on storinator2 2049 0 Y > 9657 > Self-heal Daemon on storinator2 N/A N/A Y > 9677 > > Task Status of Volume gv0 > > ------------------------------------------------------------------------------ > Task : Rebalance > ID : 28c733e9-d618-44fc-873f-405d3b29a609 > Status : completed > > > Wouldn't you know it, within a week or two of pulling the hardware > together and getting gluster installed and configured, a disk dies. Note > the dead process for brick15 on server storinator2. > > I would like to remove (not replace) the failed brick (and its replica). > (I don't have a spare disk handy, and there's plenty of room on the other > bricks.) But gluster doesn't seem to want to remove a brick if the brick > is dead: > > [root at storinator1 ~]# gluster volume remove-brick gv0 > storinator{1..2}:/export/brick15/gv0 start > volume remove-brick start: failed: Staging failed on storinator2. Error: > Found stopped brick storinator2:/export/brick15/gv0 > > > So what do I do? I can't remove the brick while the brick is bad, but I > want to remove the brick *because* the brick is bad. Bit of a Catch-22. >Hi Phil, This looks like a bug - please open a new BZ for this issue. You can remove these bricks using gluster v remove-brick <vol> <bricks> force However, that will not copy the data from the bricks that have been removed. You will need to copy the files manually from the brick to the gluster volume mount point. Thanks, Nithya> Thanks in advance for any help you can give. > > phil > > > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://www.gluster.org/mailman/listinfo/gluster-users >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://www.gluster.org/pipermail/gluster-users/attachments/20160615/a0a30c63/attachment.html>