Changliang Chen
2011-Oct-18 07:24 UTC
[Gluster-users] gluster rebalance taking three months
Hi guys, we have a rebalance running on eight bricks since July and this is what the status looks like right now: ===Tue Oct 18 13:45:01 CST 2011 ===rebalance step 1: layout fix in progress: fixed layout 223623 There are roughly 8T photos in the storage,so how long should this rebalance take? What does the number (in this case) 22362 represent? Our gluster infomation: Repository revision: v3.1.1 Volume Name: dfs Type: Distributed-Replicate Status: Started Number of Bricks: 4 x 2 = 8 Transport-type: tcp Bricks: Brick1: 10.1.1.23:/data0 Brick2: 10.1.1.24:/data0 Brick3: 10.1.1.25:/data0 Brick4: 10.1.1.26:/data0 Brick5: 10.1.1.27:/data0 Brick6: 10.1.1.28:/data0 Brick7: 10.1.1.64:/data0 Brick8: 10.1.1.65:/data0 Options Reconfigured: cluster.min-free-disk: 10% network.ping-timeout: 25 network.frame-timeout: 30 performance.cache-max-file-size: 512KB performance.cache-size: 3GB -- Regards, Cocl OM manager 19lou Operation & Maintenance Dept -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20111018/b2beb8b3/attachment.html>
Hi Chen, Can you restart the 'glusterd' and run 'gluster volume rebalance dfs migrate-data start' and check if your data migration happens? Regards, Amar On Tue, Oct 18, 2011 at 12:54 PM, Changliang Chen <hqucocl at gmail.com> wrote:> Hi guys, > > we have a rebalance running on eight bricks since July and this is > what the status looks like right now: > > ===Tue Oct 18 13:45:01 CST 2011 ===> rebalance step 1: layout fix in progress: fixed layout 223623 > > There are roughly 8T photos in the storage,so how long should this > rebalance take? > > What does the number (in this case) 22362 represent? > > Our gluster infomation: > Repository revision: v3.1.1 > Volume Name: dfs > Type: Distributed-Replicate > Status: Started > Number of Bricks: 4 x 2 = 8 > Transport-type: tcp > Bricks: > Brick1: 10.1.1.23:/data0 > Brick2: 10.1.1.24:/data0 > Brick3: 10.1.1.25:/data0 > Brick4: 10.1.1.26:/data0 > Brick5: 10.1.1.27:/data0 > Brick6: 10.1.1.28:/data0 > Brick7: 10.1.1.64:/data0 > Brick8: 10.1.1.65:/data0 > Options Reconfigured: > cluster.min-free-disk: 10% > network.ping-timeout: 25 > network.frame-timeout: 30 > performance.cache-max-file-size: 512KB > performance.cache-size: 3GB > > > > -- > > Regards, > > Cocl > OM manager > 19lou Operation & Maintenance Dept > > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users > >-------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20111018/03667517/attachment.html>
Changliang Chen
2011-Oct-21 06:53 UTC
[Gluster-users] gluster rebalance taking three months
Any help? We notice that if the below errors appears,the rebalance fixed layout will become very slow,the number just increase about 4 per five minutes. E [rpc-clnt.c:199:call_bail] dfs-client-0: bailing out frame type(GlusterFS 3.1) op(INODELK(29)) xid = 0x755696 sent = 2011-10-20 06:20:51.217782. timeout = 30 W [afr-self-heal-common.c:584:afr_sh_pending_to_delta] afr_sh_pending_to_delta: Unable to get dict value. I [dht-common.c:369:dht_revalidate_cbk] dfs-dht: subvolume 19loudfs-replicate-2 returned -1 (Invalid argument) On Tue, Oct 18, 2011 at 5:45 PM, Changliang Chen <hqucocl at gmail.com> wrote:> Thanks Amar,but it looks like that the v3.1.1 hasn't support the command > > 'gluster volume rebalance dfs migrate-data start' > > # gluster volume rebalance dfs migrate-data start > Usage: volume rebalance <VOLNAME> <start|stop|status> > Rebalance of Volume dfs failed > > On Tue, Oct 18, 2011 at 3:33 PM, Amar Tumballi <amar at gluster.com> wrote: > >> Hi Chen, >> >> Can you restart the 'glusterd' and run 'gluster volume rebalance dfs >> migrate-data start' and check if your data migration happens? >> >> Regards, >> Amar >> >> On Tue, Oct 18, 2011 at 12:54 PM, Changliang Chen <hqucocl at gmail.com>wrote: >> >>> Hi guys, >>> >>> we have a rebalance running on eight bricks since July and this is >>> what the status looks like right now: >>> >>> ===Tue Oct 18 13:45:01 CST 2011 ===>>> rebalance step 1: layout fix in progress: fixed layout 223623 >>> >>> There are roughly 8T photos in the storage,so how long should this >>> rebalance take? >>> >>> What does the number (in this case) 22362 represent? >>> >>> Our gluster infomation: >>> Repository revision: v3.1.1 >>> Volume Name: dfs >>> Type: Distributed-Replicate >>> Status: Started >>> Number of Bricks: 4 x 2 = 8 >>> Transport-type: tcp >>> Bricks: >>> Brick1: 10.1.1.23:/data0 >>> Brick2: 10.1.1.24:/data0 >>> Brick3: 10.1.1.25:/data0 >>> Brick4: 10.1.1.26:/data0 >>> Brick5: 10.1.1.27:/data0 >>> Brick6: 10.1.1.28:/data0 >>> Brick7: 10.1.1.64:/data0 >>> Brick8: 10.1.1.65:/data0 >>> Options Reconfigured: >>> cluster.min-free-disk: 10% >>> network.ping-timeout: 25 >>> network.frame-timeout: 30 >>> performance.cache-max-file-size: 512KB >>> performance.cache-size: 3GB >>> >>> >>> >>> -- >>> >>> Regards, >>> >>> Cocl >>> OM manager >>> 19lou Operation & Maintenance Dept >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org >>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>> >>> >> > > > -- > > Regards, > > Cocl > OM manager > 19lou Operation & Maintenance Dept >-- Regards, Cocl OM manager 19lou Operation & Maintenance Dept -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://supercolony.gluster.org/pipermail/gluster-users/attachments/20111021/54d5aa29/attachment.html>