Any help? We notice that if the below errors appears,the rebalance fixed layout will become very slow,the number just increase about 4 per five minutes. E [rpc-clnt.c:199:call_bail] dfs-client-0: bailing out frame type(GlusterFS 3.1) op(INODELK(29)) xid = 0x755696 sent = 2011-10-20 06:20:51.217782. timeout = 30 W [afr-self-heal-common.c:584:afr_sh_pending_to_delta] afr_sh_pending_to_delta: Unable to get dict value. I [dht-common.c:369:dht_revalidate_cbk] dfs-dht: subvolume 19loudfs-replicate-2 returned -1 (Invalid argument) On Tue, Oct 18, 2011 at 5:45 PM, Changliang Chen <hqucocl at gmail.com> wrote: > Thanks Amar,but it looks like that the v3.1.1 hasn't support the command > > 'gluster volume rebalance dfs migrate-data start' > > # gluster volume rebalance dfs migrate-data start > Usage: volume rebalance <VOLNAME> <start|stop|status> > Rebalance of Volume dfs failed > > On Tue, Oct 18, 2011 at 3:33 PM, Amar Tumballi <amar at gluster.com> wrote: > >> Hi Chen, >> >> Can you restart the 'glusterd' and run 'gluster volume rebalance dfs >> migrate-data start' and check if your data migration happens? >> >> Regards, >> Amar >> >> On Tue, Oct 18, 2011 at 12:54 PM, Changliang Chen <hqucocl at gmail.com>wrote: >> >>> Hi guys, >>> >>> we have a rebalance running on eight bricks since July and this is >>> what the status looks like right now: >>> >>> ===Tue Oct 18 13:45:01 CST 2011 ==== >>> rebalance step 1: layout fix in progress: fixed layout 223623 >>> >>> There are roughly 8T photos in the storage,so how long should this >>> rebalance take? >>> >>> What does the number (in this case) 22362 represent? >>> >>> Our gluster infomation: >>> Repository revision: v3.1.1 >>> Volume Name: dfs >>> Type: Distributed-Replicate >>> Status: Started >>> Number of Bricks: 4 x 2 = 8 >>> Transport-type: tcp >>> Bricks: >>> Brick1: 10.1.1.23:/data0 >>> Brick2: 10.1.1.24:/data0 >>> Brick3: 10.1.1.25:/data0 >>> Brick4: 10.1.1.26:/data0 >>> Brick5: 10.1.1.27:/data0 >>> Brick6: 10.1.1.28:/data0 >>> Brick7: 10.1.1.64:/data0 >>> Brick8: 10.1.1.65:/data0 >>> Options Reconfigured: >>> cluster.min-free-disk: 10% >>> network.ping-timeout: 25 >>> network.frame-timeout: 30 >>> performance.cache-max-file-size: 512KB >>> performance.cache-size: 3GB >>> >>> >>> >>> -- >>> >>> Regards, >>> >>> Cocl >>> OM manager >>> 19lou Operation & Maintenance Dept >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org >>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>> >>> >> > > > -- > > Regards, > > Cocl > OM manager > 19lou Operation & Maintenance Dept > -- Regards, Cocl OM manager 19lou Operation & Maintenance Dept -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://gluster.org/pipermail/gluster-users/attachments/20111021/54d5aa29/attachment.htm>