Hi, I changed cluster.min-free-inodes to "0". Remount the volume on clients. inode full messages not coming to syslog anymore but I see disperse-56 subvolume still not being used. Anything I can do to resolve this issue? Maybe I can destroy and recreate the volume but I am not sure It will fix this issue... Maybe the disperse size 16+4 is too big should I change it to 8+2? On Tue, May 3, 2016 at 2:36 PM, Serkan Çoban <cobanserkan@xxxxxxxxx> wrote: > I also checked the df output all 20 bricks are same like below: > /dev/sdu1 7.3T 34M 7.3T 1% /bricks/20 > > On Tue, May 3, 2016 at 1:40 PM, Raghavendra G <raghavendra@xxxxxxxxxxx> wrote: >> >> >> On Mon, May 2, 2016 at 11:41 AM, Serkan Çoban <cobanserkan@xxxxxxxxx> wrote: >>> >>> >1. What is the out put of du -hs <back-end-export>? Please get this >>> > information for each of the brick that are part of disperse. >> >> >> Sorry. I needed df output of the filesystem containing brick. Not du. Sorry >> about that. >> >>> >>> There are 20 bricks in disperse-56 and the du -hs output is like: >>> 80K /bricks/20 >>> 80K /bricks/20 >>> 80K /bricks/20 >>> 80K /bricks/20 >>> 80K /bricks/20 >>> 80K /bricks/20 >>> 80K /bricks/20 >>> 80K /bricks/20 >>> 1.8M /bricks/20 >>> 80K /bricks/20 >>> 80K /bricks/20 >>> 80K /bricks/20 >>> 80K /bricks/20 >>> 80K /bricks/20 >>> 80K /bricks/20 >>> 80K /bricks/20 >>> 80K /bricks/20 >>> 80K /bricks/20 >>> 80K /bricks/20 >>> 80K /bricks/20 >>> >>> I see that gluster is not writing to this disperse set. All other >>> disperse sets are filled 13GB but this one is empty. I see directory >>> structure created but no files in directories. >>> How can I fix the issue? I will try to rebalance but I don't think it >>> will write to this disperse set... >>> >>> >>> >>> On Sat, Apr 30, 2016 at 9:22 AM, Raghavendra G <raghavendra@xxxxxxxxxxx> >>> wrote: >>> > >>> > >>> > On Fri, Apr 29, 2016 at 12:32 AM, Serkan Çoban <cobanserkan@xxxxxxxxx> >>> > wrote: >>> >> >>> >> Hi, I cannot get an answer from user list, so asking to devel list. >>> >> >>> >> I am getting [dht-diskusage.c:277:dht_is_subvol_filled] 0-v0-dht: >>> >> inodes on subvolume 'v0-disperse-56' are at (100.00 %), consider >>> >> adding more bricks. >>> >> >>> >> message on client logs.My cluster is empty there are only a couple of >>> >> GB files for testing. Why this message appear in syslog? >>> > >>> > >>> > dht uses disk usage information from backend export. >>> > >>> > 1. What is the out put of du -hs <back-end-export>? Please get this >>> > information for each of the brick that are part of disperse. >>> > 2. Once you get du information from each brick, the value seen by dht >>> > will >>> > be based on how cluster/disperse aggregates du info (basically statfs >>> > fop). >>> > >>> > The reason for 100% disk usage may be, >>> > In case of 1, backend fs might be shared by data other than brick. >>> > In case of 2, some issues with aggregation. >>> > >>> >> Is is safe to >>> >> ignore it? >>> > >>> > >>> > dht will try not to have data files on the subvol in question >>> > (v0-disperse-56). Hence lookup cost will be two hops for files hashing >>> > to >>> > disperse-56 (note that other fops like read/write/open still have the >>> > cost >>> > of single hop and dont suffer from this penalty). Other than that there >>> > is >>> > no significant harm unless disperse-56 is really running out of space. >>> > >>> > regards, >>> > Raghavendra >>> > >>> >> _______________________________________________ >>> >> Gluster-devel mailing list >>> >> Gluster-devel@xxxxxxxxxxx >>> >> http://www.gluster.org/mailman/listinfo/gluster-devel >>> > >>> > >>> > >>> > >>> > -- >>> > Raghavendra G >>> _______________________________________________ >>> Gluster-devel mailing list >>> Gluster-devel@xxxxxxxxxxx >>> http://www.gluster.org/mailman/listinfo/gluster-devel >> >> >> >> >> -- >> Raghavendra G _______________________________________________ Gluster-devel mailing list Gluster-devel@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-devel