Re: [Gluster-users] Fwd: dht_is_subvol_filled messages on client

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, May 5, 2016 at 2:36 PM, David Gossage
<dgossage@xxxxxxxxxxxxxxxxxx> wrote:
>
>
>
> On Thu, May 5, 2016 at 3:28 AM, Serkan Çoban <cobanserkan@xxxxxxxxx> wrote:
>>
>> Hi,
>>
>> You can find the output below link:
>> https://www.dropbox.com/s/wzrh5yp494ogksc/status_detail.txt?dl=0
>>
>> Thanks,
>> Serkan
>
>
> Maybe not issue, but playing one of these things is not like the other I
> notice of all the bricks only one seems to be different at a quick glance
>
> Brick                : Brick 1.1.1.235:/bricks/20
> TCP Port             : 49170
> RDMA Port            : 0
> Online               : Y
> Pid                  : 26736
> File System          : ext4
> Device               : /dev/mapper/vol0-vol_root
> Mount Options        : rw,relatime,data=ordered
> Inode Size           : 256
> Disk Space Free      : 86.1GB
> Total Disk Space     : 96.0GB
> Inode Count          : 6406144
> Free Inodes          : 6381374
>
> Every other brick seems to be 7TB and xfs but this one.

Looks like the brick fs isn't mounted, and the root-fs is being used
instead. But that still leaves enough inodes free.

What I suspect is that one of the cluster translators is mixing up
stats when aggregating from multiple bricks.
>From the log snippet you gave in the first mail, it seems like the
disperse translator is possibly involved.

BTW, how large is the volume you have? Those are a lot of bricks!

~kaushal


>
>
>
>>
>>
>> On Thu, May 5, 2016 at 9:33 AM, Xavier Hernandez <xhernandez@xxxxxxxxxx>
>> wrote:
>> > Can you post the result of 'gluster volume status v0 detail' ?
>> >
>> >
>> > On 05/05/16 06:49, Serkan Çoban wrote:
>> >>
>> >> Hi, Can anyone suggest something for this issue? df, du has no issue
>> >> for the bricks yet one subvolume not being used by gluster..
>> >>
>> >> On Wed, May 4, 2016 at 4:40 PM, Serkan Çoban <cobanserkan@xxxxxxxxx>
>> >> wrote:
>> >>>
>> >>> Hi,
>> >>>
>> >>> I changed cluster.min-free-inodes to "0". Remount the volume on
>> >>> clients. inode full messages not coming to syslog anymore but I see
>> >>> disperse-56 subvolume still not being used.
>> >>> Anything I can do to resolve this issue? Maybe I can destroy and
>> >>> recreate the volume but I am not sure It will fix this issue...
>> >>> Maybe the disperse size 16+4 is too big should I change it to 8+2?
>> >>>
>> >>> On Tue, May 3, 2016 at 2:36 PM, Serkan Çoban <cobanserkan@xxxxxxxxx>
>> >>> wrote:
>> >>>>
>> >>>> I also checked the df output all 20 bricks are same like below:
>> >>>> /dev/sdu1 7.3T 34M 7.3T 1% /bricks/20
>> >>>>
>> >>>> On Tue, May 3, 2016 at 1:40 PM, Raghavendra G
>> >>>> <raghavendra@xxxxxxxxxxx>
>> >>>> wrote:
>> >>>>>
>> >>>>>
>> >>>>>
>> >>>>> On Mon, May 2, 2016 at 11:41 AM, Serkan Çoban
>> >>>>> <cobanserkan@xxxxxxxxx>
>> >>>>> wrote:
>> >>>>>>
>> >>>>>>
>> >>>>>>> 1. What is the out put of du -hs <back-end-export>? Please get
>> >>>>>>> this
>> >>>>>>> information for each of the brick that are part of disperse.
>> >>>>>
>> >>>>>
>> >>>>>
>> >>>>> Sorry. I needed df output of the filesystem containing brick. Not
>> >>>>> du.
>> >>>>> Sorry
>> >>>>> about that.
>> >>>>>
>> >>>>>>
>> >>>>>> There are 20 bricks in disperse-56 and the du -hs output is like:
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 1.8M /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>> 80K /bricks/20
>> >>>>>>
>> >>>>>> I see that gluster is not writing to this disperse set. All other
>> >>>>>> disperse sets are filled 13GB but this one is empty. I see
>> >>>>>> directory
>> >>>>>> structure created but no files in directories.
>> >>>>>> How can I fix the issue? I will try to rebalance but I don't think
>> >>>>>> it
>> >>>>>> will write to this disperse set...
>> >>>>>>
>> >>>>>>
>> >>>>>>
>> >>>>>> On Sat, Apr 30, 2016 at 9:22 AM, Raghavendra G
>> >>>>>> <raghavendra@xxxxxxxxxxx>
>> >>>>>> wrote:
>> >>>>>>>
>> >>>>>>>
>> >>>>>>>
>> >>>>>>> On Fri, Apr 29, 2016 at 12:32 AM, Serkan Çoban
>> >>>>>>> <cobanserkan@xxxxxxxxx>
>> >>>>>>> wrote:
>> >>>>>>>>
>> >>>>>>>>
>> >>>>>>>> Hi, I cannot get an answer from user list, so asking to devel
>> >>>>>>>> list.
>> >>>>>>>>
>> >>>>>>>> I am getting [dht-diskusage.c:277:dht_is_subvol_filled] 0-v0-dht:
>> >>>>>>>> inodes on subvolume 'v0-disperse-56' are at (100.00 %), consider
>> >>>>>>>> adding more bricks.
>> >>>>>>>>
>> >>>>>>>> message on client logs.My cluster is empty there are only a
>> >>>>>>>> couple
>> >>>>>>>> of
>> >>>>>>>> GB files for testing. Why this message appear in syslog?
>> >>>>>>>
>> >>>>>>>
>> >>>>>>>
>> >>>>>>> dht uses disk usage information from backend export.
>> >>>>>>>
>> >>>>>>> 1. What is the out put of du -hs <back-end-export>? Please get
>> >>>>>>> this
>> >>>>>>> information for each of the brick that are part of disperse.
>> >>>>>>> 2. Once you get du information from each brick, the value seen by
>> >>>>>>> dht
>> >>>>>>> will
>> >>>>>>> be based on how cluster/disperse aggregates du info (basically
>> >>>>>>> statfs
>> >>>>>>> fop).
>> >>>>>>>
>> >>>>>>> The reason for 100% disk usage may be,
>> >>>>>>> In case of 1, backend fs might be shared by data other than brick.
>> >>>>>>> In case of 2, some issues with aggregation.
>> >>>>>>>
>> >>>>>>>> Is is safe to
>> >>>>>>>> ignore it?
>> >>>>>>>
>> >>>>>>>
>> >>>>>>>
>> >>>>>>> dht will try not to have data files on the subvol in question
>> >>>>>>> (v0-disperse-56). Hence lookup cost will be two hops for files
>> >>>>>>> hashing
>> >>>>>>> to
>> >>>>>>> disperse-56 (note that other fops like read/write/open still have
>> >>>>>>> the
>> >>>>>>> cost
>> >>>>>>> of single hop and dont suffer from this penalty). Other than that
>> >>>>>>> there
>> >>>>>>> is
>> >>>>>>> no significant harm unless disperse-56 is really running out of
>> >>>>>>> space.
>> >>>>>>>
>> >>>>>>> regards,
>> >>>>>>> Raghavendra
>> >>>>>>>
>> >>>>>>>> _______________________________________________
>> >>>>>>>> Gluster-devel mailing list
>> >>>>>>>> Gluster-devel@xxxxxxxxxxx
>> >>>>>>>> http://www.gluster.org/mailman/listinfo/gluster-devel
>> >>>>>>>
>> >>>>>>>
>> >>>>>>>
>> >>>>>>>
>> >>>>>>>
>> >>>>>>> --
>> >>>>>>> Raghavendra G
>> >>>>>>
>> >>>>>> _______________________________________________
>> >>>>>> Gluster-devel mailing list
>> >>>>>> Gluster-devel@xxxxxxxxxxx
>> >>>>>> http://www.gluster.org/mailman/listinfo/gluster-devel
>> >>>>>
>> >>>>>
>> >>>>>
>> >>>>>
>> >>>>>
>> >>>>> --
>> >>>>> Raghavendra G
>> >>
>> >> _______________________________________________
>> >> Gluster-users mailing list
>> >> Gluster-users@xxxxxxxxxxx
>> >> http://www.gluster.org/mailman/listinfo/gluster-users
>> >>
>> >
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users@xxxxxxxxxxx
>> http://www.gluster.org/mailman/listinfo/gluster-users
>
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users@xxxxxxxxxxx
> http://www.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-devel mailing list
Gluster-devel@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-devel




[Index of Archives]     [Gluster Users]     [Ceph Users]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Security]     [Bugtraq]     [Linux]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux