Re: Gluter 3.12.12: performance during heal and in general

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 





On Fri, Jul 27, 2018 at 1:32 PM, Hu Bert <revirii@xxxxxxxxxxxxxx> wrote:
2018-07-27 9:22 GMT+02:00 Pranith Kumar Karampuri <pkarampu@xxxxxxxxxx>:
>
>
> On Fri, Jul 27, 2018 at 12:36 PM, Hu Bert <revirii@xxxxxxxxxxxxxx> wrote:
>>
>> 2018-07-27 8:52 GMT+02:00 Pranith Kumar Karampuri <pkarampu@xxxxxxxxxx>:
>> >
>> >
>> > On Fri, Jul 27, 2018 at 11:53 AM, Hu Bert <revirii@xxxxxxxxxxxxxx>
>> > wrote:
>> >>
>> >> > Do you already have all the 190000 directories already created? If
>> >> > not
>> >> > could you find out which of the paths need it and do a stat directly
>> >> > instead
>> >> > of find?
>> >>
>> >> Quite probable not all of them have been created (but counting how
>> >> much would take very long...). Hm, maybe running stat in a double loop
>> >> (thx to our directory structure) would help. Something like this (may
>> >> be not 100% correct):
>> >>
>> >> for a in ${100..999}; do
>> >>     for b in ${100..999}; do
>> >>         stat /$a/$b/
>> >>     done
>> >> done
>> >>
>> >> Should run stat on all directories. I think i'll give this a try.
>> >
>> >
>> > Just to prevent these served from a cache, it is probably better to do
>> > this
>> > from a fresh mount?
>> >
>> > --
>> > Pranith
>>
>> Good idea. I'll install glusterfs client on a little used machine, so
>> there should be no caching. Thx! Have a good weekend when the time
>> comes :-)
>
>
> If this proves effective, what you need to also do is unmount and mount
> again, something like:
>
> mount
> for a in ${100..999}; do
>      for b in ${100..999}; do
>          stat /$a/$b/
>      done
>   done
> unmount

I'll see what is possible over the weekend.

Btw.: i've seen in the munin stats that the disk utilization for
bricksdd1 on the healthy gluster servers is between 70% (night) and
almost 99% (daytime). So it looks like that the basic problem is the
disk which seems not to be able to work faster? If so (heal)
performance won't improve with this setup, i assume.

It could be saturating in the day. But if enough self-heals are going on, even in the night
it should have been close to 100%.

 
Maybe switching
to RAID10 (conventional hard disks), SSDs or even add 3 additional
gluster servers (distributed replicated) could help?

It definitely will give better protection against hardware failure. Failure domain will be lesser.
--
Pranith
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux