On Wed, Aug 10, 2016 at 1:58 PM, Serkan Çoban <cobanserkan@xxxxxxxxx> wrote:
Hi,
Any progress about the patch?
hi Serkan,
While testing the patch by myself, I am seeing that it is taking more than one crawl to complete heals even when there are no directory hierarchies. It is faster than before but it shouldn't take more than 1 crawl to complete the heal because all the files exist already. I am investigating why that is the case now. If you want to test things out without this patch I will give you rpms today. Otherwise we need to find until we find RCA for this crawl problem. Let me know your decision. If you are okay with testing progressive versions of this feature, that would be great. We can compare how each patch improved the performance.
Pranith
On Thu, Aug 4, 2016 at 10:16 AM, Pranith Kumar Karampuri
<pkarampu@xxxxxxxxxx> wrote:
>
>
> On Thu, Aug 4, 2016 at 11:30 AM, Serkan Çoban <cobanserkan@xxxxxxxxx> wrote:
>>
>> Thanks Pranith,
>> I am waiting for RPMs to show, I will do the tests as soon as possible
>> and inform you.
>
>
> I guess on 3.7.x the RPMs are not automatically built. Let me find how it
> can be done. I will inform you after finding that out. Give me a day.
>
>>
>>
>> On Wed, Aug 3, 2016 at 11:19 PM, Pranith Kumar Karampuri
>> <pkarampu@xxxxxxxxxx> wrote:
>> >
>> >
>> > On Thu, Aug 4, 2016 at 1:47 AM, Pranith Kumar Karampuri
>> > <pkarampu@xxxxxxxxxx> wrote:
>> >>
>> >>
>> >>
>> >> On Thu, Aug 4, 2016 at 12:51 AM, Serkan Çoban <cobanserkan@xxxxxxxxx>
>> >> wrote:
>> >>>
>> >>> I use rpms for installation. Redhat/Centos 6.8.
>> >>
>> >>
>> >> http://review.gluster.org/#/c/15084 is the patch. In some time the rpms
>> >> will be built actually.
>> >
>> >
>> > In the same URL above it will actually post the rpms for fedora/el6/el7
>> > at
>> > the end of the page.
>> >
>> >>
>> >>
>> >> Use gluster volume set <volname> disperse.shd-max-threads <num-threads
>> >> (range: 1-64)>
>> >>
>> >> While testing this I thought of ways to decrease the number of crawls
>> >> as
>> >> well. But they are a bit involved. Try to create same set of data and
>> >> see
>> >> what is the time it takes to complete heals using number of threads as
>> >> you
>> >> increase the number of parallel heals from 1 to 64.
>> >>
>> >>>
>> >>> On Wed, Aug 3, 2016 at 10:16 PM, Pranith Kumar Karampuri
>> >>> <pkarampu@xxxxxxxxxx> wrote:
>> >>> >
>> >>> >
>> >>> > On Thu, Aug 4, 2016 at 12:45 AM, Serkan Çoban
>> >>> > <cobanserkan@xxxxxxxxx>
>> >>> > wrote:
>> >>> >>
>> >>> >> I prefer 3.7 if it is ok for you. Can you also provide build
>> >>> >> instructions?
>> >>> >
>> >>> >
>> >>> > 3.7 should be fine. Do you use rpms/debs/anything-else?
>> >>> >
>> >>> >>
>> >>> >>
>> >>> >> On Wed, Aug 3, 2016 at 10:12 PM, Pranith Kumar Karampuri
>> >>> >> <pkarampu@xxxxxxxxxx> wrote:
>> >>> >> >
>> >>> >> >
>> >>> >> > On Thu, Aug 4, 2016 at 12:37 AM, Serkan Çoban
>> >>> >> > <cobanserkan@xxxxxxxxx>
>> >>> >> > wrote:
>> >>> >> >>
>> >>> >> >> Yes, but I can create 2+1(or 8+2) ec using two servers right? I
>> >>> >> >> have
>> >>> >> >> 26 disks on each server.
>> >>> >> >
>> >>> >> >
>> >>> >> > On which release-branch do you want the patch? I am testing it on
>> >>> >> > master-branch now.
>> >>> >> >
>> >>> >> >>
>> >>> >> >>
>> >>> >> >> On Wed, Aug 3, 2016 at 9:59 PM, Pranith Kumar Karampuri
>> >>> >> >> <pkarampu@xxxxxxxxxx> wrote:
>> >>> >> >> >
>> >>> >> >> >
>> >>> >> >> > On Thu, Aug 4, 2016 at 12:23 AM, Serkan Çoban
>> >>> >> >> > <cobanserkan@xxxxxxxxx>
>> >>> >> >> > wrote:
>> >>> >> >> >>
>> >>> >> >> >> I have two of my storage servers free, I think I can use them
>> >>> >> >> >> for
>> >>> >> >> >> testing. Is two server testing environment ok for you?
>> >>> >> >> >
>> >>> >> >> >
>> >>> >> >> > I think it would be better if you have at least 3. You can
>> >>> >> >> > test
>> >>> >> >> > it
>> >>> >> >> > with
>> >>> >> >> > 2+1
>> >>> >> >> > ec configuration.
>> >>> >> >> >
>> >>> >> >> >>
>> >>> >> >> >>
>> >>> >> >> >> On Wed, Aug 3, 2016 at 9:44 PM, Pranith Kumar Karampuri
>> >>> >> >> >> <pkarampu@xxxxxxxxxx> wrote:
>> >>> >> >> >> >
>> >>> >> >> >> >
>> >>> >> >> >> > On Wed, Aug 3, 2016 at 6:01 PM, Serkan Çoban
>> >>> >> >> >> > <cobanserkan@xxxxxxxxx>
>> >>> >> >> >> > wrote:
>> >>> >> >> >> >>
>> >>> >> >> >> >> Hi,
>> >>> >> >> >> >>
>> >>> >> >> >> >> May I ask if multi-threaded self heal for distributed
>> >>> >> >> >> >> disperse
>> >>> >> >> >> >> volumes
>> >>> >> >> >> >> implemented in this release?
>> >>> >> >> >> >
>> >>> >> >> >> >
>> >>> >> >> >> > Serkan,
>> >>> >> >> >> > At the moment I am a bit busy with different work,
>> >>> >> >> >> > Is
>> >>> >> >> >> > it
>> >>> >> >> >> > possible
>> >>> >> >> >> > for you to help test the feature if I provide a patch?
>> >>> >> >> >> > Actually
>> >>> >> >> >> > the
>> >>> >> >> >> > patch
>> >>> >> >> >> > should be small. Testing is where lot of time will be spent
>> >>> >> >> >> > on.
>> >>> >> >> >> >
>> >>> >> >> >> >>
>> >>> >> >> >> >>
>> >>> >> >> >> >> Thanks,
>> >>> >> >> >> >> Serkan
>> >>> >> >> >> >>
>> >>> >> >> >> >> On Tue, Aug 2, 2016 at 5:30 PM, David Gossage
>> >>> >> >> >> >> <dgossage@xxxxxxxxxxxxxxxxxx> wrote:
>> >>> >> >> >> >> > On Tue, Aug 2, 2016 at 6:01 AM, Lindsay Mathieson
>> >>> >> >> >> >> > <lindsay.mathieson@xxxxxxxxx> wrote:
>> >>> >> >> >> >> >>
>> >>> >> >> >> >> >> On 2/08/2016 5:07 PM, Kaushal M wrote:
>> >>> >> >> >> >> >>>
>> >>> >> >> >> >> >>> GlusterFS-3.7.14 has been released. This is a regular
>> >>> >> >> >> >> >>> minor
>> >>> >> >> >> >> >>> release.
>> >>> >> >> >> >> >>> The release-notes are available at
>> >>> >> >> >> >> >>>
>> >>> >> >> >> >> >>>
>> >>> >> >> >> >> >>>
>> >>> >> >> >> >> >>>
>> >>> >> >> >> >> >>>
>> >>> >> >> >> >> >>>
>> >>> >> >> >> >> >>>
>> >>> >> >> >> >> >>> https://github.com/gluster/glusterfs/blob/release-3.7/ doc/release-notes/3.7.14.md
>> >>> >> >> >> >> >>
>> >>> >> >> >> >> >>
>> >>> >> >> >> >> >> Thanks Kaushal, I'll check it out
>> >>> >> >> >> >> >>
>> >>> >> >> >> >> >
>> >>> >> >> >> >> > So far on my test box its working as expected. At least
>> >>> >> >> >> >> > the
>> >>> >> >> >> >> > issues
>> >>> >> >> >> >> > that
>> >>> >> >> >> >> > prevented it from running as before have disappeared.
>> >>> >> >> >> >> > Will
>> >>> >> >> >> >> > need
>> >>> >> >> >> >> > to
>> >>> >> >> >> >> > see
>> >>> >> >> >> >> > how
>> >>> >> >> >> >> > my test VM behaves after a few days.
>> >>> >> >> >> >> >
>> >>> >> >> >> >> >
>> >>> >> >> >> >> >
>> >>> >> >> >> >> >> --
>> >>> >> >> >> >> >> Lindsay Mathieson
>> >>> >> >> >> >> >>
>> >>> >> >> >> >> >> _______________________________________________
>> >>> >> >> >> >> >> Gluster-users mailing list
>> >>> >> >> >> >> >> Gluster-users@xxxxxxxxxxx
>> >>> >> >> >> >> >> http://www.gluster.org/mailman/listinfo/gluster-users
>> >>> >> >> >> >> >
>> >>> >> >> >> >> >
>> >>> >> >> >> >> >
>> >>> >> >> >> >> > _______________________________________________
>> >>> >> >> >> >> > Gluster-users mailing list
>> >>> >> >> >> >> > Gluster-users@xxxxxxxxxxx
>> >>> >> >> >> >> > http://www.gluster.org/mailman/listinfo/gluster-users
>> >>> >> >> >> >> _______________________________________________
>> >>> >> >> >> >> Gluster-users mailing list
>> >>> >> >> >> >> Gluster-users@xxxxxxxxxxx
>> >>> >> >> >> >> http://www.gluster.org/mailman/listinfo/gluster-users
>> >>> >> >> >> >
>> >>> >> >> >> >
>> >>> >> >> >> >
>> >>> >> >> >> >
>> >>> >> >> >> > --
>> >>> >> >> >> > Pranith
>> >>> >> >> >
>> >>> >> >> >
>> >>> >> >> >
>> >>> >> >> >
>> >>> >> >> > --
>> >>> >> >> > Pranith
>> >>> >> >
>> >>> >> >
>> >>> >> >
>> >>> >> >
>> >>> >> > --
>> >>> >> > Pranith
>> >>> >
>> >>> >
>> >>> >
>> >>> >
>> >>> > --
>> >>> > Pranith
>> >>
>> >>
>> >>
>> >>
>> >> --
>> >> Pranith
>> >
>> >
>> >
>> >
>> > --
>> > Pranith
>
>
>
>
> --
> Pranith
--
Pranith
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users