On Thu, Aug 11, 2016 at 4:29 PM, Serkan Çoban <cobanserkan@xxxxxxxxx> wrote:
I can wait for the patch to complete, please inform me when you ready.
If it will take too much time to solve the crawl issue I can test
without it too...
I don't know the Root cause for the problem, so I am not sure by when it will be ready. Let me build the rpms, I have a meeting now for around an hour. I will start building rpms after that.
Serkan
On Thu, Aug 11, 2016 at 5:52 AM, Pranith Kumar Karampuri
<pkarampu@xxxxxxxxxx> wrote:
>
>
> On Wed, Aug 10, 2016 at 1:58 PM, Serkan Çoban <cobanserkan@xxxxxxxxx> wrote:
>>
>> Hi,
>>
>> Any progress about the patch?
>
>
> hi Serkan,
> While testing the patch by myself, I am seeing that it is taking more
> than one crawl to complete heals even when there are no directory
> hierarchies. It is faster than before but it shouldn't take more than 1
> crawl to complete the heal because all the files exist already. I am
> investigating why that is the case now. If you want to test things out
> without this patch I will give you rpms today. Otherwise we need to find
> until we find RCA for this crawl problem. Let me know your decision. If you
> are okay with testing progressive versions of this feature, that would be
> great. We can compare how each patch improved the performance.
>
> Pranith
>
>>
>>
>> On Thu, Aug 4, 2016 at 10:16 AM, Pranith Kumar Karampuri
>> <pkarampu@xxxxxxxxxx> wrote:
>> >
>> >
>> > On Thu, Aug 4, 2016 at 11:30 AM, Serkan Çoban <cobanserkan@xxxxxxxxx>
>> > wrote:
>> >>
>> >> Thanks Pranith,
>> >> I am waiting for RPMs to show, I will do the tests as soon as possible
>> >> and inform you.
>> >
>> >
>> > I guess on 3.7.x the RPMs are not automatically built. Let me find how
>> > it
>> > can be done. I will inform you after finding that out. Give me a day.
>> >
>> >>
>> >>
>> >> On Wed, Aug 3, 2016 at 11:19 PM, Pranith Kumar Karampuri
>> >> <pkarampu@xxxxxxxxxx> wrote:
>> >> >
>> >> >
>> >> > On Thu, Aug 4, 2016 at 1:47 AM, Pranith Kumar Karampuri
>> >> > <pkarampu@xxxxxxxxxx> wrote:
>> >> >>
>> >> >>
>> >> >>
>> >> >> On Thu, Aug 4, 2016 at 12:51 AM, Serkan Çoban
>> >> >> <cobanserkan@xxxxxxxxx>
>> >> >> wrote:
>> >> >>>
>> >> >>> I use rpms for installation. Redhat/Centos 6.8.
>> >> >>
>> >> >>
>> >> >> http://review.gluster.org/#/c/15084 is the patch. In some time the
>> >> >> rpms
>> >> >> will be built actually.
>> >> >
>> >> >
>> >> > In the same URL above it will actually post the rpms for
>> >> > fedora/el6/el7
>> >> > at
>> >> > the end of the page.
>> >> >
>> >> >>
>> >> >>
>> >> >> Use gluster volume set <volname> disperse.shd-max-threads
>> >> >> <num-threads
>> >> >> (range: 1-64)>
>> >> >>
>> >> >> While testing this I thought of ways to decrease the number of
>> >> >> crawls
>> >> >> as
>> >> >> well. But they are a bit involved. Try to create same set of data
>> >> >> and
>> >> >> see
>> >> >> what is the time it takes to complete heals using number of threads
>> >> >> as
>> >> >> you
>> >> >> increase the number of parallel heals from 1 to 64.
>> >> >>
>> >> >>>
>> >> >>> On Wed, Aug 3, 2016 at 10:16 PM, Pranith Kumar Karampuri
>> >> >>> <pkarampu@xxxxxxxxxx> wrote:
>> >> >>> >
>> >> >>> >
>> >> >>> > On Thu, Aug 4, 2016 at 12:45 AM, Serkan Çoban
>> >> >>> > <cobanserkan@xxxxxxxxx>
>> >> >>> > wrote:
>> >> >>> >>
>> >> >>> >> I prefer 3.7 if it is ok for you. Can you also provide build
>> >> >>> >> instructions?
>> >> >>> >
>> >> >>> >
>> >> >>> > 3.7 should be fine. Do you use rpms/debs/anything-else?
>> >> >>> >
>> >> >>> >>
>> >> >>> >>
>> >> >>> >> On Wed, Aug 3, 2016 at 10:12 PM, Pranith Kumar Karampuri
>> >> >>> >> <pkarampu@xxxxxxxxxx> wrote:
>> >> >>> >> >
>> >> >>> >> >
>> >> >>> >> > On Thu, Aug 4, 2016 at 12:37 AM, Serkan Çoban
>> >> >>> >> > <cobanserkan@xxxxxxxxx>
>> >> >>> >> > wrote:
>> >> >>> >> >>
>> >> >>> >> >> Yes, but I can create 2+1(or 8+2) ec using two servers right?
>> >> >>> >> >> I
>> >> >>> >> >> have
>> >> >>> >> >> 26 disks on each server.
>> >> >>> >> >
>> >> >>> >> >
>> >> >>> >> > On which release-branch do you want the patch? I am testing it
>> >> >>> >> > on
>> >> >>> >> > master-branch now.
>> >> >>> >> >
>> >> >>> >> >>
>> >> >>> >> >>
>> >> >>> >> >> On Wed, Aug 3, 2016 at 9:59 PM, Pranith Kumar Karampuri
>> >> >>> >> >> <pkarampu@xxxxxxxxxx> wrote:
>> >> >>> >> >> >
>> >> >>> >> >> >
>> >> >>> >> >> > On Thu, Aug 4, 2016 at 12:23 AM, Serkan Çoban
>> >> >>> >> >> > <cobanserkan@xxxxxxxxx>
>> >> >>> >> >> > wrote:
>> >> >>> >> >> >>
>> >> >>> >> >> >> I have two of my storage servers free, I think I can use
>> >> >>> >> >> >> them
>> >> >>> >> >> >> for
>> >> >>> >> >> >> testing. Is two server testing environment ok for you?
>> >> >>> >> >> >
>> >> >>> >> >> >
>> >> >>> >> >> > I think it would be better if you have at least 3. You can
>> >> >>> >> >> > test
>> >> >>> >> >> > it
>> >> >>> >> >> > with
>> >> >>> >> >> > 2+1
>> >> >>> >> >> > ec configuration.
>> >> >>> >> >> >
>> >> >>> >> >> >>
>> >> >>> >> >> >>
>> >> >>> >> >> >> On Wed, Aug 3, 2016 at 9:44 PM, Pranith Kumar Karampuri
>> >> >>> >> >> >> <pkarampu@xxxxxxxxxx> wrote:
>> >> >>> >> >> >> >
>> >> >>> >> >> >> >
>> >> >>> >> >> >> > On Wed, Aug 3, 2016 at 6:01 PM, Serkan Çoban
>> >> >>> >> >> >> > <cobanserkan@xxxxxxxxx>
>> >> >>> >> >> >> > wrote:
>> >> >>> >> >> >> >>
>> >> >>> >> >> >> >> Hi,
>> >> >>> >> >> >> >>
>> >> >>> >> >> >> >> May I ask if multi-threaded self heal for distributed
>> >> >>> >> >> >> >> disperse
>> >> >>> >> >> >> >> volumes
>> >> >>> >> >> >> >> implemented in this release?
>> >> >>> >> >> >> >
>> >> >>> >> >> >> >
>> >> >>> >> >> >> > Serkan,
>> >> >>> >> >> >> > At the moment I am a bit busy with different
>> >> >>> >> >> >> > work,
>> >> >>> >> >> >> > Is
>> >> >>> >> >> >> > it
>> >> >>> >> >> >> > possible
>> >> >>> >> >> >> > for you to help test the feature if I provide a patch?
>> >> >>> >> >> >> > Actually
>> >> >>> >> >> >> > the
>> >> >>> >> >> >> > patch
>> >> >>> >> >> >> > should be small. Testing is where lot of time will be
>> >> >>> >> >> >> > spent
>> >> >>> >> >> >> > on.
>> >> >>> >> >> >> >
>> >> >>> >> >> >> >>
>> >> >>> >> >> >> >>
>> >> >>> >> >> >> >> Thanks,
>> >> >>> >> >> >> >> Serkan
>> >> >>> >> >> >> >>
>> >> >>> >> >> >> >> On Tue, Aug 2, 2016 at 5:30 PM, David Gossage
>> >> >>> >> >> >> >> <dgossage@xxxxxxxxxxxxxxxxxx> wrote:
>> >> >>> >> >> >> >> > On Tue, Aug 2, 2016 at 6:01 AM, Lindsay Mathieson
>> >> >>> >> >> >> >> > <lindsay.mathieson@xxxxxxxxx> wrote:
>> >> >>> >> >> >> >> >>
>> >> >>> >> >> >> >> >> On 2/08/2016 5:07 PM, Kaushal M wrote:
>> >> >>> >> >> >> >> >>>
>> >> >>> >> >> >> >> >>> GlusterFS-3.7.14 has been released. This is a
>> >> >>> >> >> >> >> >>> regular
>> >> >>> >> >> >> >> >>> minor
>> >> >>> >> >> >> >> >>> release.
>> >> >>> >> >> >> >> >>> The release-notes are available at
>> >> >>> >> >> >> >> >>>
>> >> >>> >> >> >> >> >>>
>> >> >>> >> >> >> >> >>>
>> >> >>> >> >> >> >> >>>
>> >> >>> >> >> >> >> >>>
>> >> >>> >> >> >> >> >>>
>> >> >>> >> >> >> >> >>>
>> >> >>> >> >> >> >> >>>
>> >> >>> >> >> >> >> >>> https://github.com/gluster/glusterfs/blob/release-3.7/ doc/release-notes/3.7.14.md
>> >> >>> >> >> >> >> >>
>> >> >>> >> >> >> >> >>
>> >> >>> >> >> >> >> >> Thanks Kaushal, I'll check it out
>> >> >>> >> >> >> >> >>
>> >> >>> >> >> >> >> >
>> >> >>> >> >> >> >> > So far on my test box its working as expected. At
>> >> >>> >> >> >> >> > least
>> >> >>> >> >> >> >> > the
>> >> >>> >> >> >> >> > issues
>> >> >>> >> >> >> >> > that
>> >> >>> >> >> >> >> > prevented it from running as before have disappeared.
>> >> >>> >> >> >> >> > Will
>> >> >>> >> >> >> >> > need
>> >> >>> >> >> >> >> > to
>> >> >>> >> >> >> >> > see
>> >> >>> >> >> >> >> > how
>> >> >>> >> >> >> >> > my test VM behaves after a few days.
>> >> >>> >> >> >> >> >
>> >> >>> >> >> >> >> >
>> >> >>> >> >> >> >> >
>> >> >>> >> >> >> >> >> --
>> >> >>> >> >> >> >> >> Lindsay Mathieson
>> >> >>> >> >> >> >> >>
>> >> >>> >> >> >> >> >> _______________________________________________
>> >> >>> >> >> >> >> >> Gluster-users mailing list
>> >> >>> >> >> >> >> >> Gluster-users@xxxxxxxxxxx
>> >> >>> >> >> >> >> >>
>> >> >>> >> >> >> >> >> http://www.gluster.org/mailman/listinfo/gluster-users
>> >> >>> >> >> >> >> >
>> >> >>> >> >> >> >> >
>> >> >>> >> >> >> >> >
>> >> >>> >> >> >> >> > _______________________________________________
>> >> >>> >> >> >> >> > Gluster-users mailing list
>> >> >>> >> >> >> >> > Gluster-users@xxxxxxxxxxx
>> >> >>> >> >> >> >> > http://www.gluster.org/mailman/listinfo/gluster-users
>> >> >>> >> >> >> >> _______________________________________________
>> >> >>> >> >> >> >> Gluster-users mailing list
>> >> >>> >> >> >> >> Gluster-users@xxxxxxxxxxx
>> >> >>> >> >> >> >> http://www.gluster.org/mailman/listinfo/gluster-users
>> >> >>> >> >> >> >
>> >> >>> >> >> >> >
>> >> >>> >> >> >> >
>> >> >>> >> >> >> >
>> >> >>> >> >> >> > --
>> >> >>> >> >> >> > Pranith
>> >> >>> >> >> >
>> >> >>> >> >> >
>> >> >>> >> >> >
>> >> >>> >> >> >
>> >> >>> >> >> > --
>> >> >>> >> >> > Pranith
>> >> >>> >> >
>> >> >>> >> >
>> >> >>> >> >
>> >> >>> >> >
>> >> >>> >> > --
>> >> >>> >> > Pranith
>> >> >>> >
>> >> >>> >
>> >> >>> >
>> >> >>> >
>> >> >>> > --
>> >> >>> > Pranith
>> >> >>
>> >> >>
>> >> >>
>> >> >>
>> >> >> --
>> >> >> Pranith
>> >> >
>> >> >
>> >> >
>> >> >
>> >> > --
>> >> > Pranith
>> >
>> >
>> >
>> >
>> > --
>> > Pranith
>
>
>
>
> --
> Pranith
--
Pranith
_______________________________________________ Gluster-devel mailing list Gluster-devel@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-devel