I can wait for the patch to complete, please inform me when you ready. If it will take too much time to solve the crawl issue I can test without it too... Serkan On Thu, Aug 11, 2016 at 5:52 AM, Pranith Kumar Karampuri <pkarampu@xxxxxxxxxx> wrote: > > > On Wed, Aug 10, 2016 at 1:58 PM, Serkan Çoban <cobanserkan@xxxxxxxxx> wrote: >> >> Hi, >> >> Any progress about the patch? > > > hi Serkan, > While testing the patch by myself, I am seeing that it is taking more > than one crawl to complete heals even when there are no directory > hierarchies. It is faster than before but it shouldn't take more than 1 > crawl to complete the heal because all the files exist already. I am > investigating why that is the case now. If you want to test things out > without this patch I will give you rpms today. Otherwise we need to find > until we find RCA for this crawl problem. Let me know your decision. If you > are okay with testing progressive versions of this feature, that would be > great. We can compare how each patch improved the performance. > > Pranith > >> >> >> On Thu, Aug 4, 2016 at 10:16 AM, Pranith Kumar Karampuri >> <pkarampu@xxxxxxxxxx> wrote: >> > >> > >> > On Thu, Aug 4, 2016 at 11:30 AM, Serkan Çoban <cobanserkan@xxxxxxxxx> >> > wrote: >> >> >> >> Thanks Pranith, >> >> I am waiting for RPMs to show, I will do the tests as soon as possible >> >> and inform you. >> > >> > >> > I guess on 3.7.x the RPMs are not automatically built. Let me find how >> > it >> > can be done. I will inform you after finding that out. Give me a day. >> > >> >> >> >> >> >> On Wed, Aug 3, 2016 at 11:19 PM, Pranith Kumar Karampuri >> >> <pkarampu@xxxxxxxxxx> wrote: >> >> > >> >> > >> >> > On Thu, Aug 4, 2016 at 1:47 AM, Pranith Kumar Karampuri >> >> > <pkarampu@xxxxxxxxxx> wrote: >> >> >> >> >> >> >> >> >> >> >> >> On Thu, Aug 4, 2016 at 12:51 AM, Serkan Çoban >> >> >> <cobanserkan@xxxxxxxxx> >> >> >> wrote: >> >> >>> >> >> >>> I use rpms for installation. Redhat/Centos 6.8. >> >> >> >> >> >> >> >> >> http://review.gluster.org/#/c/15084 is the patch. In some time the >> >> >> rpms >> >> >> will be built actually. >> >> > >> >> > >> >> > In the same URL above it will actually post the rpms for >> >> > fedora/el6/el7 >> >> > at >> >> > the end of the page. >> >> > >> >> >> >> >> >> >> >> >> Use gluster volume set <volname> disperse.shd-max-threads >> >> >> <num-threads >> >> >> (range: 1-64)> >> >> >> >> >> >> While testing this I thought of ways to decrease the number of >> >> >> crawls >> >> >> as >> >> >> well. But they are a bit involved. Try to create same set of data >> >> >> and >> >> >> see >> >> >> what is the time it takes to complete heals using number of threads >> >> >> as >> >> >> you >> >> >> increase the number of parallel heals from 1 to 64. >> >> >> >> >> >>> >> >> >>> On Wed, Aug 3, 2016 at 10:16 PM, Pranith Kumar Karampuri >> >> >>> <pkarampu@xxxxxxxxxx> wrote: >> >> >>> > >> >> >>> > >> >> >>> > On Thu, Aug 4, 2016 at 12:45 AM, Serkan Çoban >> >> >>> > <cobanserkan@xxxxxxxxx> >> >> >>> > wrote: >> >> >>> >> >> >> >>> >> I prefer 3.7 if it is ok for you. Can you also provide build >> >> >>> >> instructions? >> >> >>> > >> >> >>> > >> >> >>> > 3.7 should be fine. Do you use rpms/debs/anything-else? >> >> >>> > >> >> >>> >> >> >> >>> >> >> >> >>> >> On Wed, Aug 3, 2016 at 10:12 PM, Pranith Kumar Karampuri >> >> >>> >> <pkarampu@xxxxxxxxxx> wrote: >> >> >>> >> > >> >> >>> >> > >> >> >>> >> > On Thu, Aug 4, 2016 at 12:37 AM, Serkan Çoban >> >> >>> >> > <cobanserkan@xxxxxxxxx> >> >> >>> >> > wrote: >> >> >>> >> >> >> >> >>> >> >> Yes, but I can create 2+1(or 8+2) ec using two servers right? >> >> >>> >> >> I >> >> >>> >> >> have >> >> >>> >> >> 26 disks on each server. >> >> >>> >> > >> >> >>> >> > >> >> >>> >> > On which release-branch do you want the patch? I am testing it >> >> >>> >> > on >> >> >>> >> > master-branch now. >> >> >>> >> > >> >> >>> >> >> >> >> >>> >> >> >> >> >>> >> >> On Wed, Aug 3, 2016 at 9:59 PM, Pranith Kumar Karampuri >> >> >>> >> >> <pkarampu@xxxxxxxxxx> wrote: >> >> >>> >> >> > >> >> >>> >> >> > >> >> >>> >> >> > On Thu, Aug 4, 2016 at 12:23 AM, Serkan Çoban >> >> >>> >> >> > <cobanserkan@xxxxxxxxx> >> >> >>> >> >> > wrote: >> >> >>> >> >> >> >> >> >>> >> >> >> I have two of my storage servers free, I think I can use >> >> >>> >> >> >> them >> >> >>> >> >> >> for >> >> >>> >> >> >> testing. Is two server testing environment ok for you? >> >> >>> >> >> > >> >> >>> >> >> > >> >> >>> >> >> > I think it would be better if you have at least 3. You can >> >> >>> >> >> > test >> >> >>> >> >> > it >> >> >>> >> >> > with >> >> >>> >> >> > 2+1 >> >> >>> >> >> > ec configuration. >> >> >>> >> >> > >> >> >>> >> >> >> >> >> >>> >> >> >> >> >> >>> >> >> >> On Wed, Aug 3, 2016 at 9:44 PM, Pranith Kumar Karampuri >> >> >>> >> >> >> <pkarampu@xxxxxxxxxx> wrote: >> >> >>> >> >> >> > >> >> >>> >> >> >> > >> >> >>> >> >> >> > On Wed, Aug 3, 2016 at 6:01 PM, Serkan Çoban >> >> >>> >> >> >> > <cobanserkan@xxxxxxxxx> >> >> >>> >> >> >> > wrote: >> >> >>> >> >> >> >> >> >> >>> >> >> >> >> Hi, >> >> >>> >> >> >> >> >> >> >>> >> >> >> >> May I ask if multi-threaded self heal for distributed >> >> >>> >> >> >> >> disperse >> >> >>> >> >> >> >> volumes >> >> >>> >> >> >> >> implemented in this release? >> >> >>> >> >> >> > >> >> >>> >> >> >> > >> >> >>> >> >> >> > Serkan, >> >> >>> >> >> >> > At the moment I am a bit busy with different >> >> >>> >> >> >> > work, >> >> >>> >> >> >> > Is >> >> >>> >> >> >> > it >> >> >>> >> >> >> > possible >> >> >>> >> >> >> > for you to help test the feature if I provide a patch? >> >> >>> >> >> >> > Actually >> >> >>> >> >> >> > the >> >> >>> >> >> >> > patch >> >> >>> >> >> >> > should be small. Testing is where lot of time will be >> >> >>> >> >> >> > spent >> >> >>> >> >> >> > on. >> >> >>> >> >> >> > >> >> >>> >> >> >> >> >> >> >>> >> >> >> >> >> >> >>> >> >> >> >> Thanks, >> >> >>> >> >> >> >> Serkan >> >> >>> >> >> >> >> >> >> >>> >> >> >> >> On Tue, Aug 2, 2016 at 5:30 PM, David Gossage >> >> >>> >> >> >> >> <dgossage@xxxxxxxxxxxxxxxxxx> wrote: >> >> >>> >> >> >> >> > On Tue, Aug 2, 2016 at 6:01 AM, Lindsay Mathieson >> >> >>> >> >> >> >> > <lindsay.mathieson@xxxxxxxxx> wrote: >> >> >>> >> >> >> >> >> >> >> >>> >> >> >> >> >> On 2/08/2016 5:07 PM, Kaushal M wrote: >> >> >>> >> >> >> >> >>> >> >> >>> >> >> >> >> >>> GlusterFS-3.7.14 has been released. This is a >> >> >>> >> >> >> >> >>> regular >> >> >>> >> >> >> >> >>> minor >> >> >>> >> >> >> >> >>> release. >> >> >>> >> >> >> >> >>> The release-notes are available at >> >> >>> >> >> >> >> >>> >> >> >>> >> >> >> >> >>> >> >> >>> >> >> >> >> >>> >> >> >>> >> >> >> >> >>> >> >> >>> >> >> >> >> >>> >> >> >>> >> >> >> >> >>> >> >> >>> >> >> >> >> >>> >> >> >>> >> >> >> >> >>> >> >> >>> >> >> >> >> >>> https://github.com/gluster/glusterfs/blob/release-3.7/doc/release-notes/3.7.14.md >> >> >>> >> >> >> >> >> >> >> >>> >> >> >> >> >> >> >> >>> >> >> >> >> >> Thanks Kaushal, I'll check it out >> >> >>> >> >> >> >> >> >> >> >>> >> >> >> >> > >> >> >>> >> >> >> >> > So far on my test box its working as expected. At >> >> >>> >> >> >> >> > least >> >> >>> >> >> >> >> > the >> >> >>> >> >> >> >> > issues >> >> >>> >> >> >> >> > that >> >> >>> >> >> >> >> > prevented it from running as before have disappeared. >> >> >>> >> >> >> >> > Will >> >> >>> >> >> >> >> > need >> >> >>> >> >> >> >> > to >> >> >>> >> >> >> >> > see >> >> >>> >> >> >> >> > how >> >> >>> >> >> >> >> > my test VM behaves after a few days. >> >> >>> >> >> >> >> > >> >> >>> >> >> >> >> > >> >> >>> >> >> >> >> > >> >> >>> >> >> >> >> >> -- >> >> >>> >> >> >> >> >> Lindsay Mathieson >> >> >>> >> >> >> >> >> >> >> >>> >> >> >> >> >> _______________________________________________ >> >> >>> >> >> >> >> >> Gluster-users mailing list >> >> >>> >> >> >> >> >> Gluster-users@xxxxxxxxxxx >> >> >>> >> >> >> >> >> >> >> >>> >> >> >> >> >> http://www.gluster.org/mailman/listinfo/gluster-users >> >> >>> >> >> >> >> > >> >> >>> >> >> >> >> > >> >> >>> >> >> >> >> > >> >> >>> >> >> >> >> > _______________________________________________ >> >> >>> >> >> >> >> > Gluster-users mailing list >> >> >>> >> >> >> >> > Gluster-users@xxxxxxxxxxx >> >> >>> >> >> >> >> > http://www.gluster.org/mailman/listinfo/gluster-users >> >> >>> >> >> >> >> _______________________________________________ >> >> >>> >> >> >> >> Gluster-users mailing list >> >> >>> >> >> >> >> Gluster-users@xxxxxxxxxxx >> >> >>> >> >> >> >> http://www.gluster.org/mailman/listinfo/gluster-users >> >> >>> >> >> >> > >> >> >>> >> >> >> > >> >> >>> >> >> >> > >> >> >>> >> >> >> > >> >> >>> >> >> >> > -- >> >> >>> >> >> >> > Pranith >> >> >>> >> >> > >> >> >>> >> >> > >> >> >>> >> >> > >> >> >>> >> >> > >> >> >>> >> >> > -- >> >> >>> >> >> > Pranith >> >> >>> >> > >> >> >>> >> > >> >> >>> >> > >> >> >>> >> > >> >> >>> >> > -- >> >> >>> >> > Pranith >> >> >>> > >> >> >>> > >> >> >>> > >> >> >>> > >> >> >>> > -- >> >> >>> > Pranith >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> -- >> >> >> Pranith >> >> > >> >> > >> >> > >> >> > >> >> > -- >> >> > Pranith >> > >> > >> > >> > >> > -- >> > Pranith > > > > > -- > Pranith _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users