On Thu, Aug 4, 2016 at 1:47 AM, Pranith Kumar Karampuri <pkarampu@xxxxxxxxxx> wrote:
On Thu, Aug 4, 2016 at 12:51 AM, Serkan Çoban <cobanserkan@xxxxxxxxx> wrote:I use rpms for installation. Redhat/Centos 6.8.
In the same URL above it will actually post the rpms for fedora/el6/el7 at the end of the page.
Use gluster volume set <volname> disperse.shd-max-threads <num-threads (range: 1-64)>While testing this I thought of ways to decrease the number of crawls as well. But they are a bit involved. Try to create same set of data and see what is the time it takes to complete heals using number of threads as you increase the number of parallel heals from 1 to 64.
On Wed, Aug 3, 2016 at 10:16 PM, Pranith Kumar Karampuri
<pkarampu@xxxxxxxxxx> wrote:
>
>
> On Thu, Aug 4, 2016 at 12:45 AM, Serkan Çoban <cobanserkan@xxxxxxxxx> wrote:
>>
>> I prefer 3.7 if it is ok for you. Can you also provide build instructions?
>
>
> 3.7 should be fine. Do you use rpms/debs/anything-else?
>
>>
>>
>> On Wed, Aug 3, 2016 at 10:12 PM, Pranith Kumar Karampuri
>> <pkarampu@xxxxxxxxxx> wrote:
>> >
>> >
>> > On Thu, Aug 4, 2016 at 12:37 AM, Serkan Çoban <cobanserkan@xxxxxxxxx>
>> > wrote:
>> >>
>> >> Yes, but I can create 2+1(or 8+2) ec using two servers right? I have
>> >> 26 disks on each server.
>> >
>> >
>> > On which release-branch do you want the patch? I am testing it on
>> > master-branch now.
>> >
>> >>
>> >>
>> >> On Wed, Aug 3, 2016 at 9:59 PM, Pranith Kumar Karampuri
>> >> <pkarampu@xxxxxxxxxx> wrote:
>> >> >
>> >> >
>> >> > On Thu, Aug 4, 2016 at 12:23 AM, Serkan Çoban <cobanserkan@xxxxxxxxx>
>> >> > wrote:
>> >> >>
>> >> >> I have two of my storage servers free, I think I can use them for
>> >> >> testing. Is two server testing environment ok for you?
>> >> >
>> >> >
>> >> > I think it would be better if you have at least 3. You can test it
>> >> > with
>> >> > 2+1
>> >> > ec configuration.
>> >> >
>> >> >>
>> >> >>
>> >> >> On Wed, Aug 3, 2016 at 9:44 PM, Pranith Kumar Karampuri
>> >> >> <pkarampu@xxxxxxxxxx> wrote:
>> >> >> >
>> >> >> >
>> >> >> > On Wed, Aug 3, 2016 at 6:01 PM, Serkan Çoban
>> >> >> > <cobanserkan@xxxxxxxxx>
>> >> >> > wrote:
>> >> >> >>
>> >> >> >> Hi,
>> >> >> >>
>> >> >> >> May I ask if multi-threaded self heal for distributed disperse
>> >> >> >> volumes
>> >> >> >> implemented in this release?
>> >> >> >
>> >> >> >
>> >> >> > Serkan,
>> >> >> > At the moment I am a bit busy with different work, Is it
>> >> >> > possible
>> >> >> > for you to help test the feature if I provide a patch? Actually
>> >> >> > the
>> >> >> > patch
>> >> >> > should be small. Testing is where lot of time will be spent on.
>> >> >> >
>> >> >> >>
>> >> >> >>
>> >> >> >> Thanks,
>> >> >> >> Serkan
>> >> >> >>
>> >> >> >> On Tue, Aug 2, 2016 at 5:30 PM, David Gossage
>> >> >> >> <dgossage@xxxxxxxxxxxxxxxxxx> wrote:
>> >> >> >> > On Tue, Aug 2, 2016 at 6:01 AM, Lindsay Mathieson
>> >> >> >> > <lindsay.mathieson@xxxxxxxxx> wrote:
>> >> >> >> >>
>> >> >> >> >> On 2/08/2016 5:07 PM, Kaushal M wrote:
>> >> >> >> >>>
>> >> >> >> >>> GlusterFS-3.7.14 has been released. This is a regular minor
>> >> >> >> >>> release.
>> >> >> >> >>> The release-notes are available at
>> >> >> >> >>>
>> >> >> >> >>>
>> >> >> >> >>>
>> >> >> >> >>>
>> >> >> >> >>>
>> >> >> >> >>> https://github.com/gluster/glusterfs/blob/release-3.7/doc/release-notes/3.7.14.md
>> >> >> >> >>
>> >> >> >> >>
>> >> >> >> >> Thanks Kaushal, I'll check it out
>> >> >> >> >>
>> >> >> >> >
>> >> >> >> > So far on my test box its working as expected. At least the
>> >> >> >> > issues
>> >> >> >> > that
>> >> >> >> > prevented it from running as before have disappeared. Will
>> >> >> >> > need
>> >> >> >> > to
>> >> >> >> > see
>> >> >> >> > how
>> >> >> >> > my test VM behaves after a few days.
>> >> >> >> >
>> >> >> >> >
>> >> >> >> >
>> >> >> >> >> --
>> >> >> >> >> Lindsay Mathieson
>> >> >> >> >>
>> >> >> >> >> _______________________________________________
>> >> >> >> >> Gluster-users mailing list
>> >> >> >> >> Gluster-users@xxxxxxxxxxx
>> >> >> >> >> http://www.gluster.org/mailman/listinfo/gluster-users
>> >> >> >> >
>> >> >> >> >
>> >> >> >> >
>> >> >> >> > _______________________________________________
>> >> >> >> > Gluster-users mailing list
>> >> >> >> > Gluster-users@xxxxxxxxxxx
>> >> >> >> > http://www.gluster.org/mailman/listinfo/gluster-users
>> >> >> >> _______________________________________________
>> >> >> >> Gluster-users mailing list
>> >> >> >> Gluster-users@xxxxxxxxxxx
>> >> >> >> http://www.gluster.org/mailman/listinfo/gluster-users
>> >> >> >
>> >> >> >
>> >> >> >
>> >> >> >
>> >> >> > --
>> >> >> > Pranith
>> >> >
>> >> >
>> >> >
>> >> >
>> >> > --
>> >> > Pranith
>> >
>> >
>> >
>> >
>> > --
>> > Pranith
>
>
>
>
> --
> Pranith
--Pranith
--
Pranith
_______________________________________________ Gluster-devel mailing list Gluster-devel@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-devel