On Thu, Aug 4, 2016 at 12:51 AM, Serkan ?oban <cobanserkan at gmail.com>
wrote:
> I use rpms for installation. Redhat/Centos 6.8.
>
http://review.gluster.org/#/c/15084 is the patch. In some time the rpms
will be built actually.
Use gluster volume set <volname> disperse.shd-max-threads <num-threads
(range: 1-64)>
While testing this I thought of ways to decrease the number of crawls as
well. But they are a bit involved. Try to create same set of data and see
what is the time it takes to complete heals using number of threads as you
increase the number of parallel heals from 1 to 64.
> On Wed, Aug 3, 2016 at 10:16 PM, Pranith Kumar Karampuri
> <pkarampu at redhat.com> wrote:
> >
> >
> > On Thu, Aug 4, 2016 at 12:45 AM, Serkan ?oban <cobanserkan at
gmail.com>
> wrote:
> >>
> >> I prefer 3.7 if it is ok for you. Can you also provide build
> instructions?
> >
> >
> > 3.7 should be fine. Do you use rpms/debs/anything-else?
> >
> >>
> >>
> >> On Wed, Aug 3, 2016 at 10:12 PM, Pranith Kumar Karampuri
> >> <pkarampu at redhat.com> wrote:
> >> >
> >> >
> >> > On Thu, Aug 4, 2016 at 12:37 AM, Serkan ?oban <cobanserkan
at gmail.com>
> >> > wrote:
> >> >>
> >> >> Yes, but I can create 2+1(or 8+2) ec using two servers
right? I have
> >> >> 26 disks on each server.
> >> >
> >> >
> >> > On which release-branch do you want the patch? I am testing
it on
> >> > master-branch now.
> >> >
> >> >>
> >> >>
> >> >> On Wed, Aug 3, 2016 at 9:59 PM, Pranith Kumar Karampuri
> >> >> <pkarampu at redhat.com> wrote:
> >> >> >
> >> >> >
> >> >> > On Thu, Aug 4, 2016 at 12:23 AM, Serkan ?oban <
> cobanserkan at gmail.com>
> >> >> > wrote:
> >> >> >>
> >> >> >> I have two of my storage servers free, I think I
can use them for
> >> >> >> testing. Is two server testing environment ok
for you?
> >> >> >
> >> >> >
> >> >> > I think it would be better if you have at least 3.
You can test it
> >> >> > with
> >> >> > 2+1
> >> >> > ec configuration.
> >> >> >
> >> >> >>
> >> >> >>
> >> >> >> On Wed, Aug 3, 2016 at 9:44 PM, Pranith Kumar
Karampuri
> >> >> >> <pkarampu at redhat.com> wrote:
> >> >> >> >
> >> >> >> >
> >> >> >> > On Wed, Aug 3, 2016 at 6:01 PM, Serkan
?oban
> >> >> >> > <cobanserkan at gmail.com>
> >> >> >> > wrote:
> >> >> >> >>
> >> >> >> >> Hi,
> >> >> >> >>
> >> >> >> >> May I ask if multi-threaded self heal
for distributed disperse
> >> >> >> >> volumes
> >> >> >> >> implemented in this release?
> >> >> >> >
> >> >> >> >
> >> >> >> > Serkan,
> >> >> >> > At the moment I am a bit busy with
different work, Is it
> >> >> >> > possible
> >> >> >> > for you to help test the feature if I
provide a patch? Actually
> >> >> >> > the
> >> >> >> > patch
> >> >> >> > should be small. Testing is where lot of
time will be spent on.
> >> >> >> >
> >> >> >> >>
> >> >> >> >>
> >> >> >> >> Thanks,
> >> >> >> >> Serkan
> >> >> >> >>
> >> >> >> >> On Tue, Aug 2, 2016 at 5:30 PM, David
Gossage
> >> >> >> >> <dgossage at carouselchecks.com>
wrote:
> >> >> >> >> > On Tue, Aug 2, 2016 at 6:01 AM,
Lindsay Mathieson
> >> >> >> >> > <lindsay.mathieson at
gmail.com> wrote:
> >> >> >> >> >>
> >> >> >> >> >> On 2/08/2016 5:07 PM, Kaushal
M wrote:
> >> >> >> >> >>>
> >> >> >> >> >>> GlusterFS-3.7.14 has been
released. This is a regular minor
> >> >> >> >> >>> release.
> >> >> >> >> >>> The release-notes are
available at
> >> >> >> >> >>>
> >> >> >> >> >>>
> >> >> >> >> >>>
> >> >> >> >> >>>
> >> >> >> >> >>>
> >> >> >> >> >>>
>
https://github.com/gluster/glusterfs/blob/release-3.7/doc/release-notes/3.7.14.md
> >> >> >> >> >>
> >> >> >> >> >>
> >> >> >> >> >> Thanks Kaushal, I'll check
it out
> >> >> >> >> >>
> >> >> >> >> >
> >> >> >> >> > So far on my test box its working
as expected. At least the
> >> >> >> >> > issues
> >> >> >> >> > that
> >> >> >> >> > prevented it from running as
before have disappeared. Will
> >> >> >> >> > need
> >> >> >> >> > to
> >> >> >> >> > see
> >> >> >> >> > how
> >> >> >> >> > my test VM behaves after a few
days.
> >> >> >> >> >
> >> >> >> >> >
> >> >> >> >> >
> >> >> >> >> >> --
> >> >> >> >> >> Lindsay Mathieson
> >> >> >> >> >>
> >> >> >> >> >>
_______________________________________________
> >> >> >> >> >> Gluster-users mailing list
> >> >> >> >> >> Gluster-users at gluster.org
> >> >> >> >> >>
http://www.gluster.org/mailman/listinfo/gluster-users
> >> >> >> >> >
> >> >> >> >> >
> >> >> >> >> >
> >> >> >> >> >
_______________________________________________
> >> >> >> >> > Gluster-users mailing list
> >> >> >> >> > Gluster-users at gluster.org
> >> >> >> >> >
http://www.gluster.org/mailman/listinfo/gluster-users
> >> >> >> >>
_______________________________________________
> >> >> >> >> Gluster-users mailing list
> >> >> >> >> Gluster-users at gluster.org
> >> >> >> >>
http://www.gluster.org/mailman/listinfo/gluster-users
> >> >> >> >
> >> >> >> >
> >> >> >> >
> >> >> >> >
> >> >> >> > --
> >> >> >> > Pranith
> >> >> >
> >> >> >
> >> >> >
> >> >> >
> >> >> > --
> >> >> > Pranith
> >> >
> >> >
> >> >
> >> >
> >> > --
> >> > Pranith
> >
> >
> >
> >
> > --
> > Pranith
>
--
Pranith
-------------- next part --------------
An HTML attachment was scrubbed...
URL:
<http://www.gluster.org/pipermail/gluster-users/attachments/20160804/2fa81d1b/attachment.html>