Re: PG auto repair with BlueStore

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



> Wido den Hollander <wido@xxxxxxxx>:
> On 11/15/18 7:51 PM, koukou73gr wrote:
> > Are there any means to notify the administrator that an auto-repair has
> > taken place?
>
> I don't think so. You'll see the cluster go to HEALTH_ERR for a while
> before it turns to HEALTH_OK again after the PG has been repaired.

and I think even this is too much. No point in triggering a monitoring
system in the middle of the night when the scrubs are running just
because of some bit rot on a disk. Losing a few bits on disks here and
there is a perfectly normal and expected scenario that Ceph can take
care of all by itself without triggering an health *error*. It
certainly doesn't require immediate attention (with auto repair
enabled) like the error state indicates.
The message in the cluster log should be enough.

-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90

Am Fr., 16. Nov. 2018 um 08:26 Uhr schrieb Wido den Hollander <wido@xxxxxxxx>:
>
>
>
> On 11/15/18 7:51 PM, koukou73gr wrote:
> > Are there any means to notify the administrator that an auto-repair has
> > taken place?
>
> I don't think so. You'll see the cluster go to HEALTH_ERR for a while
> before it turns to HEALTH_OK again after the PG has been repaired.
>
> You would have to search the cluster logs to find out that a auto repair
> took place on a Placement Group.
>
> Wido
>
> >
> > -K.
> >
> >
> > On 2018-11-15 20:45, Mark Schouten wrote:
> >> As a user, I’m very surprised that this isn’t a default setting.
> >>
> >> Mark Schouten
> >>
> >>> Op 15 nov. 2018 om 18:40 heeft Wido den Hollander <wido@xxxxxxxx> het
> >>> volgende geschreven:
> >>>
> >>> Hi,
> >>>
> >>> This question is actually still outstanding. Is there any good reason to
> >>> keep auto repair for scrub errors disabled with BlueStore?
> >>>
> >>> I couldn't think of a reason when using size=3 and min_size=2, so just
> >>> wondering.
> >>>
> >>> Thanks!
> >>>
> >>> Wido
> >>>
> >>>> On 8/24/18 8:55 AM, Wido den Hollander wrote:
> >>>> Hi,
> >>>>
> >>>> osd_scrub_auto_repair still defaults to false and I was wondering
> >>>> how we
> >>>> think about enabling this feature by default.
> >>>>
> >>>> Would we say it's safe to enable this with BlueStore?
> >>>>
> >>>> Wido
> >>>> _______________________________________________
> >>>> ceph-users mailing list
> >>>> ceph-users@xxxxxxxxxxxxxx
> >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >>>>
> >>> _______________________________________________
> >>> ceph-users mailing list
> >>> ceph-users@xxxxxxxxxxxxxx
> >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >> _______________________________________________
> >> ceph-users mailing list
> >> ceph-users@xxxxxxxxxxxxxx
> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >>
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@xxxxxxxxxxxxxx
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux