hi, ceph osd set noscrub(or nodeep-scrub) would stop the scrub forever. and ceph osd unset noscrub would continue to reschedule scrub. so maybe you could use this two command in crontab to schedule scrub manually. 2015-07-30 7:59 GMT+08:00 Quentin Hartman <qhartman@xxxxxxxxxxxxxxxxxxx>: > well, that would certainly do it. I _always_ forget to twiddle the little > thing on the web page that changes the version of the docs I'm looking at. > > So I guess then my question becomes, "How do i prevent deep scrubs from > happening in the middle of the day and ruining everything?" > > QH > > > On Wed, Jul 29, 2015 at 5:55 PM, Travis Rhoden <trhoden@xxxxxxxxx> wrote: >> >> Hi Quentin, >> >> It may be the specific option you are trying to tweak. >> osd-scrub-begin-hour was first introduced in development release >> v0.93, which means it would be in 0.94.x (Hammer), but your cluster is >> 0.87.1 (Giant). >> >> Cheers, >> >> - Travis >> >> On Wed, Jul 29, 2015 at 4:28 PM, Quentin Hartman >> <qhartman@xxxxxxxxxxxxxxxxxxx> wrote: >> > I'm running a 0.87.1 cluster, and my "ceph tell" seems to not be >> > working: >> > >> > # ceph tell osd.0 injectargs '--osd-scrub-begin-hour 1' >> > failed to parse arguments: --osd-scrub-begin-hour,1 >> > >> > >> > I've also tried the daemon config set variant and it also fails: >> > >> > # ceph daemon osd.0 config set osd_scrub_begin_hour 1 >> > { "error": "error setting 'osd_scrub_begin_hour' to '1': (2) No such >> > file or >> > directory"} >> > >> > I'm guessing I have something goofed in my admin socket client config: >> > >> > [client] >> > rbd cache = true >> > rbd cache writethrough until flush = true >> > admin socket = /var/run/ceph/$cluster-$type.$id.asok >> > >> > but that seems to correlate with the structure that exists: >> > >> > # ls >> > ceph-osd.24.asok ceph-osd.25.asok ceph-osd.26.asok >> > # pwd >> > /var/run/ceph >> > >> > I can show my configs all over the place, but changing them seems to >> > always >> > fail. It behaves the same if I'm working on a local daemon, or on my >> > config >> > node trying to make changes globally. >> > >> > Thanks in advance for any ideas >> > >> > QH >> > >> > >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@xxxxxxxxxxxxxx >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > > > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > -- Regards, xinze _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com