Hi, I have all the normal options set in ceph.conf (priority and class for disk threads) however scrubs look like they are running as the standard BE/4 class in iotop. Running 10.2.3. Eg PG Dump (Shows that OSD 1 will be scrubbing) pg_stat objects mip degr misp unf bytes log disklog state state_stamp v reported up up_primary acting acting_primary last_scrub scrub_stamp last_deep_scrub deep_scrub_stamp 1.8fc 3314 0 0 0 0 13899923456 3043 3043 active+clean+scrubbing+deep 2016-11-08 07:51:25.708169 1468442'173537 1468443:139543 [15,32,1] 15 [15,32,1] 15 1435301'169725 2016-11-06 21:46:26.537632 1338333'158008 2016-10-30 23:59:50.794774 sudo iotop -o --batch --iter 1 | grep ceph 3958453 be/4 ceph 92.17 M/s 0.00 B/s 0.00 % 66.59 % ceph-osd -f --cluster ceph --id 1 --setuser ceph --setgroup ceph [tp_osd_tp] 3956201 be/4 ceph 12.47 K/s 0.00 B/s 0.00 % 4.72 % ceph-osd -f --cluster ceph --id 11 --setuser ceph --setgroup ceph [tp_osd_tp] 3954970 be/4 ceph 361.68 K/s 0.00 B/s 0.00 % 1.98 % ceph-osd -f --cluster ceph --id 6 --setuser ceph --setgroup ceph [tp_osd_tp] 3956653 be/4 ceph 12.47 K/s 0.00 B/s 0.00 % 1.26 % ceph-osd -f --cluster ceph --id 2 --setuser ceph --setgroup ceph [tp_osd_tp] 3957297 be/4 ceph 12.47 K/s 0.00 B/s 0.00 % 0.60 % ceph-osd -f --cluster ceph --id 7 --setuser ceph --setgroup ceph [tp_osd_tp] 3956648 be/4 ceph 12.47 K/s 0.00 B/s 0.00 % 0.14 % ceph-osd -f --cluster ceph --id 2 --setuser ceph --setgroup ceph [tp_osd_tp] Only OSD1 is doing IO, the rest are doing nothing in comparison, so this must be the scrubbing, but why is it not running as idle? There are idle threads though, they just don't seem to do anything iotop --batch --iter 1 | grep ceph | grep idle 3954974 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f --cluster ceph --id 6 --setuser ceph --setgroup ceph [tp_osd_disk] 3956209 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f --cluster ceph --id 11 --setuser ceph --setgroup ceph [tp_osd_disk] 3956655 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f --cluster ceph --id 2 --setuser ceph --setgroup ceph [tp_osd_disk] 3957253 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f --cluster ceph --id 8 --setuser ceph --setgroup ceph [tp_osd_disk] 3957306 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f --cluster ceph --id 7 --setuser ceph --setgroup ceph [tp_osd_disk] 3957350 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f --cluster ceph --id 10 --setuser ceph --setgroup ceph [tp_osd_disk] 3958464 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f --cluster ceph --id 1 --setuser ceph --setgroup ceph [tp_osd_disk] 3958505 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f --cluster ceph --id 4 --setuser ceph --setgroup ceph [tp_osd_disk] 3958832 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f --cluster ceph --id 3 --setuser ceph --setgroup ceph [tp_osd_disk] 3960488 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f --cluster ceph --id 5 --setuser ceph --setgroup ceph [tp_osd_disk] 3960851 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f --cluster ceph --id 9 --setuser ceph --setgroup ceph [tp_osd_disk] 3959437 idle ceph 0.00 B/s 0.00 B/s 0.00 % 0.00 % ceph-osd -f --cluster ceph --id 0 --setuser ceph --setgroup ceph [tp_osd_disk] ceph daemon osd.1 config get osd_disk_thread_ioprio_class { "osd_disk_thread_ioprio_class": "idle" } ceph daemon osd.1 config get osd_disk_thread_ioprio_priority { "osd_disk_thread_ioprio_priority": "7" } cat /sys/block/sd*/queue/scheduler noop deadline [cfq] noop deadline [cfq] noop deadline [cfq] noop deadline [cfq] noop deadline [cfq] noop deadline [cfq] noop deadline [cfq] noop deadline [cfq] noop deadline [cfq] noop deadline [cfq] noop deadline [cfq] noop deadline [cfq] Any Ideas or have I misunderstood something? Nick _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com