throttling PG removal

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi all,

We're deleting a large pool with lots of 4MB objects, and it's turning
out to be rather disruptive: >50 loadavg on osd machines, osd's
flapping, peering mess for the other pools.

In general, I'd like to be able to throttle this way back so that the
deleted PG objects slowly disappear without affecting IO to other
pools. And AFAICT there's no way to achieve that in current jewel or
master.

Here is the timing of the RemoveWQ on one of these OSDs:

2016-06-29 18:28:06.388530 7fee8a2e5700 12 OSD::disk_tp worker wq
OSD::RemoveWQ start processing 0x1 (1 active)
2016-06-29 18:28:07.954758 7fee8a2e5700 15 OSD::disk_tp worker wq
OSD::RemoveWQ done processing 0x1 (0 active)
2016-06-29 18:28:07.954766 7fee8a2e5700 12 OSD::disk_tp worker wq
OSD::RemoveWQ start processing 0x1 (1 active)
2016-06-29 18:28:10.599542 7fee8a2e5700 15 OSD::disk_tp worker wq
OSD::RemoveWQ done processing 0x1 (0 active)

I'd like to add a configurable osd_pg_remove_sleep, probably at the
top of OSD::RemoveWQ::_process ? Should that have the desired effect?

Moreover, it would be nice to be able to decrease the number of
objects removed per instance of the worker. It's currently limited to
osd_target_transaction_size (shared by build_past_intervals_parallel,
remove_dir, and handle_osd_map). Should it be safe to decrease that
(to, say, 5) or would it be better to introduce something specific to
remove_dir?

Thanks,

Dan
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux