Create a rbd image and map it to client,then stop ceph cluster through '/etc/init.d/ceph -a stop', then in client side, run command 'echo id > /sys/bus/rbd/remove', and this command can not return. Checking dmesg, seems like it enters an endless loop, try to re-connect osds and mons.Then press keys 'CTRL + C' to send an INT signal to 'echo id > /sys/bus/rbd/remove',then kernel crash. Kernel crash because ceph_osd's o_requests list is not empty. This fix is to clean ceph_osd's o_requests list when it's forced to remove a rbd device(triggered by /sys/bus/rbs/remove). Signed-off-by: Guanjun He <heguanbo@xxxxxxxxx> --- net/ceph/osd_client.c | 9 +++++++++ 1 files changed, 9 insertions(+), 0 deletions(-) diff --git a/net/ceph/osd_client.c b/net/ceph/osd_client.c index 1ffebed..4dba062 100644 --- a/net/ceph/osd_client.c +++ b/net/ceph/osd_client.c @@ -688,11 +688,20 @@ static void __remove_osd(struct ceph_osd_client *osdc, struct ceph_osd *osd) static void remove_all_osds(struct ceph_osd_client *osdc) { + struct list_head *pos, *q; + struct ceph_osd_request *req; + dout("__remove_old_osds %p\n", osdc); mutex_lock(&osdc->request_mutex); while (!RB_EMPTY_ROOT(&osdc->osds)) { struct ceph_osd *osd = rb_entry(rb_first(&osdc->osds), struct ceph_osd, o_node); + list_for_each_safe(pos, q, &osd->o_requests) { + req = list_entry(pos, struct ceph_osd_request, + r_osd_item); + list_del(pos); + kfree(req); + } __remove_osd(osdc, osd); } mutex_unlock(&osdc->request_mutex); -- 1.7.3.4 -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html