You're like the 5th person here (including me) that was hit by this. Could I get some input from someone using CEPH with RBD and thousands of OSDs? How high did you have to go? I only have ~200 OSDs and I had to bump the limit up to 10000 for VMs that have multiple volumes attached, this doesn't seem right? I understand this is the effect of striping a volume accross multiple PGs, but shouldn't this be more limited or somehow garbage collected? And to get deeper - I suppose there will be one connection from QEMU to OSD for each NCQ queue? Or how does this work? blk-mq will likely be different again... Or is it decoupled from the virtio side of things by RBD cache if that's enabled? Anyway, out of the box, at least on OpenStack installations 1) anyone having more than a few OSDs should really bump this up by default. 2) librbd should handle this situation gracefully by recycling connections, instead of hanging 3) at least we should get a warning somewhere (in the libvirt/qemu log) - I don't think there's anything when the issue hits Should I make tickets for this? Jan
|
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com