Hello, We have a ceph cluster with CephFS and RBD images enabled, from Xen-NG we connect directly to rbd images. Several times a day the VMs suffer from a high load/iowait which makes them temporarily inaccessible (arround 10~30 seconds), in the logs on xen-ng I find this: [Thu Sep 9 02:16:06 2021] rbd: rbd4: encountered watch error: -107 [Thu Sep 9 02:17:47 2021] rbd: rbd3: encountered watch error: -107 [Thu Sep 9 02:18:55 2021] rbd: rbd4: encountered watch error: -107 [Thu Sep 9 02:19:54 2021] rbd: rbd3: encountered watch error: -107 [Thu Sep 9 02:49:39 2021] rbd: rbd3: encountered watch error: -107 [Thu Sep 9 03:47:25 2021] rbd: rbd3: encountered watch error: -107 [Thu Sep 9 03:48:07 2021] rbd: rbd4: encountered watch error: -107 [Thu Sep 9 04:47:30 2021] rbd: rbd3: encountered watch error: -107 [Thu Sep 9 04:47:55 2021] rbd: rbd4: encountered watch error: -107 Version XEN: XCP-ng release 8.2.0 (xenenterprise) / Kernel 4.19.0+1 / running on 4 physical nodes. The Ceph cluster consists of 6 physical nodes, with48 osds (nvme), 3 mgr, 3 mon, 3 mds services connected with 2x10Gbps trunks from all hosts. Ceph status/detail is OK, no iowait/high cpu/network spikes. We have looked in the logs for a reason, but we are unable to match it with anything. Sometimes a scrub is in progress during these watch errors, but this does not always occur. Where is the best place to continue the search? Ceph.conf [global] fsid = ****** mon_initial_members = ceph-c01-mon-n1, ceph-c01-mon-n2, ceph-c01-mon-n3 mon_host = *.*.*.170,*.*.*.171,*.*.*.172 public network = *.*.*.0/19 cluster network = *.*.*.0/19 auth_cluster_required = cephx auth_service_required = cephx auth_client_required = cephx filestore_xattr_use_omap = true Ceph versions: Versions: { "mon": { "ceph version 14.2.22() nautilus(stable)": 3 }, "mgr": { "ceph version 14.2.22() nautilus(stable)": 3 }, "osd": { "ceph version 14.2.22() nautilus(stable)": 48 }, "mds": { "ceph version 14.2.22() nautilus(stable)": 3 }, "overall": { "ceph version 14.2.22() nautilus(stable)": 57 } } _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx