Hi guys, We are recently testing rbd-nbd using ceph N version. After map rbd image, mkfs and mount the nbd device, the rbd-nbd and dmesg will show following errors when doing some read/write testing. rbd-nbd log: 2021-05-18 11:35:08.034 7efdb8ff9700 20 []rbd-nbd: reader_entry: waiting for nbd request ... 2021-05-18 11:35:08.066 7efdb8ff9700 -1 []rbd-nbd: failed to read nbd request header: (33) Numerical argument out of domain 2021-05-18 11:35:08.066 7efdb3fff700 20 []rbd-nbd: writer_entry: no io requests, terminating 2021-05-18 11:35:08.066 7efdea8d1a00 20 []librbd::ImageState: 0x564a2be2b3c0 unregister_update_watcher: handle=0 2021-05-18 11:35:08.066 7efdea8d1a00 20 []librbd::ImageState: 0x564a2be2b4b0 ImageUpdateWatchers::unregister_watcher: handle=0 2021-05-18 11:35:08.066 7efdea8d1a00 20 []librbd::ImageState: 0x564a2be2b4b0 ImageUpdateWatchers::unregister_watcher: completing unregister 2021-05-18 11:35:08.066 7efdea8d1a00 10 []rbd-nbd: ~NBDServer: terminating 2021-05-18 11:35:08.066 7efdea8d1a00 20 []librbd::ImageState: 0x564a2be2b3c0 close dmesg: [Tue May 18 11:35:07 2021] EXT4-fs (nbd0): mounted filesystem with ordered data mode. Opts: discard [Tue May 18 11:35:07 2021] block nbd0: shutting down sockets [Tue May 18 11:35:09 2021] blk_update_request: I/O error, dev nbd0, sector 75592 op 0x0:(READ) flags 0x3000 phys_seg 1 prio class 0 client host info: centos7.x kernel 5.4.109 It looks like the kernel nbd device shutdown its socket for some reason, but we haven't figured it out. BTW, we have tried to turn on/off rbd cache, use different fs ext4/xfs, use ec pool or replicated pool, but the error remains. It is more frequent for us to reproduce when batch map, mkfs and mount rbd-nbd on different hosts simultaneously. Thanks for any suggestions. Regards, Zhi Zhang (David) Contact: zhang.david2011@xxxxxxxxx zhangz.david@xxxxxxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx