Hi Dev, Can you run the below command to check if this image is still considered as mapped by any ceph-csi nodeplugins? $ namespace=ceph-csi-rbd $ image=csi-vol-945c6a66-9129 $ for pod in $(kubectl -n $namespace get pods | grep -E 'rbdplugin|nodeplugin' | grep -v provisioner | awk '{print $1}'); do echo $pod; kubectl exec -it -n $namespace $pod -c csi-rbdplugin -- rbd device list | grep $image ; done If it pops out in the output, get into the csi-rbdplugin container of the nodeplugin pod that listed the image and unmount/unmap it: $ kubectl -n $namespace exec -ti ceph-csi-rbd-nodeplugin-xxxxx -c csi-rbdplugin -- sh <---- please adjust nodepluding pod name here sh-4.4# sh-4.4# rbd device list id pool namespace image snap device 0 k8s-rgnl-disks csi-vol-945c6a66-9129 - /dev/rbd0 sh-4.4# umount /dev/rbd/k8s-rgnl-disks/csi-vol-945c6a66-9129 sh-4.4# rbd unmap /dev/rbd/k8s-rgnl-disks/csi-vol-945c6a66-9129 sh-4.4# rbd device list sh-4.4# Hope there's no typo. Regards, Frédéric. ----- Le 21 Jan 25, à 23:33, Devender Singh devender@xxxxxxxxxx a écrit : > Hello Eugen > > Thanks for your reply. > I have the image available and it’s not under trash. > > When scaling a pod to different node using statefulset, pod gives mount issue. > > I was looking for a command if we can kill the client.id <http://client.id/> > from ceph. CEPH must have a command to kill its clients etc… > Don’t understand why pod complaining about same volume name about a k8s host > using it. Whereas its nowhere.. Not sure what to do in this situation.. > We tried upgrading csi, k8s cluster. Renamed image and blocklisted the host. And > renamed back image to its original image but still red status showing same > client host. > > > Regards > Dev > >> On Jan 21, 2025, at 12:16 PM, Eugen Block <eblock@xxxxxx> wrote: >> >> Hi, >> >> have you checked if the image is in the trash? >> >> rbd -p {pool} trash ls >> >> You can try to restore the image if there is one, then blocklist the client to >> release the watcher, then delete the image again. >> >> I have to do that from time to time on a customer’s openstack cluster. >> >> Zitat von Devender Singh <devender@xxxxxxxxxx>: >> >>> Hello >>> >>> Seeking some help if I can clean the client mounting my volume? >>> >>> rbd status pool/image >>> >>> Watchers: >>> watcher=10.160.0.245:0/2076588905 client.12541259 cookie=140446370329088 >>> >>> Issue: pod is failing in init- state. >>> Events: >>> Type Reason Age From Message >>> ---- ------ ---- ---- ------- >>> Warning FailedMount 96s (x508 over 24h) kubelet MountVolume.MountDevice >>> failed for volume "pvc-3a2048f1" : rpc error: code = Internal desc = rbd image >>> k8s-rgnl-disks/csi-vol-945c6a66-9129 is still being used >>> >>> It shows above client, but there is no such volume… >>> >>> Another similar issue… on dashboard… >>> >>> CephNodeDiskspaceWarning >>> Mountpoint /mnt/dst-volume on sea-prod-host01 will be full in less than 5 days >>> based on the 48 hour trailing fill rate. >>> >>> Whereas nothing is mounted, I mapped one image yesterday using red map and then >>> unmapped and unmounted everything but it been more than 12hours now, still >>> showing the message.. >>> >>> >>> CEPH version: 18.2.4 >>> >>> Regards >>> Dev >>> >>> >>> >>> _______________________________________________ >>> ceph-users mailing list -- ceph-users@xxxxxxx >>> To unsubscribe send an email to ceph-users-leave@xxxxxxx >> >> >> _______________________________________________ >> ceph-users mailing list -- ceph-users@xxxxxxx >> To unsubscribe send an email to ceph-users-leave@xxxxxxx > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx