No every volume. It seems that volumes with high capacity have higher probability to trigger this problem. > 2020年1月15日 下午4:28,Eugen Block <eblock@xxxxxx> 写道: > > Then it's probably something different. Does that happen with every volume/image or just this one time? > > > Zitat von 徐蕴 <yunxu@xxxxxx>: > >> Hi Eugen, >> >> Thank you for sharing your experience. I will dig into OpenStack cinder logs to check if something happened. The strange thing is the volume I deleted is not created from a snapshot, or doesn’t have any snapshot. And the rbd_id.xxx, rbd_header.xxx and rbd_object_map.xxx were deleted, just left out a lot of rbd_data objects. I plan to delete those objects manually. >> >> br, >> Xu Yun >> >>> 2020年1月15日 下午3:50,Eugen Block <eblock@xxxxxx> 写道: >>> >>> Hi, >>> >>> this might happen if you try to delete images/instances/volumes in openstack that are somehow linked, e.g. if there are snapshots etc. I have experienced this in Ocata, too. Deleting a base image worked but there were existing clones so basically just the openstack database was updated, but the base image still existed within ceph. >>> >>> Try to figure out if that is also the case. If it's something else, check the logs in your openstack environment, maybe they reveal something. Also check the ceph logs. >>> >>> Regards, >>> Eugen >>> >>> >>> Zitat von 徐蕴 <yunxu@xxxxxx>: >>> >>>> Hello, >>>> >>>> My setup is Ceph pike working with OpenStack. When I deleted an image, I found that the space was not reclaimed. I checked with rbd ls and confirmed that this image was disappeared. But when I check the objects with rados ls, most objects named rbd_data.xxx are still existed in my cluster. rbd_object_map and rbd_header were already deleted. I waited for several hours and there is no further deletion happed. Is it a known issue, or something wrong with my configuration? >>>> >>>> br, >>>> Xu Yun >>>> _______________________________________________ >>>> ceph-users mailing list -- ceph-users@xxxxxxx >>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx >>> >>> >>> _______________________________________________ >>> ceph-users mailing list -- ceph-users@xxxxxxx >>> To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx