The image size (50G) minus the fstrim size (1.7G) approximately equals the actual usage (48.19G). Therefore, I guess the question is why doesn't fstrim think it can discard more space? On a semi-related note, we should probably improve the rbd copy sparsify logic. Right now it requires the full stripe period (or object size if striping is disabled) to be zeroes before it skips the write operation to the destination. On Wed, Apr 27, 2016 at 2:26 PM, Tyler Wilson <kupo@xxxxxxxxxxxxxxxx> wrote: > Hello Jason, > > Thanks for the quick reply, this was copied from an VM instance snapshot to > my backup pool (rbd snap create, rbd cp (to backup pool), rbd snap rm). I've > tried piping through grep per your recommendation and it still reports the > same usage > > $ rbd diff backup/cd4e5d37-3023-4640-be5a-5577d3f9307e | grep data | awk '{ > SUM += $2 } END { print SUM/1024/1024 " MB" }' > 49345.4 MB > > Thanks for the help. > > On Wed, Apr 27, 2016 at 12:22 PM, Jason Dillaman <jdillama@xxxxxxxxxx> > wrote: >> >> On Wed, Apr 27, 2016 at 2:07 PM, Tyler Wilson <kupo@xxxxxxxxxxxxxxxx> >> wrote: >> > $ rbd diff backup/cd4e5d37-3023-4640-be5a-5577d3f9307e | awk '{ SUM += >> > $2 } >> > END { print SUM/1024/1024 " MB" }' >> > 49345.4 MB >> >> Is this a cloned image? That awk trick doesn't account for discarded >> regions (i.e. when column three says "zero" instead of "data"). Does >> the number change when you pipe the "rbd diff" results through "grep >> data" before piping to awk? >> >> > Could this be affected by replica counts some how? It seems to be twice >> > as >> > large as what is reported in the filesystem which matches my replica >> > count. >> >> No, the "rbd diff" output is only reporting image data and zeroed >> extents -- so the replication factor is not included. >> >> -- >> Jason > > -- Jason _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com