Correct, sorry, I have just read the first question and answered too quickly. As fas as I know the space available is "shared" (the space is a combination of OSD drives and crushmap ) between pools using the same device class but you can define quota for each pool if needed. ceph osd pool set-quota <poolname> max_objects|max_bytes <val> set object or byte limit on pool ceph osd pool get-quota <poolname> obtain object or byte limits for pool You can use "ceph df detail" to see you pools usage including quota. As the space is "shared", you can't determine a max size for just one pool (except if you have only one pool). # ceph df detail GLOBAL: SIZE AVAIL RAW USED %RAW USED OBJECTS 144T 134T 10254G 6.93 1789k POOLS: NAME ID QUOTA OBJECTS QUOTA BYTES USED %USED MAX AVAIL OBJECTS DIRTY READ WRITE RAW USED pool1 9 N/A N/A 7131G 14.70 41369G 1826183 1783k 3847k 14959k 21394G pool2 10 N/A N/A 24735M 0.06 41369G 6236 6236 1559k 226k 74205M pool3 11 N/A N/A 30188k 0 41369G 29 29 1259k 4862k 90564k pool4 12 N/A N/A 0 0 41369G 0 0 0 0 0 pool5 13 N/A N/A 0 0 41369G 0 0 0 0 0 pool6 14 N/A N/A 0 0 41369G 0 0 0 0 0 pool7 15 N/A N/A 0 0 41369G 0 0 0 0 0 pool8 16 N/A N/A 0 0 41369G 0 0 0 0 0 pool9 17 N/A N/A 0 0 41369G 0 0 0 0 0 pool10 18 N/A N/A 0 0 41369G 0 0 0 0 0 .rgw.root 19 N/A N/A 2134 0 41369G 6 6 231 6 6402 default.rgw.control 20 N/A N/A 0 0 41369G 8 8 0 0 0 default.rgw.meta 21 N/A N/A 363 0 41369G 2 2 12 3 1089 default.rgw.log 22 N/A N/A 0 0 41369G 207 207 8949k 5962k 0 You should seek for used and max sizes for images, not pools. # rbd disk-usage your_pool/your_image NAME PROVISIONED USED image-1 51200M 102400k You can see the total provisioned and used sizes for a whole pool using: # rbd disk-usage -p your_pool --format json | jq . { "images": [ { "name": "image-1", "provisioned_size": 53687091200, "used_size": 104857600 } ], "total_provisioned_size": 53687091200, "total_used_size": 104857600 } A reminder: most ceph commands can output in json format ( --format=json or -f json), useful with the jq tool.
|
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com