Re: rados bench leaves objects in tiered pool

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA256

Try:

rados -p {cachepool} cache-flush-evict-all

and see if the objects clean up.
- ----------------
Robert LeBlanc
PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1


On Tue, Nov 3, 2015 at 8:02 AM, Gregory Farnum  wrote:
> When you have a caching pool in writeback mode, updates to objects
> (including deletes) are handled by writeback rather than writethrough.
> Since there's no other activity against these pools, there is nothing
> prompting the cache pool to flush updates out to the backing pool, so
> the backing pool hasn't deleted its objects because nothing's told it
> to. You'll find that the cache pool has deleted the data for its
> objects, but it's keeping around a small "whiteout" and the object
> info metadata.
> The "rados ls" you're using has never played nicely with cache tiering
> and probably never will. :( Listings are expensive operations and
> modifying them to do more than the simple info scan would be fairly
> expensive in terms of computation and IO.
>
> I think there are some caching commands you can send to flush updates
> which would cause the objects to be entirely deleted, but I don't have
> them off-hand. You can probably search the mailing list archives or
> the docs for tiering commands. :)
> -Greg
>
> On Tue, Nov 3, 2015 at 12:40 AM, Дмитрий Глушенок  wrote:
>> Hi,
>>
>> While benchmarking tiered pool using rados bench it was noticed that objects are not being removed after test.
>>
>> Test was performed using "rados -p rbd bench 3600 write". The pool is not used by anything else.
>>
>> Just before end of test:
>> POOLS:
>>     NAME                      ID     USED       %USED     MAX AVAIL     OBJECTS
>>     rbd-cache                 36     33110M      3.41          114G        8366
>>     rbd                       37     43472M      4.47          237G       10858
>>
>> Some time later (few hundreds of writes are flushed, rados automatic cleanup finished):
>> POOLS:
>>     NAME                      ID     USED       %USED     MAX AVAIL     OBJECTS
>>     rbd-cache                 36      22998         0          157G       16342
>>     rbd                       37     46050M      4.74          234G       11503
>>
>> # rados -p rbd-cache ls | wc -l
>> 16242
>> # rados -p rbd ls | wc -l
>> 11503
>> #
>>
>> # rados -p rbd cleanup
>> error during cleanup: -2
>> error 2: (2) No such file or directory
>> #
>>
>> # rados -p rbd cleanup --run-name "" --prefix prefix ""
>>  Warning: using slow linear search
>>  Removed 0 objects
>> #
>>
>> # rados -p rbd ls | head -5
>> benchmark_data_dropbox01.tzk_7641_object10901
>> benchmark_data_dropbox01.tzk_7641_object9645
>> benchmark_data_dropbox01.tzk_7641_object10389
>> benchmark_data_dropbox01.tzk_7641_object10090
>> benchmark_data_dropbox01.tzk_7641_object11204
>> #
>>
>> #  rados -p rbd-cache ls | head -5
>> benchmark_data_dropbox01.tzk_7641_object10901
>> benchmark_data_dropbox01.tzk_7641_object9645
>> benchmark_data_dropbox01.tzk_7641_object10389
>> benchmark_data_dropbox01.tzk_7641_object5391
>> benchmark_data_dropbox01.tzk_7641_object10090
>> #
>>
>> So, it looks like the objects are still in place (in both pools?). But it is not possible to remove them:
>>
>> # rados -p rbd rm benchmark_data_dropbox01.tzk_7641_object10901
>> error removing rbd>benchmark_data_dropbox01.tzk_7641_object10901: (2) No such file or directory
>> #
>>
>> # ceph health
>> HEALTH_OK
>> #
>>
>>
>> Can somebody explain the behavior? And is it possible to cleanup the benchmark data without recreating the pools?
>>
>>
>> ceph version 0.94.5
>>
>> # ceph osd dump | grep rbd
>> pool 36 'rbd-cache' replicated size 3 min_size 1 crush_ruleset 1 object_hash rjenkins pg_num 100 pgp_num 100 last_change 755 flags hashpspool,incomplete_clones tier_of 37 cache_mode writeback target_bytes 107374182400 hit_set bloom{false_positive_probability: 0.05, target_size: 0, seed: 0} 3600s x1 stripe_width 0
>> pool 37 'rbd' erasure size 5 min_size 3 crush_ruleset 2 object_hash rjenkins pg_num 100 pgp_num 100 last_change 745 lfor 745 flags hashpspool tiers 36 read_tier 36 write_tier 36 stripe_width 4128
>> #
>>
>> # ceph osd pool get rbd-cache hit_set_type
>> hit_set_type: bloom
>> # ceph osd pool get rbd-cache hit_set_period
>> hit_set_period: 3600
>> # ceph osd pool get rbd-cache hit_set_count
>> hit_set_count: 1
>> # ceph osd pool get rbd-cache target_max_objects
>> target_max_objects: 0
>> # ceph osd pool get rbd-cache target_max_bytes
>> target_max_bytes: 107374182400
>> # ceph osd pool get rbd-cache cache_target_dirty_ratio
>> cache_target_dirty_ratio: 0.1
>> # ceph osd pool get rbd-cache cache_target_full_ratio
>> cache_target_full_ratio: 0.2
>> #
>>
>> Crush map:
>> root cache_tier {
>>         id -7           # do not change unnecessarily
>>         # weight 0.450
>>         alg straw
>>         hash 0  # rjenkins1
>>         item osd.0 weight 0.090
>>         item osd.1 weight 0.090
>>         item osd.2 weight 0.090
>>         item osd.3 weight 0.090
>>         item osd.4 weight 0.090
>> }
>> root store_tier {
>>         id -8           # do not change unnecessarily
>>         # weight 0.450
>>         alg straw
>>         hash 0  # rjenkins1
>>         item osd.5 weight 0.090
>>         item osd.6 weight 0.090
>>         item osd.7 weight 0.090
>>         item osd.8 weight 0.090
>>         item osd.9 weight 0.090
>> }
>> rule cache {
>>         ruleset 1
>>         type replicated
>>         min_size 0
>>         max_size 5
>>         step take cache_tier
>>         step chooseleaf firstn 0 type osd
>>         step emit
>> }
>> rule store {
>>         ruleset 2
>>         type erasure
>>         min_size 0
>>         max_size 5
>>         step take store_tier
>>         step chooseleaf firstn 0 type osd
>>         step emit
>> }
>>
>> Thanks
>>
>> --
>> Dmitry Glushenok
>> Jet Infosystems
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

-----BEGIN PGP SIGNATURE-----
Version: Mailvelope v1.2.3
Comment: https://www.mailvelope.com

wsFcBAEBCAAQBQJWOOqoCRDmVDuy+mK58QAAScEP/jdBEK0lH2u38S7hOwll
FA2J5+9lY0QAzxyTaRIzsZu0g+LSrLek39fFMTX/zHI0TMfSTM6dnPs6ucO2
X59wbIyJQPEzzWEa2qn4F0j/QmWkMGfuMKDjxyT6OudpZtQKDS8mt13rnqAc
QH+0bQaVfjbKGowCAHyJXTsPK4qgew7dV3JDW2hVX/vIDjYAJomkF8Ll4miT
IQ6ViV2+9u4uW99ty4aSRnYUwaf7vqycK+qUT0Uohi6iTeym7s78O42Qa0p2
WYHcXzAdYnBiR+qTIVWvKXn81tm4gmP4lSh0gpRoJ007c0hu5vTAnUvHRh0Z
070NTrmAAJXN0oZ7lkoksYZVkXDJkwBZpdif69OQU3No/HhcY9JtagEMCXcc
7/bUACaKjyKRzRmT3VNPQuMI0ix+tdi3PU4dL+16eBO832BNsqnyNHPxu570
su1m4UQVGmoXCTUOeXYe9j4jzlHO/QRXcp/soFW5DgYO6JmylZzbyNmHPjMx
CiOsdhnjAylG/zq42S4zTfd+F//aRODGJ0JNHmQYm7M2sezAvQD1HyBCAwds
VfyOcfZwyeUNubtqssmOQ+n8/EDQciK4RH/hxG8bC8xsZaNgum7Z4/zA+Efc
gMuplOsBECODAoBlfA2TP3/XixzTXoVGMmdXolOhs+Z+tT+O22eKUEK7GbMG
rIWX
=qbDR
-----END PGP SIGNATURE-----
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux