If you aren't increasing your target max bytes and target full ratio, I wouldn't bother increasing your pgs on the cache pool. It will not gain any increased size at all as its size is dictated by those settings and not the total size of the cluster. It will remain as redundant as always.
If you are set on increasing it, which I don't see a need for, then I would recommend removing it as a cache pool, increase the PG count, and then add it back in. You will need to stop all clients that are accessing the EC rbd pool during this process.
On Sun, May 28, 2017, 10:42 PM Konstantin Shalygin <k0ste@xxxxxxxx> wrote:
On 05/28/2017 09:43 PM, David Turner wrote:
> What are your pg numbers for each pool? Your % used in each pool? And
> number of OSDs?
>
GLOBAL:
SIZE AVAIL RAW USED %RAW USED
89380G 74755G 14625G 16.36
POOLS:
NAME ID USED %USED MAX AVAIL OBJECTS
replicated_rbd 1 3305G 12.10 24007G 850006
ec_rbd 2 2674G 5.83 43212G 686555
ec_cache 3 82281M 0.33 24007G 20765
pool 1 'replicated_rbd' replicated size 3 min_size 2 crush_ruleset 1 object_hash rjenkins pg_num 256 pgp_num 256 last_change 218 flags hashpspool stripe_width 0
removed_snaps [1~3,5~2,8~2,e~2]
pool 2 'ec_rbd' erasure size 5 min_size 4 crush_ruleset 0 object_hash rjenkins pg_num 256 pgp_num 256 last_change 188 lfor 107 flags hashpspool tiers 3 read_tier 3 write_tier 3 stripe_width 4128
removed_snaps [1~5]
pool 3 'ec_cache' replicated size 3 min_size 2 crush_ruleset 1 object_hash rjenkins pg_num 16 pgp_num 16 last_change 1117 flags hashpspool,incomplete_clones tier_of 2 cache_mode writeback target_bytes 107400000000 target_objects 40000 hit_set bloom{false_positive_probability: 0.05, target_size: 0, seed: 0} 0s x1 decay_rate 0 search_last_n 0 stripe_width 0
removed_snaps [1~5]
ec_cache settings:
ceph osd pool set ec_cache target_max_bytes 107400000000 # 100Gb
ceph osd pool set ec_cache cache_target_dirty_ratio 0.3
ceph osd pool set ec_cache cache_target_dirty_high_ratio 0.6
ceph osd pool set ec_cache cache_target_full_ratio 0.8
ceph osd pool set ec_cache target_max_objects 40000
Number of OSD is: 6x osd nodes * 4 osd = 24 OSDs.
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com