>> How many pgs do you have? ('ceph osd dump | grep ^pool'). > > I believe this is it. 384 PGs, but three pools of which only one (or maybe a second one, sort of) is in use. Automatically setting the right PG counts is coming some day, but until then being able to set up pools of the right size is a big gotcha. :( > Depending on how mutable the data is, recreate with larger PG counts on the pools in use. Otherwise we can do something more detailed. > -Greg hm... what would be recommended PG size per pool ? chef@cephgw:~$ ceph osd lspools 0 data,1 metadata,2 rbd, chef@cephgw:~$ ceph osd pool get data pg_num PG_NUM: 128 chef@cephgw:~$ ceph osd pool get metadata pg_num PG_NUM: 128 chef@cephgw:~$ ceph osd pool get rbd pg_num PG_NUM: 128 according to the http://ceph.com/docs/master/rados/operations/placement-groups/ (OSDs * 100) Total PGs = ------------ Replicas I have 3 OSDs and 2 replicas for each object, which gives recommended PG = 150 will it make much difference to set 150 instead of 128 or I should base on different values? btw, just one more off-topic question: chef@ceph-node03:~$ ceph pg dump| egrep -v '^(0\.|1\.|2\.)'| column -t dumped all in format plain version 113906 last_osdmap_epoch 323 last_pg_scan 1 full_ratio 0.95 nearfull_ratio 0.85 pg_stat objects mip degr unf bytes log disklog state state_stamp v reported up acting last_scrub scrub_stamp last_deep_scrub deep_scrub_stamp pool 0 74748 0 0 0 286157692336 17668034 17668034 pool 1 618 0 0 0 131846062 6414518 6414518 pool 2 0 0 0 0 0 0 0 sum 75366 0 0 0 286289538398 24082552 24082552 osdstat kbused kbavail kb hb in hb out 0 157999220 106227596 264226816 [1,2] [] 1 185604948 78621868 264226816 [0,2] [] 2 219475396 44751420 264226816 [0,1] [] sum 563079564 229600884 792680448 pool 0 (data) is used for data storage pool 1 (metadata) is used for metadata storage what is pool 2 (rbd) for? looks like it's absolutely empty. > >> >> You might also adjust the crush tunables, see >> >> http://ceph.com/docs/master/rados/operations/crush-map/?highlight=tunable#tunables >> >> sage >> Thanks for the link, Sage I set tunable values according to the doc. Btw, online document is missing magical param for crushmap which allows those scary_tunables ) -- ...WBR, Roman Hlynovskiy -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html