On Thu, 09 Apr 2015 23:25:47 -0400 Jeff Epstein wrote: > As a follow-up to this issue, I'd like to point out some other things > I've noticed. > > First, per suggestions posted here, I've reduced the number of pgs per > pool. This results in the following ceph status: > You went from one extreme to the other and clearly didn't understand the docs and usage of the pgcalc. > cluster e96e10d3-ad2b-467f-9fe4-ab5269b70206 > health HEALTH_WARN too few pgs per osd (14 < min 20) > monmap e1: 3 mons at > {a=192.168.224.4:6789/0,b=192.168.232.4:6789/0,c=192.168.240.4:6789/0}, > election epoch 8, quorum 0,1,2 a,b,c > osdmap e238: 6 osds: 6 up, 6 in > pgmap v1107: 86 pgs, 23 pools, 2511 MB data, 801 objects > 38288 MB used, 1467 GB / 1504 GB avail > 86 active+clean > > I'm not sure if I should be concerned about the HEALTH WARN. > You should. With 23 pools (again, reduce this to 1 or 2 if possible) they should have about 23 PG/PGPs per pool (if evenly sized) not 4 as your ratio up there suggests. Your performance issues are most likely related to your platform, as in actual OSD (SSD?) speed, network speed, things unique to AWS. With a truly craptastic test cluster of 4 derelict nodes and 12 ancient HDD OSDs with 1Gb/s links formatting a 50GB RBD image like this (options so the format is actually finished when this is done): "#mkfs.ext4 -E lazy_itable_init=0,lazy_journal_init=0 /dev/rbd0" takes 20 seconds... Christian -- Christian Balzer Network/Systems Engineer chibi@xxxxxxx Global OnLine Japan/Fusion Communications http://www.gol.com/ _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com