Quoting Zack Brenton (zack@xxxxxxxxxxxx): > Types of devices: > We run our Ceph pods on 3 AWS i3.2xlarge nodes. We're running 3 OSDs, 3 > Mons, and 2 MDS pods (1 active, 1 standby-replay). Currently, each pod runs > with the following resources: > - osds: 2 CPU, 6Gi RAM, 1.7Ti NVMe disk > - mds: 3 CPU, 24Gi RAM > - mons: 500m (.5) CPU, 1Gi RAM Hmm, 6 GiB of RAM is not a whole lot. Especially if you are going to increase the amount of OSDs (partitions) like Patrick suggested. By default it will take 4 GiB per OSD ... Make sure you set the "osd_memory_target" parameter accordingly [1]. Gr. Stefan [1]: http://docs.ceph.com/docs/master/rados/configuration/bluestore-config-ref/?highlight=osd%20memory%20target -- | BIT BV http://www.bit.nl/ Kamer van Koophandel 09090351 | GPG: 0xD14839C6 +31 318 648 688 / info@xxxxxx _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com