-----BEGIN PGP SIGNED MESSAGE----- Hash: SHA256 We are looking to purchase our next round of Ceph hardware and based off the work by Nick Fisk [1] our previous thought of cores over clock is being revisited. I have two camps of thoughts and would like to get some feedback, even if it is only theoretical. We currently have 12 disks per node (2 SSD/10 4TB spindle), but we may adjust that to 4/8. SSD would be used for journals and cache tier (when [2] and fstrim are resolved). We also want to stay with a single processor for cost, power and NUMA considerations. 1. For 12 disks with three threads each (2 client and 1 background), lots of slower cores would allow I/O (ceph code) to be scheduled as soon as a core is available. 2. Faster cores would get through the Ceph code faster but there would be less cores and so some I/O may have to wait to be scheduled. I'm leaning towards #2 for these reasons, please expose anything I may be missing: * The latency will only really be improved in the SSD I/O with faster clock speed, all writes and any reads from the cache tier. So 8 fast cores might be sufficient, reading from spindle and flushing the journal will have a "substantial" amount of sleep to allow other Ceph I/O to be hyperthreaded. * Even though SSDs are much faster than spindles they are still orders of magnitude slower than the processor, so it is still possible to get more lines of code executed between SSD I/O with a faster processor even with less cores. * As the Ceph code is improved through optimization and less code has to be executed for each I/O, faster clock speeds will only provide even more benefit (lower latency, less waiting for cores) as the delay shifts more from CPU to disk. Since our workload is typically small I/O 12K-18K, latency means a lot to our performance. Our current processors are Intel(R) Xeon(R) CPU E5-2640 v2 @ 2.00GHz [1] http://www.spinics.net/lists/ceph-users/msg19305.html [2] http://article.gmane.org/gmane.comp.file-systems.ceph.user/22713 Thanks, - ---------------- Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 -----BEGIN PGP SIGNATURE----- Version: Mailvelope v1.0.0 Comment: https://www.mailvelope.com wsFcBAEBCAAQBQJV16pfCRDmVDuy+mK58QAA9cgP/RwsZESriIMWZHeC0PmS CH8iEFCXCRCzvW+lYMwB9FOvPmBLlhayp39Z93Djv3sef02t3Z9NFPq7fUmb ZwZ9SnH9oVmRElbQyNtt8MfJ2cqXRU6JtYsTHnZ5G0+sFvv+BY+mYD89nULw xwbsosUCBA9Rp8geq++XLSbuEBt8AfreYaSBzY1kg51Ovtmb97R0hB7bQBWP oUgi/ET24w4sUqLSo4WBNBZ0WeWsRA4w5PEzHk28ynBY0B/GAtiGadtZWOFX 6bNz3KjMbLEWU9UF+7WyL+ppru6RIUZeayFp3tdIzqQdMbeBDPO54miOezwv 9iFNuzxj2P6jqlp18W2SZYN2JF5qCgrG5mXlU2bOM9k4IlQAqG2V3iD/rSF8 LmL/FSzU6C4k8PffaNis/grZAtjN4tCLRAoWUmsXSRW1NpSNm13l6wJfg5xq XGLQ4CfGMV/o3a1Oz1M7jfMLWb0b6TeYlqC8eeHUp9ipa8IaVKsGNDJYQOnM LvyRuyB7yIM6dEXmJjE5ZQPwbh0se3+hUhNolQ949aKrY2u8Q2kHhKqOyzuw EAAyHkeqBtAZFW+DActHYVCi9lJO8shmeWuVKxAuzKYJGYzD8yVIS+AVqZ2k OH2/NNAXzBKefsL1gd8DT4QuYqDoEN2arO+PN0vZeEruQ4vg6qZvabqeB/4o kUd4 =F5Sx -----END PGP SIGNATURE----- _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com