I am going to attempt to answer my own question here and someone can correct me if I am wrong. Looking at a few of the other OSDs that we have replaced over the last year or so it looks like they are mounted using tmpfs as well and that this is just a result of switching from filestore to bluestore and that this is really nothing to worry about. Thanks, Shain On 9/9/20, 11:16 AM, "Shain Miley" <SMiley@xxxxxxx> wrote: Hi, I recently added 3 new servers to Ceph cluster. These servers use the H740p mini raid card and I had to install the HWE kernel in Ubuntu 16.04 in order to get the drives recognized. We have a 23 node cluster and normally when we add OSDs they end up mounting like this: /dev/sde1 3.7T 2.0T 1.8T 54% /var/lib/ceph/osd/ceph-15 /dev/sdj1 3.7T 2.0T 1.7T 55% /var/lib/ceph/osd/ceph-20 /dev/sdd1 3.7T 2.1T 1.6T 58% /var/lib/ceph/osd/ceph-14 /dev/sdc1 3.7T 1.8T 1.9T 49% /var/lib/ceph/osd/ceph-13 However I noticed this morning that the 3 new servers have the OSDs mounted like this: tmpfs 47G 28K 47G 1% /var/lib/ceph/osd/ceph-246 tmpfs 47G 28K 47G 1% /var/lib/ceph/osd/ceph-240 tmpfs 47G 28K 47G 1% /var/lib/ceph/osd/ceph-248 tmpfs 47G 28K 47G 1% /var/lib/ceph/osd/ceph-237 Is this normal for deployments going forward…or did something go wrong? These are 12TB drives but they are showing up as 47G here instead. We are using ceph version 12.2.13 and I installed this using ceph-deply version 2.0.1. Thanks in advance, Shain Shain Miley | Director of Platform and Infrastructure | Digital Media | smiley@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx