Hi all, So using ceph-ansible, i built the below mentioned cluster with 2 OSD Nodes and 3 Mons Just after creating osds i started benchmarking the performance using "rbd bench" and "rados bench" and started seeing the performance drop. Checking the status shows slow requests. [root@storage-28-1 ~]# ceph -s cluster: id: 009cbed0-e5a8-4b18-a313-098e55742e85 health: HEALTH_WARN insufficient standby MDS daemons available 1264 slow requests are blocked > 32 sec services: mon: 3 daemons, quorum storage-30,storage-29,storage-28-1 mgr: storage-30(active), standbys: storage-28-1, storage-29 mds: cephfs-3/3/3 up {0=storage-30=up:active,1=storage-28-1=up:active,2=storage-29=up:active} osd: 33 osds: 33 up, 33 in tcmu-runner: 2 daemons active data: pools: 3 pools, 1536 pgs objects: 13289 objects, 42881 MB usage: 102 GB used, 55229 GB / 55331 GB avail pgs: 1536 active+clean io: client: 1694 B/s rd, 1 op/s rd, 0 op/s wr [root@storage-28-1 ~]# ceph health detail HEALTH_WARN insufficient standby MDS daemons available; 904 slow requests are blocked > 32 sec MDS_INSUFFICIENT_STANDBY insufficient standby MDS daemons available have 0; want 1 more REQUEST_SLOW 904 slow requests are blocked > 32 sec 364 ops are blocked > 1048.58 sec 212 ops are blocked > 524.288 sec 164 ops are blocked > 262.144 sec 100 ops are blocked > 131.072 sec 64 ops are blocked > 65.536 sec osd.11 has blocked requests > 524.288 sec osds 9,32 have blocked requests > 1048.58 sec osd 9 log : https://pastebin.com/ex41cFww I see that from time to time different osds are reporting blocked requests. I am not sure what could be the cause of this. Can anyone help me fix this please. [root@storage-28-1 ~]# ceph osd tree ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -1 54.03387 root default -3 27.83563 host storage-29 2 hdd 1.63739 osd.2 up 1.00000 1.00000 3 hdd 1.63739 osd.3 up 1.00000 1.00000 4 hdd 1.63739 osd.4 up 1.00000 1.00000 5 hdd 1.63739 osd.5 up 1.00000 1.00000 6 hdd 1.63739 osd.6 up 1.00000 1.00000 7 hdd 1.63739 osd.7 up 1.00000 1.00000 8 hdd 1.63739 osd.8 up 1.00000 1.00000 9 hdd 1.63739 osd.9 up 1.00000 1.00000 10 hdd 1.63739 osd.10 up 1.00000 1.00000 11 hdd 1.63739 osd.11 up 1.00000 1.00000 12 hdd 1.63739 osd.12 up 1.00000 1.00000 13 hdd 1.63739 osd.13 up 1.00000 1.00000 14 hdd 1.63739 osd.14 up 1.00000 1.00000 15 hdd 1.63739 osd.15 up 1.00000 1.00000 16 hdd 1.63739 osd.16 up 1.00000 1.00000 17 hdd 1.63739 osd.17 up 1.00000 1.00000 18 hdd 1.63739 osd.18 up 1.00000 1.00000 -5 26.19824 host storage-30 0 hdd 1.63739 osd.0 up 1.00000 1.00000 1 hdd 1.63739 osd.1 up 1.00000 1.00000 19 hdd 1.63739 osd.19 up 1.00000 1.00000 20 hdd 1.63739 osd.20 up 1.00000 1.00000 21 hdd 1.63739 osd.21 up 1.00000 1.00000 22 hdd 1.63739 osd.22 up 1.00000 1.00000 23 hdd 1.63739 osd.23 up 1.00000 1.00000 24 hdd 1.63739 osd.24 up 1.00000 1.00000 25 hdd 1.63739 osd.25 up 1.00000 1.00000 26 hdd 1.63739 osd.26 up 1.00000 1.00000 27 hdd 1.63739 osd.27 up 1.00000 1.00000 28 hdd 1.63739 osd.28 up 1.00000 1.00000 29 hdd 1.63739 osd.29 up 1.00000 1.00000 30 hdd 1.63739 osd.30 up 1.00000 1.00000 31 hdd 1.63739 osd.31 up 1.00000 1.00000 32 hdd 1.63739 osd.32 up 1.00000 1.00000 thanks On Fri, Apr 20, 2018 at 10:24 AM, Shantur Rathore <shantur.rathore@xxxxxxxxx> wrote: > > Thanks Alfredo. I will use ceph-volume. > > On Thu, Apr 19, 2018 at 4:24 PM, Alfredo Deza <adeza@xxxxxxxxxx> wrote: >> >> On Thu, Apr 19, 2018 at 11:10 AM, Shantur Rathore >> <shantur.rathore@xxxxxxxxx> wrote: >> > Hi, >> > >> > I am building my first Ceph cluster from hardware leftover from a previous >> > project. I have been reading a lot of Ceph documentation but need some help >> > to make sure I going the right way. >> > To set the stage below is what I have >> > >> > Rack-1 >> > >> > 1 x HP DL360 G9 with >> > - 256 GB Memory >> > - 5 x 300GB HDD >> > - 2 x HBA SAS >> > - 4 x 10GBe Networking Card >> > >> > 1 x SuperMicro chassis with 17 x HP Enterprise 400GB SSD and 17 x HP >> > Enterprise 1.7TB HDD >> > Chassis and HP server are connected with 2 x SAS HBA for redundancy. >> > >> > >> > Rack-2 (Same as Rack-1) >> > >> > 1 x HP DL360 G9 with >> > - 256 GB Memory >> > - 5 x 300GB HDD >> > - 2 x HBA SAS >> > - 4 x 10GBe Networking Card >> > >> > 1 x SuperMicro chassis with 17 x HP Enterprise 400GB SSD and 17 x HP >> > Enterprise 1.7TB HDD >> > Chassis and HP server are connected with 2 x SAS HBA for redundancy. >> > >> > >> > Rack-3 >> > >> > 5 x HP DL360 G8 with >> > - 128 GB Memory >> > - 2 x 400GB HP Enterprise SSD >> > - 3 x 1.7TB Enterprise HDD >> > >> > Requirements >> > - To serve storage to around 200 VMware VMs via iSCSI. VMs use disks >> > moderately. >> > - To serve storage to some docker containers using ceph volume driver >> > - To serve storage to some legacy apps using NFS >> > >> > Plan >> > >> > - Create a ceph cluster with all machines >> > - Use Bluestore as osd backing ( 3 x SSD for DB and WAL in SuperMicro >> > Chassis and 1 x SSD for DB and WAL in Rack 3 G8s) >> > - Use remaining SSDs ( 14 x in SuperMicro and 1 x Rack 3 G8s ) for Rados >> > Cache Tier >> > - Update CRUSH map to make Rack as minimum failure domain. So almost all >> > data is replicated across racks and in case one of the host dies the storage >> > still works. >> > - Single bonded network (4x10GBe) connected to ToR switches. >> > - Same public and cluster network >> > >> > Questions >> > >> > - First of all, is this kind of setup workable. >> > - I have seen that Ceph uses /dev/sdx names in guides, is it a good approach >> > considering the disks die and can come up with different /dev/sdx identifier >> > on reboot. >> >> In the case of ceph-volume, these will not matter since it uses LVM >> behind the scenes and LVM takes care of figuring out if /dev/sda1 is >> now really /dev/sdb1 after >> a reboot. >> >> If using ceph-disk however, the detection is done a bit differently, >> by reading partition labels and depending on UDEV triggers that >> sometimes can be troublesome, specially >> on reboot. In the case of a successful detection via UDEV the >> non-persistent names wouldn't matter much still. >> >> > - What should be the approx size of WAL and DB partitions for my kind of >> > setup? >> > - Can i install ceph in a VM and use other VMs on these hosts. Is Ceph too >> > CPU demanding? >> > >> > Thanks, >> > Shantur >> > >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@xxxxxxxxxxxxxx >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > > > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com