I don't use ansible anymore. But this was my config for the host onode1: ./host_vars/onode2.yml: lvm_volumes: - data: /dev/sdb db: '1' db_vg: host-2-db - data: /dev/sdc db: '2' db_vg: host-2-db - data: /dev/sde db: '3' db_vg: host-2-db - data: /dev/sdf db: '4' db_vg: host-2-db … one config file per host. The LVs were created by hand on a PV over RAID1 over two SSDs. The hosts had empty slots for hdds to be bought later. So I had to "partition" the PV by hand, because ansible uses the whole RAID1 only for the present HDDs. It is said that only certain sizes of DB & WAL partitions are sensible. I now use 58GiB LVs. The remaining space in the RAID1 is used for a faster OSD. Lars Wed, 30 Oct 2019 10:02:23 +0000 CUZA Frédéric <frederic.cuza@xxxxxx> ==> "ceph-users@xxxxxxxxxxxxxx" <ceph-users@xxxxxxxxxxxxxx> : > Hi Everyone, > > Does anyone know how to indicate block-db and block-wal to device on ansible ? > In ceph-deploy it is quite easy : > ceph-deploy osd create osd_host08 --data /dev/sdl --block-db /dev/sdm12 --block-wal /dev/sdn12 -bluestore > > On my data nodes I have 12 HDDs and 2 SSDs I use those SSDs for block-db and block-wal. > How to indicate for each osd which partition to use ? > > And finally, how do you handle the deployment if you have multiple data nodes setup ? > SSDs on sdm and sdn on one host and SSDs on sda and sdb on another ? > > Thank you for your help. > > Regards, -- Informationstechnologie Berlin-Brandenburgische Akademie der Wissenschaften Jägerstraße 22-23 10117 Berlin Tel.: +49 30 20370-352 http://www.bbaw.de _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com