On Thu, Feb 1, 2018 at 10:44 AM, Andre Goree <andre@xxxxxxxxxx> wrote: > Does ceph-volume support lvm+Bluestore? Yes it does. Since version 12.2.2 which seems to be what you have. > I ask this, bc I'm trying to use > ceph-ansible to provision an OSD node, and am getting a hang when the > 'ceph-volume create' command is trying to be run. I'm also using > ceph-ansible/master (not stable-3.0 or any other stable version), which has > the parameters for bluestore+lvm, but may not actually be implemented in > 12.2.2. ceph-ansible stable-3.0 branch does NOT contain a section for > bluestore in the ceph-volume/lvm section. Correct, stable-3.0 doesn't have it. > > Here you can see the full command(s) being run (via ansible), and an strace > on that process. I've left that process running overnight and it appears it > is still timing out (apologies if I should've used pastebin): This is not really ceph-volume. It is trying to communicate with the cluster but it is not going through. Something in your configuration is not quite right, so the command is hanging. This is the actual command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new a2ee64a4-b5ba-4ca9-8528-4205f3ad8c99 What that command is trying to do is to tell the monitor about the newly created OSD. It is easy to replicate this "hanging" problem if you modify your ceph.conf to point to an invalid IP for the monitors. > > ~# ps auxf |grep python > ceph-us+ 10772 0.0 0.0 4504 704 pts/3 Ss+ 15:56 0:00 > \_ /bin/sh -c sudo -H -S -n -u root /bin/sh -c 'echo > BECOME-SUCCESS-nusspvylzmocjedchieutgbjxlbrlyqb; CEPH_VOLUME_DEBUG=1 > /usr/bin/python > /home/ceph-user/.ansible/tmp/ansible-tmp-1517432160.32-17230757947356/ceph_volume.py; > rm -rf > "/home/ceph-user/.ansible/tmp/ansible-tmp-1517432160.32-17230757947356/" > > /dev/null 2>&1' && sleep 0 > root 10773 0.0 0.0 52700 3764 pts/3 S+ 15:56 0:00 > \_ sudo -H -S -n -u root /bin/sh -c echo > BECOME-SUCCESS-nusspvylzmocjedchieutgbjxlbrlyqb; CEPH_VOLUME_DEBUG=1 > /usr/bin/python > /home/ceph-user/.ansible/tmp/ansible-tmp-1517432160.32-17230757947356/ceph_volume.py; > rm -rf > "/home/ceph-user/.ansible/tmp/ansible-tmp-1517432160.32-17230757947356/" > > /dev/null 2>&1 > root 10774 0.0 0.0 4504 796 pts/3 S+ 15:56 0:00 > \_ /bin/sh -c echo BECOME-SUCCESS-nusspvylzmocjedchieutgbjxlbrlyqb; > CEPH_VOLUME_DEBUG=1 /usr/bin/python > /home/ceph-user/.ansible/tmp/ansible-tmp-1517432160.32-17230757947356/ceph_volume.py; > rm -rf > "/home/ceph-user/.ansible/tmp/ansible-tmp-1517432160.32-17230757947356/" > > /dev/null 2>&1 > root 10775 0.0 0.0 32328 10392 pts/3 S+ 15:56 0:00 > \_ /usr/bin/python > /home/ceph-user/.ansible/tmp/ansible-tmp-1517432160.32-17230757947356/ceph_volume.py > root 10776 0.1 0.0 36572 12480 pts/3 S+ 15:56 0:00 > \_ /usr/bin/python /tmp/ansible_N0WKmE/ansible_module_ceph_volume.py > root 10790 0.3 0.0 52236 18020 pts/3 S+ 15:56 0:00 > \_ /usr/bin/python2.7 /usr/sbin/ceph-volume --cluster ceph lvm create > --bluestore --data data-vg1/data-lv1 --block.db nvme-vg/db-lv1 --block.wal > nvme-vg/wal-lv1 > root 10797 0.2 0.0 759400 31388 pts/3 Sl+ 15:56 0:00 > \_ /usr/bin/python2.7 /usr/bin/ceph --cluster ceph --name > client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - > osd new a2ee64a4-b5ba-4ca9-8528-4205f3ad8c99 > root 10835 0.0 0.0 14224 1028 pts/2 S+ 15:56 0:00 > \_ grep --color python > root@osd-08 ~ # strace -p10797 > strace: Process 10797 attached > select(0, NULL, NULL, NULL, {0, 22600}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 50000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 35089}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 1000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 2000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 4000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 8000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 16000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 32000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 50000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 50000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 50000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 50000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 50000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 50000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 50000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 50000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 34029}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 1000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 2000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 4000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 8000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 16000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 32000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 50000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 50000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 50000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 50000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 50000}) = 0 (Timeout) > select(0, NULL, NULL, NULL, {0, 50000}^Cstrace: Process 10797 detached > <detached ...> > ~# > > > > Here is some of my lvm configuration, setup prior to running ceph-ansible: > ~# pvs > PV VG Fmt Attr PSize PFree > /dev/mapper/crypt-nvme0n1 nvme-vg lvm2 a-- 745.21g 340.71g > /dev/mapper/osd-sda data-vg1 lvm2 a-- 7.28t 0 > /dev/mapper/osd-sdb data-vg2 lvm2 a-- 7.28t 0 > /dev/mapper/osd-sdc data-vg3 lvm2 a-- 7.28t 0 > /dev/mapper/osd-sdd data-vg4 lvm2 a-- 7.28t 0 > /dev/mapper/osd-sde data-vg5 lvm2 a-- 7.28t 0 > /dev/mapper/osd-sdf data-vg6 lvm2 a-- 7.28t 0 > /dev/mapper/osd-sdg data-vg7 lvm2 a-- 7.28t 0 > /dev/mapper/osd-sdh data-vg8 lvm2 a-- 7.28t 0 > /dev/mapper/osd-sdi data-vg9 lvm2 a-- 7.28t 0 > /dev/mapper/osd-sdj data-vg10 lvm2 a-- 7.28t 0 > /dev/mapper/osd-sdk data-vg11 lvm2 a-- 7.28t 0 > /dev/mapper/osd-sdl data-vg12 lvm2 a-- 7.28t 0 > ~# vgs > VG #PV #LV #SN Attr VSize VFree > data-vg1 1 1 0 wz--n- 7.28t 0 > data-vg10 1 1 0 wz--n- 7.28t 0 > data-vg11 1 1 0 wz--n- 7.28t 0 > data-vg12 1 1 0 wz--n- 7.28t 0 > data-vg2 1 1 0 wz--n- 7.28t 0 > data-vg3 1 1 0 wz--n- 7.28t 0 > data-vg4 1 1 0 wz--n- 7.28t 0 > data-vg5 1 1 0 wz--n- 7.28t 0 > data-vg6 1 1 0 wz--n- 7.28t 0 > data-vg7 1 1 0 wz--n- 7.28t 0 > data-vg8 1 1 0 wz--n- 7.28t 0 > data-vg9 1 1 0 wz--n- 7.28t 0 > nvme-vg 1 16 0 wz--n- 745.21g 340.71g > ~# lvs > LV VG Attr LSize Pool Origin Data% Meta% Move Log > Cpy%Sync Convert > data-lv1 data-vg1 -wi-a----- 7.28t > data-lv10 data-vg10 -wi-a----- 7.28t > data-lv11 data-vg11 -wi-a----- 7.28t > data-lv12 data-vg12 -wi-a----- 7.28t > data-lv2 data-vg2 -wi-a----- 7.28t > data-lv3 data-vg3 -wi-a----- 7.28t > data-lv4 data-vg4 -wi-a----- 7.28t > data-lv5 data-vg5 -wi-a----- 7.28t > data-lv6 data-vg6 -wi-a----- 7.28t > data-lv7 data-vg7 -wi-a----- 7.28t > data-lv8 data-vg8 -wi-a----- 7.28t > data-lv9 data-vg9 -wi-a----- 7.28t > db-lv1 nvme-vg -wi-a----- 50.00g > db-lv2 nvme-vg -wi-a----- 50.00g > db-lv3 nvme-vg -wi-a----- 50.00g > db-lv4 nvme-vg -wi-a----- 50.00g > db-lv5 nvme-vg -wi-a----- 50.00g > db-lv6 nvme-vg -wi-a----- 50.00g > db-lv7 nvme-vg -wi-a----- 50.00g > db-lv8 nvme-vg -wi-a----- 50.00g > wal-lv1 nvme-vg -wi-a----- 576.00m > wal-lv2 nvme-vg -wi-a----- 576.00m > wal-lv3 nvme-vg -wi-a----- 576.00m > wal-lv4 nvme-vg -wi-a----- 576.00m > wal-lv5 nvme-vg -wi-a----- 576.00m > wal-lv6 nvme-vg -wi-a----- 576.00m > wal-lv7 nvme-vg -wi-a----- 576.00m > wal-lv8 nvme-vg -wi-a----- 576.00m > > > -- > Andre Goree > -=-=-=-=-=- > Email - andre at drenet.net > Website - http://blog.drenet.net > PGP key - http://www.drenet.net/pubkey.html > -=-=-=-=-=- > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com