On Wed, Jan 04, 2017 at 12:03:39PM +0100, Florent B wrote: > Hi everyone, > > I have a problem with automatic start of OSDs on Debian Jessie with Ceph > Jewel. > > My osd.0 is using /dev/sda5 for data and /dev/sda2 for journal, it is > listed in ceph-disk list : > > /dev/sda : > /dev/sda1 other, 21686148-6449-6e6f-744e-656564454649 > /dev/sda3 other, linux_raid_member > /dev/sda4 other, linux_raid_member > /dev/sda2 ceph journal, for /dev/sda5 > /dev/sda5 ceph data, active, cluster ceph, osd.0, journal /dev/sda2 > > It was created with ceph-disk prepare. > > When I run "ceph-disk activate /dev/sda5", it is mounted and started. > > If I run "systemctl start ceph-disk@/dev/sda5", the same, it's OK. But > this is a service that can't be "enabled" !! > > But on reboot, nothing happen. The only thing which tries to start is > ceph-osd@0 service (enabled by ceph-disk, not me), and of course it > fails because its data is not mounted. > > I think udev rules should do this, but it does not seem to. > > > root@host102:~# sgdisk -i 2 /dev/sda > Partition GUID code: 45B0969E-9B03-4F30-B4C6-B4B80CEFF106 (Unknown) > Partition unique GUID: D0F4F00F-723D-4DAD-BA2E-93D52EB564C1 > First sector: 2048 (at 1024.0 KiB) > Last sector: 9765887 (at 4.7 GiB) > Partition size: 9763840 sectors (4.7 GiB) > Attribute flags: 0000000000000000 > Partition name: 'ceph journal' > > root@host102:~# sgdisk -i 5 /dev/sda > Partition GUID code: 4FBD7E29-9D25-41B8-AFD0-062C0CEFF05D (Unknown) > Partition unique GUID: 5AB4F732-AFBE-4DEA-A4C6-AD290C1302D9 > First sector: 123047424 (at 58.7 GiB) > Last sector: 1953459199 (at 931.5 GiB) > Partition size: 1830411776 sectors (872.8 GiB) > Attribute flags: 0000000000000000 > Partition name: 'ceph data' > > > Does someone have an idea of what's going on ? > > Thank you. > > Florent are you using the packages from ceph.com? if so, you might be affected by http://tracker.ceph.com/issues/18305 (and http://tracker.ceph.com/issues/17889) did you mask the ceph.service unit generated from the ceph init script? what does "systemctl status '*ceph*'" show? what does "journalctl -b '*ceph*'" show? what happens if you run "ceph-disk activate-all"? (this is what is called last in the init script and will probably trigger mounting of the OSD disk/partition and starting of the ceph-osd@.. service) _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com