On Wed, Jan 04, 2017 at 12:55:56PM +0100, Florent B wrote: > On 01/04/2017 12:18 PM, Fabian Grünbichler wrote: > > On Wed, Jan 04, 2017 at 12:03:39PM +0100, Florent B wrote: > >> Hi everyone, > >> > >> I have a problem with automatic start of OSDs on Debian Jessie with Ceph > >> Jewel. > >> > >> My osd.0 is using /dev/sda5 for data and /dev/sda2 for journal, it is > >> listed in ceph-disk list : > >> > >> /dev/sda : > >> /dev/sda1 other, 21686148-6449-6e6f-744e-656564454649 > >> /dev/sda3 other, linux_raid_member > >> /dev/sda4 other, linux_raid_member > >> /dev/sda2 ceph journal, for /dev/sda5 > >> /dev/sda5 ceph data, active, cluster ceph, osd.0, journal /dev/sda2 > >> > >> It was created with ceph-disk prepare. > >> > >> When I run "ceph-disk activate /dev/sda5", it is mounted and started. > >> > >> If I run "systemctl start ceph-disk@/dev/sda5", the same, it's OK. But > >> this is a service that can't be "enabled" !! > >> > >> But on reboot, nothing happen. The only thing which tries to start is > >> ceph-osd@0 service (enabled by ceph-disk, not me), and of course it > >> fails because its data is not mounted. > >> > >> I think udev rules should do this, but it does not seem to. > >> > >> > >> root@host102:~# sgdisk -i 2 /dev/sda > >> Partition GUID code: 45B0969E-9B03-4F30-B4C6-B4B80CEFF106 (Unknown) > >> Partition unique GUID: D0F4F00F-723D-4DAD-BA2E-93D52EB564C1 > >> First sector: 2048 (at 1024.0 KiB) > >> Last sector: 9765887 (at 4.7 GiB) > >> Partition size: 9763840 sectors (4.7 GiB) > >> Attribute flags: 0000000000000000 > >> Partition name: 'ceph journal' > >> > >> root@host102:~# sgdisk -i 5 /dev/sda > >> Partition GUID code: 4FBD7E29-9D25-41B8-AFD0-062C0CEFF05D (Unknown) > >> Partition unique GUID: 5AB4F732-AFBE-4DEA-A4C6-AD290C1302D9 > >> First sector: 123047424 (at 58.7 GiB) > >> Last sector: 1953459199 (at 931.5 GiB) > >> Partition size: 1830411776 sectors (872.8 GiB) > >> Attribute flags: 0000000000000000 > >> Partition name: 'ceph data' > >> > >> > >> Does someone have an idea of what's going on ? > >> > >> Thank you. > >> > >> Florent > > are you using the packages from ceph.com? if so, you might be affected > > by http://tracker.ceph.com/issues/18305 (and > > http://tracker.ceph.com/issues/17889) > > > > did you mask the ceph.service unit generated from the ceph init script? > > > > what does "systemctl status '*ceph*'" show? what does "journalctl -b > > '*ceph*'" show? > > > > what happens if you run "ceph-disk activate-all"? (this is what is > > called last in the init script and will probably trigger mounting of the > > OSD disk/partition and starting of the ceph-osd@.. service) > > > > Thank you, that was the problem : I disabled ceph.service unit because I > thought it was an "old" thing, I didn't knew it is always used. > Re-enabling it did the trick. > > Isn't it an "old way" of doing things ? > I am not sure if the init script was left on purpose or if nobody realized that the existing systemd units don't cover all the activation paths because the init script was forgotten and hides this fact quite well. I assume the latter ;) IMHO the current situation is wrong, which is why I filed the bug (including a proposed fix). especially since the init script actually starts monitors using systemd-run as transient units instead of via ceph-mon@XYZ, so on monitor nodes the startup situation can get quite confusing and racy. so far there hasn't been any feedback - maybe this thread will help and get some more eyes to look at it.. _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com