On 01/04/2017 01:44 PM, Fabian Grünbichler wrote: > On Wed, Jan 04, 2017 at 12:55:56PM +0100, Florent B wrote: >> On 01/04/2017 12:18 PM, Fabian Grünbichler wrote: >>> On Wed, Jan 04, 2017 at 12:03:39PM +0100, Florent B wrote: >>>> Hi everyone, >>>> >>>> I have a problem with automatic start of OSDs on Debian Jessie with Ceph >>>> Jewel. >>>> >>>> My osd.0 is using /dev/sda5 for data and /dev/sda2 for journal, it is >>>> listed in ceph-disk list : >>>> >>>> /dev/sda : >>>> /dev/sda1 other, 21686148-6449-6e6f-744e-656564454649 >>>> /dev/sda3 other, linux_raid_member >>>> /dev/sda4 other, linux_raid_member >>>> /dev/sda2 ceph journal, for /dev/sda5 >>>> /dev/sda5 ceph data, active, cluster ceph, osd.0, journal /dev/sda2 >>>> >>>> It was created with ceph-disk prepare. >>>> >>>> When I run "ceph-disk activate /dev/sda5", it is mounted and started. >>>> >>>> If I run "systemctl start ceph-disk@/dev/sda5", the same, it's OK. But >>>> this is a service that can't be "enabled" !! >>>> >>>> But on reboot, nothing happen. The only thing which tries to start is >>>> ceph-osd@0 service (enabled by ceph-disk, not me), and of course it >>>> fails because its data is not mounted. >>>> >>>> I think udev rules should do this, but it does not seem to. >>>> >>>> >>>> root@host102:~# sgdisk -i 2 /dev/sda >>>> Partition GUID code: 45B0969E-9B03-4F30-B4C6-B4B80CEFF106 (Unknown) >>>> Partition unique GUID: D0F4F00F-723D-4DAD-BA2E-93D52EB564C1 >>>> First sector: 2048 (at 1024.0 KiB) >>>> Last sector: 9765887 (at 4.7 GiB) >>>> Partition size: 9763840 sectors (4.7 GiB) >>>> Attribute flags: 0000000000000000 >>>> Partition name: 'ceph journal' >>>> >>>> root@host102:~# sgdisk -i 5 /dev/sda >>>> Partition GUID code: 4FBD7E29-9D25-41B8-AFD0-062C0CEFF05D (Unknown) >>>> Partition unique GUID: 5AB4F732-AFBE-4DEA-A4C6-AD290C1302D9 >>>> First sector: 123047424 (at 58.7 GiB) >>>> Last sector: 1953459199 (at 931.5 GiB) >>>> Partition size: 1830411776 sectors (872.8 GiB) >>>> Attribute flags: 0000000000000000 >>>> Partition name: 'ceph data' >>>> >>>> >>>> Does someone have an idea of what's going on ? >>>> >>>> Thank you. >>>> >>>> Florent >>> are you using the packages from ceph.com? if so, you might be affected >>> by http://tracker.ceph.com/issues/18305 (and >>> http://tracker.ceph.com/issues/17889) >>> >>> did you mask the ceph.service unit generated from the ceph init script? >>> >>> what does "systemctl status '*ceph*'" show? what does "journalctl -b >>> '*ceph*'" show? >>> >>> what happens if you run "ceph-disk activate-all"? (this is what is >>> called last in the init script and will probably trigger mounting of the >>> OSD disk/partition and starting of the ceph-osd@.. service) >>> >> Thank you, that was the problem : I disabled ceph.service unit because I >> thought it was an "old" thing, I didn't knew it is always used. >> Re-enabling it did the trick. >> >> Isn't it an "old way" of doing things ? >> > I am not sure if the init script was left on purpose or if nobody > realized that the existing systemd units don't cover all the activation > paths because the init script was forgotten and hides this fact quite > well. I assume the latter ;) > > IMHO the current situation is wrong, which is why I filed the bug > (including a proposed fix). especially since the init script actually > starts monitors using systemd-run as transient units instead of via > ceph-mon@XYZ, so on monitor nodes the startup situation can get quite > confusing and racy. so far there hasn't been any feedback - maybe this > thread will help and get some more eyes to look at it.. > I agree with you, all this needs to be cleaned up ;-) _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com