Hi everyone, I have a problem with automatic start of OSDs on Debian Jessie with Ceph Jewel. My osd.0 is using /dev/sda5 for data and /dev/sda2 for journal, it is listed in ceph-disk list : /dev/sda : /dev/sda1 other, 21686148-6449-6e6f-744e-656564454649 /dev/sda3 other, linux_raid_member /dev/sda4 other, linux_raid_member /dev/sda2 ceph journal, for /dev/sda5 /dev/sda5 ceph data, active, cluster ceph, osd.0, journal /dev/sda2 It was created with ceph-disk prepare. When I run "ceph-disk activate /dev/sda5", it is mounted and started. If I run "systemctl start ceph-disk@/dev/sda5", the same, it's OK. But this is a service that can't be "enabled" !! But on reboot, nothing happen. The only thing which tries to start is ceph-osd@0 service (enabled by ceph-disk, not me), and of course it fails because its data is not mounted. I think udev rules should do this, but it does not seem to. root@host102:~# sgdisk -i 2 /dev/sda Partition GUID code: 45B0969E-9B03-4F30-B4C6-B4B80CEFF106 (Unknown) Partition unique GUID: D0F4F00F-723D-4DAD-BA2E-93D52EB564C1 First sector: 2048 (at 1024.0 KiB) Last sector: 9765887 (at 4.7 GiB) Partition size: 9763840 sectors (4.7 GiB) Attribute flags: 0000000000000000 Partition name: 'ceph journal' root@host102:~# sgdisk -i 5 /dev/sda Partition GUID code: 4FBD7E29-9D25-41B8-AFD0-062C0CEFF05D (Unknown) Partition unique GUID: 5AB4F732-AFBE-4DEA-A4C6-AD290C1302D9 First sector: 123047424 (at 58.7 GiB) Last sector: 1953459199 (at 931.5 GiB) Partition size: 1830411776 sectors (872.8 GiB) Attribute flags: 0000000000000000 Partition name: 'ceph data' Does someone have an idea of what's going on ? Thank you. Florent _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com