I have got my ceph OSDs running with keyvalue store now! Thank Mark! I have been confused for a whole week. ============ Cheers Aegeaner ? 2014-09-24 10:46, Mark Kirkwood ??: > On 24/09/14 14:29, Aegeaner wrote: >> I run ceph on Red Hat Enterprise Linux Server 6.4 Santiago, and when I >> run "service ceph start" i got: >> >> # service ceph start >> >> ERROR:ceph-disk:Failed to activate >> ceph-disk: Does not look like a Ceph OSD, or incompatible version: >> /var/lib/ceph/tmp/mnt.I71N5T >> mount: /dev/hioa1 already mounted or /var/lib/ceph/tmp/mnt.02sVHj >> busy >> ceph-disk: Mounting filesystem failed: Command '['/bin/mount', '-t', >> 'xfs', '-o', 'noatime', '--', >> '/dev/disk/by-parttypeuuid/4fbd7e29-9d25-41b8-afd0-062c0ceff05d.6d726c93-41f9-453d-858e-ab4132b5c8fd', >> '/var/lib/ceph/tmp/mnt.02sVHj']' returned non-zero exit status 32 >> ceph-disk: Error: One or more partitions failed to activate >> >> Someone told me "service ceph start" still tries to call ceph-disk which >> will create a filestore type OSD, and create a journal partition, is it >> true? >> >> ls -l /dev/disk/by-parttypeuuid/ >> >> lrwxrwxrwx. 1 root root 11 9? 23 16:56 >> 45b0969e-9b03-4f30-b4c6-b4b80ceff106.00dbee5e-fb68-47c4-aa58-924c904c4383 >> -> ../../hioa2 >> lrwxrwxrwx. 1 root root 10 9? 23 17:02 >> 45b0969e-9b03-4f30-b4c6-b4b80ceff106.c30e5b97-b914-4eb8-8306-a9649e1c20ba >> -> ../../sdb2 >> lrwxrwxrwx. 1 root root 11 9? 23 16:56 >> 4fbd7e29-9d25-41b8-afd0-062c0ceff05d.6d726c93-41f9-453d-858e-ab4132b5c8fd >> -> ../../hioa1 >> lrwxrwxrwx. 1 root root 10 9? 23 17:02 >> 4fbd7e29-9d25-41b8-afd0-062c0ceff05d.b56ec699-e134-4b90-8f55-4952453e1b7e >> -> ../../sdb1 >> lrwxrwxrwx. 1 root root 11 9? 23 16:52 >> 89c57f98-2fe5-4dc0-89c1-f3ad0ceff2be.6d726c93-41f9-453d-858e-ab4132b5c8fd >> -> ../../hioa1 >> >> There seems to be two hioa1 partitions there, maybe remained from last >> time I create the OSD using ceph-deploy osd prepare? >> > > Crap - it is fighting you, yes - looks like the startup script has > tried to build an osd for you using ceph-disk (which will make two > partitions by default). So that's toasted the setup that your script did. > > Growl - that's made it more complicated for sure. > > If you re-run your script you'll blast away the damage that 'service' > did :-) , and take a look at /etc/init.d/ceph to see why it ignored > your osd.0 arg (I'm not sure what it expects - maybe just 'osd'). > Anyway experiment. > > You can always start the osd with: > > $ sudo ceph-osd -i 0 > > which bypasses the whole system startup confusion completely :-) > > Cheers > > Mark > >