Can ceph-deploy be used with 'osd objectstore = keyvaluestore-dev' in config file ?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



After a reboot all the redundant partitions have gone, but after running 
the script I still got:

ERROR:ceph-disk:Failed to activate
ceph-disk: Does not look like a Ceph OSD, or incompatible version: 
/var/lib/ceph/tmp/mnt.SFvU7O
ceph-disk: Error: One or more partitions failed to activate


? 2014-09-24 10:29, Aegeaner ??:
> I run ceph on Red Hat Enterprise Linux Server 6.4 Santiago, and when I 
> run "service ceph start" i got:
>
> # service ceph start
>
>     ERROR:ceph-disk:Failed to activate
>     ceph-disk: Does not look like a Ceph OSD, or incompatible version:
>     /var/lib/ceph/tmp/mnt.I71N5T
>     mount: /dev/hioa1 already mounted or /var/lib/ceph/tmp/mnt.02sVHj busy
>     ceph-disk: Mounting filesystem failed: Command '['/bin/mount',
>     '-t', 'xfs', '-o', 'noatime', '--',
>     '/dev/disk/by-parttypeuuid/4fbd7e29-9d25-41b8-afd0-062c0ceff05d.6d726c93-41f9-453d-858e-ab4132b5c8fd',
>     '/var/lib/ceph/tmp/mnt.02sVHj']' returned non-zero exit status 32
>     ceph-disk: Error: One or more partitions failed to activate
>
> Someone told me "service ceph start" still tries to call ceph-disk 
> which will create a filestore type OSD, and create a journal 
> partition, is it true?
>
> ls -l /dev/disk/by-parttypeuuid/
>
>     lrwxrwxrwx. 1 root root 11 9?  23 16:56
>     45b0969e-9b03-4f30-b4c6-b4b80ceff106.00dbee5e-fb68-47c4-aa58-924c904c4383
>     -> ../../hioa2
>     lrwxrwxrwx. 1 root root 10 9?  23 17:02
>     45b0969e-9b03-4f30-b4c6-b4b80ceff106.c30e5b97-b914-4eb8-8306-a9649e1c20ba
>     -> ../../sdb2
>     lrwxrwxrwx. 1 root root 11 9?  23 16:56
>     4fbd7e29-9d25-41b8-afd0-062c0ceff05d.6d726c93-41f9-453d-858e-ab4132b5c8fd
>     -> ../../hioa1
>     lrwxrwxrwx. 1 root root 10 9?  23 17:02
>     4fbd7e29-9d25-41b8-afd0-062c0ceff05d.b56ec699-e134-4b90-8f55-4952453e1b7e
>     -> ../../sdb1
>     lrwxrwxrwx. 1 root root 11 9?  23 16:52
>     89c57f98-2fe5-4dc0-89c1-f3ad0ceff2be.6d726c93-41f9-453d-858e-ab4132b5c8fd
>     -> ../../hioa1
>
> There seems to be two hioa1 partitions there, maybe remained from last 
> time I create the OSD using ceph-deploy osd prepare?
>
>
>
> ? 2014-09-24 10:19, Mark Kirkwood ??:
>> On 24/09/14 14:07, Aegeaner wrote:
>>> I turned on the debug option, and this is what I got:
>>>
>>> # ./kv.sh
>>>
>>>     removed osd.0
>>>     removed item id 0 name 'osd.0' from crush map
>>>     0
>>>     umount: /var/lib/ceph/osd/ceph-0: not found
>>>     updated
>>>     add item id 0 name 'osd.0' weight 1 at location
>>>     {host=CVM-0-11,root=default} to crush map
>>>     meta-data=/dev/hioa              isize=256    agcount=4,
>>>     agsize=24506368 blks
>>>               =                       sectsz=512   attr=2, 
>>> projid32bit=0
>>>     data     =                       bsize=4096 blocks=98025472,
>>>     imaxpct=25
>>>               =                       sunit=0      swidth=0 blks
>>>     naming   =version 2              bsize=4096   ascii-ci=0
>>>     log      =internal log           bsize=4096 blocks=47864, version=2
>>>               =                       sectsz=512   sunit=0 blks,
>>>     lazy-count=1
>>>     realtime =none                   extsz=4096   blocks=0, rtextents=0
>>>     2014-09-24 10:02:21.049162 7fe4cf3aa7a0  0 ceph version 0.80.5
>>>     (38b73c67d375a2552d8ed67843c8a65c2c0feba6), process ceph-osd, 
>>> pid 10252
>>>     2014-09-24 10:02:21.055433 7fe4cf3aa7a0  1 mkfs in
>>>     /var/lib/ceph/osd/ceph-0
>>>     2014-09-24 10:02:21.056359 7fe4cf3aa7a0  1 mkfs generated fsid
>>>     d613a61d-a1b4-4180-aea2-552944a2f0dc
>>>     2014-09-24 10:02:21.061349 7fe4cf3aa7a0  1 keyvaluestore backend
>>>     exists/created
>>>     2014-09-24 10:02:21.061377 7fe4cf3aa7a0  1 mkfs done in
>>>     /var/lib/ceph/osd/ceph-0
>>>     2014-09-24 10:02:21.065679 7fe4cf3aa7a0 -1 created object store
>>>     /var/lib/ceph/osd/ceph-0 journal /var/lib/ceph/osd/ceph-0/journal
>>>     for osd.0 fsid d90272ca-d8cc-41eb-b525-2cffe734aec0
>>>     2014-09-24 10:02:21.065776 7fe4cf3aa7a0 -1 auth: error reading 
>>> file:
>>>     /var/lib/ceph/osd/ceph-0/keyring: can't open
>>>     /var/lib/ceph/osd/ceph-0/keyring: (2) No such file or directory
>>>     2014-09-24 10:02:21.065889 7fe4cf3aa7a0 -1 created new key in
>>>     keyring /var/lib/ceph/osd/ceph-0/keyring
>>>     added key for osd.0
>>>
>>> # ceph osd tree
>>>
>>>     # id    weight    type name    up/down    reweight
>>>     -1    1    root default
>>>     -2    1        host CVM-0-11
>>>     0    1            osd.0    down    0
>>>
>>> Also I updated my simple script to create the OSD:
>>>
>>>     ceph osd rm 0
>>>     ceph osd crush rm osd.0
>>>     ceph osd create
>>>     umount /var/lib/ceph/osd/ceph-0
>>>     rm -rf /var/lib/ceph/osd/ceph-0
>>>     rm -rf /var/lib/ceph/osd/ceph-0
>>>     mkdir /var/lib/ceph/osd/ceph-0
>>>     ceph auth del osd.0
>>>     ceph osd crush add osd.0 1 root=default host=CVM-0-11
>>>     mkfs -t xfs -f /dev/hioa
>>>     mount  /dev/hioa /var/lib/ceph/osd/ceph-0
>>>     ceph-osd --id 0 -d --mkkey --mkfs --osd-data 
>>> /var/lib/ceph/osd/ceph-0
>>>     ceph auth add osd.0 osd 'allow *' mon 'allow profile osd' -i
>>>     /var/lib/ceph/osd/ceph-0/keyring
>>>     /etc/init.d/ceph start osd.0
>>>
>>
>> From where your log stops at, it would appear that your system start 
>> script is not even trying to get osd.0 up at all.
>>
>> Can we see an ls -l of /var/lib/ceph/osd/ceph-0?
>>
>> Also what os are you on? You might need to invoke via:
>>
>> $ service ceph start
>>
>> or similar.
>>
>> Cheers
>>
>> Mark
>>
>>
>>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140924/b2cae3fe/attachment.htm>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux