Re: Adding OSDs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



> On 11 Jun 2020, at 15:21, Eugen Block <eblock@xxxxxx> wrote:
> 
> Can you share which guide and deployment strategy you're following? I didn't have any issues deploying either completely manually [3] or with cephadm.


I followed the cephadm guide at https://ceph.readthedocs.io/en/latest/cephadm/install/ <https://ceph.readthedocs.io/en/latest/cephadm/install/> as the docs said it was the recommended method.

> 
> The OSD nodes need a keyring to bootstrap OSDs:
> 


Mmkay, done that. Now I get :

Running command: /usr/bin/ceph-authtool --gen-print-key
Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new a7e9874e-883d-458c-ae27-d828c6d262da
Running command: /sbin/lvcreate --yes -l 100%FREE -n osd-block-a7e9874e-883d-458c-ae27-d828c6d262da ceph-7df92c86-ad9b-4421-974e-381c61a7d9d1
 stderr: Volume group "ceph-7df92c86-ad9b-4421-974e-381c61a7d9d1" not found
  Cannot process volume group ceph-7df92c86-ad9b-4421-974e-381c61a7d9d1
--> Was unable to complete a new OSD, will rollback changes
Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd purge-new osd.0 --yes-i-really-mean-it
 stderr: purged osd.0
-->  RuntimeError: command returned non-zero exit status: 5


.. now, this is probably my fault from too much tinkering. Where is this VG name coming from? Can I get back to a clean slate? (Or at least remove references to non-existing VGs?)

Will


> ---snip---
> admin:~ # ceph auth get client.bootstrap-osd
> exported keyring for client.bootstrap-osd
> [client.bootstrap-osd]
>        key = AQCtsWNdAAAAABAAV6g3yc7rSa0yvsfO1Xlj5w==
>        caps mgr = "allow r"
>        caps mon = "allow profile bootstrap-osd"
> 
> osd1:~ # cat  /var/lib/ceph/bootstrap-osd/ceph.keyring
> [client.bootstrap-osd]
>        key = AQCtsWNdAAAAABAAV6g3yc7rSa0yvsfO1Xlj5w==
>        caps mgr = "allow r"
>        caps mon = "allow profile bootstrap-osd"
> ---snip---
> 
> 
> The manual deployment guide describes the procedure how to create OSDs from scratch. I would recommend to setup a cluster manually to get familiar with all the components that a deployment tool like cephadm will manage for you.
> 
> 
>> Again, can WAL and DB be pointed at the same SSD?
> 
> Yes, in that case you don't need to specify a WAL device, it will be automatically placed on the DB device.
> 
> 
> [3] https://docs.ceph.com/docs/octopus/install/index_manual/#deploy-a-cluster-manually
> 
> 
> Zitat von Will Payne <will@xxxxxxxxxxxxxxxx>:
> 
>>> If you want to specify vgname/lvname you have to create them manually and run:
>>> 
>>> ceph-volume lvm create --data /dev/sdc --block.db vgname/db-sdc --block.wal vgname/wal-sdc
>> 
>> Apt helpfully told me I need to install ceph-osd, which I did. The ceph-volume command then told me :
>> 
>> Running command: /usr/bin/ceph-authtool --gen-print-key
>> Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new ...
>> stderr: [errno 13] RADOS permission denied (error connecting to the cluster)
>> -->  RuntimeError: Unable to create a new OSD id
>> 
>> I note the keyring file doesn’t exist - is this something I need to create or should exist already? Or will it be created by this command?
>> 
>>> ceph-volume lvm batch /dev/sdb ... /dev/sdj --db-devices /dev/sdk --wal-devices /dev/sdl
>>> but then you don't get to choose vgname/lvname, the batch command creates the volume groups and logical volumes for you.
>> 
>> I’m not too fussed about the vg/lv names if they’re created for me. Can the db and Wal devices be the same device?
>> 
>>> I would say the drive_groups approach is designed to automate things so you don't need to worry about creating VGs and LVs.
>> 
>> Again, can WAL and DB be pointed at the same SSD?
>> 
>> 
>> Sorry if I’m being stupid, the documentation just seems a little lacking.
>> 
>> Will
> 
> 
> 

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux