Re: Bluestore config issue with ceph orch

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

I can reproduce it with 16.2.7 (without wal_device). Although you could simply use a yaml file and run 'ceph orch apply -i osd.specs' this seems to be a bug since it's documented[1]. I'd recommend to create a tracker issue for this.

Thanks,
Eugen

[1] https://docs.ceph.com/en/pacific/cephadm/services/osd/#creating-new-osds

Zitat von Aristide Bekroundjo <bekroundjo@xxxxxxx>:

Hi dear,

I’m running a cluster under 16.2.10.

I’m trying to to configure a bluestore device with ceph orch command but I failed.



[ceph: root@monnode1 /]# ceph orch daemon add osd osdnode1:data_devices=/dev/sdd,wal_devices=/dev/sde,db_devices=/dev/sdf
Error EINVAL: Traceback (most recent call last):
  File "/usr/share/ceph/mgr/mgr_module.py", line 1446, in _handle_command
    return self.handle_command(inbuf, cmd)
File "/usr/share/ceph/mgr/orchestrator/_interface.py", line 171, in handle_command
[…]
/bin/docker: stderr [--osds-per-device OSDS_PER_DEVICE]
/bin/docker: stderr                              [--data-slots DATA_SLOTS]
/bin/docker: stderr [--block-db-size BLOCK_DB_SIZE] /bin/docker: stderr [--block-db-slots BLOCK_DB_SLOTS] /bin/docker: stderr [--block-wal-size BLOCK_WAL_SIZE]
[…]
line 1622, in call_throws
    raise RuntimeError('Failed command: %s' % ' '.join(command))
RuntimeError: Failed command: /bin/docker run --rm --ipc=host --stop-signal=SIGTERM --net=host –
[…]
11ed-8fc2-525400c2ee0b/crash:/var/lib/ceph/crash:z -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /var/lib/ceph/ca59e8b0-3398-11ed-8fc2-525400c2ee0b/selinux:/sys/fs/selinux:ro -v /:/rootfs -v /tmp/ceph-tmpa3g28z0w:/etc/ceph/ceph.conf:z -v /tmp/ceph-tmp8_s8fjc3:/var/lib/ceph/bootstrap-osd/ceph.keyring:z quay.io/ceph/ceph@sha256:c5fd9d806c54e5cc9db8efd50363e1edf7af62f101b264dccacb9d6091dcf7aa lvm batch --no-auto data_devices=/dev/sdd wal_devices=/dev/sde db_devices=/dev/sdf --yes --no-systemd

[ceph: root@monnode1 /]#



I know I can do it directly on host with ceph-volume, but how I do it with ceph orch ?

Best regards,
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux