Not sure what I'm doing wrong, I suspect its the way I'm running ceph-volume. root@drywood12:~# cephadm ceph-volume lvm create --data /dev/sda --dmcrypt Inferring fsid 1518c8e0-bbe4-11eb-9772-001e67dc85ea Using recent ceph image ceph/ceph@sha256 :54e95ae1e11404157d7b329d0bef866ebbb214b195a009e87aae4eba9d282949 /usr/bin/docker: Running command: /usr/bin/ceph-authtool --gen-print-key /usr/bin/docker: Running command: /usr/bin/ceph-authtool --gen-print-key /usr/bin/docker: --> RuntimeError: No valid ceph configuration file was loaded. Traceback (most recent call last): File "/usr/sbin/cephadm", line 8029, in <module> main() File "/usr/sbin/cephadm", line 8017, in main r = ctx.func(ctx) File "/usr/sbin/cephadm", line 1678, in _infer_fsid return func(ctx) File "/usr/sbin/cephadm", line 1738, in _infer_image return func(ctx) File "/usr/sbin/cephadm", line 4514, in command_ceph_volume out, err, code = call_throws(ctx, c.run_cmd(), verbosity=verbosity) File "/usr/sbin/cephadm", line 1464, in call_throws raise RuntimeError('Failed command: %s' % ' '.join(command)) RuntimeError: Failed command: /usr/bin/docker run --rm --ipc=host --net=host --entrypoint /usr/sbin/ceph-volume --privileged --group-add=disk --init -e CONTAINER_IMAGE=ceph/ceph@sha256:54e95ae1e11404157d7b329d0t root@drywood12:~# cephadm shell Inferring fsid 1518c8e0-bbe4-11eb-9772-001e67dc85ea Inferring config /var/lib/ceph/1518c8e0-bbe4-11eb-9772-001e67dc85ea/mon.drywood12/config Using recent ceph image ceph/ceph@sha256 :54e95ae1e11404157d7b329d0bef866ebbb214b195a009e87aae4eba9d282949 root@drywood12:/# ceph-volume lvm create --data /dev/sda --dmcrypt Running command: /usr/bin/ceph-authtool --gen-print-key Running command: /usr/bin/ceph-authtool --gen-print-key Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 70054a5c-c176-463a-a0ac-b44c5db0987c stderr: 2021-05-25T07:46:18.188+0000 7fdef8f0d700 -1 auth: unable to find a keyring on /var/lib/ceph/bootstrap-osd/ceph.keyring: (2) No such file or directory stderr: 2021-05-25T07:46:18.188+0000 7fdef8f0d700 -1 AuthRegistry(0x7fdef405b378) no keyring found at /var/lib/ceph/bootstrap-osd/ceph.keyring, disabling cephx stderr: 2021-05-25T07:46:18.188+0000 7fdef8f0d700 -1 auth: unable to find a keyring on /var/lib/ceph/bootstrap-osd/ceph.keyring: (2) No such file or directory stderr: 2021-05-25T07:46:18.188+0000 7fdef8f0d700 -1 AuthRegistry(0x7fdef405ef20) no keyring found at /var/lib/ceph/bootstrap-osd/ceph.keyring, disabling cephx stderr: 2021-05-25T07:46:18.188+0000 7fdef8f0d700 -1 auth: unable to find a keyring on /var/lib/ceph/bootstrap-osd/ceph.keyring: (2) No such file or directory stderr: 2021-05-25T07:46:18.188+0000 7fdef8f0d700 -1 AuthRegistry(0x7fdef8f0bea0) no keyring found at /var/lib/ceph/bootstrap-osd/ceph.keyring, disabling cephx stderr: 2021-05-25T07:46:18.188+0000 7fdef2d9d700 -1 monclient(hunting): handle_auth_bad_method server allowed_methods [2] but i only support [1] stderr: 2021-05-25T07:46:18.188+0000 7fdef259c700 -1 monclient(hunting): handle_auth_bad_method server allowed_methods [2] but i only support [1] stderr: 2021-05-25T07:46:18.188+0000 7fdef1d9b700 -1 monclient(hunting): handle_auth_bad_method server allowed_methods [2] but i only support [1] stderr: 2021-05-25T07:46:18.188+0000 7fdef8f0d700 -1 monclient: authenticate NOTE: no keyring found; disabled cephx authentication stderr: [errno 13] RADOS permission denied (error connecting to the cluster) --> RuntimeError: Unable to create a new OSD id root@drywood12:/# lsblk /dev/sda NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT sda 8:0 0 7.3T 0 disk As far as I can see cephadm gets a little further than this as the disks have lvm volumes on them just the osd's daemons are not created or started. So maybe I'm invoking ceph-volume incorrectly. On Tue, 25 May 2021 at 06:57, Peter Childs <pchilds@xxxxxxx> wrote: > > > On Mon, 24 May 2021, 21:08 Marc, <Marc@xxxxxxxxxxxxxxxxx> wrote: > >> > >> > I'm attempting to use cephadm and Pacific, currently on debian buster, >> > mostly because centos7 ain't supported any more and cenotos8 ain't >> > support >> > by some of my hardware. >> >> Who says centos7 is not supported any more? Afaik centos7/el7 is being >> supported till its EOL 2024. By then maybe a good alternative for >> el8/stream has surfaced. >> > > Not supported by ceph Pacific, it's our os of choice otherwise. > > My testing says the version available of podman, docker and python3, do > not work with Pacific. > > Given I've needed to upgrade docker on buster can we please have a list of > versions that work with cephadm, maybe even have cephadm say no, please > upgrade unless your running the right version or better. > > > >> > Anyway I have a few nodes with 59x 7.2TB disks but for some reason the >> > osd >> > daemons don't start, the disks get formatted and the osd are created but >> > the daemons never come up. >> >> what if you try with >> ceph-volume lvm create --data /dev/sdi --dmcrypt ? >> > > I'll have a go. > > >> > They are probably the wrong spec for ceph (48gb of memory and only 4 >> > cores) >> >> You can always start with just configuring a few disks per node. That >> should always work. >> > > That was my thought too. > > Thanks > > Peter > > >> > but I was expecting them to start and be either dirt slow or crash >> > later, >> > anyway I've got upto 30 of them, so I was hoping on getting at least get >> > 6PB of raw storage out of them. >> > >> > As yet I've not spotted any helpful error messages. >> > >> _______________________________________________ >> ceph-users mailing list -- ceph-users@xxxxxxx >> To unsubscribe send an email to ceph-users-leave@xxxxxxx >> > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx