Thank you. The image solved our problem. Jan ________________________________ Von: David Orman <ormandj@xxxxxxxxxxxx> Gesendet: Dienstag, 22. Juni 2021 17:27:12 An: Jansen, Jan Cc: ceph-users Betreff: Re: Having issues to start more than 24 OSDs per host https://tracker.ceph.com/issues/50526 https://github.com/alfredodeza/remoto/issues/62 If you're brave (YMMV, test first non-prod), we pushed an image with the issue we encountered fixed as per above here: https://hub.docker.com/repository/docker/ormandj/ceph/tags?page=1 that you can use to install with. I'm not sure when the next release is due out (I'm a little confused why a breaking install/upgrade issue like this has been allowed to sit), but it should include this fix, as well as others. David On Tue, Jun 22, 2021 at 1:16 AM <Jan.Jansen@xxxxxxxx> wrote: > > Hello > > We did try to use Cephadm with Podman to start 44 OSDs per host which consistently stop after adding 24 OSDs per host. > We did look into the cephadm.log on the problematic host and saw that the command `cephadm ceph-volume lvm list --format json` did stuck. > We were the output of the command wasn't complete. Therefore, we tried to use compacted JSON and we could increase the number to 36 OSDs per host. > > If you need more information just ask. > > > Podman version: 3.2.1 > Ceph version: 16.2.4 > OS version: Suse Leap 15.3 > > Greetings, > Jan > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx