Radoslav, I ran into the same. For Debian 10 - recent updates - you have to add 'cgroup_enable=memory swapaccount=1' to the kernel command line (/etc/default/grub). The reference I found said that Debian decided to disable this by default and make us turn it on if we want to run containers. -Dave -- Dave Hall Binghamton University kdhall@xxxxxxxxxxxxxx On Wed, Apr 14, 2021 at 12:51 PM Radoslav Milanov < radoslav.milanov@xxxxxxxxx> wrote: > Hello, > > Cluster is 3 nodes Debian 10. Started cephadm upgrade on healthy 15.2.10 > cluster. Managers were upgraded fine then first monitor went down for > upgrade and never came back. Researching at the unit files container > fails to run because of an error: > > root@host1:/var/lib/ceph/97d9f40e-9d33-11eb-8e3f-1c34da4b9fb6/mon.host1# > cat unit.run > > set -e > /usr/bin/install -d -m0770 -o 167 -g 167 > /var/run/ceph/97d9f40e-9d33-11eb-8e3f-1c34da4b9fb6 > # mon.host1 > ! /usr/bin/docker rm -f > ceph-97d9f40e-9d33-11eb-8e3f-1c34da4b9fb6-mon.host1 2> /dev/null > /usr/bin/docker run --rm --ipc=host --net=host --entrypoint > /usr/bin/ceph-mon --privileged --group-add=disk --init --name > ceph-97d9f40e-9d33-11eb-8e3f-1c34da4b9fb6-mon.host1 -e > CONTAINER_IMAGE=ceph/ceph@sha256:9b04c0f15704c49591640a37c7adfd40ffad0a4b42fecb950c3407687cb4f29a > > -e NODE_NAME=host1 -e CEPH_USE_RANDOM_NONCE=1 -v > /var/run/ceph/97d9f40e-9d33-11eb-8e3f-1c34da4b9fb6:/var/run/ceph:z -v > /var/log/ceph/97d9f40e-9d33-11eb-8e3f-1c34da4b9fb6:/var/log/ceph:z -v > /var/lib/ceph/97d9f40e-9d33-11eb-8e3f-1c34da4b9fb6/crash:/var/lib/ceph/crash:z > > -v > /var/lib/ceph/97d9f40e-9d33-11eb-8e3f-1c34da4b9fb6/mon.host1:/var/lib/ceph/mon/ceph-host1:z > > -v > /var/lib/ceph/97d9f40e-9d33-11eb-8e3f-1c34da4b9fb6/mon.host1/config:/etc/ceph/ceph.conf:z > > -v /dev:/dev -v /run/udev:/run/udev > ceph/ceph@sha256:9b04c0f15704c49591640a37c7adfd40ffad0a4b42fecb950c3407687cb4f29a > > -n mon.host1 -f --setuser ceph --setgroup ceph > --default-log-to-file=false --default-log-to-stderr=true > '--default-log-stderr-prefix=debug ' > --default-mon-cluster-log-to-file=false > --default-mon-cluster-log-to-stderr=true > > root@host1:/var/lib/ceph/97d9f40e-9d33-11eb-8e3f-1c34da4b9fb6/mon.host1# > /usr/bin/docker run --rm --ipc=host --net=host --entrypoint > /usr/bin/ceph-mon --privileged --group-add=disk --init --name > ceph-97d9f40e-9d33-11eb-8e3f-1c34da4b9fb6-mon.host1 -e > CONTAINER_IMAGE=ceph/ceph@sha256:9b04c0f15704c49591640a37c7adfd40ffad0a4b42fecb950c3407687cb4f29a > > -e NODE_NAME=host1 -e CEPH_USE_RANDOM_NONCE=1 -v > /var/run/ceph/97d9f40e-9d33-11eb-8e3f-1c34da4b9fb6:/var/run/ceph:z -v > /var/log/ceph/97d9f40e-9d33-11eb-8e3f-1c34da4b9fb6:/var/log/ceph:z -v > /var/lib/ceph/97d9f40e-9d33-11eb-8e3f-1c34da4b9fb6/crash:/var/lib/ceph/crash:z > > -v > /var/lib/ceph/97d9f40e-9d33-11eb-8e3f-1c34da4b9fb6/mon.host1:/var/lib/ceph/mon/ceph-host1:z > > -v > /var/lib/ceph/97d9f40e-9d33-11eb-8e3f-1c34da4b9fb6/mon.host1/config:/etc/ceph/ceph.conf:z > > -v /dev:/dev -v /run/udev:/run/udev > ceph/ceph@sha256:9b04c0f15704c49591640a37c7adfd40ffad0a4b42fecb950c3407687cb4f29a > > -n mon.host1 -f --setuser ceph --setgroup ceph > --default-log-to-file=false --default-log-to-stderr=true > '--default-log-stderr-prefix=debug ' > --default-mon-cluster-log-to-file=false > --default-mon-cluster-log-to-stderr=true > > > /usr/bin/docker: Error response from daemon: OCI runtime create failed: > container_linux.go:344: starting container process caused "exec: > \"/dev/init\": stat /dev/init: no such file or directory": unknown. > > Any suggestions how to resolve that ? > > Thank you. > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx