Zitat von Moritz Baumann <mo@xxxxxxxxxxxxx>:
The cluster was successfully upgraded to 19.2.1 and has been running some time with that version before it broke.As you suggested I have changed the unit.run file and replaced the sha256:XYZ.... with the one from https://quay.io/repository/ceph/ceph?tab=tags( sha256:200087c35811bf28e8a8073b15fa86c07cce85c575f1ccd62d1d6ddbfdc6770a )ceph01[1]:/var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/mgr.ceph01.ombogc# diff unit.run* -Nur--- unit.run 2025-03-12 13:33:43.455465357 +0100 +++ unit.run.bak 2025-03-12 13:32:44.713805212 +0100 @@ -5,4 +5,4 @@! /usr/bin/podman rm -f ceph-4cfb1138-d2a5-11ef-87eb-1c34da408e52-mgr-ceph01-ombogc 2> /dev/null ! /usr/bin/podman rm -f --storage ceph-4cfb1138-d2a5-11ef-87eb-1c34da408e52-mgr-ceph01-ombogc 2> /dev/null ! /usr/bin/podman rm -f --storage ceph-4cfb1138-d2a5-11ef-87eb-1c34da408e52-mgr.ceph01.ombogc 2> /dev/null -/usr/bin/podman run --rm --ipc=host --stop-signal=SIGTERM --net=host --entrypoint /usr/bin/ceph-mgr --init --name ceph-4cfb1138-d2a5-11ef-87eb-1c34da408e52-mgr-ceph01-ombogc --pids-limit=-1 -d --log-driver journald --conmon-pidfile /run/ceph-4cfb1138-d2a5-11ef-87eb-1c34da408e52@mgr.ceph01.ombogc.service-pid --cidfile /run/ceph-4cfb1138-d2a5-11ef-87eb-1c34da408e52@mgr.ceph01.ombogc.service-cid --cgroups=split -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:200087c35811bf28e8a8073b15fa86c07cce85c575f1ccd62d1d6ddbfdc6770a -e NODE_NAME=ceph01.inf.ethz.ch -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52:/var/run/ceph:z -v /var/log/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52:/var/log/ceph:z -v /var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/crash:/var/lib/ceph/crash:z -v /run/systemd/journal:/run/systemd/journal -v /var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/mgr.ceph01.ombogc:/var/lib/ceph/mgr/ceph-ceph01.ombogc:z -v /var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/mgr.ceph01.ombogc/config:/etc/ceph/ceph.conf:z -v /etc/hosts:/etc/hosts:ro quay.io/ceph/ceph@sha256:200087c35811bf28e8a8073b15fa86c07cce85c575f1ccd62d1d6ddbfdc6770a -n mgr.ceph01.ombogc -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false +/usr/bin/podman run --rm --ipc=host --stop-signal=SIGTERM --net=host --entrypoint /usr/bin/ceph-mgr --init --name ceph-4cfb1138-d2a5-11ef-87eb-1c34da408e52-mgr-ceph01-ombogc --pids-limit=-1 -d --log-driver journald --conmon-pidfile /run/ceph-4cfb1138-d2a5-11ef-87eb-1c34da408e52@mgr.ceph01.ombogc.service-pid --cidfile /run/ceph-4cfb1138-d2a5-11ef-87eb-1c34da408e52@mgr.ceph01.ombogc.service-cid --cgroups=split -e CONTAINER_IMAGE=quay.io/ceph/ceph@sha256:41d3f5e46ff7de28544cc8869fdea13fca824dcef83936cb3288ed9de935e4de -e NODE_NAME=ceph01.inf.ethz.ch -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52:/var/run/ceph:z -v /var/log/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52:/var/log/ceph:z -v /var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/crash:/var/lib/ceph/crash:z -v /run/systemd/journal:/run/systemd/journal -v /var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/mgr.ceph01.ombogc:/var/lib/ceph/mgr/ceph-ceph01.ombogc:z -v /var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/mgr.ceph01.ombogc/config:/etc/ceph/ceph.conf:z -v /etc/hosts:/etc/hosts:ro quay.io/ceph/ceph@sha256:41d3f5e46ff7de28544cc8869fdea13fca824dcef83936cb3288ed9de935e4de -n mgr.ceph01.ombogc -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=falsestopped the 19.2.1 version # cephadm unit stop --name mgr.ceph01.ombogc # cephadm unit start --name mgr.ceph01.ombogc and failed to the other mgrs # ceph mgr fail ceph01.ombogc # ceph orch host ls Error ENOENT: Module not found # ceph mgr fail ceph02.nqlpce # ceph orch host ls Error ENOENT: Module not found # podman exec -ti <id of the mgr> /bin/bash in the container ceph --versionon both nodes shows that one is running 19.2.1 and the other one is running 19.2.0but no matter which mgr is active the ceph orch commands do not work.we were only in the process of making this cluster productive, so if it comes to the wort I can copy the little data on it somewhere and reinstall it, but I hoped to find a solution (to feel better that when this should happen again there is a way to recover).On 3/10/25 22:55, Eugen Block wrote:I don’t have a better idea than to downgrade one MGR by editing the unit.run file to use a different version (19.2.0 seems to have worked, right?), restart it, then fail the MGRs until the older one takes over (or stop all newer ones). And then check if the orchestrator still misbehaves.Zitat von Moritz Baumann <mo@xxxxxxxxxxxxx>:It was installed with 19.2.0: dnf history info 20 shows: ceph01[0]:~# dnf history info 20 Transaction ID : 20 Begin time : Tue 14 Jan 2025 07:22:00 PM CETBegin rpmdb : 0fac0c11ba074f9635117d0b3d4970e6d26e8d2b40bfe8b93d22fd12147260b7End time : Tue 14 Jan 2025 07:22:04 PM CET (4 seconds)End rpmdb : 63657f6d4b27790107998bd21da76fe182d9c121177b34cf879243884deba40aUser : root <root> Return-Code : Success Releasever : 9 Command Line : install -y ceph-common cephadm Comment : Packages Altered: Install boost-program-options-1.75.0-8.el9.x86_64 @appstream Install libbabeltrace-1.5.8-10.el9.x86_64 @appstream Install libpmem-1.12.1-1.el9.x86_64 @appstream Install libpmemobj-1.12.1-1.el9.x86_64 @appstream Install librabbitmq-0.11.0-7.el9.x86_64 @appstream Install librdkafka-1.6.1-102.el9.x86_64 @appstream Install lttng-ust-2.12.0-6.el9.x86_64 @appstream Install python3-prettytable-0.7.2-27.el9.noarch @appstream Install daxctl-libs-78-2.el9.x86_64 @baseos Install librdmacm-51.0-1.el9.x86_64 @baseos Install ndctl-libs-78-2.el9.x86_64 @baseos Install ceph-common-2:19.2.0-0.el9.x86_64 @Ceph Install libcephfs2-2:19.2.0-0.el9.x86_64 @Ceph Install librados2-2:19.2.0-0.el9.x86_64 @Ceph Install libradosstriper1-2:19.2.0-0.el9.x86_64 @Ceph Install librbd1-2:19.2.0-0.el9.x86_64 @Ceph Install librgw2-2:19.2.0-0.el9.x86_64 @Ceph Install python3-ceph-argparse-2:19.2.0-0.el9.x86_64 @Ceph Install python3-ceph-common-2:19.2.0-0.el9.x86_64 @Ceph Install python3-cephfs-2:19.2.0-0.el9.x86_64 @Ceph Install python3-rados-2:19.2.0-0.el9.x86_64 @Ceph Install python3-rbd-2:19.2.0-0.el9.x86_64 @Ceph Install python3-rgw-2:19.2.0-0.el9.x86_64 @Ceph Install cephadm-2:19.2.0-0.el9.noarch @Ceph-noarch Install gperftools-libs-2.9.1-3.el9.x86_64 @epel Install libarrow-9.0.0-13.el9.x86_64 @epel Install libarrow-doc-9.0.0-13.el9.noarch @epel Install liboath-2.6.12-1.el9.x86_64 @epel Install libunwind-1.6.2-1.el9.x86_64 @epel Install parquet-libs-9.0.0-13.el9.x86_64 @epel Install re2-1:20211101-20.el9.x86_64 @epel Install thrift-0.15.0-4.el9.x86_64 @epel Scriptlet output: 1 Detected autofs mount point /home during canonicalization of /home. 2 Skipping /home I dont have that output but here is the bash history:1 [2025-01-14-19:21:39] dnf install -y https://download.ceph.com/rpm-19.2.0/el9/noarch/ceph-release-1-1.el9.noarch.rpm2 [2025-01-14-19:21:41] dnf install -y ceph-common cephadm 3 [2025-01-14-19:22:08] host 172.31.78.114 [2025-01-14-19:22:22] cephadm bootstrap --mon-ip 172.31.78.11 --cluster-network 172.31.14.224/27 --allow-fqdn-hostname | tee initial_cephadm_out 5 [2025-01-14-19:23:13] cp /etc/ceph/ceph.pub /root/.ssh/authorized_keys26 [2025-01-14-19:23:13] chmod 600 /root/.ssh/authorized_keys2 7 [2025-01-14-19:23:26] ceph telemetry on --license sharing-1-0 8 [2025-01-14-19:23:26] ceph telemetry enable channel perf9 [2025-01-14-19:24:13] nmcli connection add type vlan con-name backend ifname backend vlan.id 2987 vlan.parent bond0 ipv4.method manual ipv4.addresses 172.31.14.231/27 ipv4.may-fail no ipv6.method disabled ipv6.may-fail yes connection.zone trusted10 [2025-01-14-19:27:18] ceph mgr module disable cephadm 11 [2025-01-14-19:27:26] ceph fsid12 [2025-01-14-19:27:36] cephadm rm-cluster --force --zap-osds --fsid 7f8377c2-d2a4-11ef-adc1-1c34da408e52 13 [2025-01-14-19:28:07] cephadm bootstrap --mon-ip 172.31.78.11 --cluster-network 172.31.14.224/27 --allow-fqdn-hostname | tee initial_cephadm_out 14 [2025-01-14-19:28:46] cp /etc/ceph/ceph.pub /root/.ssh/authorized_keys215 [2025-01-14-19:28:49] chmod 600 /root/.ssh/authorized_keys2 16 [2025-01-14-19:29:05] ceph telemetry on --license sharing-1-0 17 [2025-01-14-19:29:06] ceph telemetry enable channel perf18 [2025-01-14-19:29:17] scp /root/.ssh/authorized_keys2 ceph02.inf.ethz.ch:/root/.ssh/authorized_keys2 19 [2025-01-14-19:29:19] scp /root/.ssh/authorized_keys2 ceph03.inf.ethz.ch:/root/.ssh/authorized_keys2 20 [2025-01-14-19:29:21] scp /root/.ssh/authorized_keys2 ceph04.inf.ethz.ch:/root/.ssh/authorized_keys2 21 [2025-01-14-19:29:22] scp /root/.ssh/authorized_keys2 ceph05.inf.ethz.ch:/root/.ssh/authorized_keys222 [2025-01-14-19:29:24] yes 23 [2025-01-14-19:30:15] ceph config set osd osd_memory_target 8G 24 [2025-01-14-19:30:22] ceph config set mds mds_cache_memory_limit 64G 25 [2025-01-14-19:30:30] ceph orch apply osd --all-available-devices26 [2025-01-14-19:32:27] ceph orch host add ceph02.inf.ethz.ch 172.31.78.12 --labels _admin,nfs,mgr,mds,rgw,rbd,mon,osd,grafana,iscsi 27 [2025-01-14-19:55:31] ceph orch host add ceph03.inf.ethz.ch 172.31.78.13 --labels _admin,nfs,mgr,mds,rgw,rbd,mon,osd,grafana,iscsi 28 [2025-01-14-20:11:57] ceph orch host add ceph04.inf.ethz.ch 172.31.78.14 --labels _admin,nfs,mgr,mds,rgw,rbd,mon,osd,grafana,iscsi 29 [2025-01-14-20:23:04] ceph orch host add ceph05.inf.ethz.ch 172.31.78.15 --labels _admin,nfs,mgr,mds,rgw,rbd,mon,osd,grafana,iscsi 30 [2025-01-14-20:23:27] #ceph orch apply osd --all-available-devices --unmanaged=true 31 [2025-01-14-20:47:24] ceph orch apply osd --all-available-devices --unmanaged=true32 [2025-01-14-21:24:41] ls 33 [2025-01-14-21:24:47] cat initial_cephadm_out On 3/10/25 09:30, Eugen Block wrote:(Don't drop the list in your responses)There's a lot going on in the cephadm.log, it's not clear which of these are relevant to the issue. This one looks bad though:ValueError: "'xwork.MethodAccessor.denyMethodExecution'" does not appear to be an IPv4 or IPv6 addressDid you build this cluster with 19.2.1 or did you upgrade? Do you have the output from 'ceph orch host ls' lying around (I know, currently it's not working)? Are you using IPv6? I haven't seen this issue yet.Zitat von Moritz Baumann <mo@xxxxxxxxxxxxx>:Am 09.03.2025 um 15:22 schrieb Moritz Baumann:I dont want to sound stupid, but how do I get the log? Am 09.03.2025 um 11:11 schrieb Eugen Block:ceph versionsceph01[22]:~# cephadm shell -- ceph log last cephadm Inferring fsid 4cfb1138-d2a5-11ef-87eb-1c34da408e52Inferring config /var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/mon.ceph01/config Not using image 'f2efb0401a30ec7eda97b6da76b314bd081fcb910cc5dcd826bc7c72c9dfdd7d' as it's not in list of non-dangling images with ceph=True label 2025-02-25T08:42:54.590667+0000 mgr.ceph01.ombogc (mgr.180813) 1706 : cephadm [INF] maintenance mode request for ceph05.inf.ethz.ch has SET the noout group 2025-02-25T08:43:38.380838+0000 mgr.ceph01.ombogc (mgr.180813) 1729 : cephadm [ERR] Detected invalid grafana certificate on host ceph05.inf.ethz.ch: Invalid certificate key: [('PEM routines', '', 'no start line')] 2025-02-25T08:44:40.343420+0000 mgr.ceph01.ombogc (mgr.180813) 1762 : cephadm [INF] Adjusting osd_memory_target on ceph04.inf.ethz.ch to 33423M 2025-02-25T08:44:40.860933+0000 mgr.ceph01.ombogc (mgr.180813) 1764 : cephadm [ERR] Detected invalid grafana certificate on host ceph05.inf.ethz.ch: Invalid certificate key: [('PEM routines', '', 'no start line')] 2025-02-25T08:45:41.993564+0000 mgr.ceph01.ombogc (mgr.180813) 1797 : cephadm [ERR] Detected invalid grafana certificate on host ceph05.inf.ethz.ch: Invalid certificate key: [('PEM routines', '', 'no start line')] 2025-02-25T08:46:42.758717+0000 mgr.ceph01.ombogc (mgr.180813) 1830 : cephadm [ERR] Detected invalid grafana certificate on host ceph05.inf.ethz.ch: Invalid certificate key: [('PEM routines', '', 'no start line')] 2025-02-25T08:47:44.190946+0000 mgr.ceph01.ombogc (mgr.180813) 1864 : cephadm [ERR] Unable to reach remote host ceph05.inf.ethz.ch. SSH connection closedTraceback (most recent call last):File "/usr/share/ceph/mgr/cephadm/ssh.py", line 245, in _execute_commandr = await conn.run(str(rcmd), input=stdin)File "/lib/python3.9/site-packages/asyncssh/connection.py", line 4250, in runprocess = await self.create_process(*args, **kwargs) # type: ignoreFile "/lib/python3.9/site-packages/asyncssh/connection.py", line 4128, in create_processchan, process = await self.create_session(File "/lib/python3.9/site-packages/asyncssh/connection.py", line 4018, in create_sessionchan = SSHClientChannel(self, self._loop, encoding, errors,File "/lib/python3.9/site-packages/asyncssh/channel.py", line 1085, in __init__super().__init__(conn, loop, encoding, errors, window, max_pktsize)File "/lib/python3.9/site-packages/asyncssh/channel.py", line 140, in __init__self._recv_chan: Optional[int] = conn.add_channel(self)File "/lib/python3.9/site-packages/asyncssh/connection.py", line 1307, in add_channelraise ChannelOpenError(OPEN_CONNECT_FAILED, asyncssh.misc.ChannelOpenError: SSH connection closed During handling of the above exception, another exception occurred: Traceback (most recent call last):File "/usr/share/ceph/mgr/orchestrator/_interface.py", line 138, in wrapperreturn OrchResult(f(*args, **kwargs)) File "/usr/share/ceph/mgr/cephadm/module.py", line 162, in wrapper return func(*args, **kwargs)File "/usr/share/ceph/mgr/cephadm/module.py", line 2224, in exit_host_maintenanceouts, errs, _code = self.wait_async( File "/usr/share/ceph/mgr/cephadm/module.py", line 815, in wait_async return self.event_loop.get_result(coro, timeout) File "/usr/share/ceph/mgr/cephadm/ssh.py", line 136, in get_result return future.result(timeout)File "/lib64/python3.9/concurrent/futures/_base.py", line 446, in resultreturn self.__get_result()File "/lib64/python3.9/concurrent/futures/_base.py", line 391, in __get_resultraise self._exception File "/usr/share/ceph/mgr/cephadm/serve.py", line 1669, in _run_cephadmpython = await self.mgr.ssh._check_execute_command(host, cmd, addr=addr) File "/usr/share/ceph/mgr/cephadm/ssh.py", line 300, in _check_execute_command out, err, code = await self._execute_command(host, cmd, stdin, addr, log_command) File "/usr/share/ceph/mgr/cephadm/ssh.py", line 253, in _execute_command raise HostConnectionError(f'Unable to reach remote host {host}. {str(e)}', host, address) cephadm.ssh.HostConnectionError: Unable to reach remote host ceph05.inf.ethz.ch. SSH connection closed 2025-02-25T08:47:44.762770+0000 mgr.ceph01.ombogc (mgr.180813) 1865 : cephadm [ERR] Detected invalid grafana certificate on host ceph05.inf.ethz.ch: Invalid certificate key: [('PEM routines', '', 'no start line')] 2025-02-25T08:47:55.631018+0000 mgr.ceph01.ombogc (mgr.180813) 1874 : cephadm [INF] exit maintenance request has UNSET for the noout group on host ceph05.inf.ethz.ch 2025-02-25T08:48:47.623644+0000 mgr.ceph01.ombogc (mgr.180813) 1901 : cephadm [INF] Adjusting osd_memory_target on ceph05.inf.ethz.ch to 32911M 2025-02-25T08:48:47.827026+0000 mgr.ceph01.ombogc (mgr.180813) 1903 : cephadm [ERR] Detected invalid grafana certificate on host ceph05.inf.ethz.ch: Invalid certificate key: [('PEM routines', '', 'no start line')] 2025-02-25T08:49:21.108932+0000 mgr.ceph01.ombogc (mgr.180813) 1925 : cephadm [INF] Schedule redeploy daemon grafana.ceph05 2025-02-25T08:49:21.292879+0000 mgr.ceph01.ombogc (mgr.180813) 1926 : cephadm [INF] Regenerating cephadm self-signed grafana TLS certificates 2025-02-25T08:49:21.328348+0000 mgr.ceph01.ombogc (mgr.180813) 1929 : cephadm [INF] Deploying daemon grafana.ceph05 on ceph05.inf.ethz.ch 2025-02-25T09:07:00.824841+0000 mgr.ceph01.ombogc (mgr.180813) 2461 : cephadm [INF] Detected new or changed devices on ceph05.inf.ethz.ch 2025-02-25T09:07:01.376245+0000 mgr.ceph01.ombogc (mgr.180813) 2462 : cephadm [INF] Detected new or changed devices on ceph04.inf.ethz.ch 2025-02-28T06:18:29.439683+0000 mgr.ceph01.ombogc (mgr.180813) 137418 : cephadm [ERR] [28/Feb/2025:06:18:29] ENGINE ValueError('"\'xwork.MethodAccessor.denyMethodExecution\'" does not appear to be an IPv4 or IPv6 address')Traceback (most recent call last):File "/lib/python3.9/site-packages/cheroot/server.py", line 1281, in communicatereq.parse_request()File "/lib/python3.9/site-packages/cheroot/server.py", line 714, in parse_requestsuccess = self.read_request_line()File "/lib/python3.9/site-packages/cheroot/server.py", line 815, in read_request_linescheme, authority, path, qs, fragment = urllib.parse.urlsplit(uri) File "/lib64/python3.9/urllib/parse.py", line 510, in urlsplit _check_bracketed_host(bracketed_host)File "/lib64/python3.9/urllib/parse.py", line 453, in _check_bracketed_host ip = ipaddress.ip_address(hostname) # Throws Value Error if not IPv6 or IPv4File "/lib64/python3.9/ipaddress.py", line 53, in ip_addressraise ValueError(f'{address!r} does not appear to be an IPv4 or IPv6 address') ValueError: "'xwork.MethodAccessor.denyMethodExecution'" does not appear to be an IPv4 or IPv6 address2025-02-28T06:18:29.441452+0000 mgr.ceph01.ombogc (mgr.180813) 137419 : cephadm [ERR] [28/Feb/2025:06:18:29] ENGINE ValueError('"\'xwork.MethodAccessor.denyMethodExecution\'" does not appear to be an IPv4 or IPv6 address')Traceback (most recent call last):File "/lib/python3.9/site-packages/cheroot/server.py", line 1281, in communicatereq.parse_request()File "/lib/python3.9/site-packages/cheroot/server.py", line 714, in parse_requestsuccess = self.read_request_line()File "/lib/python3.9/site-packages/cheroot/server.py", line 815, in read_request_linescheme, authority, path, qs, fragment = urllib.parse.urlsplit(uri) File "/lib64/python3.9/urllib/parse.py", line 510, in urlsplit _check_bracketed_host(bracketed_host)File "/lib64/python3.9/urllib/parse.py", line 453, in _check_bracketed_host ip = ipaddress.ip_address(hostname) # Throws Value Error if not IPv6 or IPv4File "/lib64/python3.9/ipaddress.py", line 53, in ip_addressraise ValueError(f'{address!r} does not appear to be an IPv4 or IPv6 address') ValueError: "'xwork.MethodAccessor.denyMethodExecution'" does not appear to be an IPv4 or IPv6 address2025-03-07T06:17:28.282574+0000 mgr.ceph01.ombogc (mgr.180813) 447472 : cephadm [ERR] [07/Mar/2025:06:17:28] ENGINE ValueError('"\'xwork.MethodAccessor.denyMethodExecution\'" does not appear to be an IPv4 or IPv6 address')Traceback (most recent call last):File "/lib/python3.9/site-packages/cheroot/server.py", line 1281, in communicatereq.parse_request()File "/lib/python3.9/site-packages/cheroot/server.py", line 714, in parse_requestsuccess = self.read_request_line()File "/lib/python3.9/site-packages/cheroot/server.py", line 815, in read_request_linescheme, authority, path, qs, fragment = urllib.parse.urlsplit(uri) File "/lib64/python3.9/urllib/parse.py", line 510, in urlsplit _check_bracketed_host(bracketed_host)File "/lib64/python3.9/urllib/parse.py", line 453, in _check_bracketed_host ip = ipaddress.ip_address(hostname) # Throws Value Error if not IPv6 or IPv4File "/lib64/python3.9/ipaddress.py", line 53, in ip_addressraise ValueError(f'{address!r} does not appear to be an IPv4 or IPv6 address') ValueError: "'xwork.MethodAccessor.denyMethodExecution'" does not appear to be an IPv4 or IPv6 address2025-03-09T07:36:00.522250+0000 mgr.ceph01.ombogc (mgr.180813) 536135 : cephadm [ERR] ALERT: Cannot stop active Mgr daemon, Please switch active Mgrs with 'ceph mgr fail ceph01.ombogc'Note: Warnings can be bypassed with the --force flag Traceback (most recent call last):File "/usr/share/ceph/mgr/orchestrator/_interface.py", line 138, in wrapperreturn OrchResult(f(*args, **kwargs)) File "/usr/share/ceph/mgr/cephadm/module.py", line 162, in wrapper return func(*args, **kwargs)File "/usr/share/ceph/mgr/cephadm/module.py", line 2165, in enter_host_maintenanceraise OrchestratorError(orchestrator._interface.OrchestratorError: ALERT: Cannot stop active Mgr daemon, Please switch active Mgrs with 'ceph mgr fail ceph01.ombogc'Note: Warnings can be bypassed with the --force flag
_______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx