Regarding the
ValueError: "'xwork.MethodAccessor.denyMethodExecution'" does not
appear to be an IPv4 or IPv6 address
can you check `ceph config-key get mgr/cephadm/inventory` and see
if you see something related to that (such as
"'xwork.MethodAccessor.denyMethodExecution'" being present as the
addr for a host). If nothing there, there should also be a number
of `mgr/ cephadm/host.<hostname>` entries, one of which may have
something related as well. Cephadm stores all its state in
mgr/cephadm/... entries in the config-key store, so if it keeps
failing after mgr failovers there's likely something in there that
shouldn't be, and your only recourse at this point might be to
manually edit one of those entries before failing over the mgr
again. The output of `ceph health detail` could also potentially
have something to point us in the right direction if the module is
fully crashing (sorry if I'm requesting something you've already
posted).
On Wed, Mar 12, 2025 at 2:26 PM Eugen Block <eblock@xxxxxx
<mailto:eblock@xxxxxx>> wrote:
I'm still not sure yet if we were looking at the right logs. Could you
revert the changes to the unit.run file, and then fail the mgr to get
a new set of logs. Then on the node with the active mgr run 'cephadm
logs --name mgr.<active_mgr>' and then paste the relevant output in a
pastebin or similar. Redact sensitive information. Alternatively, you
can get them with 'journalctl -u
ceph-<cluster_fsid>@mgr.<active_mgr>'. I would also like to understand
what's going on rather nuking the cluster and hope that next time it
works (in production).
Zitat von Moritz Baumann <mo@xxxxxxxxxxxxx <mailto:mo@xxxxxxxxxxxxx>>:
> The cluster was successfully upgraded to 19.2.1 and has been running
> some time with that version before it broke.
>
> As you suggested I have changed the unit.run file and replaced the
> sha256:XYZ.... with the one from
> https://quay.io/repository/ceph/ceph?tab=tags <https://quay.io/
repository/ceph/ceph?tab=tags>
>
> (
>
>
sha256:200087c35811bf28e8a8073b15fa86c07cce85c575f1ccd62d1d6ddbfdc6770a
>
> )
>
>
> ceph01[1]:/var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/
mgr.ceph01.ombogc# diff unit.run*
> -Nur
> --- unit.run 2025-03-12 13:33:43.455465357 +0100
> +++ unit.run.bak 2025-03-12 13:32:44.713805212 +0100
> @@ -5,4 +5,4 @@
> ! /usr/bin/podman rm -f
> ceph-4cfb1138-d2a5-11ef-87eb-1c34da408e52-mgr-ceph01-ombogc 2>
> /dev/null
> ! /usr/bin/podman rm -f --storage
> ceph-4cfb1138-d2a5-11ef-87eb-1c34da408e52-mgr-ceph01-ombogc 2>
> /dev/null
> ! /usr/bin/podman rm -f --storage
> ceph-4cfb1138-d2a5-11ef-87eb-1c34da408e52-mgr.ceph01.ombogc 2>
> /dev/null
> -/usr/bin/podman run --rm --ipc=host --stop-signal=SIGTERM
> --net=host --entrypoint /usr/bin/ceph-mgr --init --name
> ceph-4cfb1138-d2a5-11ef-87eb-1c34da408e52-mgr-ceph01-ombogc
> --pids-limit=-1 -d --log-driver journald --conmon-pidfile
> /run/ceph-4cfb1138-
d2a5-11ef-87eb-1c34da408e52@mgr.ceph01.ombogc.service-pid
> --cidfile
> /run/ceph-4cfb1138-
d2a5-11ef-87eb-1c34da408e52@mgr.ceph01.ombogc.service-cid
> --cgroups=split -e
> CONTAINER_IMAGE=quay.io/ceph/
ceph@sha256:200087c35811bf28e8a8073b15fa86c07cce85c575f1ccd62d1d6ddbfdc6770a <http://quay.io/ceph/ceph@sha256:200087c35811bf28e8a8073b15fa86c07cce85c575f1ccd62d1d6ddbfdc6770a> -e NODE_NAME=ceph01.inf.ethz.ch <http://ceph01.inf.ethz.ch> -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52:/var/run/ceph:z -v /var/log/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52:/var/log/ceph:z -v /var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/crash:/var/lib/ceph/crash:z -v /run/systemd/journal:/run/systemd/journal -v /var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/mgr.ceph01.ombogc:/var/lib/ceph/mgr/ceph-ceph01.ombogc:z -v /var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/mgr.ceph01.ombogc/config:/etc/ceph/ceph.conf:z -v /etc/hosts:/etc/hosts:ro quay.io/ceph/ceph@sha256:200087c35811bf28e8a8073b15fa86c07cce85c575f1ccd62d1d6ddbfdc6770a <http://quay.io/ceph/ceph@sha256:200087c35811bf28e8a8073b15fa86c07cce85c575f1ccd62d1d6ddbfdc6770a> -n mgr.ceph01.ombogc -f --setuser ceph --setgroup ceph --default-log-to-file=false
--default-log-to-journald=true
> --default-log-to-stderr=false
> +/usr/bin/podman run --rm --ipc=host --stop-signal=SIGTERM
> --net=host --entrypoint /usr/bin/ceph-mgr --init --name
> ceph-4cfb1138-d2a5-11ef-87eb-1c34da408e52-mgr-ceph01-ombogc
> --pids-limit=-1 -d --log-driver journald --conmon-pidfile
> /run/ceph-4cfb1138-
d2a5-11ef-87eb-1c34da408e52@mgr.ceph01.ombogc.service-pid
> --cidfile
> /run/ceph-4cfb1138-
d2a5-11ef-87eb-1c34da408e52@mgr.ceph01.ombogc.service-cid
> --cgroups=split -e
> CONTAINER_IMAGE=quay.io/ceph/
ceph@sha256:41d3f5e46ff7de28544cc8869fdea13fca824dcef83936cb3288ed9de935e4de <http://quay.io/ceph/ceph@sha256:41d3f5e46ff7de28544cc8869fdea13fca824dcef83936cb3288ed9de935e4de> -e NODE_NAME=ceph01.inf.ethz.ch <http://ceph01.inf.ethz.ch> -e TCMALLOC_MAX_TOTAL_THREAD_CACHE_BYTES=134217728 -v /var/run/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52:/var/run/ceph:z -v /var/log/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52:/var/log/ceph:z -v /var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/crash:/var/lib/ceph/crash:z -v /run/systemd/journal:/run/systemd/journal -v /var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/mgr.ceph01.ombogc:/var/lib/ceph/mgr/ceph-ceph01.ombogc:z -v /var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/mgr.ceph01.ombogc/config:/etc/ceph/ceph.conf:z -v /etc/hosts:/etc/hosts:ro quay.io/ceph/ceph@sha256:41d3f5e46ff7de28544cc8869fdea13fca824dcef83936cb3288ed9de935e4de <http://quay.io/ceph/ceph@sha256:41d3f5e46ff7de28544cc8869fdea13fca824dcef83936cb3288ed9de935e4de> -n mgr.ceph01.ombogc -f --setuser ceph --setgroup ceph --default-log-to-file=false
--default-log-to-journald=true
> --default-log-to-stderr=false
>
>
> stopped the 19.2.1 version
>
> # cephadm unit stop --name mgr.ceph01.ombogc
> # cephadm unit start --name mgr.ceph01.ombogc
>
> and failed to the other mgrs
>
> # ceph mgr fail ceph01.ombogc
>
> # ceph orch host ls
>
> Error ENOENT: Module not found
>
> # ceph mgr fail ceph02.nqlpce
>
> # ceph orch host ls
>
> Error ENOENT: Module not found
>
>
>
> # podman exec -ti <id of the mgr> /bin/bash
>
> in the container
>
> ceph --version
>
>
> on both nodes shows that one is running 19.2.1 and the other one is
> running 19.2.0
>
>
> but no matter which mgr is active the ceph orch commands do not work.
>
>
> we were only in the process of making this cluster productive, so if
> it comes to the wort I can copy the little data on it somewhere and
> reinstall it, but I hoped to find a solution (to feel better that
> when this should happen again there is a way to recover).
>
>
>
>
>
>
> On 3/10/25 22:55, Eugen Block wrote:
>> I don’t have a better idea than to downgrade one MGR by editing the
>> unit.run file to use a different version (19.2.0 seems to have
>> worked, right?), restart it, then fail the MGRs until the older one
>> takes over (or stop all newer ones). And then check if the
>> orchestrator still misbehaves.
>>
>>
>> Zitat von Moritz Baumann <mo@xxxxxxxxxxxxx
<mailto:mo@xxxxxxxxxxxxx>>:
>>
>>> It was installed with 19.2.0:
>>>
>>> dnf history info 20 shows:
>>>
>>> ceph01[0]:~# dnf history info 20
>>> Transaction ID : 20
>>> Begin time : Tue 14 Jan 2025 07:22:00 PM CET
>>> Begin rpmdb :
>>> 0fac0c11ba074f9635117d0b3d4970e6d26e8d2b40bfe8b93d22fd12147260b7
>>> End time : Tue 14 Jan 2025 07:22:04 PM CET (4 seconds)
>>> End rpmdb :
>>> 63657f6d4b27790107998bd21da76fe182d9c121177b34cf879243884deba40a
>>> User : root <root>
>>> Return-Code : Success
>>> Releasever : 9
>>> Command Line : install -y ceph-common cephadm
>>> Comment :
>>> Packages Altered:
>>> Install boost-program-options-1.75.0-8.el9.x86_64 @appstream
>>> Install libbabeltrace-1.5.8-10.el9.x86_64 @appstream
>>> Install libpmem-1.12.1-1.el9.x86_64 @appstream
>>> Install libpmemobj-1.12.1-1.el9.x86_64 @appstream
>>> Install librabbitmq-0.11.0-7.el9.x86_64 @appstream
>>> Install librdkafka-1.6.1-102.el9.x86_64 @appstream
>>> Install lttng-ust-2.12.0-6.el9.x86_64 @appstream
>>> Install python3-prettytable-0.7.2-27.el9.noarch @appstream
>>> Install daxctl-libs-78-2.el9.x86_64 @baseos
>>> Install librdmacm-51.0-1.el9.x86_64 @baseos
>>> Install ndctl-libs-78-2.el9.x86_64 @baseos
>>> Install ceph-common-2:19.2.0-0.el9.x86_64 @Ceph
>>> Install libcephfs2-2:19.2.0-0.el9.x86_64 @Ceph
>>> Install librados2-2:19.2.0-0.el9.x86_64 @Ceph
>>> Install libradosstriper1-2:19.2.0-0.el9.x86_64 @Ceph
>>> Install librbd1-2:19.2.0-0.el9.x86_64 @Ceph
>>> Install librgw2-2:19.2.0-0.el9.x86_64 @Ceph
>>> Install python3-ceph-argparse-2:19.2.0-0.el9.x86_64 @Ceph
>>> Install python3-ceph-common-2:19.2.0-0.el9.x86_64 @Ceph
>>> Install python3-cephfs-2:19.2.0-0.el9.x86_64 @Ceph
>>> Install python3-rados-2:19.2.0-0.el9.x86_64 @Ceph
>>> Install python3-rbd-2:19.2.0-0.el9.x86_64 @Ceph
>>> Install python3-rgw-2:19.2.0-0.el9.x86_64 @Ceph
>>> Install cephadm-2:19.2.0-0.el9.noarch @Ceph-noarch
>>> Install gperftools-libs-2.9.1-3.el9.x86_64 @epel
>>> Install libarrow-9.0.0-13.el9.x86_64 @epel
>>> Install libarrow-doc-9.0.0-13.el9.noarch @epel
>>> Install liboath-2.6.12-1.el9.x86_64 @epel
>>> Install libunwind-1.6.2-1.el9.x86_64 @epel
>>> Install parquet-libs-9.0.0-13.el9.x86_64 @epel
>>> Install re2-1:20211101-20.el9.x86_64 @epel
>>> Install thrift-0.15.0-4.el9.x86_64 @epel
>>> Scriptlet output:
>>> 1 Detected autofs mount point /home during canonicalization
of /home.
>>> 2 Skipping /home
>>>
>>> I dont have that output but here is the bash history:
>>>
>>> 1 [2025-01-14-19:21:39] dnf install -y
>>> https://download.ceph.com/rpm-19.2.0/el9/noarch/ceph-
release-1-1.el9.noarch.rpm <https://download.ceph.com/rpm-19.2.0/
el9/noarch/ceph-release-1-1.el9.noarch.rpm>
>>> 2 [2025-01-14-19:21:41] dnf install -y ceph-common cephadm
>>> 3 [2025-01-14-19:22:08] host 172.31.78.11
>>> 4 [2025-01-14-19:22:22] cephadm bootstrap --mon-ip
>>> 172.31.78.11 --cluster-network 172.31.14.224/27
<http://172.31.14.224/27>
>>> --allow-fqdn-hostname | tee initial_cephadm_out
>>> 5 [2025-01-14-19:23:13] cp /etc/ceph/ceph.pub
>>> /root/.ssh/authorized_keys2
>>> 6 [2025-01-14-19:23:13] chmod 600 /root/.ssh/authorized_keys2
>>> 7 [2025-01-14-19:23:26] ceph telemetry on --license
sharing-1-0
>>> 8 [2025-01-14-19:23:26] ceph telemetry enable channel perf
>>> 9 [2025-01-14-19:24:13] nmcli connection add type vlan
>>> con-name backend ifname backend vlan.id <http://vlan.id> 2987
vlan.parent bond0
>>> ipv4.method manual ipv4.addresses 172.31.14.231/27
<http://172.31.14.231/27> ipv4.may-fail
>>> no ipv6.method disabled ipv6.may-fail yes connection.zone trusted
>>> 10 [2025-01-14-19:27:18] ceph mgr module disable cephadm
>>> 11 [2025-01-14-19:27:26] ceph fsid
>>> 12 [2025-01-14-19:27:36] cephadm rm-cluster --force --zap-osds
>>> --fsid 7f8377c2-d2a4-11ef-adc1-1c34da408e52
>>> 13 [2025-01-14-19:28:07] cephadm bootstrap --mon-ip
>>> 172.31.78.11 --cluster-network 172.31.14.224/27
<http://172.31.14.224/27>
>>> --allow-fqdn-hostname | tee initial_cephadm_out
>>> 14 [2025-01-14-19:28:46] cp /etc/ceph/ceph.pub
>>> /root/.ssh/authorized_keys2
>>> 15 [2025-01-14-19:28:49] chmod 600 /root/.ssh/authorized_keys2
>>> 16 [2025-01-14-19:29:05] ceph telemetry on --license
sharing-1-0
>>> 17 [2025-01-14-19:29:06] ceph telemetry enable channel perf
>>> 18 [2025-01-14-19:29:17] scp /root/.ssh/authorized_keys2
>>> ceph02.inf.ethz.ch:/root/.ssh/authorized_keys2
>>> 19 [2025-01-14-19:29:19] scp /root/.ssh/authorized_keys2
>>> ceph03.inf.ethz.ch:/root/.ssh/authorized_keys2
>>> 20 [2025-01-14-19:29:21] scp /root/.ssh/authorized_keys2
>>> ceph04.inf.ethz.ch:/root/.ssh/authorized_keys2
>>> 21 [2025-01-14-19:29:22] scp /root/.ssh/authorized_keys2
>>> ceph05.inf.ethz.ch:/root/.ssh/authorized_keys2
>>> 22 [2025-01-14-19:29:24] yes
>>> 23 [2025-01-14-19:30:15] ceph config set osd
osd_memory_target 8G
>>> 24 [2025-01-14-19:30:22] ceph config set mds
mds_cache_memory_limit 64G
>>> 25 [2025-01-14-19:30:30] ceph orch apply osd --all-
available-devices
>>> 26 [2025-01-14-19:32:27] ceph orch host add
ceph02.inf.ethz.ch <http://ceph02.inf.ethz.ch>
>>> 172.31.78.12 --labels
>>> _admin,nfs,mgr,mds,rgw,rbd,mon,osd,grafana,iscsi
>>> 27 [2025-01-14-19:55:31] ceph orch host add
ceph03.inf.ethz.ch <http://ceph03.inf.ethz.ch>
>>> 172.31.78.13 --labels
>>> _admin,nfs,mgr,mds,rgw,rbd,mon,osd,grafana,iscsi
>>> 28 [2025-01-14-20:11:57] ceph orch host add
ceph04.inf.ethz.ch <http://ceph04.inf.ethz.ch>
>>> 172.31.78.14 --labels
>>> _admin,nfs,mgr,mds,rgw,rbd,mon,osd,grafana,iscsi
>>> 29 [2025-01-14-20:23:04] ceph orch host add
ceph05.inf.ethz.ch <http://ceph05.inf.ethz.ch>
>>> 172.31.78.15 --labels
>>> _admin,nfs,mgr,mds,rgw,rbd,mon,osd,grafana,iscsi
>>> 30 [2025-01-14-20:23:27] #ceph orch apply osd
>>> --all-available-devices --unmanaged=true
>>> 31 [2025-01-14-20:47:24] ceph orch apply osd
>>> --all-available-devices --unmanaged=true
>>> 32 [2025-01-14-21:24:41] ls
>>> 33 [2025-01-14-21:24:47] cat initial_cephadm_out
>>>
>>>
>>>
>>>
>>>
>>> On 3/10/25 09:30, Eugen Block wrote:
>>>> (Don't drop the list in your responses)
>>>> There's a lot going on in the cephadm.log, it's not clear which
>>>> of these are relevant to the issue. This one looks bad though:
>>>>
>>>> ValueError: "'xwork.MethodAccessor.denyMethodExecution'" does not
>>>> appear to be an IPv4 or IPv6 address
>>>>
>>>> Did you build this cluster with 19.2.1 or did you upgrade? Do you
>>>> have the output from 'ceph orch host ls' lying around (I know,
>>>> currently it's not working)? Are you using IPv6? I haven't seen
>>>> this issue yet.
>>>>
>>>> Zitat von Moritz Baumann <mo@xxxxxxxxxxxxx
<mailto:mo@xxxxxxxxxxxxx>>:
>>>>
>>>>> Am 09.03.2025 um 15:22 schrieb Moritz Baumann:
>>>>>> I dont want to sound stupid, but how do I get the log?
>>>>>>
>>>>>> Am 09.03.2025 um 11:11 schrieb Eugen Block:
>>>>>>> ceph versions
>>>>>>
>>>>>
>>>>>
>>>>> ceph01[22]:~# cephadm shell -- ceph log last cephadm
>>>>> Inferring fsid 4cfb1138-d2a5-11ef-87eb-1c34da408e52
>>>>> Inferring config
>>>>> /var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/
mon.ceph01/config
>>>>> Not using image
>>>>>
'f2efb0401a30ec7eda97b6da76b314bd081fcb910cc5dcd826bc7c72c9dfdd7d' as
>>>>> it's not in list of non-dangling images with ceph=True label
>>>>> 2025-02-25T08:42:54.590667+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 1706 : cephadm [INF] maintenance mode request for
>>>>> ceph05.inf.ethz.ch <http://ceph05.inf.ethz.ch> has SET the
noout group
>>>>> 2025-02-25T08:43:38.380838+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 1729 : cephadm [ERR] Detected invalid grafana certificate on
>>>>> host ceph05.inf.ethz.ch <http://ceph05.inf.ethz.ch>: Invalid
certificate key: [('PEM
>>>>> routines', '', 'no start line')]
>>>>> 2025-02-25T08:44:40.343420+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 1762 : cephadm [INF] Adjusting osd_memory_target on
>>>>> ceph04.inf.ethz.ch <http://ceph04.inf.ethz.ch> to 33423M
>>>>> 2025-02-25T08:44:40.860933+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 1764 : cephadm [ERR] Detected invalid grafana certificate on
>>>>> host ceph05.inf.ethz.ch <http://ceph05.inf.ethz.ch>: Invalid
certificate key: [('PEM
>>>>> routines', '', 'no start line')]
>>>>> 2025-02-25T08:45:41.993564+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 1797 : cephadm [ERR] Detected invalid grafana certificate on
>>>>> host ceph05.inf.ethz.ch <http://ceph05.inf.ethz.ch>: Invalid
certificate key: [('PEM
>>>>> routines', '', 'no start line')]
>>>>> 2025-02-25T08:46:42.758717+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 1830 : cephadm [ERR] Detected invalid grafana certificate on
>>>>> host ceph05.inf.ethz.ch <http://ceph05.inf.ethz.ch>: Invalid
certificate key: [('PEM
>>>>> routines', '', 'no start line')]
>>>>> 2025-02-25T08:47:44.190946+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 1864 : cephadm [ERR] Unable to reach remote host
>>>>> ceph05.inf.ethz.ch <http://ceph05.inf.ethz.ch>. SSH
connection closed
>>>>> Traceback (most recent call last):
>>>>> File "/usr/share/ceph/mgr/cephadm/ssh.py", line 245, in
>>>>> _execute_command
>>>>> r = await conn.run(str(rcmd), input=stdin)
>>>>> File "/lib/python3.9/site-packages/asyncssh/connection.py",
>>>>> line 4250, in run
>>>>> process = await self.create_process(*args, **kwargs) #
type: ignore
>>>>> File "/lib/python3.9/site-packages/asyncssh/connection.py",
>>>>> line 4128, in create_process
>>>>> chan, process = await self.create_session(
>>>>> File "/lib/python3.9/site-packages/asyncssh/connection.py",
>>>>> line 4018, in create_session
>>>>> chan = SSHClientChannel(self, self._loop, encoding, errors,
>>>>> File "/lib/python3.9/site-packages/asyncssh/channel.py", line
>>>>> 1085, in __init__
>>>>> super().__init__(conn, loop, encoding, errors, window,
max_pktsize)
>>>>> File "/lib/python3.9/site-packages/asyncssh/channel.py", line
>>>>> 140, in __init__
>>>>> self._recv_chan: Optional[int] = conn.add_channel(self)
>>>>> File "/lib/python3.9/site-packages/asyncssh/connection.py",
>>>>> line 1307, in add_channel
>>>>> raise ChannelOpenError(OPEN_CONNECT_FAILED,
>>>>> asyncssh.misc.ChannelOpenError: SSH connection closed
>>>>>
>>>>> During handling of the above exception, another exception
occurred:
>>>>>
>>>>> Traceback (most recent call last):
>>>>> File "/usr/share/ceph/mgr/orchestrator/_interface.py", line
>>>>> 138, in wrapper
>>>>> return OrchResult(f(*args, **kwargs))
>>>>> File "/usr/share/ceph/mgr/cephadm/module.py", line 162, in
wrapper
>>>>> return func(*args, **kwargs)
>>>>> File "/usr/share/ceph/mgr/cephadm/module.py", line 2224, in
>>>>> exit_host_maintenance
>>>>> outs, errs, _code = self.wait_async(
>>>>> File "/usr/share/ceph/mgr/cephadm/module.py", line 815, in
wait_async
>>>>> return self.event_loop.get_result(coro, timeout)
>>>>> File "/usr/share/ceph/mgr/cephadm/ssh.py", line 136, in
get_result
>>>>> return future.result(timeout)
>>>>> File "/lib64/python3.9/concurrent/futures/_base.py", line 446,
>>>>> in result
>>>>> return self.__get_result()
>>>>> File "/lib64/python3.9/concurrent/futures/_base.py", line 391,
>>>>> in __get_result
>>>>> raise self._exception
>>>>> File "/usr/share/ceph/mgr/cephadm/serve.py", line 1669, in
_run_cephadm
>>>>> python = await self.mgr.ssh._check_execute_command(host,
>>>>> cmd, addr=addr)
>>>>> File "/usr/share/ceph/mgr/cephadm/ssh.py", line 300, in
>>>>> _check_execute_command
>>>>> out, err, code = await self._execute_command(host, cmd,
>>>>> stdin, addr, log_command)
>>>>> File "/usr/share/ceph/mgr/cephadm/ssh.py", line 253, in
>>>>> _execute_command
>>>>> raise HostConnectionError(f'Unable to reach remote host
>>>>> {host}. {str(e)}', host, address)
>>>>> cephadm.ssh.HostConnectionError: Unable to reach remote host
>>>>> ceph05.inf.ethz.ch <http://ceph05.inf.ethz.ch>. SSH
connection closed
>>>>> 2025-02-25T08:47:44.762770+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 1865 : cephadm [ERR] Detected invalid grafana certificate on
>>>>> host ceph05.inf.ethz.ch <http://ceph05.inf.ethz.ch>: Invalid
certificate key: [('PEM
>>>>> routines', '', 'no start line')]
>>>>> 2025-02-25T08:47:55.631018+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 1874 : cephadm [INF] exit maintenance request has UNSET for the
>>>>> noout group on host ceph05.inf.ethz.ch <http://
ceph05.inf.ethz.ch>
>>>>> 2025-02-25T08:48:47.623644+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 1901 : cephadm [INF] Adjusting osd_memory_target on
>>>>> ceph05.inf.ethz.ch <http://ceph05.inf.ethz.ch> to 32911M
>>>>> 2025-02-25T08:48:47.827026+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 1903 : cephadm [ERR] Detected invalid grafana certificate on
>>>>> host ceph05.inf.ethz.ch <http://ceph05.inf.ethz.ch>: Invalid
certificate key: [('PEM
>>>>> routines', '', 'no start line')]
>>>>> 2025-02-25T08:49:21.108932+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 1925 : cephadm [INF] Schedule redeploy daemon grafana.ceph05
>>>>> 2025-02-25T08:49:21.292879+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 1926 : cephadm [INF] Regenerating cephadm self-signed grafana
>>>>> TLS certificates
>>>>> 2025-02-25T08:49:21.328348+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 1929 : cephadm [INF] Deploying daemon grafana.ceph05 on
>>>>> ceph05.inf.ethz.ch <http://ceph05.inf.ethz.ch>
>>>>> 2025-02-25T09:07:00.824841+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 2461 : cephadm [INF] Detected new or changed devices on
>>>>> ceph05.inf.ethz.ch <http://ceph05.inf.ethz.ch>
>>>>> 2025-02-25T09:07:01.376245+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 2462 : cephadm [INF] Detected new or changed devices on
>>>>> ceph04.inf.ethz.ch <http://ceph04.inf.ethz.ch>
>>>>> 2025-02-28T06:18:29.439683+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 137418 : cephadm [ERR] [28/Feb/2025:06:18:29] ENGINE
>>>>> ValueError('"\'xwork.MethodAccessor.denyMethodExecution\'" does
>>>>> not appear to be an IPv4 or IPv6 address')
>>>>> Traceback (most recent call last):
>>>>> File "/lib/python3.9/site-packages/cheroot/server.py", line
>>>>> 1281, in communicate
>>>>> req.parse_request()
>>>>> File "/lib/python3.9/site-packages/cheroot/server.py", line
>>>>> 714, in parse_request
>>>>> success = self.read_request_line()
>>>>> File "/lib/python3.9/site-packages/cheroot/server.py", line
>>>>> 815, in read_request_line
>>>>> scheme, authority, path, qs, fragment =
urllib.parse.urlsplit(uri)
>>>>> File "/lib64/python3.9/urllib/parse.py", line 510, in urlsplit
>>>>> _check_bracketed_host(bracketed_host)
>>>>> File "/lib64/python3.9/urllib/parse.py", line 453, in
>>>>> _check_bracketed_host
>>>>> ip = ipaddress.ip_address(hostname) # Throws Value Error if
>>>>> not IPv6 or IPv4
>>>>> File "/lib64/python3.9/ipaddress.py", line 53, in ip_address
>>>>> raise ValueError(f'{address!r} does not appear to be an IPv4
>>>>> or IPv6 address')
>>>>> ValueError: "'xwork.MethodAccessor.denyMethodExecution'" does
>>>>> not appear to be an IPv4 or IPv6 address
>>>>>
>>>>> 2025-02-28T06:18:29.441452+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 137419 : cephadm [ERR] [28/Feb/2025:06:18:29] ENGINE
>>>>> ValueError('"\'xwork.MethodAccessor.denyMethodExecution\'" does
>>>>> not appear to be an IPv4 or IPv6 address')
>>>>> Traceback (most recent call last):
>>>>> File "/lib/python3.9/site-packages/cheroot/server.py", line
>>>>> 1281, in communicate
>>>>> req.parse_request()
>>>>> File "/lib/python3.9/site-packages/cheroot/server.py", line
>>>>> 714, in parse_request
>>>>> success = self.read_request_line()
>>>>> File "/lib/python3.9/site-packages/cheroot/server.py", line
>>>>> 815, in read_request_line
>>>>> scheme, authority, path, qs, fragment =
urllib.parse.urlsplit(uri)
>>>>> File "/lib64/python3.9/urllib/parse.py", line 510, in urlsplit
>>>>> _check_bracketed_host(bracketed_host)
>>>>> File "/lib64/python3.9/urllib/parse.py", line 453, in
>>>>> _check_bracketed_host
>>>>> ip = ipaddress.ip_address(hostname) # Throws Value Error if
>>>>> not IPv6 or IPv4
>>>>> File "/lib64/python3.9/ipaddress.py", line 53, in ip_address
>>>>> raise ValueError(f'{address!r} does not appear to be an IPv4
>>>>> or IPv6 address')
>>>>> ValueError: "'xwork.MethodAccessor.denyMethodExecution'" does
>>>>> not appear to be an IPv4 or IPv6 address
>>>>>
>>>>> 2025-03-07T06:17:28.282574+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 447472 : cephadm [ERR] [07/Mar/2025:06:17:28] ENGINE
>>>>> ValueError('"\'xwork.MethodAccessor.denyMethodExecution\'" does
>>>>> not appear to be an IPv4 or IPv6 address')
>>>>> Traceback (most recent call last):
>>>>> File "/lib/python3.9/site-packages/cheroot/server.py", line
>>>>> 1281, in communicate
>>>>> req.parse_request()
>>>>> File "/lib/python3.9/site-packages/cheroot/server.py", line
>>>>> 714, in parse_request
>>>>> success = self.read_request_line()
>>>>> File "/lib/python3.9/site-packages/cheroot/server.py", line
>>>>> 815, in read_request_line
>>>>> scheme, authority, path, qs, fragment =
urllib.parse.urlsplit(uri)
>>>>> File "/lib64/python3.9/urllib/parse.py", line 510, in urlsplit
>>>>> _check_bracketed_host(bracketed_host)
>>>>> File "/lib64/python3.9/urllib/parse.py", line 453, in
>>>>> _check_bracketed_host
>>>>> ip = ipaddress.ip_address(hostname) # Throws Value Error if
>>>>> not IPv6 or IPv4
>>>>> File "/lib64/python3.9/ipaddress.py", line 53, in ip_address
>>>>> raise ValueError(f'{address!r} does not appear to be an IPv4
>>>>> or IPv6 address')
>>>>> ValueError: "'xwork.MethodAccessor.denyMethodExecution'" does
>>>>> not appear to be an IPv4 or IPv6 address
>>>>>
>>>>> 2025-03-09T07:36:00.522250+0000 mgr.ceph01.ombogc (mgr.180813)
>>>>> 536135 : cephadm [ERR] ALERT: Cannot stop active Mgr daemon,
>>>>> Please switch active Mgrs with 'ceph mgr fail ceph01.ombogc'
>>>>> Note: Warnings can be bypassed with the --force flag
>>>>> Traceback (most recent call last):
>>>>> File "/usr/share/ceph/mgr/orchestrator/_interface.py", line
>>>>> 138, in wrapper
>>>>> return OrchResult(f(*args, **kwargs))
>>>>> File "/usr/share/ceph/mgr/cephadm/module.py", line 162, in
wrapper
>>>>> return func(*args, **kwargs)
>>>>> File "/usr/share/ceph/mgr/cephadm/module.py", line 2165, in
>>>>> enter_host_maintenance
>>>>> raise OrchestratorError(
>>>>> orchestrator._interface.OrchestratorError: ALERT: Cannot stop
>>>>> active Mgr daemon, Please switch active Mgrs with 'ceph mgr fail
>>>>> ceph01.ombogc'
>>>>> Note: Warnings can be bypassed with the --force flag
>>>>
>>>>
>>>>
>>
>>
>>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx <mailto:ceph-
users@xxxxxxx>
To unsubscribe send an email to ceph-users-leave@xxxxxxx
<mailto:ceph-users-leave@xxxxxxx>