Re: ceph orch command not working anymore on squid (19.2.1)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



It was installed with 19.2.0:

dnf history info 20 shows:

ceph01[0]:~# dnf history info 20
Transaction ID : 20
Begin time     : Tue 14 Jan 2025 07:22:00 PM CET
Begin rpmdb    : 0fac0c11ba074f9635117d0b3d4970e6d26e8d2b40bfe8b93d22fd12147260b7
End time       : Tue 14 Jan 2025 07:22:04 PM CET (4 seconds)
End rpmdb      : 63657f6d4b27790107998bd21da76fe182d9c121177b34cf879243884deba40a
User           : root <root>
Return-Code    : Success
Releasever     : 9
Command Line   : install -y ceph-common cephadm
Comment        :
Packages Altered:
    Install boost-program-options-1.75.0-8.el9.x86_64   @appstream
    Install libbabeltrace-1.5.8-10.el9.x86_64           @appstream
    Install libpmem-1.12.1-1.el9.x86_64                 @appstream
    Install libpmemobj-1.12.1-1.el9.x86_64              @appstream
    Install librabbitmq-0.11.0-7.el9.x86_64             @appstream
    Install librdkafka-1.6.1-102.el9.x86_64             @appstream
    Install lttng-ust-2.12.0-6.el9.x86_64               @appstream
    Install python3-prettytable-0.7.2-27.el9.noarch     @appstream
    Install daxctl-libs-78-2.el9.x86_64                 @baseos
    Install librdmacm-51.0-1.el9.x86_64                 @baseos
    Install ndctl-libs-78-2.el9.x86_64                  @baseos
    Install ceph-common-2:19.2.0-0.el9.x86_64           @Ceph
    Install libcephfs2-2:19.2.0-0.el9.x86_64            @Ceph
    Install librados2-2:19.2.0-0.el9.x86_64             @Ceph
    Install libradosstriper1-2:19.2.0-0.el9.x86_64      @Ceph
    Install librbd1-2:19.2.0-0.el9.x86_64               @Ceph
    Install librgw2-2:19.2.0-0.el9.x86_64               @Ceph
    Install python3-ceph-argparse-2:19.2.0-0.el9.x86_64 @Ceph
    Install python3-ceph-common-2:19.2.0-0.el9.x86_64   @Ceph
    Install python3-cephfs-2:19.2.0-0.el9.x86_64        @Ceph
    Install python3-rados-2:19.2.0-0.el9.x86_64         @Ceph
    Install python3-rbd-2:19.2.0-0.el9.x86_64           @Ceph
    Install python3-rgw-2:19.2.0-0.el9.x86_64           @Ceph
    Install cephadm-2:19.2.0-0.el9.noarch @Ceph-noarch
    Install gperftools-libs-2.9.1-3.el9.x86_64          @epel
    Install libarrow-9.0.0-13.el9.x86_64                @epel
    Install libarrow-doc-9.0.0-13.el9.noarch            @epel
    Install liboath-2.6.12-1.el9.x86_64                 @epel
    Install libunwind-1.6.2-1.el9.x86_64                @epel
    Install parquet-libs-9.0.0-13.el9.x86_64            @epel
    Install re2-1:20211101-20.el9.x86_64                @epel
    Install thrift-0.15.0-4.el9.x86_64                  @epel
Scriptlet output:
   1 Detected autofs mount point /home during canonicalization of /home.
   2 Skipping /home

I dont have that output but here is the bash history:

    1  [2025-01-14-19:21:39] dnf install -y https://download.ceph.com/rpm-19.2.0/el9/noarch/ceph-release-1-1.el9.noarch.rpm
    2  [2025-01-14-19:21:41] dnf install -y ceph-common cephadm
    3  [2025-01-14-19:22:08] host 172.31.78.11
    4  [2025-01-14-19:22:22] cephadm bootstrap --mon-ip 172.31.78.11 --cluster-network 172.31.14.224/27 --allow-fqdn-hostname | tee initial_cephadm_out     5  [2025-01-14-19:23:13] cp /etc/ceph/ceph.pub /root/.ssh/authorized_keys2
    6  [2025-01-14-19:23:13] chmod 600 /root/.ssh/authorized_keys2
    7  [2025-01-14-19:23:26] ceph telemetry on --license sharing-1-0
    8  [2025-01-14-19:23:26] ceph telemetry enable channel perf
    9  [2025-01-14-19:24:13] nmcli connection add type vlan con-name backend ifname backend vlan.id 2987 vlan.parent bond0 ipv4.method manual ipv4.addresses 172.31.14.231/27 ipv4.may-fail no ipv6.method disabled ipv6.may-fail yes connection.zone trusted
   10  [2025-01-14-19:27:18] ceph mgr module disable cephadm
   11  [2025-01-14-19:27:26] ceph fsid
   12  [2025-01-14-19:27:36] cephadm rm-cluster --force --zap-osds --fsid 7f8377c2-d2a4-11ef-adc1-1c34da408e52    13  [2025-01-14-19:28:07] cephadm bootstrap --mon-ip 172.31.78.11 --cluster-network 172.31.14.224/27 --allow-fqdn-hostname | tee initial_cephadm_out    14  [2025-01-14-19:28:46] cp /etc/ceph/ceph.pub /root/.ssh/authorized_keys2
   15  [2025-01-14-19:28:49] chmod 600 /root/.ssh/authorized_keys2
   16  [2025-01-14-19:29:05] ceph telemetry on --license sharing-1-0
   17  [2025-01-14-19:29:06] ceph telemetry enable channel perf
   18  [2025-01-14-19:29:17] scp /root/.ssh/authorized_keys2 ceph02.inf.ethz.ch:/root/.ssh/authorized_keys2    19  [2025-01-14-19:29:19] scp /root/.ssh/authorized_keys2 ceph03.inf.ethz.ch:/root/.ssh/authorized_keys2    20  [2025-01-14-19:29:21] scp /root/.ssh/authorized_keys2 ceph04.inf.ethz.ch:/root/.ssh/authorized_keys2    21  [2025-01-14-19:29:22] scp /root/.ssh/authorized_keys2 ceph05.inf.ethz.ch:/root/.ssh/authorized_keys2
   22  [2025-01-14-19:29:24] yes
   23  [2025-01-14-19:30:15] ceph config set osd osd_memory_target 8G
   24  [2025-01-14-19:30:22] ceph config set mds mds_cache_memory_limit 64G
   25  [2025-01-14-19:30:30] ceph orch apply osd --all-available-devices
   26  [2025-01-14-19:32:27] ceph orch host add ceph02.inf.ethz.ch 172.31.78.12 --labels _admin,nfs,mgr,mds,rgw,rbd,mon,osd,grafana,iscsi    27  [2025-01-14-19:55:31] ceph orch host add ceph03.inf.ethz.ch 172.31.78.13 --labels _admin,nfs,mgr,mds,rgw,rbd,mon,osd,grafana,iscsi    28  [2025-01-14-20:11:57] ceph orch host add ceph04.inf.ethz.ch 172.31.78.14 --labels _admin,nfs,mgr,mds,rgw,rbd,mon,osd,grafana,iscsi    29  [2025-01-14-20:23:04] ceph orch host add ceph05.inf.ethz.ch 172.31.78.15 --labels _admin,nfs,mgr,mds,rgw,rbd,mon,osd,grafana,iscsi    30  [2025-01-14-20:23:27] #ceph orch apply osd --all-available-devices --unmanaged=true    31  [2025-01-14-20:47:24] ceph orch apply osd --all-available-devices --unmanaged=true
   32  [2025-01-14-21:24:41] ls
   33  [2025-01-14-21:24:47] cat initial_cephadm_out





On 3/10/25 09:30, Eugen Block wrote:
(Don't drop the list in your responses)
There's a lot going on in the cephadm.log, it's not clear which of these are relevant to the issue. This one looks bad though:

ValueError: "'xwork.MethodAccessor.denyMethodExecution'" does not appear to be an IPv4 or IPv6 address

Did you build this cluster with 19.2.1 or did you upgrade? Do you have the output from 'ceph orch host ls' lying around (I know, currently it's not working)? Are you using IPv6? I haven't seen this issue yet.

Zitat von Moritz Baumann <mo@xxxxxxxxxxxxx>:

Am 09.03.2025 um 15:22 schrieb Moritz Baumann:
I dont want to sound stupid, but how do I get the log?

Am 09.03.2025 um 11:11 schrieb Eugen Block:
ceph versions



ceph01[22]:~# cephadm shell -- ceph log last cephadm
Inferring fsid 4cfb1138-d2a5-11ef-87eb-1c34da408e52
Inferring config /var/lib/ceph/4cfb1138-d2a5-11ef-87eb-1c34da408e52/mon.ceph01/config Not using image 'f2efb0401a30ec7eda97b6da76b314bd081fcb910cc5dcd826bc7c72c9dfdd7d' as it's not in list of non-dangling images with ceph=True label 2025-02-25T08:42:54.590667+0000 mgr.ceph01.ombogc (mgr.180813) 1706 : cephadm [INF] maintenance mode request for ceph05.inf.ethz.ch has SET the noout group 2025-02-25T08:43:38.380838+0000 mgr.ceph01.ombogc (mgr.180813) 1729 : cephadm [ERR] Detected invalid grafana certificate on host ceph05.inf.ethz.ch: Invalid certificate key: [('PEM routines', '', 'no start line')] 2025-02-25T08:44:40.343420+0000 mgr.ceph01.ombogc (mgr.180813) 1762 : cephadm [INF] Adjusting osd_memory_target on ceph04.inf.ethz.ch to 33423M 2025-02-25T08:44:40.860933+0000 mgr.ceph01.ombogc (mgr.180813) 1764 : cephadm [ERR] Detected invalid grafana certificate on host ceph05.inf.ethz.ch: Invalid certificate key: [('PEM routines', '', 'no start line')] 2025-02-25T08:45:41.993564+0000 mgr.ceph01.ombogc (mgr.180813) 1797 : cephadm [ERR] Detected invalid grafana certificate on host ceph05.inf.ethz.ch: Invalid certificate key: [('PEM routines', '', 'no start line')] 2025-02-25T08:46:42.758717+0000 mgr.ceph01.ombogc (mgr.180813) 1830 : cephadm [ERR] Detected invalid grafana certificate on host ceph05.inf.ethz.ch: Invalid certificate key: [('PEM routines', '', 'no start line')] 2025-02-25T08:47:44.190946+0000 mgr.ceph01.ombogc (mgr.180813) 1864 : cephadm [ERR] Unable to reach remote host ceph05.inf.ethz.ch. SSH connection closed
Traceback (most recent call last):
  File "/usr/share/ceph/mgr/cephadm/ssh.py", line 245, in _execute_command
    r = await conn.run(str(rcmd), input=stdin)
  File "/lib/python3.9/site-packages/asyncssh/connection.py", line 4250, in run
    process = await self.create_process(*args, **kwargs) # type: ignore
  File "/lib/python3.9/site-packages/asyncssh/connection.py", line 4128, in create_process
    chan, process = await self.create_session(
  File "/lib/python3.9/site-packages/asyncssh/connection.py", line 4018, in create_session
    chan = SSHClientChannel(self, self._loop, encoding, errors,
  File "/lib/python3.9/site-packages/asyncssh/channel.py", line 1085, in __init__
    super().__init__(conn, loop, encoding, errors, window, max_pktsize)
  File "/lib/python3.9/site-packages/asyncssh/channel.py", line 140, in __init__
    self._recv_chan: Optional[int] = conn.add_channel(self)
  File "/lib/python3.9/site-packages/asyncssh/connection.py", line 1307, in add_channel
    raise ChannelOpenError(OPEN_CONNECT_FAILED,
asyncssh.misc.ChannelOpenError: SSH connection closed

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "/usr/share/ceph/mgr/orchestrator/_interface.py", line 138, in wrapper
    return OrchResult(f(*args, **kwargs))
  File "/usr/share/ceph/mgr/cephadm/module.py", line 162, in wrapper
    return func(*args, **kwargs)
  File "/usr/share/ceph/mgr/cephadm/module.py", line 2224, in exit_host_maintenance
    outs, errs, _code = self.wait_async(
  File "/usr/share/ceph/mgr/cephadm/module.py", line 815, in wait_async
    return self.event_loop.get_result(coro, timeout)
  File "/usr/share/ceph/mgr/cephadm/ssh.py", line 136, in get_result
    return future.result(timeout)
  File "/lib64/python3.9/concurrent/futures/_base.py", line 446, in result
    return self.__get_result()
  File "/lib64/python3.9/concurrent/futures/_base.py", line 391, in __get_result
    raise self._exception
  File "/usr/share/ceph/mgr/cephadm/serve.py", line 1669, in _run_cephadm     python = await self.mgr.ssh._check_execute_command(host, cmd, addr=addr)   File "/usr/share/ceph/mgr/cephadm/ssh.py", line 300, in _check_execute_command     out, err, code = await self._execute_command(host, cmd, stdin, addr, log_command)   File "/usr/share/ceph/mgr/cephadm/ssh.py", line 253, in _execute_command     raise HostConnectionError(f'Unable to reach remote host {host}. {str(e)}', host, address) cephadm.ssh.HostConnectionError: Unable to reach remote host ceph05.inf.ethz.ch. SSH connection closed 2025-02-25T08:47:44.762770+0000 mgr.ceph01.ombogc (mgr.180813) 1865 : cephadm [ERR] Detected invalid grafana certificate on host ceph05.inf.ethz.ch: Invalid certificate key: [('PEM routines', '', 'no start line')] 2025-02-25T08:47:55.631018+0000 mgr.ceph01.ombogc (mgr.180813) 1874 : cephadm [INF] exit maintenance request has UNSET for the noout group on host ceph05.inf.ethz.ch 2025-02-25T08:48:47.623644+0000 mgr.ceph01.ombogc (mgr.180813) 1901 : cephadm [INF] Adjusting osd_memory_target on ceph05.inf.ethz.ch to 32911M 2025-02-25T08:48:47.827026+0000 mgr.ceph01.ombogc (mgr.180813) 1903 : cephadm [ERR] Detected invalid grafana certificate on host ceph05.inf.ethz.ch: Invalid certificate key: [('PEM routines', '', 'no start line')] 2025-02-25T08:49:21.108932+0000 mgr.ceph01.ombogc (mgr.180813) 1925 : cephadm [INF] Schedule redeploy daemon grafana.ceph05 2025-02-25T08:49:21.292879+0000 mgr.ceph01.ombogc (mgr.180813) 1926 : cephadm [INF] Regenerating cephadm self-signed grafana TLS certificates 2025-02-25T08:49:21.328348+0000 mgr.ceph01.ombogc (mgr.180813) 1929 : cephadm [INF] Deploying daemon grafana.ceph05 on ceph05.inf.ethz.ch 2025-02-25T09:07:00.824841+0000 mgr.ceph01.ombogc (mgr.180813) 2461 : cephadm [INF] Detected new or changed devices on ceph05.inf.ethz.ch 2025-02-25T09:07:01.376245+0000 mgr.ceph01.ombogc (mgr.180813) 2462 : cephadm [INF] Detected new or changed devices on ceph04.inf.ethz.ch 2025-02-28T06:18:29.439683+0000 mgr.ceph01.ombogc (mgr.180813) 137418 : cephadm [ERR] [28/Feb/2025:06:18:29] ENGINE ValueError('"\'xwork.MethodAccessor.denyMethodExecution\'" does not appear to be an IPv4 or IPv6 address')
Traceback (most recent call last):
  File "/lib/python3.9/site-packages/cheroot/server.py", line 1281, in communicate
    req.parse_request()
  File "/lib/python3.9/site-packages/cheroot/server.py", line 714, in parse_request
    success = self.read_request_line()
  File "/lib/python3.9/site-packages/cheroot/server.py", line 815, in read_request_line
    scheme, authority, path, qs, fragment = urllib.parse.urlsplit(uri)
  File "/lib64/python3.9/urllib/parse.py", line 510, in urlsplit
    _check_bracketed_host(bracketed_host)
  File "/lib64/python3.9/urllib/parse.py", line 453, in _check_bracketed_host     ip = ipaddress.ip_address(hostname) # Throws Value Error if not IPv6 or IPv4
  File "/lib64/python3.9/ipaddress.py", line 53, in ip_address
    raise ValueError(f'{address!r} does not appear to be an IPv4 or IPv6 address') ValueError: "'xwork.MethodAccessor.denyMethodExecution'" does not appear to be an IPv4 or IPv6 address

2025-02-28T06:18:29.441452+0000 mgr.ceph01.ombogc (mgr.180813) 137419 : cephadm [ERR] [28/Feb/2025:06:18:29] ENGINE ValueError('"\'xwork.MethodAccessor.denyMethodExecution\'" does not appear to be an IPv4 or IPv6 address')
Traceback (most recent call last):
  File "/lib/python3.9/site-packages/cheroot/server.py", line 1281, in communicate
    req.parse_request()
  File "/lib/python3.9/site-packages/cheroot/server.py", line 714, in parse_request
    success = self.read_request_line()
  File "/lib/python3.9/site-packages/cheroot/server.py", line 815, in read_request_line
    scheme, authority, path, qs, fragment = urllib.parse.urlsplit(uri)
  File "/lib64/python3.9/urllib/parse.py", line 510, in urlsplit
    _check_bracketed_host(bracketed_host)
  File "/lib64/python3.9/urllib/parse.py", line 453, in _check_bracketed_host     ip = ipaddress.ip_address(hostname) # Throws Value Error if not IPv6 or IPv4
  File "/lib64/python3.9/ipaddress.py", line 53, in ip_address
    raise ValueError(f'{address!r} does not appear to be an IPv4 or IPv6 address') ValueError: "'xwork.MethodAccessor.denyMethodExecution'" does not appear to be an IPv4 or IPv6 address

2025-03-07T06:17:28.282574+0000 mgr.ceph01.ombogc (mgr.180813) 447472 : cephadm [ERR] [07/Mar/2025:06:17:28] ENGINE ValueError('"\'xwork.MethodAccessor.denyMethodExecution\'" does not appear to be an IPv4 or IPv6 address')
Traceback (most recent call last):
  File "/lib/python3.9/site-packages/cheroot/server.py", line 1281, in communicate
    req.parse_request()
  File "/lib/python3.9/site-packages/cheroot/server.py", line 714, in parse_request
    success = self.read_request_line()
  File "/lib/python3.9/site-packages/cheroot/server.py", line 815, in read_request_line
    scheme, authority, path, qs, fragment = urllib.parse.urlsplit(uri)
  File "/lib64/python3.9/urllib/parse.py", line 510, in urlsplit
    _check_bracketed_host(bracketed_host)
  File "/lib64/python3.9/urllib/parse.py", line 453, in _check_bracketed_host     ip = ipaddress.ip_address(hostname) # Throws Value Error if not IPv6 or IPv4
  File "/lib64/python3.9/ipaddress.py", line 53, in ip_address
    raise ValueError(f'{address!r} does not appear to be an IPv4 or IPv6 address') ValueError: "'xwork.MethodAccessor.denyMethodExecution'" does not appear to be an IPv4 or IPv6 address

2025-03-09T07:36:00.522250+0000 mgr.ceph01.ombogc (mgr.180813) 536135 : cephadm [ERR] ALERT: Cannot stop active Mgr daemon, Please switch active Mgrs with 'ceph mgr fail ceph01.ombogc'
Note: Warnings can be bypassed with the --force flag
Traceback (most recent call last):
  File "/usr/share/ceph/mgr/orchestrator/_interface.py", line 138, in wrapper
    return OrchResult(f(*args, **kwargs))
  File "/usr/share/ceph/mgr/cephadm/module.py", line 162, in wrapper
    return func(*args, **kwargs)
  File "/usr/share/ceph/mgr/cephadm/module.py", line 2165, in enter_host_maintenance
    raise OrchestratorError(
orchestrator._interface.OrchestratorError: ALERT: Cannot stop active Mgr daemon, Please switch active Mgrs with 'ceph mgr fail ceph01.ombogc'
Note: Warnings can be bypassed with the --force flag



_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux