Hi,
I don't really know how the ceph-exporter gets into your Quincy
bootstrap, when I deploy it with Quincy (also cephadm is from Quincy
repo) it doesn't try to deploy ceph-exporter. When I use cephadm from
Reef repo it does deploy a Reef cluster including ceph-exporter
successfully.
As a workaround you should be able to deploy the cluster if you skip
the monitoring stack during bootstrap and add it later after you
upgraded to Reef. The parameter is --skip-monitoring-stack. Or you
deploy directly with Reef with the --image option:
cephadm --image quay.io/ceph/ceph:v18 bootstrap ...
The quincy docs [1] already contain the information about the ceph-exporter:
With the introduction of ceph-exporter daemon, the prometheus module
will no longer export Ceph daemon perf counters as prometheus
metrics by default.
But when trying to apply it in Quincy it fails as well:
[root@quincy-3 ~]# ceph orch apply ceph-exporter
Error EINVAL: Usage:
ceph orch apply -i <yaml spec> [--dry-run]
ceph orch apply <service_type> [--placement=<placement_string>]
[--unmanaged]
I'll check if there's an existing tracker issue.
Thanks,
Eugen
[1]
https://docs.ceph.com/en/quincy/mgr/prometheus/#ceph-daemon-performance-counters-metrics
Zitat von ankit@xxxxxxxxx:
Hi Guys,
I am newbie and trying to install Ceph Storage cluster and following
this
https://docs.ceph.com/en/latest/cephadm/install/#cephadm-deploying-new-cluster
=============================================================
OS - Ubuntu 22.04.3 LTS (Jammy Jellyfish)
4 node Cluster - mon1,mgr1,2 OSD nodes
mon1 node can ssh all nodes via root to sudo ceph-user and ceph-user
to ceph-user on other nodes
basic requirements are done like podman, python3, systemd,ntp, lvm.
===================================================================
cephadm bootstrap --mon-ip 192.168.2.125 - after running this i am
getting following error.
ceph-user@mon1:~$ sudo cephadm bootstrap --mon-ip 192.168.2.125
Creating directory /etc/ceph for ceph.conf
Verifying podman|docker is present...
Verifying lvm2 is present...
Verifying time synchronization is in place...
Unit chrony.service is enabled and running
Repeating the final host check...
podman (/usr/bin/podman) version 3.4.4 is present
systemctl is present
lvcreate is present
Unit chrony.service is enabled and running
Host looks OK
Cluster fsid: 90813682-c656-11ee-9ca3-0800274ff361
Verifying IP 192.168.2.125 port 3300 ...
Verifying IP 192.168.2.125 port 6789 ...
Mon IP `192.168.2.125` is in CIDR network `192.168.2.0/24`
Mon IP `192.168.2.125` is in CIDR network `192.168.2.0/24`
Internal network (--cluster-network) has not been provided, OSD
replication will default to the public_network
Pulling container image quay.io/ceph/ceph:v17...
Ceph version: ceph version 17.2.7
(b12291d110049b2f35e32e0de30d70e9a4c060d2) quincy (stable)
Extracting ceph user uid/gid from container image...
Creating initial keys...
Creating initial monmap...
Creating mon...
Waiting for mon to start...
Waiting for mon...
mon is available
Assimilating anything we can from ceph.conf...
Generating new minimal ceph.conf...
Restarting the monitor...
Setting mon public_network to 192.168.2.0/24
Wrote config to /etc/ceph/ceph.conf
Wrote keyring to /etc/ceph/ceph.client.admin.keyring
Creating mgr...
Verifying port 9283 ...
Waiting for mgr to start...
Waiting for mgr...
mgr not available, waiting (1/15)...
mgr not available, waiting (2/15)...
mgr not available, waiting (3/15)...
mgr not available, waiting (4/15)...
mgr not available, waiting (5/15)...
mgr not available, waiting (6/15)...
mgr not available, waiting (7/15)...
mgr is available
Enabling cephadm module...
Waiting for the mgr to restart...
Waiting for mgr epoch 5...
mgr epoch 5 is available
Setting orchestrator backend to cephadm...
Generating ssh key...
Wrote public SSH key to /etc/ceph/ceph.pub
Adding key to root@localhost authorized_keys...
Adding host mon1...
Deploying mon service with default placement...
Deploying mgr service with default placement...
Deploying crash service with default placement...
Deploying ceph-exporter service with default placement...
Non-zero exit code 22 from /usr/bin/podman run --rm --ipc=host
--stop-signal=SIGTERM --net=host --entrypoint /usr/bin/ceph --init
-e CONTAINER_IMAGE=quay.io/ceph/ceph:v17 -e NODE_NAME=mon1 -e
CEPH_USE_RANDOM_NONCE=1 -v
/var/log/ceph/90813682-c656-11ee-9ca3-0800274ff361:/var/log/ceph:z
-v /tmp/ceph-tmpnjonhex7:/etc/ceph/ceph.client.admin.keyring:z -v
/tmp/ceph-tmp3gil6lbb:/etc/ceph/ceph.conf:z quay.io/ceph/ceph:v17
orch apply ceph-exporter
/usr/bin/ceph: stderr Error EINVAL: Usage:
/usr/bin/ceph: stderr ceph orch apply -i <yaml spec> [--dry-run]
/usr/bin/ceph: stderr ceph orch apply <service_type>
[--placement=<placement_string>] [--unmanaged]
/usr/bin/ceph: stderr
Traceback (most recent call last):
File "/usr/sbin/cephadm", line 9653, in <module>
main()
File "/usr/sbin/cephadm", line 9641, in main
r = ctx.func(ctx)
File "/usr/sbin/cephadm", line 2205, in _default_image
return func(ctx)
File "/usr/sbin/cephadm", line 5774, in command_bootstrap
prepare_ssh(ctx, cli, wait_for_mgr_restart)
File "/usr/sbin/cephadm", line 5275, in prepare_ssh
cli(['orch', 'apply', t])
File "/usr/sbin/cephadm", line 5708, in cli
return CephContainer(
File "/usr/sbin/cephadm", line 4144, in run
out, _, _ = call_throws(self.ctx, self.run_cmd(),
File "/usr/sbin/cephadm", line 1853, in call_throws
raise RuntimeError('Failed command: %s' % ' '.join(command))
RuntimeError: Failed command: /usr/bin/podman run --rm --ipc=host
--stop-signal=SIGTERM --net=host --entrypoint /usr/bin/ceph --init
-e CONTAINER_IMAGE=quay.io/ceph/ceph:v17 -e NODE_NAME=mon1 -e
CEPH_USE_RANDOM_NONCE=1 -v
/var/log/ceph/90813682-c656-11ee-9ca3-0800274ff361:/var/log/ceph:z
-v /tmp/ceph-tmpnjonhex7:/etc/ceph/ceph.client.admin.keyring:z -v
/tmp/ceph-tmp3gil6lbb:/etc/ceph/ceph.conf:z quay.io/ceph/ceph:v17
orch apply ceph-exporter
What i am doing wrong or missing? Please help.
Many Thanks
AS
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx