Ah, I checked on a newer test cluster (Squid) and now I see what you
mean. The alert is shown per OSD in the dashboard, if you open the
dropdown you see which daemons are affected. I think it works a bit
different in Pacific (that's what the customer is still running) when
I last had to modify this. How many OSDs do you have? I noticed that
it takes a few seconds for prometheus to clear the warning with only 3
OSDs in my lab cluster. Maybe you could share a screenshot (with
redacted sensitive data) showing the alerts? And the status of the
affected OSDs as well.
Zitat von "Devin A. Bougie" <devin.bougie@xxxxxxxxxxx>:
Hi Eugen,
No, as far as I can tell I only have one prometheus service running.
———
[root@cephman2 ~]# ceph orch ls prometheus --export
service_type: prometheus
service_name: prometheus
placement:
count: 1
label: _admin
[root@cephman2 ~]# ceph orch ps --daemon-type prometheus
NAME HOST PORTS STATUS
REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID
CONTAINER ID
prometheus.cephman2 cephman2.classe.cornell.edu *:9095 running
(12h) 4m ago 3w 350M - 2.43.0 a07b618ecd1d
5a8d88682c28
———
Anything else I can check or do?
Thanks,
Devin
On Jan 13, 2025, at 6:39 PM, Eugen Block <eblock@xxxxxx> wrote:
Do you have two Prometheus instances? Maybe you could share
ceph orch ls prometheus --export
Or alternatively:
ceph orch ps --daemon-type prometheus
You can use two instances for HA, but then you need to change the
threshold for both, of course.
Zitat von "Devin A. Bougie"
<devin.bougie@xxxxxxxxxxx<mailto:devin.bougie@xxxxxxxxxxx>>:
Thanks, Eugen! Just incase you have any more suggestions, this
still isn’t quite working for us.
Perhaps one clue is that in the Alerts view of the cephadm
dashboard, every alert is listed twice. We see two CephPGImbalance
alerts, both set to 30% after redeploying the service. If I then
follow your procedure, one of the alerts updates to 50% as
configured, but the other stays at 30. Is it normal to see each
alert listed twice, or did I somehow make a mess of things when
trying to change the default alerts?
No problem if it’s not an obvious answer, we can live with and
ignore the spurious CephPGImbalance alerts.
Thanks again,
Devin
On Jan 7, 2025, at 2:14 AM, Eugen Block <eblock@xxxxxx> wrote:
Hi,
sure thing, here's the diff how I changed it to 50% deviation instead of 30%:
---snip---
diff -u
/var/lib/ceph/{FSID}/prometheus.host1/etc/prometheus/alerting/ceph_alerts.yml
/var/lib/ceph/{FSID}/prometheus.host1/etc/prometheus/alerting/ceph_alerts.yml.dist
---
/var/lib/ceph/{FSID}/prometheus.host1/etc/prometheus/alerting/ceph_alerts.yml 2024-12-17 10:03:23.540179209
+0100
+++
/var/lib/ceph/{FSID}/prometheus.host1/etc/prometheus/alerting/ceph_alerts.yml.dist 2024-12-17 10:03:00.380883413
+0100
@@ -237,13 +237,13 @@
type: "ceph_default"
- alert: "CephPGImbalance"
annotations:
- description: "OSD {{ $labels.ceph_daemon }} on {{
$labels.hostname }} deviates by more than 50% from average PG count."
+ description: "OSD {{ $labels.ceph_daemon }} on {{
$labels.hostname }} deviates by more than 30% from average PG count."
summary: "PGs are not balanced across OSDs"
expr: |
abs(
((ceph_osd_numpg > 0) - on (job) group_left
avg(ceph_osd_numpg > 0) by (job)) /
on (job) group_left avg(ceph_osd_numpg > 0) by (job)
- ) * on (ceph_daemon) group_left(hostname) ceph_osd_metadata > 0.50
+ ) * on (ceph_daemon) group_left(hostname) ceph_osd_metadata > 0.30
---snip---
Then you restart prometheus ('ceph orch ps --daemon-type prometheus'
shows you the exact daemon name):
ceph orch daemon restart prometheus.host1
This will only work until you upgrade prometheus, of course.
Regards,
Eugen
Zitat von "Devin A. Bougie" <devin.bougie@xxxxxxxxxxx>:
Thanks, Eugen. I’m afraid I haven’t yet found a way to either
disable the CephPGImbalance alert or change it to handle different
OSD sizes. Changing
/var/lib/ceph/<cluster_id>/home/ceph_default_alerts.yml doesn’t seem
to have any effect, and I haven’t even managed to change the
behavior from within the running prometheus container.
If you have a functioning workaround, can you give a little more
detail on exactly what yaml file you’re changing and where?
Thanks again,
Devin
On Dec 30, 2024, at 12:39 PM, Eugen Block <eblock@xxxxxx> wrote:
Funny, I wanted to take a look next week how to deal with different
OSD sizes or if somebody already has a fix for that. My workaround
is changing the yaml file for Prometheus as well.
Zitat von "Devin A. Bougie" <devin.bougie@xxxxxxxxxxx>:
Hi, All. We are using cephadm to manage a 19.2.0 cluster on
fully-updated AlmaLinux 9 hosts, and would greatly appreciate help
modifying or overriding the alert rules in ceph_default_alerts.yml.
Is the best option to simply update the
/var/lib/ceph/<cluster_id>/home/ceph_default_alerts.yml file?
In particular, we’d like to either disable the CephPGImbalance alert
or change it to calculate averages per-pool or per-crush_rule
instead of globally as in [1].
We currently have PG autoscaling enabled, and have two separate
crush_rules (one with large spinning disks, one with much smaller
nvme drives). Although I don’t believe it causes any technical
issues with our configuration, our dashboard is full of
CephPGImbalance alerts that would be nice to clean up without having
to create periodic silences.
Any help or suggestions would be greatly appreciated.
Many thanks,
Devin
[1]
https://nam12.safelinks.protection.outlook.com/?url=https%3A%2F%2Fgithub.com%2Frook%2Frook%2Fdiscussions%2F13126%23discussioncomment-10043490&data=05%7C02%7Cdevin.bougie%40cornell.edu%7C27dddfa00b2e4475b30e08dd342b82cc%7C5d7e43661b9b45cf8e79b14b27df46e1%7C0%7C0%7C638724083682129542%7CUnknown%7CTWFpbGZsb3d8eyJFbXB0eU1hcGkiOnRydWUsIlYiOiIwLjAuMDAwMCIsIlAiOiJXaW4zMiIsIkFOIjoiTWFpbCIsIldUIjoyfQ%3D%3D%7C0%7C%7C%7C&sdata=%2FrmKNJ1hVwWdQ5U05CvNXX1Df3f4SR2HAxTyZA3PJKw%3D&reserved=0<https://github.com/rook/rook/discussions/13126#discussioncomment-10043490>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx>
To unsubscribe send an email to
ceph-users-leave@xxxxxxx<mailto:ceph-users-leave@xxxxxxx>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx