Autoscaler stopped working after upgrade Octopus -> Pacific

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Dear all,

just upgraded our cluster from Octopus to Pacific (16.2.10). This
introduced an error in autoscaler:

2022-10-11T14:47:40.421+0200 7f3ec2d03700  0 [pg_autoscaler ERROR root] pool 17 has overlapping roots: {-4, -1}
2022-10-11T14:47:40.423+0200 7f3ec2d03700  0 [pg_autoscaler ERROR root] pool 22 has overlapping roots: {-4, -1}
2022-10-11T14:47:40.423+0200 7f3ec2d03700  0 [pg_autoscaler ERROR root] pool 23 has overlapping roots: {-4, -1}
2022-10-11T14:47:40.427+0200 7f3ec2d03700  0 [pg_autoscaler ERROR root] pool 27 has overlapping roots: {-6, -4, -1}
2022-10-11T14:47:40.428+0200 7f3ec2d03700  0 [pg_autoscaler ERROR root] pool 28 has overlapping roots: {-6, -4, -1}

Autoscaler status is empty:

[cephmon1] /root # ceph osd pool autoscale-status
[cephmon1] /root # 


On https://forum.proxmox.com/threads/ceph-overlapping-roots.104199/ I
found something similar:

---
I assume that you have at least one pool that still has the
"replicated_rule" assigned, which does not make a distinction between the
device class of the OSDs.

This is why you see this error. The autoscaler cannot decide how many PGs
the pools need. Make sure that all pools are assigned a rule that limit
them to a device class and the errors should stop.
---

Indeed, we have a mixed cluster (hdd + ssd) with some pools spanning hdd-
only, some ssd-only and some both (ec & replicated) which don't care about
the storage device class (e.g. via default "replicated_rule"):

[cephmon1] /root # ceph osd crush rule ls
replicated_rule
ssd_only_replicated_rule
hdd_only_replicated_rule
default.rgw.buckets.data.ec42
test.ec42
[cephmon1] /root #


That worked flawlessly until Octopus. Any idea how to make autoscaler work
again with that kind of setup? Can I really have pools on one device class
only in Pacific in order to get a functional autoscaler?

Thanks,
Andreas
-- 
| Andreas Haupt            | E-Mail: andreas.haupt@xxxxxxx
|  DESY Zeuthen            | WWW:    http://www-zeuthen.desy.de/~ahaupt
|  Platanenallee 6         | Phone:  +49/33762/7-7359
|  D-15738 Zeuthen         | Fax:    +49/33762/7-7216

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux