Hi,
there's a tracker issue [0] for that. I was assisting with the same
issue in a different thread [1].
Thanks,
Eugen
[0] https://tracker.ceph.com/issues/67329
[1]
https://lists.ceph.io/hyperkitty/list/ceph-users@xxxxxxx/thread/SRJPC5ZYTPXF63AKGIIOA2LLLBBWCIT4/
Zitat von Alex Sanderson <alex@xxxxxxxxxxxx>:
Hi everyone,
I recently upgraded from Quincy to Reef v18.2.4 and my dashboard and
mgr systems have been broken since. Since the upgrade I was slowly
removing and zapping osd's that still had the 64k
"bluestore_bdev_block_size" and decided to have a look at the
dashboard problem. I restarted the mgrs one at a time and they
showed in status that they working but actually the cephadm module
was failing. The systems were all upgraded to 18 via orch from
17.2.7 and are running the official docker images.
This is the error message:
debug 2024-08-13T10:08:11.736+0000 7fd30dbe0640 -1 mgr load Failed
to construct class in 'cephadm'
debug 2024-08-13T10:08:11.736+0000 7fd30dbe0640 -1 mgr load
Traceback (most recent call last):
File "/usr/share/ceph/mgr/cephadm/module.py", line 619, in __init__
self.to_remove_osds.load_from_store()
File "/usr/share/ceph/mgr/cephadm/services/osd.py", line 924, in
load_from_store
osd_obj = OSD.from_json(osd, rm_util=self.rm_util)
File "/usr/share/ceph/mgr/cephadm/services/osd.py", line 789, in from_json
return cls(**inp)
TypeError: __init__() got an unexpected keyword argument 'original_weight'
debug 2024-08-13T10:08:11.736+0000 7fd30dbe0640 -1 mgr operator()
Failed to run module in active mode ('cephadm')
The config-key responsible was mgr/cephadm/osd_remove_queue
This is what it looked like before. After removing the
original_weight field and setting the variable again, the cephadm
module loads and orch works. It seems like a bug.
[{"osd_id": 89, "started": true, "draining": true, "stopped": false,
"replace": false, "force": true, "zap": true, "hostname": "goanna",
"original_weight": 0.930999755859375, "drain_started_at":
"2024-08-12T13:21:04.458019Z", "drain_stopped_at": null,
"drain_done_at": null, "process_started_at":
"2024-08-12T13:20:40.021185Z"}, {"osd_id": 37, "started": true,
"draining": true, "stopped": false, "replace": false, "force": true,
"zap": true, "hostname": "gsceph1osd05", "original_weight": 4,
"drain_started_at": "2024-08-10T06:30:37.569931Z",
"drain_stopped_at": null, "drain_done_at": null,
"process_started_at": "2024-08-10T06:30:19.729143Z"}, {"osd_id": 47,
"started": true, "draining": true, "stopped": false, "replace":
false, "force": true, "zap": true, "hostname": "gsceph1osd07",
"original_weight": 4, "drain_started_at":
"2024-08-10T09:54:49.132830Z", "drain_stopped_at": null,
"drain_done_at": null, "process_started_at":
"2024-08-10T09:54:34.367655Z"}]
I thought I should put this out there in case anyone else was having
a weird issue with a keyword argument problem. It did not fix the
problem with the dashboard, still working on that.
Alex
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx