Hi everyone,
I recently upgraded from Quincy to Reef v18.2.4 and my dashboard and mgr
systems have been broken since. Since the upgrade I was slowly removing
and zapping osd's that still had the 64k "bluestore_bdev_block_size" and
decided to have a look at the dashboard problem. I restarted the mgrs
one at a time and they showed in status that they working but actually
the cephadm module was failing. The systems were all upgraded to 18 via
orch from 17.2.7 and are running the official docker images.
This is the error message:
debug 2024-08-13T10:08:11.736+0000 7fd30dbe0640 -1 mgr load Failed to
construct class in 'cephadm'
debug 2024-08-13T10:08:11.736+0000 7fd30dbe0640 -1 mgr load Traceback
(most recent call last):
File "/usr/share/ceph/mgr/cephadm/module.py", line 619, in __init__
self.to_remove_osds.load_from_store()
File "/usr/share/ceph/mgr/cephadm/services/osd.py", line 924, in
load_from_store
osd_obj = OSD.from_json(osd, rm_util=self.rm_util)
File "/usr/share/ceph/mgr/cephadm/services/osd.py", line 789, in
from_json
return cls(**inp)
TypeError: __init__() got an unexpected keyword argument 'original_weight'
debug 2024-08-13T10:08:11.736+0000 7fd30dbe0640 -1 mgr operator() Failed
to run module in active mode ('cephadm')
The config-key responsible was mgr/cephadm/osd_remove_queue
This is what it looked like before. After removing the original_weight
field and setting the variable again, the cephadm module loads and orch
works. It seems like a bug.
[{"osd_id": 89, "started": true, "draining": true, "stopped": false,
"replace": false, "force": true, "zap": true, "hostname": "goanna",
"original_weight": 0.930999755859375, "drain_started_at":
"2024-08-12T13:21:04.458019Z", "drain_stopped_at": null,
"drain_done_at": null, "process_started_at":
"2024-08-12T13:20:40.021185Z"}, {"osd_id": 37, "started": true,
"draining": true, "stopped": false, "replace": false, "force": true,
"zap": true, "hostname": "gsceph1osd05", "original_weight": 4,
"drain_started_at": "2024-08-10T06:30:37.569931Z", "drain_stopped_at":
null, "drain_done_at": null, "process_started_at":
"2024-08-10T06:30:19.729143Z"}, {"osd_id": 47, "started": true,
"draining": true, "stopped": false, "replace": false, "force": true,
"zap": true, "hostname": "gsceph1osd07", "original_weight": 4,
"drain_started_at": "2024-08-10T09:54:49.132830Z", "drain_stopped_at":
null, "drain_done_at": null, "process_started_at":
"2024-08-10T09:54:34.367655Z"}]
I thought I should put this out there in case anyone else was having a
weird issue with a keyword argument problem. It did not fix the problem
with the dashboard, still working on that.
Alex
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx