Hi Frédéric, Thank you for the response. I tried disabling and re-enabling the module, and it seems that the recovery metrics are indeed being collected, but still not displayed on the new dashboard. Interestingly, I have another environment running Ceph 19.2.0 where the recovery throughput is displayed correctly, but I’m unable to identify any major differences between the two setups. Do you have any additional suggestions for troubleshooting this issue further? Thanks again for your help! Best regards, Sanjay Mohan Software Defined Storage Engineer sanjaymohan@xxxxxxxxxxxxx ________________________________ From: Frédéric Nass <frederic.nass@xxxxxxxxxxxxxxxx> Sent: 21 October 2024 1:22 PM To: Sanjay Mohan <sanjaymohan@xxxxxxxxxxxxx> Cc: ceph-users <ceph-users@xxxxxxx> Subject: Re: Issue with Recovery Throughput Not Visible in Ceph Dashboard After Upgrade to 19.2.0 (Squid) Hi Sanjay, I've just checked the dashboard of a v19.2.0 cluster, and the recovery throughput is displayed correctly, as shown in the screenshot here [1]. You might want to consider redeploying the dashboard. Regards, Frédéric. [1] https://docs.ceph.com/en/latest/mgr/dashboard/ ----- Le 19 Oct 24, à 19:23, Sanjay Mohan sanjaymohan@xxxxxxxxxxxxx a écrit : > Dear Ceph Users, > I hope this message finds you well. > I recently performed an upgrade of my Ceph cluster, moving through the following > versions: > > * 17.2.7 >> 18.2.0 >> 18.2.2 >> 18.2.4 >> 19.2.0 (Squid) > > After successfully upgrading to Ceph 19.2.0, I noticed an issue where the > recovery throughput is no longer visible in the new Ceph dashboard. However, > Old dashboard metrics and features seem to be working as expected. It is > important to note that the recovery throughput was displayed properly in the > previous version of the Ceph dashboard. I am using Cephadm for the > installation, not Rook. > Current behavior: > > * Recovery throughput metrics are not displayed in the new dashboard after > upgrading to 19.2.0. > > Expected behavior: > > * The recovery throughput should be visible, as it was in previous versions of > the Ceph dashboard. > > I am reaching out to inquire if there are any known issues, workarounds, or > upcoming fixes for this. Your assistance in this matter would be greatly > appreciated. > Thank you for your time and support. I look forward to hearing from you soon. > Best regards, > Sanjay Mohan > Software Defined Storage Engineer > sanjaymohan@xxxxxxxxxxxxx > > [Amrita University] > Disclaimer: The information transmitted in this email, including attachments, is > intended only for the person(s) or entity to which it is addressed and may > contain confidential and/or privileged material. Any review, retransmission, > dissemination or other use of, or taking of any action in reliance upon this > information by persons or entities other than the intended recipient is > prohibited. Any views expressed in any message are those of the individual > sender and may not necessarily reflect the views of Amrita Vishwa Vidyapeetham. > If you received this in error, please contact the sender and destroy any copies > of this information. > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx [Amrita University] Disclaimer: The information transmitted in this email, including attachments, is intended only for the person(s) or entity to which it is addressed and may contain confidential and/or privileged material. Any review, retransmission, dissemination or other use of, or taking of any action in reliance upon this information by persons or entities other than the intended recipient is prohibited. Any views expressed in any message are those of the individual sender and may not necessarily reflect the views of Amrita Vishwa Vidyapeetham. If you received this in error, please contact the sender and destroy any copies of this information. _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx