(Again to the mailing list, ups) Hi Travis, Thanks for your input – it's greatly appreciated. I assume that my deployment was using v17.2.6, as I hadn't explicitly specified a version in my provided rook-ceph-cluster/values.yaml <https://github.com/rook/rook/blob/master/deploy/charts/rook-ceph-cluster/values.yaml#L96> However, due to an issue with the Dashboard (could not use the object-gateway tab), I updated to v17.2.7 so I cant confirm this anymore. I'm considering completely destroying the cluster and redeploying with potentially even v16. Any further advice or insights in this regard would be very helpful. Best regards, P On Mon, 27 Nov 2023 at 20:32, Travis Nielsen <tnielsen@xxxxxxxxxx> wrote: > Sounds like you're hitting a known issue with v17.2.7. > https://github.com/rook/rook/issues/13136 > > The fix will be in v18.2.1 if it's an option to upgrade to Reef. If not, > you'll need to use v17.2.6 until the fix comes out for quincy in v17.2.8. > > Travis > > On Thu, Nov 23, 2023 at 4:06 PM P Wagner-Beccard < > wagner-kerschbaumer@xxxxxxxxxxxxx> wrote: > >> Hi Mailing-Lister's, >> >> I am reaching out for assistance regarding a deployment issue I am facing >> with Ceph on a 4 node RKE2 cluster. We are attempting to deploy Ceph via >> the rook helm chart, but we are encountering an issue that apparently >> seems >> related to a known bug (https://tracker.ceph.com/issues/61597). >> >> During the OSD preparation phase, the deployment consistently fails with >> an >> IndexError: list index out of range. The logs indicate a problem occurs >> when configuring new Disks, specifically using /dev/dm-3 as a metadata >> device. It's important to note that /dev/dm-3 is an LVM on top of an mdadm >> RAID, which might or might not be contributing to this issue. (I swear, >> this setup worked already) >> >> Here is a snippet of the error from the deployment logs: >> > 2023-11-23 23:11:30.196913 D | exec: IndexError: list index out of range >> > 2023-11-23 23:11:30.236962 C | rookcmd: failed to configure devices: >> failed to initialize osd: failed ceph-volume report: exit status 1 >> https://paste.openstack.org/show/bileqRFKbolrBlTqszmC/ >> >> We have attempted different configurations, including specifying devices >> explicitly and using the useAllDevices: true option with a specified >> metadata device (/dev/dm-3 or the /dev/pv_md0/lv_md0 path). However, the >> issue persists across multiple configurations. >> >> tested configurations are as follows: >> >> Explicit device specification: >> >> ```yaml >> nodes: >> - name: "ceph01.maas" >> devices: >> - name: /dev/dm-1 >> - name: /dev/dm-2 >> - name: "sdb" >> config: >> metadataDevice: "/dev/dm-3" >> - name: "sdc" >> config: >> metadataDevice: "/dev/dm-3" >> ``` >> >> General device specification with metadata device: >> ```yaml >> storage: >> useAllNodes: true >> useAllDevices: true >> config: >> metadataDevice: /dev/dm-3 >> ``` >> >> I would greatly appreciate any insights or recommendations on how to >> proceed or work around this issue. >> Is there a halfway decent way to apply the fix or maybe a workaround that >> we can apply to successfully deploy Ceph in our environment? >> >> Kind regards, >> _______________________________________________ >> ceph-users mailing list -- ceph-users@xxxxxxx >> To unsubscribe send an email to ceph-users-leave@xxxxxxx >> >> _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx