Any data that can be retrieved would be a plus. There is much data on this array that I don't mind being trashed. The older drives are WD Red, they are pre-SHMR. I have made sure after that to use WD Red Plus and WD Red Pro drives. From what I found online, they should be CMR too. Unless they quietly changed those too. No, the conversion definitely did not stop at 0%. It ran for several hours. It stopped during the night, so I can't tell you more. I am worried that the processes are hung, though. Is that normal? Thank you for your time! On Mon, Apr 24, 2023 at 9:41 AM Wols Lists <antlists@xxxxxxxxxxxxxxx> wrote: > > On 23/04/2023 20:09, Jove wrote: > > # mdadm --version > > mdadm - v4.2 - 2021-12-30 - 8 > > > > # mdadm -D /dev/md0 > > /dev/md0: > > Version : 1.2 > > Creation Time : Sat Oct 21 01:57:20 2017 > > Raid Level : raid6 > > Array Size : 7813771264 (7.28 TiB 8.00 TB) > > Used Dev Size : 3906885632 (3.64 TiB 4.00 TB) > > Raid Devices : 4 > > Total Devices : 5 > > Persistence : Superblock is persistent > > > > Intent Bitmap : Internal > > > > Update Time : Sun Apr 23 10:32:01 2023 > > State : clean, degraded > > Active Devices : 3 > > Working Devices : 5 > > Failed Devices : 0 > > Spare Devices : 2 > > > > Layout : left-symmetric-6 > > Chunk Size : 512K > > > > Consistency Policy : bitmap > > > > New Layout : left-symmetric > > > > Name : atom:0 (local to host atom) > > UUID : 8c56384e:ba1a3cec:aaf34c17:d0cd9318 > > Events : 669453 > > > > Number Major Minor RaidDevice State > > 0 8 33 0 active sync /dev/sdc1 > > 1 8 97 1 active sync /dev/sdg1 > > 3 8 49 2 active sync /dev/sdd1 > > 5 8 80 3 spare rebuilding /dev/sdf > > > > 4 8 64 - spare /dev/sde > > This bit looks good. You have three active drives, so I'm HOPEFUL your > data hasn't actually been damaged. > > I've cc'd two people more experienced than me who I hope can help. > > Cheers, > Wol