Re: Fwd: RAID 5 growing hanged at 0.0%

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

Thanks for the reply. I didn't do much. I put the HDDs in another
machine with a fresh installation of debian/OMV, to rule out any
hardware issue. The mdadm --detail detected the array with the 5
drives and state was reshaping paused or something. I resumed with
mdadm --readwrite /dev/md127 it started for a couple of seconds, just
to stop again, more or less in the same block...

root@openmediavault:~# cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5]
[raid4] [raid10]
md127 : active raid5 sda[6] sde[4] sdd[7] sdb[5]
      11720661504 blocks super 1.2 level 5, 512k chunk, algorithm 2
[5/4] [UUUU_]
      [>....................]  reshape =  0.0% (2303100/3906887168)
finish=3315.9min speed=4K/sec
      bitmap: 6/30 pages [24KB], 65536KB chunk


Then tried to assemble it back with just 4 HDDs, the original 4 HDDs:

root@openmediavault:~# mdadm --assemble --force /dev/md127 /dev/sda
/dev/sdb /dev/sdd /dev/sde
mdadm: Marking array /dev/md127 as 'clean'
mdadm: /dev/md127 has been started with 4 drives (out of 5).
root@openmediavault:~# mdadm --detail /dev/md127
/dev/md127:
Version : 1.2
Creation Time : Sat Apr  7 18:19:22 2018
Raid Level : raid5
Array Size : 11720661504 (11177.69 GiB 12001.96 GB)
Used Dev Size : 3906887168 (3725.90 GiB 4000.65 GB)
Raid Devices : 5
Total Devices : 4
Persistence : Superblock is persistent


Intent Bitmap : Internal


Update Time : Sat Jan 14 13:45:57 2023
State : clean, degraded
Active Devices : 4
Working Devices : 4
Failed Devices : 0
Spare Devices : 0


Layout : left-symmetric
Chunk Size : 512K


Consistency Policy : bitmap


Delta Devices : 1, (4->5)


Name : helios4:Helios4
UUID : b6d959d8:96a3a0e6:29da39c1:15407b98
Events : 44240


Number   Major   Minor   RaidDevice State
6       8        0        0      active sync   /dev/sda
5       8       16        1      active sync   /dev/sdb
7       8       48        2      active sync   /dev/sdd
4       8       64        3      active sync   /dev/sde
-       0        0        4      removed
root@openmediavault:~# cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5]
[raid4] [raid10]
md127 : active (auto-read-only) raid5 sda[6] sde[4] sdd[7] sdb[5]
11720661504 blocks super 1.2 level 5, 512k chunk, algorithm 2 [5/4] [UUUU_]
bitmap: 6/30 pages [24KB], 65536KB chunk


unused devices: <none>

root@openmediavault:~# cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5]
[raid4] [raid10]
md127 : active raid5 sda[6] sde[4] sdd[7] sdb[5]
      11720661504 blocks super 1.2 level 5, 512k chunk, algorithm 2
[5/4] [UUUU_]
      [>....................]  reshape =  0.0% (2303100/3906887168)
finish=3315.9min speed=19624K/sec
      bitmap: 6/30 pages [24KB], 65536KB chunk

Then it just stopped again when reshaping.

I'm now at this state.
Regarding (2): Yeah, I know that. I'm slowly replacing them with REDs.

Yeah, I saw that "assemble no-bad-blocks force" in the wiki, but I was
also afraid to try.

Thanks a lot for the help.

BR,
Tiago

On Mon, 16 Jan 2023 at 11:44, Wols Lists <antlists@xxxxxxxxxxxxxxx> wrote:
>
> On 16/01/2023 11:12, Tiago Afonso wrote:
> > Hi,
> >
> > Long story short, the bigger story is here:
> > https://forum.openmediavault.org/index.php?thread/45829-raid-5-growing-hanged-at-0-0/
> > I had a 4x4TB raid5 configuration and I added a new 4TB drive in order
> > to grow the array as 5x4TB raid5. I did this via openmediavault GUI
> > (my mistake, maybe I should have checked tutorials and more info on
> > how to do it properly). The reshape started but hung right at the
> > beginning.
> >
> > root@openmediavault:~# cat /proc/mdstat Personalities : [raid6]
> > [raid5] [raid4] [linear] [multipath] [raid0] [raid1] [raid10] md127 :
> > active raid5 sdh[4] sdg[7] sdf[6] sde[5] sdi[8] 11720661504 blocks
> > super 1.2 level 5, 512k chunk, algorithm 2 [5/5] [UUUUU]
> > [>....................] reshape = 0.0% (1953276/3906887168)
> > finish=2008857.0min speed=32K/sec bitmap: 6/30 pages [24KB], 65536KB
> > chunk
> >
> > I forced shutdown the process by pulling the power, as it was not
> > letting me stop the process otherwise. Even put the disks in another
> > machine to no avail. The problem seems to be that the array was not
> > healthy and there were bad-blocks, and I think that is why it keeps
> > stopping the reshape. The bad-blocks are in two of the HDDs and it
> > seems they are the same block. Now I'm stuck. I'm unable to reshape
> > back to 4 HDDs (I did try this) or access the data.
> >
> > How can I get out of this situation and recover the data back even if
> > not all data?
> >
> > I read the wiki, but I don't have much experience in linux or raid.
> > I'm afraid of doing something that puts me in an even worse situation.
> >
> > Attached are some commands output.
> >
> >
> > Thank you.
>
> Thanks. Looks like you've done your homework :-)
>
> I was about to say this looks like a well-known problem, and then I saw
> the mdadm version and the kernel. You should not be getting that problem
> with something this up-to-date.
>
> The good news is this still looks like that problem, but the question is
> what on earth is going on. Can you boot into a rescue disk rather than
> the mediavault stuff?
>
> The array is clean but degraded. Not knowing what you've done, I'm not
> sure how to put the array together again, but the really good news is it
> looks like - because the reshape never started - it shouldn't be a hard
> job to retrieve everything. Then we can start sorting things out from there.
>
> I'm going to bring in the two experts - they might take a little while
> to respond - but in the meantime two points to ponder ...
>
> (1) raid badblocks are a mis-feature, as soon as we get the array back,
> we want them gone.
> (2) "SCT Error Recovery not supported" - Your blues are not suitable
> raid drives. I've got Barracudas, and you shouldn't use those in raid
> either, so it's not immediately serious, but you want something like a
> Seagate IronWolf, or Toshiba N300. I avoid WD for precisely this reason
> - they tend to advertise stuff as suitable when it isn't.
>
> If you can afford it (they're not cheap) you might consider getting four
> decent raid drives, backing up your existing drives, and then fixing it
> when the experts chime in. I'm guessing an "assemble no-bad-blocks
> force" will fix everything, but I'm hesitant to say "go ahead and try it".
>
> Cheers,
> Wol



[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux