Re: Need help with degraded raid 5

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



William Morgan <therealbrewer@xxxxxxxxx> 于2020年3月5日周四 上午1:33写道:
>
> Hello,
>
> I'm working with a 4 disk raid 5. In the past I have experienced a
> problem that resulted in the array being set to "inactive", but with
> some guidance from the list, I was able to rebuild with no loss of
> data. Recently I have a slightly different situation where one disk
> was "removed" and marked as "spare", so the array is still active, but
> degraded.
>
> I've been monitoring the array, and I got a "Fail event" notification
> right after a power blip which showed this mdstat:
>
> Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5]
> [raid4] [raid10]
> md0 : active raid5 sdm1[4](F) sdj1[0] sdk1[1] sdl1[2]
>       23441679360 blocks super 1.2 level 5, 64k chunk, algorithm 2 [4/3] [UUU_]
>       bitmap: 0/59 pages [0KB], 65536KB chunk
>
> unused devices: <none>
>
> A little while later I got a "DegradedArray event" notification with
> the following mdstat:
>
> Personalities : [raid6] [raid5] [raid4] [linear] [multipath] [raid0]
> [raid1] [raid10]
> md0 : active raid5 sdl1[4] sdj1[1] sdk1[2] sdi1[0]
>       23441679360 blocks super 1.2 level 5, 64k chunk, algorithm 2 [4/3] [UUU_]
>       [>....................]  recovery =  0.0% (12600/7813893120)
> finish=113621.8min speed=1145K/sec
>       bitmap: 2/59 pages [8KB], 65536KB chunk
>
> unused devices: <none>
>
> which seemed to imply that sdl was being rebuilt, but then I got
> another "DegradedArray event" notification with this:
>
> Personalities : [raid6] [raid5] [raid4] [linear] [multipath] [raid0]
> [raid1] [raid10]
> md0 : active raid5 sdl1[4](S) sdj1[1] sdk1[2] sdi1[0]
>       23441679360 blocks super 1.2 level 5, 64k chunk, algorithm 2 [4/3] [UUU_]
>       bitmap: 2/59 pages [8KB], 65536KB chunk
>
> unused devices: <none>
>
>
> I don't think anything is really wrong with the removed disk however.
> So assuming I've got backups, what do I need to do to reinsert the
> disk and get the array back to a normal state? Or does that disk's
> data need to be completely rebuilt? And how do I initiate that?
>
> I'm using the latest mdadm and a very recent kernel. Currently I get this:
>
> bill@bill-desk:~$ sudo mdadm --detail /dev/md0
> /dev/md0:
>            Version : 1.2
>      Creation Time : Sat Sep 22 19:10:10 2018
>         Raid Level : raid5
>         Array Size : 23441679360 (22355.73 GiB 24004.28 GB)
>      Used Dev Size : 7813893120 (7451.91 GiB 8001.43 GB)
>       Raid Devices : 4
>      Total Devices : 4
>        Persistence : Superblock is persistent
>
>      Intent Bitmap : Internal
>
>        Update Time : Mon Mar  2 17:41:32 2020
>              State : clean, degraded
>     Active Devices : 3
>    Working Devices : 4
>     Failed Devices : 0
>      Spare Devices : 1
>
>             Layout : left-symmetric
>         Chunk Size : 64K
>
> Consistency Policy : bitmap
>
>               Name : bill-desk:0  (local to host bill-desk)
>               UUID : 06ad8de5:3a7a15ad:88116f44:fcdee150
>             Events : 10407
>
>     Number   Major   Minor   RaidDevice State
>        0       8      129        0      active sync   /dev/sdi1
>        1       8      145        1      active sync   /dev/sdj1
>        2       8      161        2      active sync   /dev/sdk1
>        -       0        0        3      removed
>
>        4       8      177        -      spare   /dev/sdl1

"mdadm /dev/md0 -a /dev/sdl1"  should work for you to add the disk
back to array, maybe you can check first with "mdadm -E /dev/sdl1" to
make sure.

Regards,
Jack Wang




[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux