Re: All disk ar reported as spare disks

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Phil & Wol  and everyone else.

I just wanted to say a big thank you, --assemble --force solved the
problem and I got the raid running again :-D

And now after a fsck I am copying all the data to my new raid1.
And what I can see so far I don't seem to have lost anything :-)

The new disks were purchased since before (WD Red NAS 10TB), but
fortunately they have support for SCT "Error Recovery Control" ,
"Feature Control" , "Data Table".
And "Recovery Control" is set to 70.70, just as mentioned on:
https://raid.wiki.kernel.org/index.php/Timeout_Mismatch

But I will still put that script into the startup of my new server.


Once again, a big thanks for all the help!

Best regards Rickard

Den fre 31 jan. 2020 kl 14:57 skrev Phil Turmel <philip@xxxxxxxxxx>:
>
> Hi Rickard,
>
> Good report.
>
> On 1/30/20 6:48 PM, Rickard Svensson wrote:
> > Hello
> >
> > Excuse me for asking again.
> >
> > But this is a simpler(?) follow-up question to:
> > https://marc.info/?t=157895855400002&r=1&w=2
> >
> > In short summary. I had a raid 1 0, there were too many write errors
> > on one disk (I call it DiskError1), which I did not notice, and then
> > two days later the same problem on another disk (I call it
> > DiskError2).
> >
> > I got good help here, and copy the disk portions of the 2 working
> > disks as well as disk DiskError2 with ddrescue to new disks.
> > Later I'll create a new raid 1, so I don't plan reuse the same raid 1 0 again.
> >
> >
> > My questions:
> > 1) I haven't copied the disk DiskError1, because it is older data, and
> > it shouldn't be needed.   Or is it better to add that one as well?
> >
> > 2) Everything looks pretty good :)
> > But all disk ar reported as spare disks in /proc/mdstat
> > A assume that is because "Events" count is not the same. It is same on
> > the good disks(2864) but not DiskError2 (2719).
>
> No, the array isn't running, so /proc/mdstat isn't complete.  Your three
> disks all have proper "Active device" roles per --examine.
>
> > I have been looking how I can "force add" disk DiskError2, use
> > "--force" or "--- zero-superblock"?
>
> Neither --add nor --zero-superblock is appropriate.  They will break
> your otherwise very good condition.
>
> > But would prefer to avoid making a mistake now,   what has the
> > greatest chance of being right :)
>
> First, ensure you do not have a timeout mismatch as evidenced in your
> original thread's smartctl output.  The wiki has some advice.  Hopefully
> your new drives are "NAS" rated and you need no special action.
>
> Then you should simply use --assemble --force with those three devices.
>
> That should get you running degraded.  Then immediately backup the most
> valuable data in the array before doing anything else.
>
> Finally, --add a fourth device and let your raid rebuild its redundancy.
>
> When all is safe, consider converting to a more durable redundancy
> setup, like raid6, or raid10,near=3.
>
> Phil



[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux