Vreme: 12/19/2011 10:58 PM, Paul Heinlein piše: > On Mon, 19 Dec 2011, aurfalien@xxxxxxxxx wrote: > >>> I'm interested to know if you used mdadm to fail and remove the bad >>> disk from the array when it first started acting up. >> >> No, I should have but left it alone. >> >> I know, my bad. > > I was merely interested. > > Recently I had a RAID-1 device get marked as bad, but I couldn't see > any SMART errors. So I failed, removed, and then re-added the device. > It worked for about a week, then it failed again, but time the SMART > errors were obvious. > > I'd ordered a new drive at the first failure, so I was ready when it > failed the second time. > > I guess the point is that I've seen "bad" drives go "good" again, at > least for short periods of time. > Les, I will reply to your mail here. This would better explain my troubles. I've had to manually re-add them few times, but I've had little experience and I thought these things happen, never gave it much thought. I am still not 100% on what actually happened, but my guess would be that since boot partition was active only on one drive, and that one was creating problems, I went with line of easier resistance and just patch it up. I am going to use C6 ability to boot from full raid partition, and maybe even add IDE DOM module for boot partition. I am now good and watching like a hawk. -- Ljubomir Ljubojevic (Love is in the Air) PL Computers Serbia, Europe Google is the Mother, Google is the Father, and traceroute is your trusty Spiderman... StarOS, Mikrotik and CentOS/RHEL/Linux consultant _______________________________________________ CentOS mailing list CentOS@xxxxxxxxxx http://lists.centos.org/mailman/listinfo/centos