On Mon, Apr 25, 2011 at 3:32 PM, Mark Knecht <markknecht@xxxxxxxxx> wrote: > I did a drive check today, first time in months, and found I have a > high mismatch count on my RAID1 root device. What's the best way to > handle getting this cleaned up? > > 1) I'm running some smartctl tests as I write this. > > 2) Do I just do an > > echo repair > > to md126 or do I have to boot a rescue CD before I do that? > > If you need more info please let me know. > > Thanks, > Mark > > c2stable ~ # cat /sys/block/md3/md/mismatch_cnt > 0 > c2stable ~ # cat /sys/block/md6/md/mismatch_cnt > 0 > c2stable ~ # cat /sys/block/md7/md/mismatch_cnt > 0 > c2stable ~ # cat /sys/block/md126/md/mismatch_cnt > 222336 > c2stable ~ # df > Filesystem      1K-blocks   ÂUsed Available Use% Mounted on > /dev/md126      Â51612920 Â26159408 Â22831712 Â54% / > udev           10240    432   Â9808  5% /dev > /dev/md7       389183252 144979184 224434676 Â40% /VirtualMachines > shm          Â6151452     0  6151452  0% /dev/shm > c2stable ~ # cat /proc/mdstat > Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] > md6 : active raid1 sdc6[2] sda6[0] sdb6[1] >   Â247416933 blocks super 1.1 [3/3] [UUU] > > md7 : active raid6 sdc7[2] sda7[0] sdb7[1] sdd2[3] sde2[4] >   Â395387904 blocks super 1.2 level 6, 16k chunk, algorithm 2 [5/5] [UUUUU] > > md3 : active raid6 sdc3[2] sda3[0] sdb3[1] sdd3[3] sde3[4] >   Â157305168 blocks super 1.2 level 6, 16k chunk, algorithm 2 [5/5] [UUUUU] > > md126 : active raid1 sdc5[2] sda5[0] sdb5[1] >   Â52436032 blocks [3/3] [UUU] > > unused devices: <none> > c2stable ~ # > The smartctl tests that I ran (long) completed without error on all 5 drives in the system: SMART Error Log Version: 1 No Errors Logged SMART Self-test log structure revision number 1 Num Test_Description Status Remaining LifeTime(hours) LBA_of_first_error # 1 Extended offline Completed without error 00% 2887 - # 2 Extended offline Completed without error 00% 2046 - So, if I understand correctly the next step I'd do would be something like echo repair >/sys/block/md126/md/sync_action but I'm unclear about the need to do this when mdadm seems to think the RAID is clean: c2stable ~ # mdadm -D /dev/md126 /dev/md126: Version : 0.90 Creation Time : Tue Apr 13 09:02:34 2010 Raid Level : raid1 Array Size : 52436032 (50.01 GiB 53.69 GB) Used Dev Size : 52436032 (50.01 GiB 53.69 GB) Raid Devices : 3 Total Devices : 3 Preferred Minor : 126 Persistence : Superblock is persistent Update Time : Mon Apr 25 18:29:39 2011 State : clean Active Devices : 3 Working Devices : 3 Failed Devices : 0 Spare Devices : 0 UUID : edb0ed65:6e87b20e:dc0d88ba:780ef6a3 Events : 0.248880 Number Major Minor RaidDevice State 0 8 5 0 active sync /dev/sda5 1 8 21 1 active sync /dev/sdb5 2 8 37 2 active sync /dev/sdc5 c2stable ~ # Thanks in advance. Cheers, Mark -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html