Re: What the heck happened to my array? (No apparent data loss).

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



what kernel version? more informations about your linux box?

2011/4/3 Brad Campbell <lists2009@xxxxxxxxxxxxxxx>:
> 2.6.38.2
> x86_64
> 10 x 1TB SATA drives in a single RAID-6
>
> Here is the chain of events.
>
> Saturday morning I started a reshape on a 10 element RAID-6. Simply changing
> the Chunk size from 512k to 64k. This was going to take about 4.5 days
> according to the initial estimates.
>
> I then went away for the weekend and came home to a wedged array.
> Here is the chain of events that caused it.
>
> This occurred about 1 minute after my scheduled morning SMART long (it is
> Sunday after all) began.
>
> Apr  3 03:19:08 srv kernel: [288180.455339] sd 0:0:12:0: [sdd] Unhandled
> error code
> Apr  3 03:19:08 srv kernel: [288180.455359] sd 0:0:12:0: [sdd]  Result:
> hostbyte=0x04 driverbyte=0x00
> Apr  3 03:19:08 srv kernel: [288180.455377] sd 0:0:12:0: [sdd] CDB:
> cdb[0]=0x2a: 2a 00 00 00 00 08 00 00 02 00
> Apr  3 03:19:08 srv kernel: [288180.455415] end_request: I/O error, dev sdd,
> sector 8
> Apr  3 03:19:08 srv kernel: [288180.455449] end_request: I/O error, dev sdd,
> sector 8
> Apr  3 03:19:08 srv kernel: [288180.455462] md: super_written gets error=-5,
> uptodate=0
> Apr  3 03:19:08 srv kernel: [288180.455477] md/raid:md0: Disk failure on
> sdd, disabling device.
> Apr  3 03:19:08 srv kernel: [288180.455480] md/raid:md0: Operation
> continuing on 9 devices.
> Apr  3 03:19:08 srv kernel: [288180.472914] md: md0: reshape done.
> Apr  3 03:19:08 srv kernel: [288180.472983] md: delaying data-check of md5
> until md3 has finished (they share one or more physical units)
> Apr  3 03:19:08 srv kernel: [288180.473002] md: delaying data-check of md4
> until md6 has finished (they share one or more physical units)
> Apr  3 03:19:08 srv kernel: [288180.473030] md: delaying data-check of md6
> until md5 has finished (they share one or more physical units)
> Apr  3 03:19:08 srv kernel: [288180.473047] md: delaying data-check of md3
> until md1 has finished (they share one or more physical units)
> Apr  3 03:19:08 srv kernel: [288180.551450] md: reshape of RAID array md0
> Apr  3 03:19:08 srv kernel: [288180.551468] md: minimum _guaranteed_ speed:
> 200000 KB/sec/disk.
> Apr  3 03:19:08 srv kernel: [288180.551483] md: using maximum available idle
> IO bandwidth (but not more than 200000 KB/sec) for reshape.
> Apr  3 03:19:08 srv kernel: [288180.551514] md: using 128k window, over a
> total of 976759808 blocks.
> Apr  3 03:19:08 srv kernel: [288180.620089] sd 0:0:12:0: [sdd] Synchronizing
> SCSI cache
> Apr  3 03:19:08 srv mdadm[4803]: RebuildFinished event detected on md device
> /dev/md0
> Apr  3 03:19:08 srv mdadm[4803]: Fail event detected on md device /dev/md0,
> component device /dev/sdd
> Apr  3 03:19:08 srv mdadm[4803]: RebuildStarted event detected on md device
> /dev/md0
> Apr  3 03:19:10 srv kernel: [288182.614918] scsi 0:0:12:0: Direct-Access
> ATA      MAXTOR STM310003 MX1A PQ: 0 ANSI: 5
> Apr  3 03:19:10 srv kernel: [288182.615312] sd 0:0:12:0: Attached scsi
> generic sg3 type 0
> Apr  3 03:19:10 srv kernel: [288182.618262] sd 0:0:12:0: [sdq] 1953525168
> 512-byte logical blocks: (1.00 TB/931 GiB)
> Apr  3 03:19:10 srv kernel: [288182.736998] sd 0:0:12:0: [sdq] Write Protect
> is off
> Apr  3 03:19:10 srv kernel: [288182.737019] sd 0:0:12:0: [sdq] Mode Sense:
> 73 00 00 08
> Apr  3 03:19:10 srv kernel: [288182.740521] sd 0:0:12:0: [sdq] Write cache:
> enabled, read cache: enabled, doesn't support DPO or FUA
> Apr  3 03:19:10 srv kernel: [288182.848999]  sdq: unknown partition table
> Apr  3 03:19:10 srv ata_id[28453]: HDIO_GET_IDENTITY failed for '/dev/sdq'
> Apr  3 03:19:10 srv kernel: [288182.970091] sd 0:0:12:0: [sdq] Attached SCSI
> disk
> Apr  3 03:20:01 srv /USR/SBIN/CRON[28624]: (brad) CMD ([ -z "`/usr/bin/pgrep
> -u brad collect`" ] && /usr/bin/screen -X -S brad-bot screen
> /home/brad/bin/collect-thermostat)
> Apr  3 03:20:01 srv /USR/SBIN/CRON[28625]: (root) CMD ([ -z `/usr/bin/pgrep
> -u root keepalive` ] && /home/brad/bin/launch-keepalive)
> Apr  3 03:20:01 srv /USR/SBIN/CRON[28626]: (brad) CMD ([ -z "`screen -list |
> grep brad-bot`" ] && /home/brad/bin/botstart)
> Apr  3 03:20:01 srv /USR/SBIN/CRON[28628]: (root) CMD (if [ -x /usr/bin/mrtg
> ] && [ -r /etc/mrtg.cfg ]; then mkdir -p /var/log/mrtg ; env LANG=C
> /usr/bin/mrtg /etc/mrtg.cfg 2>&1 | tee -a /var/log/mrtg/mrtg.log ; fi)
> Apr  3 03:20:01 srv /USR/SBIN/CRON[28627]: (brad) CMD
> (/home/brad/rrd/rrd-create-graphs)
> Apr  3 03:20:01 srv /USR/SBIN/CRON[28590]: (CRON) error (grandchild #28625
> failed with exit status 1)
> Apr  3 03:20:01 srv /USR/SBIN/CRON[28589]: (CRON) error (grandchild #28626
> failed with exit status 1)
> Apr  3 03:20:01 srv /USR/SBIN/CRON[28587]: (CRON) error (grandchild #28624
> failed with exit status 1)
> Apr  3 03:22:10 srv kernel: [288363.070094] INFO: task jbd2/md0-8:2647
> blocked for more than 120 seconds.
> Apr  3 03:22:10 srv kernel: [288363.070114] "echo 0 >
> /proc/sys/kernel/hung_task_timeout_secs" disables this message.
> Apr  3 03:22:10 srv kernel: [288363.070132] jbd2/md0-8      D
> ffff88041aa52948     0  2647      2 0x00000000
> Apr  3 03:22:10 srv kernel: [288363.070154]  ffff88041aa526f0
> 0000000000000046 0000000000000000 ffff8804196769b0
> Apr  3 03:22:10 srv kernel: [288363.070178]  0000000000011180
> ffff88041bdc5fd8 0000000000004000 0000000000011180
> Apr  3 03:22:10 srv kernel: [288363.070201]  ffff88041bdc4010
> ffff88041aa52950 ffff88041bdc5fd8 ffff88041aa52948
> Apr  3 03:22:10 srv kernel: [288363.070224] Call Trace:
> Apr  3 03:22:10 srv kernel: [288363.070246]  [<ffffffff8104e4c6>] ?
> queue_work_on+0x16/0x20
> Apr  3 03:22:10 srv kernel: [288363.070266]  [<ffffffff812e6bfd>] ?
> md_write_start+0xad/0x190
> Apr  3 03:22:10 srv kernel: [288363.070283]  [<ffffffff81052b90>] ?
> autoremove_wake_function+0x0/0x30
> Apr  3 03:22:10 srv kernel: [288363.070299]  [<ffffffff812e16f5>] ?
> make_request+0x35/0x600
> Apr  3 03:22:10 srv kernel: [288363.070317]  [<ffffffff8108463b>] ?
> __alloc_pages_nodemask+0x10b/0x810
> Apr  3 03:22:10 srv kernel: [288363.070335]  [<ffffffff81142042>] ?
> T.1015+0x32/0x90
> Apr  3 03:22:10 srv kernel: [288363.070350]  [<ffffffff812e6a24>] ?
> md_make_request+0xd4/0x200
> Apr  3 03:22:10 srv kernel: [288363.070366]  [<ffffffff81142218>] ?
> ext4_map_blocks+0x178/0x210
> Apr  3 03:22:10 srv kernel: [288363.070382]  [<ffffffff811b6e84>] ?
> generic_make_request+0x144/0x2f0
> Apr  3 03:22:10 srv kernel: [288363.070397]  [<ffffffff8116e89d>] ?
> jbd2_journal_file_buffer+0x3d/0x70
> Apr  3 03:22:10 srv kernel: [288363.070413]  [<ffffffff811b708c>] ?
> submit_bio+0x5c/0xd0
> Apr  3 03:22:10 srv kernel: [288363.070430]  [<ffffffff810e61d5>] ?
> submit_bh+0xe5/0x120
> Apr  3 03:22:10 srv kernel: [288363.070445]  [<ffffffff811709b1>] ?
> jbd2_journal_commit_transaction+0x441/0x1180
> Apr  3 03:22:10 srv kernel: [288363.070466]  [<ffffffff81044893>] ?
> lock_timer_base+0x33/0x70
> Apr  3 03:22:10 srv kernel: [288363.070480]  [<ffffffff81052b90>] ?
> autoremove_wake_function+0x0/0x30
> Apr  3 03:22:10 srv kernel: [288363.070498]  [<ffffffff81174871>] ?
> kjournald2+0xb1/0x1e0
> Apr  3 03:22:10 srv kernel: [288363.070511]  [<ffffffff81052b90>] ?
> autoremove_wake_function+0x0/0x30
> Apr  3 03:22:10 srv kernel: [288363.070527]  [<ffffffff811747c0>] ?
> kjournald2+0x0/0x1e0
> Apr  3 03:22:10 srv kernel: [288363.070544]  [<ffffffff811747c0>] ?
> kjournald2+0x0/0x1e0
> Apr  3 03:22:10 srv kernel: [288363.070557]  [<ffffffff81052716>] ?
> kthread+0x96/0xa0
> Apr  3 03:22:10 srv kernel: [288363.070573]  [<ffffffff81003154>] ?
> kernel_thread_helper+0x4/0x10
> Apr  3 03:22:10 srv kernel: [288363.070588]  [<ffffffff81052680>] ?
> kthread+0x0/0xa0
> Apr  3 03:22:10 srv kernel: [288363.070602]  [<ffffffff81003150>] ?
> kernel_thread_helper+0x0/0x10
>
> So apparently sdd suffered an unknown failure (it happens) and the array
> kicked it out (as it should). But 120 seconds later all tasks accessing that
> array trigger their 120 second hangcheck warning and are all suck in the D
> state.
>
> At the time the array was 12.1% of the way through a reshape. I had to
> reboot the machine to get it back up and it's now continuing the reshape on
> 9 drives.
>
> brad@srv:~$ cat /proc/mdstat
> Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4]
> md0 : active raid6 sdc[0] sdh[9] sda[8] sde[7] sdg[5] sdb[4] sdf[3] sdm[2]
> sdl[1]
>      7814078464 blocks super 1.2 level 6, 512k chunk, algorithm 2 [10/9]
> [UUUUUU_UUU]
>      [===>.................]  reshape = 16.5% (162091008/976759808)
> finish=5778.6min speed=2349K/sec
>
>
>
> To make matters more confusing the other arrays on the machine were in the
> middle of their "Debians first Sunday of every month" "check" scrub.
>
> I have the full syslog and can probably procure any other information that
> might be useful. I don't think I've lost any data, the machine continued
> reshaping and we're all moving along nicely. I just wanted to report it and
> offer assistance in diagnosing it should that be requested.
>
> Regards,
> Brad
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>



-- 
Roberto Spadim
Spadim Technology / SPAEmpresarial
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux