On Tue, Jun 11, 2019 at 12:15 PM Colt Boyd <coltboyd@xxxxxxxxx> wrote: > > On Tue, Jun 11, 2019 at 9:16 AM Andreas Klauer > <Andreas.Klauer@xxxxxxxxxxxxxx> wrote: > > > > On Sat, Jun 08, 2019 at 10:47:30AM -0500, Colt Boyd wrote: > > > I’ve also since tried to reassemble it with the following create but > > > the XFS fs is not accessible: > > > 'mdadm --create /dev/md0 --data-offset=1024 --level=6 --raid-devices=5 > > > --chunk=1024K --name=OMV:0 /dev/sdb1 /dev/sde1 /dev/sdc1 /dev/sdd1 > > > /dev/sdf1 --assume-clean --readonly' > > > > Well, all sorts of things can go wrong with a re-create. > > You should be using overlays for such experiments. > > > > https://raid.wiki.kernel.org/index.php/Recovering_a_failed_software_RAID#Making_the_harddisks_read-only_using_an_overlay_file > > Thanks, I am now using overlays, but was not during the first hour > after disaster. I still have the superblock intact on the 6th device > (raid device 5) but the superblocks were overwritten on raid devices > 0-4 when I tried the create. > > > Also, which kernel version are you running? > > 4.19.0-0.bpo.2-amd64 > > > I think there was a RAID6 bug recently in kernel 5.1.3 or so. > > > > https://www.spinics.net/lists/raid/msg62645.html > > > > > /dev/sdg1: > > > Magic : a92b4efc > > > Version : 1.2 > > > Feature Map : 0x1 > > > Array UUID : f8fdf8d4:d173da32:eaa97186:eaf88ded > > > Name : OMV:0 > > > Creation Time : Mon Feb 24 18:19:36 2014 > > > Raid Level : raid6 > > > Raid Devices : 6 > > > > > > Avail Dev Size : 5858529280 (2793.56 GiB 2999.57 GB) > > > Array Size : 11717054464 (11174.25 GiB 11998.26 GB) > > > Used Dev Size : 5858527232 (2793.56 GiB 2999.57 GB) > > > Data Offset : 2048 sectors > > > Super Offset : 8 sectors > > > Unused Space : before=1960 sectors, after=2048 sectors > > > State : clean > > > Device UUID : 92e022c9:ee6fbc26:74da4bcc:5d0e0409 > > > > > > Internal Bitmap : 8 sectors from superblock > > > Update Time : Thu Jun 6 10:24:34 2019 > > > Bad Block Log : 512 entries available at offset 72 sectors > > > Checksum : 8f0d9eb9 - correct > > > Events : 1010399 > > > > > > Layout : left-symmetric > > > Chunk Size : 1024K > > > > > > Device Role : Active device 5 > > > Array State : AAAAAA ('A' == active, '.' == missing, 'R' == replacing) > > > > This already believes to have 6 drives, not in mid-reshape. > > What you created has 5 drives... that's a bit odd. > > > > It could still be normal, metadata for drives that get kicked out > > is no longer updated after all, and I haven't tested it myself... > > > > --examine of the other drives (before re-create) would be interesting. > > If those are not available, maybe syslogs of the original assembly, > > reshape and subsequent recreate. > > The other drives got their superblocks overwritten (with the exception > of raid device 5 that was being added). Here are the applicable > sections from the syslogs. > > Jun 6 10:12:25 OMV1 kernel: [ 2.141772] md/raid:md0: device sde1 > operational as raid disk 1 > Jun 6 10:12:25 OMV1 kernel: [ 2.141774] md/raid:md0: device sdc1 > operational as raid disk 2 > Jun 6 10:12:25 OMV1 kernel: [ 2.141789] md/raid:md0: device sdd1 > operational as raid disk 3 > Jun 6 10:12:25 OMV1 kernel: [ 2.141792] md/raid:md0: device sdf1 > operational as raid disk 4 > Jun 6 10:12:25 OMV1 kernel: [ 2.141796] md/raid:md0: device sdb1 > operational as raid disk 0 > Jun 6 10:12:25 OMV1 kernel: [ 2.142877] md/raid:md0: raid level 6 > active with 5 out of 5 devices, algorithm 2 > Jun 6 10:12:25 OMV1 kernel: [ 2.196783] md0: detected capacity > change from 0 to 8998697828352 > Jun 6 10:12:25 OMV1 kernel: [ 3.885628] XFS (md0): Mounting V4 Filesystem > Jun 6 10:12:25 OMV1 kernel: [ 4.213947] XFS (md0): Ending clean mount > Jun 6 10:12:25 OMV1 kernel: [ 4.220989] XFS (md0): Quotacheck > needed: Please wait. > Jun 6 10:12:25 OMV1 kernel: [ 7.200429] XFS (md0): Quotacheck: Done. > <snip> > Jun 6 10:17:40 OMV1 kernel: [ 321.232145] md: reshape of RAID array md0 > Jun 6 10:17:40 OMV1 systemd[1]: Created slice > system-mdadm\x2dgrow\x2dcontinue.slice. > Jun 6 10:17:40 OMV1 systemd[1]: Started Manage MD Reshape on /dev/md0. > Jun 6 10:17:40 OMV1 systemd[1]: mdadm-grow-continue@md0.service: Main > process exited, code=exited, status=2/INVALIDARGUMENT > Jun 6 10:17:40 OMV1 systemd[1]: mdadm-grow-continue@md0.service: Unit > entered failed state. > Jun 6 10:17:40 OMV1 systemd[1]: mdadm-grow-continue@md0.service: > Failed with result 'exit-code'. > Jun 6 10:18:02 OMV1 monit[1170]: > 'filesystem_media_4e2b0464-e81b-49d9-a520-b574799452f8' space usage > 88.1% matches resource limit [space usage>85.0%] > Jun 6 10:18:32 OMV1 monit[1170]: > 'filesystem_media_4e2b0464-e81b-49d9-a520-b574799452f8' space usage > 88.1% matches resource limit [space usage>85.0%] > Jun 6 10:19:02 OMV1 monit[1170]: > 'filesystem_media_4e2b0464-e81b-49d9-a520-b574799452f8' space usage > 88.1% matches resource limit [space usage>85.0%] > Jun 6 10:19:32 OMV1 monit[1170]: > 'filesystem_media_4e2b0464-e81b-49d9-a520-b574799452f8' space usage > 88.1% matches resource limit [space usage>85.0%] > Jun 6 10:20:02 OMV1 monit[1170]: > 'filesystem_media_4e2b0464-e81b-49d9-a520-b574799452f8' space usage > 88.1% matches resource limit [space usage>85.0%] > Jun 6 10:20:32 OMV1 monit[1170]: > 'filesystem_media_4e2b0464-e81b-49d9-a520-b574799452f8' space usage > 88.1% matches resource limit [space usage>85.0%] > Jun 6 10:21:02 OMV1 monit[1170]: > 'filesystem_media_4e2b0464-e81b-49d9-a520-b574799452f8' space usage > 88.1% matches resource limit [space usage>85.0%] > Jun 6 10:21:32 OMV1 monit[1170]: > 'filesystem_media_4e2b0464-e81b-49d9-a520-b574799452f8' space usage > 88.1% matches resource limit [space usage>85.0%] > Jun 6 10:22:02 OMV1 monit[1170]: > 'filesystem_media_4e2b0464-e81b-49d9-a520-b574799452f8' space usage > 88.1% matches resource limit [space usage>85.0%] > Jun 6 10:22:32 OMV1 monit[1170]: > 'filesystem_media_4e2b0464-e81b-49d9-a520-b574799452f8' space usage > 88.1% matches resource limit [space usage>85.0%] > Jun 6 10:23:02 OMV1 monit[1170]: > 'filesystem_media_4e2b0464-e81b-49d9-a520-b574799452f8' space usage > 88.1% matches resource limit [space usage>85.0%] > Jun 6 10:23:32 OMV1 monit[1170]: > 'filesystem_media_4e2b0464-e81b-49d9-a520-b574799452f8' space usage > 88.1% matches resource limit [space usage>85.0%] > Jun 6 10:24:02 OMV1 monit[1170]: > 'filesystem_media_4e2b0464-e81b-49d9-a520-b574799452f8' space usage > 88.1% matches resource limit [space usage>85.0%] > Jun 6 10:24:28 OMV1 systemd[1]: Unmounting /export/Shared... > Jun 6 10:24:28 OMV1 systemd[1]: Removed slice > system-mdadm\x2dgrow\x2dcontinue.slice. > <snip> - server shutting down > Jun 6 10:24:28 OMV1 systemd[1]: openmediavault-engined.service: > Killing process 1214 (omv-engined) with signal SIGKILL. There is also these: Jun 6 10:39:18 OMV1 kernel: [ 120.022361] md0: detected capacity change from 11998263771136 to 0 Jun 6 10:39:18 OMV1 kernel: [ 120.022371] md: md0 stopped. Jun 6 10:40:39 OMV1 kernel: [ 200.948894] md: md0 stopped. Jun 6 10:40:39 OMV1 kernel: [ 201.692933] md: md0 stopped. Jun 6 10:41:11 OMV1 kernel: [ 233.338779] md: md0 stopped. Jun 6 10:43:11 OMV1 kernel: [ 353.828080] md: md0 stopped. Jun 6 10:43:55 OMV1 kernel: [ 397.202545] md: md0 stopped. Jun 6 10:44:12 OMV1 kernel: [ 414.195911] md: md0 stopped. Jun 6 10:44:15 OMV1 kernel: [ 417.271865] md: md0 stopped. Jun 6 10:44:45 OMV1 kernel: [ 447.195016] md: md0 stopped. Jun 6 10:44:47 OMV1 kernel: [ 449.499203] md: md0 stopped. Jun 6 10:44:47 OMV1 kernel: [ 449.511291] md/raid:md0: device sdb1 operational as raid disk 0 Jun 6 10:44:47 OMV1 kernel: [ 449.511294] md/raid:md0: device sdf1 operational as raid disk 4 Jun 6 10:44:47 OMV1 kernel: [ 449.511297] md/raid:md0: device sdd1 operational as raid disk 3 Jun 6 10:44:47 OMV1 kernel: [ 449.511299] md/raid:md0: device sdc1 operational as raid disk 2 Jun 6 10:44:47 OMV1 kernel: [ 449.511302] md/raid:md0: device sde1 operational as raid disk 1 Jun 6 10:44:47 OMV1 kernel: [ 449.512553] md/raid:md0: raid level 6 active with 5 out of 6 devices, algorithm 2 Jun 6 10:44:47 OMV1 kernel: [ 449.554738] md0: detected capacity change from 0 to 11998263771136 Jun 6 10:44:48 OMV1 postfix/smtp[2514]: 9672F6B4: replace: header Subject: DegradedArray event on /dev/md0:OMV1: Subject: [OMV1.veldanet.local] DegradedArray event on /dev/md0:OMV1 Jun 6 10:46:25 OMV1 kernel: [ 547.047912] XFS (md0): Mounting V4 Filesystem Jun 6 10:46:28 OMV1 kernel: [ 550.226215] XFS (md0): Log inconsistent (didn't find previous header) Jun 6 10:46:28 OMV1 kernel: [ 550.226224] XFS (md0): failed to find log head Jun 6 10:46:28 OMV1 kernel: [ 550.226227] XFS (md0): log mount/recovery failed: error -5 Jun 6 10:46:28 OMV1 kernel: [ 550.226264] XFS (md0): log mount failed Jun 6 10:48:29 OMV1 kernel: [ 671.505718] md0: detected capacity change from 11998263771136 to 0 Jun 6 10:48:29 OMV1 kernel: [ 671.505731] md: md0 stopped. Jun 6 10:48:53 OMV1 kernel: [ 695.670961] md: md0 stopped. Jun 6 10:48:53 OMV1 kernel: [ 695.682191] md/raid:md0: device sdb1 operational as raid disk 0 Jun 6 10:48:53 OMV1 kernel: [ 695.682193] md/raid:md0: device sdg1 operational as raid disk 5 Jun 6 10:48:53 OMV1 kernel: [ 695.682195] md/raid:md0: device sdf1 operational as raid disk 4 Jun 6 10:48:53 OMV1 kernel: [ 695.682197] md/raid:md0: device sdd1 operational as raid disk 3 Jun 6 10:48:53 OMV1 kernel: [ 695.682199] md/raid:md0: device sdc1 operational as raid disk 2 Jun 6 10:48:53 OMV1 kernel: [ 695.682200] md/raid:md0: device sde1 operational as raid disk 1 Jun 6 10:48:53 OMV1 kernel: [ 695.683023] md/raid:md0: raid level 6 active with 6 out of 6 devices, algorithm 2 Jun 6 10:48:53 OMV1 kernel: [ 695.760209] md0: detected capacity change from 0 to 11998263771136 Jun 6 10:51:24 OMV1 kernel: [ 846.532095] XFS (md0): Mounting V4 Filesystem Jun 6 10:51:27 OMV1 kernel: [ 849.779906] XFS (md0): Log inconsistent (didn't find previous header) Jun 6 10:51:27 OMV1 kernel: [ 849.779924] XFS (md0): failed to find log head Jun 6 10:51:27 OMV1 kernel: [ 849.779931] XFS (md0): log mount/recovery failed: error -5 Jun 6 10:51:27 OMV1 kernel: [ 849.780000] XFS (md0): log mount failed Jun 6 10:52:22 OMV1 systemd[1]: Starting Cleanup of Temporary Directories... Jun 6 10:52:22 OMV1 systemd[1]: Started Cleanup of Temporary Directories. Jun 6 10:52:52 OMV1 kernel: [ 934.000472] md0: detected capacity change from 11998263771136 to 0 Jun 6 10:52:52 OMV1 kernel: [ 934.000480] md: md0 stopped. > > > Otherwise you have to look at the raw data (or try blindly) > > to figure out the data layout. > > > > Please use overlays for experiments... > > > > Good luck. > > I am now, if only I was to start with this may be easier. Anyway to > rebuild superblocks from the remaing drive and / or the backup file? > And if so, would that be better? > > Thanks, > Colt > > > Regards > > Andreas Klauer