Thank you. I replaced 2 month ago one disk of the 3 disk raid 5 and just collected some informations from it: /dev/sde6: Magic : a92b4efc Version : 1.2 Feature Map : 0x1 Array UUID : 342ec726:3804270d:5917dd5f:c24883a9 Name : TS-XLB6C:2 Creation Time : Fri Dec 23 17:58:59 2011 Raid Level : raid5 Raid Devices : 3 Avail Dev Size : 1923497952 (917.20 GiB 984.83 GB) Array Size : 1923496960 (1834.39 GiB 1969.66 GB) Used Dev Size : 1923496960 (917.19 GiB 984.83 GB) Data Offset : 2048 sectors Super Offset : 8 sectors Unused Space : before=1968 sectors, after=992 sectors State : active Device UUID : d27a69d0:456f3704:8e17ac75:78939886 Internal Bitmap : 8 sectors from superblock Update Time : Wed Jul 27 19:08:08 2016 Checksum : de9dbd10 - correct Events : 11543 Layout : left-symmetric Chunk Size : 512K Device Role : Active device 0 Array State : AAA ('A' == active, '.' == missing, 'R' == replacing) So it was a 3 disk raid5 with 512K chunk size. Anyway this disk should be of no help as the "offset" i too big but it was of help to see how the geometry is. Using fdisk on that "old" disk also confirms that the disk layout is identically to the 3 disk i had in place: Disk /dev/sde: 931.5 GiB, 1000204886016 bytes, 1953525168 sectors Units: sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disklabel type: gpt Disk identifier: B84660EA-8691-4C7C-B914-DA2769BDD6D7 Device Start End Sectors Size Type /dev/sde1 2048 2002943 2000896 977M Microsoft basic data /dev/sde2 2002944 12003327 10000384 4.8G Microsoft basic data /dev/sde3 12003328 12005375 2048 1M Microsoft basic data /dev/sde4 12005376 12007423 2048 1M Microsoft basic data /dev/sde5 12007424 14008319 2000896 977M Microsoft basic data /dev/sde6 14008320 1937508319 1923500000 917.2G Microsoft basic data So on the partitions size nothing has changed. So i will now only work with overlay as i do not have that much space available to copy all disks. What would be the next steps? just create a new raid 5 array with sd[a-c]6 ? Thank you. Simon 2016-09-21 17:38 GMT+02:00 Chris Murphy <lists@xxxxxxxxxxxxxxxxx>: > On Wed, Sep 21, 2016 at 4:39 AM, Simon Becks <beckssimon5@xxxxxxxxx> wrote: >> Dear Developers & Gurus & Gods, >> >> i had a 3 disk software raid 5 (mdadm) on a buffalo terrastation. By >> accident I reset the raid and the NAS put on a xfs filesystem on each >> of the 3 partitions. >> >> sda6 sdb6 and sdc6 have been the raid5 member partitions. >> >> Now sda6 sdb6 and sdc6 only contain a xfs filesystem with some empty >> default folder structure - my NAS created during the "reset". > > OK that doesn't really make sense, if it's going to do a reset I'd > expect it to format all the other partitions and even repartition > those drives. Why does it format just the sixth partition on these > three drives? The point is, you need to make sure this "new" sda6 is > really exactly the same as the old one. As in, the same start and end > sector values. If this is a different partition scheme on any of the > drives, you have to fix that problem first. > > >> >> mdadm --examine /dev/sda6 >> mdadm: No md superblock detected on /dev/sda6 >> mdadm --examine /dev/sdb6 >> mdadm: No md superblock detected on /dev/sdb6 >> mdadm --examine /dev/sdc6 >> mdadm: No md superblock detected on /dev/sdc6 > > mkfs.xfs doesn't write that much metadata but it does write a lot of > zeros, about 60MB of writes per mkfs depending on how many AG's are > created. So no matter what the resulting array is going to have about > 200MB of data loss spread around. It's hard to say what that will have > stepped on, if you're lucky it'll be only data. If you're unluckly it > will have hit the file system in a way that'll make it difficult to > extract your data. > > So no matter what, this is now a scraping operation. And that will > take some iteration. Invariably it will take less time to just create > a new array and restore from backup. If you don't have a backup for > this data, then it's not important data. Either way, it's not worth > your time. > > However... > > To make it possible to iterate mdadm metadata version, and chunk size, > and device order without more damage, you need to work on files of > these partitions, or use an overlay. > https://raid.wiki.kernel.org/index.php/Recovering_a_failed_software_RAID#Making_the_harddisks_read-only_using_an_overlay_file > > The overlay file option is better because you can iterate and throw > them away and quickly start again. If you dd the partitions to files, > and you directly change those files, to start over you have to dd > those partitions yet again. So you're probably going to want the > overlay option no matter what. > > Each iteration will produce an assembled array, but only one > combination will produce an array that's your array. And even that > array might not be mountable due to mkfs damage. So you'll need some > tests to find out whether you have a file system on that array or if > it's just garbage. fsck -n *might* recognize the filesystem even if > it's badly damaged, and tell you how badly damaged it is, without > trying to fix it. You're almost certainly best off not fixing it for > starters, and just mounting it read only and getting off as much data > as you can elsewhere. i.e. making the backup you should already have. > > > > > >> >> Disk /dev/sdc: 931.5 GiB, 1000204886016 bytes, 1953525168 sectors >> Units: sectors of 1 * 512 = 512 bytes >> Sector size (logical/physical): 512 bytes / 4096 bytes >> I/O size (minimum/optimal): 4096 bytes / 4096 bytes >> Disklabel type: gpt >> Disk identifier: FAB30D96-11C4-477E-ADAA-9448A087E124 >> >> Device Start End Sectors Size Type >> /dev/sdc1 2048 2002943 2000896 977M Microsoft basic data >> /dev/sdc2 2002944 12003327 10000384 4.8G Microsoft basic data >> /dev/sdc3 12003328 12005375 2048 1M Microsoft basic data >> /dev/sdc4 12005376 12007423 2048 1M Microsoft basic data >> /dev/sdc5 12007424 14008319 2000896 977M Microsoft basic data >> /dev/sdc6 14008320 1937508319 1923500000 917.2G Microsoft basic data >> >> XFS-Log attached for reference. >> >> Am I screwed or is there a chance to recreate the raid with the 3 >> disks end up with the raid and the filesystem i had before? > > It's pretty unlikely you'll totally avoid data loss, it's just matter > of what damage has happened and that's not knowable in advance. You'll > just have to try it out. > > If the file system can't be mounted ro; if the fsck can't make it > "good enough" to mount ro; then you will want to take a look at > testdisk which can scrape the array (not the individual drives) for > file signatures. So long as the files are in contiguous blocks, it can > enable you to scrape off things like photos and documents. Smaller > files tend to recover better than big files. > > If that also fails, well yeah you can try test disk pointed at > individual partitions or the whole drives and see what it finds. If > the chunk size is 512KiB that sorta improves the chances you'll get > some files but they'll definitely only be small files that'll be > recognized. Any file broken up by raid striping will be on some other > drive. So it's a huge jigsaw puzzle, which is why raid is not a > backup, etc. > > > -- > Chris Murphy -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html