I am not sure what the problem is then. If it is growing your raid10,n2 to a raid10,n3 - which may not be doable with mdadm grow - then you could try out creating a raid10,n3 array on your new disk, with only 1 disk. copy the stuff, and then adding the 2 old drives. I think it is a insight that raid1 only - mostly - performs out of one disk, regardslessly of how many disks you have. I have used multi-disk raid1 to have redundancy for booting, so some use can be found. Best regards Keld On Wed, Nov 02, 2016 at 01:56:02PM -0600, Robert LeBlanc wrote: > Yes, we can have any number of disks in a RAID1 (we currently have > three), but reads only ever come from the first drive. We want to move > to RAID10 so that all drives can service reads and provide performance > as well. We just need the option to grow a RAID10 like we can with > RAID1. We don't need the "extra" space by growing a RAID10 without > changing '-p n'. Basically, we want to be super paranoid with several > identical copies of the data and get extra read performance. We know > that we will be limited in write performance which is kind of counter > intuitive for RAID10, but our workload is OK with that. > > I hope that makes sense. I could provide some test data on n-disk > RAID1, but my experience says there is little value to it, it is very > similar to 2 disk RAID1. If I have time, I'll supply something. > ---------------- > Robert LeBlanc > PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 > > > On Wed, Nov 2, 2016 at 1:48 PM, <keld@xxxxxxxxxx> wrote: > > If you want all your disks to be identical, then you only can chose between > > raid1 and raid10 near. I believe then the raid10 near is the better layout, as some > > stats say you will have better random performance. I don't know why. Probably a driver issue > > I believe you can have raid1 in a 3-disk solution. You should try it out, and then please report the > > stats back to the list, then I will add it to the wiki (it seems unacessibe at the moment, tho) > > > > best regards > > Keld > > > > On Wed, Nov 02, 2016 at 01:02:29PM -0600, Robert LeBlanc wrote: > >> My boss basically wants RAID1 with all drives able to be read from. He > >> has a requirement to have all the drives identical (minus the > >> superblock) hence the 'near' option being used. From my rudimentary > >> tests, sequential reds do seem to use all drives, but random reads > >> don't. I wonder what logic is preventing the spreading out of random > >> workloads for 'near'. 'far' is using all disks in random read and > >> getting better performance on both random and sequential. I'm testing > >> loopbacks on an NVME drive so seek latency should not be a major > >> concern. > >> ---------------- > >> Robert LeBlanc > >> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 > >> > >> > >> On Wed, Nov 2, 2016 at 12:19 PM, <keld@xxxxxxxxxx> wrote: > >> > There is some speed limits om raid10,n2 as also reported in > >> > https://raid.wiki.kernel.org/index.php/Performance > >> > > >> > f you want speed, I suggest you use raid10,f2. > >> > > >> > Unfortunatlely you cannot grow "far" layouts, Neil says it is too complicated. > >> > > >> > But in your case you should be able to disable one of your raid10,N2 drives, > >> > then build a raid10,n2 array for 3 disks, but only with the disk you removed from > >> > your N2 disk plus your new disk. Then you can copy the contents of the remaining > >> > old disk to the new "far" disk, and when complete, add the old raid10,n2 disk to the > >> > new Far raid, with 3 disks. This should give you about 3 times the speed > >> > of your old raid10,n2 array. > >> > > >> > Best regards > >> > keld > >> > > >> > > >> > > >> > On Wed, Nov 02, 2016 at 11:59:25AM -0600, Robert LeBlanc wrote: > >> >> We would like to add read performance to our RAID10 volume by adding > >> >> another drive (we don't care about space), so I did the following test > >> >> with poor results. > >> >> > >> >> # mdadm --create /dev/md13 --level 10 --run --assume-clean -p n2 > >> >> --raid-devices 2 /dev/loop{2..3} > >> >> mdadm: /dev/loop2 appears to be part of a raid array: > >> >> level=raid10 devices=3 ctime=Wed Nov 2 11:25:22 2016 > >> >> mdadm: /dev/loop3 appears to be part of a raid array: > >> >> level=raid10 devices=3 ctime=Wed Nov 2 11:25:22 2016 > >> >> mdadm: Defaulting to version 1.2 metadata > >> >> mdadm: array /dev/md13 started. > >> >> > >> >> # mdadm --detail /dev/md13 > >> >> /dev/md13: > >> >> Version : 1.2 > >> >> Creation Time : Wed Nov 2 11:47:48 2016 > >> >> Raid Level : raid10 > >> >> Array Size : 10477568 (9.99 GiB 10.73 GB) > >> >> Used Dev Size : 10477568 (9.99 GiB 10.73 GB) > >> >> Raid Devices : 2 > >> >> Total Devices : 2 > >> >> Persistence : Superblock is persistent > >> >> > >> >> Update Time : Wed Nov 2 11:47:48 2016 > >> >> State : clean > >> >> Active Devices : 2 > >> >> Working Devices : 2 > >> >> Failed Devices : 0 > >> >> Spare Devices : 0 > >> >> > >> >> Layout : near=2 > >> >> Chunk Size : 512K > >> >> > >> >> Name : rleblanc-pc:13 (local to host rleblanc-pc) > >> >> UUID : 1eb66d7c:21308453:1e731c8b:1c43dd55 > >> >> Events : 0 > >> >> > >> >> Number Major Minor RaidDevice State > >> >> 0 7 2 0 active sync set-A /dev/loop2 > >> >> 1 7 3 1 active sync set-B /dev/loop3 > >> >> > >> >> # mdadm /dev/md13 -a /dev/loop4 > >> >> mdadm: added /dev/loop4 > >> >> > >> >> # mdadm --detail /dev/md13 > >> >> /dev/md13: > >> >> Version : 1.2 > >> >> Creation Time : Wed Nov 2 11:47:48 2016 > >> >> Raid Level : raid10 > >> >> Array Size : 10477568 (9.99 GiB 10.73 GB) > >> >> Used Dev Size : 10477568 (9.99 GiB 10.73 GB) > >> >> Raid Devices : 2 > >> >> Total Devices : 3 > >> >> Persistence : Superblock is persistent > >> >> > >> >> Update Time : Wed Nov 2 11:48:13 2016 > >> >> State : clean > >> >> Active Devices : 2 > >> >> Working Devices : 3 > >> >> Failed Devices : 0 > >> >> Spare Devices : 1 > >> >> > >> >> Layout : near=2 > >> >> Chunk Size : 512K > >> >> > >> >> Name : rleblanc-pc:13 (local to host rleblanc-pc) > >> >> UUID : 1eb66d7c:21308453:1e731c8b:1c43dd55 > >> >> Events : 1 > >> >> > >> >> Number Major Minor RaidDevice State > >> >> 0 7 2 0 active sync set-A /dev/loop2 > >> >> 1 7 3 1 active sync set-B /dev/loop3 > >> >> > >> >> 2 7 4 - spare /dev/loop4 > >> >> > >> >> # mdadm --grow /dev/md13 -p n3 --raid-devices 3 > >> >> mdadm: Cannot change number of copies when reshaping RAID10 > >> >> > >> >> I also tried to add the device, grow raid-devices, let it reshape, > >> >> then try to change the number of copies and it didn't like that > >> >> either. It would be nice to supply -p nX and --raid-devices X at the > >> >> same time to prevent the reshape and only copy the data over to the > >> >> new drive (or drop a drive out completely). I could see changing -p > >> >> separately or at a different rate of drives added/removed could be > >> >> difficult, but for lockstep changes, it seems that it would be rather > >> >> easy. > >> >> > >> >> Any ideas? > >> >> > >> >> Thanks, > >> >> > >> >> ---------------- > >> >> Robert LeBlanc > >> >> PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 > >> >> -- > >> >> To unsubscribe from this list: send the line "unsubscribe linux-raid" in > >> >> the body of a message to majordomo@xxxxxxxxxxxxxxx > >> >> More majordomo info at http://vger.kernel.org/majordomo-info.html > >> -- > >> To unsubscribe from this list: send the line "unsubscribe linux-raid" in > >> the body of a message to majordomo@xxxxxxxxxxxxxxx > >> More majordomo info at http://vger.kernel.org/majordomo-info.html > > -- > > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > > the body of a message to majordomo@xxxxxxxxxxxxxxx > > More majordomo info at http://vger.kernel.org/majordomo-info.html > -- > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html