On 10/01/2014 04:56 AM, NeilBrown wrote:On Wed, 24 Sep 2014 13:02:28 +0200 Heinz Mauelshagen <heinzm@xxxxxxxxxx> wrote:Martin, thanks for the good explanation of the state of the discard union. Do you have an ETA for the 'zeroout, deallocate' ... support you mentioned? I was planning to have a followup patch for dm-raid supporting a dm-raid table line argument to prohibit discard passdown. In lieu of the fuzzy field situation wrt SSD fw and discard_zeroes_data support related to RAID4/5/6, we need that in upstream together with the initial patch. That 'no_discard_passdown' table line can be added to dm-raid RAID4/5/6 table lines to avoid possible data corruption but can be avoided on RAID1/10 table lines, because the latter are not suffering from any discard_zeroes_data flaw. Neil, are you going to disable discards in RAID4/5/6 shortly or rather go with your bitmap solution?Can I just close my eyes and hope it goes away? I'd think that it actually would. Sane deployments are unlikely to base MD RAID4/5/6 mappings on ancient/unreliable SSDs. Your "md_mod.willing_to_risk_discard=Y" mentioned below is the proper way against all insane odds. BTW: we're referring to SSD discard_zeroes_data issues only so far it seems? Can we be sure that SCSI unmap with the TPRZ bit set will always return zeroes on reads for arbitrary storage devices? The idea of a bitmap of uninitialised areas is not a short-term solution. But I'm not really keen on simply disabling discard for RAID4/5/6 either. It would mean that people with good sensible hardware wouldn't be able to use it properly. I would really rather that discard_zeroes_data were only set on devices where it was actually true. Then it wouldn't be my problem any more. Maybe I could do a loud warning "Not enabling DISCARD on RAID5 because we cannot trust committees. Set "md_mod.willing_to_risk_discard=Y" if your devices reads discarded sectors as zeros" Yes, we are thinking the same for dm-raid (like a "force_discard" parameter). and add an appropriate module parameter...... While we are on the topic, maybe I should write down my thoughts about the bitmap thing in case someone wants to contribute. There are 3 states that a 'region' can be in: 1- known to be in-sync 2- possibly not in sync, but it should be 3- probably not in sync, contains no valuable data. A read from '3' should return zeroes. A write to '3' should change the region to be '2'. It could either write zeros before allowing the write to start, or it could just start a normal resync. Here is a question: if a region has been discarded, are we guaranteed that reads are at least stable. i.e. if I read twice will I definitely get the same value? We can never be sure taking the standards and implementation hassle into account. Thus RAID system OEMs have whitelists for supported drives. It would be simplest to store all the states in the one bitmap. i.e. have a new version of the current bitmap (which distinguishes between '1' and '2') which allows the 3rd state. Probably just 2 bits per region, though we could squeeze state for 20 regions into 32 bits if we really wanted :-) Then we set the bitmap to all '3's when it is created and set regions to '3' when discarding. One problem with this approach is that it forces the same region size. Regions for the current bitmap can conveniently be 10s of Megabytes. For state '3', we ideally want one region per stripe. For 4TB drives with a 512K chunk size (and smaller is not uncommon) that is 8 million regions which is a megabyte of bits. I guess that isn't all that much. One problem with small regions for the 1/2 distinction is that we end up updating the bitmap more often, but that could be avoided by setting multiple bits at one. i.e. keep the internal counters with a larger granularity, and when a counter (of outstanding writes) becomes non-zero, set quite a few bits in the bitmap. How close can you get with update patterns on bitmaps for that new approach to what we have with the current one? Will it be negligible or performance impacting? So that is probably what I would do: - new version for bitmap which has 2 bits per region and encodes 3 states - bitmap granularity matches chunk size (by default) - decouple region size of bitmap from region size for internal 'dirty' accounting - write to a 'state 3' region sets it to 'state 2' and kicks off resync - 'discard' sets state to '3'. Makes sense but we'd rather avoid it based on the aforementioned "willing" option. Heinz NeilBrown |
-- dm-devel mailing list dm-devel@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/dm-devel