On 11/12/2014 1:55 PM, Artem Bityutskiy wrote:
On Tue, 2014-11-11 at 22:36 +0200, Tanya Brokhman wrote:
Unfortunately none. This is done for a new device that we received just
now. The development was done on a virtual machine with nandsim. Testing
was more of stability and regression
OK. So the implementation is theory-driven and misses the experimental
prove. This means that building a product based on this implementation
has certain amount of risk involved.
And from where I am, the theoretical base for the solution also does not
look very strong.
The advantages of the "read all periodically" approach were:
1. Simple, no modifications needed
2. No need to write if the media is read-only, except when scrubbing
happens.
3. Should cover all the NAND effects, including the "radiation" one.
Disadvantages (as I see it):
1. performance hit: when do you trigger the "read-all"? will effect
performance
Right. We do not know how often, just like we do not know how often and
how much (read counter threshold) in your proposal.
Performance - sure, matter of experiment, just like the performance of
your solution. And as I notice, energy too (read - battery life).
In your solution you have to do more work maintaining the counters and
writing them. With read solution you do more work reading data.
But the maintaining work is minimal here. ++the counter on every read is
all that is required and verify it's value. O(1)...
Saving them on fastmap also doesn't add any more maintenance work. They
are saved as part of fastmap. I didn't increase the number of events
that trigger saving fastmat to flash. So all is changes is that the
number of scubbing events increased
The promise that reading may be done in background, when there is no
other I/O.
2. finds bitflips only when they are present instead of preventing them
from happening
But is this true? I do not see how is this true in your case. Yo want to
scrub by threshold, which is a theoretical value with very large
deviation from the real one. And there may be no real one even - the
real one depends on the erase block, it depends on the I/O patterns, and
it depends on the temperature.
I know... We got the threshold value (that is exposed in my patches as a
define you just missed it) from NAND manufacturer asking to take into
consideration the temperature the device will operate at. I know its
still an estimation but so is the program/erase threshold. Since it was
set by manufacturer - I think its the best one we can hope for.
You will end up scrubbing a lot earlier than needed. Here comes the
performance loss too (and energy). And you will eventually end up
scrubbing too late.
I don't see why I would end up scrubbing too late?
I do not see how your solution provides any hard guarantee. Please,
explain how do you guarantee that my PEB does not bit-rot earlier than
read counter reaches the threshold? It may bit-rot earlier because it is
close to be worn out, or because of just higher temperature, or because
it has a nano-defect.
I can't guarantee it wont bit-flip, I don't think any one could but I
can say that with my implementation the chance of bit-flip is reduced.
Even if not all the scenarios are covered. For example in the bellow
case I reduce the chance of data loss:
In an endless loop - read page 3 of PEB-A.
This will effect near by pages (say 4 and 2 for simplicity). But if I
scrub the whole PEB according to read-counter I will save data of pages
2 and 4.
If I do nothing: when reading eventually page 4 it will produce
bit-flips that may not be fixable.
Perhaps our design is an overkill for this and not covering 100% of te
usecases. But it was requested by our customers to handle read-disturb
and data retention specifically (as in "prevent" and not just "fix").
This is due to a new NAND device that should operate in high temperature
and last for ~15-20 years.
I understand the whole customer orientation concept. But for me so far
the solution does not feel like something suitable to a customer I could
imagine. I mean, if I think about me as a potential customer, I would
just want my data to be safe and covered from all the NAND effects.
I'm not sure that at the moment "all NAND effects" can be covered. In
our case the result is that we reduce the chance of loosing data. not to
0% unfortunately but still reduce.
And from the tests we ran we didn't observe performance hit with this
implementation. And the customer doesn't really care how this was done.
I do not know about power. Its possible that our implementation will
have negative effect on power consumption. I don't have the equipment to
verify that unfortunately.
There are plans to test this implementation in extreme temperature
conditions and get some real numbers and statistics on endurance. It
wasn't done yet and wont be done by us. When I get the results I'll try
to share (if allowed to by legal)
I
would not want counters, I'd want the result. And in the proposed
solution I would not see how I'd get the guaranteed result. But of
course I do not know the customer requirements that you've got.
Thanks,
Tanya Brokhman
--
Qualcomm Israel, on behalf of Qualcomm Innovation Center, Inc.
The Qualcomm Innovation Center, Inc. is a member of the Code Aurora
Forum, a Linux Foundation Collaborative Project
--
To unsubscribe from this list: send the line "unsubscribe linux-arm-msm" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html