On Thu, May 16, 2013 at 04:07:32PM +0400, Dmitry Monakhov wrote: > Parallels team have old good tool called hwflush-check which is server/client > application for testing data integrity under system/disk failure conditions. > Usually we run hwflush-check on two different hosts and use PMU to trigger real > power failure of the client as a whole unit. This tests may be used for > SSD checking (some of them are known to have probelms with hwflush). > I hope it will be good to share it with community. > > This tests simulate just one disk failure while client system should > survive this failure. This test extend idea of shared/305. > 1) Run hwflush-check server and client on same host as usual > 2) Simulare disk failure via blkdev failt injection API aka 'make-it-fail' > 3) Umount failed device > 4) Makes disk operatable again > 5) Mount filesystem > 3) Check data integrity So, for local disk failure, why do we need a client/server network architecture? That just complicates the code, and AFAICT all the client does is send report report packets to server which contain an id number that is kept in memory. If on restart of the client after failure the ID in the report packet doesn't match what the server wants, then it fails the test. So, why is the server needed here? Just dump the IDs the client writes to the file on a device not being tested, and either diff them against a golden image or run a check to see all the IDs are monotonically increasing. That removes all the networking code from the test, the need for a client/server architecture, etc, and makes the test far easier to review. Cheers, Dave. -- Dave Chinner david@xxxxxxxxxxxxx -- To unsubscribe from this list: send the line "unsubscribe linux-ext4" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html