On Thu 12-11-15 17:40:59, Baolin Wang wrote: > On 12 November 2015 at 17:17, Jan Kara <jack@xxxxxxx> wrote: > > On Thu 12-11-15 10:15:32, Baolin Wang wrote: > >> On 11 November 2015 at 17:48, Christoph Hellwig <hch@xxxxxxxxxxxxx> wrote: > >> > On Wed, Nov 11, 2015 at 05:31:43PM +0800, Baolin Wang wrote: > >> >> Now the dm-crypt code only implemented the 'based-bio' method to encrypt/ > >> >> decrypt block data, which can only hanle one bio at one time. As we know, > >> >> one bio must use the sequential physical address and it also has a limitation > >> >> of length. Thus it may limit the big block encyrtion/decryption when some > >> >> hardware support the big block data encryption. > >> >> > >> >> This patch series introduc the 'based-request' method to handle the data > >> >> encryption/decryption. One request can contain multiple bios, so it can > >> >> handle big block data to improve the efficiency. > >> > > >> > NAK for more request based stacking or DM drivers. They are a major > >> > pain to deal with, and adding more with different requirements then > >> > dm-multipath is not helping in actually making that one work properly. > >> > >> But now many vendors supply the hardware engine to handle the > >> encyrtion/decryption. The hardware really need a big block to indicate > >> its performance with request based things. Another thing is now the > >> request based things is used by many vendors (Qualcomm, Spreadtrum and > >> so on) to improve their performance and there's a real performance > >> requirement here (I can show the performance result later). > > > > So you've mentioned several times that hardware needs big blocks. How big > > those blocks need to be? Ideally, can you give some numbers on how the > > throughput of the encryption hw grows with the block size? > > It depends on the hardware design. My beaglebone black board's AES > engine can handle 1M at one time which is not big. As I know some > other AES engine can handle 16M data at one time or more. Well, one question is "can handle" and other question is how big gain in throughput it will bring compared to say 1M chunks. I suppose there's some constant overhead to issue a request to the crypto hw and by the time it is encrypting 1M it may be that this overhead is well amortized by the cost of the encryption itself which is in principle linear in the size of the block. That's why I'd like to get idea of the real numbers... > > You mentioned that you use requests because of size limitations on bios - I > > had a look and current struct bio can easily describe 1MB requests (that's > > assuming 64-bit architecture, 4KB pages) when we have 1 page worth of > > struct bio_vec. Is that not enough? > > Usually one bio does not always use the full 1M, maybe some 1k/2k/8k > or some other small chunks. But request can combine some sequential > small bios to be a big block and it is better than bio at least. As Christoph mentions 4.3 should be better in submitting larger bios. Did you check it? Honza -- Jan Kara <jack@xxxxxxxx> SUSE Labs, CR -- dm-devel mailing list dm-devel@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/dm-devel