Both the SCSI and NVMe subsystems receive user data from the block layer in scatterlist_s (aka scatter gather lists (sgl) which are often arrays). If drivers in those subsystems represent storage (e.g. a ramdisk) or cache "hot" user data then they may also choose to use scatterlist_s. Currently there are no sgl to sgl operations in the kernel. Start with a copy. Signed-off-by: Douglas Gilbert <dgilbert@xxxxxxxxxxxx> --- include/linux/scatterlist.h | 4 ++ lib/scatterlist.c | 86 +++++++++++++++++++++++++++++++++++++ 2 files changed, 90 insertions(+) diff --git a/include/linux/scatterlist.h b/include/linux/scatterlist.h index 80178afc2a4a..6649414c0749 100644 --- a/include/linux/scatterlist.h +++ b/include/linux/scatterlist.h @@ -321,6 +321,10 @@ size_t sg_pcopy_to_buffer(struct scatterlist *sgl, unsigned int nents, size_t sg_zero_buffer(struct scatterlist *sgl, unsigned int nents, size_t buflen, off_t skip); +size_t sgl_copy_sgl(struct scatterlist *d_sgl, unsigned int d_nents, off_t d_skip, + struct scatterlist *s_sgl, unsigned int s_nents, off_t s_skip, + size_t n_bytes); + /* * Maximum number of entries that will be allocated in one piece, if * a list larger than this is required then chaining will be utilized. diff --git a/lib/scatterlist.c b/lib/scatterlist.c index d5770e7f1030..1ec2c909c8d4 100644 --- a/lib/scatterlist.c +++ b/lib/scatterlist.c @@ -974,3 +974,89 @@ size_t sg_zero_buffer(struct scatterlist *sgl, unsigned int nents, return offset; } EXPORT_SYMBOL(sg_zero_buffer); + +/** + * sgl_copy_sgl - Copy over a destination sgl from a source sgl + * @d_sgl: Destination sgl + * @d_nents: Number of SG entries in destination sgl + * @d_skip: Number of bytes to skip in destination before copying + * @s_sgl: Source sgl + * @s_nents: Number of SG entries in source sgl + * @s_skip: Number of bytes to skip in source before copying + * @n_bytes: The number of bytes to copy + * + * Returns the number of copied bytes. + * + * Notes: + * Destination arguments appear before the source arguments, as with memcpy(). + * + * Stops copying if the end of d_sgl or s_sgl is reached. + * + * Since memcpy() is used, overlapping copies (where d_sgl and s_sgl belong + * to the same sgl and the copy regions overlap) are not supported. + * + * If d_skip is large, potentially spanning multiple d_nents then some + * integer arithmetic to adjust d_sgl may improve performance. For example + * if d_sgl is built using sgl_alloc_order(chainable=false) then the sgl + * will be an array with equally sized segments facilitating that + * arithmetic. The suggestion applies to s_skip, s_sgl and s_nents as well. + * + **/ +size_t sgl_copy_sgl(struct scatterlist *d_sgl, unsigned int d_nents, off_t d_skip, + struct scatterlist *s_sgl, unsigned int s_nents, off_t s_skip, + size_t n_bytes) +{ + size_t d_off, s_off, len, d_len, s_len; + size_t offset = 0; + struct sg_mapping_iter d_iter; + struct sg_mapping_iter s_iter; + + if (n_bytes == 0) + return 0; + sg_miter_start(&d_iter, d_sgl, d_nents, SG_MITER_ATOMIC | SG_MITER_TO_SG); + sg_miter_start(&s_iter, s_sgl, s_nents, SG_MITER_ATOMIC | SG_MITER_FROM_SG); + if (!sg_miter_skip(&d_iter, d_skip)) + goto fini; + if (!sg_miter_skip(&s_iter, s_skip)) + goto fini; + + for (d_off = 0, s_off = 0; true ; ) { + /* Assume d_iter.length and s_iter.length can never be 0 */ + if (d_off == 0) { + if (!sg_miter_next(&d_iter)) + break; + d_len = d_iter.length; + } else { + d_len = d_iter.length - d_off; + } + if (s_off == 0) { + if (!sg_miter_next(&s_iter)) + break; + s_len = s_iter.length; + } else { + s_len = s_iter.length - s_off; + } + len = min3(d_len, s_len, n_bytes - offset); + + memcpy(d_iter.addr + d_off, s_iter.addr + s_off, len); + offset += len; + if (offset >= n_bytes) + break; + if (d_len == s_len) { + d_off = 0; + s_off = 0; + } else if (d_len < s_len) { + d_off = 0; + s_off += len; + } else { + d_off += len; + s_off = 0; + } + } +fini: + sg_miter_stop(&d_iter); + sg_miter_stop(&s_iter); + return offset; +} +EXPORT_SYMBOL(sgl_copy_sgl); + -- 2.25.1