On 07/11/2023 16:55, Corentin Labbe wrote: > RK3588 have a new crypto IP, this patch adds basic support for it. > Only hashes and cipher are handled for the moment. > > Signed-off-by: Corentin Labbe <clabbe@xxxxxxxxxxxx> > --- > drivers/crypto/Kconfig | 29 + > drivers/crypto/rockchip/Makefile | 5 + > drivers/crypto/rockchip/rk2_crypto.c | 739 ++++++++++++++++++ > drivers/crypto/rockchip/rk2_crypto.h | 246 ++++++ > drivers/crypto/rockchip/rk2_crypto_ahash.c | 344 ++++++++ > drivers/crypto/rockchip/rk2_crypto_skcipher.c | 576 ++++++++++++++ > 6 files changed, 1939 insertions(+) > create mode 100644 drivers/crypto/rockchip/rk2_crypto.c > create mode 100644 drivers/crypto/rockchip/rk2_crypto.h > create mode 100644 drivers/crypto/rockchip/rk2_crypto_ahash.c > create mode 100644 drivers/crypto/rockchip/rk2_crypto_skcipher.c > > diff --git a/drivers/crypto/Kconfig b/drivers/crypto/Kconfig > index 79c3bb9c99c3..b6a2027b1f9a 100644 > --- a/drivers/crypto/Kconfig > +++ b/drivers/crypto/Kconfig > @@ -660,6 +660,35 @@ config CRYPTO_DEV_ROCKCHIP_DEBUG > the number of requests per algorithm and other internal stats. > > > +config CRYPTO_DEV_ROCKCHIP2 > + tristate "Rockchip's cryptographic offloader V2" > + depends on OF && ARCH_ROCKCHIP > + depends on PM > + select CRYPTO_ECB > + select CRYPTO_CBC > + select CRYPTO_AES > + select CRYPTO_MD5 > + select CRYPTO_SHA1 > + select CRYPTO_SHA256 > + select CRYPTO_SHA512 > + select CRYPTO_SM3_GENERIC > + select CRYPTO_HASH > + select CRYPTO_SKCIPHER > + select CRYPTO_ENGINE > + > + help > + This driver interfaces with the hardware crypto offloader present > + on RK3566, RK3568 and RK3588. > + > +config CRYPTO_DEV_ROCKCHIP2_DEBUG > + bool "Enable Rockchip V2 crypto stats" > + depends on CRYPTO_DEV_ROCKCHIP2 > + depends on DEBUG_FS > + help > + Say y to enable Rockchip crypto debug stats. > + This will create /sys/kernel/debug/rk3588_crypto/stats for displaying > + the number of requests per algorithm and other internal stats. > + > config CRYPTO_DEV_ZYNQMP_AES > tristate "Support for Xilinx ZynqMP AES hw accelerator" > depends on ZYNQMP_FIRMWARE || COMPILE_TEST > diff --git a/drivers/crypto/rockchip/Makefile b/drivers/crypto/rockchip/Makefile > index 785277aca71e..452a12ff6538 100644 > --- a/drivers/crypto/rockchip/Makefile > +++ b/drivers/crypto/rockchip/Makefile > @@ -3,3 +3,8 @@ obj-$(CONFIG_CRYPTO_DEV_ROCKCHIP) += rk_crypto.o > rk_crypto-objs := rk3288_crypto.o \ > rk3288_crypto_skcipher.o \ > rk3288_crypto_ahash.o > + > +obj-$(CONFIG_CRYPTO_DEV_ROCKCHIP2) += rk_crypto2.o > +rk_crypto2-objs := rk2_crypto.o \ > + rk2_crypto_skcipher.o \ > + rk2_crypto_ahash.o > diff --git a/drivers/crypto/rockchip/rk2_crypto.c b/drivers/crypto/rockchip/rk2_crypto.c > new file mode 100644 > index 000000000000..f3b8d27924da > --- /dev/null > +++ b/drivers/crypto/rockchip/rk2_crypto.c > @@ -0,0 +1,739 @@ > +// SPDX-License-Identifier: GPL-2.0 > +/* > + * hardware cryptographic offloader for RK3568/RK3588 SoC > + * > + * Copyright (c) 2022-2023, Corentin Labbe <clabbe@xxxxxxxxxxxx> > + */ > + > +#include "rk2_crypto.h" > +#include <linux/clk.h> > +#include <linux/crypto.h> > +#include <linux/dma-mapping.h> > +#include <linux/module.h> > +#include <linux/platform_device.h> > +#include <linux/of.h> > +#include <linux/of_device.h> > +#include <linux/reset.h> > + > +static struct rockchip_ip rocklist = { > + .dev_list = LIST_HEAD_INIT(rocklist.dev_list), > + .lock = __SPIN_LOCK_UNLOCKED(rocklist.lock), Drop it, not needed. > +}; > + Exposed interfaces need kerneldoc. > +struct rk2_crypto_dev *get_rk2_crypto(void) > +{ > + struct rk2_crypto_dev *first; > + > + spin_lock(&rocklist.lock); > + first = list_first_entry_or_null(&rocklist.dev_list, > + struct rk2_crypto_dev, list); > + list_rotate_left(&rocklist.dev_list); > + spin_unlock(&rocklist.lock); > + return first; This is some spaghetti... Why do need a global list of devices to pass to crypto callbacks? The context is for that. This does not scale. Drop entire function. > +} > + > +static const struct rk2_variant rk3568_variant = { > + .num_clks = 3, > +}; > + > +static const struct rk2_variant rk3588_variant = { > + .num_clks = 3, > +}; > + > +static int rk2_crypto_get_clks(struct rk2_crypto_dev *dev) > +{ > + int i, j, err; > + unsigned long cr; > + > + dev->num_clks = devm_clk_bulk_get_all(dev->dev, &dev->clks); > + if (dev->num_clks < dev->variant->num_clks) { > + dev_err(dev->dev, "Missing clocks, got %d instead of %d\n", > + dev->num_clks, dev->variant->num_clks); > + return -EINVAL; > + } > + > + for (i = 0; i < dev->num_clks; i++) { > + cr = clk_get_rate(dev->clks[i].clk); > + for (j = 0; j < ARRAY_SIZE(dev->variant->rkclks); j++) { > + if (dev->variant->rkclks[j].max == 0) > + continue; > + if (strcmp(dev->variant->rkclks[j].name, dev->clks[i].id)) > + continue; > + if (cr > dev->variant->rkclks[j].max) { > + err = clk_set_rate(dev->clks[i].clk, > + dev->variant->rkclks[j].max); > + if (err) > + dev_err(dev->dev, "Fail downclocking %s from %lu to %lu\n", > + dev->variant->rkclks[j].name, cr, > + dev->variant->rkclks[j].max); > + else > + dev_info(dev->dev, "Downclocking %s from %lu to %lu\n", > + dev->variant->rkclks[j].name, cr, > + dev->variant->rkclks[j].max); > + } > + } > + } > + return 0; > +} > + > +static int rk2_crypto_enable_clk(struct rk2_crypto_dev *dev) > +{ > + int err; > + > + err = clk_bulk_prepare_enable(dev->num_clks, dev->clks); > + if (err) > + dev_err(dev->dev, "Could not enable clock clks\n"); Why do you create abstract wrappers over single call? > + > + return err; > +} > + > +static void rk2_crypto_disable_clk(struct rk2_crypto_dev *dev) > +{ > + clk_bulk_disable_unprepare(dev->num_clks, dev->clks); What is the benefit of this? I know that downstream code likes abstraction layers. You are supposed to remove all of them when upstreaming. > +} > + > +/* > + * Power management strategy: The device is suspended until a request > + * is handled. For avoiding suspend/resume yoyo, the autosuspend is set to 2s. > + */ > +static int rk2_crypto_pm_suspend(struct device *dev) > +{ ... > + } > +} > + > +static const struct of_device_id crypto_of_id_table[] = { > + { .compatible = "rockchip,rk3568-crypto", > + .data = &rk3568_variant, > + }, > + { .compatible = "rockchip,rk3588-crypto", > + .data = &rk3588_variant, > + }, > + {} > +}; > +MODULE_DEVICE_TABLE(of, crypto_of_id_table); > + > +static int rk2_crypto_probe(struct platform_device *pdev) > +{ > + struct device *dev = &pdev->dev; > + struct rk2_crypto_dev *rkc, *first; > + int err = 0; > + > + rkc = devm_kzalloc(&pdev->dev, sizeof(*rkc), GFP_KERNEL); > + if (!rkc) { > + err = -ENOMEM; return -ENOMEM > + goto err_crypto; > + } > + > + rkc->dev = &pdev->dev; > + platform_set_drvdata(pdev, rkc); > + > + rkc->variant = of_device_get_match_data(&pdev->dev); > + if (!rkc->variant) { > + dev_err(&pdev->dev, "Missing variant\n"); How is this possible? > + return -EINVAL; > + } > + > + rkc->rst = devm_reset_control_array_get_exclusive(dev); > + if (IS_ERR(rkc->rst)) { > + err = PTR_ERR(rkc->rst); > + dev_err(&pdev->dev, "Fail to get resets err=%d\n", err); Nope, this is just wrong. You do not have any cleanup path. And regardless of cleanup path, why do you print "Accelerator not successfully registered? It's an error! Syntax is: return dev_err_probe > + goto err_crypto; > + } > + > + rkc->tl = dma_alloc_coherent(rkc->dev, > + sizeof(struct rk2_crypto_lli) * MAX_LLI, > + &rkc->t_phy, GFP_KERNEL); > + if (!rkc->tl) { > + dev_err(rkc->dev, "Cannot get DMA memory for task\n"); Run coccinelle. > + err = -ENOMEM; Nope, return dev_err_probe > + goto err_crypto; > + } > + > + reset_control_assert(rkc->rst); > + usleep_range(10, 20); > + reset_control_deassert(rkc->rst); > + > + rkc->reg = devm_platform_ioremap_resource(pdev, 0); > + if (IS_ERR(rkc->reg)) { > + err = PTR_ERR(rkc->reg); > + dev_err(&pdev->dev, "Fail to get resources\n"); > + goto err_crypto; > + } > + > + err = rk2_crypto_get_clks(rkc); > + if (err) > + goto err_crypto; > + > + rkc->irq = platform_get_irq(pdev, 0); > + if (rkc->irq < 0) { > + dev_err(&pdev->dev, "control Interrupt is not available.\n"); > + err = rkc->irq; > + goto err_crypto; > + } > + > + err = devm_request_irq(&pdev->dev, rkc->irq, > + rk2_crypto_irq_handle, IRQF_SHARED, > + "rk-crypto", pdev); > + > + if (err) { > + dev_err(&pdev->dev, "irq request failed.\n"); > + goto err_crypto; > + } > + > + rkc->engine = crypto_engine_alloc_init(&pdev->dev, true); > + crypto_engine_start(rkc->engine); > + init_completion(&rkc->complete); > + > + err = rk2_crypto_pm_init(rkc); > + if (err) > + goto err_pm; > + > + err = pm_runtime_resume_and_get(&pdev->dev); > + > + spin_lock(&rocklist.lock); > + first = list_first_entry_or_null(&rocklist.dev_list, > + struct rk2_crypto_dev, list); > + list_add_tail(&rkc->list, &rocklist.dev_list); > + spin_unlock(&rocklist.lock); > + > + if (!first) { > + dev_info(dev, "Registers crypto algos\n"); > + err = rk2_crypto_register(rkc); > + if (err) { > + dev_err(dev, "Fail to register crypto algorithms"); > + goto err_register_alg; > + } > + > + register_debugfs(rkc); > + } > + > + return 0; > + > +err_register_alg: > + rk2_crypto_pm_exit(rkc); > +err_pm: > + crypto_engine_exit(rkc->engine); > +err_crypto: > + dev_err(dev, "Crypto Accelerator not successfully registered\n"); Drop useless probe success messages. > + return err; > +} > + > +static int rk2_crypto_remove(struct platform_device *pdev) > +{ > + struct rk2_crypto_dev *rkc = platform_get_drvdata(pdev); > + struct rk2_crypto_dev *first; > + > + spin_lock_bh(&rocklist.lock); > + list_del(&rkc->list); > + first = list_first_entry_or_null(&rocklist.dev_list, > + struct rk2_crypto_dev, list); > + spin_unlock_bh(&rocklist.lock); > + > + if (!first) { > +#ifdef CONFIG_CRYPTO_DEV_ROCKCHIP2_DEBUG > + debugfs_remove_recursive(rocklist.dbgfs_dir); > +#endif > + rk2_crypto_unregister(); > + } > + rk2_crypto_pm_exit(rkc); > + crypto_engine_exit(rkc->engine); > + return 0; > +} > + > +static struct platform_driver crypto_driver = { > + .probe = rk2_crypto_probe, > + .remove = rk2_crypto_remove, > + .driver = { > + .name = "rk2-crypto", > + .pm = &rk2_crypto_pm_ops, > + .of_match_table = crypto_of_id_table, > + }, > +}; > + > +module_platform_driver(crypto_driver); > + > +MODULE_DESCRIPTION("Rockchip Crypto Engine cryptographic offloader"); > +MODULE_LICENSE("GPL"); > +MODULE_AUTHOR("Corentin Labbe <clabbe@xxxxxxxxxxxx>"); > diff --git a/drivers/crypto/rockchip/rk2_crypto.h b/drivers/crypto/rockchip/rk2_crypto.h > new file mode 100644 > index 000000000000..59cd8be59f70 > --- /dev/null > +++ b/drivers/crypto/rockchip/rk2_crypto.h > @@ -0,0 +1,246 @@ > +/* SPDX-License-Identifier: GPL-2.0 */ > + > +#include <crypto/aes.h> > +#include <crypto/xts.h> > +#include <crypto/engine.h> > +#include <crypto/internal/des.h> > +#include <crypto/internal/hash.h> > +#include <crypto/internal/skcipher.h> > +#include <crypto/algapi.h> > +#include <crypto/md5.h> > +#include <crypto/sha1.h> > +#include <crypto/sha2.h> > +#include <crypto/sm3.h> > +#include <linux/dma-mapping.h> > +#include <linux/interrupt.h> > +#include <linux/debugfs.h> > +#include <linux/delay.h> > +#include <linux/pm_runtime.h> > +#include <linux/scatterlist.h> > +#include <linux/hw_random.h> Nope. This is some insane list of unneeded headers. You include here only the headers which this header uses. I see just few. > + > +#define RK2_CRYPTO_CLK_CTL 0x0000 > +#define RK2_CRYPTO_RST_CTL 0x0004 > + > +#define RK2_CRYPTO_DMA_INT_EN 0x0008 > +/* values for RK2_CRYPTO_DMA_INT_EN */ > +#define RK2_CRYPTO_DMA_INT_LISTDONE BIT(0) > + > +#define RK2_CRYPTO_DMA_INT_ST 0x000C > +/* values in RK2_CRYPTO_DMA_INT_ST are the same than in RK2_CRYPTO_DMA_INT_EN */ > + > +#define RK2_CRYPTO_DMA_CTL 0x0010 > +#define RK2_CRYPTO_DMA_CTL_START BIT(0) > + > +#define RK2_CRYPTO_DMA_LLI_ADDR 0x0014 > +#define RK2_CRYPTO_DMA_ST 0x0018 > +#define RK2_CRYPTO_DMA_STATE 0x001C > +#define RK2_CRYPTO_DMA_LLI_RADDR 0x0020 > +#define RK2_CRYPTO_DMA_SRC_RADDR 0x0024 > +#define RK2_CRYPTO_DMA_DST_WADDR 0x0028 > +#define RK2_CRYPTO_DMA_ITEM_ID 0x002C > + > +#define RK2_CRYPTO_FIFO_CTL 0x0040 > + > +#define RK2_CRYPTO_BC_CTL 0x0044 > +#define RK2_CRYPTO_AES (0 << 8) > +#define RK2_CRYPTO_MODE_ECB (0 << 4) > +#define RK2_CRYPTO_MODE_CBC (1 << 4) > +#define RK2_CRYPTO_XTS (6 << 4) > + > +#define RK2_CRYPTO_HASH_CTL 0x0048 > +#define RK2_CRYPTO_HW_PAD BIT(2) > +#define RK2_CRYPTO_SHA1 (0 << 4) > +#define RK2_CRYPTO_MD5 (1 << 4) > +#define RK2_CRYPTO_SHA224 (3 << 4) > +#define RK2_CRYPTO_SHA256 (2 << 4) > +#define RK2_CRYPTO_SHA384 (9 << 4) > +#define RK2_CRYPTO_SHA512 (8 << 4) > +#define RK2_CRYPTO_SM3 (4 << 4) > + > +#define RK2_CRYPTO_AES_ECB (RK2_CRYPTO_AES | RK2_CRYPTO_MODE_ECB) > +#define RK2_CRYPTO_AES_CBC (RK2_CRYPTO_AES | RK2_CRYPTO_MODE_CBC) > +#define RK2_CRYPTO_AES_XTS (RK2_CRYPTO_AES | RK2_CRYPTO_XTS) > +#define RK2_CRYPTO_AES_CTR_MODE 3 > +#define RK2_CRYPTO_AES_128BIT_key (0 << 2) > +#define RK2_CRYPTO_AES_192BIT_key (1 << 2) > +#define RK2_CRYPTO_AES_256BIT_key (2 << 2) > + > +#define RK2_CRYPTO_DEC BIT(1) > +#define RK2_CRYPTO_ENABLE BIT(0) > + > +#define RK2_CRYPTO_CIPHER_ST 0x004C > +#define RK2_CRYPTO_CIPHER_STATE 0x0050 > + > +#define RK2_CRYPTO_CH0_IV_0 0x0100 > + > +#define RK2_CRYPTO_KEY0 0x0180 > +#define RK2_CRYPTO_KEY1 0x0184 > +#define RK2_CRYPTO_KEY2 0x0188 > +#define RK2_CRYPTO_KEY3 0x018C > +#define RK2_CRYPTO_KEY4 0x0190 > +#define RK2_CRYPTO_KEY5 0x0194 > +#define RK2_CRYPTO_KEY6 0x0198 > +#define RK2_CRYPTO_KEY7 0x019C > +#define RK2_CRYPTO_CH4_KEY0 0x01c0 > + > +#define RK2_CRYPTO_CH0_PC_LEN_0 0x0280 > + > +#define RK2_CRYPTO_CH0_IV_LEN 0x0300 > + > +#define RK2_CRYPTO_HASH_DOUT_0 0x03A0 > +#define RK2_CRYPTO_HASH_VALID 0x03E4 > + > +#define RK2_CRYPTO_TRNG_CTL 0x0400 > +#define RK2_CRYPTO_TRNG_START BIT(0) > +#define RK2_CRYPTO_TRNG_ENABLE BIT(1) > +#define RK2_CRYPTO_TRNG_256 (0x3 << 4) > +#define RK2_CRYPTO_TRNG_SAMPLE_CNT 0x0404 > +#define RK2_CRYPTO_TRNG_DOUT 0x0410 > + > +#define CRYPTO_AES_VERSION 0x0680 > +#define CRYPTO_DES_VERSION 0x0684 > +#define CRYPTO_SM4_VERSION 0x0688 > +#define CRYPTO_HASH_VERSION 0x068C > +#define CRYPTO_HMAC_VERSION 0x0690 > +#define CRYPTO_RNG_VERSION 0x0694 > +#define CRYPTO_PKA_VERSION 0x0698 > +#define CRYPTO_CRYPTO_VERSION 0x06F0 > + > +#define RK2_LLI_DMA_CTRL_SRC_INT BIT(10) > +#define RK2_LLI_DMA_CTRL_DST_INT BIT(9) > +#define RK2_LLI_DMA_CTRL_LIST_INT BIT(8) > +#define RK2_LLI_DMA_CTRL_LAST BIT(0) > + > +#define RK2_LLI_STRING_LAST BIT(2) > +#define RK2_LLI_STRING_FIRST BIT(1) > +#define RK2_LLI_CIPHER_START BIT(0) > + > +#define RK2_MAX_CLKS 4 > + > +#define MAX_LLI 20 > + > +struct rk2_crypto_lli { > + __le32 src_addr; > + __le32 src_len; > + __le32 dst_addr; > + __le32 dst_len; > + __le32 user; > + __le32 iv; > + __le32 dma_ctrl; > + __le32 next; > +}; > + > +/* > + * struct rockchip_ip - struct for managing a list of RK crypto instance > + * @dev_list: Used for doing a list of rk2_crypto_dev > + * @lock: Control access to dev_list > + * @dbgfs_dir: Debugfs dentry for statistic directory > + * @dbgfs_stats: Debugfs dentry for statistic counters > + */ > +struct rockchip_ip { > + struct list_head dev_list; > + spinlock_t lock; /* Control access to dev_list */ > + struct dentry *dbgfs_dir; > + struct dentry *dbgfs_stats; > +}; > + > +struct rk2_clks { > + const char *name; > + unsigned long max; > +}; > + > +struct rk2_variant { > + int num_clks; > + struct rk2_clks rkclks[RK2_MAX_CLKS]; > +}; > + > +struct rk2_crypto_dev { > + struct list_head list; > + struct device *dev; > + struct clk_bulk_data *clks; > + int num_clks; > + struct reset_control *rst; > + void __iomem *reg; > + int irq; > + const struct rk2_variant *variant; > + unsigned long nreq; > + struct crypto_engine *engine; > + struct completion complete; > + int status; > + struct rk2_crypto_lli *tl; > + dma_addr_t t_phy; > +}; > + > +/* the private variable of hash */ > +struct rk2_ahash_ctx { > + /* for fallback */ > + struct crypto_ahash *fallback_tfm; > +}; > + > +/* the private variable of hash for fallback */ > +struct rk2_ahash_rctx { > + struct rk2_crypto_dev *dev; > + struct ahash_request fallback_req; > + u32 mode; > + int nrsgs; > +}; > + > +/* the private variable of cipher */ > +struct rk2_cipher_ctx { > + unsigned int keylen; > + u8 key[AES_MAX_KEY_SIZE * 2]; > + u8 iv[AES_BLOCK_SIZE]; > + struct crypto_skcipher *fallback_tfm; > +}; > + > +struct rk2_cipher_rctx { > + struct rk2_crypto_dev *dev; > + u8 backup_iv[AES_BLOCK_SIZE]; > + u32 mode; > + struct skcipher_request fallback_req; // keep at the end > +}; > + > +struct rk2_crypto_template { > + u32 type; > + u32 rk2_mode; > + bool is_xts; > + struct rk2_crypto_dev *dev; Your indentation is a mess. > + union { > + struct skcipher_engine_alg skcipher; > + struct ahash_engine_alg hash; > + } alg; > + unsigned long stat_req; > + unsigned long stat_fb; > + unsigned long stat_fb_len; > + unsigned long stat_fb_sglen; > + unsigned long stat_fb_align; > + unsigned long stat_fb_sgdiff; > +}; > + > +struct rk2_crypto_dev *get_rk2_crypto(void); > +int rk2_cipher_run(struct crypto_engine *engine, void *async_req); > +int rk2_hash_run(struct crypto_engine *engine, void *breq); > + > +int rk2_cipher_tfm_init(struct crypto_skcipher *tfm); > +void rk2_cipher_tfm_exit(struct crypto_skcipher *tfm); > +int rk2_aes_setkey(struct crypto_skcipher *cipher, const u8 *key, > + unsigned int keylen); > +int rk2_aes_xts_setkey(struct crypto_skcipher *cipher, const u8 *key, > + unsigned int keylen); > +int rk2_skcipher_encrypt(struct skcipher_request *req); > +int rk2_skcipher_decrypt(struct skcipher_request *req); > +int rk2_aes_ecb_encrypt(struct skcipher_request *req); > +int rk2_aes_ecb_decrypt(struct skcipher_request *req); > +int rk2_aes_cbc_encrypt(struct skcipher_request *req); > +int rk2_aes_cbc_decrypt(struct skcipher_request *req); > + > +int rk2_ahash_init(struct ahash_request *req); > +int rk2_ahash_update(struct ahash_request *req); > +int rk2_ahash_final(struct ahash_request *req); > +int rk2_ahash_finup(struct ahash_request *req); > +int rk2_ahash_import(struct ahash_request *req, const void *in); > +int rk2_ahash_export(struct ahash_request *req, void *out); > +int rk2_ahash_digest(struct ahash_request *req); > +int rk2_hash_init_tfm(struct crypto_ahash *tfm); > +void rk2_hash_exit_tfm(struct crypto_ahash *tfm); Missing guards. > diff --git a/drivers/crypto/rockchip/rk2_crypto_ahash.c b/drivers/crypto/rockchip/rk2_crypto_ahash.c > new file mode 100644 > index 000000000000..75b8d9893447 > --- /dev/null > +++ b/drivers/crypto/rockchip/rk2_crypto_ahash.c > @@ -0,0 +1,344 @@ > +// SPDX-License-Identifier: GPL-2.0-only > +/* > + * Crypto offloader support for Rockchip RK3568/RK3588 > + * > + * Copyright (c) 2022-2023 Corentin Labbe <clabbe@xxxxxxxxxxxx> > + */ > +#include <asm/unaligned.h> > +#include <linux/iopoll.h> There is no way this file needs only two headers. Each unit must include everything it uses. > +#include "rk2_crypto.h" So here you stuffed all the includes... no. > + > +int rk2_hash_run(struct crypto_engine *engine, void *breq) > +{ > + struct ahash_request *areq = container_of(breq, struct ahash_request, base); > + struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq); > + struct rk2_ahash_rctx *rctx = ahash_request_ctx(areq); > + struct ahash_alg *alg = crypto_ahash_alg(tfm); > + struct rk2_crypto_template *algt = container_of(alg, struct rk2_crypto_template, alg.hash.base); > + struct scatterlist *sgs = areq->src; > + struct rk2_crypto_dev *rkc = rctx->dev; > + struct rk2_crypto_lli *dd = &rkc->tl[0]; > + int ddi = 0; > + int err = 0; > + unsigned int len = areq->nbytes; > + unsigned int todo; > + u32 v; > + int i; > + > + err = rk2_hash_prepare(engine, breq); > + > + err = pm_runtime_resume_and_get(rkc->dev); > + if (err) > + return err; > + > + dev_dbg(rkc->dev, "%s %s len=%d\n", __func__, > + crypto_tfm_alg_name(areq->base.tfm), areq->nbytes); > + > + algt->stat_req++; > + rkc->nreq++; > + > + rctx->mode = algt->rk2_mode; > + rctx->mode |= 0xffff0000; > + rctx->mode |= RK2_CRYPTO_ENABLE | RK2_CRYPTO_HW_PAD; > + writel(rctx->mode, rkc->reg + RK2_CRYPTO_HASH_CTL); > + > + while (sgs && len > 0) { > + dd = &rkc->tl[ddi]; > + > + todo = min(sg_dma_len(sgs), len); > + dd->src_addr = sg_dma_address(sgs); > + dd->src_len = todo; > + dd->dst_addr = 0; > + dd->dst_len = 0; > + dd->dma_ctrl = ddi << 24; > + dd->iv = 0; > + dd->next = rkc->t_phy + sizeof(struct rk2_crypto_lli) * (ddi + 1); > + > + if (ddi == 0) > + dd->user = RK2_LLI_CIPHER_START | RK2_LLI_STRING_FIRST; > + else > + dd->user = 0; > + > + len -= todo; > + dd->dma_ctrl |= RK2_LLI_DMA_CTRL_SRC_INT; > + if (len == 0) { > + dd->user |= RK2_LLI_STRING_LAST; > + dd->dma_ctrl |= RK2_LLI_DMA_CTRL_LAST; > + } > + dev_dbg(rkc->dev, "HASH SG %d sglen=%d user=%x dma=%x mode=%x len=%d todo=%d phy=%llx\n", > + ddi, sgs->length, dd->user, dd->dma_ctrl, rctx->mode, len, todo, rkc->t_phy); > + > + sgs = sg_next(sgs); > + ddi++; > + } > + dd->next = 1; > + writel(RK2_CRYPTO_DMA_INT_LISTDONE | 0x7F, rkc->reg + RK2_CRYPTO_DMA_INT_EN); > + > + writel(rkc->t_phy, rkc->reg + RK2_CRYPTO_DMA_LLI_ADDR); > + > + reinit_completion(&rkc->complete); > + rkc->status = 0; > + > + writel(RK2_CRYPTO_DMA_CTL_START | RK2_CRYPTO_DMA_CTL_START << 16, rkc->reg + RK2_CRYPTO_DMA_CTL); Wrap your code according to Coding Style document (so 80). > + > + wait_for_completion_interruptible_timeout(&rkc->complete, > + msecs_to_jiffies(2000)); > + if (!rkc->status) { > + dev_err(rkc->dev, "DMA timeout\n"); > + err = -EFAULT; > + goto theend; Please use something closer to Linux. goto exit; goto out; Best regards, Krzysztof