Am Samstag, 5. Dezember 2015, 14:30:25 schrieb Zain Wang: Hi Zain, >Add md5 sha1 sha256 support for crypto engine in rk3288. >This patch can't support multiple updatings because of limited of IC, >as result, it can't support import and export too. > >Signed-off-by: Zain Wang <zain.wang@xxxxxxxxxxxxxx> >--- > drivers/crypto/rockchip/Makefile | 1 + > drivers/crypto/rockchip/rk3288_crypto.c | 33 +- > drivers/crypto/rockchip/rk3288_crypto.h | 50 ++- > drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c | 20 +- > drivers/crypto/rockchip/rk3288_crypto_ahash.c | 369 >+++++++++++++++++++++ 5 files changed, 455 insertions(+), 18 deletions(-) > create mode 100644 drivers/crypto/rockchip/rk3288_crypto_ahash.c > >diff --git a/drivers/crypto/rockchip/Makefile >b/drivers/crypto/rockchip/Makefile index 7051c6c..30f9129 100644 >--- a/drivers/crypto/rockchip/Makefile >+++ b/drivers/crypto/rockchip/Makefile >@@ -1,3 +1,4 @@ > obj-$(CONFIG_CRYPTO_DEV_ROCKCHIP) += rk_crypto.o > rk_crypto-objs := rk3288_crypto.o \ > rk3288_crypto_ablkcipher.o \ >+ rk3288_crypto_ahash.o >diff --git a/drivers/crypto/rockchip/rk3288_crypto.c >b/drivers/crypto/rockchip/rk3288_crypto.c index 82f3044..67d69d2 100644 >--- a/drivers/crypto/rockchip/rk3288_crypto.c >+++ b/drivers/crypto/rockchip/rk3288_crypto.c >@@ -190,7 +190,6 @@ static void rk_crypto_tasklet_cb(unsigned long data) > { > struct rk_crypto_info *dev = (struct rk_crypto_info *)data; > struct crypto_async_request *async_req, *backlog; >- struct rk_cipher_reqctx *ablk_reqctx; > int err = 0; > unsigned long flags; > >@@ -207,10 +206,10 @@ static void rk_crypto_tasklet_cb(unsigned long data) > backlog = NULL; > } > >- if (crypto_tfm_alg_type(async_req->tfm) == CRYPTO_ALG_TYPE_ABLKCIPHER) { >+ if (crypto_tfm_alg_type(async_req->tfm) == CRYPTO_ALG_TYPE_ABLKCIPHER) > dev->ablk_req = ablkcipher_request_cast(async_req); >- ablk_reqctx = ablkcipher_request_ctx(dev->ablk_req); >- } >+ else >+ dev->ahash_req = ahash_request_cast(async_req); > err = dev->start(dev); > if (err) > dev->complete(dev, err); >@@ -223,6 +222,9 @@ static struct rk_crypto_tmp *rk_cipher_algs[] = { > &rk_cbc_des_alg, > &rk_ecb_des3_ede_alg, > &rk_cbc_des3_ede_alg, >+ &rk_ahash_sha1, >+ &rk_ahash_sha256, >+ &rk_ahash_md5, > }; > > static int rk_crypto_register(struct rk_crypto_info *crypto_info) >@@ -232,15 +234,24 @@ static int rk_crypto_register(struct rk_crypto_info >*crypto_info) > > for (i = 0; i < ARRAY_SIZE(rk_cipher_algs); i++) { > rk_cipher_algs[i]->dev = crypto_info; >- err = crypto_register_alg(&rk_cipher_algs[i]->alg); >+ if (rk_cipher_algs[i]->type == ALG_TYPE_CIPHER) >+ err = crypto_register_alg( >+ &rk_cipher_algs[i]->alg.crypto); >+ else >+ err = crypto_register_ahash( >+ &rk_cipher_algs[i]->alg.hash); > if (err) > goto err_cipher_algs; > } > return 0; > > err_cipher_algs: >- for (k = 0; k < i; k++) >- crypto_unregister_alg(&rk_cipher_algs[k]->alg); >+ for (k = 0; k < i; k++) { >+ if (rk_cipher_algs[i]->type == ALG_TYPE_CIPHER) >+ crypto_unregister_alg(&rk_cipher_algs[k]->alg.crypto); >+ else >+ crypto_unregister_ahash(&rk_cipher_algs[i]->alg.hash); >+ } > return err; > } > >@@ -248,8 +259,12 @@ static void rk_crypto_unregister(void) > { > unsigned int i; > >- for (i = 0; i < ARRAY_SIZE(rk_cipher_algs); i++) >- crypto_unregister_alg(&rk_cipher_algs[i]->alg); >+ for (i = 0; i < ARRAY_SIZE(rk_cipher_algs); i++) { >+ if (rk_cipher_algs[i]->type == ALG_TYPE_CIPHER) >+ crypto_unregister_alg(&rk_cipher_algs[i]->alg.crypto); >+ else >+ crypto_unregister_ahash(&rk_cipher_algs[i]->alg.hash); >+ } > } > > static void rk_crypto_action(void *data) >diff --git a/drivers/crypto/rockchip/rk3288_crypto.h >b/drivers/crypto/rockchip/rk3288_crypto.h index 604ffe7..453a00f 100644 >--- a/drivers/crypto/rockchip/rk3288_crypto.h >+++ b/drivers/crypto/rockchip/rk3288_crypto.h >@@ -6,6 +6,10 @@ > #include <crypto/algapi.h> > #include <linux/interrupt.h> > #include <linux/delay.h> >+#include <crypto/internal/hash.h> >+ >+#include "crypto/md5.h" >+#include "crypto/sha.h" > > #define _SBF(v, f) ((v) << (f)) > >@@ -149,6 +153,28 @@ > #define RK_CRYPTO_TDES_KEY3_0 0x0130 > #define RK_CRYPTO_TDES_KEY3_1 0x0134 > >+/* HASH */ >+#define RK_CRYPTO_HASH_CTRL 0x0180 >+#define RK_CRYPTO_HASH_SWAP_DO BIT(3) >+#define RK_CRYPTO_HASH_SWAP_DI BIT(2) >+#define RK_CRYPTO_HASH_SHA1 _SBF(0x00, 0) >+#define RK_CRYPTO_HASH_MD5 _SBF(0x01, 0) >+#define RK_CRYPTO_HASH_SHA256 _SBF(0x02, 0) >+#define RK_CRYPTO_HASH_PRNG _SBF(0x03, 0) >+ >+#define RK_CRYPTO_HASH_STS 0x0184 >+#define RK_CRYPTO_HASH_DONE BIT(0) >+ >+#define RK_CRYPTO_HASH_MSG_LEN 0x0188 >+#define RK_CRYPTO_HASH_DOUT_0 0x018c >+#define RK_CRYPTO_HASH_DOUT_1 0x0190 >+#define RK_CRYPTO_HASH_DOUT_2 0x0194 >+#define RK_CRYPTO_HASH_DOUT_3 0x0198 >+#define RK_CRYPTO_HASH_DOUT_4 0x019c >+#define RK_CRYPTO_HASH_DOUT_5 0x01a0 >+#define RK_CRYPTO_HASH_DOUT_6 0x01a4 >+#define RK_CRYPTO_HASH_DOUT_7 0x01a8 >+ > #define CRYPTO_READ(dev, offset) \ > readl_relaxed(((dev)->reg + (offset))) > #define CRYPTO_WRITE(dev, offset, val) \ >@@ -166,6 +192,7 @@ struct rk_crypto_info { > struct crypto_queue queue; > struct tasklet_struct crypto_tasklet; > struct ablkcipher_request *ablk_req; >+ struct ahash_request *ahash_req; > /* device lock */ > spinlock_t lock; > >@@ -194,6 +221,12 @@ struct rk_crypto_info { > struct scatterlist *sg_dst); > void (*unload_data)(struct rk_crypto_info *dev); > }; >+/* the private variable of hash */ >+struct rk_ahash_ctx { >+ struct rk_crypto_info *dev; >+ int FLAG_FINUP; >+ int first_op; >+}; > > /* the private variable of cipher */ > struct rk_cipher_ctx { >@@ -201,9 +234,18 @@ struct rk_cipher_ctx { > unsigned int keylen; > }; > >+enum alg_type { >+ ALG_TYPE_HASH, >+ ALG_TYPE_CIPHER, >+}; >+ > struct rk_crypto_tmp { >- struct rk_crypto_info *dev; >- struct crypto_alg alg; >+ struct rk_crypto_info *dev; >+ union { >+ struct crypto_alg crypto; >+ struct ahash_alg hash; >+ } alg; >+ enum alg_type type; > }; > > extern struct rk_crypto_tmp rk_ecb_aes_alg; >@@ -213,4 +255,8 @@ extern struct rk_crypto_tmp rk_cbc_des_alg; > extern struct rk_crypto_tmp rk_ecb_des3_ede_alg; > extern struct rk_crypto_tmp rk_cbc_des3_ede_alg; > >+extern struct rk_crypto_tmp rk_ahash_sha1; >+extern struct rk_crypto_tmp rk_ahash_sha256; >+extern struct rk_crypto_tmp rk_ahash_md5; >+ > #endif >diff --git a/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c >b/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c index 3624080..a260203 >100644 >--- a/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c >+++ b/drivers/crypto/rockchip/rk3288_crypto_ablkcipher.c >@@ -322,7 +322,7 @@ static int rk_ablk_cra_init(struct crypto_tfm *tfm) > struct crypto_alg *alg = tfm->__crt_alg; > struct rk_crypto_tmp *algt; > >- algt = container_of(alg, struct rk_crypto_tmp, alg); >+ algt = container_of(alg, struct rk_crypto_tmp, alg.crypto); > > ctx->dev = algt->dev; > ctx->dev->align_size = crypto_tfm_alg_alignmask(tfm) + 1; >@@ -343,7 +343,8 @@ static void rk_ablk_cra_exit(struct crypto_tfm *tfm) > } > > struct rk_crypto_tmp rk_ecb_aes_alg = { >- .alg = { >+ .type = ALG_TYPE_CIPHER, >+ .alg.crypto = { > .cra_name = "ecb(aes)", > .cra_driver_name = "ecb-aes-rk", > .cra_priority = 300, >@@ -367,7 +368,8 @@ struct rk_crypto_tmp rk_ecb_aes_alg = { > }; > > struct rk_crypto_tmp rk_cbc_aes_alg = { >- .alg = { >+ .type = ALG_TYPE_CIPHER, >+ .alg.crypto = { > .cra_name = "cbc(aes)", > .cra_driver_name = "cbc-aes-rk", > .cra_priority = 300, >@@ -392,7 +394,8 @@ struct rk_crypto_tmp rk_cbc_aes_alg = { > }; > > struct rk_crypto_tmp rk_ecb_des_alg = { >- .alg = { >+ .type = ALG_TYPE_CIPHER, >+ .alg.crypto = { > .cra_name = "ecb(des)", > .cra_driver_name = "ecb-des-rk", > .cra_priority = 300, >@@ -416,7 +419,8 @@ struct rk_crypto_tmp rk_ecb_des_alg = { > }; > > struct rk_crypto_tmp rk_cbc_des_alg = { >- .alg = { >+ .type = ALG_TYPE_CIPHER, >+ .alg.crypto = { > .cra_name = "cbc(des)", > .cra_driver_name = "cbc-des-rk", > .cra_priority = 300, >@@ -441,7 +445,8 @@ struct rk_crypto_tmp rk_cbc_des_alg = { > }; > > struct rk_crypto_tmp rk_ecb_des3_ede_alg = { >- .alg = { >+ .type = ALG_TYPE_CIPHER, >+ .alg.crypto = { > .cra_name = "ecb(des3_ede)", > .cra_driver_name = "ecb-des3-ede-rk", > .cra_priority = 300, >@@ -466,7 +471,8 @@ struct rk_crypto_tmp rk_ecb_des3_ede_alg = { > }; > > struct rk_crypto_tmp rk_cbc_des3_ede_alg = { >- .alg = { >+ .type = ALG_TYPE_CIPHER, >+ .alg.crypto = { > .cra_name = "cbc(des3_ede)", > .cra_driver_name = "cbc-des3-ede-rk", > .cra_priority = 300, >diff --git a/drivers/crypto/rockchip/rk3288_crypto_ahash.c >b/drivers/crypto/rockchip/rk3288_crypto_ahash.c new file mode 100644 >index 0000000..c049656 >--- /dev/null >+++ b/drivers/crypto/rockchip/rk3288_crypto_ahash.c >@@ -0,0 +1,369 @@ >+/* >+ * Crypto acceleration support for Rockchip RK3288 >+ * >+ * Copyright (c) 2015, Fuzhou Rockchip Electronics Co., Ltd >+ * >+ * Author: Zain Wang <zain.wang@xxxxxxxxxxxxxx> >+ * >+ * This program is free software; you can redistribute it and/or modify it >+ * under the terms and conditions of the GNU General Public License, >+ * version 2, as published by the Free Software Foundation. >+ * >+ * Some ideas are from marvell/cesa.c and s5p-sss.c driver. >+ */ >+#include "rk3288_crypto.h" >+ >+static u8 *outdata[3] = { >+ "\xda\x39\xa3\xee\x5e\x6b\x4b\x0d\x32\x55" >+ "\xbf\xef\x95\x60\x18\x90\xaf\xd8\x07\x09", >+ >+ "\xe3\xb0\xc4\x42\x98\xfc\x1c\x14" >+ "\x9a\xfb\xf4\xc8\x99\x6f\xb9\x24" >+ "\x27\xae\x41\xe4\x64\x9b\x93\x4c" >+ "\xa4\x95\x99\x1b\x78\x52\xb8\x55", >+ >+ "\xd4\x1d\x8c\xd9\x8f\x00\xb2\x04" >+ "\xe9\x80\x09\x98\xec\xf8\x42\x7e", >+}; >+ >+static void nodata_process(struct ahash_request *req) >+{ >+ struct crypto_ahash *tfm = crypto_ahash_reqtfm(req); >+ int rk_digest_size; >+ >+ rk_digest_size = crypto_ahash_digestsize(tfm); >+ >+ if (rk_digest_size == SHA1_DIGEST_SIZE) >+ memcpy(req->result, outdata[0], rk_digest_size); >+ else if (rk_digest_size == SHA256_DIGEST_SIZE) >+ memcpy(req->result, outdata[1], rk_digest_size); >+ else if (rk_digest_size == MD5_DIGEST_SIZE) >+ memcpy(req->result, outdata[2], rk_digest_size); >+} What is the purpose of this code? >+ >+static void rk_ahash_crypto_complete(struct rk_crypto_info *dev, int err) >+{ >+ if (dev->ahash_req->base.complete) >+ dev->ahash_req->base.complete(&dev->ahash_req->base, err); >+} >+ >+static void rk_ahash_hw_init(struct rk_crypto_info *dev) >+{ >+ int reg_status = 0; >+ >+ reg_status = CRYPTO_READ(dev, RK_CRYPTO_CTRL) | >+ RK_CRYPTO_HASH_FLUSH | >+ _SBF(0xffff, 16); >+ CRYPTO_WRITE(dev, RK_CRYPTO_CTRL, reg_status); >+ >+ reg_status = CRYPTO_READ(dev, RK_CRYPTO_CTRL); >+ reg_status &= (~RK_CRYPTO_HASH_FLUSH); >+ reg_status |= _SBF(0xffff, 16); >+ CRYPTO_WRITE(dev, RK_CRYPTO_CTRL, reg_status); >+ >+ memset_io(dev->reg + RK_CRYPTO_HASH_DOUT_0, 0, 32); >+} >+ >+static void rk_ahash_reg_init(struct rk_crypto_info *dev) >+{ >+ rk_ahash_hw_init(dev); >+ >+ CRYPTO_WRITE(dev, RK_CRYPTO_INTENA, RK_CRYPTO_HRDMA_ERR_ENA | >+ RK_CRYPTO_HRDMA_DONE_ENA); >+ >+ CRYPTO_WRITE(dev, RK_CRYPTO_INTSTS, RK_CRYPTO_HRDMA_ERR_INT | >+ RK_CRYPTO_HRDMA_DONE_INT); >+ >+ CRYPTO_WRITE(dev, RK_CRYPTO_HASH_CTRL, dev->mode | >+ RK_CRYPTO_HASH_SWAP_DO); >+ >+ CRYPTO_WRITE(dev, RK_CRYPTO_CONF, RK_CRYPTO_BYTESWAP_HRFIFO | >+ RK_CRYPTO_BYTESWAP_BRFIFO | >+ RK_CRYPTO_BYTESWAP_BTFIFO); >+} >+ >+static int rk_ahash_init(struct ahash_request *req) >+{ >+ struct crypto_ahash *tfm = crypto_ahash_reqtfm(req); >+ struct rk_ahash_ctx *tctx = crypto_tfm_ctx(req->base.tfm); >+ struct rk_crypto_info *dev = NULL; >+ int rk_digest_size; >+ >+ dev = tctx->dev; >+ dev->left_bytes = 0; >+ dev->aligned = 0; >+ dev->ahash_req = req; >+ dev->mode = 0; >+ dev->align_size = 4; >+ dev->sg_dst = NULL; >+ >+ tctx->first_op = 1; >+ >+ rk_digest_size = crypto_ahash_digestsize(tfm); >+ if (!rk_digest_size) >+ dev_err(dev->dev, "can't get digestsize\n"); >+ if (rk_digest_size == SHA1_DIGEST_SIZE) >+ dev->mode = RK_CRYPTO_HASH_SHA1; >+ else if (rk_digest_size == SHA256_DIGEST_SIZE) >+ dev->mode = RK_CRYPTO_HASH_SHA256; >+ else if (rk_digest_size == MD5_DIGEST_SIZE) >+ dev->mode = RK_CRYPTO_HASH_MD5; >+ >+ rk_ahash_reg_init(dev); >+ return 0; >+} >+ >+static int rk_ahash_final(struct ahash_request *req) >+{ >+ struct rk_ahash_ctx *tctx = crypto_tfm_ctx(req->base.tfm); >+ struct rk_crypto_info *dev = tctx->dev; >+ struct crypto_ahash *tfm = crypto_ahash_reqtfm(dev->ahash_req); >+ >+ if (!dev->total) { >+ nodata_process(dev->ahash_req); >+ return 0; >+ } >+ >+ /* >+ * IC should process the result again after last dma interrupt. >+ * And the last processing is very quick so than it may entry >+ * interrupt before finishing last interrupt. >+ * So I don't use interrupt finished hash. >+ */ >+ while (!CRYPTO_READ(dev, RK_CRYPTO_HASH_STS)) >+ usleep_range(50, 100); >+ >+ memcpy_fromio(dev->ahash_req->result, dev->reg + RK_CRYPTO_HASH_DOUT_0, >+ crypto_ahash_digestsize(tfm)); >+ return 0; >+} >+ >+static int rk_ahash_update(struct ahash_request *req) >+{ >+ struct rk_ahash_ctx *tctx = crypto_tfm_ctx(req->base.tfm); >+ struct rk_crypto_info *dev = tctx->dev; >+ int ret; >+ >+ dev->total = req->nbytes; >+ dev->left_bytes = req->nbytes; >+ dev->sg_src = req->src; >+ dev->first = req->src; >+ dev->nents = sg_nents(req->src); >+ >+ /* IC can calculate 0 data hash, so it should finish update here */ >+ if (!dev->total) { >+ pr_err("[%s:%d] no data\n", __func__, __LINE__); >+ return 0; >+ } >+ >+ if (tctx->first_op) { >+ tctx->first_op = 0; >+ CRYPTO_WRITE(dev, RK_CRYPTO_HASH_MSG_LEN, dev->total); >+ } else { >+ /* >+ * IC must know the length of total data at first, >+ * multiple updatings cannot support this variable. >+ */ >+ dev_warn(dev->dev, "Cannot carry multiple updatings!\n"); >+ return 0; >+ } >+ spin_lock(&dev->lock); >+ ret = crypto_enqueue_request(&dev->queue, &req->base); >+ spin_unlock(&dev->lock); >+ >+ tasklet_schedule(&dev->crypto_tasklet); >+ >+ return ret; >+} >+ >+static int rk_ahash_finup(struct ahash_request *req) >+{ >+ struct rk_ahash_ctx *tctx = crypto_tfm_ctx(req->base.tfm); >+ int err; >+ >+ /* >+ * finup should should process one updating and final. >+ * and we should wait for updating in finup so that we can >+ * fetching result by calling rk_ahash_final in finup. >+ */ >+ >+ tctx->FLAG_FINUP = 1; >+ err = rk_ahash_update(req); >+ if (err == -EINPROGRESS || err == -EBUSY) >+ while (tctx->FLAG_FINUP) >+ usleep_range(50, 500); >+ >+ return rk_ahash_final(req); >+} >+ >+static int rk_ahash_digest(struct ahash_request *req) >+{ >+ return rk_ahash_init(req) ? -EINVAL : rk_ahash_finup(req); >+} >+ >+static void crypto_ahash_dma_start(struct rk_crypto_info *dev) >+{ >+ CRYPTO_WRITE(dev, RK_CRYPTO_HRDMAS, dev->addr_in); >+ CRYPTO_WRITE(dev, RK_CRYPTO_HRDMAL, (dev->count + 3) / 4); >+ CRYPTO_WRITE(dev, RK_CRYPTO_CTRL, RK_CRYPTO_HASH_START | >+ (RK_CRYPTO_HASH_START << 16)); >+} >+ >+static int rk_ahash_set_data_start(struct rk_crypto_info *dev) >+{ >+ int err; >+ >+ err = dev->load_data(dev, dev->sg_src, NULL); >+ if (!err) >+ crypto_ahash_dma_start(dev); >+ return err; >+} >+ >+static int rk_ahash_start(struct rk_crypto_info *dev) >+{ >+ return rk_ahash_set_data_start(dev); >+} >+ >+/* >+ * return: >+ * true: some err was occurred >+ * fault: no err, please continue >+ */ >+static int rk_ahash_crypto_rx(struct rk_crypto_info *dev) >+{ >+ int err = 0; >+ struct rk_ahash_ctx *tctx = crypto_tfm_ctx(dev->ahash_req->base.tfm); >+ >+ dev->unload_data(dev); >+ if (dev->left_bytes) { >+ if (dev->aligned) { >+ if (sg_is_last(dev->sg_src)) { >+ dev_warn(dev->dev, "[%s:%d], lack of data\n", >+ __func__, __LINE__); >+ err = -ENOMEM; >+ goto out_rx; >+ } >+ dev->sg_src = sg_next(dev->sg_src); >+ } >+ err = rk_ahash_set_data_start(dev); >+ } else { >+ tctx->FLAG_FINUP = 0; >+ dev->complete(dev, 0); >+ return 0; >+ } >+ >+out_rx: >+ return err; >+} >+ >+static int rk_cra_hash_init(struct crypto_tfm *tfm) >+{ >+ struct rk_ahash_ctx *tctx = crypto_tfm_ctx(tfm); >+ struct rk_crypto_tmp *algt; >+ struct ahash_alg *alg = __crypto_ahash_alg(tfm->__crt_alg); >+ >+ algt = container_of(alg, struct rk_crypto_tmp, alg.hash); >+ >+ tctx->dev = algt->dev; >+ tctx->dev->addr_vir = (void *)__get_free_page(GFP_KERNEL); >+ if (!tctx->dev->addr_vir) { >+ pr_err("failed to kmalloc for addr_vir\n"); >+ return -ENOMEM; >+ } >+ tctx->dev->start = rk_ahash_start; >+ tctx->dev->update = rk_ahash_crypto_rx; >+ tctx->dev->complete = rk_ahash_crypto_complete; >+ return tctx->dev->enable_clk(tctx->dev); >+} >+ >+static void rk_cra_hash_exit(struct crypto_tfm *tfm) >+{ >+ struct rk_ahash_ctx *tctx = crypto_tfm_ctx(tfm); >+ >+ free_page((unsigned long)tctx->dev->addr_vir); >+ return tctx->dev->disable_clk(tctx->dev); >+} >+ >+struct rk_crypto_tmp rk_ahash_sha1 = { >+ .type = ALG_TYPE_HASH, >+ .alg.hash = { >+ .init = rk_ahash_init, >+ .update = rk_ahash_update, >+ .final = rk_ahash_final, >+ .finup = rk_ahash_finup, >+ .digest = rk_ahash_digest, >+ .halg = { >+ .digestsize = SHA1_DIGEST_SIZE, >+ .statesize = sizeof(struct sha1_state), >+ .base = { >+ .cra_name = "sha1", >+ .cra_driver_name = "rk-sha1", >+ .cra_priority = 300, >+ .cra_flags = CRYPTO_ALG_ASYNC | >+ CRYPTO_ALG_NEED_FALLBACK, >+ .cra_blocksize = SHA1_BLOCK_SIZE, >+ .cra_ctxsize = sizeof(struct rk_ahash_ctx), >+ .cra_alignmask = 3, >+ .cra_init = rk_cra_hash_init, >+ .cra_exit = rk_cra_hash_exit, >+ .cra_module = THIS_MODULE, >+ } >+ } >+ } >+}; >+ >+struct rk_crypto_tmp rk_ahash_sha256 = { >+ .type = ALG_TYPE_HASH, >+ .alg.hash = { >+ .init = rk_ahash_init, >+ .update = rk_ahash_update, >+ .final = rk_ahash_final, >+ .finup = rk_ahash_finup, >+ .digest = rk_ahash_digest, >+ .halg = { >+ .digestsize = SHA256_DIGEST_SIZE, >+ .statesize = sizeof(struct sha256_state), >+ .base = { >+ .cra_name = "sha256", >+ .cra_driver_name = "rk-sha256", >+ .cra_priority = 300, >+ .cra_flags = CRYPTO_ALG_ASYNC | >+ CRYPTO_ALG_NEED_FALLBACK, >+ .cra_blocksize = SHA256_BLOCK_SIZE, >+ .cra_ctxsize = sizeof(struct rk_ahash_ctx), >+ .cra_alignmask = 0, >+ .cra_init = rk_cra_hash_init, >+ .cra_exit = rk_cra_hash_exit, >+ .cra_module = THIS_MODULE, >+ } >+ } >+ } >+}; >+ >+struct rk_crypto_tmp rk_ahash_md5 = { >+ .type = ALG_TYPE_HASH, >+ .alg.hash = { >+ .init = rk_ahash_init, >+ .update = rk_ahash_update, >+ .final = rk_ahash_final, >+ .finup = rk_ahash_finup, >+ .digest = rk_ahash_digest, >+ .halg = { >+ .digestsize = MD5_DIGEST_SIZE, >+ .statesize = sizeof(struct md5_state), >+ .base = { >+ .cra_name = "md5", >+ .cra_driver_name = "rk-md5", >+ .cra_priority = 300, >+ .cra_flags = CRYPTO_ALG_ASYNC | >+ CRYPTO_ALG_NEED_FALLBACK, >+ .cra_blocksize = SHA1_BLOCK_SIZE, >+ .cra_ctxsize = sizeof(struct rk_ahash_ctx), >+ .cra_alignmask = 0, >+ .cra_init = rk_cra_hash_init, >+ .cra_exit = rk_cra_hash_exit, >+ .cra_module = THIS_MODULE, >+ } >+ } >+ } >+}; Ciao Stephan -- To unsubscribe from this list: send the line "unsubscribe linux-crypto" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html