[RFC PATCH] crypto: Alchemy AES engine driver

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Driver for the AES engine in ECB and CBC modes found on some
Alchemy Au1200/Au1300 SoCs.

Signed-off-by: Manuel Lauss <manuel.lauss@xxxxxxxxx>
---

lightly "tested" with the tcrypt module on Au1200;  I have no idea whether
it really works correctly:

# modprobe alchemy-aes
alg: skcipher: setkey failed on test 2 for ecb-aes-alchemy: flags=200000
# modprobe tcrypt mode=10
alg: skcipher: setkey failed on test 3 for cbc-aes-alchemy: flags=0
alg: skcipher: Failed to load transform for cbc(aes): -2
alg: skcipher: Failed to load transform for cbc(aes): -2
tcrypt: one or more tests failed!
FATAL: Error inserting tcrypt (/lib/modules/2.6.34-rc6-db1200-00214-g9f84af9/kernel/crypto/tcrypt.ko): Unknown symbol in module, or unknown parameter (see dmesg)

The error in "test 3 for cbc-aes-alchemy" probably comes from the inability
to process keys larger than 128 bits.

Please have a look.
 Thanks!

 drivers/crypto/Kconfig       |    8 +
 drivers/crypto/Makefile      |    1 +
 drivers/crypto/alchemy-aes.c |  579 ++++++++++++++++++++++++++++++++++++++++++
 3 files changed, 588 insertions(+), 0 deletions(-)
 create mode 100644 drivers/crypto/alchemy-aes.c

diff --git a/drivers/crypto/Kconfig b/drivers/crypto/Kconfig
index b08403d..7705b13 100644
--- a/drivers/crypto/Kconfig
+++ b/drivers/crypto/Kconfig
@@ -222,4 +222,12 @@ config CRYPTO_DEV_PPC4XX
 	help
 	  This option allows you to have support for AMCC crypto acceleration.
 
+config CRYPTO_DEV_ALCHEMY_AES
+	tristate "Au1200/Au1300 AES engine"
+	depends on MACH_ALCHEMY
+	select CRYPTO_ALGAPI
+	select CRYPTO_BLKCIPHER
+	help
+	  Driver for the AES engine in Alchemy Au1200/Au1300 series SoCs.
+
 endif # CRYPTO_HW
diff --git a/drivers/crypto/Makefile b/drivers/crypto/Makefile
index 6ffcb3f..624777f 100644
--- a/drivers/crypto/Makefile
+++ b/drivers/crypto/Makefile
@@ -6,3 +6,4 @@ obj-$(CONFIG_CRYPTO_DEV_MV_CESA) += mv_cesa.o
 obj-$(CONFIG_CRYPTO_DEV_TALITOS) += talitos.o
 obj-$(CONFIG_CRYPTO_DEV_IXP4XX) += ixp4xx_crypto.o
 obj-$(CONFIG_CRYPTO_DEV_PPC4XX) += amcc/
+obj-$(CONFIG_CRYPTO_DEV_ALCHEMY_AES) += alchemy-aes.o
diff --git a/drivers/crypto/alchemy-aes.c b/drivers/crypto/alchemy-aes.c
new file mode 100644
index 0000000..14e8ace
--- /dev/null
+++ b/drivers/crypto/alchemy-aes.c
@@ -0,0 +1,579 @@
+/*
+ * alchemy-aes.c -- Driver for the Alchemy Au1200/Au1300 AES engine.
+ *
+ * (c) 2010 Manuel Lauss <manuel.lauss@xxxxxxxxx>
+ *
+ * Licensed under the GPLv2.
+ */
+
+#include <linux/crypto.h>
+#include <linux/kernel.h>
+#include <linux/list.h>
+#include <linux/slab.h>
+#include <linux/interrupt.h>
+#include <linux/completion.h>
+#include <linux/module.h>
+#include <linux/platform_device.h>
+#include <linux/resource.h>
+
+#include <crypto/algapi.h>
+#include <crypto/aes.h>
+
+#include <asm/mach-au1x00/au1000.h>
+#include <asm/mach-au1x00/au1xxx_dbdma.h>
+
+/*#define DEBUG*/
+
+#ifdef DEBUG
+#define DBG(x...) printk(KERN_ALERT "AES: " x)
+#define DBGHEX(x...)	print_hex_dump(x)
+#else
+#define DBG(x...)
+#define DBGHEX(x...)
+#endif
+
+#define DRVNAME	"alchemy-aes"
+
+/* the AES engine likes the number 16 */
+#define ALCHEMY_AES_KEY_SIZE	16
+#define ALCHEMY_AES_IV_SIZE	16
+#define ALCHEMY_AES_BLK_SIZE	16
+#define ALCHEMY_AES_DDMA_DSCRS	16
+
+/* register offsets */
+#define AES_STATUS	0x00
+#define AES_INDATA	0x04
+#define AES_OUTDATA	0x08
+#define AES_INTCAUSE	0x0c
+#define AES_CONFIG	0x10
+
+#define AES_S_PS	0x01	/* start crypto */
+#define AES_S_IE	0x02	/* int enable */
+#define AES_S_CR_1	0x00	/* periph. clock  */
+#define AES_S_CR_2	0x04	/* periph. clock div 2 */
+#define AES_S_CR_4	0x08	/* periph. clock div 4 */
+#define AES_S_CR_8	0x0c	/* periph. clock div 8 */
+#define AES_S_CR_MASK	0x0c
+#define AES_S_OUT	0x10	/* set if out fifo has space */
+#define AES_S_IN	0x20	/* set if in fifo has space */
+
+#define AES_CAUSE_RDY	0x01	/* process complete */
+#define AES_CAUSE_OVR	0x02	/* in fifo overflow */
+#define AES_CAUSE_UND	0x04	/* out fifo underflow */
+
+#define AES_C_ED	0x01	/* set to DE/clear to ENcrypt */
+#define AES_C_IKG	0x02	/* do internal key generation */
+#define AES_C_RPK	0x04	/* replay key (output 10th key) */
+#define AES_C_RK	0x08	/* reuse internal key store */
+#define AES_C_UC	0x10	/* undefined block count */
+#define AES_C_OP_ECB	0x00
+#define AES_C_OP_CBC	0x20
+#define AES_C_OP_CFB	0x40
+#define AES_C_OP_OFB	0x60
+#define AES_C_OP_MASK	0x60
+
+
+struct alchemy_aes_priv {
+	void __iomem *base;
+	int irq;
+
+	/* dbdma */
+	u32 dmatx;
+	u32 dmarx;
+	int txid;
+	int rxid;
+
+	unsigned char key[ALCHEMY_AES_KEY_SIZE];
+
+	struct completion done;
+	struct list_head alg_list;
+	struct resource *ioarea;
+};
+
+/* need to improvise to stash private data */
+struct alchemy_aes_alg {
+	struct crypto_alg alg;
+	struct alchemy_aes_priv *priv;
+	struct list_head entry;
+};
+
+#define crypto_alg_to_alchemy(a) \
+	container_of(a, struct alchemy_aes_alg, alg)
+
+
+/* 8 bit memory device */
+static dbdev_tab_t alchemy_aes_mem_dbdev = {
+	.dev_id		= DSCR_CMD0_ALWAYS,
+	.dev_flags	= DEV_FLAGS_ANYUSE,
+	.dev_tsize	= 0,
+	.dev_devwidth	= 8,
+	.dev_physaddr	= 0x00000000,
+	.dev_intlevel	= 0,
+	.dev_intpolarity = 0,
+};
+static int __alchemy_aes_memid;
+
+/**********************************************************************/
+
+static inline void aes_write(struct alchemy_aes_priv *p, int reg,
+			     unsigned long val)
+{
+	__raw_writel(val, p->base + reg);
+}
+
+static inline unsigned long aes_read(struct alchemy_aes_priv *p, int reg)
+{
+	return __raw_readl(p->base + reg);
+}
+
+static inline void aes_infifo_write(struct alchemy_aes_priv *p,
+				    unsigned long data)
+{
+	/* poll for open space */
+	while (!(aes_read(p, AES_STATUS) & AES_S_IN))
+		cpu_relax();
+	aes_write(p, AES_INDATA, data);
+}
+
+/**********************************************************************/
+
+static void alchemy_aes_dmacb(int irq, void *data)
+{
+	struct alchemy_aes_priv *p = data;
+
+	au1xxx_dbdma_stop(p->dmarx);
+	au1xxx_dbdma_stop(p->dmatx);
+
+	complete(&p->done);
+}
+
+static irqreturn_t alchemy_aes_irq(int irq, void *data)
+{
+	struct alchemy_aes_priv *p = data;
+	unsigned long istat;
+
+	istat = aes_read(p, AES_INTCAUSE);
+	aes_write(p, AES_INTCAUSE, 0);
+
+	if (istat & AES_CAUSE_RDY)
+		complete(&p->done);
+	/* fifo over/underruns shouldn't happen with DMA */
+
+	return IRQ_HANDLED;
+}
+
+/**********************************************************************/
+
+static int alchemy_aes_crypt(struct blkcipher_desc *desc,
+			     struct scatterlist *sdst,
+			     struct scatterlist *ssrc,
+			     unsigned int nbytes,
+			     unsigned long cryptmode)
+{
+	struct crypto_alg *alg = desc->tfm->base.__crt_alg;
+	struct alchemy_aes_alg *aa = crypto_alg_to_alchemy(alg);
+	struct alchemy_aes_priv *p = aa->priv;
+	struct blkcipher_walk walk;
+	int ret, i, blks, len;
+	unsigned long k;
+	dma_addr_t src, dst;
+	unsigned char *iv;
+
+	DBG("crypt: ALG %s base %p nbytes %d mode %08lx\n",
+	    alg->cra_name, p->base, nbytes, cryptmode);
+
+	blkcipher_walk_init(&walk, sdst, ssrc, nbytes);
+	ret = blkcipher_walk_virt(desc, &walk);
+	iv = walk.iv;
+
+	/* create DMA descriptor lists */
+	i = blks = 0;
+	au1xxx_dbdma_reset(p->dmatx);
+	au1xxx_dbdma_reset(p->dmarx);
+	while ((nbytes = walk.nbytes)) {
+		src = (dma_addr_t)virt_to_phys(walk.src.virt.addr);
+		dst = (dma_addr_t)virt_to_phys(walk.dst.virt.addr);
+
+		blks += nbytes / ALCHEMY_AES_BLK_SIZE;
+		len = nbytes - (nbytes % ALCHEMY_AES_BLK_SIZE);
+
+		DBG("SRC %p DST %p bytes %d blks %d dscr %d len %d\n",
+		    (void *)src, (void *)dst, nbytes, blks, i, len);
+		DBGHEX(KERN_ERR, "AES DATA ", 0, 16, 1, walk.src.virt.addr,
+		       len, false);
+
+		nbytes -= len;
+		ret = blkcipher_walk_done(desc, &walk, nbytes);
+
+		/* let DBDMA interrupt when last block of data has been
+		 * fetched from AES output fifo.
+		 */
+		au1xxx_dbdma_put_dest(p->dmarx, dst, len,
+				      walk.nbytes ? 0 : DDMA_FLAGS_IE);
+		au1xxx_dbdma_put_source(p->dmatx, src, len, 0);
+
+		if (++i > ALCHEMY_AES_DDMA_DSCRS) {
+			ret = -ENOMEM;
+			printk(KERN_ERR DRVNAME ": DMA descriptor limit"
+				" reached!\n");
+			goto out;
+		}
+	}
+
+	/* drain output fifo, clear interrupts, configure and enable */
+	init_completion(&p->done);
+	for (i = 0; i < 8; i++)
+		k = aes_read(p, AES_OUTDATA);
+	aes_write(p, AES_INTCAUSE, 0);
+	aes_write(p, AES_CONFIG, cryptmode);
+	aes_write(p, AES_STATUS, AES_S_PS | AES_S_IE);	/* start engine */
+
+	/* block count comes first */
+	if (!(cryptmode & AES_C_UC)) {
+		DBG("BLKCNT %d\n", blks);
+		aes_infifo_write(p, blks);
+	}
+
+	/* then new key if necessary */
+	if (!(cryptmode & AES_C_RK)) {
+		DBGHEX(KERN_ERR, "AES: KEY ", 0, 16, 1, p->key, 16, false);
+
+		k = (p->key[0] << 24) | (p->key[1] << 16) |
+		    (p->key[2] <<  8) |  p->key[3];
+		aes_infifo_write(p, k);
+		k = (p->key[4] << 24) | (p->key[5] << 16) |
+		    (p->key[6] <<  8) |  p->key[7];
+		aes_infifo_write(p, k);
+		k = (p->key[8]  << 24) | (p->key[9]  << 16) |
+		    (p->key[10] <<  8) |  p->key[11];
+		aes_infifo_write(p, k);
+		k = (p->key[12] << 24) | (p->key[13] << 16) |
+		    (p->key[14] <<  8) |  p->key[15];
+		aes_infifo_write(p, k);
+	}
+
+	/* then IV for non-ECB modes */
+	if ((cryptmode & AES_C_OP_MASK) != AES_C_OP_ECB) {
+		DBGHEX(KERN_ERR, "AES: IV  ", 0, 16, 1, iv, 16, false);
+
+		k = (iv[0] << 24) | (iv[1] << 16) |
+		    (iv[2] <<  8) |  iv[3];
+		aes_infifo_write(p, k);
+		k = (iv[4] << 24) | (iv[5] << 16) |
+		    (iv[6] <<  8) |  iv[7];
+		aes_infifo_write(p, k);
+		k = (iv[8]  << 24) | (iv[9]  << 16) |
+		    (iv[10] <<  8) |  iv[11];
+		aes_infifo_write(p, k);
+		k = (iv[12] << 24) | (iv[13] << 16) |
+		    (iv[14] <<  8) |  iv[15];
+		aes_infifo_write(p, k);
+	}
+
+	/* DMA engine does data transfer */
+	au1xxx_dbdma_start(p->dmarx);
+	au1xxx_dbdma_start(p->dmatx);
+	ret = wait_for_completion_interruptible(&p->done);
+
+out:
+	aes_write(p, AES_STATUS, 0);
+	au1xxx_dbdma_stop(p->dmatx);
+	au1xxx_dbdma_stop(p->dmarx);
+
+	/* read replayed key if necessary */
+	if ((!ret) && ((cryptmode & AES_C_OP_MASK) == AES_C_OP_ECB) &&
+	    (cryptmode & (AES_C_ED | AES_C_RPK)) &&
+	    (!(cryptmode & AES_C_UC))) {
+		k = aes_read(p, AES_OUTDATA);
+		k = aes_read(p, AES_OUTDATA);
+		k = aes_read(p, AES_OUTDATA);
+		k = aes_read(p, AES_OUTDATA);
+		/* FIXME: does cryptoapi acutally want it? */
+	}
+
+	return ret;
+}
+
+static int alchemy_aes_setkey(struct crypto_tfm *tfm, const u8 *key,
+			      unsigned int len)
+{
+	struct crypto_alg *alg = tfm->__crt_alg;
+	struct alchemy_aes_alg *aa = crypto_alg_to_alchemy(alg);
+	struct alchemy_aes_priv *p = aa->priv;
+
+	DBGHEX(KERN_ERR, "AES: SETKEY ", 0, 16, 1, key, len, false);
+
+	if (len != AES_KEYSIZE_128) {
+		tfm->crt_flags |= CRYPTO_TFM_RES_BAD_KEY_LEN;
+		return -EINVAL;
+	}
+
+	memcpy(p->key, key, ALCHEMY_AES_KEY_SIZE);
+	return 0;
+}
+
+static int alchemy_aes_ecb_enc(struct blkcipher_desc *desc,
+			       struct scatterlist *dst,
+			       struct scatterlist *src,
+			       unsigned int nbytes)
+{
+	return alchemy_aes_crypt(desc, dst, src, nbytes,
+				 AES_C_OP_ECB);
+}
+
+static int alchemy_aes_ecb_dec(struct blkcipher_desc *desc,
+			       struct scatterlist *dst,
+			       struct scatterlist *src,
+			       unsigned int nbytes)
+{
+	return alchemy_aes_crypt(desc, dst, src, nbytes,
+				 AES_C_OP_ECB | AES_C_ED);
+}
+
+static struct crypto_alg alchemy_aes_ecb_alg = {
+	.cra_name		=	"ecb(aes)",
+	.cra_driver_name	=	"ecb-aes-alchemy",
+	.cra_u = {
+		.blkcipher = {
+			.min_keysize	= ALCHEMY_AES_KEY_SIZE,
+			.max_keysize	= ALCHEMY_AES_KEY_SIZE,
+			.setkey		= alchemy_aes_setkey,
+			.encrypt	= alchemy_aes_ecb_enc,
+			.decrypt	= alchemy_aes_ecb_dec,
+		},
+	},
+};
+
+static int alchemy_aes_cbc_enc(struct blkcipher_desc *desc,
+			       struct scatterlist *dst,
+			       struct scatterlist *src,
+			       unsigned int nbytes)
+{
+	return alchemy_aes_crypt(desc, dst, src, nbytes,
+				 AES_C_OP_CBC);
+}
+
+static int alchemy_aes_cbc_dec(struct blkcipher_desc *desc,
+			       struct scatterlist *dst,
+			       struct scatterlist *src,
+			       unsigned int nbytes)
+{
+	return alchemy_aes_crypt(desc, dst, src, nbytes,
+				 AES_C_OP_CBC | AES_C_ED);
+}
+
+static struct crypto_alg alchemy_aes_cbc_alg = {
+	.cra_name		=	"cbc(aes)",
+	.cra_driver_name	=	"cbc-aes-alchemy",
+	.cra_u = {
+		.blkcipher = {
+			.min_keysize	= ALCHEMY_AES_KEY_SIZE,
+			.max_keysize	= ALCHEMY_AES_KEY_SIZE,
+			.ivsize		= ALCHEMY_AES_IV_SIZE,
+			.setkey		= alchemy_aes_setkey,
+			.encrypt	= alchemy_aes_cbc_enc,
+			.decrypt	= alchemy_aes_cbc_dec,
+		},
+	},
+};
+
+static int __init add_algo(struct alchemy_aes_priv *p,
+			   struct crypto_alg *calg)
+{
+	struct alchemy_aes_alg *alg;
+	int ret;
+
+	alg = kzalloc(sizeof(struct alchemy_aes_alg), GFP_KERNEL);
+	if (!alg)
+		return -ENOMEM;
+
+	alg->priv = p;
+	alg->alg.cra_priority = 300;
+	alg->alg.cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER | CRYPTO_ALG_ASYNC;
+	alg->alg.cra_blocksize = ALCHEMY_AES_BLK_SIZE;
+	alg->alg.cra_ctxsize = 0;
+	alg->alg.cra_alignmask = 3;
+	alg->alg.cra_type = &crypto_blkcipher_type;
+	alg->alg.cra_module = THIS_MODULE;
+	INIT_LIST_HEAD(&alg->alg.cra_list);
+	alg->alg.cra_u.blkcipher = calg->cra_u.blkcipher;
+	sprintf(alg->alg.cra_name, "%s", calg->cra_name);
+	sprintf(alg->alg.cra_driver_name, "%s", calg->cra_driver_name);
+
+	list_add_tail(&alg->entry, &p->alg_list);
+
+	ret = crypto_register_alg(&alg->alg);
+	if (ret) {
+		list_del(&alg->entry);
+		kfree(alg);
+	}
+	return ret;
+}
+
+static void delete_algos(struct alchemy_aes_priv *p)
+{
+	struct alchemy_aes_alg *a, *n;
+
+	list_for_each_entry_safe(a, n, &p->alg_list, entry) {
+		list_del(&a->entry);
+		crypto_unregister_alg(&a->alg);
+		kfree(a);
+	}
+}
+
+static int __init alchemy_aes_probe(struct platform_device *pdev)
+{
+	struct resource *r;
+	struct alchemy_aes_priv *p;
+	int ret;
+
+	p = kzalloc(sizeof(struct alchemy_aes_priv), GFP_KERNEL);
+	if (!p)
+		return -ENOMEM;
+
+	ret = -ENODEV;
+	r = platform_get_resource(pdev, IORESOURCE_MEM, 0);
+	if (!r)
+		goto out1;
+
+	p->ioarea = request_mem_region(r->start, resource_size(r),
+					DRVNAME);
+	if (!p->ioarea)
+		goto out1;
+
+	p->base = ioremap_nocache(r->start, resource_size(r));
+	if (!p->base)
+		goto out2;
+
+	r = platform_get_resource(pdev, IORESOURCE_IRQ, 0);
+	if (!r)
+		goto out3;
+	p->irq = r->start;
+
+	r = platform_get_resource_byname(pdev, IORESOURCE_DMA, "dmatx");
+	if (!r) {
+		r = platform_get_resource(pdev, IORESOURCE_DMA, 0);
+		if (!r)
+			goto out3;
+	}
+	p->txid = r->start;
+
+	r = platform_get_resource_byname(pdev, IORESOURCE_DMA, "dmarx");
+	if (!r) {
+		r = platform_get_resource(pdev, IORESOURCE_DMA, 1);
+		if (!r)
+			goto out3;
+	}
+	p->rxid = r->start;
+
+	p->dmatx = au1xxx_dbdma_chan_alloc(__alchemy_aes_memid, p->txid,
+					    alchemy_aes_dmacb, p);
+	if (!p->dmatx)
+		goto out3;
+
+	p->dmarx = au1xxx_dbdma_chan_alloc(p->rxid, __alchemy_aes_memid,
+					    alchemy_aes_dmacb, p);
+	if (!p->dmarx)
+		goto out4;
+
+	au1xxx_dbdma_ring_alloc(p->dmatx, ALCHEMY_AES_DDMA_DSCRS);
+	au1xxx_dbdma_ring_alloc(p->dmarx, ALCHEMY_AES_DDMA_DSCRS);
+
+	aes_write(p, AES_STATUS, 0);
+	aes_write(p, AES_CONFIG, 0);
+	aes_write(p, AES_INTCAUSE, 0);
+
+	ret = request_irq(p->irq, alchemy_aes_irq, IRQF_DISABLED,
+			  DRVNAME, p);
+	if (ret)
+		goto out5;
+
+	INIT_LIST_HEAD(&p->alg_list);
+
+	ret = add_algo(p, &alchemy_aes_ecb_alg);
+	if (ret)
+		goto out6;
+
+	ret = add_algo(p, &alchemy_aes_cbc_alg);
+	if (ret)
+		goto out6;
+
+	platform_set_drvdata(pdev, p);
+	return 0;
+
+out6:
+	delete_algos(p);
+	free_irq(p->irq, p);
+out5:
+	au1xxx_dbdma_chan_free(p->dmarx);
+out4:
+	au1xxx_dbdma_chan_free(p->dmatx);
+out3:
+	iounmap(p->base);
+out2:
+	release_resource(p->ioarea);
+	kfree(p->ioarea);
+out1:
+	kfree(p);
+	return ret;
+}
+
+static int __exit alchemy_aes_remove(struct platform_device *pdev)
+{
+	struct alchemy_aes_priv *p = platform_get_drvdata(pdev);
+
+	delete_algos(p);
+
+	au1xxx_dbdma_stop(p->dmatx);
+	au1xxx_dbdma_stop(p->dmarx);
+
+	aes_write(p, AES_STATUS, 0);
+	aes_write(p, AES_CONFIG, 0);
+	aes_write(p, AES_INTCAUSE, 0);
+
+	free_irq(p->irq, p);
+	au1xxx_dbdma_chan_free(p->dmarx);
+	au1xxx_dbdma_chan_free(p->dmatx);
+	iounmap(p->base);
+	release_resource(p->ioarea);
+	kfree(p->ioarea);
+	kfree(p);
+
+	return 0;
+}
+
+static struct platform_driver alchemy_aes_driver = {
+	.driver.name	= DRVNAME,
+	.driver.owner	= THIS_MODULE,
+	.probe		= alchemy_aes_probe,
+	.remove		= __devexit_p(alchemy_aes_remove),
+};
+
+static int __init alchemy_aes_load(void)
+{
+	/* FIXME: hier sollte auch noch der PRId des prozessors getestet
+	 * werden; Au1210 (0x0503xxxx) und einige Au1300 haben lt. Daten-
+	 * blatt keine AES engine.
+	 */
+
+	/* need to do 8bit accesses to memory to get correct data */
+	__alchemy_aes_memid = au1xxx_ddma_add_device(&alchemy_aes_mem_dbdev);
+	if (!__alchemy_aes_memid)
+		return -ENODEV;
+
+	return platform_driver_register(&alchemy_aes_driver);
+}
+
+static void __exit alchemy_aes_unload(void)
+{
+	if (__alchemy_aes_memid)
+		au1xxx_ddma_del_device(__alchemy_aes_memid);
+	platform_driver_unregister(&alchemy_aes_driver);
+}
+
+module_init(alchemy_aes_load);
+module_exit(alchemy_aes_unload);
+
+MODULE_LICENSE("GPL");
+MODULE_DESCRIPTION("Alchemy Au1200 AES engine driver");
+MODULE_AUTHOR("Manuel Lauss <manuel.lauss@xxxxxxxxx>");
-- 
1.7.0.4

--
To unsubscribe from this list: send the line "unsubscribe linux-crypto" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Kernel]     [Gnu Classpath]     [Gnu Crypto]     [DM Crypt]     [Netfilter]     [Bugtraq]

  Powered by Linux