Re: [PATCH v3 2/3] crypto: inside-secure: add SafeXcel EIP197 crypto engine driver

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Am Montag, 24. April 2017, 09:54:06 CEST schrieb Antoine Tenart:

Hi Antoine,

> +struct safexcel_cipher_ctx {
> +	struct safexcel_context base;
> +	struct safexcel_crypto_priv *priv;
> +
> +	enum safexcel_cipher_direction direction;
> +	u32 mode;
> +
> +	__le32 key[8];

Can you please help me find the location where this memory is zeroized when 
released?


> +	unsigned int key_len;
> +};
> +
> +static void safexcel_cipher_token(struct safexcel_cipher_ctx *ctx,
> +				  struct crypto_async_request *async,
> +				  struct safexcel_command_desc *cdesc,
> +				  u32 length)
> +{
> +	struct ablkcipher_request *req = ablkcipher_request_cast(async);
> +	struct safexcel_token *token;
> +	unsigned offset = 0;
> +
> +	if (ctx->mode == CONTEXT_CONTROL_CRYPTO_MODE_CBC) {
> +		offset = AES_BLOCK_SIZE / sizeof(u32);
> +		memcpy(cdesc->control_data.token, req->info, AES_BLOCK_SIZE);
> +
> +		cdesc->control_data.options |= EIP197_OPTION_4_TOKEN_IV_CMD;
> +	}
> +
> +	token = (struct safexcel_token *)(cdesc->control_data.token + offset);
> +
> +	token[0].opcode = EIP197_TOKEN_OPCODE_DIRECTION;
> +	token[0].packet_length = length;
> +	token[0].stat = EIP197_TOKEN_STAT_LAST_PACKET;
> +	token[0].instructions = EIP197_TOKEN_INS_LAST |
> +				EIP197_TOKEN_INS_TYPE_CRYTO |
> +				EIP197_TOKEN_INS_TYPE_OUTPUT;
> +}
> +
> +static int safexcel_aes_setkey(struct crypto_ablkcipher *ctfm, const u8
> *key, +			       unsigned int len)
> +{

You still use ablkcipher. I thought that it is on its way out in favor of the 
skcipher API. Why do you stick to ablkcipher?

Note, a change could be as simple as s/ablkcipher/skcipher/g

> +	struct crypto_tfm *tfm = crypto_ablkcipher_tfm(ctfm);
> +	struct safexcel_cipher_ctx *ctx = crypto_tfm_ctx(tfm);
> +	struct crypto_aes_ctx aes;
> +	int ret, i;
> +
> +	ret = crypto_aes_expand_key(&aes, key, len);
> +	if (ret) {
> +		crypto_ablkcipher_set_flags(ctfm, CRYPTO_TFM_RES_BAD_KEY_LEN);
> +		return ret;
> +	}
> +
> +	for (i = 0; i < len / sizeof(u32); i++) {
> +		if (ctx->key[i] != cpu_to_le32(aes.key_enc[i])) {
> +			ctx->base.needs_inv = true;
> +			break;
> +		}
> +	}
> +
> +	for (i = 0; i < len / sizeof(u32); i++)
> +		ctx->key[i] = cpu_to_le32(aes.key_enc[i]);
> +
> +	ctx->key_len = len;

memzero_explicit(aes)?

> +
> +	return 0;
> +}
> +
> +static int safexcel_context_control(struct safexcel_cipher_ctx *ctx,
> +				    struct safexcel_command_desc *cdesc)
> +{
> +	struct safexcel_crypto_priv *priv = ctx->priv;
> +	int ctrl_size;
> +
> +	if (ctx->direction == SAFEXCEL_ENCRYPT)
> +		cdesc->control_data.control0 |= CONTEXT_CONTROL_TYPE_CRYPTO_OUT;
> +	else
> +		cdesc->control_data.control0 |= CONTEXT_CONTROL_TYPE_CRYPTO_IN;
> +
> +	cdesc->control_data.control0 |= CONTEXT_CONTROL_KEY_EN;
> +	cdesc->control_data.control1 |= ctx->mode;
> +
> +	switch (ctx->key_len) {
> +	case AES_KEYSIZE_128:
> +		cdesc->control_data.control0 |= CONTEXT_CONTROL_CRYPTO_ALG_AES128;
> +		ctrl_size = 4;
> +		break;
> +	case AES_KEYSIZE_192:
> +		cdesc->control_data.control0 |= CONTEXT_CONTROL_CRYPTO_ALG_AES192;
> +		ctrl_size = 6;
> +		break;
> +	case AES_KEYSIZE_256:
> +		cdesc->control_data.control0 |= CONTEXT_CONTROL_CRYPTO_ALG_AES256;
> +		ctrl_size = 8;
> +		break;
> +	default:
> +		dev_err(priv->dev, "aes keysize not supported: %u\n",
> +			ctx->key_len);
> +		return -EINVAL;
> +	}
> +	cdesc->control_data.control0 |= CONTEXT_CONTROL_SIZE(ctrl_size);
> +
> +	return 0;
> +}
> +
> +static int safexcel_handle_result(struct safexcel_crypto_priv *priv, int
> ring, +				  struct crypto_async_request *async,
> +				  bool *should_complete, int *ret)
> +{
> +	struct ablkcipher_request *req = ablkcipher_request_cast(async);
> +	struct safexcel_result_desc *rdesc;
> +	int ndesc = 0;
> +
> +	*ret = 0;
> +
> +	spin_lock_bh(&priv->ring[ring].egress_lock);
> +	do {
> +		rdesc = safexcel_ring_next_rptr(priv, &priv->ring[ring].rdr);
> +		if (IS_ERR(rdesc)) {
> +			dev_err(priv->dev,
> +				"cipher: result: could not retrieve the result descriptor\n");
> +			*ret = PTR_ERR(rdesc);
> +			break;
> +		}
> +
> +		if (rdesc->result_data.error_code) {
> +			dev_err(priv->dev,
> +				"cipher: result: result descriptor error (%d)\n",
> +				rdesc->result_data.error_code);
> +			*ret = -EIO;
> +		}
> +
> +		ndesc++;
> +	} while (!rdesc->last_seg);
> +
> +	safexcel_complete(priv, ring);
> +	spin_unlock_bh(&priv->ring[ring].egress_lock);
> +
> +	if (req->src == req->dst) {
> +		dma_unmap_sg(priv->dev, req->src,
> +			     sg_nents_for_len(req->src, req->nbytes),
> +			     DMA_BIDIRECTIONAL);
> +	} else {
> +		dma_unmap_sg(priv->dev, req->src,
> +			     sg_nents_for_len(req->src, req->nbytes),
> +			     DMA_TO_DEVICE);
> +		dma_unmap_sg(priv->dev, req->dst,
> +			     sg_nents_for_len(req->dst, req->nbytes),
> +			     DMA_FROM_DEVICE);
> +	}
> +
> +	*should_complete = true;
> +
> +	return ndesc;
> +}
> +
> +static int safexcel_aes_send(struct crypto_async_request *async,
> +			     int ring, struct safexcel_request *request,
> +			     int *commands, int *results)
> +{
> +	struct ablkcipher_request *req = ablkcipher_request_cast(async);
> +	struct safexcel_cipher_ctx *ctx = crypto_tfm_ctx(req->base.tfm);
> +	struct safexcel_crypto_priv *priv = ctx->priv;
> +	struct safexcel_command_desc *cdesc;
> +	struct safexcel_result_desc *rdesc;
> +	struct scatterlist *sg;
> +	int nr_src, nr_dst, n_cdesc = 0, n_rdesc = 0, queued = req->nbytes;
> +	int i, ret = 0;
> +
> +	request->req = &req->base;
> +
> +	if (req->src == req->dst) {
> +		nr_src = dma_map_sg(priv->dev, req->src,
> +				    sg_nents_for_len(req->src, req->nbytes),
> +				    DMA_BIDIRECTIONAL);
> +		nr_dst = nr_src;
> +		if (!nr_src)
> +			return -EINVAL;
> +	} else {
> +		nr_src = dma_map_sg(priv->dev, req->src,
> +				    sg_nents_for_len(req->src, req->nbytes),
> +				    DMA_TO_DEVICE);
> +		if (!nr_src)
> +			return -EINVAL;
> +
> +		nr_dst = dma_map_sg(priv->dev, req->dst,
> +				    sg_nents_for_len(req->dst, req->nbytes),
> +				    DMA_FROM_DEVICE);
> +		if (!nr_dst) {
> +			dma_unmap_sg(priv->dev, req->src,
> +				     sg_nents_for_len(req->src, req->nbytes),
> +				     DMA_TO_DEVICE);
> +			return -EINVAL;
> +		}
> +	}
> +
> +	memcpy(ctx->base.ctxr->data, ctx->key, ctx->key_len);

Is ctxr->data properly zeroized?


Ciao
Stephan



[Index of Archives]     [Kernel]     [Gnu Classpath]     [Gnu Crypto]     [DM Crypt]     [Netfilter]     [Bugtraq]

  Powered by Linux