Re: [PATCH] crypto/caam: add backlogging support

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 9/9/2015 9:57 AM, Alex Porosanu wrote:
> caam_jr_enqueue() function returns -EBUSY once there are no
> more slots available in the JR, but it doesn't actually save
> the current request. This breaks the functionality of users
> that expect that even if there is no more space for the request,
> it is at least queued for later execution. In other words, all
> crypto transformations that request backlogging
> (i.e. have CRYPTO_TFM_REQ_MAY_BACKLOG set), will hang. Such an
> example is dm-crypt.
> The current patch solves this issue by setting a threshold after
> which caam_jr_enqueue() returns -EBUSY, but since the HW job ring
> isn't actually full, the job is enqueued.

You should mention the reason of not using the functions and mechanism
available in the Crypto API, i.e. having a 0-length crypto_queue used
only for backlogging.

> Caveat: if the users of the driver don't obey the API contract which
> states that once -EBUSY is received, no more requests are to be
> sent, eventually the driver will reject the enqueues.
> 
> Signed-off-by: Alex Porosanu <alexandru.porosanu@xxxxxxxxxxxxx>
> ---
>  drivers/crypto/caam/caamalg.c | 233 ++++++++++++++++++++++++++++++++++--------
>  drivers/crypto/caam/intern.h  |   7 ++
>  drivers/crypto/caam/jr.c      | 190 +++++++++++++++++++++++++++-------
>  drivers/crypto/caam/jr.h      |   5 +
>  4 files changed, 352 insertions(+), 83 deletions(-)

The patch updates only caamalg.c. What about the others (caamhash.c etc.)?

> 
> diff --git a/drivers/crypto/caam/caamalg.c b/drivers/crypto/caam/caamalg.c
> index ba79d63..c281483 100644
> --- a/drivers/crypto/caam/caamalg.c
> +++ b/drivers/crypto/caam/caamalg.c
> @@ -1815,9 +1815,14 @@ static void aead_encrypt_done(struct device *jrdev, u32 *desc, u32 err,
>  
>  	edesc = container_of(desc, struct aead_edesc, hw_desc[0]);
>  
> -	if (err)
> +	if (err && (err != -EINPROGRESS))
>  		caam_jr_strstatus(jrdev, err);
>  
> +	if (err == -EINPROGRESS) {
> +		aead_request_complete(req, err);
> +		return;
> +	}

Logic can be simplified by reversing the conditions:

if (err == -EINPROGRESS)
	goto out;

if (err)
	caam_jr_strstatus(jrdev, err);

[...]
out:
	aead_request_complete(req, err);

Same for the other places.

> +
>  	aead_unmap(jrdev, edesc, req);
>  
>  	kfree(edesc);
> @@ -1837,9 +1842,14 @@ static void aead_decrypt_done(struct device *jrdev, u32 *desc, u32 err,
>  
>  	edesc = container_of(desc, struct aead_edesc, hw_desc[0]);
>  
> -	if (err)
> +	if (err && (err != -EINPROGRESS))
>  		caam_jr_strstatus(jrdev, err);
>  
> +	if (err == -EINPROGRESS) {
> +		aead_request_complete(req, err);
> +		return;
> +	}
> +
>  	aead_unmap(jrdev, edesc, req);
>  
>  	/*
> @@ -1864,13 +1874,17 @@ static void ablkcipher_encrypt_done(struct device *jrdev, u32 *desc, u32 err,
>  
>  	dev_err(jrdev, "%s %d: err 0x%x\n", __func__, __LINE__, err);
>  #endif
> -
>  	edesc = (struct ablkcipher_edesc *)((char *)desc -
>  		 offsetof(struct ablkcipher_edesc, hw_desc));
>  
> -	if (err)
> +	if (err && (err != -EINPROGRESS))
>  		caam_jr_strstatus(jrdev, err);
>  
> +	if (err == -EINPROGRESS) {
> +		ablkcipher_request_complete(req, err);
> +		return;
> +	}
> +
>  #ifdef DEBUG
>  	print_hex_dump(KERN_ERR, "dstiv  @"__stringify(__LINE__)": ",
>  		       DUMP_PREFIX_ADDRESS, 16, 4, req->info,
> @@ -1900,9 +1914,14 @@ static void ablkcipher_decrypt_done(struct device *jrdev, u32 *desc, u32 err,
>  
>  	edesc = (struct ablkcipher_edesc *)((char *)desc -
>  		 offsetof(struct ablkcipher_edesc, hw_desc));
> -	if (err)
> +	if (err && (err != -EINPROGRESS))
>  		caam_jr_strstatus(jrdev, err);
>  
> +	if (err == -EINPROGRESS) {
> +		ablkcipher_request_complete(req, err);
> +		return;
> +	}
> +
>  #ifdef DEBUG
>  	print_hex_dump(KERN_ERR, "dstiv  @"__stringify(__LINE__)": ",
>  		       DUMP_PREFIX_ADDRESS, 16, 4, req->info,
> @@ -2294,12 +2313,30 @@ static int gcm_encrypt(struct aead_request *req)
>  #endif
>  
>  	desc = edesc->hw_desc;
> -	ret = caam_jr_enqueue(jrdev, desc, aead_encrypt_done, req);
> -	if (!ret) {
> -		ret = -EINPROGRESS;
> +	if (req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG) {
> +		ret = caam_jr_enqueue_bklog(jrdev, desc, aead_encrypt_done,
> +					    req);
> +		switch (ret) {
> +		case 0:
> +			ret = -EINPROGRESS;
> +			break;
> +
> +		case -EBUSY:
> +			break;
> +
> +		default:
> +			aead_unmap(jrdev, edesc, req);
> +			kfree(edesc);
> +			break;
> +		}
>  	} else {
> -		aead_unmap(jrdev, edesc, req);
> -		kfree(edesc);
> +		ret = caam_jr_enqueue(jrdev, desc, aead_encrypt_done, req);
> +		if (!ret) {
> +			ret = -EINPROGRESS;
> +		} else {
> +			aead_unmap(jrdev, edesc, req);
> +			kfree(edesc);
> +		}
>  	}

Again, this should be simplified:

if (req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG) {
	ret = caam_jr_enqueue_bklog(jrdev, desc, aead_encrypt_done,
				    req);
	if (ret == -EBUSY)
		return ret;
} else {
	ret = caam_jr_enqueue(jrdev, desc, aead_encrypt_done, req);

}

if (!ret) {
	ret = -EINPROGRESS;
} else {
	aead_unmap(jrdev, edesc, req);
	kfree(edesc);
}

>  
>  	return ret;
> @@ -2338,12 +2375,30 @@ static int aead_encrypt(struct aead_request *req)
>  #endif
>  
>  	desc = edesc->hw_desc;
> -	ret = caam_jr_enqueue(jrdev, desc, aead_encrypt_done, req);
> -	if (!ret) {
> -		ret = -EINPROGRESS;
> +	if (req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG) {
> +		ret = caam_jr_enqueue_bklog(jrdev, desc, aead_encrypt_done,
> +					    req);
> +		switch (ret) {
> +		case 0:
> +			ret = -EINPROGRESS;
> +			break;
> +
> +		case -EBUSY:
> +			break;
> +
> +		default:
> +			aead_unmap(jrdev, edesc, req);
> +			kfree(edesc);
> +			break;
> +		}
>  	} else {
> -		aead_unmap(jrdev, edesc, req);
> -		kfree(edesc);
> +		ret = caam_jr_enqueue(jrdev, desc, aead_encrypt_done, req);
> +		if (!ret) {
> +			ret = -EINPROGRESS;
> +		} else {
> +			aead_unmap(jrdev, edesc, req);
> +			kfree(edesc);
> +		}
>  	}
>  
>  	return ret;
> @@ -2373,12 +2428,30 @@ static int gcm_decrypt(struct aead_request *req)
>  #endif
>  
>  	desc = edesc->hw_desc;
> -	ret = caam_jr_enqueue(jrdev, desc, aead_decrypt_done, req);
> -	if (!ret) {
> -		ret = -EINPROGRESS;
> +	if (req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG) {
> +		ret = caam_jr_enqueue_bklog(jrdev, desc, aead_decrypt_done,
> +					    req);
> +		switch (ret) {
> +		case 0:
> +			ret = -EINPROGRESS;
> +			break;
> +
> +		case -EBUSY:
> +			break;
> +
> +		default:
> +			aead_unmap(jrdev, edesc, req);
> +			kfree(edesc);
> +			break;
> +		}
>  	} else {
> -		aead_unmap(jrdev, edesc, req);
> -		kfree(edesc);
> +		ret = caam_jr_enqueue(jrdev, desc, aead_decrypt_done, req);
> +		if (!ret) {
> +			ret = -EINPROGRESS;
> +		} else {
> +			aead_unmap(jrdev, edesc, req);
> +			kfree(edesc);
> +		}
>  	}
>  
>  	return ret;
> @@ -2423,12 +2496,30 @@ static int aead_decrypt(struct aead_request *req)
>  #endif
>  
>  	desc = edesc->hw_desc;
> -	ret = caam_jr_enqueue(jrdev, desc, aead_decrypt_done, req);
> -	if (!ret) {
> -		ret = -EINPROGRESS;
> +	if (req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG) {
> +		ret = caam_jr_enqueue_bklog(jrdev, desc, aead_decrypt_done,
> +					    req);
> +		switch (ret) {
> +		case 0:
> +			ret = -EINPROGRESS;
> +			break;
> +
> +		case -EBUSY:
> +			break;
> +
> +		default:
> +			aead_unmap(jrdev, edesc, req);
> +			kfree(edesc);
> +			break;
> +		}
>  	} else {
> -		aead_unmap(jrdev, edesc, req);
> -		kfree(edesc);
> +		ret = caam_jr_enqueue(jrdev, desc, aead_decrypt_done, req);
> +		if (!ret) {
> +			ret = -EINPROGRESS;
> +		} else {
> +			aead_unmap(jrdev, edesc, req);
> +			kfree(edesc);
> +		}
>  	}
>  
>  	return ret;
> @@ -2575,13 +2666,31 @@ static int ablkcipher_encrypt(struct ablkcipher_request *req)
>  		       desc_bytes(edesc->hw_desc), 1);
>  #endif
>  	desc = edesc->hw_desc;
> -	ret = caam_jr_enqueue(jrdev, desc, ablkcipher_encrypt_done, req);
> -
> -	if (!ret) {
> -		ret = -EINPROGRESS;
> +	if (req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG) {
> +		ret = caam_jr_enqueue_bklog(jrdev, desc,
> +					    ablkcipher_encrypt_done, req);
> +		switch (ret) {
> +		case 0:
> +			ret = -EINPROGRESS;
> +			break;
> +
> +		case -EBUSY:
> +			break;
> +
> +		default:
> +			ablkcipher_unmap(jrdev, edesc, req);
> +			kfree(edesc);
> +			break;
> +		}
>  	} else {
> -		ablkcipher_unmap(jrdev, edesc, req);
> -		kfree(edesc);
> +		ret = caam_jr_enqueue(jrdev, desc, ablkcipher_encrypt_done,
> +				      req);
> +		if (!ret) {
> +			ret = -EINPROGRESS;
> +		} else {
> +			ablkcipher_unmap(jrdev, edesc, req);
> +			kfree(edesc);
> +		}
>  	}
>  
>  	return ret;
> @@ -2612,15 +2721,32 @@ static int ablkcipher_decrypt(struct ablkcipher_request *req)
>  		       DUMP_PREFIX_ADDRESS, 16, 4, edesc->hw_desc,
>  		       desc_bytes(edesc->hw_desc), 1);
>  #endif
> -
> -	ret = caam_jr_enqueue(jrdev, desc, ablkcipher_decrypt_done, req);
> -	if (!ret) {
> -		ret = -EINPROGRESS;
> +	if (req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG) {
> +		ret = caam_jr_enqueue_bklog(jrdev, desc,
> +					    ablkcipher_decrypt_done, req);
> +		switch (ret) {
> +		case 0:
> +			ret = -EINPROGRESS;
> +			break;
> +
> +		case -EBUSY:
> +			break;
> +
> +		default:
> +			ablkcipher_unmap(jrdev, edesc, req);
> +			kfree(edesc);
> +			break;
> +		}
>  	} else {
> -		ablkcipher_unmap(jrdev, edesc, req);
> -		kfree(edesc);
> +		ret = caam_jr_enqueue_bklog(jrdev, desc,
> +					    ablkcipher_decrypt_done, req);

Typo, s/caam_jr_enqueue_bklog/caam_jr_enqueue.
What testing has been performed?

> +		if (!ret) {
> +			ret = -EINPROGRESS;
> +		} else {
> +			ablkcipher_unmap(jrdev, edesc, req);
> +			kfree(edesc);
> +		}
>  	}
> -
>  	return ret;
>  }
>  
> @@ -2757,13 +2883,32 @@ static int ablkcipher_givencrypt(struct skcipher_givcrypt_request *creq)
>  		       desc_bytes(edesc->hw_desc), 1);
>  #endif
>  	desc = edesc->hw_desc;
> -	ret = caam_jr_enqueue(jrdev, desc, ablkcipher_encrypt_done, req);
> -
> -	if (!ret) {
> -		ret = -EINPROGRESS;
> +	if (req->base.flags & CRYPTO_TFM_REQ_MAY_BACKLOG) {
> +		ret = caam_jr_enqueue_bklog(jrdev, desc,
> +					    ablkcipher_encrypt_done, req);
> +		switch (ret) {
> +		case 0:
> +			ret = -EINPROGRESS;
> +			break;
> +
> +		case -EBUSY:
> +			break;
> +
> +		default:
> +			ablkcipher_unmap(jrdev, edesc, req);
> +			kfree(edesc);
> +			break;
> +		}
>  	} else {
> -		ablkcipher_unmap(jrdev, edesc, req);
> -		kfree(edesc);
> +		ret = caam_jr_enqueue(jrdev, desc, ablkcipher_encrypt_done,
> +				      req);
> +
> +		if (!ret) {
> +			ret = -EINPROGRESS;
> +		} else {
> +			ablkcipher_unmap(jrdev, edesc, req);
> +			kfree(edesc);
> +		}
>  	}
>  
>  	return ret;
> diff --git a/drivers/crypto/caam/intern.h b/drivers/crypto/caam/intern.h
> index e2bcacc..13e63ef 100644
> --- a/drivers/crypto/caam/intern.h
> +++ b/drivers/crypto/caam/intern.h
> @@ -11,6 +11,12 @@
>  
>  /* Currently comes from Kconfig param as a ^2 (driver-required) */
>  #define JOBR_DEPTH (1 << CONFIG_CRYPTO_DEV_FSL_CAAM_RINGSIZE)
> +/*
> + * If the user tries to enqueue a job and the number of slots available
> + * is less than this value, then the job will be backlogged (if the user
> + * allows for it) or it will be dropped.
> + */
> +#define JOBR_THRESH 16

Why 16?
What happens when user configures CONFIG_CRYPTO_DEV_FSL_CAAM_RINGSIZE to
{2, 3, 4}?
Threshold should depend on JOBR_DEPTH.

>  
>  /* Kconfig params for interrupt coalescing if selected (else zero) */
>  #ifdef CONFIG_CRYPTO_DEV_FSL_CAAM_INTC
> @@ -33,6 +39,7 @@ struct caam_jrentry_info {
>  	u32 *desc_addr_virt;	/* Stored virt addr for postprocessing */
>  	dma_addr_t desc_addr_dma;	/* Stored bus addr for done matching */
>  	u32 desc_size;	/* Stored size for postprocessing, header derived */
> +	bool is_backlogged; /* True if the request has been backlogged */
>  };
>  
>  /* Private sub-storage for a single JobR */
> diff --git a/drivers/crypto/caam/jr.c b/drivers/crypto/caam/jr.c
> index f7e0d8d..916288d 100644
> --- a/drivers/crypto/caam/jr.c
> +++ b/drivers/crypto/caam/jr.c
> @@ -168,6 +168,7 @@ static void caam_jr_dequeue(unsigned long devarg)
>  	void (*usercall)(struct device *dev, u32 *desc, u32 status, void *arg);
>  	u32 *userdesc, userstatus;
>  	void *userarg;
> +	bool is_backlogged;
>  
>  	while (rd_reg32(&jrp->rregs->outring_used)) {
>  
> @@ -201,6 +202,7 @@ static void caam_jr_dequeue(unsigned long devarg)
>  		userarg = jrp->entinfo[sw_idx].cbkarg;
>  		userdesc = jrp->entinfo[sw_idx].desc_addr_virt;
>  		userstatus = jrp->outring[hw_idx].jrstatus;
> +		is_backlogged = jrp->entinfo[sw_idx].is_backlogged;
>  
>  		/*
>  		 * Make sure all information from the job has been obtained
> @@ -231,6 +233,20 @@ static void caam_jr_dequeue(unsigned long devarg)
>  
>  		spin_unlock(&jrp->outlock);
>  
> +		if (is_backlogged)
> +			/*
> +			 * For backlogged requests, the user callback needs to
> +			 * be called twice: once when starting to process it
> +			 * (with a status of -EINPROGRESS and once when it's
> +			 * done. Since SEC cheats by enqueuing the request in
> +			 * its HW ring but returning -EBUSY, the time when the
> +			 * request's processing has started is not known.
> +			 * Thus notify here the user. The second call is on the
> +			 * normal path (i.e. the one that is called even for
> +			 * non-backlogged requests.
						  ^ missing parenthesis

> +			 */
> +			usercall(dev, userdesc, -EINPROGRESS, userarg);
> +
>  		/* Finally, execute user's callback */
>  		usercall(dev, userdesc, userstatus, userarg);
>  	}
> @@ -292,6 +308,84 @@ void caam_jr_free(struct device *rdev)
>  }
>  EXPORT_SYMBOL(caam_jr_free);
>  
> +static inline int __caam_jr_enqueue(struct caam_drv_private_jr *jrp, u32 *desc,
> +				    int desc_size, dma_addr_t desc_dma,
> +				    void (*cbk)(struct device *dev, u32 *desc,
> +						u32 status, void *areq),
> +				    void *areq,
> +				    bool can_be_backlogged)
> +{
> +	int head, tail;
> +	struct caam_jrentry_info *head_entry;
> +	int ret = 0, hw_slots, sw_slots;
> +
> +	spin_lock_bh(&jrp->inplock);
> +
> +	head = jrp->head;
> +	tail = ACCESS_ONCE(jrp->tail);
> +
> +	head_entry = &jrp->entinfo[head];
> +
> +	/* Reset backlogging status here */
> +	head_entry->is_backlogged = false;
> +
> +	hw_slots = rd_reg32(&jrp->rregs->inpring_avail);
> +	sw_slots = CIRC_SPACE(head, tail, JOBR_DEPTH);
> +
> +	if (hw_slots <= JOBR_THRESH || sw_slots <= JOBR_THRESH) {
> +		/*
> +		 * The state below can be reached in three cases:
> +		 * 1) A badly behaved backlogging user doesn't back off when
> +		 *    told so by the -EBUSY return code
> +		 * 2) More than JOBR_THRESH backlogging users requests
> +		 * 3) Due to the high system load, the entries reserved for the
> +		 *    backlogging users are being filled (slowly) in between
> +		 *    the successive calls to the user callback (the first one
> +		 *    with -EINPROGRESS and the 2nd one with the real result.
> +		 * The code below is a last-resort measure which will DROP
> +		 * any request if there is physically no more space. This will
> +		 * lead to data-loss for disk-related users.
> +		 */
> +		if (!hw_slots || sw_slots <= 0) {

sw_slots cannot be negative.

> +			spin_unlock_bh(&jrp->inplock);
> +			return -EIO;

Or:
ret = -EIO;
goto out_unlock;

> +		}
> +
> +		if (can_be_backlogged) {
> +			head_entry->is_backlogged = true;
> +			ret = -EBUSY;
> +		} else {
> +			spin_unlock_bh(&jrp->inplock);
> +			return -EBUSY;
> +		}

Or:
ret = -EBUSY;
if (!can_be_backlogged)
	goto out_unlock;
head_entry->is_backlogged = true;

> +	}
> +
> +	head_entry->desc_addr_virt = desc;
> +	head_entry->desc_size = desc_size;
> +	head_entry->callbk = (void *)cbk;
> +	head_entry->cbkarg = areq;
> +	head_entry->desc_addr_dma = desc_dma;
> +
> +	jrp->inpring[jrp->inp_ring_write_index] = desc_dma;
> +
> +	/*
> +	 * Guarantee that the descriptor's DMA address has been written to
> +	 * the next slot in the ring before the write index is updated, since
> +	 * other cores may update this index independently.
> +	 */
> +	smp_wmb();
> +
> +	jrp->inp_ring_write_index = (jrp->inp_ring_write_index + 1) &
> +				    (JOBR_DEPTH - 1);
> +	jrp->head = (head + 1) & (JOBR_DEPTH - 1);
> +
> +	wr_reg32(&jrp->rregs->inpring_jobadd, 1);
> +

out_unlock:

> +	spin_unlock_bh(&jrp->inplock);
> +
> +	return ret;
> +}
> +
>  /**
>   * caam_jr_enqueue() - Enqueue a job descriptor head. Returns 0 if OK,
>   * -EBUSY if the queue is full, -EIO if it cannot map the caller's
> @@ -326,8 +420,7 @@ int caam_jr_enqueue(struct device *dev, u32 *desc,
>  		    void *areq)
>  {
>  	struct caam_drv_private_jr *jrp = dev_get_drvdata(dev);
> -	struct caam_jrentry_info *head_entry;
> -	int head, tail, desc_size;
> +	int desc_size, ret;
>  	dma_addr_t desc_dma;
>  
>  	desc_size = (*desc & HDR_JD_LENGTH_MASK) * sizeof(u32);
> @@ -337,51 +430,70 @@ int caam_jr_enqueue(struct device *dev, u32 *desc,
>  		return -EIO;
>  	}
>  
> -	spin_lock_bh(&jrp->inplock);
> -
> -	head = jrp->head;
> -	tail = ACCESS_ONCE(jrp->tail);
> -
> -	if (!rd_reg32(&jrp->rregs->inpring_avail) ||
> -	    CIRC_SPACE(head, tail, JOBR_DEPTH) <= 0) {
> -		spin_unlock_bh(&jrp->inplock);
> +	ret = __caam_jr_enqueue(jrp, desc, desc_size, desc_dma, cbk, areq,
> +				false);
> +	if (unlikely(ret))
>  		dma_unmap_single(dev, desc_dma, desc_size, DMA_TO_DEVICE);
> -		return -EBUSY;
> -	}
>  
> -	head_entry = &jrp->entinfo[head];
> -	head_entry->desc_addr_virt = desc;
> -	head_entry->desc_size = desc_size;
> -	head_entry->callbk = (void *)cbk;
> -	head_entry->cbkarg = areq;
> -	head_entry->desc_addr_dma = desc_dma;
> -
> -	jrp->inpring[jrp->inp_ring_write_index] = desc_dma;
> -
> -	/*
> -	 * Guarantee that the descriptor's DMA address has been written to
> -	 * the next slot in the ring before the write index is updated, since
> -	 * other cores may update this index independently.
> -	 */
> -	smp_wmb();
> +	return ret;
> +}
> +EXPORT_SYMBOL(caam_jr_enqueue);
>  
> -	jrp->inp_ring_write_index = (jrp->inp_ring_write_index + 1) &
> -				    (JOBR_DEPTH - 1);
> -	jrp->head = (head + 1) & (JOBR_DEPTH - 1);
> +/**
> + * caam_jr_enqueue_bklog() - Enqueue a job descriptor head, returns 0 if OK, or
> + * -EINPROGRESS if the number of available entries in the Job Ring is less

The function actually returns -EBUSY, not -EINPROGRESS.

> + * than the threshold configured through CONFIG_CRYPTO_DEV_FSL_CAAM_BKLOG_SIZE,

Leftover, threshold is not configurable.

> + * and -EIO if it cannot map the caller's descriptor or if the threshold has
> + * been exceeded.
> + * @dev:  device of the job ring to be used. This device should have
> + *        been assigned prior by caam_jr_register().
> + * @desc: points to a job descriptor that execute our request. All
> + *        descriptors (and all referenced data) must be in a DMAable
> + *        region, and all data references must be physical addresses
> + *        accessible to CAAM (i.e. within a PAMU window granted
> + *        to it).
> + * @cbk:  pointer to a callback function to be invoked upon completion
> + *        of this request. This has the form:
> + *        callback(struct device *dev, u32 *desc, u32 stat, void *arg)
> + *        where:
> + *        @dev:    contains the job ring device that processed this
> + *                 response.
> + *        @desc:   descriptor that initiated the request, same as
> + *                 "desc" being argued to caam_jr_enqueue().
> + *        @status: untranslated status received from CAAM. See the
> + *                 reference manual for a detailed description of
> + *                 error meaning, or see the JRSTA definitions in the
> + *                 register header file
> + *        @areq:   optional pointer to an argument passed with the
> + *                 original request

Though I haven't checked, I am pretty sure that kernel-doc is not smart
enough to handle the description of function/callback parameters.

> + * @areq: optional pointer to a user argument for use at callback
> + *        time.
> + **/
> +int caam_jr_enqueue_bklog(struct device *dev, u32 *desc,
> +			  void (*cbk)(struct device *dev, u32 *desc,
> +				      u32 status, void *areq),
> +			  void *areq)
> +{
> +	struct caam_drv_private_jr *jrp = dev_get_drvdata(dev);
> +	int desc_size, ret;
> +	dma_addr_t desc_dma;
>  
> -	/*
> -	 * Ensure that all job information has been written before
> -	 * notifying CAAM that a new job was added to the input ring.
> -	 */
> -	wmb();
> +	desc_size = (*desc & HDR_JD_LENGTH_MASK) * sizeof(u32);
> +	desc_dma = dma_map_single(dev, desc, desc_size, DMA_TO_DEVICE);
> +	if (dma_mapping_error(dev, desc_dma)) {
> +		dev_err(dev, "caam_jr_enqueue(): can't map jobdesc\n");
> +		return -EIO;
> +	}
>  
> -	wr_reg32(&jrp->rregs->inpring_jobadd, 1);
> +	ret = __caam_jr_enqueue(jrp, desc, desc_size, desc_dma, cbk, areq,
> +				true);
> +	if (unlikely(ret && (ret != -EBUSY)))
> +		dma_unmap_single(dev, desc_dma, desc_size, DMA_TO_DEVICE);
>  
> -	spin_unlock_bh(&jrp->inplock);
> +	return ret;
>  
> -	return 0;
>  }
> -EXPORT_SYMBOL(caam_jr_enqueue);
> +EXPORT_SYMBOL(caam_jr_enqueue_bklog);
>  
>  /*
>   * Init JobR independent of platform property detection
> diff --git a/drivers/crypto/caam/jr.h b/drivers/crypto/caam/jr.h
> index 97113a6..21558df 100644
> --- a/drivers/crypto/caam/jr.h
> +++ b/drivers/crypto/caam/jr.h
> @@ -15,4 +15,9 @@ int caam_jr_enqueue(struct device *dev, u32 *desc,
>  				void *areq),
>  		    void *areq);
>  
> +int caam_jr_enqueue_bklog(struct device *dev, u32 *desc,
> +			  void (*cbk)(struct device *dev, u32 *desc, u32 status,
> +				      void *areq),
> +			  void *areq);
> +
>  #endif /* JR_H */
> 


--
To unsubscribe from this list: send the line "unsubscribe linux-crypto" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [Kernel]     [Gnu Classpath]     [Gnu Crypto]     [DM Crypt]     [Netfilter]     [Bugtraq]

  Powered by Linux