[dpdk-dev] [PATCH 6/8] raw/dpaa2_qdma: support enq and deq operations

Shreyansh Jain shreyansh.jain at nxp.com
Mon Apr 16 16:01:30 CEST 2018


On Saturday 07 April 2018 08:47 PM, Nipun Gupta wrote:
> Signed-off-by: Nipun Gupta <nipun.gupta at nxp.com>
> ---
>   drivers/raw/dpaa2_qdma/dpaa2_qdma.c                | 332 +++++++++++++++++++++
>   drivers/raw/dpaa2_qdma/dpaa2_qdma.h                |  21 ++
>   drivers/raw/dpaa2_qdma/rte_pmd_dpaa2_qdma.h        |  70 +++++
>   .../raw/dpaa2_qdma/rte_pmd_dpaa2_qdma_version.map  |   4 +
>   4 files changed, 427 insertions(+)
> 
> diff --git a/drivers/raw/dpaa2_qdma/dpaa2_qdma.c b/drivers/raw/dpaa2_qdma/dpaa2_qdma.c
> index b5f6bd9..5cd18ec 100644
> --- a/drivers/raw/dpaa2_qdma/dpaa2_qdma.c
> +++ b/drivers/raw/dpaa2_qdma/dpaa2_qdma.c
> @@ -344,6 +344,338 @@
>   	return i;
>   }
>   
> +static void
> +dpaa2_qdma_populate_fle(struct qbman_fle *fle,
> +			uint64_t src, uint64_t dest,
> +			size_t len, uint32_t flags)
> +{
> +	struct qdma_sdd *sdd;
> +
> +	DPAA2_QDMA_FUNC_TRACE();
> +
> +	sdd = (struct qdma_sdd *)((uint8_t *)(fle) +
> +		(DPAA2_QDMA_MAX_FLE * sizeof(struct qbman_fle)));
> +
> +	/* first frame list to source descriptor */
> +	DPAA2_SET_FLE_ADDR(fle, DPAA2_VADDR_TO_IOVA(sdd));
> +	DPAA2_SET_FLE_LEN(fle, (2 * (sizeof(struct qdma_sdd))));
> +
> +	/* source and destination descriptor */
> +	DPAA2_SET_SDD_RD_COHERENT(sdd); /* source descriptor CMD */
> +	sdd++;
> +	DPAA2_SET_SDD_WR_COHERENT(sdd); /* dest descriptor CMD */
> +
> +	fle++;
> +	/* source frame list to source buffer */
> +	if (flags & RTE_QDMA_JOB_SRC_PHY) {
> +		DPAA2_SET_FLE_ADDR(fle, src);
> +		DPAA2_SET_FLE_BMT(fle);
> +	} else {
> +		DPAA2_SET_FLE_ADDR(fle, DPAA2_VADDR_TO_IOVA(src));
> +	}
> +	DPAA2_SET_FLE_LEN(fle, len);
> +
> +	fle++;
> +	/* destination frame list to destination buffer */
> +	if (flags & RTE_QDMA_JOB_DEST_PHY) {
> +		DPAA2_SET_FLE_BMT(fle);
> +		DPAA2_SET_FLE_ADDR(fle, dest);
> +	} else {
> +		DPAA2_SET_FLE_ADDR(fle, DPAA2_VADDR_TO_IOVA(dest));
> +	}
> +	DPAA2_SET_FLE_LEN(fle, len);
> +
> +	/* Final bit: 1, for last frame list */
> +	DPAA2_SET_FLE_FIN(fle);
> +}
> +
> +static int
> +dpdmai_dev_enqueue(struct dpaa2_dpdmai_dev *dpdmai_dev,
> +		   uint16_t txq_id,
> +		   uint16_t vq_id,
> +		   struct rte_qdma_job *job)
> +{
> +	struct qdma_io_meta *io_meta;
> +	struct qbman_fd fd = {};
> +	struct dpaa2_queue *txq;
> +	struct qbman_fle *fle;
> +	struct qbman_eq_desc eqdesc;
> +	struct qbman_swp *swp;
> +	int ret;
> +
> +	DPAA2_QDMA_FUNC_TRACE();
> +
> +	if (unlikely(!DPAA2_PER_LCORE_DPIO)) {
> +		ret = dpaa2_affine_qbman_swp();
> +		if (ret) {
> +			DPAA2_QDMA_ERR("Failure in affining portal\n");
> +			return 0;
> +		}
> +	}
> +	swp = DPAA2_PER_LCORE_PORTAL;
> +
> +	txq = &(dpdmai_dev->tx_queue[txq_id]);
> +
> +	/* Prepare enqueue descriptor */
> +	qbman_eq_desc_clear(&eqdesc);
> +	qbman_eq_desc_set_fq(&eqdesc, txq->fqid);
> +	qbman_eq_desc_set_no_orp(&eqdesc, 0);
> +	qbman_eq_desc_set_response(&eqdesc, 0, 0);
> +
> +	/*
> +	 * Get an FLE/SDD from FLE pool.
> +	 * Note: IO metadata is before the FLE and SDD memory.
> +	 */
> +	ret = rte_mempool_get(qdma_dev.fle_pool, (void **)(&io_meta));
> +	if (ret) {
> +		DPAA2_QDMA_ERR("Memory alloc failed for FLE\n");
> +		return ret;
> +	}
> +
> +	/* Set the metadata */
> +	io_meta->cnxt = (size_t)job;
> +	io_meta->id = vq_id;
> +
> +	fle = (struct qbman_fle *)(io_meta + 1);
> +
> +	/* populate Frame descriptor */
> +	DPAA2_SET_FD_ADDR(&fd, DPAA2_VADDR_TO_IOVA(fle));
> +	DPAA2_SET_FD_COMPOUND_FMT(&fd);
> +	DPAA2_SET_FD_FRC(&fd, QDMA_SER_CTX);
> +
> +	/* Populate FLE */
> +	memset(fle, 0, QDMA_FLE_POOL_SIZE);
> +	dpaa2_qdma_populate_fle(fle, job->src, job->dest, job->len, job->flags);
> +
> +	/* Enqueue the packet to the QBMAN */
> +	do {
> +		ret = qbman_swp_enqueue_multiple(swp, &eqdesc, &fd, NULL, 1);
> +		if (ret < 0 && ret != -EBUSY)
> +			DPAA2_QDMA_ERR("Transmit failure with err: %d\n", ret);

Similar comment to one of the other patch - your macro definition has 
'\n' already in it. So, double new lines.

> +	} while (ret == -EBUSY);
> +
> +	DPAA2_QDMA_DEBUG("Successfully transmitted a packet\n");

Macros like the above are compiled into the code unlike the DP_ 
variants. If this is I/O path and such conditionals impact performance, 
maybe DP_ variant would be more efficient.

> +
> +	return ret;
> +}
> +
> +int
> +rte_qdma_vq_enqueue_multi(uint16_t vq_id,
> +			  struct rte_qdma_job **job,
> +			  uint16_t nb_jobs)
> +{
> +	int i, ret;
> +
> +	DPAA2_QDMA_FUNC_TRACE();
> +
> +	for (i = 0; i < nb_jobs; i++) {
> +		ret = rte_qdma_vq_enqueue(vq_id, job[i]);
> +		if (ret < 0)
> +			break;
> +	}
> +
> +	return i;
> +}
> +
> +int
> +rte_qdma_vq_enqueue(uint16_t vq_id,
> +		    struct rte_qdma_job *job)
> +{
> +	struct qdma_virt_queue *qdma_vq = &qdma_vqs[vq_id];
> +	struct qdma_hw_queue *qdma_pq = qdma_vq->hw_queue;
> +	struct dpaa2_dpdmai_dev *dpdmai_dev = qdma_pq->dpdmai_dev;
> +	int ret;
> +
> +	DPAA2_QDMA_FUNC_TRACE();
> +
> +	/* Return error in case of wrong lcore_id */
> +	if (rte_lcore_id() != qdma_vq->lcore_id) {
> +		DPAA2_QDMA_WARN("QDMA enqueue for vqid %d on wrong core",
                /^^^^^^^^^^^^^^^
               /
Should this be DPAA2_QDMA_ERR?

> +				vq_id);
> +		return -EINVAL;
> +	}
> +
> +	ret = dpdmai_dev_enqueue(dpdmai_dev, qdma_pq->queue_id, vq_id, job);
> +	if (ret < 0) {
> +		DPAA2_QDMA_ERR("DPDMAI device enqueue failed: %d\n", ret);
> +		return ret;
> +	}
> +
> +	qdma_vq->num_enqueues++;
> +
> +	return 1;
> +}
> +

[...]

> +
> +struct rte_qdma_job *
> +rte_qdma_vq_dequeue(uint16_t vq_id)
> +{
> +	struct qdma_virt_queue *qdma_vq = &qdma_vqs[vq_id];
> +	struct qdma_hw_queue *qdma_pq = qdma_vq->hw_queue;
> +	struct dpaa2_dpdmai_dev *dpdmai_dev = qdma_pq->dpdmai_dev;
> +	struct rte_qdma_job *job = NULL;
> +	struct qdma_virt_queue *temp_qdma_vq;
> +	int dequeue_budget = QDMA_DEQUEUE_BUDGET;
> +	int ring_count, ret, i;
> +	uint16_t temp_vq_id;
> +
> +	DPAA2_QDMA_FUNC_TRACE();
> +
> +	/* Return error in case of wrong lcore_id */
> +	if (rte_lcore_id() != (unsigned int)(qdma_vq->lcore_id)) {
> +		DPAA2_QDMA_WARN("QDMA dequeue for vqid %d on wrong core",
> +				vq_id);
> +		return NULL;
> +	}
> +
> +	/* Only dequeue when there are pending jobs on VQ */
> +	if (qdma_vq->num_enqueues == qdma_vq->num_dequeues)
> +		return NULL;
> +
> +	if (qdma_vq->exclusive_hw_queue) {
> +		/* In case of exclusine queue directly fetch from HW queue */

for above comment s/exclusine/exclusive/

[...]

-
Shreyansh



More information about the dev mailing list