[PATCH v4 12/14] app/dma-perf: fix on-flight DMA when verify data

Chengwen Feng fengchengwen at huawei.com
Mon Oct 20 06:11:03 CEST 2025


There maybe on-flight DMA when verify_data() because the DMA device
may still working when worker exit.

This commit add wait DMA complete stage before worker exit.

Fixes: 623dc9364dc6 ("app/dma-perf: introduce DMA performance test")
Cc: stable at dpdk.org

Signed-off-by: Chengwen Feng <fengchengwen at huawei.com>
---
 app/test-dma-perf/benchmark.c | 71 ++++++++++++++++++++++-------------
 1 file changed, 44 insertions(+), 27 deletions(-)

diff --git a/app/test-dma-perf/benchmark.c b/app/test-dma-perf/benchmark.c
index 6643ccc95f..4ce95d0f7b 100644
--- a/app/test-dma-perf/benchmark.c
+++ b/app/test-dma-perf/benchmark.c
@@ -19,7 +19,6 @@
 #define MAX_DMA_CPL_NB 255
 
 #define TEST_WAIT_U_SECOND 10000
-#define POLL_MAX 1000
 
 #define CSV_LINE_DMA_FMT "Scenario %u,%u,%s,%u,%u,%u,%u,%.2lf,%" PRIu64 ",%.3lf,%.3lf\n"
 #define CSV_LINE_CPU_FMT "Scenario %u,%u,NA,NA,NA,%u,%u,%.2lf,%" PRIu64 ",%.3lf,%.3lf\n"
@@ -293,6 +292,45 @@ do_dma_submit_and_poll(uint16_t dev_id, uint64_t *async_cnt,
 	worker_info->total_cpl += nr_cpl;
 }
 
+static int
+do_dma_submit_and_wait_cpl(uint16_t dev_id, uint64_t async_cnt, bool use_ops)
+{
+#define MAX_WAIT_MSEC	1000
+#define MAX_POLL	1000
+#define DEQ_SZ		64
+	struct rte_dma_op *op[DEQ_SZ];
+	enum rte_dma_vchan_status st;
+	uint32_t poll_cnt = 0;
+	uint32_t wait_ms = 0;
+	uint16_t nr_cpl;
+
+	if (!use_ops)
+		rte_dma_submit(dev_id, 0);
+
+	if (rte_dma_vchan_status(dev_id, 0, &st) < 0) {
+		rte_delay_ms(MAX_WAIT_MSEC);
+		goto wait_cpl;
+	}
+
+	while (st == RTE_DMA_VCHAN_ACTIVE && wait_ms++ < MAX_WAIT_MSEC) {
+		rte_delay_ms(1);
+		rte_dma_vchan_status(dev_id, 0, &st);
+	}
+
+wait_cpl:
+	while ((async_cnt > 0) && (poll_cnt++ < MAX_POLL)) {
+		if (use_ops)
+			nr_cpl = rte_dma_dequeue_ops(dev_id, 0, op, DEQ_SZ);
+		else
+			nr_cpl = rte_dma_completed(dev_id, 0, MAX_DMA_CPL_NB, NULL, NULL);
+		async_cnt -= nr_cpl;
+	}
+	if (async_cnt > 0)
+		PRINT_ERR("Error: wait DMA %u failed!\n", dev_id);
+
+	return async_cnt == 0 ? 0 : -1;
+}
+
 static inline int
 do_dma_plain_mem_copy(void *p)
 {
@@ -304,10 +342,8 @@ do_dma_plain_mem_copy(void *p)
 	const uint32_t buf_size = para->buf_size;
 	struct rte_mbuf **srcs = para->srcs;
 	struct rte_mbuf **dsts = para->dsts;
-	uint16_t nr_cpl;
 	uint64_t async_cnt = 0;
 	uint32_t i;
-	uint32_t poll_cnt = 0;
 	int ret;
 
 	worker_info->stop_flag = false;
@@ -338,13 +374,7 @@ do_dma_plain_mem_copy(void *p)
 			break;
 	}
 
-	rte_dma_submit(dev_id, 0);
-	while ((async_cnt > 0) && (poll_cnt++ < POLL_MAX)) {
-		nr_cpl = rte_dma_completed(dev_id, 0, MAX_DMA_CPL_NB, NULL, NULL);
-		async_cnt -= nr_cpl;
-	}
-
-	return 0;
+	return do_dma_submit_and_wait_cpl(dev_id, async_cnt, false);
 }
 
 static inline int
@@ -360,8 +390,6 @@ do_dma_sg_mem_copy(void *p)
 	const uint16_t dev_id = para->dev_id;
 	uint32_t nr_buf = para->nr_buf;
 	uint64_t async_cnt = 0;
-	uint32_t poll_cnt = 0;
-	uint16_t nr_cpl;
 	uint32_t i, j;
 	int ret;
 
@@ -397,13 +425,7 @@ do_dma_sg_mem_copy(void *p)
 			break;
 	}
 
-	rte_dma_submit(dev_id, 0);
-	while ((async_cnt > 0) && (poll_cnt++ < POLL_MAX)) {
-		nr_cpl = rte_dma_completed(dev_id, 0, MAX_DMA_CPL_NB, NULL, NULL);
-		async_cnt -= nr_cpl;
-	}
-
-	return 0;
+	return do_dma_submit_and_wait_cpl(dev_id, async_cnt, false);
 }
 
 static inline int
@@ -414,11 +436,11 @@ do_dma_enq_deq_mem_copy(void *p)
 	volatile struct worker_info *worker_info = &(para->worker_info);
 	struct rte_dma_op **dma_ops = para->dma_ops;
 	uint16_t kick_batch = para->kick_batch, sz;
-	uint16_t enq, deq, poll_cnt;
-	uint64_t tenq, tdeq;
 	const uint16_t dev_id = para->dev_id;
 	uint32_t nr_buf = para->nr_buf;
 	struct rte_dma_op *op[DEQ_SZ];
+	uint64_t tenq, tdeq;
+	uint16_t enq, deq;
 	uint32_t i;
 
 	worker_info->stop_flag = false;
@@ -454,11 +476,7 @@ do_dma_enq_deq_mem_copy(void *p)
 			break;
 	}
 
-	poll_cnt = 0;
-	while ((tenq != tdeq) && (poll_cnt++ < POLL_MAX))
-		tdeq += rte_dma_dequeue_ops(dev_id, 0, op, DEQ_SZ);
-
-	return 0;
+	return do_dma_submit_and_wait_cpl(dev_id, tenq - tdeq, true);
 }
 
 static inline int
@@ -614,7 +632,6 @@ setup_memory_env(struct test_configure *cfg, uint32_t nr_buf,
 		}
 
 		if (cfg->use_ops) {
-
 			nr_buf /= RTE_MAX(nb_src_sges, nb_dst_sges);
 			*dma_ops = rte_zmalloc(NULL, nr_buf * (sizeof(struct rte_dma_op *)),
 					       RTE_CACHE_LINE_SIZE);
-- 
2.17.1



More information about the dev mailing list