[dpdk-dev] [PATCH 2/3] driver/net/mpipe: optimize mpipe buffer return mechanism.

Tony Lu zlu at ezchip.com
Fri Jan 8 04:04:53 CET 2016


>-----Original Message-----
>From: dev [mailto:dev-bounces at dpdk.org] On Behalf Of Liming Sun
>Sent: Tuesday, December 15, 2015 11:38 PM
>To: dev at dpdk.org
>Subject: [dpdk-dev] [PATCH 2/3] driver/net/mpipe: optimize mpipe buffer
return
>mechanism.
>
>This submit has changes to optimize the mpipe buffer return. When
>a packet is received, instead of allocating and refilling the
>buffer stack right away, it tracks the number of pending buffers,
>and use HW buffer return as an optimization when the pending
>number is below certain threshold, thus save two MMIO writes and
>improves performance especially for bidirectional traffic case.
>
>Signed-off-by: Liming Sun <lsun at ezchip.com>
>---
> drivers/net/mpipe/mpipe_tilegx.c |   50
>++++++++++++++++++++++++++++++-------
> 1 files changed, 40 insertions(+), 10 deletions(-)
>
>diff --git a/drivers/net/mpipe/mpipe_tilegx.c
>b/drivers/net/mpipe/mpipe_tilegx.c
>index 35134ba..be7b6f2 100644
>--- a/drivers/net/mpipe/mpipe_tilegx.c
>+++ b/drivers/net/mpipe/mpipe_tilegx.c
>@@ -78,6 +78,13 @@ struct mpipe_context {
> 	struct mpipe_channel_config channels[MPIPE_MAX_CHANNELS];
> };
>
>+/* Per-core local data. */
>+struct mpipe_local {
>+	int mbuf_push_debt[RTE_MAX_ETHPORTS];	/* Buffer push debt. */
>+} __rte_cache_aligned;
>+
>+#define MPIPE_BUF_DEBT_THRESHOLD	32
>+static __thread struct mpipe_local mpipe_local;
> static struct mpipe_context mpipe_contexts[GXIO_MPIPE_INSTANCE_MAX];
> static int mpipe_instances;
> static const char *drivername = "MPIPE PMD";
>@@ -137,7 +144,7 @@ struct mpipe_dev_priv {
> 	int first_bucket;		/* mPIPE bucket start index. */
> 	int first_ring;			/* mPIPE notif ring start index. */
> 	int notif_group;		/* mPIPE notif group. */
>-	rte_atomic32_t dp_count;	/* Active datapath thread count. */
>+	rte_atomic32_t dp_count __rte_cache_aligned;	/* DP Entry count.
*/
> 	int tx_stat_mapping[RTE_ETHDEV_QUEUE_STAT_CNTRS];
> 	int rx_stat_mapping[RTE_ETHDEV_QUEUE_STAT_CNTRS];
> };
>@@ -461,6 +468,14 @@ mpipe_dp_wait(struct mpipe_dev_priv *priv)
> 	}
> }
>
>+static inline int
>+mpipe_mbuf_stack_index(struct mpipe_dev_priv *priv, struct rte_mbuf *mbuf)
>+{
>+	return (mbuf->port < RTE_MAX_ETHPORTS)?
>+		mpipe_priv(&rte_eth_devices[mbuf->port])->stack :
>+		priv->stack;
>+}
>+
> static inline struct rte_mbuf *
> mpipe_recv_mbuf(struct mpipe_dev_priv *priv, gxio_mpipe_idesc_t *idesc,
> 		int in_port)
>@@ -1267,6 +1282,7 @@ mpipe_do_xmit(struct mpipe_tx_queue *tx_queue,
>struct rte_mbuf **tx_pkts,
> 	unsigned nb_bytes = 0;
> 	unsigned nb_sent = 0;
> 	int nb_slots, i;
>+	uint8_t port_id;
>
> 	PMD_DEBUG_TX("Trying to transmit %d packets on %s:%d.\n",
> 		     nb_pkts, mpipe_name(tx_queue->q.priv),
>@@ -1315,14 +1331,23 @@ mpipe_do_xmit(struct mpipe_tx_queue *tx_queue,
>struct rte_mbuf **tx_pkts,
> 			if (priv->tx_comps[idx])
> 				rte_pktmbuf_free_seg(priv->tx_comps[idx]);
>
>+			port_id = (mbuf->port < RTE_MAX_ETHPORTS)?
>+						mbuf->port : priv->port_id;
> 			desc = (gxio_mpipe_edesc_t) { {
> 				.va        = rte_pktmbuf_mtod(mbuf,
uintptr_t),
> 				.xfer_size = rte_pktmbuf_data_len(mbuf),
> 				.bound     = next ? 0 : 1,
>+				.stack_idx = mpipe_mbuf_stack_index(priv,
mbuf),
> 			} };
>+			if (mpipe_local.mbuf_push_debt[port_id] > 0) {
>+				mpipe_local.mbuf_push_debt[port_id]--;
>+				desc.hwb = 1;
>+				priv->tx_comps[idx] = NULL;
>+			}
>+			else
>+				priv->tx_comps[idx] = mbuf;
>
> 			nb_bytes += mbuf->data_len;
>-			priv->tx_comps[idx] = mbuf;
> 			gxio_mpipe_equeue_put_at(equeue, desc, slot + i);
>
> 			PMD_DEBUG_TX("%s:%d: Sending packet %p, len %d\n",
>@@ -1443,17 +1468,22 @@ mpipe_do_recv(struct mpipe_rx_queue *rx_queue,
>struct rte_mbuf **rx_pkts,
> 				continue;
> 			}
>
>-			mbuf = __rte_mbuf_raw_alloc(priv->rx_mpool);
>-			if (unlikely(!mbuf)) {
>-				nb_nomem++;
>-				gxio_mpipe_iqueue_drop(iqueue, idesc);
>-				PMD_DEBUG_RX("%s:%d: RX alloc failure\n",
>+			if (mpipe_local.mbuf_push_debt[in_port] <
>+					MPIPE_BUF_DEBT_THRESHOLD)
>+				mpipe_local.mbuf_push_debt[in_port]++;
>+			else {
>+				mbuf = __rte_mbuf_raw_alloc(priv->rx_mpool);
>+				if (unlikely(!mbuf)) {
>+					nb_nomem++;
>+					gxio_mpipe_iqueue_drop(iqueue,
idesc);
>+					PMD_DEBUG_RX("%s:%d: alloc
failure\n",
> 					     mpipe_name(rx_queue->q.priv),
> 					     rx_queue->q.queue_idx);
>-				continue;
>-			}
>+					continue;
>+				}
>
>-			mpipe_recv_push(priv, mbuf);
>+				mpipe_recv_push(priv, mbuf);
>+			}
>
> 			/* Get and setup the mbuf for the received packet.
*/
> 			mbuf = mpipe_recv_mbuf(priv, idesc, in_port);
>--
>1.7.1

Acked-by: Zhigang Lu <zlu at ezchip.com>



More information about the dev mailing list