[PATCH v3 06/36] net/intel: add common fn to calculate needed descriptors
Loftus, Ciara
ciara.loftus at intel.com
Fri Feb 6 11:25:36 CET 2026
> Subject: [PATCH v3 06/36] net/intel: add common fn to calculate needed
> descriptors
>
> Multiple drivers used the same logic to calculate how many Tx data
> descriptors were needed. Move that calculation to common code. In the
> process of updating drivers, fix idpf driver calculation for the TSO
> case.
Should this fix be split out into a separate patch with a Fixes tag?
>
> Signed-off-by: Bruce Richardson <bruce.richardson at intel.com>
> ---
> drivers/net/intel/common/tx_scalar_fns.h | 21 +++++++++++++++++++++
> drivers/net/intel/i40e/i40e_rxtx.c | 18 +-----------------
> drivers/net/intel/iavf/iavf_rxtx.c | 17 +----------------
> drivers/net/intel/ice/ice_rxtx.c | 18 +-----------------
> drivers/net/intel/idpf/idpf_common_rxtx.c | 21 +++++++++++++++++----
> 5 files changed, 41 insertions(+), 54 deletions(-)
>
> diff --git a/drivers/net/intel/common/tx_scalar_fns.h
> b/drivers/net/intel/common/tx_scalar_fns.h
> index c79210d084..f894cea616 100644
> --- a/drivers/net/intel/common/tx_scalar_fns.h
> +++ b/drivers/net/intel/common/tx_scalar_fns.h
> @@ -64,4 +64,25 @@ ci_tx_xmit_cleanup(struct ci_tx_queue *txq)
> return 0;
> }
>
> +static inline uint16_t
> +ci_div_roundup16(uint16_t x, uint16_t y)
> +{
> + return (uint16_t)((x + y - 1) / y);
> +}
> +
> +/* Calculate the number of TX descriptors needed for each pkt */
> +static inline uint16_t
> +ci_calc_pkt_desc(const struct rte_mbuf *tx_pkt)
> +{
> + uint16_t count = 0;
> +
> + while (tx_pkt != NULL) {
> + count += ci_div_roundup16(tx_pkt->data_len,
> CI_MAX_DATA_PER_TXD);
> + tx_pkt = tx_pkt->next;
> + }
> +
> + return count;
> +}
> +
> +
> #endif /* _COMMON_INTEL_TX_SCALAR_FNS_H_ */
> diff --git a/drivers/net/intel/i40e/i40e_rxtx.c
> b/drivers/net/intel/i40e/i40e_rxtx.c
> index f96c5c7f1e..b75306931a 100644
> --- a/drivers/net/intel/i40e/i40e_rxtx.c
> +++ b/drivers/net/intel/i40e/i40e_rxtx.c
> @@ -1029,21 +1029,6 @@ i40e_set_tso_ctx(struct rte_mbuf *mbuf, union
> ci_tx_offload tx_offload)
> return ctx_desc;
> }
>
> -/* Calculate the number of TX descriptors needed for each pkt */
> -static inline uint16_t
> -i40e_calc_pkt_desc(struct rte_mbuf *tx_pkt)
> -{
> - struct rte_mbuf *txd = tx_pkt;
> - uint16_t count = 0;
> -
> - while (txd != NULL) {
> - count += DIV_ROUND_UP(txd->data_len,
> CI_MAX_DATA_PER_TXD);
> - txd = txd->next;
> - }
> -
> - return count;
> -}
> -
> uint16_t
> i40e_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t
> nb_pkts)
> {
> @@ -1106,8 +1091,7 @@ i40e_xmit_pkts(void *tx_queue, struct rte_mbuf
> **tx_pkts, uint16_t nb_pkts)
> * per tx desc.
> */
> if (ol_flags & RTE_MBUF_F_TX_TCP_SEG)
> - nb_used = (uint16_t)(i40e_calc_pkt_desc(tx_pkt) +
> - nb_ctx);
> + nb_used = (uint16_t)(ci_calc_pkt_desc(tx_pkt) +
> nb_ctx);
> else
> nb_used = (uint16_t)(tx_pkt->nb_segs + nb_ctx);
> tx_last = (uint16_t)(tx_id + nb_used - 1);
> diff --git a/drivers/net/intel/iavf/iavf_rxtx.c
> b/drivers/net/intel/iavf/iavf_rxtx.c
> index 947b6c24d2..885d9309cc 100644
> --- a/drivers/net/intel/iavf/iavf_rxtx.c
> +++ b/drivers/net/intel/iavf/iavf_rxtx.c
> @@ -2666,21 +2666,6 @@ iavf_build_data_desc_cmd_offset_fields(volatile
> uint64_t *qw1,
> ((uint64_t)l2tag1 << IAVF_TXD_DATA_QW1_L2TAG1_SHIFT));
> }
>
> -/* Calculate the number of TX descriptors needed for each pkt */
> -static inline uint16_t
> -iavf_calc_pkt_desc(struct rte_mbuf *tx_pkt)
> -{
> - struct rte_mbuf *txd = tx_pkt;
> - uint16_t count = 0;
> -
> - while (txd != NULL) {
> - count += (txd->data_len + CI_MAX_DATA_PER_TXD - 1) /
> CI_MAX_DATA_PER_TXD;
> - txd = txd->next;
> - }
> -
> - return count;
> -}
> -
> static inline void
> iavf_fill_data_desc(volatile struct ci_tx_desc *desc,
> uint64_t desc_template, uint16_t buffsz,
> @@ -2766,7 +2751,7 @@ iavf_xmit_pkts(void *tx_queue, struct rte_mbuf
> **tx_pkts, uint16_t nb_pkts)
> * per tx desc.
> */
> if (mb->ol_flags & RTE_MBUF_F_TX_TCP_SEG)
> - nb_desc_required = iavf_calc_pkt_desc(mb) +
> nb_desc_ctx + nb_desc_ipsec;
> + nb_desc_required = ci_calc_pkt_desc(mb) +
> nb_desc_ctx + nb_desc_ipsec;
> else
> nb_desc_required = nb_desc_data + nb_desc_ctx +
> nb_desc_ipsec;
>
> diff --git a/drivers/net/intel/ice/ice_rxtx.c b/drivers/net/intel/ice/ice_rxtx.c
> index 52bbf95967..2a53b614b2 100644
> --- a/drivers/net/intel/ice/ice_rxtx.c
> +++ b/drivers/net/intel/ice/ice_rxtx.c
> @@ -3075,21 +3075,6 @@ ice_set_tso_ctx(struct rte_mbuf *mbuf, union
> ci_tx_offload tx_offload)
> return ctx_desc;
> }
>
> -/* Calculate the number of TX descriptors needed for each pkt */
> -static inline uint16_t
> -ice_calc_pkt_desc(struct rte_mbuf *tx_pkt)
> -{
> - struct rte_mbuf *txd = tx_pkt;
> - uint16_t count = 0;
> -
> - while (txd != NULL) {
> - count += DIV_ROUND_UP(txd->data_len,
> CI_MAX_DATA_PER_TXD);
> - txd = txd->next;
> - }
> -
> - return count;
> -}
> -
> uint16_t
> ice_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts)
> {
> @@ -3152,8 +3137,7 @@ ice_xmit_pkts(void *tx_queue, struct rte_mbuf
> **tx_pkts, uint16_t nb_pkts)
> * per tx desc.
> */
> if (ol_flags & (RTE_MBUF_F_TX_TCP_SEG |
> RTE_MBUF_F_TX_UDP_SEG))
> - nb_used = (uint16_t)(ice_calc_pkt_desc(tx_pkt) +
> - nb_ctx);
> + nb_used = (uint16_t)(ci_calc_pkt_desc(tx_pkt) +
> nb_ctx);
> else
> nb_used = (uint16_t)(tx_pkt->nb_segs + nb_ctx);
> tx_last = (uint16_t)(tx_id + nb_used - 1);
> diff --git a/drivers/net/intel/idpf/idpf_common_rxtx.c
> b/drivers/net/intel/idpf/idpf_common_rxtx.c
> index 587871b54a..11d6848430 100644
> --- a/drivers/net/intel/idpf/idpf_common_rxtx.c
> +++ b/drivers/net/intel/idpf/idpf_common_rxtx.c
> @@ -934,7 +934,16 @@ idpf_dp_splitq_xmit_pkts(void *tx_queue, struct
> rte_mbuf **tx_pkts,
> tx_offload.tso_segsz = tx_pkt->tso_segsz;
> /* Calculate the number of context descriptors needed. */
> nb_ctx = idpf_calc_context_desc(ol_flags);
> - nb_used = tx_pkt->nb_segs + nb_ctx;
> +
> + /* Calculate the number of TX descriptors needed for
> + * each packet. For TSO packets, use ci_calc_pkt_desc as
> + * the mbuf data size might exceed max data size that hw
> allows
> + * per tx desc.
> + */
> + if (ol_flags & RTE_MBUF_F_TX_TCP_SEG)
> + nb_used = ci_calc_pkt_desc(tx_pkt) + nb_ctx;
> + else
> + nb_used = tx_pkt->nb_segs + nb_ctx;
>
> if (ol_flags & IDPF_TX_CKSUM_OFFLOAD_MASK)
> cmd_dtype = IDPF_TXD_FLEX_FLOW_CMD_CS_EN;
> @@ -1382,10 +1391,14 @@ idpf_dp_singleq_xmit_pkts(void *tx_queue,
> struct rte_mbuf **tx_pkts,
> nb_ctx = idpf_calc_context_desc(ol_flags);
>
> /* The number of descriptors that must be allocated for
> - * a packet equals to the number of the segments of that
> - * packet plus 1 context descriptor if needed.
> + * a packet. For TSO packets, use ci_calc_pkt_desc as
> + * the mbuf data size might exceed max data size that hw
> allows
> + * per tx desc.
> */
> - nb_used = (uint16_t)(tx_pkt->nb_segs + nb_ctx);
> + if (ol_flags & RTE_MBUF_F_TX_TCP_SEG)
> + nb_used = (uint16_t)(ci_calc_pkt_desc(tx_pkt) +
> nb_ctx);
> + else
> + nb_used = (uint16_t)(tx_pkt->nb_segs + nb_ctx);
> tx_last = (uint16_t)(tx_id + nb_used - 1);
>
> /* Circular ring */
> --
> 2.51.0
More information about the dev
mailing list