dpdk Tx falling short
Ivan Malov
ivan.malov at arknetworks.am
Mon Jul 7 23:49:44 CEST 2025
Hi Ed,
On Mon, 7 Jul 2025, Lombardo, Ed wrote:
> Hi Stephen,
> I ran a perf diff on two perf records and reveals the real problem with the tx thread in transmitting packets.
>
> The comparison is traffic received on ifn3 and transmit ifn4 to traffic received on ifn3, ifn5 and transmit on ifn4, ifn6.
> When transmit packets on one port the performance is better, however when transmit on two ports the performance across the two drops dramatically.
>
> There is increase of 55.29% of the CPU spent in common_ring_mp_enqueue and 54.18% less time in i40e_xmit_pkts (was E810 tried x710).
> The common_ring_mp_enqueue is multi-producer, is the enqueue of mbuf pointers passed in to rte_eth_tx_burst() have to be multi-producer?
I may be wrong, but rte_eth_tx_burst(), as part of what is known as "reap"
process, should check for "done" Tx descriptors resulting from previous
invocations and free (enqueue) the associated mbufs into respective mempools.
In your case, you say you only have a single mempool shared between the port
pairs, which, as I understand, are served by concurrent threads, so it might be
logical to use a multi-producer mempool in this case. Or am I missing something?
The pktmbuf API for mempool allocation is a wrapper around generic API and it
might request multi-producer multi-consumer by default (see [1], 'flags').
According to your original mempool monitor printout, the per-lcore cache size is
512. On the premise that separate lcores serve the two port pairs, and taking
into account the burst size, it should be OK, yet you may want to play with the
per-lcore cache size argument when creating the pool. Does it change anything?
Regarding separate mempools, -- I saw Stephen's response about those making CPU
cache behaviour worse and not better. Makes sense and I won't argue. And yet,
why not just try an make sure this indeed holds in this particular case? Also,
since you're seeking single-producer behaviour, having separate per-port-pair
mempools might allow to create such (again, see 'flags' at [1]), provided that
API [1] is used for mempool creation. Please correct me in case I'm mistaken.
Also, PMDs can support "fast free" Tx offload. Please see [2] to check whether
the application asks for this offload flag or not. It may be worth enabling.
[1] https://doc.dpdk.org/api-25.03/rte__mempool_8h.html#a0b64d611bc140a4d2a0c94911580efd5
[2] https://doc.dpdk.org/api-25.03/rte__ethdev_8h.html#a43f198c6b59d965130d56fd8f40ceac1
Thank you.
>
> Is there a way to change dpdk to use single-producer?
>
> # Event 'cycles'
> #
> # Baseline Delta Abs Shared Object Symbol
> # ........ ......... ................. ......................................
> #
> 36.37% +55.29% test [.] common_ring_mp_enqueue
> 62.36% -54.18% test [.] i40e_xmit_pkts
> 1.10% -0.94% test [.] dpdk_tx_thread
> 0.01% -0.01% [kernel.kallsyms] [k] native_sched_clock
> +0.00% [kernel.kallsyms] [k] fill_pmd
> +0.00% [kernel.kallsyms] [k] perf_sample_event_took
> 0.00% +0.00% [kernel.kallsyms] [k] __flush_smp_call_function_queue
> 0.02% [kernel.kallsyms] [k] __intel_pmu_enable_all.constprop.0
> 0.02% [kernel.kallsyms] [k] native_irq_return_iret
> 0.02% [kernel.kallsyms] [k] native_tss_update_io_bitmap
> 0.01% [kernel.kallsyms] [k] ktime_get
> 0.01% [kernel.kallsyms] [k] perf_adjust_freq_unthr_context
> 0.01% [kernel.kallsyms] [k] __update_blocked_fair
> 0.01% [kernel.kallsyms] [k] perf_adjust_freq_unthr_events
>
> Thanks,
> Ed
>
> -----Original Message-----
> From: Lombardo, Ed
> Sent: Sunday, July 6, 2025 1:45 PM
> To: Stephen Hemminger <stephen at networkplumber.org>
> Cc: Ivan Malov <ivan.malov at arknetworks.am>; users <users at dpdk.org>
> Subject: RE: dpdk Tx falling short
>
> Hi Stephen,
> If using dpdk rings comes with this penalty then what should I use, is there an alterative to rings. We do not want to use shared memory and do buffer copies?
>
> Thanks,
> Ed
>
> -----Original Message-----
> From: Stephen Hemminger <stephen at networkplumber.org>
> Sent: Sunday, July 6, 2025 12:03 PM
> To: Lombardo, Ed <Ed.Lombardo at netscout.com>
> Cc: Ivan Malov <ivan.malov at arknetworks.am>; users <users at dpdk.org>
> Subject: Re: dpdk Tx falling short
>
> External Email: This message originated outside of NETSCOUT. Do not click links or open attachments unless you recognize the sender and know the content is safe.
>
> On Sun, 6 Jul 2025 00:03:16 +0000
> "Lombardo, Ed" <Ed.Lombardo at netscout.com> wrote:
>
>> Hi Stephen,
>> Here are comments to the list of obvious causes of cache misses you mentiond.
>>
>> Obvious cache misses.
>> - passing packets to worker with ring - we use lots of rings to pass mbuf pointers. If I skip the rte_eth_tx_burst() and just free mbuf bulk, the tx ring does not fill up.
>> - using spinlocks (cost 16ns) - The driver does not use spinlocks, other than what dpdk uses.
>> - fetching TSC - We don't do this, we let Rx offload timestamp packets.
>> - syscalls? - No syscalls are done in our driver fast path.
>>
>> You mention "passing packets to worker with ring", do you mean using rings to pass mbuf pointers causes cache misses and should be avoided?
>
> Rings do cause data to be modified by one core and examined by another so they are a cache miss.
>
>
More information about the users
mailing list