rte_pktmbuf_alloc() out of rte_mbufs

amit sehas cun23 at yahoo.com
Wed Nov 27 02:21:04 CET 2024


looking at code for rte_eth_tx_buffer_flush(), in the error case where some buffers were not sent, it has a default callback which will enqueue the buffers back to the mempool, and in the non error case as well when the call is done it will enqueue the buffers back to the mempool ...

this is what we have relied on for ever, otherwise we would not be able to utilize it at all ... in the transmit case we never explicitly free the buffers ...and we are able to run the product through 100s of millions of packet transmits, its only under special circumstances that we run into the said issue ..

I hope i am not misunderstanding something. 


regards







On Tuesday, November 26, 2024 at 04:51:09 PM PST, Stephen Hemminger <stephen at networkplumber.org> wrote: 





On Tue, 26 Nov 2024 23:50:25 +0000 (UTC)

amit sehas <cun23 at yahoo.com> wrote:

> Dumping the stats every 10 minutes suggest that there is no slow leak of buffers. The problem arises when the system is under stress and starts performing extra disk i/o. In this situation dpdk accumulates the buffers and does not return them back to the mempool right away thereby accumulating all the 4k buffers allocated to the queue.
> 
> rte_eth_tx_buffer_flush() should be flushing the buffers and returning them to the mempool ... is there any additional API that can make sure that this happens.


If you read the code in rte_ethdev.h
The rte_eth_tx_buffer_flush is just does a send of the packets that application has aggregated
via rte_eth_tx_buffer.

It does nothing vis-a-vis mempools are causing the driver (PMD) to complete transmits.
There are some tuneables such as tx_free_thresh which control when driver should
start freeing sent mbufs. 

Have you isolated the CPU's used by DPDK threads?
Is the application stalling because it starts swapping. You may have to mlockall to keep the
pages of application from swapping out.



More information about the users mailing list