[dpdk-dev] 8023ad bond tx crashed if one port has 2 more tx queues

hfli at netitest.com hfli at netitest.com
Tue Mar 19 10:41:25 CET 2019

Hi Guys,


I found a bug in dpdk bond code, while one port has 2 more tx queues, 8023ad
bond port will be crashed in tx burst.


Just analyzed the code below, if 2 more CPU cores send packets on a port by
different tx queue, the arrays like 

slave_port_ids/ dist_slave_port_ids/ slave_tx_fail_coun/ slave_bufs will be
shared by all of cores, it will be crashed in this function.


Is there any better solution for this? For now, I just add lock for



static uint16_t

bond_ethdev_tx_burst_8023ad(void *queue, struct rte_mbuf **bufs,

                   uint16_t nb_bufs)


         struct bond_tx_queue *bd_tx_q = (struct bond_tx_queue *)queue;

         struct bond_dev_private *internals = bd_tx_q->dev_private;


         uint16_t [RTE_MAX_ETHPORTS];

         uint16_t slave_count;


         uint16_t dist_slave_port_ids[RTE_MAX_ETHPORTS];

         uint16_t dist_slave_count;


         /* 2-D array to sort mbufs for transmission on each slave into */

         struct rte_mbuf *slave_bufs[RTE_MAX_ETHPORTS][nb_bufs];

         /* Number of mbufs for transmission on each slave */

         uint16_t slave_nb_bufs[RTE_MAX_ETHPORTS] = { 0 };

         /* Mapping array generated by hash function to map mbufs to slaves

         uint16_t bufs_slave_port_idxs[RTE_MAX_ETHPORTS] = { 0 };


         uint16_t slave_tx_count, slave_tx_fail_count[RTE_MAX_ETHPORTS] = {
0 };

         uint16_t total_tx_count = 0, total_tx_fail_count = 0;




Thanks and Regards,


More information about the dev mailing list