[dpdk-dev] [PATCH] igb: fix crash with offload on 82575 chipset
Ananyev, Konstantin
konstantin.ananyev at intel.com
Fri Mar 25 15:06:51 CET 2016
> -----Original Message-----
> From: dev [mailto:dev-bounces at dpdk.org] On Behalf Of Olivier Matz
> Sent: Friday, March 25, 2016 10:32 AM
> To: dev at dpdk.org
> Cc: Lu, Wenzhuo
> Subject: [dpdk-dev] [PATCH] igb: fix crash with offload on 82575 chipset
>
> On the 82575 chipset, there is a pool of global TX contexts instead of 2
> per queues on 82576. See Table A-1 "Changes in Programming Interface
> Relative to 82575" of Intel® 82576EB GbE Controller datasheet (*).
>
> In the driver, the contexts are attributed to a TX queue: 0-1 for txq0,
> 2-3 for txq1, and so on.
>
> In igbe_set_xmit_ctx(), the variable ctx_curr contains the index of the
> per-queue context (0 or 1), and ctx_idx contains the index to be given
> to the hardware (0 to 7). The size of txq->ctx_cache[] is 2, and must
> be indexed with ctx_curr to avoid an out-of-bound access.
>
> Also, the index returned by what_advctx_update() is the per-queue
> index (0 or 1), so we need to add txq->ctx_start before sending it
> to the hardware.
>
> (*) The datasheets says 16 global contexts, however the IDX fields in TX
> descriptors are 3 bits, which gives a total of 8 contexts. The
> driver assumes there are 8 contexts on 82575: 2 per queues, 4 txqs.
>
> Fixes: 4c8db5f09a ("igb: enable TSO support")
> Fixes: af75078fec ("first public release")
> Signed-off-by: Olivier Matz <olivier.matz at 6wind.com>
> ---
> drivers/net/e1000/igb_rxtx.c | 6 +++---
> 1 file changed, 3 insertions(+), 3 deletions(-)
>
> diff --git a/drivers/net/e1000/igb_rxtx.c b/drivers/net/e1000/igb_rxtx.c
> index e527895..529dba4 100644
> --- a/drivers/net/e1000/igb_rxtx.c
> +++ b/drivers/net/e1000/igb_rxtx.c
> @@ -325,9 +325,9 @@ igbe_set_xmit_ctx(struct igb_tx_queue* txq,
> }
>
> txq->ctx_cache[ctx_curr].flags = ol_flags;
> - txq->ctx_cache[ctx_idx].tx_offload.data =
> + txq->ctx_cache[ctx_curr].tx_offload.data =
> tx_offload_mask.data & tx_offload.data;
> - txq->ctx_cache[ctx_idx].tx_offload_mask = tx_offload_mask;
> + txq->ctx_cache[ctx_curr].tx_offload_mask = tx_offload_mask;
>
> ctx_txd->type_tucmd_mlhl = rte_cpu_to_le_32(type_tucmd_mlhl);
> vlan_macip_lens = (uint32_t)tx_offload.data;
> @@ -450,7 +450,7 @@ eth_igb_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
> ctx = what_advctx_update(txq, tx_ol_req, tx_offload);
> /* Only allocate context descriptor if required*/
> new_ctx = (ctx == IGB_CTX_NUM);
> - ctx = txq->ctx_curr;
> + ctx = txq->ctx_curr + txq->ctx_start;
> tx_last = (uint16_t) (tx_last + new_ctx);
> }
> if (tx_last >= txq->nb_tx_desc)
> --
Acked-by: Konstantin Ananyev <konstantin.ananyev at intel.com>
> 2.1.4
More information about the dev
mailing list