[dpdk-users] X710 DA2 (2x10G) performance 64B packet

Paul T paultop6 at outlook.com
Thu Mar 21 12:50:38 CET 2019


Hi Tomas,

I would isolate the CPUs in which the dpdk threads are running from the linux schedular.  The low packet drop at 64B makes me thing its context switching happen on the core because of the linux scheduler.

Use the following command in the linux command line params in your grub config:
isolcpus=cpus to isolate, e.g. 1,3,4 or 1-4

Regards

Paul

Message: 3
Date: Thu, 21 Mar 2019 10:53:34 +0100
From: Tom?? J?nsk? <tomas.jansky at flowmon.com>
To: users at dpdk.org
Subject: [dpdk-users] X710 DA2 (2x10G) performance 64B packets
Message-ID:
        <CAPP7y6z13qFR-34+-Xn97ru5jOnaVAV7s=6WPgk_j=9CLMQrSQ at mail.gmail.com>
Content-Type: text/plain; charset="UTF-8"

Hello DPDK users,

I am having an issue concerning the performance of X710 DA2 (2x10G) NIC
when using testpmd (and also l2fwd) application on both ports.

HW and SW parameters:
CPUs: Intel(R) Xeon(R) CPU E5-2620 v4 @ 2.10GHz x16
Disabled hyperthreading.
All used lcores and ports are on the same NUMA node (0).
Hugepages: 1024x 2MB on the NUMA node 0.
RAM: 64 GB

DPDK version: 18.05.1
Modue: IGB UIO
GCC version: 4.8.5

When using testpmd application only on one port:
./testpmd -b 0000:04:00.0 -n 4 --lcore=0 at 0,2 at 2 -- --socket-num=0
--nb-cores=1 --nb-ports=1 --numa --forward-mode=rxonly

14.63 Mpps (64B packet length) - 0.01% packets dropped

When using testmpd on both ports:
./testpmd -n 4 --lcore=0 at 0,2 at 2,4 at 4 -- --socket-num=0 --nb-cores=2
--nb-ports=2 --numa --forward-mode=rxonly

28.08 Mpps (64B packet length) - 3.47% packets dropped

Does anybody have an explanation why am I experiencing this performance
drop?
Any suggestion would be much appreciated.

Thank you
Tomas


More information about the users mailing list