[dpdk-users] l2fwd performance in VM with SR-IOV

Furong WBAHACER at 126.com
Sat Dec 19 07:23:20 CET 2015

Hello, everybody.
     I have measured performace of example/l2fwd in VM with SR-IOV.
     My experiment server: CPU: 32 core Intel Xeon E5-4603 v2 @ 
2.20GHz,  NIC: 10G Intel 82599ES, OS:ubuntu14.04.3.
     I started a VM with this command:
         # qemu-system-x86_64 -enable-kvm -cpu host -m 4G -smp 4 -net 
none -device vfio-pci,host=<vf1-pcie-addr> -device 
vfio-pci,host=<vf2-pcie-addr> -hda vm.img -vnc :1
     In VM:
         I bound vf1 & vf2 to igb_uio, then started a example/l2fwd in VM.
     Then i started a pktgen in another server (same hardware & os with 
this server) to send packets (small packet - 64bit).
     The results is :
         1. when i sent packets with pktgen from only 1 port , the 
throughput (measured by pktgen rx/tx rates) was 7.0Gbps.
         2. when i sent packets from both 2 port, the throughput was 
7.2Gbps (3.6Gbps each port).

     But, i have measured l2fwd performance in host with SR-IOV (binding 
vf1 & vf2 to vfio-pci & starting l2fwd in host).
     The result is :
         when i sent packets from both 2 port, the throughput was 
14.4Gbps (7.2Gbps each port).

     I want to ask when i ran l2fwd in VM, Can i achieve similar 
performance with host? or, there are some methods to tune the performance ?

     Thanks a lot!

More information about the users mailing list