[dpdk-dev] [PATCH v6 3/3] app/testpmd: enable TCP/IPv4 GRO

Yao, Lei A lei.a.yao at intel.com
Sat Jun 24 10:01:02 CEST 2017



> -----Original Message-----
> From: Hu, Jiayu
> Sent: Friday, June 23, 2017 10:43 PM
> To: dev at dpdk.org
> Cc: Ananyev, Konstantin <konstantin.ananyev at intel.com>; Tan, Jianfeng
> <jianfeng.tan at intel.com>; stephen at networkplumber.org;
> yliu at fridaylinux.org; Wiles, Keith <keith.wiles at intel.com>; Bie, Tiwei
> <tiwei.bie at intel.com>; Yao, Lei A <lei.a.yao at intel.com>; Hu, Jiayu
> <jiayu.hu at intel.com>
> Subject: [PATCH v6 3/3] app/testpmd: enable TCP/IPv4 GRO
> 
> This patch enables TCP/IPv4 GRO library in csum forwarding engine.
> By default, GRO is turned off. Users can use command "gro (on|off)
> (port_id)" to enable or disable GRO for a given port. If a port is
> enabled GRO, all TCP/IPv4 packets received from the port are performed
> GRO. Besides, users can set max flow number and packets number per-flow
> by command "gro set (max_flow_num) (max_item_num_per_flow)
> (port_id)".
> 
> Signed-off-by: Jiayu Hu <jiayu.hu at intel.com>
Tested-By: Lei Yao<lei.a.yao at intel.com>
This patch is tested on the following test bench:
OS: Ubuntu 16.04
CPU: Intel(R) Xeon(R) CPU E5-2680 v2 @ 2.80GHz
NIC: XXV710 25G
We can see the iperf result improve a lot after enable GRO. The data
flow is NIC1->NIC2->testpmd(GRO on/off)->vhost->virtio-net(in VM) 

> ---
>  app/test-pmd/cmdline.c                      | 125
> ++++++++++++++++++++++++++++
>  app/test-pmd/config.c                       |  37 ++++++++
>  app/test-pmd/csumonly.c                     |   5 ++
>  app/test-pmd/testpmd.c                      |   3 +
>  app/test-pmd/testpmd.h                      |  11 +++
>  doc/guides/testpmd_app_ug/testpmd_funcs.rst |  34 ++++++++
>  6 files changed, 215 insertions(+)
> 
> diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c
> index ff8ffd2..cb359e1 100644
> --- a/app/test-pmd/cmdline.c
> +++ b/app/test-pmd/cmdline.c
> @@ -76,6 +76,7 @@
>  #include <rte_devargs.h>
>  #include <rte_eth_ctrl.h>
>  #include <rte_flow.h>
> +#include <rte_gro.h>
> 
>  #include <cmdline_rdline.h>
>  #include <cmdline_parse.h>
> @@ -419,6 +420,14 @@ static void cmd_help_long_parsed(void
> *parsed_result,
>  			"tso show (portid)"
>  			"    Display the status of TCP Segmentation
> Offload.\n\n"
> 
> +			"gro (on|off) (port_id)"
> +			"    Enable or disable Generic Receive Offload in io"
> +			" forward engine.\n\n"
> +
> +			"gro set (max_flow_num)
> (max_item_num_per_flow) (port_id)\n"
> +			"    Set max flow number and max packet number
> per-flow"
> +			" for GRO.\n\n"
> +
>  			"set fwd (%s)\n"
>  			"    Set packet forwarding mode.\n\n"
> 
> @@ -3827,6 +3836,120 @@ cmdline_parse_inst_t cmd_tunnel_tso_show = {
>  	},
>  };
> 
> +/* *** SET GRO FOR A PORT *** */
> +struct cmd_gro_result {
> +	cmdline_fixed_string_t cmd_keyword;
> +	cmdline_fixed_string_t mode;
> +	uint8_t port_id;
> +};
> +
> +static void
> +cmd_enable_gro_parsed(void *parsed_result,
> +		__attribute__((unused)) struct cmdline *cl,
> +		__attribute__((unused)) void *data)
> +{
> +	struct cmd_gro_result *res;
> +
> +	res = parsed_result;
> +	setup_gro(res->mode, res->port_id);
> +}
> +
> +cmdline_parse_token_string_t cmd_gro_keyword =
> +	TOKEN_STRING_INITIALIZER(struct cmd_gro_result,
> +			cmd_keyword, "gro");
> +cmdline_parse_token_string_t cmd_gro_mode =
> +	TOKEN_STRING_INITIALIZER(struct cmd_gro_result,
> +			mode, "on#off");
> +cmdline_parse_token_num_t cmd_gro_pid =
> +	TOKEN_NUM_INITIALIZER(struct cmd_gro_result,
> +			port_id, UINT8);
> +
> +cmdline_parse_inst_t cmd_enable_gro = {
> +	.f = cmd_enable_gro_parsed,
> +	.data = NULL,
> +	.help_str = "gro (on|off) (port_id)",
> +	.tokens = {
> +		(void *)&cmd_gro_keyword,
> +		(void *)&cmd_gro_mode,
> +		(void *)&cmd_gro_pid,
> +		NULL,
> +	},
> +};
> +
> +/* *** SET MAX FLOW NUMBER AND ITEM NUM PER FLOW FOR GRO ***
> */
> +struct cmd_gro_set_result {
> +	cmdline_fixed_string_t gro;
> +	cmdline_fixed_string_t mode;
> +	uint16_t flow_num;
> +	uint16_t item_num_per_flow;
> +	uint8_t port_id;
> +};
> +
> +static void
> +cmd_gro_set_parsed(void *parsed_result,
> +		       __attribute__((unused)) struct cmdline *cl,
> +		       __attribute__((unused)) void *data)
> +{
> +	struct cmd_gro_set_result *res = parsed_result;
> +
> +	if (port_id_is_invalid(res->port_id, ENABLED_WARN))
> +		return;
> +	if (test_done == 0) {
> +		printf("Before set GRO flow_num and item_num_per_flow,"
> +				" please stop forwarding first\n");
> +		return;
> +	}
> +
> +	if (!strcmp(res->mode, "set")) {
> +		if (res->flow_num == 0)
> +			printf("Invalid flow number. Revert to default value:"
> +					" %u.\n",
> GRO_DEFAULT_FLOW_NUM);
> +		else
> +			gro_ports[res->port_id].param.max_flow_num =
> +				res->flow_num;
> +
> +		if (res->item_num_per_flow == 0)
> +			printf("Invalid item number per-flow. Revert"
> +					" to default value:%u.\n",
> +
> 	GRO_DEFAULT_ITEM_NUM_PER_FLOW);
> +		else
> +			gro_ports[res->port_id].param.max_item_per_flow
> =
> +				res->item_num_per_flow;
> +	}
> +}
> +
> +cmdline_parse_token_string_t cmd_gro_set_gro =
> +	TOKEN_STRING_INITIALIZER(struct cmd_gro_set_result,
> +				gro, "gro");
> +cmdline_parse_token_string_t cmd_gro_set_mode =
> +	TOKEN_STRING_INITIALIZER(struct cmd_gro_set_result,
> +				mode, "set");
> +cmdline_parse_token_num_t cmd_gro_set_flow_num =
> +	TOKEN_NUM_INITIALIZER(struct cmd_gro_set_result,
> +				flow_num, UINT16);
> +cmdline_parse_token_num_t cmd_gro_set_item_num_per_flow =
> +	TOKEN_NUM_INITIALIZER(struct cmd_gro_set_result,
> +				item_num_per_flow, UINT16);
> +cmdline_parse_token_num_t cmd_gro_set_portid =
> +	TOKEN_NUM_INITIALIZER(struct cmd_gro_set_result,
> +				port_id, UINT8);
> +
> +cmdline_parse_inst_t cmd_gro_set = {
> +	.f = cmd_gro_set_parsed,
> +	.data = NULL,
> +	.help_str = "gro set <max_flow_num> <max_item_num_per_flow>
> "
> +		"<port_id>: set max flow number and max packet number
> per-flow "
> +		"for GRO",
> +	.tokens = {
> +		(void *)&cmd_gro_set_gro,
> +		(void *)&cmd_gro_set_mode,
> +		(void *)&cmd_gro_set_flow_num,
> +		(void *)&cmd_gro_set_item_num_per_flow,
> +		(void *)&cmd_gro_set_portid,
> +		NULL,
> +	},
> +};
> +
>  /* *** ENABLE/DISABLE FLUSH ON RX STREAMS *** */
>  struct cmd_set_flush_rx {
>  	cmdline_fixed_string_t set;
> @@ -13732,6 +13855,8 @@ cmdline_parse_ctx_t main_ctx[] = {
>  	(cmdline_parse_inst_t *)&cmd_tso_show,
>  	(cmdline_parse_inst_t *)&cmd_tunnel_tso_set,
>  	(cmdline_parse_inst_t *)&cmd_tunnel_tso_show,
> +	(cmdline_parse_inst_t *)&cmd_enable_gro,
> +	(cmdline_parse_inst_t *)&cmd_gro_set,
>  	(cmdline_parse_inst_t *)&cmd_link_flow_control_set,
>  	(cmdline_parse_inst_t *)&cmd_link_flow_control_set_rx,
>  	(cmdline_parse_inst_t *)&cmd_link_flow_control_set_tx,
> diff --git a/app/test-pmd/config.c b/app/test-pmd/config.c
> index b0b340e..2a33a63 100644
> --- a/app/test-pmd/config.c
> +++ b/app/test-pmd/config.c
> @@ -71,6 +71,7 @@
>  #ifdef RTE_LIBRTE_BNXT_PMD
>  #include <rte_pmd_bnxt.h>
>  #endif
> +#include <rte_gro.h>
> 
>  #include "testpmd.h"
> 
> @@ -2414,6 +2415,42 @@ set_tx_pkt_segments(unsigned *seg_lengths,
> unsigned nb_segs)
>  	tx_pkt_nb_segs = (uint8_t) nb_segs;
>  }
> 
> +void
> +setup_gro(const char *mode, uint8_t port_id)
> +{
> +	if (!rte_eth_dev_is_valid_port(port_id)) {
> +		printf("invalid port id %u\n", port_id);
> +		return;
> +	}
> +	if (test_done == 0) {
> +		printf("Before enable/disable GRO,"
> +				" please stop forwarding first\n");
> +		return;
> +	}
> +	if (strcmp(mode, "on") == 0) {
> +		if (gro_ports[port_id].enable) {
> +			printf("port %u has enabled GRO\n", port_id);
> +			return;
> +		}
> +		gro_ports[port_id].enable = 1;
> +		gro_ports[port_id].param.desired_gro_types =
> GRO_TCP_IPV4;
> +		gro_ports[port_id].param.max_packet_size = UINT16_MAX;
> +
> +		if (gro_ports[port_id].param.max_flow_num == 0)
> +			gro_ports[port_id].param.max_flow_num =
> +				GRO_DEFAULT_FLOW_NUM;
> +		if (gro_ports[port_id].param.max_item_per_flow == 0)
> +			gro_ports[port_id].param.max_item_per_flow =
> +				GRO_DEFAULT_ITEM_NUM_PER_FLOW;
> +	} else {
> +		if (gro_ports[port_id].enable == 0) {
> +			printf("port %u has disabled GRO\n", port_id);
> +			return;
> +		}
> +		gro_ports[port_id].enable = 0;
> +	}
> +}
> +
>  char*
>  list_pkt_forwarding_modes(void)
>  {
> diff --git a/app/test-pmd/csumonly.c b/app/test-pmd/csumonly.c
> index 66fc9a0..430bd8b 100644
> --- a/app/test-pmd/csumonly.c
> +++ b/app/test-pmd/csumonly.c
> @@ -71,6 +71,7 @@
>  #include <rte_prefetch.h>
>  #include <rte_string_fns.h>
>  #include <rte_flow.h>
> +#include <rte_gro.h>
>  #include "testpmd.h"
> 
>  #define IP_DEFTTL  64   /* from RFC 1340. */
> @@ -658,6 +659,10 @@ pkt_burst_checksum_forward(struct fwd_stream
> *fs)
>  				 nb_pkt_per_burst);
>  	if (unlikely(nb_rx == 0))
>  		return;
> +	if (unlikely(gro_ports[fs->rx_port].enable))
> +		nb_rx = rte_gro_reassemble_burst(pkts_burst,
> +				nb_rx,
> +				gro_ports[fs->rx_port].param);
> 
>  #ifdef RTE_TEST_PMD_RECORD_BURST_STATS
>  	fs->rx_burst_stats.pkt_burst_spread[nb_rx]++;
> diff --git a/app/test-pmd/testpmd.c b/app/test-pmd/testpmd.c
> index b29328a..ed27c7a 100644
> --- a/app/test-pmd/testpmd.c
> +++ b/app/test-pmd/testpmd.c
> @@ -90,6 +90,7 @@
>  #ifdef RTE_LIBRTE_LATENCY_STATS
>  #include <rte_latencystats.h>
>  #endif
> +#include <rte_gro.h>
> 
>  #include "testpmd.h"
> 
> @@ -378,6 +379,8 @@ lcoreid_t bitrate_lcore_id;
>  uint8_t bitrate_enabled;
>  #endif
> 
> +struct gro_status gro_ports[RTE_MAX_ETHPORTS];
> +
>  /* Forward function declarations */
>  static void map_port_queue_stats_mapping_registers(uint8_t pi, struct
> rte_port *port);
>  static void check_all_ports_link_status(uint32_t port_mask);
> diff --git a/app/test-pmd/testpmd.h b/app/test-pmd/testpmd.h
> index 364502d..377d933 100644
> --- a/app/test-pmd/testpmd.h
> +++ b/app/test-pmd/testpmd.h
> @@ -34,6 +34,8 @@
>  #ifndef _TESTPMD_H_
>  #define _TESTPMD_H_
> 
> +#include <rte_gro.h>
> +
>  #define RTE_PORT_ALL            (~(portid_t)0x0)
> 
>  #define RTE_TEST_RX_DESC_MAX    2048
> @@ -428,6 +430,14 @@ extern struct ether_addr
> peer_eth_addrs[RTE_MAX_ETHPORTS];
>  extern uint32_t burst_tx_delay_time; /**< Burst tx delay time(us) for mac-
> retry. */
>  extern uint32_t burst_tx_retry_num;  /**< Burst tx retry number for mac-
> retry. */
> 
> +#define GRO_DEFAULT_FLOW_NUM 4
> +#define GRO_DEFAULT_ITEM_NUM_PER_FLOW DEF_PKT_BURST
> +struct gro_status {
> +	struct rte_gro_param param;
> +	uint8_t enable;
> +};
> +extern struct gro_status gro_ports[RTE_MAX_ETHPORTS];
> +
>  static inline unsigned int
>  lcore_num(void)
>  {
> @@ -626,6 +636,7 @@ void get_2tuple_filter(uint8_t port_id, uint16_t
> index);
>  void get_5tuple_filter(uint8_t port_id, uint16_t index);
>  int rx_queue_id_is_invalid(queueid_t rxq_id);
>  int tx_queue_id_is_invalid(queueid_t txq_id);
> +void setup_gro(const char *mode, uint8_t port_id);
> 
>  /* Functions to manage the set of filtered Multicast MAC addresses */
>  void mcast_addr_add(uint8_t port_id, struct ether_addr *mc_addr);
> diff --git a/doc/guides/testpmd_app_ug/testpmd_funcs.rst
> b/doc/guides/testpmd_app_ug/testpmd_funcs.rst
> index 2b9a1ea..528c833 100644
> --- a/doc/guides/testpmd_app_ug/testpmd_funcs.rst
> +++ b/doc/guides/testpmd_app_ug/testpmd_funcs.rst
> @@ -884,6 +884,40 @@ Display the status of TCP Segmentation Offload::
> 
>     testpmd> tso show (port_id)
> 
> +gro
> +~~~~~~~~
> +
> +Enable or disable GRO in ``csum`` forwarding engine::
> +
> +   testpmd> gro (on|off) (port_id)
> +
> +If enabled, the csum forwarding engine will perform GRO on the TCP/IPv4
> +packets received from the given port.
> +
> +If disabled, packets received from the given port won't be performed
> +GRO. By default, GRO is disabled for all ports.
> +
> +.. note::
> +
> +   When enable GRO for a port, TCP/IPv4 packets received from the port
> +   will be performed GRO. After GRO, the merged packets are multi-
> segments.
> +   But csum forwarding engine doesn't support to calculate TCP checksum
> +   for multi-segment packets in SW. So please select TCP HW checksum
> +   calculation for the port which GROed packets are transmitted to.
> +
> +gro set
> +~~~~~~~~
> +
> +Set max flow number and max packet number per-flow for GRO::
> +
> +   testpmd> gro set (max_flow_num) (max_item_num_per_flow) (port_id)
> +
> +The product of ``max_flow_num`` and ``max_item_num_per_flow`` is the
> max
> +number of packets a GRO table can store.
> +
> +If current packet number is greater than or equal to the max value, GRO
> +will stop processing incoming packets.
> +
>  mac_addr add
>  ~~~~~~~~~~~~
> 
> --
> 2.7.4



More information about the dev mailing list