[dpdk-dev] [PATCH v13] eventdev: simplify Rx adapter event vector config

Jayatheerthan, Jay jay.jayatheerthan at intel.com
Wed Aug 18 10:22:49 CEST 2021


HI Harish,
Could you review this patch ?

-Jay


> -----Original Message-----
> From: pbhagavatula at marvell.com <pbhagavatula at marvell.com>
> Sent: Wednesday, August 18, 2021 12:27 PM
> To: jerinj at marvell.com; Ray Kinsella <mdr at ashroe.eu>; Pavan Nikhilesh <pbhagavatula at marvell.com>; Shijith Thotton
> <sthotton at marvell.com>; Jayatheerthan, Jay <jay.jayatheerthan at intel.com>
> Cc: dev at dpdk.org
> Subject: [dpdk-dev] [PATCH v13] eventdev: simplify Rx adapter event vector config
> 
> From: Pavan Nikhilesh <pbhagavatula at marvell.com>
> 
> Include vector configuration into the structure
> ``rte_event_eth_rx_adapter_queue_conf`` that is used to configure
> Rx adapter ethernet device Rx queue parameters.
> This simplifies event vector configuration as it avoids splitting
> configuration per Rx queue.
> 
> Signed-off-by: Pavan Nikhilesh <pbhagavatula at marvell.com>
> Acked-by: Jay Jayatheerthan <jay.jayatheerthan at intel.com>
> ---
>  v13 Changes:
>  - Fix cnxk driver compilation.
>  v12 Changes:
>  - Remove deprication notice.
>  - Remove unnecessary change Id.
> 
>  app/test-eventdev/test_pipeline_common.c |  16 +-
>  doc/guides/rel_notes/deprecation.rst     |   9 --
>  drivers/event/cnxk/cn10k_eventdev.c      |  77 ----------
>  drivers/event/cnxk/cnxk_eventdev_adptr.c |  41 ++++++
>  lib/eventdev/eventdev_pmd.h              |  29 ----
>  lib/eventdev/rte_event_eth_rx_adapter.c  | 179 ++++++++---------------
>  lib/eventdev/rte_event_eth_rx_adapter.h  |  30 ----
>  lib/eventdev/version.map                 |   1 -
>  8 files changed, 104 insertions(+), 278 deletions(-)
> 
> diff --git a/app/test-eventdev/test_pipeline_common.c b/app/test-eventdev/test_pipeline_common.c
> index 6ee530d4cd..2697547641 100644
> --- a/app/test-eventdev/test_pipeline_common.c
> +++ b/app/test-eventdev/test_pipeline_common.c
> @@ -332,7 +332,6 @@ pipeline_event_rx_adapter_setup(struct evt_options *opt, uint8_t stride,
>  	uint16_t prod;
>  	struct rte_mempool *vector_pool = NULL;
>  	struct rte_event_eth_rx_adapter_queue_conf queue_conf;
> -	struct rte_event_eth_rx_adapter_event_vector_config vec_conf;
> 
>  	memset(&queue_conf, 0,
>  			sizeof(struct rte_event_eth_rx_adapter_queue_conf));
> @@ -398,8 +397,12 @@ pipeline_event_rx_adapter_setup(struct evt_options *opt, uint8_t stride,
>  			}
> 
>  			if (cap & RTE_EVENT_ETH_RX_ADAPTER_CAP_EVENT_VECTOR) {
> +				queue_conf.vector_sz = opt->vector_size;
> +				queue_conf.vector_timeout_ns =
> +					opt->vector_tmo_nsec;
>  				queue_conf.rx_queue_flags |=
>  				RTE_EVENT_ETH_RX_ADAPTER_QUEUE_EVENT_VECTOR;
> +				queue_conf.vector_mp = vector_pool;
>  			} else {
>  				evt_err("Rx adapter doesn't support event vector");
>  				return -EINVAL;
> @@ -419,17 +422,6 @@ pipeline_event_rx_adapter_setup(struct evt_options *opt, uint8_t stride,
>  			return ret;
>  		}
> 
> -		if (opt->ena_vector) {
> -			vec_conf.vector_sz = opt->vector_size;
> -			vec_conf.vector_timeout_ns = opt->vector_tmo_nsec;
> -			vec_conf.vector_mp = vector_pool;
> -			if (rte_event_eth_rx_adapter_queue_event_vector_config(
> -				    prod, prod, -1, &vec_conf) < 0) {
> -				evt_err("Failed to configure event vectorization for Rx adapter");
> -				return -EINVAL;
> -			}
> -		}
> -
>  		if (!(cap & RTE_EVENT_ETH_RX_ADAPTER_CAP_INTERNAL_PORT)) {
>  			uint32_t service_id = -1U;
> 
> diff --git a/doc/guides/rel_notes/deprecation.rst b/doc/guides/rel_notes/deprecation.rst
> index 76a4abfd6b..2c37d7222c 100644
> --- a/doc/guides/rel_notes/deprecation.rst
> +++ b/doc/guides/rel_notes/deprecation.rst
> @@ -257,15 +257,6 @@ Deprecation Notices
>    An 8-byte reserved field will be added to the structure ``rte_event_timer`` to
>    support future extensions.
> 
> -* eventdev: The structure ``rte_event_eth_rx_adapter_queue_conf`` will be
> -  extended to include ``rte_event_eth_rx_adapter_event_vector_config`` elements
> -  and the function ``rte_event_eth_rx_adapter_queue_event_vector_config`` will
> -  be removed in DPDK 21.11.
> -
> -  An application can enable event vectorization by passing the desired vector
> -  values to the function ``rte_event_eth_rx_adapter_queue_add`` using
> -  the structure ``rte_event_eth_rx_adapter_queue_add``.
> -
>  * eventdev: Reserved bytes of ``rte_event_crypto_request`` is a space holder
>    for ``response_info``. Both should be decoupled for better clarity.
>    New space for ``response_info`` can be made by changing
> diff --git a/drivers/event/cnxk/cn10k_eventdev.c b/drivers/event/cnxk/cn10k_eventdev.c
> index 6f37c5bd23..160192bb27 100644
> --- a/drivers/event/cnxk/cn10k_eventdev.c
> +++ b/drivers/event/cnxk/cn10k_eventdev.c
> @@ -696,81 +696,6 @@ cn10k_sso_rx_adapter_vector_limits(
>  	return 0;
>  }
> 
> -static int
> -cnxk_sso_rx_adapter_vwqe_enable(struct cnxk_eth_dev *cnxk_eth_dev,
> -				uint16_t port_id, uint16_t rq_id, uint16_t sz,
> -				uint64_t tmo_ns, struct rte_mempool *vmp)
> -{
> -	struct roc_nix_rq *rq;
> -
> -	rq = &cnxk_eth_dev->rqs[rq_id];
> -
> -	if (!rq->sso_ena)
> -		return -EINVAL;
> -	if (rq->flow_tag_width == 0)
> -		return -EINVAL;
> -
> -	rq->vwqe_ena = 1;
> -	rq->vwqe_first_skip = 0;
> -	rq->vwqe_aura_handle = roc_npa_aura_handle_to_aura(vmp->pool_id);
> -	rq->vwqe_max_sz_exp = rte_log2_u32(sz);
> -	rq->vwqe_wait_tmo =
> -		tmo_ns /
> -		((roc_nix_get_vwqe_interval(&cnxk_eth_dev->nix) + 1) * 100);
> -	rq->tag_mask = (port_id & 0xF) << 20;
> -	rq->tag_mask |=
> -		(((port_id >> 4) & 0xF) | (RTE_EVENT_TYPE_ETHDEV_VECTOR << 4))
> -		<< 24;
> -
> -	return roc_nix_rq_modify(&cnxk_eth_dev->nix, rq, 0);
> -}
> -
> -static int
> -cn10k_sso_rx_adapter_vector_config(
> -	const struct rte_eventdev *event_dev, const struct rte_eth_dev *eth_dev,
> -	int32_t rx_queue_id,
> -	const struct rte_event_eth_rx_adapter_event_vector_config *config)
> -{
> -	struct cnxk_eth_dev *cnxk_eth_dev;
> -	struct cnxk_sso_evdev *dev;
> -	int i, rc;
> -
> -	rc = strncmp(eth_dev->device->driver->name, "net_cn10k", 8);
> -	if (rc)
> -		return -EINVAL;
> -
> -	dev = cnxk_sso_pmd_priv(event_dev);
> -	cnxk_eth_dev = cnxk_eth_pmd_priv(eth_dev);
> -	if (rx_queue_id < 0) {
> -		for (i = 0; i < eth_dev->data->nb_rx_queues; i++) {
> -			cnxk_sso_updt_xae_cnt(dev, config->vector_mp,
> -					      RTE_EVENT_TYPE_ETHDEV_VECTOR);
> -			rc = cnxk_sso_xae_reconfigure(
> -				(struct rte_eventdev *)(uintptr_t)event_dev);
> -			rc = cnxk_sso_rx_adapter_vwqe_enable(
> -				cnxk_eth_dev, eth_dev->data->port_id, i,
> -				config->vector_sz, config->vector_timeout_ns,
> -				config->vector_mp);
> -			if (rc)
> -				return -EINVAL;
> -		}
> -	} else {
> -
> -		cnxk_sso_updt_xae_cnt(dev, config->vector_mp,
> -				      RTE_EVENT_TYPE_ETHDEV_VECTOR);
> -		rc = cnxk_sso_xae_reconfigure(
> -			(struct rte_eventdev *)(uintptr_t)event_dev);
> -		rc = cnxk_sso_rx_adapter_vwqe_enable(
> -			cnxk_eth_dev, eth_dev->data->port_id, rx_queue_id,
> -			config->vector_sz, config->vector_timeout_ns,
> -			config->vector_mp);
> -		if (rc)
> -			return -EINVAL;
> -	}
> -
> -	return 0;
> -}
> -
>  static int
>  cn10k_sso_tx_adapter_caps_get(const struct rte_eventdev *dev,
>  			      const struct rte_eth_dev *eth_dev, uint32_t *caps)
> @@ -841,8 +766,6 @@ static struct rte_eventdev_ops cn10k_sso_dev_ops = {
>  	.eth_rx_adapter_stop = cnxk_sso_rx_adapter_stop,
> 
>  	.eth_rx_adapter_vector_limits_get = cn10k_sso_rx_adapter_vector_limits,
> -	.eth_rx_adapter_event_vector_config =
> -		cn10k_sso_rx_adapter_vector_config,
> 
>  	.eth_tx_adapter_caps_get = cn10k_sso_tx_adapter_caps_get,
>  	.eth_tx_adapter_queue_add = cn10k_sso_tx_adapter_queue_add,
> diff --git a/drivers/event/cnxk/cnxk_eventdev_adptr.c b/drivers/event/cnxk/cnxk_eventdev_adptr.c
> index baf2f2aa6b..80f5602286 100644
> --- a/drivers/event/cnxk/cnxk_eventdev_adptr.c
> +++ b/drivers/event/cnxk/cnxk_eventdev_adptr.c
> @@ -156,6 +156,35 @@ cnxk_sso_rxq_disable(struct cnxk_eth_dev *cnxk_eth_dev, uint16_t rq_id)
>  	return roc_nix_rq_modify(&cnxk_eth_dev->nix, rq, 0);
>  }
> 
> +static int
> +cnxk_sso_rx_adapter_vwqe_enable(struct cnxk_eth_dev *cnxk_eth_dev,
> +				uint16_t port_id, uint16_t rq_id, uint16_t sz,
> +				uint64_t tmo_ns, struct rte_mempool *vmp)
> +{
> +	struct roc_nix_rq *rq;
> +
> +	rq = &cnxk_eth_dev->rqs[rq_id];
> +
> +	if (!rq->sso_ena)
> +		return -EINVAL;
> +	if (rq->flow_tag_width == 0)
> +		return -EINVAL;
> +
> +	rq->vwqe_ena = 1;
> +	rq->vwqe_first_skip = 0;
> +	rq->vwqe_aura_handle = roc_npa_aura_handle_to_aura(vmp->pool_id);
> +	rq->vwqe_max_sz_exp = rte_log2_u32(sz);
> +	rq->vwqe_wait_tmo =
> +		tmo_ns /
> +		((roc_nix_get_vwqe_interval(&cnxk_eth_dev->nix) + 1) * 100);
> +	rq->tag_mask = (port_id & 0xF) << 20;
> +	rq->tag_mask |=
> +		(((port_id >> 4) & 0xF) | (RTE_EVENT_TYPE_ETHDEV_VECTOR << 4))
> +		<< 24;
> +
> +	return roc_nix_rq_modify(&cnxk_eth_dev->nix, rq, 0);
> +}
> +
>  int
>  cnxk_sso_rx_adapter_queue_add(
>  	const struct rte_eventdev *event_dev, const struct rte_eth_dev *eth_dev,
> @@ -183,6 +212,18 @@ cnxk_sso_rx_adapter_queue_add(
>  			&queue_conf->ev,
>  			!!(queue_conf->rx_queue_flags &
>  			   RTE_EVENT_ETH_RX_ADAPTER_CAP_OVERRIDE_FLOW_ID));
> +		if (queue_conf->rx_queue_flags &
> +		    RTE_EVENT_ETH_RX_ADAPTER_QUEUE_EVENT_VECTOR) {
> +			cnxk_sso_updt_xae_cnt(dev, queue_conf->vector_mp,
> +					      RTE_EVENT_TYPE_ETHDEV_VECTOR);
> +			rc |= cnxk_sso_xae_reconfigure(
> +				(struct rte_eventdev *)(uintptr_t)event_dev);
> +			rc |= cnxk_sso_rx_adapter_vwqe_enable(
> +				cnxk_eth_dev, port, rx_queue_id,
> +				queue_conf->vector_sz,
> +				queue_conf->vector_timeout_ns,
> +				queue_conf->vector_mp);
> +		}
>  		rox_nix_fc_npa_bp_cfg(&cnxk_eth_dev->nix,
>  				      rxq_sp->qconf.mp->pool_id, true,
>  				      dev->force_ena_bp);
> diff --git a/lib/eventdev/eventdev_pmd.h b/lib/eventdev/eventdev_pmd.h
> index 0f724ac85d..63b3bc4b51 100644
> --- a/lib/eventdev/eventdev_pmd.h
> +++ b/lib/eventdev/eventdev_pmd.h
> @@ -667,32 +667,6 @@ typedef int (*eventdev_eth_rx_adapter_vector_limits_get_t)(
>  	const struct rte_eventdev *dev, const struct rte_eth_dev *eth_dev,
>  	struct rte_event_eth_rx_adapter_vector_limits *limits);
> 
> -struct rte_event_eth_rx_adapter_event_vector_config;
> -/**
> - * Enable event vector on an given Rx queue of a ethernet devices belonging to
> - * the Rx adapter.
> - *
> - * @param dev
> - *   Event device pointer
> - *
> - * @param eth_dev
> - *   Ethernet device pointer
> - *
> - * @param rx_queue_id
> - *   The Rx queue identifier
> - *
> - * @param config
> - *   Pointer to the event vector configuration structure.
> - *
> - * @return
> - *   - 0: Success.
> - *   - <0: Error code returned by the driver function.
> - */
> -typedef int (*eventdev_eth_rx_adapter_event_vector_config_t)(
> -	const struct rte_eventdev *dev, const struct rte_eth_dev *eth_dev,
> -	int32_t rx_queue_id,
> -	const struct rte_event_eth_rx_adapter_event_vector_config *config);
> -
>  typedef uint32_t rte_event_pmd_selftest_seqn_t;
>  extern int rte_event_pmd_selftest_seqn_dynfield_offset;
> 
> @@ -1118,9 +1092,6 @@ struct rte_eventdev_ops {
>  	eventdev_eth_rx_adapter_vector_limits_get_t
>  		eth_rx_adapter_vector_limits_get;
>  	/**< Get event vector limits for the Rx adapter */
> -	eventdev_eth_rx_adapter_event_vector_config_t
> -		eth_rx_adapter_event_vector_config;
> -	/**< Configure Rx adapter with event vector */
> 
>  	eventdev_timer_adapter_caps_get_t timer_adapter_caps_get;
>  	/**< Get timer adapter capabilities */
> diff --git a/lib/eventdev/rte_event_eth_rx_adapter.c b/lib/eventdev/rte_event_eth_rx_adapter.c
> index 13dfb28401..2b2dd688fc 100644
> --- a/lib/eventdev/rte_event_eth_rx_adapter.c
> +++ b/lib/eventdev/rte_event_eth_rx_adapter.c
> @@ -1895,6 +1895,24 @@ rxa_add_queue(struct rte_event_eth_rx_adapter *rx_adapter,
>  	} else
>  		qi_ev->flow_id = 0;
> 
> +	if (conf->rx_queue_flags &
> +	    RTE_EVENT_ETH_RX_ADAPTER_QUEUE_EVENT_VECTOR) {
> +		queue_info->ena_vector = 1;
> +		qi_ev->event_type = RTE_EVENT_TYPE_ETH_RX_ADAPTER_VECTOR;
> +		rxa_set_vector_data(queue_info, conf->vector_sz,
> +				    conf->vector_timeout_ns, conf->vector_mp,
> +				    rx_queue_id, dev_info->dev->data->port_id);
> +		rx_adapter->ena_vector = 1;
> +		rx_adapter->vector_tmo_ticks =
> +			rx_adapter->vector_tmo_ticks ?
> +				      RTE_MIN(queue_info->vector_data
> +							.vector_timeout_ticks >>
> +						1,
> +					rx_adapter->vector_tmo_ticks) :
> +				queue_info->vector_data.vector_timeout_ticks >>
> +					1;
> +	}
> +
>  	rxa_update_queue(rx_adapter, dev_info, rx_queue_id, 1);
>  	if (rxa_polled_queue(dev_info, rx_queue_id)) {
>  		rx_adapter->num_rx_polled += !pollq;
> @@ -1920,42 +1938,6 @@ rxa_add_queue(struct rte_event_eth_rx_adapter *rx_adapter,
>  	}
>  }
> 
> -static void
> -rxa_sw_event_vector_configure(
> -	struct rte_event_eth_rx_adapter *rx_adapter, uint16_t eth_dev_id,
> -	int rx_queue_id,
> -	const struct rte_event_eth_rx_adapter_event_vector_config *config)
> -{
> -	struct eth_device_info *dev_info = &rx_adapter->eth_devices[eth_dev_id];
> -	struct eth_rx_queue_info *queue_info;
> -	struct rte_event *qi_ev;
> -
> -	if (rx_queue_id == -1) {
> -		uint16_t nb_rx_queues;
> -		uint16_t i;
> -
> -		nb_rx_queues = dev_info->dev->data->nb_rx_queues;
> -		for (i = 0; i < nb_rx_queues; i++)
> -			rxa_sw_event_vector_configure(rx_adapter, eth_dev_id, i,
> -						      config);
> -		return;
> -	}
> -
> -	queue_info = &dev_info->rx_queue[rx_queue_id];
> -	qi_ev = (struct rte_event *)&queue_info->event;
> -	queue_info->ena_vector = 1;
> -	qi_ev->event_type = RTE_EVENT_TYPE_ETH_RX_ADAPTER_VECTOR;
> -	rxa_set_vector_data(queue_info, config->vector_sz,
> -			    config->vector_timeout_ns, config->vector_mp,
> -			    rx_queue_id, dev_info->dev->data->port_id);
> -	rx_adapter->ena_vector = 1;
> -	rx_adapter->vector_tmo_ticks =
> -		rx_adapter->vector_tmo_ticks ?
> -			      RTE_MIN(config->vector_timeout_ns >> 1,
> -				rx_adapter->vector_tmo_ticks) :
> -			      config->vector_timeout_ns >> 1;
> -}
> -
>  static int rxa_sw_add(struct rte_event_eth_rx_adapter *rx_adapter,
>  		uint16_t eth_dev_id,
>  		int rx_queue_id,
> @@ -2270,6 +2252,7 @@ rte_event_eth_rx_adapter_queue_add(uint8_t id,
>  	struct rte_event_eth_rx_adapter *rx_adapter;
>  	struct rte_eventdev *dev;
>  	struct eth_device_info *dev_info;
> +	struct rte_event_eth_rx_adapter_vector_limits limits;
> 
>  	RTE_EVENT_ETH_RX_ADAPTER_ID_VALID_OR_ERR_RET(id, -EINVAL);
>  	RTE_ETH_VALID_PORTID_OR_ERR_RET(eth_dev_id, -EINVAL);
> @@ -2297,13 +2280,46 @@ rte_event_eth_rx_adapter_queue_add(uint8_t id,
>  		return -EINVAL;
>  	}
> 
> -	if ((cap & RTE_EVENT_ETH_RX_ADAPTER_CAP_EVENT_VECTOR) == 0 &&
> -	    (queue_conf->rx_queue_flags &
> -	     RTE_EVENT_ETH_RX_ADAPTER_QUEUE_EVENT_VECTOR)) {
> -		RTE_EDEV_LOG_ERR("Event vectorization is not supported,"
> -				 " eth port: %" PRIu16 " adapter id: %" PRIu8,
> -				 eth_dev_id, id);
> -		return -EINVAL;
> +	if (queue_conf->rx_queue_flags &
> +	    RTE_EVENT_ETH_RX_ADAPTER_QUEUE_EVENT_VECTOR) {
> +
> +		if ((cap & RTE_EVENT_ETH_RX_ADAPTER_CAP_EVENT_VECTOR) == 0) {
> +			RTE_EDEV_LOG_ERR("Event vectorization is not supported,"
> +					 " eth port: %" PRIu16
> +					 " adapter id: %" PRIu8,
> +					 eth_dev_id, id);
> +			return -EINVAL;
> +		}
> +
> +		ret = rte_event_eth_rx_adapter_vector_limits_get(
> +			rx_adapter->eventdev_id, eth_dev_id, &limits);
> +		if (ret < 0) {
> +			RTE_EDEV_LOG_ERR("Failed to get event device vector limits,"
> +					 " eth port: %" PRIu16
> +					 " adapter id: %" PRIu8,
> +					 eth_dev_id, id);
> +			return -EINVAL;
> +		}
> +		if (queue_conf->vector_sz < limits.min_sz ||
> +		    queue_conf->vector_sz > limits.max_sz ||
> +		    queue_conf->vector_timeout_ns < limits.min_timeout_ns ||
> +		    queue_conf->vector_timeout_ns > limits.max_timeout_ns ||
> +		    queue_conf->vector_mp == NULL) {
> +			RTE_EDEV_LOG_ERR("Invalid event vector configuration,"
> +					 " eth port: %" PRIu16
> +					 " adapter id: %" PRIu8,
> +					 eth_dev_id, id);
> +			return -EINVAL;
> +		}
> +		if (queue_conf->vector_mp->elt_size <
> +		    (sizeof(struct rte_event_vector) +
> +		     (sizeof(uintptr_t) * queue_conf->vector_sz))) {
> +			RTE_EDEV_LOG_ERR("Invalid event vector configuration,"
> +					 " eth port: %" PRIu16
> +					 " adapter id: %" PRIu8,
> +					 eth_dev_id, id);
> +			return -EINVAL;
> +		}
>  	}
> 
>  	if ((cap & RTE_EVENT_ETH_RX_ADAPTER_CAP_MULTI_EVENTQ) == 0 &&
> @@ -2499,83 +2515,6 @@ rte_event_eth_rx_adapter_queue_del(uint8_t id, uint16_t eth_dev_id,
>  	return ret;
>  }
> 
> -int
> -rte_event_eth_rx_adapter_queue_event_vector_config(
> -	uint8_t id, uint16_t eth_dev_id, int32_t rx_queue_id,
> -	struct rte_event_eth_rx_adapter_event_vector_config *config)
> -{
> -	struct rte_event_eth_rx_adapter_vector_limits limits;
> -	struct rte_event_eth_rx_adapter *rx_adapter;
> -	struct rte_eventdev *dev;
> -	uint32_t cap;
> -	int ret;
> -
> -	RTE_EVENT_ETH_RX_ADAPTER_ID_VALID_OR_ERR_RET(id, -EINVAL);
> -	RTE_ETH_VALID_PORTID_OR_ERR_RET(eth_dev_id, -EINVAL);
> -
> -	rx_adapter = rxa_id_to_adapter(id);
> -	if ((rx_adapter == NULL) || (config == NULL))
> -		return -EINVAL;
> -
> -	dev = &rte_eventdevs[rx_adapter->eventdev_id];
> -	ret = rte_event_eth_rx_adapter_caps_get(rx_adapter->eventdev_id,
> -						eth_dev_id, &cap);
> -	if (ret) {
> -		RTE_EDEV_LOG_ERR("Failed to get adapter caps edev %" PRIu8
> -				 "eth port %" PRIu16,
> -				 id, eth_dev_id);
> -		return ret;
> -	}
> -
> -	if (!(cap & RTE_EVENT_ETH_RX_ADAPTER_CAP_EVENT_VECTOR)) {
> -		RTE_EDEV_LOG_ERR("Event vectorization is not supported,"
> -				 " eth port: %" PRIu16 " adapter id: %" PRIu8,
> -				 eth_dev_id, id);
> -		return -EINVAL;
> -	}
> -
> -	ret = rte_event_eth_rx_adapter_vector_limits_get(
> -		rx_adapter->eventdev_id, eth_dev_id, &limits);
> -	if (ret) {
> -		RTE_EDEV_LOG_ERR("Failed to get vector limits edev %" PRIu8
> -				 "eth port %" PRIu16,
> -				 rx_adapter->eventdev_id, eth_dev_id);
> -		return ret;
> -	}
> -
> -	if (config->vector_sz < limits.min_sz ||
> -	    config->vector_sz > limits.max_sz ||
> -	    config->vector_timeout_ns < limits.min_timeout_ns ||
> -	    config->vector_timeout_ns > limits.max_timeout_ns ||
> -	    config->vector_mp == NULL) {
> -		RTE_EDEV_LOG_ERR("Invalid event vector configuration,"
> -				 " eth port: %" PRIu16 " adapter id: %" PRIu8,
> -				 eth_dev_id, id);
> -		return -EINVAL;
> -	}
> -	if (config->vector_mp->elt_size <
> -	    (sizeof(struct rte_event_vector) +
> -	     (sizeof(uintptr_t) * config->vector_sz))) {
> -		RTE_EDEV_LOG_ERR("Invalid event vector configuration,"
> -				 " eth port: %" PRIu16 " adapter id: %" PRIu8,
> -				 eth_dev_id, id);
> -		return -EINVAL;
> -	}
> -
> -	if (cap & RTE_EVENT_ETH_RX_ADAPTER_CAP_INTERNAL_PORT) {
> -		RTE_FUNC_PTR_OR_ERR_RET(
> -			*dev->dev_ops->eth_rx_adapter_event_vector_config,
> -			-ENOTSUP);
> -		ret = dev->dev_ops->eth_rx_adapter_event_vector_config(
> -			dev, &rte_eth_devices[eth_dev_id], rx_queue_id, config);
> -	} else {
> -		rxa_sw_event_vector_configure(rx_adapter, eth_dev_id,
> -					      rx_queue_id, config);
> -	}
> -
> -	return ret;
> -}
> -
>  int
>  rte_event_eth_rx_adapter_vector_limits_get(
>  	uint8_t dev_id, uint16_t eth_port_id,
> diff --git a/lib/eventdev/rte_event_eth_rx_adapter.h b/lib/eventdev/rte_event_eth_rx_adapter.h
> index 182dd2e5dd..3f8b362295 100644
> --- a/lib/eventdev/rte_event_eth_rx_adapter.h
> +++ b/lib/eventdev/rte_event_eth_rx_adapter.h
> @@ -171,9 +171,6 @@ struct rte_event_eth_rx_adapter_queue_conf {
>  	 * The event adapter sets ev.event_type to RTE_EVENT_TYPE_ETHDEV in the
>  	 * enqueued event.
>  	 */
> -};
> -
> -struct rte_event_eth_rx_adapter_event_vector_config {
>  	uint16_t vector_sz;
>  	/**<
>  	 * Indicates the maximum number for mbufs to combine and form a vector.
> @@ -548,33 +545,6 @@ int rte_event_eth_rx_adapter_vector_limits_get(
>  	uint8_t dev_id, uint16_t eth_port_id,
>  	struct rte_event_eth_rx_adapter_vector_limits *limits);
> 
> -/**
> - * Configure event vectorization for a given ethernet device queue, that has
> - * been added to a event eth Rx adapter.
> - *
> - * @param id
> - *  The identifier of the ethernet Rx event adapter.
> - *
> - * @param eth_dev_id
> - *  The identifier of the ethernet device.
> - *
> - * @param rx_queue_id
> - *  Ethernet device receive queue index.
> - *  If rx_queue_id is -1, then all Rx queues configured for the ethernet device
> - *  are configured with event vectorization.
> - *
> - * @param config
> - *  Event vector configuration structure.
> - *
> - * @return
> - *  - 0: Success, Receive queue configured correctly.
> - *  - <0: Error code on failure.
> - */
> -__rte_experimental
> -int rte_event_eth_rx_adapter_queue_event_vector_config(
> -	uint8_t id, uint16_t eth_dev_id, int32_t rx_queue_id,
> -	struct rte_event_eth_rx_adapter_event_vector_config *config);
> -
>  #ifdef __cplusplus
>  }
>  #endif
> diff --git a/lib/eventdev/version.map b/lib/eventdev/version.map
> index 88625621ec..cd86d2d908 100644
> --- a/lib/eventdev/version.map
> +++ b/lib/eventdev/version.map
> @@ -142,7 +142,6 @@ EXPERIMENTAL {
>  	#added in 21.05
>  	rte_event_vector_pool_create;
>  	rte_event_eth_rx_adapter_vector_limits_get;
> -	rte_event_eth_rx_adapter_queue_event_vector_config;
>  	__rte_eventdev_trace_crypto_adapter_enqueue;
>  };
> 
> --
> 2.17.1



More information about the dev mailing list