[dpdk-dev] [PATCH v1 3/5] bpf: introduce basic RX/TX BPF filters

Jerin Jacob jerin.jacob at caviumnetworks.com
Tue Mar 13 14:39:42 CET 2018


-----Original Message-----
> Date: Fri, 9 Mar 2018 16:42:03 +0000
> From: Konstantin Ananyev <konstantin.ananyev at intel.com>
> To: dev at dpdk.org
> CC: Konstantin Ananyev <konstantin.ananyev at intel.com>
> Subject: [dpdk-dev] [PATCH v1 3/5] bpf: introduce basic RX/TX BPF filters
> X-Mailer: git-send-email 1.7.0.7
> 
> Introduce API to install BPF based filters on ethdev RX/TX path.
> Current implementation is pure SW one, based on ethdev RX/TX
> callback mechanism.
> 
> Signed-off-by: Konstantin Ananyev <konstantin.ananyev at intel.com>
> ---
>  lib/librte_bpf/Makefile            |   2 +
>  lib/librte_bpf/bpf_pkt.c           | 524 +++++++++++++++++++++++++++++++++++++
>  lib/librte_bpf/rte_bpf_ethdev.h    |  50 ++++
>  lib/librte_bpf/rte_bpf_version.map |   4 +
>  4 files changed, 580 insertions(+)
>  create mode 100644 lib/librte_bpf/bpf_pkt.c
>  create mode 100644 lib/librte_bpf/rte_bpf_ethdev.h
> 
> diff --git a/lib/librte_bpf/Makefile b/lib/librte_bpf/Makefile
> +
> +/*
> + * information about all installed BPF rx/tx callbacks
> + */
> +
> +struct bpf_eth_cbi {
> +	uint32_t use;    /*usage counter */
> +	void *cb;        /* callback handle */
> +	struct rte_bpf *bpf;
> +	struct rte_bpf_jit jit;
> +} __rte_cache_aligned;
> +
> +/*
> + * Odd number means that callback is used by datapath.
> + * Even number means that callback is not used by datapath.
> + */
> +#define BPF_ETH_CBI_INUSE  1
> +
> +static struct bpf_eth_cbi rx_cbi[RTE_MAX_ETHPORTS][RTE_MAX_QUEUES_PER_PORT];
> +static struct bpf_eth_cbi tx_cbi[RTE_MAX_ETHPORTS][RTE_MAX_QUEUES_PER_PORT];

How about allocating this memory from huge page?

> +
> +/*
> + * Marks given callback as used by datapath.
> + */
> +static __rte_always_inline void
> +bpf_eth_cbi_inuse(struct bpf_eth_cbi *cbi)
> +{
> +	cbi->use++;
> +	/* make sure no store/load reordering could happen */
> +	rte_smp_mb();

This is an full barrier on non x86. How about a light version of this
logic? See below.

> +}
> +
> +/*
> + * Marks given callback list as not used by datapath.
> + */
> +static __rte_always_inline void
> +bpf_eth_cbi_unuse(struct bpf_eth_cbi *cbi)
> +{
> +	/* make sure all previous loads are completed */
> +	rte_smp_rmb();
> +	cbi->use++;
> +}
> +
> +/*
> + * Waits till datapath finished using given callback.
> + */
> +static void
> +bpf_eth_cbi_wait(const struct bpf_eth_cbi *cbi)
> +{
> +	uint32_t nuse, puse;
> +
> +	/* make sure all previous loads and stores are completed */
> +	rte_smp_mb();
> +

Read conjunction with below change

#if 0
> +	puse = cbi->use;
> +
> +	/* in use, busy wait till current RX/TX iteration is finished */
> +	if ((puse & BPF_ETH_CBI_INUSE) != 0) {
> +		do {
> +			rte_pause();
> +			rte_compiler_barrier();
> +			nuse = cbi->use;
> +		} while (nuse == puse);
> +	}
#else
	cbi->cb = NULL;
	while (likely(cb->done != 1)) {
		rte_pause();
		rte_smb_rmb();
	}

or any other logic using flag to wait until callback completes.
#endif

> +}
> +
> +
> +/*
> + * RX/TX callbacks for raw data bpf.
> + */
> +
> +static uint16_t
> +bpf_rx_callback_vm(__rte_unused uint16_t port, __rte_unused uint16_t queue,
> +	struct rte_mbuf *pkt[], uint16_t nb_pkts,
> +	__rte_unused uint16_t max_pkts, void *user_param)
> +{
> +	struct bpf_eth_cbi *cbi;
> +	uint16_t rc;
> +
> +	cbi = user_param;
> +

Read conjunction with above change

#if 0
> +	bpf_eth_cbi_inuse(cbi);
> +	rc = (cbi->cb != NULL) ?
> +		pkt_filter_vm(cbi->bpf, pkt, nb_pkts, 1) :
> +		nb_pkts;
> +	bpf_eth_cbi_unuse(cbi);
#else
	if (likely(cbi->cb != NULL))
		return pkt_filter_vm(cbi->bpf, pkt, nb_pkts, 1) :
	else {
		cbi->done = 1;
		rte_smb_wmb();
		return nb_pkts;
	}
#endif

> +	return rc;
> +}
> +


More information about the dev mailing list