[PATCH v4] mempool: fix mempool cache flushing algorithm
Bruce Richardson
bruce.richardson at intel.com
Thu Apr 7 11:14:28 CEST 2022
On Thu, Apr 07, 2022 at 11:04:53AM +0200, Morten Brørup wrote:
> > From: Morten Brørup [mailto:mb at smartsharesystems.com]
> > Sent: Wednesday, 2 February 2022 11.34
> >
> > This patch fixes the rte_mempool_do_generic_put() caching algorithm,
> > which was fundamentally wrong, causing multiple performance issues when
> > flushing.
> >
>
> [...]
>
> Olivier,
>
> Will you please consider this patch [1] and the other one [2].
>
> The primary bug here is this: When a mempool cache becomes full (i.e. exceeds the "flush threshold"), and is flushed to the backing ring, it is still full afterwards; but it should be empty afterwards. It is not flushed entirely, only the elements exceeding "size" are flushed.
>
I don't believe it should be flushed entirely, there should always be some
elements left so that even after flush we can still allocate an additional
burst. We want to avoid the situation where a flush of all elements is
immediately followed by a refill of new elements. However, we can flush to
maybe size/2, and improve things. In short, this not emptying is by design
rather than a bug, though we can look to tweak the behaviour.
> E.g. pipelined applications having ingress threads and egress threads running on different lcores are affected by this bug.
>
If we are looking at improvements for pipelined applications, I think a
bigger win would be to change the default mempool from ring-based to
stack-based. For apps using a run-to-completion model, they should run out
of cache and should therefore be largely unaffected by such a change.
> I don't think the real performance impact is very big, but these algorithm level bugs really annoy me.
>
> I'm still wondering how the patch introducing the mempool cache flush threshold could pass internal code review with so many bugs.
>
> [1] https://patchwork.dpdk.org/project/dpdk/patch/20220202103354.79832-1-mb@smartsharesystems.com/
> [2] https://patchwork.dpdk.org/project/dpdk/patch/20220202081426.77975-1-mb@smartsharesystems.com/
>
> -Morten
>
> > Signed-off-by: Morten Brørup <mb at smartsharesystems.com>
> > ---
> > lib/mempool/rte_mempool.h | 34 ++++++++++++++++++++++------------
> > 1 file changed, 22 insertions(+), 12 deletions(-)
> >
> > diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
> > index 1e7a3c1527..e7e09e48fc 100644
> > --- a/lib/mempool/rte_mempool.h
> > +++ b/lib/mempool/rte_mempool.h
> > @@ -1344,31 +1344,41 @@ rte_mempool_do_generic_put(struct rte_mempool
> > *mp, void * const *obj_table,
> > if (unlikely(cache == NULL || n > RTE_MEMPOOL_CACHE_MAX_SIZE))
> > goto ring_enqueue;
> >
> > - cache_objs = &cache->objs[cache->len];
> > + /* If the request itself is too big for the cache */
> > + if (unlikely(n > cache->flushthresh))
> > + goto ring_enqueue;
> >
> > /*
> > * The cache follows the following algorithm
> > - * 1. Add the objects to the cache
> > - * 2. Anything greater than the cache min value (if it crosses
> > the
> > - * cache flush threshold) is flushed to the ring.
>
> In the code, "the cache min value" is actually "the cache size". This indicates an intention to do something more. Perhaps the patch introducing the "flush threshold" was committed while still incomplete, and just never got completed?
>
> > + * 1. If the objects cannot be added to the cache without
> > + * crossing the flush threshold, flush the cache to the ring.
> > + * 2. Add the objects to the cache.
> > */
> >
> > - /* Add elements back into the cache */
> > - rte_memcpy(&cache_objs[0], obj_table, sizeof(void *) * n);
> > + if (cache->len + n <= cache->flushthresh) {
> > + cache_objs = &cache->objs[cache->len];
> >
> > - cache->len += n;
> > + cache->len += n;
> > + } else {
> > + cache_objs = &cache->objs[0];
> >
> > - if (cache->len >= cache->flushthresh) {
> > - rte_mempool_ops_enqueue_bulk(mp, &cache->objs[cache->size],
> > - cache->len - cache->size);
> > - cache->len = cache->size;
> > +#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
> > + if (rte_mempool_ops_enqueue_bulk(mp, cache_objs, cache-
> > >len) < 0)
> > + rte_panic("cannot put objects in mempool\n");
> > +#else
> > + rte_mempool_ops_enqueue_bulk(mp, cache_objs, cache->len);
> > +#endif
> > + cache->len = n;
> > }
> >
> > + /* Add the objects to the cache. */
> > + rte_memcpy(cache_objs, obj_table, sizeof(void *) * n);
> > +
> > return;
> >
> > ring_enqueue:
> >
> > - /* push remaining objects in ring */
> > + /* Put the objects into the ring */
> > #ifdef RTE_LIBRTE_MEMPOOL_DEBUG
> > if (rte_mempool_ops_enqueue_bulk(mp, obj_table, n) < 0)
> > rte_panic("cannot put objects in mempool\n");
> > --
> > 2.17.1
>
More information about the dev
mailing list