[dpdk-dev] [PATCH v3 1/6] mem: add function for checking memsegs IOVAs addresses
Tu, Lijuan
lijuan.tu at intel.com
Wed Oct 10 10:56:05 CEST 2018
Hi
> -----Original Message-----
> From: dev [mailto:dev-bounces at dpdk.org] On Behalf Of Alejandro Lucero
> Sent: Friday, October 5, 2018 8:45 PM
> To: dev at dpdk.org
> Subject: [dpdk-dev] [PATCH v3 1/6] mem: add function for checking
> memsegs IOVAs addresses
>
> A device can suffer addressing limitations. This function checks memsegs
> have iovas within the supported range based on dma mask.
>
> PMDs should use this function during initialization if device suffers
> addressing limitations, returning an error if this function returns memsegs
> out of range.
>
> Another usage is for emulated IOMMU hardware with addressing limitations.
>
> It is necessary to save the most restricted dma mask for checking out
> memory allocated dynamically after initialization.
>
> Signed-off-by: Alejandro Lucero <alejandro.lucero at netronome.com>
> Reviewed-by: Anatoly Burakov <anatoly.burakov at intel.com>
> ---
> doc/guides/rel_notes/release_18_11.rst | 10 ++++
> lib/librte_eal/common/eal_common_memory.c | 60
> +++++++++++++++++++++++
> lib/librte_eal/common/include/rte_eal_memconfig.h | 3 ++
> lib/librte_eal/common/include/rte_memory.h | 3 ++
> lib/librte_eal/common/malloc_heap.c | 12 +++++
> lib/librte_eal/linuxapp/eal/eal.c | 2 +
> lib/librte_eal/rte_eal_version.map | 1 +
> 7 files changed, 91 insertions(+)
>
> diff --git a/doc/guides/rel_notes/release_18_11.rst
> b/doc/guides/rel_notes/release_18_11.rst
> index 2133a5b..c806dc6 100644
> --- a/doc/guides/rel_notes/release_18_11.rst
> +++ b/doc/guides/rel_notes/release_18_11.rst
> @@ -104,6 +104,14 @@ New Features
> the specified port. The port must be stopped before the command call in
> order
> to reconfigure queues.
>
> +* **Added check for ensuring allocated memory addressable by devices.**
> +
> + Some devices can have addressing limitations so a new function,
> + ``rte_eal_check_dma_mask``, has been added for checking allocated
> + memory is not out of the device range. Because now memory can be
> + dynamically allocated after initialization, a dma mask is kept and
> + any new allocated memory will be checked out against that dma mask
> + and rejected if out of range. If more than one device has addressing
> limitations, the dma mask is the more restricted one.
>
> API Changes
> -----------
> @@ -156,6 +164,8 @@ ABI Changes
> ``rte_config`` structure on account of improving DPDK usability
> when
> using either ``--legacy-mem`` or ``--single-file-segments`` flags.
>
> +* eal: added ``dma_maskbits`` to ``rte_mem_config`` for keeping more
> restricted
> + dma mask based on devices addressing limitations.
>
> Removed Items
> -------------
> diff --git a/lib/librte_eal/common/eal_common_memory.c
> b/lib/librte_eal/common/eal_common_memory.c
> index 0b69804..c482f0d 100644
> --- a/lib/librte_eal/common/eal_common_memory.c
> +++ b/lib/librte_eal/common/eal_common_memory.c
> @@ -385,6 +385,66 @@ struct virtiova {
> rte_memseg_walk(dump_memseg, f);
> }
>
> +static int
> +check_iova(const struct rte_memseg_list *msl __rte_unused,
> + const struct rte_memseg *ms, void *arg) {
> + uint64_t *mask = arg;
> + rte_iova_t iova;
> +
> + /* higher address within segment */
> + iova = (ms->iova + ms->len) - 1;
> + if (!(iova & *mask))
> + return 0;
> +
> + RTE_LOG(DEBUG, EAL, "memseg iova %"PRIx64", len %zx, out of
> range\n",
> + ms->iova, ms->len);
> +
> + RTE_LOG(DEBUG, EAL, "\tusing dma mask %"PRIx64"\n", *mask);
> + return 1;
> +}
> +
> +#if defined(RTE_ARCH_64)
> +#define MAX_DMA_MASK_BITS 63
> +#else
> +#define MAX_DMA_MASK_BITS 31
> +#endif
> +
> +/* check memseg iovas are within the required range based on dma mask
> +*/ int __rte_experimental rte_eal_check_dma_mask(uint8_t maskbits) {
> + struct rte_mem_config *mcfg =
> rte_eal_get_configuration()->mem_config;
> + uint64_t mask;
> +
> + /* sanity check */
> + if (maskbits > MAX_DMA_MASK_BITS) {
> + RTE_LOG(ERR, EAL, "wrong dma mask size %u (Max: %u)\n",
> + maskbits, MAX_DMA_MASK_BITS);
> + return -1;
> + }
> +
> + /* create dma mask */
> + mask = ~((1ULL << maskbits) - 1);
> +
> + if (rte_memseg_walk(check_iova, &mask))
[Lijuan]In my environment, testpmd halts at rte_memseg_walk() when maskbits is 0.
> + /*
> + * Dma mask precludes hugepage usage.
> + * This device can not be used and we do not need to keep
> + * the dma mask.
> + */
> + return 1;
> +
> + /*
> + * we need to keep the more restricted maskbit for checking
> + * potential dynamic memory allocation in the future.
> + */
> + mcfg->dma_maskbits = mcfg->dma_maskbits == 0 ? maskbits :
> + RTE_MIN(mcfg->dma_maskbits, maskbits);
> +
> + return 0;
> +}
> +
> /* return the number of memory channels */ unsigned
> rte_memory_get_nchannel(void) { diff --git
> a/lib/librte_eal/common/include/rte_eal_memconfig.h
> b/lib/librte_eal/common/include/rte_eal_memconfig.h
> index 62a21c2..b5dff70 100644
> --- a/lib/librte_eal/common/include/rte_eal_memconfig.h
> +++ b/lib/librte_eal/common/include/rte_eal_memconfig.h
> @@ -81,6 +81,9 @@ struct rte_mem_config {
> /* legacy mem and single file segments options are shared */
> uint32_t legacy_mem;
> uint32_t single_file_segments;
> +
> + /* keeps the more restricted dma mask */
> + uint8_t dma_maskbits;
> } __attribute__((__packed__));
>
>
> diff --git a/lib/librte_eal/common/include/rte_memory.h
> b/lib/librte_eal/common/include/rte_memory.h
> index 14bd277..c349d6c 100644
> --- a/lib/librte_eal/common/include/rte_memory.h
> +++ b/lib/librte_eal/common/include/rte_memory.h
> @@ -454,6 +454,9 @@ typedef int (*rte_memseg_list_walk_t)(const struct
> rte_memseg_list *msl,
> */
> unsigned rte_memory_get_nrank(void);
>
> +/* check memsegs iovas are within a range based on dma mask */ int
> +rte_eal_check_dma_mask(uint8_t maskbits);
> +
> /**
> * Drivers based on uio will not load unless physical
> * addresses are obtainable. It is only possible to get diff --git
> a/lib/librte_eal/common/malloc_heap.c
> b/lib/librte_eal/common/malloc_heap.c
> index ac7bbb3..3b5b2b6 100644
> --- a/lib/librte_eal/common/malloc_heap.c
> +++ b/lib/librte_eal/common/malloc_heap.c
> @@ -259,11 +259,13 @@ struct malloc_elem *
> int socket, unsigned int flags, size_t align, size_t bound,
> bool contig, struct rte_memseg **ms, int n_segs) {
> + struct rte_mem_config *mcfg =
> rte_eal_get_configuration()->mem_config;
> struct rte_memseg_list *msl;
> struct malloc_elem *elem = NULL;
> size_t alloc_sz;
> int allocd_pages;
> void *ret, *map_addr;
> + uint64_t mask;
>
> alloc_sz = (size_t)pg_sz * n_segs;
>
> @@ -291,6 +293,16 @@ struct malloc_elem *
> goto fail;
> }
>
> + if (mcfg->dma_maskbits) {
> + mask = ~((1ULL << mcfg->dma_maskbits) - 1);
> + if (rte_eal_check_dma_mask(mask)) {
> + RTE_LOG(ERR, EAL,
> + "%s(): couldn't allocate memory due to DMA mask\n",
> + __func__);
> + goto fail;
> + }
> + }
> +
> /* add newly minted memsegs to malloc heap */
> elem = malloc_heap_add_memory(heap, msl, map_addr, alloc_sz);
>
> diff --git a/lib/librte_eal/linuxapp/eal/eal.c
> b/lib/librte_eal/linuxapp/eal/eal.c
> index 4a55d3b..dfe1b8c 100644
> --- a/lib/librte_eal/linuxapp/eal/eal.c
> +++ b/lib/librte_eal/linuxapp/eal/eal.c
> @@ -263,6 +263,8 @@ enum rte_iova_mode
> * processes could later map the config into this exact location */
> rte_config.mem_config->mem_cfg_addr = (uintptr_t)
> rte_mem_cfg_addr;
>
> + rte_config.mem_config->dma_maskbits = 0;
> +
> }
>
> /* attach to an existing shared memory config */ diff --git
> a/lib/librte_eal/rte_eal_version.map b/lib/librte_eal/rte_eal_version.map
> index 73282bb..2baefce 100644
> --- a/lib/librte_eal/rte_eal_version.map
> +++ b/lib/librte_eal/rte_eal_version.map
> @@ -291,6 +291,7 @@ EXPERIMENTAL {
> rte_devargs_parsef;
> rte_devargs_remove;
> rte_devargs_type_count;
> + rte_eal_check_dma_mask;
> rte_eal_cleanup;
> rte_eal_hotplug_add;
> rte_eal_hotplug_remove;
> --
> 1.9.1
More information about the dev
mailing list