[RFC 09/27] vhost: add page size info to IOTLB entry
Maxime Coquelin
maxime.coquelin at redhat.com
Wed May 3 15:57:48 CEST 2023
Hi Chenbo,
On 4/25/23 08:20, Xia, Chenbo wrote:
> Hi Maxime,
>
>> -----Original Message-----
>> From: Maxime Coquelin <maxime.coquelin at redhat.com>
>> Sent: Friday, March 31, 2023 11:43 PM
>> To: dev at dpdk.org; david.marchand at redhat.com; Xia, Chenbo
>> <chenbo.xia at intel.com>; mkp at redhat.com; fbl at redhat.com;
>> jasowang at redhat.com; Liang, Cunming <cunming.liang at intel.com>; Xie, Yongji
>> <xieyongji at bytedance.com>; echaudro at redhat.com; eperezma at redhat.com;
>> amorenoz at redhat.com
>> Cc: Maxime Coquelin <maxime.coquelin at redhat.com>
>> Subject: [RFC 09/27] vhost: add page size info to IOTLB entry
>>
>> VDUSE will close the file descriptor after having mapped
>> the shared memory, so it will not be possible to get the
>> page size afterwards.
>>
>> This patch adds an new page_shift field to the IOTLB entry,
>> so that the information will be passed at IOTLB cache
>> insertion time. The information is stored as a bit shift
>> value so that IOTLB entry keeps fitting in a single
>> cacheline.
>>
>> Signed-off-by: Maxime Coquelin <maxime.coquelin at redhat.com>
>> ---
>> lib/vhost/iotlb.c | 46 ++++++++++++++++++++----------------------
>> lib/vhost/iotlb.h | 2 +-
>> lib/vhost/vhost.h | 1 -
>> lib/vhost/vhost_user.c | 8 +++++---
>> 4 files changed, 28 insertions(+), 29 deletions(-)
>>
>> diff --git a/lib/vhost/iotlb.c b/lib/vhost/iotlb.c
>> index 51f118bc48..188dfb8e38 100644
>> --- a/lib/vhost/iotlb.c
>> +++ b/lib/vhost/iotlb.c
>> @@ -19,14 +19,14 @@ struct vhost_iotlb_entry {
>> uint64_t uaddr;
>> uint64_t uoffset;
>> uint64_t size;
>> + uint8_t page_shift;
>> uint8_t perm;
>> };
>>
>> #define IOTLB_CACHE_SIZE 2048
>>
>> static bool
>> -vhost_user_iotlb_share_page(struct vhost_iotlb_entry *a, struct
>> vhost_iotlb_entry *b,
>> - uint64_t align)
>> +vhost_user_iotlb_share_page(struct vhost_iotlb_entry *a, struct
>> vhost_iotlb_entry *b)
>> {
>> uint64_t a_start, a_end, b_start;
>>
>> @@ -38,44 +38,41 @@ vhost_user_iotlb_share_page(struct vhost_iotlb_entry
>> *a, struct vhost_iotlb_entr
>>
>> /* Assumes entry a lower than entry b */
>> RTE_ASSERT(a_start < b_start);
>> - a_end = RTE_ALIGN_CEIL(a_start + a->size, align);
>> - b_start = RTE_ALIGN_FLOOR(b_start, align);
>> + a_end = RTE_ALIGN_CEIL(a_start + a->size, RTE_BIT64(a->page_shift));
>> + b_start = RTE_ALIGN_FLOOR(b_start, RTE_BIT64(b->page_shift));
>>
>> return a_end > b_start;
>> }
>>
>> static void
>> -vhost_user_iotlb_set_dump(struct virtio_net *dev, struct
>> vhost_iotlb_entry *node)
>> +vhost_user_iotlb_set_dump(struct vhost_iotlb_entry *node)
>> {
>> - uint64_t align, start;
>> + uint64_t start;
>>
>> start = node->uaddr + node->uoffset;
>> - align = hua_to_alignment(dev->mem, (void *)(uintptr_t)start);
>> -
>> - mem_set_dump((void *)(uintptr_t)start, node->size, false, align);
>> + mem_set_dump((void *)(uintptr_t)start, node->size, false,
>> RTE_BIT64(node->page_shift));
>> }
>>
>> static void
>> -vhost_user_iotlb_clear_dump(struct virtio_net *dev, struct
>> vhost_iotlb_entry *node,
>> +vhost_user_iotlb_clear_dump(struct vhost_iotlb_entry *node,
>> struct vhost_iotlb_entry *prev, struct vhost_iotlb_entry *next)
>> {
>> - uint64_t align, start, end;
>> + uint64_t start, end;
>>
>> start = node->uaddr + node->uoffset;
>> end = start + node->size;
>>
>> - align = hua_to_alignment(dev->mem, (void *)(uintptr_t)start);
>> -
>> /* Skip first page if shared with previous entry. */
>> - if (vhost_user_iotlb_share_page(prev, node, align))
>> - start = RTE_ALIGN_CEIL(start, align);
>> + if (vhost_user_iotlb_share_page(prev, node))
>> + start = RTE_ALIGN_CEIL(start, RTE_BIT64(node->page_shift));
>>
>> /* Skip last page if shared with next entry. */
>> - if (vhost_user_iotlb_share_page(node, next, align))
>> - end = RTE_ALIGN_FLOOR(end, align);
>> + if (vhost_user_iotlb_share_page(node, next))
>> + end = RTE_ALIGN_FLOOR(end, RTE_BIT64(node->page_shift));
>>
>> if (end > start)
>> - mem_set_dump((void *)(uintptr_t)start, end - start, false,
>> align);
>> + mem_set_dump((void *)(uintptr_t)start, end - start, false,
>> + RTE_BIT64(node->page_shift));
>> }
>>
>> static struct vhost_iotlb_entry *
>> @@ -198,7 +195,7 @@ vhost_user_iotlb_cache_remove_all(struct virtio_net
>> *dev)
>> vhost_user_iotlb_wr_lock_all(dev);
>>
>> RTE_TAILQ_FOREACH_SAFE(node, &dev->iotlb_list, next, temp_node) {
>> - vhost_user_iotlb_set_dump(dev, node);
>> + vhost_user_iotlb_set_dump(node);
>>
>> TAILQ_REMOVE(&dev->iotlb_list, node, next);
>> vhost_user_iotlb_pool_put(dev, node);
>> @@ -223,7 +220,7 @@ vhost_user_iotlb_cache_random_evict(struct virtio_net
>> *dev)
>> if (!entry_idx) {
>> struct vhost_iotlb_entry *next_node =
>> RTE_TAILQ_NEXT(node, next);
>>
>> - vhost_user_iotlb_clear_dump(dev, node, prev_node,
>> next_node);
>> + vhost_user_iotlb_clear_dump(node, prev_node, next_node);
>>
>> TAILQ_REMOVE(&dev->iotlb_list, node, next);
>> vhost_user_iotlb_pool_put(dev, node);
>> @@ -239,7 +236,7 @@ vhost_user_iotlb_cache_random_evict(struct virtio_net
>> *dev)
>>
>> void
>> vhost_user_iotlb_cache_insert(struct virtio_net *dev, uint64_t iova,
>> uint64_t uaddr,
>> - uint64_t uoffset, uint64_t size, uint8_t perm)
>> + uint64_t uoffset, uint64_t size, uint64_t
>> page_size, uint8_t perm)
>> {
>> struct vhost_iotlb_entry *node, *new_node;
>>
>> @@ -263,6 +260,7 @@ vhost_user_iotlb_cache_insert(struct virtio_net *dev,
>> uint64_t iova, uint64_t ua
>> new_node->uaddr = uaddr;
>> new_node->uoffset = uoffset;
>> new_node->size = size;
>> + new_node->page_shift = __builtin_ctz(page_size);
>
> __builtin_ctzll ?
Indeed, that's better. Weird I don't get a warning!
Fixed in v1.
Thanks,
Maxime
> Thanks,
> Chenbo
More information about the dev
mailing list