[dpdk-dev] [PATCH v2] net/kni: calc mbuf&mtu according to given mb_pool
Ferruh Yigit
ferruh.yigit at intel.com
Thu Mar 14 10:28:07 CET 2019
On 3/14/2019 6:37 AM, Liron Himi wrote:
>
>
> -----Original Message-----
> From: Ferruh Yigit <ferruh.yigit at intel.com>
> Sent: Wednesday, March 13, 2019 18:58
> To: Liron Himi <lironh at marvell.com>
> Cc: dev at dpdk.org; Alan Winkowski <walan at marvell.com>
> Subject: Re: [PATCH v2] net/kni: calc mbuf&mtu according to given mb_pool
>
> On 3/10/2019 2:27 PM, Liron Himi wrote:
>> Adding Alan.
>>
>> -----Original Message-----
>> From: Liron Himi
>> Sent: Monday, February 25, 2019 13:30
>> To: ferruh.yigit at intel.com
>> Cc: dev at dpdk.org; Liron Himi <lironh at marvell.com>; Liron Himi
>> <lironh at marvell.com>
>> Subject: RE: [PATCH v2] net/kni: calc mbuf&mtu according to given
>> mb_pool
>>
>> Hi,
>>
>> Kind reminder
>
> Sorry for late response.
>
>>
>> Regards,
>> Liron
>>
>> -----Original Message-----
>> From: lironh at marvell.com <lironh at marvell.com>
>> Sent: Saturday, February 23, 2019 22:15
>> To: ferruh.yigit at intel.com
>> Cc: dev at dpdk.org; Liron Himi <lironh at marvell.com>
>> Subject: [PATCH v2] net/kni: calc mbuf&mtu according to given mb_pool
>>
>> From: Liron Himi <lironh at marvell.com>
>>
>> - mbuf_size and mtu are now being calculated according to the given mb-pool.
>
> +1 to have dynamic size instead of fixed "MAX_PACKET_SZ"
>
>>
>> - max_mtu is now being set according to the given mtu
>>
>> the above two changes provide the ability to work with jumbo frames
>
> From kernel -> userspace, if the data length is bigger than mbuf->buffer_len (-
> headroom) the packet is dropped. I guess you are trying to solve that issue?
> [L.H.] correct
>
> By providing larger mbuf buffer, it should be possible to send larger (jumbo) packets?
> [L.H.] correct
>
> Another option can be adding multi segment send support, that also lets sending large packets from kernel to userspace, and it can co-exits with your patch.
> What do you think, can you work on that support?
> [L.H.] I suggest to first go with this patch, and then prepare multi-segment patch if possible
Yes, I was hoping both can go in a same patchset, can it be possible?
> Multi segment support already exists in userspace to kernel path, but otherway around is missing.
>
>>
>> Signed-off-by: Liron Himi <lironh at marvell.com>
>> ---
>> drivers/net/kni/rte_eth_kni.c | 10 +++++++---
>> kernel/linux/kni/compat.h | 4 ++++
>> kernel/linux/kni/kni_misc.c | 3 +++
>
> It can be good to update release notes / kni documentation to document new feature.
> [L.H.] okay
>
>> 3 files changed, 14 insertions(+), 3 deletions(-)
>>
>> diff --git a/drivers/net/kni/rte_eth_kni.c
>> b/drivers/net/kni/rte_eth_kni.c index a1e9970..5e02224 100644
>> --- a/drivers/net/kni/rte_eth_kni.c
>> +++ b/drivers/net/kni/rte_eth_kni.c
>> @@ -16,9 +16,11 @@
>> /* Only single queue supported */
>> #define KNI_MAX_QUEUE_PER_PORT 1
>>
>> -#define MAX_PACKET_SZ 2048
>> #define MAX_KNI_PORTS 8
>>
>> +#define KNI_ETHER_MTU(mbuf_size) \
>> + ((mbuf_size) - ETHER_HDR_LEN) /**< Ethernet MTU. */
>> +
>> #define ETH_KNI_NO_REQUEST_THREAD_ARG "no_request_thread"
>> static const char * const valid_arguments[] = {
>> ETH_KNI_NO_REQUEST_THREAD_ARG,
>> @@ -123,11 +125,13 @@ eth_kni_start(struct rte_eth_dev *dev)
>> struct rte_kni_conf conf;
>> const char *name = dev->device->name + 4; /* remove net_ */
>>
>> + mb_pool = internals->rx_queues[0].mb_pool;
>> snprintf(conf.name, RTE_KNI_NAMESIZE, "%s", name);
>> conf.force_bind = 0;
>> conf.group_id = port_id;
>> - conf.mbuf_size = MAX_PACKET_SZ;
>> - mb_pool = internals->rx_queues[0].mb_pool;
>> + conf.mbuf_size =
>> + rte_pktmbuf_data_room_size(mb_pool) - RTE_PKTMBUF_HEADROOM;
>> + conf.mtu = KNI_ETHER_MTU(conf.mbuf_size);
>
> Can you please do "conf.mbuf_size" changes also to kni sample application?
> kni sample application gets mtu from physical device, so I believe better to not change that but I think mbuf_size can be dynamic instead of hardcoded.
> [L.H.] okay
>
> Another question, for the case mbuf size < ETHER_MTU, should we keep MTU ETHER_MTU, what do you think?
> [L.H.] in any case we need to set the MTU according to the mbuf-size until multi-segment support will be available, right?
Right.
>
>>
>> internals->kni = rte_kni_alloc(mb_pool, &conf, NULL);
>> if (internals->kni == NULL) {
>> diff --git a/kernel/linux/kni/compat.h b/kernel/linux/kni/compat.h
>> index 3c575c7..b9f9a6f 100644
>> --- a/kernel/linux/kni/compat.h
>> +++ b/kernel/linux/kni/compat.h
>> @@ -117,3 +117,7 @@
>> #if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 11, 0) #define
>> HAVE_SIGNAL_FUNCTIONS_OWN_HEADER #endif
>> +
>> +#if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 10, 0) #define
>> +HAVE_MAX_MTU_PARAM #endif
>> diff --git a/kernel/linux/kni/kni_misc.c b/kernel/linux/kni/kni_misc.c
>> index 522ae23..04c78eb 100644
>> --- a/kernel/linux/kni/kni_misc.c
>> +++ b/kernel/linux/kni/kni_misc.c
>> @@ -459,6 +459,9 @@ kni_ioctl_create(struct net *net, uint32_t
>> ioctl_num,
>>
>> if (dev_info.mtu)
>> net_dev->mtu = dev_info.mtu;
>> +#ifdef HAVE_MAX_MTU_PARAM
>> + net_dev->max_mtu = net_dev->mtu;
>> +#endif
>
> Do we need to set 'max_mtu'? I guess this is not really required for large packet support, if so what do you think making this separate patch?
> [L.H.] 'max_mtu' is set by default to '1500', so in order to be able to modify the interface MTU to support jumbo (or even any size > 1500) the 'max_mtu' must be updated to the larger supported value.
I missed that it set by default to '1500', I was thinking it is zero by default.
Can you please point where its default value set in Linux?
More information about the dev
mailing list