<html xmlns:v="urn:schemas-microsoft-com:vml" xmlns:o="urn:schemas-microsoft-com:office:office" xmlns:w="urn:schemas-microsoft-com:office:word" xmlns:m="http://schemas.microsoft.com/office/2004/12/omml" xmlns="http://www.w3.org/TR/REC-html40">
<head>
<meta http-equiv="Content-Type" content="text/html; charset=utf-8">
<meta name="Generator" content="Microsoft Word 15 (filtered medium)">
<style><!--
/* Font Definitions */
@font-face
{font-family:"Cambria Math";
panose-1:2 4 5 3 5 4 6 3 2 4;}
@font-face
{font-family:Calibri;
panose-1:2 15 5 2 2 2 4 3 2 4;}
@font-face
{font-family:Aptos;
panose-1:2 11 0 4 2 2 2 2 2 4;}
@font-face
{font-family:"Roboto Mono";
panose-1:0 0 0 9 0 0 0 0 0 0;}
/* Style Definitions */
p.MsoNormal, li.MsoNormal, div.MsoNormal
{margin:0cm;
font-size:10.0pt;
font-family:"Calibri",sans-serif;}
a:link, span.MsoHyperlink
{mso-style-priority:99;
color:blue;
text-decoration:underline;}
.MsoChpDefault
{mso-style-type:export-only;
font-size:10.0pt;
mso-ligatures:none;}
@page WordSection1
{size:612.0pt 792.0pt;
margin:72.0pt 72.0pt 72.0pt 72.0pt;}
div.WordSection1
{page:WordSection1;}
--></style>
</head>
<body lang="EN-IN" link="blue" vlink="purple" style="word-wrap:break-word">
<div class="WordSection1">
<p class="MsoNormal"><span style="font-size:11.0pt;mso-fareast-language:EN-US">Hey Darius,<o:p></o:p></span></p>
<p class="MsoNormal"><span style="font-size:11.0pt;mso-fareast-language:EN-US"><o:p> </o:p></span></p>
<p class="MsoNormal"><span style="font-size:11.0pt;mso-fareast-language:EN-US">Thanks for the help. Actually the issue was very silly – our SriovNetwork resource had a
</span><b><span style="font-family:"Roboto Mono";color:#0072EB;background:#F9F9F9">maxTxRate
</span></b>and<b><span style="font-family:"Roboto Mono";color:#0072EB;background:#F9F9F9"> minTxRate
</span></b><span style="font-size:11.0pt">specified due to which it was capping the generation in the first place.
<o:p></o:p></span></p>
<p class="MsoNormal"><span style="font-size:11.0pt;mso-fareast-language:EN-US"><o:p> </o:p></span></p>
<p class="MsoNormal"><span style="font-size:11.0pt;mso-fareast-language:EN-US">Removing that fixed the generation cap.
<o:p></o:p></span></p>
<p class="MsoNormal"><span style="font-size:11.0pt;mso-fareast-language:EN-US"><o:p> </o:p></span></p>
<div>
<div>
<div>
<p class="MsoNormal"><span style="font-size:11.0pt">Regards<o:p></o:p></span></p>
<p class="MsoNormal"><span style="font-size:11.0pt">Tanmay <o:p></o:p></span></p>
</div>
</div>
</div>
<p class="MsoNormal"><span style="font-size:11.0pt;mso-fareast-language:EN-US"><o:p> </o:p></span></p>
<p class="MsoNormal"><span style="font-size:11.0pt;mso-fareast-language:EN-US"><o:p> </o:p></span></p>
<div id="mail-editor-reference-message-container">
<div>
<div style="border:none;border-top:solid #B5C4DF 1.0pt;padding:3.0pt 0cm 0cm 0cm">
<p class="MsoNormal" style="margin-bottom:12.0pt"><b><span style="font-size:12.0pt;font-family:"Aptos",sans-serif;color:black">From:
</span></b><span style="font-size:12.0pt;font-family:"Aptos",sans-serif;color:black">Dariusz Sosnowski <dsosnowski@nvidia.com><br>
<b>Date: </b>Thursday, 25 April 2024 at 10:08</span><span style="font-size:12.0pt;font-family:"Arial",sans-serif;color:black"> </span><span style="font-size:12.0pt;font-family:"Aptos",sans-serif;color:black">PM<br>
<b>To: </b>Tanmay Pandey <tanmay@voereir.com><br>
<b>Cc: </b>users@dpdk.org <users@dpdk.org><br>
<b>Subject: </b>RE: Performance Bottleneck at NIC with Openshift<o:p></o:p></span></p>
</div>
<div>
<p class="MsoNormal"><span style="font-size:11.0pt">Hi,<br>
<br>
Since as you mentioned, the similar HW, the same DPDK and PRoX versions were able to achieve much better performance,<br>
I'd guess that the problem might be related to how processes in pods are being scheduled with OpenShift. Specifically, I would:<br>
<br>
- Check if both pods are not scheduled on the same cores.<br>
- Verify if cores on which these pods are running, are isolated.<br>
<br>
Anything interrupting threads responsible for generating traffic will hurt the performance.<br>
<br>
Best regards,<br>
Dariusz Sosnowski<br>
<br>
> From: Tanmay Pandey <tanmay@voereir.com> <br>
> Sent: Friday, April 5, 2024 14:41<br>
> To: users@dpdk.org<br>
> Subject: Performance Bottleneck at NIC with Openshift<br>
> <br>
> External email: Use caution opening links or attachments <br>
> <br>
> Hi,<br>
> <br>
> I am using DPDK version 22.11 for performance evaluation running PRoX on an Openshift Cluster where I have created two pods - I am sending traffic from one and receiving on the other and I've found that I'm unable to utilize more than 6GB of bandwidth in
the server at the packet generation level. I have tested with a 64-byte frame size and achieved a maximum of 6.99 MPPS.<br>
> I've attempted to address this issue by adhering to the recommendations outlined in the DPDK 22.11 NVIDIA Mellanox NIC performance report available at
</span><a href="https://fast.dpdk.org/doc/perf/DPDK_22_11_NVIDIA_Mellanox_NIC_performance_report.pdf"><span style="font-size:11.0pt">https://fast.dpdk.org/doc/perf/DPDK_22_11_NVIDIA_Mellanox_NIC_performance_report.pdf</span></a><span style="font-size:11.0pt">
. However, the problem persists.<br>
> Additionally, I've investigated packet loss at the NIC interface level and found no anomalies. The bottleneck appears to be related to packet generation, but I'm uncertain about the underlying cause.<br>
> I am very new to DPDK so don't really know how to debug this issue. I believe there is something happening between the NIC layer and Openshift.
<br>
> Additionally, I used the same hardware running kubeadm where I was using the same DPDK and PRoX version with a similar setup and was able to achieve much better performance(at least for the packet generation part - where my current bottleneck occurs).<br>
> Can someone point me in the right direction? <br>
> I would be happy to provide any other required information<br>
> Below are the SUT details:<br>
> Nic Model: Ethernet controller: Mellanox Technologies MT2892 Family [ConnectX-6 Dx]<br>
> uname -r<br>
> 5.14.0-284.54.1.rt14.339.el9_2.x86_64<br>
> <br>
> ethtool -i enp216s0f0np0<br>
> driver: mlx5_core<br>
> version: 5.14.0-284.54.1.rt14.339.el9_2.<br>
> firmware-version: 22.35.2000 (MT_0000000359)<br>
> expansion-rom-version:<br>
> bus-info: 0000:d8:00.0<br>
> supports-statistics: yes<br>
> supports-test: yes<br>
> supports-eeprom-access: no<br>
> supports-register-dump: no<br>
> supports-priv-flags: yes<br>
> <br>
> ## CPU<br>
> Architecture: x86_64<br>
> CPU op-mode(s): 32-bit, 64-bit<br>
> Address sizes: 46 bits physical, 48 bits virtual<br>
> Byte Order: Little Endian<br>
> CPU(s): 104<br>
> On-line CPU(s) list: 0-103<br>
> Vendor ID: GenuineIntel<br>
> BIOS Vendor ID: Intel<br>
> Model name: Intel(R) Xeon(R) Gold 6230R CPU @ 2.10GHz<br>
> BIOS Model name: Intel(R) Xeon(R) Gold 6230R CPU @ 2.10GHz<br>
> Operating System:<br>
> cat /etc/os-release<br>
> NAME="Red Hat Enterprise Linux CoreOS"<br>
> ID="rhcos"<br>
> ID_LIKE="rhel fedora"<br>
> VERSION="415.92.202402201450-0"<br>
> VERSION_ID="4.15"<br>
> VARIANT="CoreOS"<br>
> VARIANT_ID=coreos<br>
> PLATFORM_ID="platform:el9"<br>
> PRETTY_NAME="Red Hat Enterprise Linux CoreOS 415.92.202402201450-0 (Plow)"<br>
> ANSI_COLOR="0;31"<br>
> CPE_NAME="cpe:/o:redhat:enterprise_linux:9::coreos"<br>
> HOME_URL="</span><a href="https://www.redhat.com/"><span style="font-size:11.0pt">https://www.redhat.com/</span></a><span style="font-size:11.0pt">"<br>
> DOCUMENTATION_URL="</span><a href="https://docs.openshift.com/container-platform/4.15/"><span style="font-size:11.0pt">https://docs.openshift.com/container-platform/4.15/</span></a><span style="font-size:11.0pt">"<br>
> BUG_REPORT_URL="</span><a href="https://bugzilla.redhat.com/"><span style="font-size:11.0pt">https://bugzilla.redhat.com/</span></a><span style="font-size:11.0pt">"<br>
> REDHAT_BUGZILLA_PRODUCT="OpenShift Container Platform"<br>
> REDHAT_BUGZILLA_PRODUCT_VERSION="4.15"<br>
> REDHAT_SUPPORT_PRODUCT="OpenShift Container Platform"<br>
> REDHAT_SUPPORT_PRODUCT_VERSION="4.15"<br>
> OPENSHIFT_VERSION="4.15"<br>
> RHEL_VERSION="9.2"<br>
> OSTREE_VERSION="415.92.202402201450-0"<br>
> OCP Cluster<br>
> oc version<br>
> Client Version: 4.15.0-202402070507.p0.g48dcf59.assembly.stream-48dcf59<br>
> Kustomize Version: v5.0.4-0.20230601165947-6ce0bf390ce3<o:p></o:p></span></p>
</div>
</div>
</div>
</div>
</body>
</html>