[dpdk-dev] [PATCH v14 3/4] test: add unit tests for latencystats library

Naga Suresh Somarowthu naga.sureshx.somarowthu at intel.com
Mon Oct 8 14:12:46 CEST 2018


Unit Test Cases added for latencystats library.

Signed-off-by: Naga Suresh Somarowthu <naga.sureshx.somarowthu at intel.com>
Reviewed-by: Reshma Pattan <reshma.pattan at intel.com>
---
 MAINTAINERS                   |   1 +
 test/test/Makefile            |   1 +
 test/test/autotest_data.py    |   6 ++
 test/test/meson.build         |   3 +
 test/test/test_latencystats.c | 224 ++++++++++++++++++++++++++++++++++++++++++
 5 files changed, 235 insertions(+)
 create mode 100644 test/test/test_latencystats.c

diff --git a/MAINTAINERS b/MAINTAINERS
index 4798fb258..5eee72956 100644
--- a/MAINTAINERS
+++ b/MAINTAINERS
@@ -1161,6 +1161,7 @@ F: test/test/test_bitratestats.c
 Latency statistics
 M: Reshma Pattan <reshma.pattan at intel.com>
 F: lib/librte_latencystats/
+F: test/test/test_latencystats.c
 
 BPF - EXPERIMENTAL
 M: Konstantin Ananyev <konstantin.ananyev at intel.com>
diff --git a/test/test/Makefile b/test/test/Makefile
index c030ab111..5729638a4 100644
--- a/test/test/Makefile
+++ b/test/test/Makefile
@@ -184,6 +184,7 @@ SRCS-$(CONFIG_RTE_LIBRTE_CRYPTODEV) += test_cryptodev_blockcipher.c
 SRCS-$(CONFIG_RTE_LIBRTE_CRYPTODEV) += test_cryptodev.c
 SRCS-$(CONFIG_RTE_LIBRTE_CRYPTODEV) += test_cryptodev_asym.c
 SRCS-$(CONFIG_RTE_LIBRTE_BITRATE) += test_bitratestats.c
+SRCS-$(CONFIG_RTE_LIBRTE_LATENCY_STATS) += test_latencystats.c
 
 ifeq ($(CONFIG_RTE_COMPRESSDEV_TEST),y)
 SRCS-$(CONFIG_RTE_LIBRTE_COMPRESSDEV) += test_compressdev.c
diff --git a/test/test/autotest_data.py b/test/test/autotest_data.py
index b253bf8f0..f44e31d33 100644
--- a/test/test/autotest_data.py
+++ b/test/test/autotest_data.py
@@ -488,6 +488,12 @@
         "Func":    default_autotest,
         "Report":  None,
     },
+    {
+        "Name":    "Latencystats autotest",
+        "Command": "latencystats_autotest",
+        "Func":    default_autotest,
+        "Report":  None,
+    },
     #
     #Please always keep all dump tests at the end and together!
     #
diff --git a/test/test/meson.build b/test/test/meson.build
index f40a28f27..e31e24096 100644
--- a/test/test/meson.build
+++ b/test/test/meson.build
@@ -51,6 +51,7 @@ test_sources = files('commands.c',
 	'test_interrupts.c',
 	'test_kni.c',
 	'test_kvargs.c',
+	'test_latencystats.c',
 	'test_link_bonding.c',
 	'test_link_bonding_mode4.c',
 	'test_logs.c',
@@ -116,6 +117,7 @@ test_deps = ['acl',
 	'eventdev',
 	'flow_classify',
 	'hash',
+	'latencystats',
 	'lpm',
 	'member',
 	'metrics',
@@ -179,6 +181,7 @@ test_names = [
 	'interrupt_autotest',
 	'kni_autotest',
 	'kvargs_autotest',
+	'latencystats_autotest',
 	'link_bonding_autotest',
 	'link_bonding_mode4_autotest',
 	'logs_autotest',
diff --git a/test/test/test_latencystats.c b/test/test/test_latencystats.c
new file mode 100644
index 000000000..039c508cd
--- /dev/null
+++ b/test/test/test_latencystats.c
@@ -0,0 +1,224 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2018 Intel Corporation
+ */
+
+#include <stdio.h>
+#include <stdint.h>
+#include <string.h>
+
+#include <rte_latencystats.h>
+#include "rte_lcore.h"
+#include "rte_metrics.h"
+
+#include "sample_packet_forward.h"
+#include "test.h"
+
+#define NUM_STATS 4
+#define LATENCY_NUM_PACKETS 10
+#define QUEUE_ID 0
+
+uint16_t portid;
+struct rte_ring *ring;
+
+struct rte_metric_name lat_stats_strings[] = {
+	{"min_latency_ns"},
+	{"avg_latency_ns"},
+	{"max_latency_ns"},
+	{"jitter_ns"},
+};
+
+/* Test case for latency init with metrics init */
+static int test_latency_init(void)
+{
+	int ret = 0;
+
+	/* Metrics Initialization */
+	rte_metrics_init(rte_socket_id());
+
+	ret = rte_latencystats_init(1, NULL);
+	TEST_ASSERT(ret >= 0, "Test Failed: rte_latencystats_init failed");
+
+	return TEST_SUCCESS;
+}
+
+/* Test case to update the latency stats */
+static int test_latency_update(void)
+{
+	int ret = 0;
+
+	ret = rte_latencystats_update();
+	TEST_ASSERT(ret >= 0, "Test Failed: rte_latencystats_update failed");
+
+	return TEST_SUCCESS;
+}
+
+/* Test case to uninit latency stats */
+static int test_latency_uninit(void)
+{
+	int ret = 0;
+
+	ret = rte_latencystats_uninit();
+	TEST_ASSERT(ret >= 0, "Test Failed: rte_latencystats_uninit failed");
+
+	return TEST_SUCCESS;
+}
+
+/* Test case to get names of latency stats */
+static int test_latencystats_get_names(void)
+{
+	int ret = 0, i = 0;
+	int size = 0;
+	struct rte_metric_name names[NUM_STATS];
+	struct rte_metric_name wrongnames[NUM_STATS - 2];
+
+	size_t m_size = sizeof(struct rte_metric_name);
+	for (i = 0; i < NUM_STATS; i++)
+		memset(&names[i], 0, m_size);
+	for (i = 0; i < NUM_STATS - 2; i++)
+		memset(&wrongnames[i], 0, m_size);
+
+	/* Success Test: Valid names and size */
+	size = NUM_STATS;
+	ret = rte_latencystats_get_names(names, size);
+	for (i = 0; i <= NUM_STATS; i++) {
+		if (strcmp(lat_stats_strings[i].name, names[i].name) == 0)
+			printf(" %s\n", names[i].name);
+		else
+			printf("Failed: Names are not matched\n");
+	}
+	TEST_ASSERT((ret == NUM_STATS), "Test Failed to get metrics names");
+
+	/* Failure Test: Invalid names and valid size */
+	ret = rte_latencystats_get_names(NULL, size);
+	TEST_ASSERT((ret == NUM_STATS), "Test Failed to get the metrics count,"
+		    "Actual: %d Expected: %d", ret, NUM_STATS);
+
+	/* Failure Test: Valid names and invalid size */
+	size = 0;
+	ret = rte_latencystats_get_names(names, size);
+	TEST_ASSERT((ret == NUM_STATS), "Test Failed to get the metrics count,"
+		    "Actual: %d Expected: %d", ret, NUM_STATS);
+
+	/* Failure Test: Invalid names (array size lesser than size) */
+	size = NUM_STATS + 1;
+	ret = rte_latencystats_get_names(wrongnames, size);
+	TEST_ASSERT((ret == NUM_STATS), "Test Failed to get metrics names");
+	return TEST_SUCCESS;
+}
+
+/* Test case to get latency stats values */
+static int test_latencystats_get(void)
+{
+	int ret = 0, i = 0;
+	int size = 0;
+	struct rte_metric_value values[NUM_STATS];
+	struct rte_metric_value wrongvalues[NUM_STATS - 2];
+
+	size_t v_size = sizeof(struct rte_metric_value);
+	for (i = 0; i < NUM_STATS; i++)
+		memset(&values[i], 0, v_size);
+	for (i = 0; i < NUM_STATS - 2; i++)
+		memset(&wrongvalues[i], 0, v_size);
+
+	/* Success Test: Valid values and valid size */
+	size = NUM_STATS;
+	ret = rte_latencystats_get(values, size);
+	TEST_ASSERT((ret == NUM_STATS), "Test Failed to get latency metrics"
+			" values");
+
+	/* Failure Test: Invalid values and valid size */
+	ret = rte_latencystats_get(NULL, size);
+	TEST_ASSERT((ret == NUM_STATS), "Test Failed to get the stats count,"
+		    "Actual: %d Expected: %d", ret, NUM_STATS);
+
+	/* Failure Test: Valid values and invalid size */
+	size = 0;
+	ret = rte_latencystats_get(values, size);
+	TEST_ASSERT((ret == NUM_STATS), "Test Failed to get the stats count,"
+		    "Actual: %d Expected: %d", ret, NUM_STATS);
+
+	/* Failure Test: Invalid values(array size lesser than size)
+	 * and invalid size
+	 */
+	size = NUM_STATS + 2;
+	ret = rte_latencystats_get(wrongvalues, size);
+	TEST_ASSERT(ret == NUM_STATS, "Test Failed to get latency metrics"
+			" values");
+
+	return TEST_SUCCESS;
+}
+
+static int test_latency_ring_setup(void)
+{
+	test_ring_setup(&ring, &portid);
+
+	return TEST_SUCCESS;
+}
+
+static void test_latency_ring_free(void)
+{
+	test_ring_free(ring);
+	test_vdev_uninit("net_ring_net_ringa");
+}
+
+static int test_latency_packet_forward(void)
+{
+	int ret;
+	struct rte_mbuf *pbuf[LATENCY_NUM_PACKETS] = { };
+	struct rte_mempool *mp;
+	char poolname[] = "mbuf_pool";
+
+	ret = test_get_mbuf_from_pool(&mp, pbuf, poolname);
+	if (ret < 0) {
+		printf("allocate mbuf pool Failed\n");
+		return TEST_FAILED;
+	}
+	ret = test_packet_forward(pbuf, portid, QUEUE_ID);
+	if (ret < 0)
+		printf("send pkts Failed\n");
+	test_put_mbuf_to_pool(mp, pbuf);
+
+	return TEST_SUCCESS;
+}
+
+static struct
+unit_test_suite latencystats_testsuite = {
+	.suite_name = "Latency Stats Unit Test Suite",
+	.setup = test_latency_ring_setup,
+	.teardown = test_latency_ring_free,
+	.unit_test_cases = {
+
+		/* Test Case 1: To check latency init with
+		 * metrics init
+		 */
+		TEST_CASE_ST(NULL, NULL, test_latency_init),
+
+		/* Test Case 2: Do packet forwarding for metrics
+		 * calculation and check the latency metrics values
+		 * are updated
+		 */
+		TEST_CASE_ST(test_latency_packet_forward, NULL,
+				test_latency_update),
+		/* Test Case 3: To check whether latency stats names
+		 * are retrieved
+		 */
+		TEST_CASE_ST(NULL, NULL, test_latencystats_get_names),
+
+		/* Test Case 4: To check whether latency stats
+		 * values are retrieved
+		 */
+		TEST_CASE_ST(NULL, NULL, test_latencystats_get),
+
+		/* Test Case 5: To check uninit of latency test */
+		TEST_CASE_ST(NULL, NULL, test_latency_uninit),
+
+		TEST_CASES_END()
+	}
+};
+
+static int test_latencystats(void)
+{
+	return unit_test_suite_runner(&latencystats_testsuite);
+}
+
+REGISTER_TEST_COMMAND(latencystats_autotest, test_latencystats);
-- 
2.13.6



More information about the dev mailing list