Introduce rawdev driver support for ZXDH which<br />can help to connect two separate hosts with each other.<br /> <br />Signed-off-by: Yong Zhang <zhang.yong25@zte.com.cn> <br />---<br /> MAINTAINERS                    |   5 +<br /> doc/guides/rawdevs/index.rst   |   1 +<br /> doc/guides/rawdevs/zxdh.rst    |  29 ++<br /> drivers/raw/meson.build        |   1 +<br /> drivers/raw/zxdh/meson.build   |   5 +<br /> drivers/raw/zxdh/zxdh_pci.c    | 562 ++++++++++++++++++++<br /> drivers/raw/zxdh/zxdh_pci.h    |  41 ++<br /> drivers/raw/zxdh/zxdh_rawdev.c | 915 +++++++++++++++++++++++++++++++++<br /> drivers/raw/zxdh/zxdh_rawdev.h | 167 ++++++<br /> 9 files changed, 1726 insertions(+)<br /> create mode 100644 doc/guides/rawdevs/zxdh.rst<br /> create mode 100644 drivers/raw/zxdh/meson.build<br /> create mode 100644 drivers/raw/zxdh/zxdh_pci.c<br /> create mode 100644 drivers/raw/zxdh/zxdh_pci.h<br /> create mode 100644 drivers/raw/zxdh/zxdh_rawdev.c<br /> create mode 100644 drivers/raw/zxdh/zxdh_rawdev.h<br /> <br />diff --git a/MAINTAINERS b/MAINTAINERS<br />index c9adff9846..4fcf137c10 100644<br />--- a/MAINTAINERS<br />+++ b/MAINTAINERS<br />@@ -1478,6 +1478,11 @@ M: Gagandeep Singh <g.singh@nxp.com> <br /> F: drivers/raw/dpaa2_cmdif/<br /> F: doc/guides/rawdevs/dpaa2_cmdif.rst<br />  <br />+ZXDH<br />+M: Yong Zhang <zhang.yong25@zte.com.cn> <br />+F: drivers/raw/zxdh/<br />+F: doc/guides/rawdevs/zxdh.rst<br />+<br />  <br /> Packet processing<br /> -----------------<br />diff --git a/doc/guides/rawdevs/index.rst b/doc/guides/rawdevs/index.rst<br />index f34315f051..d85a4b7148 100644<br />--- a/doc/guides/rawdevs/index.rst<br />+++ b/doc/guides/rawdevs/index.rst<br />@@ -16,3 +16,4 @@ application through rawdev API.<br />     dpaa2_cmdif<br />     ifpga<br />     ntb<br />+    zxdh<br />diff --git a/doc/guides/rawdevs/zxdh.rst b/doc/guides/rawdevs/zxdh.rst<br />new file mode 100644<br />index 0000000000..a75fd8a748<br />--- /dev/null<br />+++ b/doc/guides/rawdevs/zxdh.rst<br />@@ -0,0 +1,29 @@<br />+..  SPDX-License-Identifier: BSD-3-Clause<br />+    Copyright 2024 ZTE Corporation<br />+<br />+ZXDH Rawdev Driver<br />+======================<br />+<br />+The ``zxdh`` rawdev driver is an implementation of the rawdev API,<br />+that provides communication between two separate hosts.<br />+This is achieved via using the GDMA controller of Dinghai SoC,<br />+which can be configured through exposed MPF devices.<br />+<br />+Device Setup<br />+-------------<br />+<br />+It is recommended to bind the ZXDH MPF kernel driver for MPF devices (Not mandatory).<br />+The kernel drivers can be downloaded at `ZTE Official Website<br />+<https://enterprise.zte.com.cn/>`_.<br />+<br />+Initialization<br />+--------------<br />+<br />+The ``zxdh`` rawdev driver needs to work in IOVA PA mode.<br />+Consider using ``--iova-mode=pa`` in the EAL options.<br />+<br />+Platform Requirement<br />+~~~~~~~~~~~~~~~~~~~~<br />+<br />+This PMD is only supported on ZTE Neo Platforms:<br />+- Neo X510/X512<br />diff --git a/drivers/raw/meson.build b/drivers/raw/meson.build<br />index 05cad143fe..237d1bdd80 100644<br />--- a/drivers/raw/meson.build<br />+++ b/drivers/raw/meson.build<br />@@ -12,5 +12,6 @@ drivers = [<br />         'ifpga',<br />         'ntb',<br />         'skeleton',<br />+        'zxdh',<br /> ]<br /> std_deps = ['rawdev']<br />diff --git a/drivers/raw/zxdh/meson.build b/drivers/raw/zxdh/meson.build<br />new file mode 100644<br />index 0000000000..c9e5c83f1b<br />--- /dev/null<br />+++ b/drivers/raw/zxdh/meson.build<br />@@ -0,0 +1,5 @@<br />+#SPDX-License-Identifier: BSD-3-Clause<br />+#Copyright 2024 ZTE Corporation<br />+<br />+deps += ['rawdev', 'kvargs', 'mbuf', 'bus_pci']<br />+sources = files('zxdh_rawdev.c', 'zxdh_pci.c')<br />diff --git a/drivers/raw/zxdh/zxdh_pci.c b/drivers/raw/zxdh/zxdh_pci.c<br />new file mode 100644<br />index 0000000000..884e39109a<br />--- /dev/null<br />+++ b/drivers/raw/zxdh/zxdh_pci.c<br />@@ -0,0 +1,562 @@<br />+/* SPDX-License-Identifier: BSD-3-Clause<br />+ * Copyright 2024 ZTE Corporation<br />+ */<br />+<br />+#include <stdio.h> <br />+#include <stdlib.h> <br />+#include <string.h> <br />+#include <getopt.h> <br />+#include <sys/mman.h> <br />+#include <fcntl.h> <br />+#include <errno.h> <br />+#include <dirent.h> <br />+#include <unistd.h> <br />+<br />+#include "zxdh_rawdev.h" <br />+#include "zxdh_pci.h" <br />+<br />+#define PCI_DEVICES_DIR          "/sys/bus/pci/devices" <br />+<br />+#define BAR0_IDX                 (0)<br />+#define BAR2_IDX                 (2)<br />+<br />+#define IORESOURCE_MEM           (0x00000200)<br />+#define FILE_FMT_NVAL            (2)<br />+<br />+#define STR_BUFF_LEN             (128)<br />+<br />+#define BYTES_NO_SWAP            (0)<br />+#define BYTES_SWAP               (1)<br />+<br />+#define PCI_CMD_OFFSET           (0x04)<br />+#define PCI_CMD_BYTES            (2)<br />+#define PCI_CMD_MSE_BIT          (1)<br />+#define FPGA_VER_OFFSET          (0x420)<br />+#define FPGA_VER_BYTES           (4)<br />+#define BOM_ID_OFFSET            (0x424)<br />+#define BOM_ID_BYTES             (1)<br />+#define FPGA_PR_FLAG_OFFSET      (0x425)<br />+#define FPGA_PR_FLAG_BYTES       (1)<br />+#define BOARD_ID_OFFSET          (0x426)<br />+#define BOARD_ID_BYTES           (2)<br />+#define FPGA_MAKE_TIME_OFFSET    (0x428)<br />+#define FPGA_MAKE_TIME_BYTES     (4)<br />+<br />+#define PARA_PR_FLAG             (0)<br />+#define PARA_FPGA_VER            (1)<br />+#define PARA_FPGA_MAKE_TIME      (2)<br />+#define PARA_BOARD_ID            (3)<br />+#define PARA_BOM_ID              (4)<br />+#define PARA_PCI_CMD             (5)<br />+<br />+#define PCI_READ                 (0)<br />+#define PCI_WRITE                (1)<br />+<br />+struct zxdh_pci_dev gdev;<br />+<br />+static int<br />+zxdh_gdma_rw_pci_config(struct zxdh_pci_dev *dev, uint8_t rw, uint offset, uint count, uint8_t *buf)<br />+{<br />+    int fd = -1;<br />+    uint res = 0;<br />+    int ret = -1;<br />+    char filename[FILE_PATH_LEN] = {0};<br />+<br />+    snprintf(filename, sizeof(filename), "/proc/bus/pci/%02x/%02x.%d",<br />+            dev->bus, dev->devid, dev->function);<br />+    fd = open(filename, O_RDWR);<br />+    if (fd < 0) {<br />+        snprintf(filename, sizeof(filename), "/proc/bus/pci/%04x:%02x/%02x.%d",<br />+                dev->domain, dev->bus, dev->devid, dev->function);<br />+        fd = open(filename, O_RDWR);<br />+        if (fd < 0) {<br />+            ZXDH_PMD_LOG(ERR, "Failed to open file:%s, fd:%d!", filename, fd);<br />+            return -1;<br />+        }<br />+    }<br />+<br />+    res = lseek(fd, offset, SEEK_SET);<br />+    if (res != offset) {<br />+        close(fd);<br />+        ZXDH_PMD_LOG(ERR, "Failed to lseek pci, res:%d!", res);<br />+        return -1;<br />+    }<br />+<br />+    if (rw == PCI_READ)<br />+        ret = read(fd, buf, count);<br />+    else<br />+        ret = write(fd, buf, count);<br />+<br />+    if (ret < 0) {<br />+        close(fd);<br />+        ZXDH_PMD_LOG(ERR, "Failed to rw pci:%d, ret:%d!", rw, ret);<br />+        return -1;<br />+    }<br />+<br />+    close(fd);<br />+    return 0;<br />+}<br />+<br />+static int<br />+zxdh_gdma_cfg_space_read(struct zxdh_pci_dev *dev, uint8_t ParaType, uint *pParaVer)<br />+{<br />+    int ret = 0;<br />+    uint8_t aRegVal[sizeof(uint)] = {0};<br />+    uint8_t ucLoop = 0;<br />+    uint8_t ucSwap = BYTES_NO_SWAP;<br />+    uint dwRegOffset = 0;<br />+    uint dwRegLen = 0;<br />+<br />+    if ((dev == NULL) || (pParaVer == NULL)) {<br />+        ZXDH_PMD_LOG(ERR, "Param is invalid!");<br />+        return -EINVAL;<br />+    }<br />+<br />+    switch (ParaType) {<br />+    case PARA_PR_FLAG:<br />+        dwRegOffset = FPGA_PR_FLAG_OFFSET;<br />+        dwRegLen    = FPGA_PR_FLAG_BYTES;<br />+        ucSwap      = BYTES_NO_SWAP;<br />+        break;<br />+    case PARA_FPGA_VER:<br />+        dwRegOffset = FPGA_VER_OFFSET;<br />+        dwRegLen    = FPGA_VER_BYTES;<br />+        ucSwap      = BYTES_NO_SWAP;<br />+        break;<br />+    case PARA_FPGA_MAKE_TIME:<br />+        dwRegOffset = FPGA_MAKE_TIME_OFFSET;<br />+        dwRegLen    = FPGA_MAKE_TIME_BYTES;<br />+        ucSwap      = BYTES_NO_SWAP;<br />+        break;<br />+    case PARA_BOARD_ID:<br />+        dwRegOffset = BOARD_ID_OFFSET;<br />+        dwRegLen    = BOARD_ID_BYTES;<br />+        ucSwap      = BYTES_NO_SWAP;<br />+        break;<br />+    case PARA_BOM_ID:<br />+        dwRegOffset = BOM_ID_OFFSET;<br />+        dwRegLen    = BOM_ID_BYTES;<br />+        ucSwap      = BYTES_NO_SWAP;<br />+        break;<br />+    case PARA_PCI_CMD:<br />+        dwRegOffset = PCI_CMD_OFFSET;<br />+        dwRegLen    = PCI_CMD_BYTES;<br />+        ucSwap      = BYTES_SWAP;<br />+        break;<br />+    default:<br />+        ZXDH_PMD_LOG(ERR, "ParaType %u not support!", ParaType);<br />+        return -EINVAL;<br />+    }<br />+<br />+    if (dwRegLen > sizeof(uint)) {<br />+        ZXDH_PMD_LOG(ERR, "dwRegLen %u is invalid", dwRegLen);<br />+        return -1;<br />+    }<br />+<br />+    *pParaVer = 0;<br />+    ret = zxdh_gdma_rw_pci_config(dev, PCI_READ, dwRegOffset, dwRegLen, aRegVal);<br />+    if (ret != 0) {<br />+        ZXDH_PMD_LOG(ERR, "ParaType %u, zxdh_gdma_rw_pci_config failed!", ParaType);<br />+        return ret;<br />+    }<br />+<br />+    if (ucSwap == BYTES_SWAP) {<br />+        for (ucLoop = 0; ucLoop < dwRegLen; ucLoop++)<br />+            *pParaVer = (*pParaVer << 8) | aRegVal[dwRegLen-1-ucLoop];<br />+    } else {<br />+        for (ucLoop = 0; ucLoop < dwRegLen; ucLoop++)<br />+            *pParaVer = (*pParaVer << 8) | aRegVal[ucLoop];<br />+    }<br />+<br />+    return ret;<br />+}<br />+<br />+static int<br />+zxdh_gdma_cfg_space_write(struct zxdh_pci_dev *dev, uint8_t ParaType, uint *pParaVer)<br />+{<br />+    int ret = 0;<br />+    uint8_t aRegVal[sizeof(uint)] = {0};<br />+    uint8_t ucLoop = 0;<br />+    uint8_t ucSwap = BYTES_NO_SWAP;<br />+    uint dwRegOffset = 0;<br />+    uint dwRegLen = 0;<br />+<br />+    if ((dev == NULL) || (pParaVer == NULL)) {<br />+        ZXDH_PMD_LOG(ERR, "Param is invalid");<br />+        return -EINVAL;<br />+    }<br />+<br />+    if (ParaType != PARA_PCI_CMD) {<br />+        ZXDH_PMD_LOG(ERR, "ParaType %u not support!", ParaType);<br />+        return -EINVAL;<br />+    }<br />+<br />+    dwRegOffset = PCI_CMD_OFFSET;<br />+    dwRegLen = PCI_CMD_BYTES;<br />+    ucSwap = BYTES_SWAP;<br />+<br />+    if (dwRegLen > sizeof(uint)) {<br />+        ZXDH_PMD_LOG(ERR, "dwRegLen %u is invalid", dwRegLen);<br />+        return -1;<br />+    }<br />+<br />+    if (ucSwap == BYTES_SWAP) {<br />+        for (ucLoop = 0; ucLoop < dwRegLen; ucLoop++)<br />+            aRegVal[ucLoop] = (*pParaVer >> 8*ucLoop) & 0xff;<br />+    } else {<br />+        for (ucLoop = 0; ucLoop < dwRegLen; ucLoop++)<br />+            aRegVal[ucLoop] = (*pParaVer >> 8*(dwRegLen-1-ucLoop)) & 0xff;<br />+    }<br />+<br />+    ret = zxdh_gdma_rw_pci_config(dev, PCI_WRITE, dwRegOffset, dwRegLen, aRegVal);<br />+    if (ret != 0) {<br />+        ZXDH_PMD_LOG(ERR, "ParaType %u, zxdh_gdma_rw_pci_config failed!", ParaType);<br />+        return ret;<br />+    }<br />+<br />+    return ret;<br />+}<br />+<br />+static int<br />+zxdh_gdma_str_split(char *string, int stringlen, char **tokens, int maxtokens, char delim)<br />+{<br />+    int loop = 0;<br />+    int tok = 0;<br />+    int tokstart = 1; /* first token is right at start of string */<br />+<br />+    if (string == NULL || tokens == NULL) {<br />+        ZXDH_PMD_LOG(ERR, "Param is invalid!");<br />+        return -1;<br />+    }<br />+<br />+    for (loop = 0; loop < stringlen; loop++) {<br />+        if (string[loop] == '\0' || tok >= maxtokens)<br />+            break;<br />+<br />+        if (tokstart) {<br />+            tokstart = 0;<br />+            tokens[tok++] = &string[loop];<br />+        }<br />+<br />+        if (string[loop] == delim) {<br />+            string[loop] = '\0';<br />+            tokstart = 1;<br />+        }<br />+    }<br />+<br />+    return tok;<br />+}<br />+<br />+static int<br />+zxdh_gdma_devfs_parse(const char *filename, unsigned long *val)<br />+{<br />+    FILE *f = NULL;<br />+    char *end = NULL;<br />+    char buf[STR_BUFF_LEN] = {0};<br />+<br />+    f = fopen(filename, "r");<br />+    if (f == NULL) {<br />+        ZXDH_PMD_LOG(ERR, "Cannot open sysfs %s", filename);<br />+        return -1;<br />+    }<br />+<br />+    if (fgets(buf, sizeof(buf), f) == NULL) {<br />+        ZXDH_PMD_LOG(ERR, "Cannot read sysfs value %s", filename);<br />+        fclose(f);<br />+        return -1;<br />+    }<br />+<br />+    *val = strtoul(buf, &end, 0);<br />+    if ((buf[0] == '\0') || (end == NULL) || (*end != '\n')) {<br />+        ZXDH_PMD_LOG(ERR, "Cannot parse sysfs value %s", filename);<br />+        fclose(f);<br />+        return -1;<br />+    }<br />+<br />+    fclose(f);<br />+    return 0;<br />+}<br />+<br />+static int<br />+zxdh_gdma_resfs_parse(const char *filename, struct zxdh_pci_dev *dev)<br />+{<br />+    FILE *fp = NULL;<br />+    char buf[STR_BUFF_LEN] = {0};<br />+    uint8_t  loop = 0;<br />+    uint64_t phys_addr = 0;<br />+    uint64_t end_addr = 0;<br />+    uint64_t flags = 0;<br />+    int ret = 0;<br />+    union pci_resource_info {<br />+        struct {<br />+            char *phys_addr;<br />+            char *end_addr;<br />+            char *flags;<br />+        };<br />+        char *ptrs[PCI_RESOURCE_FMT_NVAL];<br />+    } res_info;<br />+<br />+    fp = fopen(filename, "r");<br />+    if (fp == NULL) {<br />+        ZXDH_PMD_LOG(ERR, "Failed to open file %s", filename);<br />+        return -1;<br />+    }<br />+<br />+    for (loop = 0; loop < PCI_MAX_RESOURCE; loop++) {<br />+        if (fgets(buf, sizeof(buf), fp) == NULL) {<br />+            ZXDH_PMD_LOG(ERR, "Failed to gets file %s", filename);<br />+            goto err_exit;<br />+        }<br />+<br />+        ret = zxdh_gdma_str_split(buf, sizeof(buf), res_info.ptrs,<br />+                                    PCI_RESOURCE_FMT_NVAL, ' ');<br />+        if (ret != PCI_RESOURCE_FMT_NVAL) {<br />+            ZXDH_PMD_LOG(ERR, "file %s:zxdh_gdma_str_split failed!", filename);<br />+            goto err_exit;<br />+        }<br />+        errno = 0;<br />+        phys_addr = strtoull(res_info.phys_addr, NULL, 16);<br />+        end_addr  = strtoull(res_info.end_addr, NULL, 16);<br />+        flags     = strtoull(res_info.flags, NULL, 16);<br />+<br />+        if (errno != 0) {<br />+            ZXDH_PMD_LOG(ERR, "file %s:bad resource format!", filename);<br />+            goto err_exit;<br />+        }<br />+<br />+        if (flags & IORESOURCE_MEM) {<br />+            if (loop == BAR0_IDX) {<br />+                dev->bar_pa[BAR0_IDX] = phys_addr;<br />+                dev->bar_len[BAR0_IDX] = end_addr - phys_addr + 1;<br />+            }<br />+<br />+            if (loop == BAR2_IDX) {<br />+                dev->bar_pa[BAR2_IDX] = phys_addr;<br />+                dev->bar_len[BAR2_IDX] = end_addr - phys_addr + 1;<br />+                fclose(fp);<br />+                return 0;<br />+            }<br />+        }<br />+    }<br />+<br />+    ZXDH_PMD_LOG(ERR, "file %s: Not found IO resource memory!", filename);<br />+<br />+err_exit:<br />+    fclose(fp);<br />+    return -1;<br />+}<br />+<br />+static int<br />+zxdh_gdma_pci_addr_parse(const char *buf, int buf_size, struct zxdh_pci_dev *dev)<br />+{<br />+    char *buf_copy = NULL;<br />+    int ret = 0;<br />+    union splitaddr {<br />+        struct {<br />+            char *domain;<br />+            char *bus;<br />+            char *devid;<br />+            char *function;<br />+        };<br />+        char *str[PCI_FMT_NVAL];<br />+    } splitaddr;<br />+<br />+    buf_copy = strndup(buf, buf_size);<br />+    if (buf_copy == NULL) {<br />+        ZXDH_PMD_LOG(ERR, "buf %s: strndup failed!", buf);<br />+        return -1;<br />+    }<br />+<br />+    /* first split on ':' */<br />+    ret = zxdh_gdma_str_split(buf_copy, buf_size, splitaddr.str, PCI_FMT_NVAL, ':');<br />+    if (ret != (PCI_FMT_NVAL - 1)) {<br />+        ZXDH_PMD_LOG(ERR, "buf %s: zxdh_gdma_str_split failed!", buf);<br />+        goto err_exit;<br />+    }<br />+<br />+    /* final split is on '.' between devid and function */<br />+    splitaddr.function = strchr(splitaddr.devid, '.');<br />+    if (splitaddr.function == NULL) {<br />+        ZXDH_PMD_LOG(ERR, "buf %s: strchr failed!", buf);<br />+        goto err_exit;<br />+    }<br />+    *splitaddr.function++ = '\0';<br />+<br />+    /* now convert to int values */<br />+    errno = 0;<br />+    dev->domain = (uint16_t)strtoul(splitaddr.domain, NULL, 16);<br />+    dev->bus = (uint8_t)strtoul(splitaddr.bus, NULL, 16);<br />+    dev->devid = (uint8_t)strtoul(splitaddr.devid, NULL, 16);<br />+    dev->function = (uint8_t)strtoul(splitaddr.function, NULL, 10);<br />+    if (errno != 0) {<br />+        ZXDH_PMD_LOG(ERR, "buf %s: bad format!", buf);<br />+        goto err_exit;<br />+    }<br />+    free(buf_copy);<br />+    return 0;<br />+<br />+err_exit:<br />+    free(buf_copy);<br />+    return -1;<br />+}<br />+<br />+static int<br />+zxdh_gdma_pci_dev_mmap(const char *filename, struct zxdh_pci_dev *dev, uint8_t bar_idx)<br />+{<br />+    int fd = -1;<br />+<br />+    if (dev->bar_va[bar_idx] == NULL) {<br />+        fd = open(filename, O_RDWR);<br />+        if (fd < 0) {<br />+            ZXDH_PMD_LOG(ERR, "Failed to open file %s", filename);<br />+            return -1;<br />+        }<br />+<br />+        dev->bar_va[bar_idx] = mmap((void *)dev->bar_pa[bar_idx],<br />+                                    dev->bar_len[bar_idx],<br />+                                    PROT_READ | PROT_WRITE,<br />+                                    MAP_SHARED, fd, 0);<br />+<br />+        if (dev->bar_va[bar_idx] == MAP_FAILED) {<br />+            ZXDH_PMD_LOG(ERR, "Failed to mmap file %s!", filename);<br />+            goto err_exit;<br />+        }<br />+        close(fd);<br />+    } else<br />+        ZXDH_PMD_LOG(ERR, "BarVirtAddr is not NULL!");<br />+<br />+    return 0;<br />+<br />+err_exit:<br />+    close(fd);<br />+    return -1;<br />+}<br />+<br />+void<br />+zxdh_gdma_pci_dev_munmap(void)<br />+{<br />+    if (gdev.bar_va[BAR0_IDX] != NULL) {<br />+        munmap(gdev.bar_va[BAR0_IDX], gdev.bar_len[BAR0_IDX]);<br />+        gdev.bar_va[BAR0_IDX] = NULL;<br />+    }<br />+<br />+    if (gdev.bar_va[BAR2_IDX] != NULL) {<br />+        munmap(gdev.bar_va[BAR2_IDX], gdev.bar_len[BAR2_IDX]);<br />+        gdev.bar_va[BAR2_IDX] = NULL;<br />+    }<br />+}<br />+<br />+static int<br />+zxdh_gdma_pci_mse_en(struct zxdh_pci_dev *dev)<br />+{<br />+    int ret = 0;<br />+    uint RegVal = 0;<br />+<br />+    ret = zxdh_gdma_cfg_space_read(dev, PARA_PCI_CMD, &RegVal);<br />+    if (ret != 0) {<br />+        ZXDH_PMD_LOG(ERR, "Failed to read %04x:%02x:%02x.%01x pci config space!",<br />+                        dev->domain, dev->bus, dev->devid, dev->function);<br />+        return ret;<br />+    }<br />+<br />+    if ((RegVal & (1 << PCI_CMD_MSE_BIT)) == 0) {<br />+        RegVal = RegVal | (1 << PCI_CMD_MSE_BIT);<br />+<br />+        ret = zxdh_gdma_cfg_space_write(dev, PARA_PCI_CMD, &RegVal);<br />+        if (ret != 0) {<br />+            ZXDH_PMD_LOG(ERR, "Failed to write %04x:%02x:%02x.%01x pci config space!",<br />+                            dev->domain, dev->bus,<br />+                            dev->devid, dev->function);<br />+            return ret;<br />+        }<br />+    }<br />+<br />+    return ret;<br />+}<br />+<br />+int<br />+zxdh_gdma_pci_scan(void)<br />+{<br />+    struct dirent *e = NULL;<br />+    DIR *dir = NULL;<br />+    char dirname[FILE_PATH_LEN] = {0};<br />+    char filename[FILE_PATH_LEN] = {0};<br />+    uint16_t vendor_id = 0;<br />+    uint16_t device_id = 0;<br />+    unsigned long tmp = 0;<br />+    bool found = false;<br />+    int ret = 0;<br />+<br />+    dir = opendir(PCI_DEVICES_DIR);<br />+    if (dir == NULL) {<br />+        ZXDH_PMD_LOG(ERR, "Failed to opendir %s", PCI_DEVICES_DIR);<br />+        return -1;<br />+    }<br />+<br />+    while ((e = readdir(dir)) != NULL) {<br />+        if (e->d_name[0] == '.')<br />+            continue;<br />+<br />+        memset(dirname, 0, FILE_PATH_LEN);<br />+        snprintf(dirname, FILE_PATH_LEN, "%s/%s", PCI_DEVICES_DIR, e->d_name);<br />+<br />+        snprintf(filename, sizeof(filename), "%s/vendor", dirname);<br />+        ret = zxdh_gdma_devfs_parse(filename, &tmp);<br />+        if (ret != 0)<br />+            goto out;<br />+<br />+        vendor_id = (uint16_t)tmp;<br />+<br />+        snprintf(filename, sizeof(filename), "%s/device", dirname);<br />+        ret = zxdh_gdma_devfs_parse(filename, &tmp);<br />+        if (ret != 0)<br />+            goto out;<br />+<br />+        device_id = (uint16_t)tmp;<br />+<br />+        if ((vendor_id == ZXDH_GDMA_VENDORID) && (device_id == ZXDH_GDMA_DEVICEID)) {<br />+            found = true;<br />+            break;<br />+        }<br />+    }<br />+<br />+    if (found != true) {<br />+        ZXDH_PMD_LOG(ERR, "Failed to found gdma pci dev");<br />+        ret = -1;<br />+        goto out;<br />+    }<br />+<br />+    gdev.vendor_id = vendor_id;<br />+    gdev.device_id = device_id;<br />+    memcpy(gdev.d_name, e->d_name, PCI_BUFF_LEN);<br />+    memcpy(gdev.dirname, dirname, FILE_PATH_LEN);<br />+    ZXDH_PMD_LOG(INFO, "Found gdma pci dev %s", e->d_name);<br />+<br />+    /* Parse pci addr */<br />+    ret = zxdh_gdma_pci_addr_parse(e->d_name, sizeof(e->d_name), &gdev);<br />+    if (ret != 0)<br />+        goto out;<br />+<br />+    /* Enable MSE */<br />+    ret = zxdh_gdma_pci_mse_en(&gdev);<br />+    if (ret != 0)<br />+        goto out;<br />+<br />+    /* Get bar0 phyaddr and len */<br />+    snprintf(filename, sizeof(filename), "%s/resource", dirname);<br />+    ret = zxdh_gdma_resfs_parse(filename, &gdev);<br />+    if (ret != 0)<br />+        goto out;<br />+<br />+    /* Mmap bar0 virtaddr */<br />+    snprintf(filename, sizeof(filename), "%s/resource0", dirname);<br />+    ret = zxdh_gdma_pci_dev_mmap(filename, &gdev, BAR0_IDX);<br />+    if (ret != 0)<br />+        goto out;<br />+<br />+    ZXDH_PMD_LOG(INFO, "Found pci_scan success");<br />+<br />+out:<br />+    closedir(dir);<br />+    return ret;<br />+}<br />diff --git a/drivers/raw/zxdh/zxdh_pci.h b/drivers/raw/zxdh/zxdh_pci.h<br />new file mode 100644<br />index 0000000000..6ecaaeaa2f<br />--- /dev/null<br />+++ b/drivers/raw/zxdh/zxdh_pci.h<br />@@ -0,0 +1,41 @@<br />+/* SPDX-License-Identifier: BSD-3-Clause<br />+ * Copyright 2024 ZTE Corporation<br />+ */<br />+<br />+#ifndef __ZXDH_PCI_H__<br />+#define __ZXDH_PCI_H__<br />+<br />+#ifdef __cplusplus<br />+extern "C" {<br />+#endif<br />+<br />+#include <rte_pci.h> <br />+<br />+#define FILE_PATH_LEN                       (100)<br />+#define PCI_BUFF_LEN                        (16)<br />+<br />+struct zxdh_pci_dev {<br />+    uint16_t    vendor_id;<br />+    uint16_t    device_id;<br />+    uint16_t    domain;<br />+    uint8_t     bus;<br />+    uint8_t     devid;<br />+    uint8_t     function;<br />+    char        dirname[FILE_PATH_LEN];<br />+    char        d_name[PCI_BUFF_LEN];<br />+    void       *bar_va[PCI_MAX_RESOURCE];<br />+    uint64_t    bar_pa[PCI_MAX_RESOURCE];<br />+    uint64_t    bar_len[PCI_MAX_RESOURCE];<br />+};<br />+<br />+extern struct zxdh_pci_dev gdev;<br />+<br />+void zxdh_gdma_pci_dev_munmap(void);<br />+int zxdh_gdma_pci_scan(void);<br />+<br />+#ifdef __cplusplus<br />+}<br />+#endif<br />+<br />+#endif /* __ZXDH_PCI_H__ */<br />+<br />diff --git a/drivers/raw/zxdh/zxdh_rawdev.c b/drivers/raw/zxdh/zxdh_rawdev.c<br />new file mode 100644<br />index 0000000000..07d2784468<br />--- /dev/null<br />+++ b/drivers/raw/zxdh/zxdh_rawdev.c<br />@@ -0,0 +1,915 @@<br />+/* SPDX-License-Identifier: BSD-3-Clause<br />+ * Copyright 2024 ZTE Corporation<br />+ */<br />+<br />+#include <assert.h> <br />+#include <stdio.h> <br />+#include <stdbool.h> <br />+#include <errno.h> <br />+#include <stdint.h> <br />+#include <inttypes.h> <br />+#include <string.h> <br />+#include <time.h> <br />+#include <sys/types.h> <br />+<br />+#include <rte_byteorder.h> <br />+#include <rte_errno.h> <br />+#include <rte_common.h> <br />+#include <rte_debug.h> <br />+#include <rte_dev.h> <br />+#include <rte_eal.h> <br />+#include <rte_kvargs.h> <br />+#include <rte_log.h> <br />+#include <rte_malloc.h> <br />+#include <rte_memory.h> <br />+#include <rte_memcpy.h> <br />+#include <rte_lcore.h> <br />+#include <rte_cycles.h> <br />+#include <rte_memzone.h> <br />+#include <rte_atomic.h> <br />+#include <rte_rawdev.h> <br />+#include <rte_rawdev_pmd.h> <br />+#include <rte_pci.h> <br />+#include <bus_pci_driver.h> <br />+<br />+#include "zxdh_rawdev.h" <br />+#include "zxdh_pci.h" <br />+<br />+/*<br />+ * User define:<br />+ * ep_id-bit[15:12] vfunc_num-bit[11:4] func_num-bit[3:1] vfunc_active-bit0<br />+ * host ep_id:5~8   zf ep_id:9<br />+ */<br />+#define ZXDH_GDMA_ZF_USER                       0x9000      /* ep4 pf0 */<br />+#define ZXDH_GDMA_PF_NUM_SHIFT                  1<br />+#define ZXDH_GDMA_VF_NUM_SHIFT                  4<br />+#define ZXDH_GDMA_EP_ID_SHIFT                   12<br />+#define ZXDH_GDMA_VF_EN                         1<br />+#define ZXDH_GDMA_EPID_OFFSET                   5<br />+<br />+/* Register offset */<br />+#define ZXDH_GDMA_BASE_OFFSET                   0x100000<br />+#define ZXDH_GDMA_EXT_ADDR_OFFSET               0x218<br />+#define ZXDH_GDMA_SAR_LOW_OFFSET                0x200<br />+#define ZXDH_GDMA_DAR_LOW_OFFSET                0x204<br />+#define ZXDH_GDMA_SAR_HIGH_OFFSET               0x234<br />+#define ZXDH_GDMA_DAR_HIGH_OFFSET               0x238<br />+#define ZXDH_GDMA_XFERSIZE_OFFSET               0x208<br />+#define ZXDH_GDMA_CONTROL_OFFSET                0x230<br />+#define ZXDH_GDMA_TC_STATUS_OFFSET              0x0<br />+#define ZXDH_GDMA_STATUS_CLEAN_OFFSET           0x80<br />+#define ZXDH_GDMA_LLI_L_OFFSET                  0x21c<br />+#define ZXDH_GDMA_LLI_H_OFFSET                  0x220<br />+#define ZXDH_GDMA_CHAN_CONTINUE_OFFSET          0x224<br />+#define ZXDH_GDMA_TC_CNT_OFFSET                 0x23c<br />+#define ZXDH_GDMA_LLI_USER_OFFSET               0x228<br />+<br />+/* Control register */<br />+#define ZXDH_GDMA_CHAN_ENABLE                   0x1<br />+#define ZXDH_GDMA_CHAN_DISABLE                  0<br />+#define ZXDH_GDMA_SOFT_CHAN                     0x2<br />+#define ZXDH_GDMA_TC_INTR_ENABLE                0x10<br />+#define ZXDH_GDMA_ALL_INTR_ENABLE               0x30<br />+#define ZXDH_GDMA_SBS_SHIFT                     6           /* src burst size */<br />+#define ZXDH_GDMA_SBL_SHIFT                     9           /* src burst length */<br />+#define ZXDH_GDMA_DBS_SHIFT                     13          /* dest burst size */<br />+#define ZXDH_GDMA_BURST_SIZE_MIN                0x1         /* 1 byte */<br />+#define ZXDH_GDMA_BURST_SIZE_MEDIUM             0x4         /* 4 word */<br />+#define ZXDH_GDMA_BURST_SIZE_MAX                0x6         /* 16 word */<br />+#define ZXDH_GDMA_DEFAULT_BURST_LEN             0xf         /* 16 beats */<br />+#define ZXDH_GDMA_TC_CNT_ENABLE                 (1 << 27)<br />+#define ZXDH_GDMA_CHAN_FORCE_CLOSE              (1 << 31)<br />+<br />+/* TC count & Error interrupt status register */<br />+#define ZXDH_GDMA_SRC_LLI_ERR                   (1 << 16)<br />+#define ZXDH_GDMA_SRC_DATA_ERR                  (1 << 17)<br />+#define ZXDH_GDMA_DST_ADDR_ERR                  (1 << 18)<br />+#define ZXDH_GDMA_ERR_STATUS                    (1 << 19)<br />+#define ZXDH_GDMA_ERR_INTR_ENABLE               (1 << 20)<br />+#define ZXDH_GDMA_TC_CNT_CLEAN                  (1)<br />+<br />+#define ZXDH_GDMA_CHAN_SHIFT                    0x80<br />+#define ZXDH_GDMA_LINK_END_NODE                 (1 << 30)<br />+#define ZXDH_GDMA_CHAN_CONTINUE                 (1)<br />+<br />+#define LOW32_MASK                              0xffffffff<br />+#define LOW16_MASK                              0xffff<br />+<br />+#define ZXDH_GDMA_WAIT_TIMES_MAX                100<br />+#define ZXDH_GDMA_TC_CNT_MAX                    0x10000<br />+<br />+#define IDX_TO_ADDR(addr, idx, t) \<br />+    ((t)((uint8_t *)(addr) + (idx) * sizeof(struct zxdh_gdma_buff_desc)))<br />+<br />+static int zxdh_gdma_queue_init(struct rte_rawdev *dev, uint16_t queue_id);<br />+static int zxdh_gdma_queue_free(struct rte_rawdev *dev, uint16_t queue_id);<br />+<br />+char zxdh_gdma_driver_name[] = "rawdev_zxdh_gdma";<br />+char dev_name[] = "zxdh_gdma";<br />+<br />+static inline struct zxdh_gdma_queue *<br />+zxdh_gdma_get_queue(struct rte_rawdev *dev, uint16_t queue_id)<br />+{<br />+    struct zxdh_gdma_rawdev *gdmadev = zxdh_gdma_rawdev_get_priv(dev);<br />+<br />+    if (queue_id >= ZXDH_GDMA_TOTAL_CHAN_NUM) {<br />+        ZXDH_PMD_LOG(ERR, "queue id %d is invalid", queue_id);<br />+        return NULL;<br />+    }<br />+<br />+    return &(gdmadev->vqs[queue_id]);<br />+}<br />+<br />+uint<br />+zxdh_gdma_read_reg(struct rte_rawdev *dev, uint16_t queue_id, uint offset)<br />+{<br />+    struct zxdh_gdma_rawdev *gdmadev = zxdh_gdma_rawdev_get_priv(dev);<br />+    uint addr = 0;<br />+    uint val = 0;<br />+<br />+    addr = offset + queue_id * ZXDH_GDMA_CHAN_SHIFT;<br />+    val = *(uint *)(gdmadev->base_addr + addr);<br />+<br />+    return val;<br />+}<br />+<br />+void<br />+zxdh_gdma_write_reg(struct rte_rawdev *dev, uint16_t queue_id, uint offset, uint val)<br />+{<br />+    struct zxdh_gdma_rawdev *gdmadev = zxdh_gdma_rawdev_get_priv(dev);<br />+    uint addr = 0;<br />+<br />+    addr = offset + queue_id * ZXDH_GDMA_CHAN_SHIFT;<br />+    *(uint *)(gdmadev->base_addr + addr) = val;<br />+}<br />+<br />+int<br />+zxdh_gdma_debug_info_dump(struct rte_rawdev *dev, uint16_t queue_id)<br />+{<br />+    struct zxdh_gdma_queue *queue = NULL;<br />+    struct zxdh_gdma_buff_desc *bd = NULL;<br />+    struct zxdh_gdma_job *job = NULL;<br />+    uint16_t i = 0;<br />+<br />+    if (dev == NULL)<br />+        return -EINVAL;<br />+<br />+    queue = zxdh_gdma_get_queue(dev, queue_id);<br />+    if (queue == NULL)<br />+        return -EINVAL;<br />+<br />+    ZXDH_PMD_LOG(INFO, "###dump sw_ring info###");<br />+    ZXDH_PMD_LOG(INFO, "free_cnt:%u deq_cnt:%u",<br />+                        queue->sw_ring.free_cnt, queue->sw_ring.deq_cnt);<br />+    ZXDH_PMD_LOG(INFO, "enq_idx:%u deq_idx:%u used_idx:%u",<br />+                        queue->sw_ring.enq_idx, queue->sw_ring.deq_idx,<br />+                        queue->sw_ring.used_idx);<br />+    for (i = 0; i < ZXDH_GDMA_QUEUE_SIZE; i++) {<br />+        if (queue->sw_ring.job[i] != NULL) {<br />+            job = queue->sw_ring.job[i];<br />+            ZXDH_PMD_LOG(INFO, "idx:%d, SrcAddr:0x%"PRIx64" DstAddr:0x%"PRIx64" len:%u",<br />+                                i, job->src, job->dest, job->len);<br />+        }<br />+    }<br />+<br />+    ZXDH_PMD_LOG(INFO, "###dump ring info###");<br />+    ZXDH_PMD_LOG(INFO, "avail_idx:%u tc_cnt:%u", queue->ring.avail_idx, queue->tc_cnt);<br />+    for (i = 0; i < ZXDH_GDMA_RING_SIZE; i++) {<br />+        bd = IDX_TO_ADDR(queue->ring.desc, i, struct zxdh_gdma_buff_desc*);<br />+        ZXDH_PMD_LOG(INFO, "idx:%d Src:0x%"PRIx64" Dst:0x%"PRIx64" LLI_L:0x%x LLI_H:0x%x ctrl:0x%x user:0x%x",<br />+                i, bd->SrcAddr_L | ((uint64_t)bd->SrcAddr_H << 32),<br />+                bd->DstAddr_L | ((uint64_t)bd->DstAddr_H << 32),<br />+                bd->LLI_Addr_L, bd->LLI_Addr_H, bd->Control, bd->ExtAddr);<br />+    }<br />+<br />+    return 0;<br />+}<br />+<br />+static int<br />+zxdh_gdma_rawdev_info_get(struct rte_rawdev *dev,<br />+                          __rte_unused rte_rawdev_obj_t dev_info,<br />+                          __rte_unused size_t dev_info_size)<br />+{<br />+    if (dev == NULL)<br />+        return -EINVAL;<br />+<br />+    return 0;<br />+}<br />+<br />+static int<br />+zxdh_gdma_rawdev_configure(const struct rte_rawdev *dev,<br />+                           rte_rawdev_obj_t config,<br />+                           size_t config_size)<br />+{<br />+    struct zxdh_gdma_config *gdma_config = NULL;<br />+<br />+    if ((dev == NULL) ||<br />+        (config == NULL) ||<br />+        (config_size != sizeof(struct zxdh_gdma_config)))<br />+        return -EINVAL;<br />+<br />+    gdma_config = (struct zxdh_gdma_config *)config;<br />+    if (gdma_config->max_vqs != ZXDH_GDMA_TOTAL_CHAN_NUM) {<br />+        ZXDH_PMD_LOG(ERR, "gdma only support queue num %d", ZXDH_GDMA_TOTAL_CHAN_NUM);<br />+        return -EINVAL;<br />+    }<br />+<br />+    return 0;<br />+}<br />+<br />+static int<br />+zxdh_gdma_rawdev_start(struct rte_rawdev *dev)<br />+{<br />+    struct zxdh_gdma_rawdev *gdmadev = NULL;<br />+<br />+    if (dev == NULL)<br />+        return -EINVAL;<br />+<br />+    gdmadev = zxdh_gdma_rawdev_get_priv(dev);<br />+    gdmadev->device_state = ZXDH_GDMA_DEV_RUNNING;<br />+<br />+    return 0;<br />+}<br />+<br />+static void<br />+zxdh_gdma_rawdev_stop(struct rte_rawdev *dev)<br />+{<br />+    struct zxdh_gdma_rawdev *gdmadev = NULL;<br />+<br />+    if (dev == NULL)<br />+        return;<br />+<br />+    gdmadev = zxdh_gdma_rawdev_get_priv(dev);<br />+    gdmadev->device_state = ZXDH_GDMA_DEV_STOPPED;<br />+}<br />+<br />+static int<br />+zxdh_gdma_rawdev_reset(struct rte_rawdev *dev)<br />+{<br />+    if (dev == NULL)<br />+        return -EINVAL;<br />+<br />+    return 0;<br />+}<br />+<br />+static int<br />+zxdh_gdma_rawdev_close(struct rte_rawdev *dev)<br />+{<br />+    struct zxdh_gdma_rawdev *gdmadev = NULL;<br />+    struct zxdh_gdma_queue *queue = NULL;<br />+    uint16_t queue_id = 0;<br />+<br />+    if (dev == NULL)<br />+        return -EINVAL;<br />+<br />+    for (queue_id = 0; queue_id < ZXDH_GDMA_TOTAL_CHAN_NUM; queue_id++) {<br />+        queue = zxdh_gdma_get_queue(dev, queue_id);<br />+        if ((queue == NULL) || (queue->enable == 0))<br />+            continue;<br />+<br />+        zxdh_gdma_queue_free(dev, queue_id);<br />+    }<br />+    gdmadev = zxdh_gdma_rawdev_get_priv(dev);<br />+    gdmadev->device_state = ZXDH_GDMA_DEV_STOPPED;<br />+<br />+    return 0;<br />+}<br />+<br />+static int<br />+zxdh_gdma_rawdev_queue_setup(struct rte_rawdev *dev,<br />+                             uint16_t queue_id,<br />+                             rte_rawdev_obj_t queue_conf,<br />+                             size_t conf_size)<br />+{<br />+    struct zxdh_gdma_rawdev *gdmadev = NULL;<br />+    struct zxdh_gdma_queue *queue = NULL;<br />+    struct zxdh_gdma_queue_config *qconfig = NULL;<br />+    struct zxdh_gdma_rbp *rbp = NULL;<br />+    uint16_t i = 0;<br />+    uint8_t is_txq = 0;<br />+    uint src_user = 0;<br />+    uint dst_user = 0;<br />+<br />+    if (dev == NULL)<br />+        return -EINVAL;<br />+<br />+    if ((queue_conf == NULL) || (conf_size != sizeof(struct zxdh_gdma_queue_config)))<br />+        return -EINVAL;<br />+<br />+    gdmadev = zxdh_gdma_rawdev_get_priv(dev);<br />+    qconfig = (struct zxdh_gdma_queue_config *)queue_conf;<br />+<br />+    for (i = 0; i < ZXDH_GDMA_TOTAL_CHAN_NUM; i++) {<br />+        if (gdmadev->vqs[i].enable == 0)<br />+            break;<br />+    }<br />+    if (i >= ZXDH_GDMA_TOTAL_CHAN_NUM) {<br />+        ZXDH_PMD_LOG(ERR, "Failed to setup queue, no avail queues");<br />+        return -1;<br />+    }<br />+    queue_id = i;<br />+    if (zxdh_gdma_queue_init(dev, queue_id) != 0) {<br />+        ZXDH_PMD_LOG(ERR, "Failed to init queue");<br />+        return -1;<br />+    }<br />+    queue = &(gdmadev->vqs[queue_id]);<br />+<br />+    rbp = qconfig->rbp;<br />+    if ((rbp->srbp != 0) && (rbp->drbp == 0)) {<br />+        is_txq = 0;<br />+        dst_user = ZXDH_GDMA_ZF_USER;<br />+        src_user = ((rbp->spfid << ZXDH_GDMA_PF_NUM_SHIFT) |<br />+            ((rbp->sportid + ZXDH_GDMA_EPID_OFFSET) << ZXDH_GDMA_EP_ID_SHIFT));<br />+<br />+        if (rbp->svfid != 0)<br />+            src_user |= (ZXDH_GDMA_VF_EN |<br />+                         ((rbp->svfid - 1) << ZXDH_GDMA_VF_NUM_SHIFT));<br />+<br />+        ZXDH_PMD_LOG(DEBUG, "rxq->qidx:%d setup src_user(ep:%d pf:%d vf:%d) success",<br />+                    queue_id, (uint8_t)rbp->sportid, (uint8_t)rbp->spfid,<br />+                    (uint8_t)rbp->svfid);<br />+    } else if ((rbp->srbp == 0) && (rbp->drbp != 0)) {<br />+        is_txq = 1;<br />+        src_user = ZXDH_GDMA_ZF_USER;<br />+        dst_user = ((rbp->dpfid << ZXDH_GDMA_PF_NUM_SHIFT) |<br />+            ((rbp->dportid + ZXDH_GDMA_EPID_OFFSET) << ZXDH_GDMA_EP_ID_SHIFT));<br />+<br />+        if (rbp->dvfid != 0)<br />+            dst_user |= (ZXDH_GDMA_VF_EN |<br />+                         ((rbp->dvfid - 1) << ZXDH_GDMA_VF_NUM_SHIFT));<br />+<br />+        ZXDH_PMD_LOG(DEBUG, "txq->qidx:%d setup dst_user(ep:%d pf:%d vf:%d) success",<br />+                    queue_id, (uint8_t)rbp->dportid, (uint8_t)rbp->dpfid,<br />+                    (uint8_t)rbp->dvfid);<br />+    } else {<br />+        ZXDH_PMD_LOG(ERR, "Failed to setup queue, srbp/drbp is invalid");<br />+        return -EINVAL;<br />+    }<br />+    queue->is_txq = is_txq;<br />+<br />+    /* setup queue user info */<br />+    queue->user = (src_user & LOW16_MASK) | (dst_user << 16);<br />+<br />+    zxdh_gdma_write_reg(dev, queue_id, ZXDH_GDMA_EXT_ADDR_OFFSET, queue->user);<br />+    gdmadev->used_num++;<br />+<br />+    return queue_id;<br />+}<br />+<br />+static int<br />+zxdh_gdma_rawdev_queue_release(struct rte_rawdev *dev, uint16_t queue_id)<br />+{<br />+    struct zxdh_gdma_queue *queue = NULL;<br />+<br />+    if (dev == NULL)<br />+        return -EINVAL;<br />+<br />+    queue = zxdh_gdma_get_queue(dev, queue_id);<br />+    if ((queue == NULL) || (queue->enable == 0))<br />+        return -EINVAL;<br />+<br />+    zxdh_gdma_queue_free(dev, queue_id);<br />+<br />+    return 0;<br />+}<br />+<br />+static int<br />+zxdh_gdma_rawdev_get_attr(struct rte_rawdev *dev,<br />+                          __rte_unused const char *attr_name,<br />+                          uint64_t *attr_value)<br />+{<br />+    struct zxdh_gdma_rawdev *gdmadev = NULL;<br />+    struct zxdh_gdma_attr *gdma_attr = NULL;<br />+<br />+    if ((dev == NULL) || (attr_value == NULL))<br />+        return -EINVAL;<br />+<br />+    gdmadev   = zxdh_gdma_rawdev_get_priv(dev);<br />+    gdma_attr = (struct zxdh_gdma_attr *)attr_value;<br />+    gdma_attr->num_hw_queues = gdmadev->used_num;<br />+<br />+    return 0;<br />+}<br />+<br />+static inline void<br />+zxdh_gdma_control_cal(uint *val, uint8_t tc_enable)<br />+{<br />+    *val = (ZXDH_GDMA_CHAN_ENABLE |<br />+            ZXDH_GDMA_SOFT_CHAN |<br />+            (ZXDH_GDMA_DEFAULT_BURST_LEN << ZXDH_GDMA_SBL_SHIFT) |<br />+            (ZXDH_GDMA_BURST_SIZE_MAX << ZXDH_GDMA_SBS_SHIFT) |<br />+            (ZXDH_GDMA_BURST_SIZE_MAX << ZXDH_GDMA_DBS_SHIFT));<br />+<br />+    if (tc_enable != 0)<br />+        *val |= ZXDH_GDMA_TC_CNT_ENABLE;<br />+}<br />+<br />+static inline uint<br />+zxdh_gdma_user_get(struct zxdh_gdma_queue *queue, struct zxdh_gdma_job *job)<br />+{<br />+    uint src_user = 0;<br />+    uint dst_user = 0;<br />+<br />+    if ((job->flags & ZXDH_GDMA_JOB_DIR_MASK) == 0) {<br />+        ZXDH_PMD_LOG(DEBUG, "job flags:0x%x default user:0x%x",<br />+                            job->flags, queue->user);<br />+        return queue->user;<br />+    } else if ((job->flags & ZXDH_GDMA_JOB_DIR_TX) != 0) {<br />+        src_user = ZXDH_GDMA_ZF_USER;<br />+        dst_user = ((job->pf_id << ZXDH_GDMA_PF_NUM_SHIFT) |<br />+            ((job->ep_id + ZXDH_GDMA_EPID_OFFSET) << ZXDH_GDMA_EP_ID_SHIFT));<br />+<br />+        if (job->vf_id != 0)<br />+            dst_user |= (ZXDH_GDMA_VF_EN |<br />+                         ((job->vf_id - 1) << ZXDH_GDMA_VF_NUM_SHIFT));<br />+    } else {<br />+        dst_user = ZXDH_GDMA_ZF_USER;<br />+        src_user = ((job->pf_id << ZXDH_GDMA_PF_NUM_SHIFT) |<br />+            ((job->ep_id + ZXDH_GDMA_EPID_OFFSET) << ZXDH_GDMA_EP_ID_SHIFT));<br />+<br />+        if (job->vf_id != 0)<br />+            src_user |= (ZXDH_GDMA_VF_EN |<br />+                         ((job->vf_id - 1) << ZXDH_GDMA_VF_NUM_SHIFT));<br />+    }<br />+    ZXDH_PMD_LOG(DEBUG, "job flags:0x%x ep_id:%u, pf_id:%u, vf_id:%u, user:0x%x",<br />+                        job->flags, job->ep_id, job->pf_id, job->vf_id,<br />+                        (src_user & LOW16_MASK) | (dst_user << 16));<br />+<br />+    return (src_user & LOW16_MASK) | (dst_user << 16);<br />+}<br />+<br />+static inline void<br />+zxdh_gdma_fill_bd(struct zxdh_gdma_queue *queue, struct zxdh_gdma_job *job)<br />+{<br />+    struct zxdh_gdma_buff_desc *bd = NULL;<br />+    uint val = 0;<br />+    uint64_t next_bd_addr = 0;<br />+    uint16_t avail_idx = 0;<br />+<br />+    avail_idx = queue->ring.avail_idx;<br />+    bd = &(queue->ring.desc[avail_idx]);<br />+    memset(bd, 0, sizeof(struct zxdh_gdma_buff_desc));<br />+<br />+    /* data bd */<br />+    if (job != NULL) {<br />+        zxdh_gdma_control_cal(&val, 1);<br />+        next_bd_addr   = IDX_TO_ADDR(queue->ring.ring_mem,<br />+                            (avail_idx + 1) % ZXDH_GDMA_RING_SIZE,<br />+                            uint64_t);<br />+        bd->SrcAddr_L  = job->src & LOW32_MASK;<br />+        bd->DstAddr_L  = job->dest & LOW32_MASK;<br />+        bd->SrcAddr_H  = (job->src >> 32) & LOW32_MASK;<br />+        bd->DstAddr_H  = (job->dest >> 32) & LOW32_MASK;<br />+        bd->Xpara      = job->len;<br />+        bd->ExtAddr    = zxdh_gdma_user_get(queue, job);<br />+        bd->LLI_Addr_L = (next_bd_addr >> 6) & LOW32_MASK;<br />+        bd->LLI_Addr_H = next_bd_addr >> 38;<br />+        bd->LLI_User   = ZXDH_GDMA_ZF_USER;<br />+        bd->Control    = val;<br />+    } else {<br />+        zxdh_gdma_control_cal(&val, 0);<br />+        next_bd_addr   = IDX_TO_ADDR(queue->ring.ring_mem, avail_idx, uint64_t);<br />+        bd->ExtAddr    = queue->user;<br />+        bd->LLI_User   = ZXDH_GDMA_ZF_USER;<br />+        bd->Control    = val;<br />+        bd->LLI_Addr_L = (next_bd_addr >> 6) & LOW32_MASK;<br />+        bd->LLI_Addr_H = (next_bd_addr >> 38) | ZXDH_GDMA_LINK_END_NODE;<br />+        if (queue->flag != 0) {<br />+            bd = IDX_TO_ADDR(queue->ring.desc,<br />+                    queue->ring.last_avail_idx,<br />+                    struct zxdh_gdma_buff_desc*);<br />+            next_bd_addr = IDX_TO_ADDR(queue->ring.ring_mem,<br />+                    (queue->ring.last_avail_idx + 1) % ZXDH_GDMA_RING_SIZE,<br />+                    uint64_t);<br />+            bd->LLI_Addr_L  = (next_bd_addr >> 6) & LOW32_MASK;<br />+            bd->LLI_Addr_H  = next_bd_addr >> 38;<br />+            rte_wmb();<br />+            bd->LLI_Addr_H &= ~ZXDH_GDMA_LINK_END_NODE;<br />+        }<br />+        /* Record the index of empty bd for dynamic chaining */<br />+        queue->ring.last_avail_idx = avail_idx;<br />+    }<br />+<br />+    if (++avail_idx >= ZXDH_GDMA_RING_SIZE)<br />+        avail_idx -= ZXDH_GDMA_RING_SIZE;<br />+<br />+    queue->ring.avail_idx = avail_idx;<br />+}<br />+<br />+static int<br />+zxdh_gdma_rawdev_enqueue_bufs(struct rte_rawdev *dev,<br />+                        __rte_unused struct rte_rawdev_buf **buffers,<br />+                        uint count,<br />+                        rte_rawdev_obj_t context)<br />+{<br />+    struct zxdh_gdma_rawdev *gdmadev = NULL;<br />+    struct zxdh_gdma_queue *queue = NULL;<br />+    struct zxdh_gdma_enqdeq *e_context = NULL;<br />+    struct zxdh_gdma_job *job = NULL;<br />+    uint16_t queue_id = 0;<br />+    uint val = 0;<br />+    uint16_t i = 0;<br />+    uint16_t free_cnt = 0;<br />+<br />+    if (dev == NULL)<br />+        return -EINVAL;<br />+<br />+    if (unlikely((count < 1) || (context == NULL)))<br />+        return -EINVAL;<br />+<br />+    gdmadev = zxdh_gdma_rawdev_get_priv(dev);<br />+    if (gdmadev->device_state == ZXDH_GDMA_DEV_STOPPED) {<br />+        ZXDH_PMD_LOG(ERR, "gdma dev is stop");<br />+        return 0;<br />+    }<br />+<br />+    e_context = (struct zxdh_gdma_enqdeq *)context;<br />+    queue_id = e_context->vq_id;<br />+    queue = zxdh_gdma_get_queue(dev, queue_id);<br />+    if ((queue == NULL) || (queue->enable == 0))<br />+        return -EINVAL;<br />+<br />+    free_cnt = queue->sw_ring.free_cnt;<br />+    if (free_cnt == 0) {<br />+        ZXDH_PMD_LOG(ERR, "queue %u is full, enq_idx:%u deq_idx:%u used_idx:%u",<br />+                           queue_id, queue->sw_ring.enq_idx,<br />+                           queue->sw_ring.deq_idx, queue->sw_ring.used_idx);<br />+        return 0;<br />+    } else if (free_cnt < count) {<br />+        ZXDH_PMD_LOG(DEBUG, "job num %u > free_cnt, change to %u", count, free_cnt);<br />+        count = free_cnt;<br />+    }<br />+<br />+    rte_spinlock_lock(&queue->enqueue_lock);<br />+<br />+    /* Build bd list, the last bd is empty bd */<br />+    for (i = 0; i < count; i++) {<br />+        job = e_context->job[i];<br />+        zxdh_gdma_fill_bd(queue, job);<br />+    }<br />+    zxdh_gdma_fill_bd(queue, NULL);<br />+<br />+    if (unlikely(queue->flag == 0)) {<br />+        zxdh_gdma_write_reg(dev, queue_id, ZXDH_GDMA_LLI_L_OFFSET,<br />+                            (queue->ring.ring_mem >> 6) & LOW32_MASK);<br />+        zxdh_gdma_write_reg(dev, queue_id, ZXDH_GDMA_LLI_H_OFFSET,<br />+                             queue->ring.ring_mem >> 38);<br />+        /* Start hardware handling */<br />+        zxdh_gdma_write_reg(dev, queue_id, ZXDH_GDMA_XFERSIZE_OFFSET, 0);<br />+        zxdh_gdma_control_cal(&val, 0);<br />+        zxdh_gdma_write_reg(dev, queue_id, ZXDH_GDMA_CONTROL_OFFSET, val);<br />+        queue->flag = 1;<br />+    } else {<br />+        val = ZXDH_GDMA_CHAN_CONTINUE;<br />+        zxdh_gdma_write_reg(dev, queue->vq_id, ZXDH_GDMA_CHAN_CONTINUE_OFFSET, val);<br />+    }<br />+<br />+    /* job enqueue */<br />+    for (i = 0; i < count; i++) {<br />+        queue->sw_ring.job[queue->sw_ring.enq_idx] = e_context->job[i];<br />+        if (++queue->sw_ring.enq_idx >= queue->queue_size)<br />+            queue->sw_ring.enq_idx -= queue->queue_size;<br />+<br />+        free_cnt--;<br />+    }<br />+    queue->sw_ring.free_cnt = free_cnt;<br />+    queue->sw_ring.pend_cnt += count;<br />+    rte_spinlock_unlock(&queue->enqueue_lock);<br />+<br />+    return count;<br />+}<br />+<br />+static inline void<br />+zxdh_gdma_used_idx_update(struct zxdh_gdma_queue *queue, uint16_t cnt, uint8_t data_bd_err)<br />+{<br />+    uint16_t idx = 0;<br />+<br />+    if (queue->sw_ring.used_idx + cnt < queue->queue_size)<br />+        queue->sw_ring.used_idx += cnt;<br />+    else<br />+        queue->sw_ring.used_idx = queue->sw_ring.used_idx + cnt - queue->queue_size;<br />+<br />+    if (data_bd_err == 1) {<br />+        /* Update job status, the last job status is error */<br />+        if (queue->sw_ring.used_idx == 0)<br />+            idx = queue->queue_size - 1;<br />+        else<br />+            idx = queue->sw_ring.used_idx - 1;<br />+<br />+        queue->sw_ring.job[idx]->status = 1;<br />+    }<br />+}<br />+<br />+static int<br />+zxdh_gdma_rawdev_dequeue_bufs(struct rte_rawdev *dev,<br />+                        __rte_unused struct rte_rawdev_buf **buffers,<br />+                        uint count,<br />+                        rte_rawdev_obj_t context)<br />+{<br />+    struct zxdh_gdma_queue *queue = NULL;<br />+    struct zxdh_gdma_enqdeq *e_context = NULL;<br />+    uint16_t queue_id = 0;<br />+    uint val = 0;<br />+    uint16_t tc_cnt = 0;<br />+    uint16_t diff_cnt = 0;<br />+    uint16_t i = 0;<br />+    uint16_t bd_idx = 0;<br />+    uint64_t next_bd_addr = 0;<br />+    uint8_t data_bd_err = 0;<br />+<br />+    if ((dev == NULL) || (context == NULL))<br />+        return -EINVAL;<br />+<br />+    e_context = (struct zxdh_gdma_enqdeq *)context;<br />+    queue_id = e_context->vq_id;<br />+    queue = zxdh_gdma_get_queue(dev, queue_id);<br />+    if ((queue == NULL) || (queue->enable == 0))<br />+        return -EINVAL;<br />+<br />+    if (queue->sw_ring.pend_cnt == 0)<br />+        goto deq_job;<br />+<br />+    /* Get data transmit count */<br />+    val = zxdh_gdma_read_reg(dev, queue_id, ZXDH_GDMA_TC_CNT_OFFSET);<br />+    tc_cnt = val & LOW16_MASK;<br />+    if (tc_cnt >= queue->tc_cnt)<br />+        diff_cnt = tc_cnt - queue->tc_cnt;<br />+    else<br />+        diff_cnt = tc_cnt + ZXDH_GDMA_TC_CNT_MAX - queue->tc_cnt;<br />+<br />+    queue->tc_cnt = tc_cnt;<br />+<br />+    /* Data transmit error, channel stopped */<br />+    if ((val & ZXDH_GDMA_ERR_STATUS) != 0) {<br />+        next_bd_addr  = zxdh_gdma_read_reg(dev, queue_id, ZXDH_GDMA_LLI_L_OFFSET);<br />+        next_bd_addr |= ((uint64_t)zxdh_gdma_read_reg(dev, queue_id,<br />+                            ZXDH_GDMA_LLI_H_OFFSET) << 32);<br />+        next_bd_addr  = next_bd_addr << 6;<br />+        bd_idx = (next_bd_addr - queue->ring.ring_mem) / sizeof(struct zxdh_gdma_buff_desc);<br />+        if ((val & ZXDH_GDMA_SRC_DATA_ERR) || (val & ZXDH_GDMA_DST_ADDR_ERR)) {<br />+            diff_cnt++;<br />+            data_bd_err = 1;<br />+        }<br />+        ZXDH_PMD_LOG(INFO, "queue%d is err(0x%x) next_bd_idx:%u ll_addr:0x%"PRIx64" def user:0x%x",<br />+                    queue_id, val, bd_idx, next_bd_addr, queue->user);<br />+        zxdh_gdma_debug_info_dump(dev, queue_id);<br />+<br />+        ZXDH_PMD_LOG(INFO, "Clean up error status");<br />+        val = ZXDH_GDMA_ERR_STATUS | ZXDH_GDMA_ERR_INTR_ENABLE;<br />+        zxdh_gdma_write_reg(dev, queue_id, ZXDH_GDMA_TC_CNT_OFFSET, val);<br />+<br />+        ZXDH_PMD_LOG(INFO, "Restart channel");<br />+        zxdh_gdma_write_reg(dev, queue_id, ZXDH_GDMA_XFERSIZE_OFFSET, 0);<br />+        zxdh_gdma_control_cal(&val, 0);<br />+        zxdh_gdma_write_reg(dev, queue_id, ZXDH_GDMA_CONTROL_OFFSET, val);<br />+    }<br />+<br />+    if (diff_cnt != 0) {<br />+        zxdh_gdma_used_idx_update(queue, diff_cnt, data_bd_err);<br />+        queue->sw_ring.deq_cnt += diff_cnt;<br />+        queue->sw_ring.pend_cnt -= diff_cnt;<br />+    }<br />+<br />+deq_job:<br />+    if (queue->sw_ring.deq_cnt == 0)<br />+        return 0;<br />+    else if (queue->sw_ring.deq_cnt < count)<br />+        count = queue->sw_ring.deq_cnt;<br />+<br />+    queue->sw_ring.deq_cnt -= count;<br />+<br />+    for (i = 0; i < count; i++) {<br />+        e_context->job[i] = queue->sw_ring.job[queue->sw_ring.deq_idx];<br />+        queue->sw_ring.job[queue->sw_ring.deq_idx] = NULL;<br />+        if (++queue->sw_ring.deq_idx >= queue->queue_size)<br />+            queue->sw_ring.deq_idx -= queue->queue_size;<br />+    }<br />+    queue->sw_ring.free_cnt += count;<br />+<br />+    return count;<br />+}<br />+<br />+static const struct rte_rawdev_ops zxdh_gdma_rawdev_ops = {<br />+    .dev_info_get = zxdh_gdma_rawdev_info_get,<br />+    .dev_configure = zxdh_gdma_rawdev_configure,<br />+    .dev_start = zxdh_gdma_rawdev_start,<br />+    .dev_stop = zxdh_gdma_rawdev_stop,<br />+    .dev_close = zxdh_gdma_rawdev_close,<br />+    .dev_reset = zxdh_gdma_rawdev_reset,<br />+<br />+    .queue_setup = zxdh_gdma_rawdev_queue_setup,<br />+    .queue_release = zxdh_gdma_rawdev_queue_release,<br />+<br />+    .attr_get = zxdh_gdma_rawdev_get_attr,<br />+<br />+    .enqueue_bufs = zxdh_gdma_rawdev_enqueue_bufs,<br />+    .dequeue_bufs = zxdh_gdma_rawdev_dequeue_bufs,<br />+};<br />+<br />+static int<br />+zxdh_gdma_queue_init(struct rte_rawdev *dev, uint16_t queue_id)<br />+{<br />+    char name[RTE_RAWDEV_NAME_MAX_LEN];<br />+    struct zxdh_gdma_queue *queue = NULL;<br />+    const struct rte_memzone *mz = NULL;<br />+    uint size = 0;<br />+    uint val = 0;<br />+    int ret = 0;<br />+<br />+    queue = zxdh_gdma_get_queue(dev, queue_id);<br />+    if (queue == NULL)<br />+        return -EINVAL;<br />+<br />+    queue->enable = 1;<br />+    queue->vq_id  = queue_id;<br />+    queue->flag   = 0;<br />+    queue->tc_cnt = 0;<br />+<br />+    /* Init sw_ring */<br />+    memset(name, 0, sizeof(name));<br />+    snprintf(name, RTE_RAWDEV_NAME_MAX_LEN, "gdma_vq%d_sw_ring", queue_id);<br />+    size = queue->queue_size * sizeof(struct zxdh_gdma_job *);<br />+    queue->sw_ring.job = rte_zmalloc(name, size, 0);<br />+    if (queue->sw_ring.job == NULL) {<br />+        ZXDH_PMD_LOG(ERR, "can not allocate sw_ring %s", name);<br />+        ret = -ENOMEM;<br />+        goto free_queue;<br />+    }<br />+<br />+    /* Cache up to size-1 job in the ring to prevent overwriting hardware prefetching */<br />+    queue->sw_ring.free_cnt = queue->queue_size - 1;<br />+    queue->sw_ring.deq_cnt  = 0;<br />+    queue->sw_ring.pend_cnt = 0;<br />+    queue->sw_ring.enq_idx  = 0;<br />+    queue->sw_ring.deq_idx  = 0;<br />+    queue->sw_ring.used_idx = 0;<br />+<br />+    /* Init ring */<br />+    memset(name, 0, sizeof(name));<br />+    snprintf(name, RTE_RAWDEV_NAME_MAX_LEN, "gdma_vq%d_ring", queue_id);<br />+    size = ZXDH_GDMA_RING_SIZE * sizeof(struct zxdh_gdma_buff_desc);<br />+    mz = rte_memzone_reserve_aligned(name, size, rte_socket_id(),<br />+                            RTE_MEMZONE_IOVA_CONTIG, size);<br />+    if (mz == NULL) {<br />+        if (rte_errno == EEXIST)<br />+            mz = rte_memzone_lookup(name);<br />+        if (mz == NULL) {<br />+            ZXDH_PMD_LOG(ERR, "can not allocate ring %s", name);<br />+            ret = -ENOMEM;<br />+            goto free_queue;<br />+        }<br />+    }<br />+    memset(mz->addr, 0, size);<br />+    queue->ring.ring_mz   = mz;<br />+    queue->ring.desc      = (struct zxdh_gdma_buff_desc *)(mz->addr);<br />+    queue->ring.ring_mem  = mz->iova;<br />+    queue->ring.avail_idx = 0;<br />+    ZXDH_PMD_LOG(INFO, "queue%u ring phy addr:0x%"PRIx64" virt addr:%p",<br />+                        queue_id, mz->iova, mz->addr);<br />+<br />+    /* clean gdma channel */<br />+    val = ZXDH_GDMA_CHAN_FORCE_CLOSE;<br />+    zxdh_gdma_write_reg(dev, queue_id, ZXDH_GDMA_CONTROL_OFFSET, val);<br />+<br />+    val = ZXDH_GDMA_ERR_INTR_ENABLE | ZXDH_GDMA_ERR_STATUS | ZXDH_GDMA_TC_CNT_CLEAN;<br />+    zxdh_gdma_write_reg(dev, queue_id, ZXDH_GDMA_TC_CNT_OFFSET, val);<br />+<br />+    val = ZXDH_GDMA_ZF_USER;<br />+    zxdh_gdma_write_reg(dev, queue_id, ZXDH_GDMA_LLI_USER_OFFSET, val);<br />+<br />+    return 0;<br />+<br />+free_queue:<br />+    zxdh_gdma_queue_free(dev, queue_id);<br />+    return ret;<br />+}<br />+<br />+static int<br />+zxdh_gdma_queue_free(struct rte_rawdev *dev, uint16_t queue_id)<br />+{<br />+    struct zxdh_gdma_rawdev *gdmadev = NULL;<br />+    struct zxdh_gdma_queue *queue = NULL;<br />+    uint val = 0;<br />+<br />+    queue = zxdh_gdma_get_queue(dev, queue_id);<br />+    if (queue == NULL)<br />+        return -EINVAL;<br />+<br />+    gdmadev = zxdh_gdma_rawdev_get_priv(dev);<br />+    gdmadev->used_num--;<br />+<br />+    /* disable gdma channel */<br />+    val = ZXDH_GDMA_CHAN_FORCE_CLOSE;<br />+    zxdh_gdma_write_reg(dev, queue_id, ZXDH_GDMA_CONTROL_OFFSET, val);<br />+<br />+    queue->enable           = 0;<br />+    queue->is_txq           = 0;<br />+    queue->flag             = 0;<br />+    queue->user             = 0;<br />+    queue->tc_cnt           = 0;<br />+    queue->ring.avail_idx   = 0;<br />+    queue->sw_ring.free_cnt = 0;<br />+    queue->sw_ring.deq_cnt  = 0;<br />+    queue->sw_ring.pend_cnt = 0;<br />+    queue->sw_ring.enq_idx  = 0;<br />+    queue->sw_ring.deq_idx  = 0;<br />+    queue->sw_ring.used_idx = 0;<br />+<br />+    if (queue->sw_ring.job != NULL)<br />+        rte_free(queue->sw_ring.job);<br />+<br />+    if (queue->ring.ring_mz != NULL)<br />+        rte_memzone_free(queue->ring.ring_mz);<br />+<br />+    return 0;<br />+}<br />+<br />+static int<br />+zxdh_gdma_rawdev_probe(struct rte_pci_driver *pci_drv __rte_unused,<br />+                        struct rte_pci_device *pci_dev)<br />+{<br />+    struct rte_rawdev *dev = NULL;<br />+    struct zxdh_gdma_rawdev *gdmadev = NULL;<br />+    struct zxdh_gdma_queue *queue = NULL;<br />+    uint8_t i = 0;<br />+<br />+    if (zxdh_gdma_pci_scan() != 0) {<br />+        ZXDH_PMD_LOG(ERR, "Failed to scan gdma pci device!");<br />+        return -1;<br />+    }<br />+<br />+    if ((gdev.bar_pa[0]) == 0) {<br />+        ZXDH_PMD_LOG(ERR, "Empty bars 0x%"PRIx64,<br />+                            (uint64_t)gdev.bar_pa[0]);<br />+        zxdh_gdma_pci_dev_munmap();<br />+        return -ENODEV;<br />+    }<br />+    ZXDH_PMD_LOG(INFO, "%04x:%02x:%02x.%01x Bar0 PhyAddr: 0x%"PRIx64,<br />+                    gdev.domain, gdev.bus, gdev.devid, gdev.function,<br />+                    (uint64_t)gdev.bar_pa[0]);<br />+<br />+    dev = rte_rawdev_pmd_allocate(dev_name, sizeof(struct zxdh_gdma_rawdev), rte_socket_id());<br />+    if (dev == NULL) {<br />+        ZXDH_PMD_LOG(ERR, "Unable to allocate gdma rawdev");<br />+        zxdh_gdma_pci_dev_munmap();<br />+        return -1;<br />+    }<br />+    ZXDH_PMD_LOG(INFO, "Init %s on NUMA node %d, dev_id is %d",<br />+                        dev_name, rte_socket_id(), dev->dev_id);<br />+<br />+    dev->dev_ops = &zxdh_gdma_rawdev_ops;<br />+    dev->device = &pci_dev->device;<br />+    dev->driver_name = zxdh_gdma_driver_name;<br />+    gdmadev = zxdh_gdma_rawdev_get_priv(dev);<br />+    gdmadev->device_state = ZXDH_GDMA_DEV_STOPPED;<br />+    gdmadev->rawdev = dev;<br />+    gdmadev->queue_num = ZXDH_GDMA_TOTAL_CHAN_NUM;<br />+    gdmadev->used_num = 0;<br />+    gdmadev->base_addr = (uintptr_t)gdev.bar_va[0] + ZXDH_GDMA_BASE_OFFSET;<br />+<br />+    for (i = 0; i < ZXDH_GDMA_TOTAL_CHAN_NUM; i++) {<br />+        queue = &(gdmadev->vqs[i]);<br />+        queue->enable = 0;<br />+        queue->queue_size = ZXDH_GDMA_QUEUE_SIZE;<br />+        rte_spinlock_init(&(queue->enqueue_lock));<br />+    }<br />+<br />+    return 0;<br />+}<br />+<br />+static int<br />+zxdh_gdma_rawdev_remove(__rte_unused struct rte_pci_device *pci_dev)<br />+{<br />+    struct rte_rawdev *dev = NULL;<br />+    int ret = 0;<br />+<br />+    dev = rte_rawdev_pmd_get_named_dev(dev_name);<br />+    if (dev == NULL)<br />+        return -EINVAL;<br />+<br />+    /* rte_rawdev_close is called by pmd_release */<br />+    ret = rte_rawdev_pmd_release(dev);<br />+    if (ret != 0) {<br />+        ZXDH_PMD_LOG(ERR, "Device cleanup failed");<br />+        return -1;<br />+    }<br />+    ZXDH_PMD_LOG(DEBUG, "rawdev %s remove done!", dev_name);<br />+<br />+    return ret;<br />+}<br />+<br />+static const struct rte_pci_id zxdh_gdma_rawdev_map[] = {<br />+    { RTE_PCI_DEVICE(ZXDH_GDMA_VENDORID, ZXDH_GDMA_DEVICEID) },<br />+    { .vendor_id = 0, /* sentinel */ },<br />+};<br />+<br />+static struct rte_pci_driver zxdh_gdma_rawdev_pmd = {<br />+    .id_table = zxdh_gdma_rawdev_map,<br />+    .drv_flags = 0,<br />+    .probe = zxdh_gdma_rawdev_probe,<br />+    .remove = zxdh_gdma_rawdev_remove,<br />+};<br />+<br />+RTE_PMD_REGISTER_PCI(zxdh_gdma_rawdev_pci_driver, zxdh_gdma_rawdev_pmd);<br />+RTE_PMD_REGISTER_PCI_TABLE(zxdh_gdma_rawdev_pci_driver, zxdh_gdma_rawdev_map);<br />+RTE_LOG_REGISTER_DEFAULT(zxdh_gdma_rawdev_logtype, NOTICE);<br />diff --git a/drivers/raw/zxdh/zxdh_rawdev.h b/drivers/raw/zxdh/zxdh_rawdev.h<br />new file mode 100644<br />index 0000000000..e2e0ffa667<br />--- /dev/null<br />+++ b/drivers/raw/zxdh/zxdh_rawdev.h<br />@@ -0,0 +1,167 @@<br />+/* SPDX-License-Identifier: BSD-3-Clause<br />+ * Copyright 2024 ZTE Corporation<br />+ */<br />+<br />+#ifndef __ZXDH_RAWDEV_H__<br />+#define __ZXDH_RAWDEV_H__<br />+<br />+#ifdef __cplusplus<br />+extern "C" {<br />+#endif<br />+<br />+#include <rte_rawdev.h> <br />+#include <rte_spinlock.h> <br />+<br />+extern int zxdh_gdma_rawdev_logtype;<br />+#define RTE_LOGTYPE_ZXDH_GDMA                   zxdh_gdma_rawdev_logtype<br />+<br />+#define ZXDH_PMD_LOG(level, ...) \<br />+    RTE_LOG_LINE_PREFIX(level, ZXDH_GDMA, \<br />+        "%s() line %u: ", __func__ RTE_LOG_COMMA __LINE__, __VA_ARGS__)<br />+<br />+#define ZXDH_GDMA_VENDORID                      0x1cf2<br />+#define ZXDH_GDMA_DEVICEID                      0x8044<br />+<br />+#define ZXDH_GDMA_TOTAL_CHAN_NUM                58<br />+#define ZXDH_GDMA_QUEUE_SIZE                    16384 /* >= 65*64*3 */<br />+#define ZXDH_GDMA_RING_SIZE                     32768<br />+<br />+/* States if the source addresses is physical. */<br />+#define ZXDH_GDMA_JOB_SRC_PHY                   (1UL)<br />+<br />+/* States if the destination addresses is physical. */<br />+#define ZXDH_GDMA_JOB_DEST_PHY                  (1UL << 1)<br />+<br />+/* ZF->HOST */<br />+#define ZXDH_GDMA_JOB_DIR_TX                    (1UL << 2)<br />+<br />+/* HOST->ZF */<br />+#define ZXDH_GDMA_JOB_DIR_RX                    (1UL << 3)<br />+<br />+#define ZXDH_GDMA_JOB_DIR_MASK                  (ZXDH_GDMA_JOB_DIR_TX | ZXDH_GDMA_JOB_DIR_RX)<br />+<br />+enum zxdh_gdma_device_state {<br />+    ZXDH_GDMA_DEV_RUNNING,<br />+    ZXDH_GDMA_DEV_STOPPED<br />+};<br />+<br />+struct zxdh_gdma_buff_desc {<br />+    uint SrcAddr_L;<br />+    uint DstAddr_L;<br />+    uint Xpara;<br />+    uint ZY_para;<br />+    uint ZY_SrcStep;<br />+    uint ZY_DstStep;<br />+    uint ExtAddr;<br />+    uint LLI_Addr_L;<br />+    uint LLI_Addr_H;<br />+    uint ChCont;<br />+    uint LLI_User;<br />+    uint ErrAddr;<br />+    uint Control;<br />+    uint SrcAddr_H;<br />+    uint DstAddr_H;<br />+    uint Reserved;<br />+};<br />+<br />+struct zxdh_gdma_queue {<br />+    uint8_t   enable;<br />+    uint8_t   is_txq;<br />+    uint16_t  vq_id;<br />+    uint16_t  queue_size;<br />+    /* 0:GDMA needs to be configured through the APB interface */<br />+    uint16_t  flag;<br />+    uint      user;<br />+    uint16_t  tc_cnt;<br />+    rte_spinlock_t enqueue_lock;<br />+    struct {<br />+        uint16_t avail_idx;<br />+        uint16_t last_avail_idx;<br />+        rte_iova_t ring_mem;<br />+        const struct rte_memzone *ring_mz;<br />+        struct zxdh_gdma_buff_desc *desc;<br />+    } ring;<br />+    struct {<br />+        uint16_t  free_cnt;<br />+        uint16_t  deq_cnt;<br />+        uint16_t  pend_cnt;<br />+        uint16_t  enq_idx;<br />+        uint16_t  deq_idx;<br />+        uint16_t  used_idx;<br />+        struct zxdh_gdma_job **job;<br />+    } sw_ring;<br />+};<br />+<br />+struct zxdh_gdma_rawdev {<br />+    struct rte_device *device;<br />+    struct rte_rawdev *rawdev;<br />+    uintptr_t base_addr;<br />+    uint8_t queue_num; /* total queue num */<br />+    uint8_t used_num;  /* used  queue num */<br />+    enum zxdh_gdma_device_state device_state;<br />+    struct zxdh_gdma_queue vqs[ZXDH_GDMA_TOTAL_CHAN_NUM];<br />+};<br />+<br />+struct zxdh_gdma_job {<br />+    uint64_t src;<br />+    uint64_t dest;<br />+    uint len;<br />+    uint flags;<br />+    uint64_t cnxt;<br />+    uint16_t status;<br />+    uint16_t vq_id;<br />+    void *usr_elem;<br />+    uint8_t ep_id;<br />+    uint8_t pf_id;<br />+    uint16_t vf_id;<br />+};<br />+<br />+struct zxdh_gdma_enqdeq {<br />+    uint16_t vq_id;<br />+    struct zxdh_gdma_job **job;<br />+};<br />+<br />+struct zxdh_gdma_config {<br />+    uint16_t max_hw_queues_per_core;<br />+    uint16_t max_vqs;<br />+    int fle_queue_pool_cnt;<br />+};<br />+<br />+struct zxdh_gdma_rbp {<br />+    uint use_ultrashort:1;<br />+    uint enable:1;<br />+    uint dportid:3;<br />+    uint dpfid:3;<br />+    uint dvfid:8; /*using route by port for destination */<br />+    uint drbp:1;<br />+    uint sportid:3;<br />+    uint spfid:3;<br />+    uint svfid:8;<br />+    uint srbp:1;<br />+};<br />+<br />+struct zxdh_gdma_queue_config {<br />+    uint lcore_id;<br />+    uint flags;<br />+    struct zxdh_gdma_rbp *rbp;<br />+};<br />+<br />+struct zxdh_gdma_attr {<br />+    uint16_t num_hw_queues;<br />+};<br />+<br />+static inline struct zxdh_gdma_rawdev *<br />+zxdh_gdma_rawdev_get_priv(const struct rte_rawdev *rawdev)<br />+{<br />+    return rawdev->dev_private;<br />+}<br />+<br />+uint zxdh_gdma_read_reg(struct rte_rawdev *dev, uint16_t qidx, uint offset);<br />+void zxdh_gdma_write_reg(struct rte_rawdev *dev, uint16_t qidx, uint offset, uint val);<br />+int zxdh_gdma_debug_info_dump(struct rte_rawdev *dev, uint16_t queue_id);<br />+<br />+#ifdef __cplusplus<br />+}<br />+#endif<br />+<br />+#endif /* __ZXDH_RAWDEV_H__ */<br />--  <br />2.43.0<br />