[PATCH v8 01/13] vdpa/ifc: add support for virtio blk device
Xia, Chenbo
chenbo.xia at intel.com
Mon May 23 05:41:48 CEST 2022
Hi Andy,
> -----Original Message-----
> From: Pei, Andy <andy.pei at intel.com>
> Sent: Wednesday, May 18, 2022 8:14 PM
> To: dev at dpdk.org
> Cc: Xia, Chenbo <chenbo.xia at intel.com>; maxime.coquelin at redhat.com; Cao,
> Gang <gang.cao at intel.com>; Liu, Changpeng <changpeng.liu at intel.com>; Xu,
> Rosen <rosen.xu at intel.com>; Xiao, QimaiX <qimaix.xiao at intel.com>
> Subject: [PATCH v8 01/13] vdpa/ifc: add support for virtio blk device
>
> Re-use the vdpa/ifc code, distinguish blk and net device by pci_device_id.
> Blk and net device are implemented with proper feature and ops.
>
> Signed-off-by: Andy Pei <andy.pei at intel.com>
> ---
> drivers/vdpa/ifc/base/ifcvf.h | 16 +++++++-
> drivers/vdpa/ifc/ifcvf_vdpa.c | 91
> +++++++++++++++++++++++++++++++++++++++----
> 2 files changed, 98 insertions(+), 9 deletions(-)
>
> diff --git a/drivers/vdpa/ifc/base/ifcvf.h b/drivers/vdpa/ifc/base/ifcvf.h
> index 573a35f..483d38b 100644
> --- a/drivers/vdpa/ifc/base/ifcvf.h
> +++ b/drivers/vdpa/ifc/base/ifcvf.h
> @@ -5,8 +5,17 @@
> #ifndef _IFCVF_H_
> #define _IFCVF_H_
>
> +#include <linux/virtio_blk.h>
> #include "ifcvf_osdep.h"
>
> +#define IFCVF_NET 0
> +#define IFCVF_BLK 1
> +
> +/* for BLK */
> +#define IFCVF_BLK_TRANSITIONAL_DEVICE_ID 0x1001
> +#define IFCVF_BLK_MODERN_DEVICE_ID 0x1042
> +#define IFCVF_BLK_DEVICE_ID 0x0002
> +
> #define IFCVF_VENDOR_ID 0x1AF4
> #define IFCVF_DEVICE_ID 0x1041
> #define IFCVF_SUBSYS_VENDOR_ID 0x8086
Let's rename IFCVF_DEVICE_ID to IFCVF_NET_DEVICE_ID as it's only used for net now.
> @@ -126,13 +135,18 @@ struct ifcvf_hw {
> u8 notify_region;
> u32 notify_off_multiplier;
> struct ifcvf_pci_common_cfg *common_cfg;
> - struct ifcvf_net_config *dev_cfg;
> + union {
> + struct ifcvf_net_config *net_cfg;
> + struct virtio_blk_config *blk_cfg;
> + void *dev_cfg;
> + };
> u8 *isr;
> u16 *notify_base;
> u16 *notify_addr[IFCVF_MAX_QUEUES * 2];
> u8 *lm_cfg;
> struct vring_info vring[IFCVF_MAX_QUEUES * 2];
> u8 nr_vring;
> + int device_type;
> struct ifcvf_pci_mem_resource mem_resource[IFCVF_PCI_MAX_RESOURCE];
> };
>
> diff --git a/drivers/vdpa/ifc/ifcvf_vdpa.c b/drivers/vdpa/ifc/ifcvf_vdpa.c
> index 9f05595..be0efd3 100644
> --- a/drivers/vdpa/ifc/ifcvf_vdpa.c
> +++ b/drivers/vdpa/ifc/ifcvf_vdpa.c
> @@ -75,6 +75,12 @@ struct internal_list {
> struct ifcvf_internal *internal;
> };
>
> +/* vdpa device info includes device features and devcic operation. */
> +struct rte_vdpa_dev_info {
> + uint64_t features;
> + struct rte_vdpa_dev_ops *ops;
> +};
> +
> TAILQ_HEAD(internal_list_head, internal_list);
> static struct internal_list_head internal_list =
> TAILQ_HEAD_INITIALIZER(internal_list);
> @@ -1167,6 +1173,48 @@ struct internal_list {
> return 0;
> }
>
> +static int16_t
> +ifcvf_pci_get_device_type(struct rte_pci_device *pci_dev)
> +{
> + uint16_t pci_device_id = pci_dev->id.device_id;
> + uint16_t device_id;
> +
> + if (pci_device_id < 0x1000 || pci_device_id > 0x107f) {
> + DRV_LOG(ERR, "Probe device is not a virtio device\n");
> + return -1;
> + }
> +
> + if (pci_device_id < 0x1040) {
> + /* Transitional devices: use the PCI subsystem device id as
> + * virtio device id, same as legacy driver always did.
> + */
> + device_id = pci_dev->id.subsystem_device_id;
> + } else {
> + /* Modern devices: simply use PCI device id,
> + * but start from 0x1040.
> + */
> + device_id = pci_device_id - 0x1040;
> + }
> +
> + return device_id;
> +}
> +
> +struct rte_vdpa_dev_info dev_info[] = {
> + {
> + .features = (1ULL << VIRTIO_NET_F_GUEST_ANNOUNCE) |
> + (1ULL << VIRTIO_NET_F_CTRL_VQ) |
> + (1ULL << VIRTIO_NET_F_STATUS) |
> + (1ULL << VHOST_USER_F_PROTOCOL_FEATURES) |
> + (1ULL << VHOST_F_LOG_ALL),
> + .ops = &ifcvf_ops,
Rename ifcvf_ops -> ifcvf_net_ops
Overall the patch LGTM. With above fixed:
Reviewed-by: Chenbo Xia <chenbo.xia at intel.com>
> + },
> + {
> + .features = (1ULL << VHOST_USER_F_PROTOCOL_FEATURES) |
> + (1ULL << VHOST_F_LOG_ALL),
> + .ops = NULL,
> + },
> +};
> +
> static int
> ifcvf_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
> struct rte_pci_device *pci_dev)
> @@ -1178,6 +1226,7 @@ struct internal_list {
> int sw_fallback_lm = 0;
> struct rte_kvargs *kvlist = NULL;
> int ret = 0;
> + int16_t device_id;
>
> if (rte_eal_process_type() != RTE_PROC_PRIMARY)
> return 0;
> @@ -1227,13 +1276,24 @@ struct internal_list {
> internal->configured = 0;
> internal->max_queues = IFCVF_MAX_QUEUES;
> features = ifcvf_get_features(&internal->hw);
> - internal->features = (features &
> - ~(1ULL << VIRTIO_F_IOMMU_PLATFORM)) |
> - (1ULL << VIRTIO_NET_F_GUEST_ANNOUNCE) |
> - (1ULL << VIRTIO_NET_F_CTRL_VQ) |
> - (1ULL << VIRTIO_NET_F_STATUS) |
> - (1ULL << VHOST_USER_F_PROTOCOL_FEATURES) |
> - (1ULL << VHOST_F_LOG_ALL);
> +
> + device_id = ifcvf_pci_get_device_type(pci_dev);
> + if (device_id < 0) {
> + DRV_LOG(ERR, "failed to get device %s type", pci_dev->name);
> + goto error;
> + }
> +
> + if (device_id == VIRTIO_ID_NET) {
> + internal->hw.device_type = IFCVF_NET;
> + internal->features = features &
> + ~(1ULL << VIRTIO_F_IOMMU_PLATFORM);
> + internal->features |= dev_info[IFCVF_NET].features;
> + } else if (device_id == VIRTIO_ID_BLOCK) {
> + internal->hw.device_type = IFCVF_BLK;
> + internal->features = features &
> + ~(1ULL << VIRTIO_F_IOMMU_PLATFORM);
> + internal->features |= dev_info[IFCVF_BLK].features;
> + }
>
> list->internal = internal;
>
> @@ -1245,7 +1305,8 @@ struct internal_list {
> }
> internal->sw_lm = sw_fallback_lm;
>
> - internal->vdev = rte_vdpa_register_device(&pci_dev->device,
> &ifcvf_ops);
> + internal->vdev = rte_vdpa_register_device(&pci_dev->device,
> + dev_info[internal->hw.device_type].ops);
> if (internal->vdev == NULL) {
> DRV_LOG(ERR, "failed to register device %s", pci_dev->name);
> goto error;
> @@ -1313,6 +1374,20 @@ struct internal_list {
> .subsystem_device_id = IFCVF_SUBSYS_DEVICE_ID,
> },
>
> + { .class_id = RTE_CLASS_ANY_ID,
> + .vendor_id = IFCVF_VENDOR_ID,
> + .device_id = IFCVF_BLK_TRANSITIONAL_DEVICE_ID,
> + .subsystem_vendor_id = IFCVF_SUBSYS_VENDOR_ID,
> + .subsystem_device_id = IFCVF_BLK_DEVICE_ID,
> + },
> +
> + { .class_id = RTE_CLASS_ANY_ID,
> + .vendor_id = IFCVF_VENDOR_ID,
> + .device_id = IFCVF_BLK_MODERN_DEVICE_ID,
> + .subsystem_vendor_id = IFCVF_SUBSYS_VENDOR_ID,
> + .subsystem_device_id = IFCVF_BLK_DEVICE_ID,
> + },
> +
> { .vendor_id = 0, /* sentinel */
> },
> };
> --
> 1.8.3.1
More information about the dev
mailing list