[dpdk-dev] [PATCH v5 resend 04/12] vhost: rxtx: prepare work for multiple queue support

Michael S. Tsirkin mst at redhat.com
Mon Sep 21 11:04:16 CEST 2015


On Mon, Sep 21, 2015 at 10:25:18AM +0800, Yuanhan Liu wrote:
> On Sun, Sep 20, 2015 at 12:29:17PM +0300, Michael S. Tsirkin wrote:
> > On Fri, Sep 18, 2015 at 11:10:53PM +0800, Yuanhan Liu wrote:
> > > From: Changchun Ouyang <changchun.ouyang at intel.com>
> > > 
> > > Do not use VIRTIO_RXQ or VIRTIO_TXQ anymore; use the queue_id,
> > > instead, which will be set to a proper value for a specific queue
> > > when we have multiple queue support enabled.
> > > 
> > > For now, queue_id is still set with VIRTIO_RXQ or VIRTIO_TXQ,
> > > so it should not break anything.
> > > 
> > > Signed-off-by: Changchun Ouyang <changchun.ouyang at intel.com>
> > > Signed-off-by: Yuanhan Liu <yuanhan.liu at linux.intel.com>
> > 
> > What I'm missing is the logic that tracks which tx vq
> > did a specific flow use, to send traffic back on the correct
> > rx vq.
> > 
> > Is this something added by one of the follow-up patches?
> 
> Yeah, and also, they are offloaded to the application (examples/
> vhost/vhost-switch): patch 11 does the job.
> 
> 	--yliu

Couldn't find anything dealing with flows there ...
Could you be more specific pls?

> > 
> > 
> > > ---
> > >  lib/librte_vhost/vhost_rxtx.c | 46 ++++++++++++++++++++++++++++++-------------
> > >  1 file changed, 32 insertions(+), 14 deletions(-)
> > > 
> > > diff --git a/lib/librte_vhost/vhost_rxtx.c b/lib/librte_vhost/vhost_rxtx.c
> > > index b2b2bcc..a4ab6ca 100644
> > > --- a/lib/librte_vhost/vhost_rxtx.c
> > > +++ b/lib/librte_vhost/vhost_rxtx.c
> > > @@ -42,6 +42,16 @@
> > >  
> > >  #define MAX_PKT_BURST 32
> > >  
> > > +static inline int __attribute__((always_inline))
> > > +is_valid_virt_queue_idx(uint32_t virtq_idx, int is_tx, uint32_t max_qp_idx)
> > > +{
> > > +	if ((is_tx ^ (virtq_idx & 0x1)) ||
> > > +	    (virtq_idx >= max_qp_idx * VIRTIO_QNUM))
> > > +		return 0;
> > > +
> > > +	return 1;
> > > +}
> > > +
> > >  /**
> > >   * This function adds buffers to the virtio devices RX virtqueue. Buffers can
> > >   * be received from the physical port or from another virtio device. A packet
> > > @@ -68,12 +78,14 @@ virtio_dev_rx(struct virtio_net *dev, uint16_t queue_id,
> > >  	uint8_t success = 0;
> > >  
> > >  	LOG_DEBUG(VHOST_DATA, "(%"PRIu64") virtio_dev_rx()\n", dev->device_fh);
> > > -	if (unlikely(queue_id != VIRTIO_RXQ)) {
> > > -		LOG_DEBUG(VHOST_DATA, "mq isn't supported in this version.\n");
> > > +	if (unlikely(!is_valid_virt_queue_idx(queue_id, 0, dev->virt_qp_nb))) {
> > > +		RTE_LOG(ERR, VHOST_DATA,
> > > +			"%s (%"PRIu64"): virtqueue idx:%d invalid.\n",
> > > +			__func__, dev->device_fh, queue_id);
> > >  		return 0;
> > >  	}
> > >  
> > > -	vq = dev->virtqueue[VIRTIO_RXQ];
> > > +	vq = dev->virtqueue[queue_id];
> > >  	count = (count > MAX_PKT_BURST) ? MAX_PKT_BURST : count;
> > >  
> > >  	/*
> > > @@ -235,8 +247,9 @@ virtio_dev_rx(struct virtio_net *dev, uint16_t queue_id,
> > >  }
> > >  
> > >  static inline uint32_t __attribute__((always_inline))
> > > -copy_from_mbuf_to_vring(struct virtio_net *dev, uint16_t res_base_idx,
> > > -	uint16_t res_end_idx, struct rte_mbuf *pkt)
> > > +copy_from_mbuf_to_vring(struct virtio_net *dev, uint32_t queue_id,
> > > +			uint16_t res_base_idx, uint16_t res_end_idx,
> > > +			struct rte_mbuf *pkt)
> > >  {
> > >  	uint32_t vec_idx = 0;
> > >  	uint32_t entry_success = 0;
> > > @@ -264,7 +277,7 @@ copy_from_mbuf_to_vring(struct virtio_net *dev, uint16_t res_base_idx,
> > >  	 * Convert from gpa to vva
> > >  	 * (guest physical addr -> vhost virtual addr)
> > >  	 */
> > > -	vq = dev->virtqueue[VIRTIO_RXQ];
> > > +	vq = dev->virtqueue[queue_id];
> > >  	vb_addr = gpa_to_vva(dev, vq->buf_vec[vec_idx].buf_addr);
> > >  	vb_hdr_addr = vb_addr;
> > >  
> > > @@ -464,11 +477,14 @@ virtio_dev_merge_rx(struct virtio_net *dev, uint16_t queue_id,
> > >  
> > >  	LOG_DEBUG(VHOST_DATA, "(%"PRIu64") virtio_dev_merge_rx()\n",
> > >  		dev->device_fh);
> > > -	if (unlikely(queue_id != VIRTIO_RXQ)) {
> > > -		LOG_DEBUG(VHOST_DATA, "mq isn't supported in this version.\n");
> > > +	if (unlikely(!is_valid_virt_queue_idx(queue_id, 0, dev->virt_qp_nb))) {
> > > +		RTE_LOG(ERR, VHOST_DATA,
> > > +			"%s (%"PRIu64"): virtqueue idx:%d invalid.\n",
> > > +			__func__, dev->device_fh, queue_id);
> > > +		return 0;
> > >  	}
> > >  
> > > -	vq = dev->virtqueue[VIRTIO_RXQ];
> > > +	vq = dev->virtqueue[queue_id];
> > >  	count = RTE_MIN((uint32_t)MAX_PKT_BURST, count);
> > >  
> > >  	if (count == 0)
> > > @@ -509,8 +525,8 @@ virtio_dev_merge_rx(struct virtio_net *dev, uint16_t queue_id,
> > >  							res_cur_idx);
> > >  		} while (success == 0);
> > >  
> > > -		entry_success = copy_from_mbuf_to_vring(dev, res_base_idx,
> > > -			res_cur_idx, pkts[pkt_idx]);
> > > +		entry_success = copy_from_mbuf_to_vring(dev, queue_id,
> > > +			res_base_idx, res_cur_idx, pkts[pkt_idx]);
> > >  
> > >  		rte_compiler_barrier();
> > >  
> > > @@ -559,12 +575,14 @@ rte_vhost_dequeue_burst(struct virtio_net *dev, uint16_t queue_id,
> > >  	uint16_t free_entries, entry_success = 0;
> > >  	uint16_t avail_idx;
> > >  
> > > -	if (unlikely(queue_id != VIRTIO_TXQ)) {
> > > -		LOG_DEBUG(VHOST_DATA, "mq isn't supported in this version.\n");
> > > +	if (unlikely(!is_valid_virt_queue_idx(queue_id, 1, dev->virt_qp_nb))) {
> > > +		RTE_LOG(ERR, VHOST_DATA,
> > > +			"%s (%"PRIu64"): virtqueue idx:%d invalid.\n",
> > > +			__func__, dev->device_fh, queue_id);
> > >  		return 0;
> > >  	}
> > >  
> > > -	vq = dev->virtqueue[VIRTIO_TXQ];
> > > +	vq = dev->virtqueue[queue_id];
> > >  	avail_idx =  *((volatile uint16_t *)&vq->avail->idx);
> > >  
> > >  	/* If there are no available buffers then return. */
> > > -- 
> > > 1.9.0


More information about the dev mailing list