[dpdk-dev] [PATCH 40/56] net/sfc: maintain management event queue
Andrew Rybchenko
arybchenko at solarflare.com
Mon Nov 21 16:00:54 CET 2016
The event queue is required for device level events (e.g. link status
change) and flush events.
Provide thread-safe function to poll the event queue since it may be
really done from different contexts.
Reviewed-by: Andy Moreton <amoreton at solarflare.com>
Signed-off-by: Andrew Rybchenko <arybchenko at solarflare.com>
---
drivers/net/sfc/efx/sfc.h | 1 +
drivers/net/sfc/efx/sfc_ev.c | 50 ++++++++++++++++++++++++++++++++++++++++----
drivers/net/sfc/efx/sfc_ev.h | 2 ++
3 files changed, 49 insertions(+), 4 deletions(-)
diff --git a/drivers/net/sfc/efx/sfc.h b/drivers/net/sfc/efx/sfc.h
index eb8c071..39f15b1 100644
--- a/drivers/net/sfc/efx/sfc.h
+++ b/drivers/net/sfc/efx/sfc.h
@@ -144,6 +144,7 @@ struct sfc_adapter {
struct sfc_evq_info *evq_info;
unsigned int mgmt_evq_index;
+ rte_spinlock_t mgmt_evq_lock;
};
/*
diff --git a/drivers/net/sfc/efx/sfc_ev.c b/drivers/net/sfc/efx/sfc_ev.c
index aa04b34..1734b1e 100644
--- a/drivers/net/sfc/efx/sfc_ev.c
+++ b/drivers/net/sfc/efx/sfc_ev.c
@@ -213,6 +213,19 @@ sfc_ev_qpoll(struct sfc_evq *evq)
/* Poll-mode driver does not re-prime the event queue for interrupts */
}
+void
+sfc_ev_mgmt_qpoll(struct sfc_adapter *sa)
+{
+ if (rte_spinlock_trylock(&sa->mgmt_evq_lock)) {
+ struct sfc_evq *mgmt_evq = sa->evq_info[sa->mgmt_evq_index].evq;
+
+ if (mgmt_evq->init_state == SFC_EVQ_STARTED)
+ sfc_ev_qpoll(mgmt_evq);
+
+ rte_spinlock_unlock(&sa->mgmt_evq_lock);
+ }
+}
+
int
sfc_ev_qprime(struct sfc_evq *evq)
{
@@ -324,13 +337,26 @@ sfc_ev_start(struct sfc_adapter *sa)
if (rc != 0)
goto fail_ev_init;
+ /* Start management EVQ used for global events */
+ rte_spinlock_lock(&sa->mgmt_evq_lock);
+
+ rc = sfc_ev_qstart(sa, sa->mgmt_evq_index);
+ if (rc != 0)
+ goto fail_mgmt_evq_start;
+
+ rte_spinlock_unlock(&sa->mgmt_evq_lock);
+
/*
- * Rx/Tx event queues are started/stopped when corresponding queue
- * is started/stopped.
+ * Rx/Tx event queues are started/stopped when corresponding
+ * Rx/Tx queue is started/stopped.
*/
return 0;
+fail_mgmt_evq_start:
+ rte_spinlock_unlock(&sa->mgmt_evq_lock);
+ efx_ev_fini(sa->nic);
+
fail_ev_init:
sfc_log_init(sa, "failed %d", rc);
return rc;
@@ -345,8 +371,17 @@ sfc_ev_stop(struct sfc_adapter *sa)
/* Make sure that all event queues are stopped */
sw_index = sa->evq_count;
- while (--sw_index >= 0)
- sfc_ev_qstop(sa, sw_index);
+ while (--sw_index >= 0) {
+ if (sw_index == sa->mgmt_evq_index) {
+ /* Locks are required for the management EVQ */
+ rte_spinlock_lock(&sa->mgmt_evq_lock);
+ sfc_ev_qstop(sa, sa->mgmt_evq_index);
+ rte_spinlock_unlock(&sa->mgmt_evq_lock);
+ } else {
+ sfc_ev_qstop(sa, sw_index);
+ }
+ }
+
efx_ev_fini(sa->nic);
}
@@ -444,6 +479,7 @@ sfc_ev_init(struct sfc_adapter *sa)
sa->evq_count = sfc_ev_qcount(sa);
sa->mgmt_evq_index = 0;
+ rte_spinlock_init(&sa->mgmt_evq_lock);
/* Allocate EVQ info array */
rc = ENOMEM;
@@ -459,6 +495,11 @@ sfc_ev_init(struct sfc_adapter *sa)
goto fail_ev_qinit_info;
}
+ rc = sfc_ev_qinit(sa, sa->mgmt_evq_index, SFC_MGMT_EVQ_ENTRIES,
+ sa->socket_id);
+ if (rc != 0)
+ goto fail_mgmt_evq_init;
+
/*
* Rx/Tx event queues are created/destroyed when corresponding
* Rx/Tx queue is created/destroyed.
@@ -466,6 +507,7 @@ sfc_ev_init(struct sfc_adapter *sa)
return 0;
+fail_mgmt_evq_init:
fail_ev_qinit_info:
while (sw_index-- > 0)
sfc_ev_qfini_info(sa, sw_index);
diff --git a/drivers/net/sfc/efx/sfc_ev.h b/drivers/net/sfc/efx/sfc_ev.h
index 140a436..f7bcf01 100644
--- a/drivers/net/sfc/efx/sfc_ev.h
+++ b/drivers/net/sfc/efx/sfc_ev.h
@@ -132,6 +132,8 @@ void sfc_ev_qstop(struct sfc_adapter *sa, unsigned int sw_index);
int sfc_ev_qprime(struct sfc_evq *evq);
void sfc_ev_qpoll(struct sfc_evq *evq);
+void sfc_ev_mgmt_qpoll(struct sfc_adapter *sa);
+
#ifdef __cplusplus
}
#endif
--
2.5.5
More information about the dev
mailing list