X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=drivers%2Fnet%2Fsfc%2Fsfc_repr_proxy.c;h=535b07ea524c82ab1551e8c9c84159bf2da2b387;hb=6ded2e01380e5f3bd48154ca8f23eaad35ffe839;hp=ea03d5afddf0a23c47d21cfd22a2373f9d1ad994;hpb=75f080fdf74a92ad023fc8081c29b9ff6e9702e3;p=dpdk.git diff --git a/drivers/net/sfc/sfc_repr_proxy.c b/drivers/net/sfc/sfc_repr_proxy.c index ea03d5afdd..535b07ea52 100644 --- a/drivers/net/sfc/sfc_repr_proxy.c +++ b/drivers/net/sfc/sfc_repr_proxy.c @@ -18,6 +18,7 @@ #include "sfc_ev.h" #include "sfc_rx.h" #include "sfc_tx.h" +#include "sfc_dp_rx.h" /** * Amount of time to wait for the representor proxy routine (which is @@ -25,6 +26,14 @@ */ #define SFC_REPR_PROXY_MBOX_POLL_TIMEOUT_MS 1000 +/** + * Amount of time to wait for the representor proxy routine (which is + * running on a service core) to terminate after service core is stopped. + */ +#define SFC_REPR_PROXY_ROUTINE_TERMINATE_TIMEOUT_MS 10000 + +#define SFC_REPR_INVALID_ROUTE_PORT_ID (UINT16_MAX) + static struct sfc_repr_proxy * sfc_repr_proxy_by_adapter(struct sfc_adapter *sa) { @@ -148,16 +157,181 @@ sfc_repr_proxy_mbox_handle(struct sfc_repr_proxy *rp) __atomic_store_n(&mbox->ack, true, __ATOMIC_RELEASE); } +static void +sfc_repr_proxy_handle_tx(struct sfc_repr_proxy_dp_txq *rp_txq, + struct sfc_repr_proxy_txq *repr_txq) +{ + /* + * With multiple representor proxy queues configured it is + * possible that not all of the corresponding representor + * queues were created. Skip the queues that do not exist. + */ + if (repr_txq->ring == NULL) + return; + + if (rp_txq->available < RTE_DIM(rp_txq->tx_pkts)) { + rp_txq->available += + rte_ring_sc_dequeue_burst(repr_txq->ring, + (void **)(&rp_txq->tx_pkts[rp_txq->available]), + RTE_DIM(rp_txq->tx_pkts) - rp_txq->available, + NULL); + + if (rp_txq->available == rp_txq->transmitted) + return; + } + + rp_txq->transmitted += rp_txq->pkt_burst(rp_txq->dp, + &rp_txq->tx_pkts[rp_txq->transmitted], + rp_txq->available - rp_txq->transmitted); + + if (rp_txq->available == rp_txq->transmitted) { + rp_txq->available = 0; + rp_txq->transmitted = 0; + } +} + +static struct sfc_repr_proxy_port * +sfc_repr_proxy_rx_route_mbuf(struct sfc_repr_proxy *rp, struct rte_mbuf *m) +{ + struct sfc_repr_proxy_port *port; + efx_mport_id_t mport_id; + + mport_id.id = *RTE_MBUF_DYNFIELD(m, sfc_dp_mport_offset, + typeof(&((efx_mport_id_t *)0)->id)); + + TAILQ_FOREACH(port, &rp->ports, entries) { + if (port->egress_mport.id == mport_id.id) { + m->port = port->rte_port_id; + m->ol_flags &= ~sfc_dp_mport_override; + return port; + } + } + + return NULL; +} + +/* + * Returns true if a packet is encountered which should be forwarded to a + * port which is different from the one that is currently routed. + */ +static bool +sfc_repr_proxy_rx_route(struct sfc_repr_proxy *rp, + struct sfc_repr_proxy_dp_rxq *rp_rxq) +{ + unsigned int i; + + for (i = rp_rxq->routed; + i < rp_rxq->available && !rp_rxq->stop_route; + i++, rp_rxq->routed++) { + struct sfc_repr_proxy_port *port; + struct rte_mbuf *m = rp_rxq->pkts[i]; + + port = sfc_repr_proxy_rx_route_mbuf(rp, m); + /* Cannot find destination representor */ + if (port == NULL) { + /* Effectively drop the packet */ + rp_rxq->forwarded++; + continue; + } + + /* Currently routed packets are mapped to a different port */ + if (port->repr_id != rp_rxq->route_port_id && + rp_rxq->route_port_id != SFC_REPR_INVALID_ROUTE_PORT_ID) + return true; + + rp_rxq->route_port_id = port->repr_id; + } + + return false; +} + +static void +sfc_repr_proxy_rx_forward(struct sfc_repr_proxy *rp, + struct sfc_repr_proxy_dp_rxq *rp_rxq) +{ + struct sfc_repr_proxy_port *port; + + if (rp_rxq->route_port_id != SFC_REPR_INVALID_ROUTE_PORT_ID) { + port = sfc_repr_proxy_find_port(rp, rp_rxq->route_port_id); + + if (port != NULL && port->started) { + rp_rxq->forwarded += + rte_ring_sp_enqueue_burst(port->rxq[0].ring, + (void **)(&rp_rxq->pkts[rp_rxq->forwarded]), + rp_rxq->routed - rp_rxq->forwarded, NULL); + } else { + /* Drop all routed packets if the port is not started */ + rp_rxq->forwarded = rp_rxq->routed; + } + } + + if (rp_rxq->forwarded == rp_rxq->routed) { + rp_rxq->route_port_id = SFC_REPR_INVALID_ROUTE_PORT_ID; + rp_rxq->stop_route = false; + } else { + /* Stall packet routing if not all packets were forwarded */ + rp_rxq->stop_route = true; + } + + if (rp_rxq->available == rp_rxq->forwarded) + rp_rxq->available = rp_rxq->forwarded = rp_rxq->routed = 0; +} + +static void +sfc_repr_proxy_handle_rx(struct sfc_repr_proxy *rp, + struct sfc_repr_proxy_dp_rxq *rp_rxq) +{ + bool route_again; + + if (rp_rxq->available < RTE_DIM(rp_rxq->pkts)) { + rp_rxq->available += rp_rxq->pkt_burst(rp_rxq->dp, + &rp_rxq->pkts[rp_rxq->available], + RTE_DIM(rp_rxq->pkts) - rp_rxq->available); + if (rp_rxq->available == rp_rxq->forwarded) + return; + } + + do { + route_again = sfc_repr_proxy_rx_route(rp, rp_rxq); + sfc_repr_proxy_rx_forward(rp, rp_rxq); + } while (route_again && !rp_rxq->stop_route); +} + static int32_t sfc_repr_proxy_routine(void *arg) { + struct sfc_repr_proxy_port *port; struct sfc_repr_proxy *rp = arg; + unsigned int i; sfc_repr_proxy_mbox_handle(rp); + TAILQ_FOREACH(port, &rp->ports, entries) { + if (!port->started) + continue; + + for (i = 0; i < rp->nb_txq; i++) + sfc_repr_proxy_handle_tx(&rp->dp_txq[i], &port->txq[i]); + } + + for (i = 0; i < rp->nb_rxq; i++) + sfc_repr_proxy_handle_rx(rp, &rp->dp_rxq[i]); + return 0; } +static struct sfc_txq_info * +sfc_repr_proxy_txq_info_get(struct sfc_adapter *sa, unsigned int repr_queue_id) +{ + struct sfc_adapter_shared *sas = sfc_sa2shared(sa); + struct sfc_repr_proxy_dp_txq *dp_txq; + + SFC_ASSERT(repr_queue_id < sfc_repr_nb_txq(sas)); + dp_txq = &sa->repr_proxy.dp_txq[repr_queue_id]; + + return &sas->txq_info[dp_txq->sw_index]; +} + static int sfc_repr_proxy_txq_attach(struct sfc_adapter *sa) { @@ -289,11 +463,20 @@ sfc_repr_proxy_txq_fini(struct sfc_adapter *sa) static int sfc_repr_proxy_txq_start(struct sfc_adapter *sa) { + struct sfc_adapter_shared * const sas = sfc_sa2shared(sa); struct sfc_repr_proxy *rp = &sa->repr_proxy; + unsigned int i; sfc_log_init(sa, "entry"); - RTE_SET_USED(rp); + for (i = 0; i < sfc_repr_nb_txq(sas); i++) { + struct sfc_repr_proxy_dp_txq *txq = &rp->dp_txq[i]; + + txq->dp = sfc_repr_proxy_txq_info_get(sa, i)->dp; + txq->pkt_burst = sa->eth_dev->tx_pkt_burst; + txq->available = 0; + txq->transmitted = 0; + } sfc_log_init(sa, "done"); @@ -342,6 +525,18 @@ sfc_repr_proxy_rxq_detach(struct sfc_adapter *sa) sfc_log_init(sa, "done"); } +static struct sfc_rxq_info * +sfc_repr_proxy_rxq_info_get(struct sfc_adapter *sa, unsigned int repr_queue_id) +{ + struct sfc_adapter_shared *sas = sfc_sa2shared(sa); + struct sfc_repr_proxy_dp_rxq *dp_rxq; + + SFC_ASSERT(repr_queue_id < sfc_repr_nb_rxq(sas)); + dp_rxq = &sa->repr_proxy.dp_rxq[repr_queue_id]; + + return &sas->rxq_info[dp_rxq->sw_index]; +} + static int sfc_repr_proxy_rxq_init(struct sfc_adapter *sa, struct sfc_repr_proxy_dp_rxq *rxq) @@ -469,6 +664,14 @@ sfc_repr_proxy_rxq_start(struct sfc_adapter *sa) i); goto fail_start; } + + rxq->dp = sfc_repr_proxy_rxq_info_get(sa, i)->dp; + rxq->pkt_burst = sa->eth_dev->rx_pkt_burst; + rxq->available = 0; + rxq->routed = 0; + rxq->forwarded = 0; + rxq->stop_route = false; + rxq->route_port_id = SFC_REPR_INVALID_ROUTE_PORT_ID; } sfc_log_init(sa, "done"); @@ -922,6 +1125,9 @@ sfc_repr_proxy_start(struct sfc_adapter *sa) if (rc != 0) goto fail_txq_start; + rp->nb_txq = sfc_repr_nb_txq(sas); + rp->nb_rxq = sfc_repr_nb_rxq(sas); + /* Service core may be in "stopped" state, start it */ rc = rte_service_lcore_start(rp->service_core_id); if (rc != 0 && rc != -EALREADY) { @@ -1007,6 +1213,9 @@ sfc_repr_proxy_stop(struct sfc_adapter *sa) struct sfc_adapter_shared * const sas = sfc_sa2shared(sa); struct sfc_repr_proxy *rp = &sa->repr_proxy; struct sfc_repr_proxy_port *port; + const unsigned int wait_ms_total = + SFC_REPR_PROXY_ROUTINE_TERMINATE_TIMEOUT_MS; + unsigned int i; int rc; sfc_log_init(sa, "entry"); @@ -1050,6 +1259,17 @@ sfc_repr_proxy_stop(struct sfc_adapter *sa) /* Service lcore may be shared and we never stop it */ + /* + * Wait for the representor proxy routine to finish the last iteration. + * Give up on timeout. + */ + for (i = 0; i < wait_ms_total; i++) { + if (rte_service_may_be_active(rp->service_id) == 0) + break; + + rte_delay_ms(1); + } + sfc_repr_proxy_rxq_stop(sa); sfc_repr_proxy_txq_stop(sa);