event/octeontx2: fix unconditional Tx flush
authorPavan Nikhilesh <pbhagavatula@marvell.com>
Fri, 20 Nov 2020 10:41:46 +0000 (16:11 +0530)
committerJerin Jacob <jerinj@marvell.com>
Fri, 20 Nov 2020 17:53:47 +0000 (18:53 +0100)
Fix unconditional Tx flush, in case of Tx only we need to check if
work slot is non-empty before issuing flush.
Also, in packet retransmit cases added check for the reference
count and flush the work slot only for the last packet.

Fixes: cb7ee83b6365 ("event/octeontx2: improve single flow performance")
Cc: stable@dpdk.org
Signed-off-by: Pavan Nikhilesh <pbhagavatula@marvell.com>
drivers/event/octeontx2/otx2_evdev.h
drivers/event/octeontx2/otx2_worker.c
drivers/event/octeontx2/otx2_worker.h

index 210ee89..d0f5b3f 100644 (file)
@@ -79,6 +79,7 @@
 #define SSOW_LF_GWS_OP_GWC_INVAL            (0xe00ull)
 
 #define OTX2_SSOW_GET_BASE_ADDR(_GW)        ((_GW) - SSOW_LF_GWS_OP_GET_WORK)
+#define OTX2_SSOW_TT_FROM_TAG(x)           (((x) >> 32) & SSO_TT_EMPTY)
 
 #define NSEC2USEC(__ns)                        ((__ns) / 1E3)
 #define USEC2NSEC(__us)                 ((__us) * 1E3)
index 1d427e4..b098407 100644 (file)
@@ -274,12 +274,14 @@ otx2_ssogws_tx_adptr_enq_ ## name(void *port, struct rte_event ev[],      \
 {                                                                      \
        struct otx2_ssogws *ws = port;                                  \
        uint64_t cmd[sz];                                               \
+       int i;                                                          \
                                                                        \
-       RTE_SET_USED(nb_events);                                        \
-       return otx2_ssogws_event_tx(ws, ev, cmd, (const uint64_t        \
+       for (i = 0; i < nb_events; i++)                                 \
+               otx2_ssogws_event_tx(ws, &ev[i], cmd, (const uint64_t   \
                                    (*)[RTE_MAX_QUEUES_PER_PORT])       \
                                    &ws->tx_adptr_data,                 \
                                    flags);                             \
+       return nb_events;                                               \
 }
 SSO_TX_ADPTR_ENQ_FASTPATH_FUNC
 #undef T
@@ -289,14 +291,16 @@ uint16_t __rte_hot                                                        \
 otx2_ssogws_tx_adptr_enq_seg_ ## name(void *port, struct rte_event ev[],\
                                      uint16_t nb_events)               \
 {                                                                      \
-       struct otx2_ssogws *ws = port;                                  \
        uint64_t cmd[(sz) + NIX_TX_MSEG_SG_DWORDS - 2];                 \
+       struct otx2_ssogws *ws = port;                                  \
+       int i;                                                          \
                                                                        \
-       RTE_SET_USED(nb_events);                                        \
-       return otx2_ssogws_event_tx(ws, ev, cmd, (const uint64_t        \
+       for (i = 0; i < nb_events; i++)                                 \
+               otx2_ssogws_event_tx(ws, &ev[i], cmd, (const uint64_t   \
                                    (*)[RTE_MAX_QUEUES_PER_PORT])       \
                                    &ws->tx_adptr_data,                 \
                                    (flags) | NIX_TX_MULTI_SEG_F);      \
+       return nb_events;                                               \
 }
 SSO_TX_ADPTR_ENQ_FASTPATH_FUNC
 #undef T
index 3efd3ba..0a7d667 100644 (file)
@@ -198,6 +198,10 @@ otx2_ssogws_swtag_untag(struct otx2_ssogws *ws)
 static __rte_always_inline void
 otx2_ssogws_swtag_flush(struct otx2_ssogws *ws)
 {
+       if (OTX2_SSOW_TT_FROM_TAG(otx2_read64(ws->tag_op)) == SSO_TT_EMPTY) {
+               ws->cur_tt = SSO_SYNC_EMPTY;
+               return;
+       }
        otx2_write64(0, ws->swtag_flush_op);
        ws->cur_tt = SSO_SYNC_EMPTY;
 }
@@ -272,13 +276,14 @@ otx2_ssogws_prepare_pkt(const struct otx2_eth_txq *txq, struct rte_mbuf *m,
 }
 
 static __rte_always_inline uint16_t
-otx2_ssogws_event_tx(struct otx2_ssogws *ws, struct rte_event ev[],
-                    uint64_t *cmd, const uint64_t
-                    txq_data[][RTE_MAX_QUEUES_PER_PORT],
+otx2_ssogws_event_tx(struct otx2_ssogws *ws, struct rte_event *ev,
+                    uint64_t *cmd,
+                    const uint64_t txq_data[][RTE_MAX_QUEUES_PER_PORT],
                     const uint32_t flags)
 {
-       struct rte_mbuf *m = ev[0].mbuf;
+       struct rte_mbuf *m = ev->mbuf;
        const struct otx2_eth_txq *txq;
+       uint16_t ref_cnt = m->refcnt;
 
        if ((flags & NIX_TX_OFFLOAD_SECURITY_F) &&
            (m->ol_flags & PKT_TX_SEC_OFFLOAD)) {
@@ -329,7 +334,12 @@ otx2_ssogws_event_tx(struct otx2_ssogws *ws, struct rte_event ev[],
                }
        }
 
-       otx2_write64(0, ws->swtag_flush_op);
+       if (flags & NIX_TX_OFFLOAD_MBUF_NOFF_F) {
+               if (ref_cnt > 1)
+                       return 1;
+       }
+
+       otx2_ssogws_swtag_flush(ws);
 
        return 1;
 }