mempool: fix slow allocation of large mempools
[dpdk.git] / app / test-pmd / macswap.c
index 1786095..71af916 100644 (file)
 #include <rte_cycles.h>
 #include <rte_memory.h>
 #include <rte_memcpy.h>
-#include <rte_memzone.h>
 #include <rte_launch.h>
-#include <rte_tailq.h>
 #include <rte_eal.h>
 #include <rte_per_lcore.h>
 #include <rte_lcore.h>
 #include <rte_atomic.h>
 #include <rte_branch_prediction.h>
-#include <rte_ring.h>
-#include <rte_memory.h>
 #include <rte_mempool.h>
 #include <rte_mbuf.h>
 #include <rte_interrupts.h>
 #include <rte_ethdev.h>
 #include <rte_ip.h>
 #include <rte_string_fns.h>
+#include <rte_flow.h>
 
 #include "testpmd.h"
+#if defined(RTE_ARCH_X86)
+#include "macswap_sse.h"
+#elif defined(RTE_MACHINE_CPUFLAG_NEON)
+#include "macswap_neon.h"
+#else
+#include "macswap.h"
+#endif
 
 /*
  * MAC swap forwarding mode: Swap the source and the destination Ethernet
@@ -79,12 +83,9 @@ pkt_burst_mac_swap(struct fwd_stream *fs)
 {
        struct rte_mbuf  *pkts_burst[MAX_PKT_BURST];
        struct rte_port  *txp;
-       struct rte_mbuf  *mb;
-       struct ether_hdr *eth_hdr;
-       struct ether_addr addr;
        uint16_t nb_rx;
        uint16_t nb_tx;
-       uint16_t i;
+       uint32_t retry;
 #ifdef RTE_TEST_PMD_RECORD_CORE_CYCLES
        uint64_t start_tsc;
        uint64_t end_tsc;
@@ -108,21 +109,21 @@ pkt_burst_mac_swap(struct fwd_stream *fs)
 #endif
        fs->rx_packets += nb_rx;
        txp = &ports[fs->tx_port];
-       for (i = 0; i < nb_rx; i++) {
-               mb = pkts_burst[i];
-               eth_hdr = rte_pktmbuf_mtod(mb, struct ether_hdr *);
 
-               /* Swap dest and src mac addresses. */
-               ether_addr_copy(&eth_hdr->d_addr, &addr);
-               ether_addr_copy(&eth_hdr->s_addr, &eth_hdr->d_addr);
-               ether_addr_copy(&addr, &eth_hdr->s_addr);
+       do_macswap(pkts_burst, nb_rx, txp);
 
-               mb->ol_flags = txp->tx_ol_flags;
-               mb->l2_len = sizeof(struct ether_hdr);
-               mb->l3_len = sizeof(struct ipv4_hdr);
-               mb->vlan_tci = txp->tx_vlan_id;
-       }
        nb_tx = rte_eth_tx_burst(fs->tx_port, fs->tx_queue, pkts_burst, nb_rx);
+       /*
+        * Retry if necessary
+        */
+       if (unlikely(nb_tx < nb_rx) && fs->retry_enabled) {
+               retry = 0;
+               while (nb_tx < nb_rx && retry++ < burst_tx_retry_num) {
+                       rte_delay_us(burst_tx_delay_time);
+                       nb_tx += rte_eth_tx_burst(fs->tx_port, fs->tx_queue,
+                                       &pkts_burst[nb_tx], nb_rx - nb_tx);
+               }
+       }
        fs->tx_packets += nb_tx;
 #ifdef RTE_TEST_PMD_RECORD_BURST_STATS
        fs->tx_burst_stats.pkt_burst_spread[nb_tx]++;