mempool: rename functions with confusing names
[dpdk.git] / examples / multi_process / l2fwd_fork / main.c
index 03fba55..2d951d9 100644 (file)
@@ -1,7 +1,7 @@
 /*-
  *   BSD LICENSE
  *
- *   Copyright(c) 2010-2014 Intel Corporation. All rights reserved.
+ *   Copyright(c) 2010-2016 Intel Corporation. All rights reserved.
  *   All rights reserved.
  *
  *   Redistribution and use in source and binary forms, with or without
@@ -52,7 +52,6 @@
 #include <rte_memory.h>
 #include <rte_memcpy.h>
 #include <rte_memzone.h>
-#include <rte_tailq.h>
 #include <rte_eal.h>
 #include <rte_per_lcore.h>
 #include <rte_launch.h>
 #include <rte_mbuf.h>
 #include <rte_malloc.h>
 
-#include "main.h"
 #include "flib.h"
 
 #define RTE_LOGTYPE_L2FWD RTE_LOGTYPE_USER1
 #define MBUF_NAME      "mbuf_pool_%d"
-#define MBUF_SIZE (2048 + sizeof(struct rte_mbuf) + RTE_PKTMBUF_HEADROOM)
+#define MBUF_SIZE      \
+(RTE_MBUF_DEFAULT_DATAROOM + sizeof(struct rte_mbuf) + RTE_PKTMBUF_HEADROOM)
 #define NB_MBUF   8192
 #define RING_MASTER_NAME       "l2fwd_ring_m2s_"
 #define RING_SLAVE_NAME                "l2fwd_ring_s2m_"
@@ -96,25 +95,6 @@ enum l2fwd_cmd{
        CMD_STOP,
 };
 
-/*
- * RX and TX Prefetch, Host, and Write-back threshold values should be
- * carefully set for optimal performance. Consult the network
- * controller's datasheet and supporting DPDK documentation for guidance
- * on how these parameters should be set.
- */
-#define RX_PTHRESH 8 /**< Default values of RX prefetch threshold reg. */
-#define RX_HTHRESH 8 /**< Default values of RX host threshold reg. */
-#define RX_WTHRESH 4 /**< Default values of RX write-back threshold reg. */
-
-/*
- * These default values are optimized for use with the Intel(R) 82599 10 GbE
- * Controller and the DPDK ixgbe PMD. Consider using other values for other
- * network controllers and/or network drivers.
- */
-#define TX_PTHRESH 36 /**< Default values of TX prefetch threshold reg. */
-#define TX_HTHRESH 0  /**< Default values of TX host threshold reg. */
-#define TX_WTHRESH 0  /**< Default values of TX write-back threshold reg. */
-
 #define MAX_PKT_BURST 32
 #define BURST_TX_DRAIN_US 100 /* TX drain every ~100us */
 
@@ -147,11 +127,11 @@ struct mbuf_table {
 struct lcore_queue_conf {
        unsigned n_rx_port;
        unsigned rx_port_list[MAX_RX_QUEUE_PER_LCORE];
-       struct mbuf_table tx_mbufs[RTE_MAX_ETHPORTS];
-
 } __rte_cache_aligned;
 struct lcore_queue_conf lcore_queue_conf[RTE_MAX_LCORE];
 
+struct rte_eth_dev_tx_buffer *tx_buffer[RTE_MAX_ETHPORTS];
+
 struct lcore_resource_struct {
        int enabled;    /* Only set in case this lcore involved into packet forwarding */
        int flags;          /* Set only slave need to restart or recreate */
@@ -190,25 +170,6 @@ static const struct rte_eth_conf port_conf = {
        },
 };
 
-static const struct rte_eth_rxconf rx_conf = {
-       .rx_thresh = {
-               .pthresh = RX_PTHRESH,
-               .hthresh = RX_HTHRESH,
-               .wthresh = RX_WTHRESH,
-       },
-};
-
-static const struct rte_eth_txconf tx_conf = {
-       .tx_thresh = {
-               .pthresh = TX_PTHRESH,
-               .hthresh = TX_HTHRESH,
-               .wthresh = TX_WTHRESH,
-       },
-       .tx_free_thresh = 0, /* Use PMD default values */
-       .tx_rs_thresh = 0, /* Use PMD default values */
-       .txq_flags = ETH_TXQ_FLAGS_NOMULTSEGS | ETH_TXQ_FLAGS_NOOFFLOADS,
-};
-
 static struct rte_mempool * l2fwd_pktmbuf_pool[RTE_MAX_ETHPORTS];
 
 /* Per-port statistics struct */
@@ -481,7 +442,8 @@ reset_slave_all_ports(unsigned slaveid)
                pool = rte_mempool_lookup(buf_name);
                if (pool)
                        printf("Port %d mempool free object is %u(%u)\n", slave->port[i],
-                               rte_mempool_count(pool), (unsigned)NB_MBUF);
+                               rte_mempool_avail_count(pool),
+                               (unsigned int)NB_MBUF);
                else
                        printf("Can't find mempool %s\n", buf_name);
 
@@ -622,58 +584,14 @@ slave_exit_cb(unsigned slaveid, __attribute__((unused))int stat)
        rte_spinlock_unlock(&res_lock);
 }
 
-/* Send the packet on an output interface */
-static int
-l2fwd_send_burst(struct lcore_queue_conf *qconf, unsigned n, uint8_t port)
-{
-       struct rte_mbuf **m_table;
-       unsigned ret;
-       unsigned queueid =0;
-
-       m_table = (struct rte_mbuf **)qconf->tx_mbufs[port].m_table;
-
-       ret = rte_eth_tx_burst(port, (uint16_t) queueid, m_table, (uint16_t) n);
-       port_statistics[port].tx += ret;
-       if (unlikely(ret < n)) {
-               port_statistics[port].dropped += (n - ret);
-               do {
-                       rte_pktmbuf_free(m_table[ret]);
-               } while (++ret < n);
-       }
-
-       return 0;
-}
-
-/* Send the packet on an output interface */
-static int
-l2fwd_send_packet(struct rte_mbuf *m, uint8_t port)
-{
-       unsigned lcore_id, len;
-       struct lcore_queue_conf *qconf;
-
-       lcore_id = rte_lcore_id();
-
-       qconf = &lcore_queue_conf[lcore_id];
-       len = qconf->tx_mbufs[port].len;
-       qconf->tx_mbufs[port].m_table[len] = m;
-       len++;
-
-       /* enough pkts to be sent */
-       if (unlikely(len == MAX_PKT_BURST)) {
-               l2fwd_send_burst(qconf, MAX_PKT_BURST, port);
-               len = 0;
-       }
-
-       qconf->tx_mbufs[port].len = len;
-       return 0;
-}
-
 static void
 l2fwd_simple_forward(struct rte_mbuf *m, unsigned portid)
 {
        struct ether_hdr *eth;
        void *tmp;
        unsigned dst_port;
+       int sent;
+       struct rte_eth_dev_tx_buffer *buffer;
 
        dst_port = l2fwd_dst_ports[portid];
        eth = rte_pktmbuf_mtod(m, struct ether_hdr *);
@@ -685,7 +603,10 @@ l2fwd_simple_forward(struct rte_mbuf *m, unsigned portid)
        /* src addr */
        ether_addr_copy(&l2fwd_ports_eth_addr[dst_port], &eth->s_addr);
 
-       l2fwd_send_packet(m, (uint8_t) dst_port);
+       buffer = tx_buffer[dst_port];
+       sent = rte_eth_tx_buffer(dst_port, 0, buffer, m);
+       if (sent)
+               port_statistics[dst_port].tx += sent;
 }
 
 /* main processing loop */
@@ -694,11 +615,14 @@ l2fwd_main_loop(void)
 {
        struct rte_mbuf *pkts_burst[MAX_PKT_BURST];
        struct rte_mbuf *m;
+       int sent;
        unsigned lcore_id;
        uint64_t prev_tsc, diff_tsc, cur_tsc;
        unsigned i, j, portid, nb_rx;
        struct lcore_queue_conf *qconf;
-       const uint64_t drain_tsc = (rte_get_tsc_hz() + US_PER_S - 1) / US_PER_S * BURST_TX_DRAIN_US;
+       const uint64_t drain_tsc = (rte_get_tsc_hz() + US_PER_S - 1) / US_PER_S *
+                       BURST_TX_DRAIN_US;
+       struct rte_eth_dev_tx_buffer *buffer;
 
        prev_tsc = 0;
 
@@ -738,13 +662,15 @@ l2fwd_main_loop(void)
                diff_tsc = cur_tsc - prev_tsc;
                if (unlikely(diff_tsc > drain_tsc)) {
 
-                       for (portid = 0; portid < RTE_MAX_ETHPORTS; portid++) {
-                               if (qconf->tx_mbufs[portid].len == 0)
-                                       continue;
-                               l2fwd_send_burst(&lcore_queue_conf[lcore_id],
-                                                qconf->tx_mbufs[portid].len,
-                                                (uint8_t) portid);
-                               qconf->tx_mbufs[portid].len = 0;
+                       for (i = 0; i < qconf->n_rx_port; i++) {
+
+                               portid = l2fwd_dst_ports[qconf->rx_port_list[i]];
+                               buffer = tx_buffer[portid];
+
+                               sent = rte_eth_tx_buffer_flush(portid, 0, buffer);
+                               if (sent)
+                                       port_statistics[portid].tx += sent;
+
                        }
                }
 
@@ -975,7 +901,7 @@ check_all_ports_link_status(uint8_t port_num, uint32_t port_mask)
                                continue;
                        }
                        /* clear all_ports_up flag if any link down */
-                       if (link.link_status == 0) {
+                       if (link.link_status == ETH_LINK_DOWN) {
                                all_ports_up = 0;
                                break;
                        }
@@ -999,7 +925,7 @@ check_all_ports_link_status(uint8_t port_num, uint32_t port_mask)
 }
 
 int
-MAIN(int argc, char **argv)
+main(int argc, char **argv)
 {
        struct lcore_queue_conf *qconf;
        struct rte_eth_dev_info dev_info;
@@ -1050,16 +976,10 @@ MAIN(int argc, char **argv)
        for (i = 0; i < RTE_MAX_LCORE; i++)
                lcore_resource[i].lcore_id = i;
 
-       if (rte_eal_pci_probe() < 0)
-               rte_exit(EXIT_FAILURE, "Cannot probe PCI\n");
-
        nb_ports = rte_eth_dev_count();
        if (nb_ports == 0)
                rte_exit(EXIT_FAILURE, "No Ethernet ports - bye\n");
 
-       if (nb_ports > RTE_MAX_ETHPORTS)
-               nb_ports = RTE_MAX_ETHPORTS;
-
        /* create the mbuf pool */
        for (portid = 0; portid < nb_ports; portid++) {
                /* skip ports that are not enabled */
@@ -1170,7 +1090,8 @@ MAIN(int argc, char **argv)
                /* init one RX queue */
                fflush(stdout);
                ret = rte_eth_rx_queue_setup(portid, 0, nb_rxd,
-                                            rte_eth_dev_socket_id(portid), &rx_conf,
+                                            rte_eth_dev_socket_id(portid),
+                                            NULL,
                                             l2fwd_pktmbuf_pool[portid]);
                if (ret < 0)
                        rte_exit(EXIT_FAILURE, "rte_eth_rx_queue_setup:err=%d, port=%u\n",
@@ -1179,11 +1100,29 @@ MAIN(int argc, char **argv)
                /* init one TX queue on each port */
                fflush(stdout);
                ret = rte_eth_tx_queue_setup(portid, 0, nb_txd,
-                               rte_eth_dev_socket_id(portid), &tx_conf);
+                               rte_eth_dev_socket_id(portid),
+                               NULL);
                if (ret < 0)
                        rte_exit(EXIT_FAILURE, "rte_eth_tx_queue_setup:err=%d, port=%u\n",
                                ret, (unsigned) portid);
 
+               /* Initialize TX buffers */
+               tx_buffer[portid] = rte_zmalloc_socket("tx_buffer",
+                               RTE_ETH_TX_BUFFER_SIZE(MAX_PKT_BURST), 0,
+                               rte_eth_dev_socket_id(portid));
+               if (tx_buffer[portid] == NULL)
+                       rte_exit(EXIT_FAILURE, "Cannot allocate buffer for tx on port %u\n",
+                                       (unsigned) portid);
+
+               rte_eth_tx_buffer_init(tx_buffer[portid], MAX_PKT_BURST);
+
+               ret = rte_eth_tx_buffer_set_err_callback(tx_buffer[portid],
+                               rte_eth_tx_buffer_count_callback,
+                               &port_statistics[portid].dropped);
+               if (ret < 0)
+                               rte_exit(EXIT_FAILURE, "Cannot set error callback for "
+                                               "tx buffer on port %u\n", (unsigned) portid);
+
                /* Start device */
                ret = rte_eth_dev_start(portid);
                if (ret < 0)