remove unused ring includes
[dpdk.git] / examples / l3fwd-acl / main.c
index 4487c95..3cfbb40 100644 (file)
@@ -1,7 +1,7 @@
 /*-
  *   BSD LICENSE
  *
- *   Copyright(c) 2010-2014 Intel Corporation. All rights reserved.
+ *   Copyright(c) 2010-2016 Intel Corporation. All rights reserved.
  *   All rights reserved.
  *
  *   Redistribution and use in source and binary forms, with or without
@@ -48,7 +48,6 @@
 #include <rte_memory.h>
 #include <rte_memcpy.h>
 #include <rte_memzone.h>
-#include <rte_tailq.h>
 #include <rte_eal.h>
 #include <rte_per_lcore.h>
 #include <rte_launch.h>
@@ -64,7 +63,6 @@
 #include <rte_debug.h>
 #include <rte_ether.h>
 #include <rte_ethdev.h>
-#include <rte_ring.h>
 #include <rte_mempool.h>
 #include <rte_mbuf.h>
 #include <rte_ip.h>
@@ -73,6 +71,9 @@
 #include <rte_string_fns.h>
 #include <rte_acl.h>
 
+#if RTE_LOG_LEVEL >= RTE_LOG_DEBUG
+#define L3FWDACL_DEBUG
+#endif
 #define DO_RFC_1812_CHECKS
 
 #define RTE_LOGTYPE_L3FWD RTE_LOGTYPE_USER1
@@ -81,8 +82,6 @@
 
 #define MEMPOOL_CACHE_SIZE 256
 
-#define MBUF_SIZE (2048 + sizeof(struct rte_mbuf) + RTE_PKTMBUF_HEADROOM)
-
 /*
  * This expression is used to calculate the number of mbufs needed
  * depending on user input, taking into account memory for rx and tx hardware
@@ -122,11 +121,6 @@ static uint32_t enabled_port_mask;
 static int promiscuous_on; /**< Ports set in promiscuous mode off by default. */
 static int numa_on = 1; /**< NUMA is enabled by default. */
 
-struct mbuf_table {
-       uint16_t len;
-       struct rte_mbuf *m_table[MAX_PKT_BURST];
-};
-
 struct lcore_rx_queue {
        uint8_t port_id;
        uint8_t queue_id;
@@ -174,11 +168,8 @@ static struct rte_eth_conf port_conf = {
        .rx_adv_conf = {
                .rss_conf = {
                        .rss_key = NULL,
-                       .rss_hf = ETH_RSS_IPV4 | ETH_RSS_IPV4_TCP
-                               | ETH_RSS_IPV4_UDP
-                               | ETH_RSS_IPV6 | ETH_RSS_IPV6_EX
-                               | ETH_RSS_IPV6_TCP | ETH_RSS_IPV6_TCP_EX
-                               | ETH_RSS_IPV6_UDP | ETH_RSS_IPV6_UDP_EX,
+                       .rss_hf = ETH_RSS_IP | ETH_RSS_UDP |
+                               ETH_RSS_TCP | ETH_RSS_SCTP,
                },
        },
        .txmode = {
@@ -193,7 +184,7 @@ static struct rte_mempool *pktmbuf_pool[NB_SOCKETS];
 static inline int
 is_valid_ipv4_pkt(struct ipv4_hdr *pkt, uint32_t link_len);
 #endif
-static inline int
+static inline void
 send_single_packet(struct rte_mbuf *m, uint8_t port);
 
 #define MAX_ACL_RULE_NUM       100000
@@ -222,9 +213,9 @@ send_single_packet(struct rte_mbuf *m, uint8_t port);
 #define OFF_IPV42PROTO (offsetof(struct ipv4_hdr, next_proto_id))
 #define OFF_IPV62PROTO (offsetof(struct ipv6_hdr, proto))
 #define MBUF_IPV4_2PROTO(m)    \
-       (rte_pktmbuf_mtod((m), uint8_t *) + OFF_ETHHEAD + OFF_IPV42PROTO)
+       rte_pktmbuf_mtod_offset((m), uint8_t *, OFF_ETHHEAD + OFF_IPV42PROTO)
 #define MBUF_IPV6_2PROTO(m)    \
-       (rte_pktmbuf_mtod((m), uint8_t *) + OFF_ETHHEAD + OFF_IPV62PROTO)
+       rte_pktmbuf_mtod_offset((m), uint8_t *, OFF_ETHHEAD + OFF_IPV62PROTO)
 
 #define GET_CB_FIELD(in, fd, base, lim, dlm)   do {            \
        unsigned long val;                                      \
@@ -267,6 +258,23 @@ enum {
        NUM_FIELDS_IPV4
 };
 
+/*
+ * That effectively defines order of IPV4VLAN classifications:
+ *  - PROTO
+ *  - VLAN (TAG and DOMAIN)
+ *  - SRC IP ADDRESS
+ *  - DST IP ADDRESS
+ *  - PORTS (SRC and DST)
+ */
+enum {
+       RTE_ACL_IPV4VLAN_PROTO,
+       RTE_ACL_IPV4VLAN_VLAN,
+       RTE_ACL_IPV4VLAN_SRC,
+       RTE_ACL_IPV4VLAN_DST,
+       RTE_ACL_IPV4VLAN_PORTS,
+       RTE_ACL_IPV4VLAN_NUM
+};
+
 struct rte_acl_field_def ipv4_defs[NUM_FIELDS_IPV4] = {
        {
                .type = RTE_ACL_FIELD_TYPE_BITMASK,
@@ -570,9 +578,9 @@ dump_acl4_rule(struct rte_mbuf *m, uint32_t sig)
 {
        uint32_t offset = sig & ~ACL_DENY_SIGNATURE;
        unsigned char a, b, c, d;
-       struct ipv4_hdr *ipv4_hdr = (struct ipv4_hdr *)
-                                       (rte_pktmbuf_mtod(m, unsigned char *) +
-                                       sizeof(struct ether_hdr));
+       struct ipv4_hdr *ipv4_hdr = rte_pktmbuf_mtod_offset(m,
+                                                           struct ipv4_hdr *,
+                                                           sizeof(struct ether_hdr));
 
        uint32_t_to_char(rte_bswap32(ipv4_hdr->src_addr), &a, &b, &c, &d);
        printf("Packet Src:%hhu.%hhu.%hhu.%hhu ", a, b, c, d);
@@ -594,9 +602,9 @@ dump_acl6_rule(struct rte_mbuf *m, uint32_t sig)
 {
        unsigned i;
        uint32_t offset = sig & ~ACL_DENY_SIGNATURE;
-       struct ipv6_hdr *ipv6_hdr = (struct ipv6_hdr *)
-                                       (rte_pktmbuf_mtod(m, unsigned char *) +
-                                       sizeof(struct ether_hdr));
+       struct ipv6_hdr *ipv6_hdr = rte_pktmbuf_mtod_offset(m,
+                                                           struct ipv6_hdr *,
+                                                           sizeof(struct ether_hdr));
 
        printf("Packet Src");
        for (i = 0; i < RTE_DIM(ipv6_hdr->src_addr); i += sizeof(uint16_t))
@@ -651,12 +659,9 @@ prepare_one_packet(struct rte_mbuf **pkts_in, struct acl_search_t *acl,
        struct ipv4_hdr *ipv4_hdr;
        struct rte_mbuf *pkt = pkts_in[index];
 
-       int type = pkt->ol_flags & (PKT_RX_IPV4_HDR | PKT_RX_IPV6_HDR);
-
-       if (type == PKT_RX_IPV4_HDR) {
-
-               ipv4_hdr = (struct ipv4_hdr *)(rte_pktmbuf_mtod(pkt,
-                       unsigned char *) + sizeof(struct ether_hdr));
+       if (RTE_ETH_IS_IPV4_HDR(pkt->packet_type)) {
+               ipv4_hdr = rte_pktmbuf_mtod_offset(pkt, struct ipv4_hdr *,
+                                                  sizeof(struct ether_hdr));
 
                /* Check to make sure the packet is valid (RFC1812) */
                if (is_valid_ipv4_pkt(ipv4_hdr, pkt->pkt_len) >= 0) {
@@ -673,9 +678,7 @@ prepare_one_packet(struct rte_mbuf **pkts_in, struct acl_search_t *acl,
                        /* Not a valid IPv4 packet */
                        rte_pktmbuf_free(pkt);
                }
-
-       } else if (type == PKT_RX_IPV6_HDR) {
-
+       } else if (RTE_ETH_IS_IPV6_HDR(pkt->packet_type)) {
                /* Fill acl structure */
                acl->data_ipv6[acl->num_ipv6] = MBUF_IPV6_2PROTO(pkt);
                acl->m_ipv6[(acl->num_ipv6)++] = pkt;
@@ -693,17 +696,12 @@ prepare_one_packet(struct rte_mbuf **pkts_in, struct acl_search_t *acl,
 {
        struct rte_mbuf *pkt = pkts_in[index];
 
-       int type = pkt->ol_flags & (PKT_RX_IPV4_HDR | PKT_RX_IPV6_HDR);
-
-       if (type == PKT_RX_IPV4_HDR) {
-
+       if (RTE_ETH_IS_IPV4_HDR(pkt->packet_type)) {
                /* Fill acl structure */
                acl->data_ipv4[acl->num_ipv4] = MBUF_IPV4_2PROTO(pkt);
                acl->m_ipv4[(acl->num_ipv4)++] = pkt;
 
-
-       } else if (type == PKT_RX_IPV6_HDR) {
-
+       } else if (RTE_ETH_IS_IPV6_HDR(pkt->packet_type)) {
                /* Fill acl structure */
                acl->data_ipv6[acl->num_ipv6] = MBUF_IPV6_2PROTO(pkt);
                acl->m_ipv6[(acl->num_ipv6)++] = pkt;
@@ -751,9 +749,9 @@ send_one_packet(struct rte_mbuf *m, uint32_t res)
                /* in the ACL list, drop it */
 #ifdef L3FWDACL_DEBUG
                if ((res & ACL_DENY_SIGNATURE) != 0) {
-                       if (m->ol_flags & PKT_RX_IPV4_HDR)
+                       if (RTE_ETH_IS_IPV4_HDR(m->packet_type))
                                dump_acl4_rule(m, res);
-                       else
+                       else if (RTE_ETH_IS_IPV6_HDR(m->packet_type))
                                dump_acl6_rule(m, res);
                }
 #endif
@@ -1047,13 +1045,13 @@ add_rules(const char *rule_path,
 
        fseek(fh, 0, SEEK_SET);
 
-       acl_rules = (uint8_t *)calloc(acl_num, rule_size);
+       acl_rules = calloc(acl_num, rule_size);
 
        if (NULL == acl_rules)
                rte_exit(EXIT_FAILURE, "%s: failed to malloc memory\n",
                        __func__);
 
-       route_rules = (uint8_t *)calloc(route_num, rule_size);
+       route_rules = calloc(route_num, rule_size);
 
        if (NULL == route_rules)
                rte_exit(EXIT_FAILURE, "%s: failed to malloc memory\n",
@@ -1178,8 +1176,9 @@ setup_acl(struct rte_acl_rule *route_base,
                        rte_exit(EXIT_FAILURE, "add rules failed\n");
 
        /* Perform builds */
-       acl_build_param.num_categories = DEFAULT_MAX_CATEGORIES;
+       memset(&acl_build_param, 0, sizeof(acl_build_param));
 
+       acl_build_param.num_categories = DEFAULT_MAX_CATEGORIES;
        acl_build_param.num_fields = dim;
        memcpy(&acl_build_param.defs, ipv6 ? ipv6_defs : ipv4_defs,
                ipv6 ? sizeof(ipv6_defs) : sizeof(ipv4_defs));
@@ -1247,6 +1246,10 @@ app_acl_init(void)
                                acl_log("Socket %d of lcore %u is out "
                                        "of range %d\n",
                                        socketid, lcore_id, NB_SOCKETS);
+                               free(route_base_ipv4);
+                               free(route_base_ipv6);
+                               free(acl_base_ipv4);
+                               free(acl_base_ipv6);
                                return -1;
                        }
 
@@ -1285,56 +1288,26 @@ app_acl_init(void)
 struct lcore_conf {
        uint16_t n_rx_queue;
        struct lcore_rx_queue rx_queue_list[MAX_RX_QUEUE_PER_LCORE];
+       uint16_t n_tx_port;
+       uint16_t tx_port_id[RTE_MAX_ETHPORTS];
        uint16_t tx_queue_id[RTE_MAX_ETHPORTS];
-       struct mbuf_table tx_mbufs[RTE_MAX_ETHPORTS];
+       struct rte_eth_dev_tx_buffer *tx_buffer[RTE_MAX_ETHPORTS];
 } __rte_cache_aligned;
 
 static struct lcore_conf lcore_conf[RTE_MAX_LCORE];
 
-/* Send burst of packets on an output interface */
-static inline int
-send_burst(struct lcore_conf *qconf, uint16_t n, uint8_t port)
-{
-       struct rte_mbuf **m_table;
-       int ret;
-       uint16_t queueid;
-
-       queueid = qconf->tx_queue_id[port];
-       m_table = (struct rte_mbuf **)qconf->tx_mbufs[port].m_table;
-
-       ret = rte_eth_tx_burst(port, queueid, m_table, n);
-       if (unlikely(ret < n)) {
-               do {
-                       rte_pktmbuf_free(m_table[ret]);
-               } while (++ret < n);
-       }
-
-       return 0;
-}
-
 /* Enqueue a single packet, and send burst if queue is filled */
-static inline int
+static inline void
 send_single_packet(struct rte_mbuf *m, uint8_t port)
 {
        uint32_t lcore_id;
-       uint16_t len;
        struct lcore_conf *qconf;
 
        lcore_id = rte_lcore_id();
 
        qconf = &lcore_conf[lcore_id];
-       len = qconf->tx_mbufs[port].len;
-       qconf->tx_mbufs[port].m_table[len] = m;
-       len++;
-
-       /* enough pkts to be sent */
-       if (unlikely(len == MAX_PKT_BURST)) {
-               send_burst(qconf, MAX_PKT_BURST, port);
-               len = 0;
-       }
-
-       qconf->tx_mbufs[port].len = len;
-       return 0;
+       rte_eth_tx_buffer(port, qconf->tx_queue_id[port],
+                       qconf->tx_buffer[port], m);
 }
 
 #ifdef DO_RFC_1812_CHECKS
@@ -1422,20 +1395,12 @@ main_loop(__attribute__((unused)) void *dummy)
                 */
                diff_tsc = cur_tsc - prev_tsc;
                if (unlikely(diff_tsc > drain_tsc)) {
-
-                       /*
-                        * This could be optimized (use queueid instead of
-                        * portid), but it is not called so often
-                        */
-                       for (portid = 0; portid < RTE_MAX_ETHPORTS; portid++) {
-                               if (qconf->tx_mbufs[portid].len == 0)
-                                       continue;
-                               send_burst(&lcore_conf[lcore_id],
-                                       qconf->tx_mbufs[portid].len,
-                                       portid);
-                               qconf->tx_mbufs[portid].len = 0;
+                       for (i = 0; i < qconf->n_tx_port; ++i) {
+                               portid = qconf->tx_port_id[i];
+                               rte_eth_tx_buffer_flush(portid,
+                                               qconf->tx_queue_id[portid],
+                                               qconf->tx_buffer[portid]);
                        }
-
                        prev_tsc = cur_tsc;
                }
 
@@ -1847,12 +1812,10 @@ init_mem(unsigned nb_mbuf)
                if (pktmbuf_pool[socketid] == NULL) {
                        snprintf(s, sizeof(s), "mbuf_pool_%d", socketid);
                        pktmbuf_pool[socketid] =
-                               rte_mempool_create(s, nb_mbuf, MBUF_SIZE,
-                                       MEMPOOL_CACHE_SIZE,
-                                       sizeof(struct rte_pktmbuf_pool_private),
-                                       rte_pktmbuf_pool_init, NULL,
-                                       rte_pktmbuf_init, NULL,
-                                       socketid, 0);
+                               rte_pktmbuf_pool_create(s, nb_mbuf,
+                                       MEMPOOL_CACHE_SIZE, 0,
+                                       RTE_MBUF_DEFAULT_BUF_SIZE,
+                                       socketid);
                        if (pktmbuf_pool[socketid] == NULL)
                                rte_exit(EXIT_FAILURE,
                                        "Cannot init mbuf pool on socket %d\n",
@@ -1897,7 +1860,7 @@ check_all_ports_link_status(uint8_t port_num, uint32_t port_mask)
                                continue;
                        }
                        /* clear all_ports_up flag if any link down */
-                       if (link.link_status == 0) {
+                       if (link.link_status == ETH_LINK_DOWN) {
                                all_ports_up = 0;
                                break;
                        }
@@ -1953,8 +1916,6 @@ main(int argc, char **argv)
                rte_exit(EXIT_FAILURE, "init_lcore_rx_queues failed\n");
 
        nb_ports = rte_eth_dev_count();
-       if (nb_ports > RTE_MAX_ETHPORTS)
-               nb_ports = RTE_MAX_ETHPORTS;
 
        if (check_port_config(nb_ports) < 0)
                rte_exit(EXIT_FAILURE, "check_port_config failed\n");
@@ -1999,6 +1960,22 @@ main(int argc, char **argv)
                if (ret < 0)
                        rte_exit(EXIT_FAILURE, "init_mem failed\n");
 
+               for (lcore_id = 0; lcore_id < RTE_MAX_LCORE; lcore_id++) {
+                       if (rte_lcore_is_enabled(lcore_id) == 0)
+                               continue;
+
+                       /* Initialize TX buffers */
+                       qconf = &lcore_conf[lcore_id];
+                       qconf->tx_buffer[portid] = rte_zmalloc_socket("tx_buffer",
+                                       RTE_ETH_TX_BUFFER_SIZE(MAX_PKT_BURST), 0,
+                                       rte_eth_dev_socket_id(portid));
+                       if (qconf->tx_buffer[portid] == NULL)
+                               rte_exit(EXIT_FAILURE, "Can't allocate tx buffer for port %u\n",
+                                               (unsigned) portid);
+
+                       rte_eth_tx_buffer_init(qconf->tx_buffer[portid], MAX_PKT_BURST);
+               }
+
                /* init one TX queue per couple (lcore,port) */
                queueid = 0;
                for (lcore_id = 0; lcore_id < RTE_MAX_LCORE; lcore_id++) {
@@ -2028,6 +2005,9 @@ main(int argc, char **argv)
                        qconf = &lcore_conf[lcore_id];
                        qconf->tx_queue_id[portid] = queueid;
                        queueid++;
+
+                       qconf->tx_port_id[qconf->n_tx_port] = portid;
+                       qconf->n_tx_port++;
                }
                printf("\n");
        }