X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=examples%2Fqos_meter%2Fmain.c;h=6d057abfe3e746d3ec0d594ded5d1ea895fc4a7e;hb=ddcd7640ca48a13ac211f356495ec894651b87b4;hp=f5df4486521d59db3503e23398b31aab9515d986;hpb=e9d48c0072d36eb6423b45fba4ec49d0def6c36f;p=dpdk.git diff --git a/examples/qos_meter/main.c b/examples/qos_meter/main.c old mode 100755 new mode 100644 index f5df448652..6d057abfe3 --- a/examples/qos_meter/main.c +++ b/examples/qos_meter/main.c @@ -1,34 +1,5 @@ -/*- - * BSD LICENSE - * - * Copyright(c) 2010-2014 Intel Corporation. All rights reserved. - * All rights reserved. - * - * Redistribution and use in source and binary forms, with or without - * modification, are permitted provided that the following conditions - * are met: - * - * * Redistributions of source code must retain the above copyright - * notice, this list of conditions and the following disclaimer. - * * Redistributions in binary form must reproduce the above copyright - * notice, this list of conditions and the following disclaimer in - * the documentation and/or other materials provided with the - * distribution. - * * Neither the name of Intel Corporation nor the names of its - * contributors may be used to endorse or promote products derived - * from this software without specific prior written permission. - * - * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS - * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT - * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR - * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT - * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, - * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT - * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, - * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY - * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT - * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE - * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2010-2016 Intel Corporation */ #include @@ -36,6 +7,7 @@ #include #include +#include #include #include #include @@ -67,10 +39,9 @@ #endif /* - * Buffer pool configuration + * Buffer pool configuration * ***/ -#define MBUF_SIZE (2048 + sizeof(struct rte_mbuf) + RTE_PKTMBUF_HEADROOM) #define NB_MBUF 8192 #define MEMPOOL_CACHE_SIZE 256 @@ -82,18 +53,15 @@ static struct rte_mempool *pool = NULL; ***/ static struct rte_eth_conf port_conf = { .rxmode = { - .max_rx_pkt_len = ETHER_MAX_LEN, + .mq_mode = ETH_MQ_RX_RSS, + .max_rx_pkt_len = RTE_ETHER_MAX_LEN, .split_hdr_size = 0, - .header_split = 0, - .hw_ip_checksum = 1, - .hw_vlan_filter = 0, - .jumbo_frame = 0, - .hw_strip_crc = 0, + .offloads = DEV_RX_OFFLOAD_CHECKSUM, }, .rx_adv_conf = { .rss_conf = { .rss_key = NULL, - .rss_hf = ETH_RSS_IPV4 | ETH_RSS_IPV6, + .rss_hf = ETH_RSS_IP, }, }, .txmode = { @@ -101,28 +69,8 @@ static struct rte_eth_conf port_conf = { }, }; -static const struct rte_eth_rxconf rx_conf = { - .rx_thresh = { - .pthresh = 8, /* RX prefetch threshold reg */ - .hthresh = 8, /* RX host threshold reg */ - .wthresh = 4, /* RX write-back threshold reg */ - }, - .rx_free_thresh = 32, -}; - -static const struct rte_eth_txconf tx_conf = { - .tx_thresh = { - .pthresh = 36, /* TX prefetch threshold reg */ - .hthresh = 0, /* TX host threshold reg */ - .wthresh = 0, /* TX write-back threshold reg */ - }, - .tx_free_thresh = 0, - .tx_rs_thresh = 0, - .txq_flags = 0x0, -}; - -#define NIC_RX_QUEUE_DESC 128 -#define NIC_TX_QUEUE_DESC 512 +#define NIC_RX_QUEUE_DESC 1024 +#define NIC_TX_QUEUE_DESC 1024 #define NIC_RX_QUEUE 0 #define NIC_TX_QUEUE 0 @@ -135,34 +83,55 @@ static const struct rte_eth_txconf tx_conf = { #define PKT_TX_BURST_MAX 32 #define TIME_TX_DRAIN 200000ULL -static uint8_t port_rx; -static uint8_t port_tx; +static uint16_t port_rx; +static uint16_t port_tx; static struct rte_mbuf *pkts_rx[PKT_RX_BURST_MAX]; -static struct rte_mbuf *pkts_tx[PKT_TX_BURST_MAX]; -static uint16_t pkts_tx_len = 0; +struct rte_eth_dev_tx_buffer *tx_buffer; - -struct rte_meter_srtcm_params app_srtcm_params[] = { - {.cir = 1000000 * 46, .cbs = 2048, .ebs = 2048}, +struct rte_meter_srtcm_params app_srtcm_params = { + .cir = 1000000 * 46, + .cbs = 2048, + .ebs = 2048 }; -struct rte_meter_trtcm_params app_trtcm_params[] = { - {.cir = 1000000 * 46, .pir = 1500000 * 46, .cbs = 2048, .pbs = 2048}, +struct rte_meter_srtcm_profile app_srtcm_profile; + +struct rte_meter_trtcm_params app_trtcm_params = { + .cir = 1000000 * 46, + .pir = 1500000 * 46, + .cbs = 2048, + .pbs = 2048 }; -#define DIM(a) (sizeof (a) / sizeof ((a)[0])) +struct rte_meter_trtcm_profile app_trtcm_profile; + #define APP_FLOWS_MAX 256 FLOW_METER app_flows[APP_FLOWS_MAX]; -static void +static int app_configure_flow_table(void) { - uint32_t i, j; + uint32_t i; + int ret; + + ret = rte_meter_srtcm_profile_config(&app_srtcm_profile, + &app_srtcm_params); + if (ret) + return ret; - for (i = 0, j = 0; i < APP_FLOWS_MAX; i ++, j = (j + 1) % DIM(PARAMS)){ - FUNC_CONFIG(&app_flows[i], &PARAMS[j]); + ret = rte_meter_trtcm_profile_config(&app_trtcm_profile, + &app_trtcm_params); + if (ret) + return ret; + + for (i = 0; i < APP_FLOWS_MAX; i++) { + ret = FUNC_CONFIG(&app_flows[i], &PROFILE); + if (ret) + return ret; } + + return 0; } static inline void @@ -176,14 +145,18 @@ app_pkt_handle(struct rte_mbuf *pkt, uint64_t time) { uint8_t input_color, output_color; uint8_t *pkt_data = rte_pktmbuf_mtod(pkt, uint8_t *); - uint32_t pkt_len = rte_pktmbuf_pkt_len(pkt) - sizeof(struct ether_hdr); + uint32_t pkt_len = rte_pktmbuf_pkt_len(pkt) - + sizeof(struct rte_ether_hdr); uint8_t flow_id = (uint8_t)(pkt_data[APP_PKT_FLOW_POS] & (APP_FLOWS_MAX - 1)); input_color = pkt_data[APP_PKT_COLOR_POS]; enum policer_action action; /* color input is not used for blind modes */ - output_color = (uint8_t) FUNC_METER(&app_flows[flow_id], time, pkt_len, - (enum rte_meter_color) input_color); + output_color = (uint8_t) FUNC_METER(&app_flows[flow_id], + &PROFILE, + time, + pkt_len, + (enum rte_color) input_color); /* Apply policing and set the output color */ action = policer_table[input_color][output_color]; @@ -193,8 +166,8 @@ app_pkt_handle(struct rte_mbuf *pkt, uint64_t time) } -static __attribute__((noreturn)) int -main_loop(__attribute__((unused)) void *dummy) +static __rte_noreturn int +main_loop(__rte_unused void *dummy) { uint64_t current_time, last_time = rte_rdtsc(); uint32_t lcore_id = rte_lcore_id(); @@ -204,65 +177,28 @@ main_loop(__attribute__((unused)) void *dummy) while (1) { uint64_t time_diff; int i, nb_rx; - + /* Mechanism to avoid stale packets in the output buffer */ current_time = rte_rdtsc(); time_diff = current_time - last_time; if (unlikely(time_diff > TIME_TX_DRAIN)) { - int ret; - - if (pkts_tx_len == 0) { - last_time = current_time; - - continue; - } - - /* Write packet burst to NIC TX */ - ret = rte_eth_tx_burst(port_tx, NIC_TX_QUEUE, pkts_tx, pkts_tx_len); - - /* Free buffers for any packets not written successfully */ - if (unlikely(ret < pkts_tx_len)) { - for ( ; ret < pkts_tx_len; ret ++) { - rte_pktmbuf_free(pkts_tx[ret]); - } - } - - /* Empty the output buffer */ - pkts_tx_len = 0; - + /* Flush tx buffer */ + rte_eth_tx_buffer_flush(port_tx, NIC_TX_QUEUE, tx_buffer); last_time = current_time; } - + /* Read packet burst from NIC RX */ nb_rx = rte_eth_rx_burst(port_rx, NIC_RX_QUEUE, pkts_rx, PKT_RX_BURST_MAX); - + /* Handle packets */ for (i = 0; i < nb_rx; i ++) { struct rte_mbuf *pkt = pkts_rx[i]; - + /* Handle current packet */ if (app_pkt_handle(pkt, current_time) == DROP) rte_pktmbuf_free(pkt); - else { - pkts_tx[pkts_tx_len] = pkt; - pkts_tx_len ++; - } - - /* Write packets from output buffer to NIC TX when full burst is available */ - if (unlikely(pkts_tx_len == PKT_TX_BURST_MAX)) { - /* Write packet burst to NIC TX */ - int ret = rte_eth_tx_burst(port_tx, NIC_TX_QUEUE, pkts_tx, PKT_TX_BURST_MAX); - - /* Free buffers for any packets not written successfully */ - if (unlikely(ret < PKT_TX_BURST_MAX)) { - for ( ; ret < PKT_TX_BURST_MAX; ret ++) { - rte_pktmbuf_free(pkts_tx[ret]); - } - } - - /* Empty the output buffer */ - pkts_tx_len = 0; - } + else + rte_eth_tx_buffer(port_tx, NIC_TX_QUEUE, tx_buffer, pkt); } } } @@ -303,7 +239,7 @@ parse_args(int argc, char **argv) static struct option lgopts[] = { {NULL, 0, 0, 0} }; - uint64_t port_mask, i, mask; + uint64_t port_mask, i, mask; argvopt = argv; @@ -316,7 +252,7 @@ parse_args(int argc, char **argv) print_usage(prgname); return -1; } - + for (i = 0, mask = 1; i < 64; i ++, mask <<= 1){ if (mask & port_mask){ port_rx = i; @@ -332,14 +268,14 @@ parse_args(int argc, char **argv) break; } } - + if (port_mask != 0) { printf("invalid port mask (more than 2 ports)\n"); print_usage(prgname); return -1; } break; - + default: print_usage(prgname); return -1; @@ -353,14 +289,20 @@ parse_args(int argc, char **argv) argv[optind-1] = prgname; - optind = 0; /* reset getopt lib */ + optind = 1; /* reset getopt lib */ return 0; } int -MAIN(int argc, char **argv) +main(int argc, char **argv) { uint32_t lcore_id; + uint16_t nb_rxd = NIC_RX_QUEUE_DESC; + uint16_t nb_txd = NIC_TX_QUEUE_DESC; + struct rte_eth_conf conf; + struct rte_eth_rxconf rxq_conf; + struct rte_eth_txconf txq_conf; + struct rte_eth_dev_info dev_info; int ret; /* EAL init */ @@ -370,68 +312,149 @@ MAIN(int argc, char **argv) argc -= ret; argv += ret; if (rte_lcore_count() != 1) { - rte_exit(EXIT_FAILURE, "This application does not accept more than one core. " + rte_exit(EXIT_FAILURE, "This application does not accept more than one core. " "Please adjust the \"-c COREMASK\" parameter accordingly.\n"); } - + /* Application non-EAL arguments parse */ ret = parse_args(argc, argv); if (ret < 0) rte_exit(EXIT_FAILURE, "Invalid input arguments\n"); /* Buffer pool init */ - pool = rte_mempool_create("pool", NB_MBUF, MBUF_SIZE, MEMPOOL_CACHE_SIZE, - sizeof(struct rte_pktmbuf_pool_private), rte_pktmbuf_pool_init, NULL, - rte_pktmbuf_init, NULL, rte_socket_id(), 0); + pool = rte_pktmbuf_pool_create("pool", NB_MBUF, MEMPOOL_CACHE_SIZE, + 0, RTE_MBUF_DEFAULT_BUF_SIZE, rte_socket_id()); if (pool == NULL) rte_exit(EXIT_FAILURE, "Buffer pool creation error\n"); - /* PMD init */ - if (rte_pmd_init_all() < 0) - rte_exit(EXIT_FAILURE, "PMD init error\n"); - - if (rte_eal_pci_probe() < 0) - rte_exit(EXIT_FAILURE, "PCI probe error\n"); - /* NIC init */ - ret = rte_eth_dev_configure(port_rx, 1, 1, &port_conf); + conf = port_conf; + + ret = rte_eth_dev_info_get(port_rx, &dev_info); + if (ret != 0) + rte_exit(EXIT_FAILURE, + "Error during getting device (port %u) info: %s\n", + port_rx, strerror(-ret)); + + if (dev_info.tx_offload_capa & DEV_TX_OFFLOAD_MBUF_FAST_FREE) + conf.txmode.offloads |= DEV_TX_OFFLOAD_MBUF_FAST_FREE; + + conf.rx_adv_conf.rss_conf.rss_hf &= dev_info.flow_type_rss_offloads; + if (conf.rx_adv_conf.rss_conf.rss_hf != + port_conf.rx_adv_conf.rss_conf.rss_hf) { + printf("Port %u modified RSS hash function based on hardware support," + "requested:%#"PRIx64" configured:%#"PRIx64"\n", + port_rx, + port_conf.rx_adv_conf.rss_conf.rss_hf, + conf.rx_adv_conf.rss_conf.rss_hf); + } + + ret = rte_eth_dev_configure(port_rx, 1, 1, &conf); if (ret < 0) rte_exit(EXIT_FAILURE, "Port %d configuration error (%d)\n", port_rx, ret); - ret = rte_eth_rx_queue_setup(port_rx, NIC_RX_QUEUE, NIC_RX_QUEUE_DESC, rte_eth_dev_socket_id(port_rx), &rx_conf, pool); + ret = rte_eth_dev_adjust_nb_rx_tx_desc(port_rx, &nb_rxd, &nb_txd); + if (ret < 0) + rte_exit(EXIT_FAILURE, "Port %d adjust number of descriptors error (%d)\n", + port_rx, ret); + + rxq_conf = dev_info.default_rxconf; + rxq_conf.offloads = conf.rxmode.offloads; + ret = rte_eth_rx_queue_setup(port_rx, NIC_RX_QUEUE, nb_rxd, + rte_eth_dev_socket_id(port_rx), + &rxq_conf, pool); if (ret < 0) rte_exit(EXIT_FAILURE, "Port %d RX queue setup error (%d)\n", port_rx, ret); - - ret = rte_eth_tx_queue_setup(port_rx, NIC_TX_QUEUE, NIC_TX_QUEUE_DESC, rte_eth_dev_socket_id(port_rx), &tx_conf); + + txq_conf = dev_info.default_txconf; + txq_conf.offloads = conf.txmode.offloads; + ret = rte_eth_tx_queue_setup(port_rx, NIC_TX_QUEUE, nb_txd, + rte_eth_dev_socket_id(port_rx), + &txq_conf); if (ret < 0) rte_exit(EXIT_FAILURE, "Port %d TX queue setup error (%d)\n", port_rx, ret); - ret = rte_eth_dev_configure(port_tx, 1, 1, &port_conf); + conf = port_conf; + + ret = rte_eth_dev_info_get(port_tx, &dev_info); + if (ret != 0) + rte_exit(EXIT_FAILURE, + "Error during getting device (port %u) info: %s\n", + port_tx, strerror(-ret)); + + if (dev_info.tx_offload_capa & DEV_TX_OFFLOAD_MBUF_FAST_FREE) + conf.txmode.offloads |= DEV_TX_OFFLOAD_MBUF_FAST_FREE; + + conf.rx_adv_conf.rss_conf.rss_hf &= dev_info.flow_type_rss_offloads; + if (conf.rx_adv_conf.rss_conf.rss_hf != + port_conf.rx_adv_conf.rss_conf.rss_hf) { + printf("Port %u modified RSS hash function based on hardware support," + "requested:%#"PRIx64" configured:%#"PRIx64"\n", + port_tx, + port_conf.rx_adv_conf.rss_conf.rss_hf, + conf.rx_adv_conf.rss_conf.rss_hf); + } + + ret = rte_eth_dev_configure(port_tx, 1, 1, &conf); if (ret < 0) rte_exit(EXIT_FAILURE, "Port %d configuration error (%d)\n", port_tx, ret); - ret = rte_eth_rx_queue_setup(port_tx, NIC_RX_QUEUE, NIC_RX_QUEUE_DESC, rte_eth_dev_socket_id(port_tx), &rx_conf, pool); + nb_rxd = NIC_RX_QUEUE_DESC; + nb_txd = NIC_TX_QUEUE_DESC; + ret = rte_eth_dev_adjust_nb_rx_tx_desc(port_tx, &nb_rxd, &nb_txd); + if (ret < 0) + rte_exit(EXIT_FAILURE, "Port %d adjust number of descriptors error (%d)\n", + port_tx, ret); + + rxq_conf = dev_info.default_rxconf; + rxq_conf.offloads = conf.rxmode.offloads; + ret = rte_eth_rx_queue_setup(port_tx, NIC_RX_QUEUE, nb_rxd, + rte_eth_dev_socket_id(port_tx), + NULL, pool); if (ret < 0) rte_exit(EXIT_FAILURE, "Port %d RX queue setup error (%d)\n", port_tx, ret); - ret = rte_eth_tx_queue_setup(port_tx, NIC_TX_QUEUE, NIC_TX_QUEUE_DESC, rte_eth_dev_socket_id(port_tx), &tx_conf); + txq_conf = dev_info.default_txconf; + txq_conf.offloads = conf.txmode.offloads; + ret = rte_eth_tx_queue_setup(port_tx, NIC_TX_QUEUE, nb_txd, + rte_eth_dev_socket_id(port_tx), + NULL); if (ret < 0) rte_exit(EXIT_FAILURE, "Port %d TX queue setup error (%d)\n", port_tx, ret); + tx_buffer = rte_zmalloc_socket("tx_buffer", + RTE_ETH_TX_BUFFER_SIZE(PKT_TX_BURST_MAX), 0, + rte_eth_dev_socket_id(port_tx)); + if (tx_buffer == NULL) + rte_exit(EXIT_FAILURE, "Port %d TX buffer allocation error\n", + port_tx); + + rte_eth_tx_buffer_init(tx_buffer, PKT_TX_BURST_MAX); + ret = rte_eth_dev_start(port_rx); if (ret < 0) rte_exit(EXIT_FAILURE, "Port %d start error (%d)\n", port_rx, ret); - + ret = rte_eth_dev_start(port_tx); if (ret < 0) rte_exit(EXIT_FAILURE, "Port %d start error (%d)\n", port_tx, ret); - rte_eth_promiscuous_enable(port_rx); + ret = rte_eth_promiscuous_enable(port_rx); + if (ret != 0) + rte_exit(EXIT_FAILURE, + "Port %d promiscuous mode enable error (%s)\n", + port_rx, rte_strerror(-ret)); + + ret = rte_eth_promiscuous_enable(port_tx); + if (ret != 0) + rte_exit(EXIT_FAILURE, + "Port %d promiscuous mode enable error (%s)\n", + port_rx, rte_strerror(-ret)); - rte_eth_promiscuous_enable(port_tx); - /* App configuration */ - app_configure_flow_table(); + ret = app_configure_flow_table(); + if (ret < 0) + rte_exit(EXIT_FAILURE, "Invalid configure flow table\n"); /* Launch per-lcore init on every lcore */ rte_eal_mp_remote_launch(main_loop, NULL, CALL_MASTER);