X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=examples%2Flink_status_interrupt%2Fmain.c;h=9997c58f4a81770e794aa3ac8d5ec0c351d6a50a;hb=6d13ea8e8e49ab957deae2bba5ecf4a4bfe747d1;hp=b6a46223d7a63eade958a8a6bf4088de08ae3c4e;hpb=041cdbf947fa7eb732e67b1c691c2572c7513461;p=dpdk.git diff --git a/examples/link_status_interrupt/main.c b/examples/link_status_interrupt/main.c index b6a46223d7..9997c58f4a 100644 --- a/examples/link_status_interrupt/main.c +++ b/examples/link_status_interrupt/main.c @@ -1,35 +1,5 @@ -/*- - * BSD LICENSE - * - * Copyright(c) 2010-2012 Intel Corporation. All rights reserved. - * All rights reserved. - * - * Redistribution and use in source and binary forms, with or without - * modification, are permitted provided that the following conditions - * are met: - * - * * Redistributions of source code must retain the above copyright - * notice, this list of conditions and the following disclaimer. - * * Redistributions in binary form must reproduce the above copyright - * notice, this list of conditions and the following disclaimer in - * the documentation and/or other materials provided with the - * distribution. - * * Neither the name of Intel Corporation nor the names of its - * contributors may be used to endorse or promote products derived - * from this software without specific prior written permission. - * - * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS - * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT - * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR - * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT - * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, - * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT - * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, - * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY - * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT - * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE - * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - * +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2010-2016 Intel Corporation */ #include @@ -38,7 +8,6 @@ #include #include #include -#include #include #include #include @@ -49,12 +18,10 @@ #include #include +#include #include #include -#include -#include #include -#include #include #include #include @@ -63,58 +30,30 @@ #include #include #include -#include #include #include #include #include -#include #include #include -#include "main.h" - #define RTE_LOGTYPE_LSI RTE_LOGTYPE_USER1 -#define LSI_MAX_PORTS 32 - -#define MBUF_SIZE (2048 + sizeof(struct rte_mbuf) + RTE_PKTMBUF_HEADROOM) #define NB_MBUF 8192 -/* - * RX and TX Prefetch, Host, and Write-back threshold values should be - * carefully set for optimal performance. Consult the network - * controller's datasheet and supporting DPDK documentation for guidance - * on how these parameters should be set. - */ -#define RX_PTHRESH 8 /**< Default values of RX prefetch threshold reg. */ -#define RX_HTHRESH 8 /**< Default values of RX host threshold reg. */ -#define RX_WTHRESH 4 /**< Default values of RX write-back threshold reg. */ - -/* - * These default values are optimized for use with the Intel(R) 82599 10 GbE - * Controller and the DPDK ixgbe PMD. Consider using other values for other - * network controllers and/or network drivers. - */ -#define TX_PTHRESH 36 /**< Default values of TX prefetch threshold reg. */ -#define TX_HTHRESH 0 /**< Default values of TX host threshold reg. */ -#define TX_WTHRESH 0 /**< Default values of TX write-back threshold reg. */ - #define MAX_PKT_BURST 32 -#define BURST_TX_DRAIN 200000ULL /* around 100us at 2 Ghz */ - -#define SOCKET0 0 +#define BURST_TX_DRAIN_US 100 /* TX drain every ~100us */ /* * Configurable number of RX/TX ring descriptors */ -#define RTE_TEST_RX_DESC_DEFAULT 128 -#define RTE_TEST_TX_DESC_DEFAULT 512 +#define RTE_TEST_RX_DESC_DEFAULT 1024 +#define RTE_TEST_TX_DESC_DEFAULT 1024 static uint16_t nb_rxd = RTE_TEST_RX_DESC_DEFAULT; static uint16_t nb_txd = RTE_TEST_TX_DESC_DEFAULT; /* ethernet addresses of ports */ -static struct ether_addr lsi_ports_eth_addr[LSI_MAX_PORTS]; +static struct rte_ether_addr lsi_ports_eth_addr[RTE_MAX_ETHPORTS]; /* mask of enabled ports */ static uint32_t lsi_enabled_port_mask = 0; @@ -122,60 +61,33 @@ static uint32_t lsi_enabled_port_mask = 0; static unsigned int lsi_rx_queue_per_lcore = 1; /* destination port for L2 forwarding */ -static unsigned lsi_dst_ports[LSI_MAX_PORTS] = {0}; +static unsigned lsi_dst_ports[RTE_MAX_ETHPORTS] = {0}; #define MAX_PKT_BURST 32 -struct mbuf_table { - unsigned len; - struct rte_mbuf *m_table[MAX_PKT_BURST]; -}; #define MAX_RX_QUEUE_PER_LCORE 16 #define MAX_TX_QUEUE_PER_PORT 16 struct lcore_queue_conf { - unsigned n_rx_queue; - unsigned rx_queue_list[MAX_RX_QUEUE_PER_LCORE]; + unsigned n_rx_port; + unsigned rx_port_list[MAX_RX_QUEUE_PER_LCORE]; unsigned tx_queue_id; - struct mbuf_table tx_mbufs[LSI_MAX_PORTS]; - } __rte_cache_aligned; struct lcore_queue_conf lcore_queue_conf[RTE_MAX_LCORE]; -static const struct rte_eth_conf port_conf = { +struct rte_eth_dev_tx_buffer *tx_buffer[RTE_MAX_ETHPORTS]; + +static struct rte_eth_conf port_conf = { .rxmode = { .split_hdr_size = 0, - .header_split = 0, /**< Header Split disabled */ - .hw_ip_checksum = 0, /**< IP checksum offload disabled */ - .hw_vlan_filter = 0, /**< VLAN filtering disabled */ - .jumbo_frame = 0, /**< Jumbo Frame Support disabled */ - .hw_strip_crc = 0, /**< CRC stripped by hardware */ }, .txmode = { - .mq_mode = ETH_DCB_NONE, + .mq_mode = ETH_MQ_TX_NONE, }, .intr_conf = { .lsc = 1, /**< lsc interrupt feature enabled */ }, }; -static const struct rte_eth_rxconf rx_conf = { - .rx_thresh = { - .pthresh = RX_PTHRESH, - .hthresh = RX_HTHRESH, - .wthresh = RX_WTHRESH, - }, -}; - -static const struct rte_eth_txconf tx_conf = { - .tx_thresh = { - .pthresh = TX_PTHRESH, - .hthresh = TX_HTHRESH, - .wthresh = TX_WTHRESH, - }, - .tx_free_thresh = 0, /* Use PMD default values */ - .tx_rs_thresh = 0, /* Use PMD default values */ -}; - struct rte_mempool * lsi_pktmbuf_pool = NULL; /* Per-port statistics struct */ @@ -184,7 +96,7 @@ struct lsi_port_statistics { uint64_t rx; uint64_t dropped; } __rte_cache_aligned; -struct lsi_port_statistics port_statistics[LSI_MAX_PORTS]; +struct lsi_port_statistics port_statistics[RTE_MAX_ETHPORTS]; /* A tsc-based timer responsible for triggering statistics printout */ #define TIMER_MILLISECOND 2000000ULL /* around 1ms at 2 Ghz */ @@ -197,7 +109,7 @@ print_stats(void) { struct rte_eth_link link; uint64_t total_packets_dropped, total_packets_tx, total_packets_rx; - unsigned portid; + uint16_t portid; total_packets_dropped = 0; total_packets_tx = 0; @@ -211,13 +123,13 @@ print_stats(void) printf("\nPort statistics ===================================="); - for (portid = 0; portid < LSI_MAX_PORTS; portid++) { + for (portid = 0; portid < RTE_MAX_ETHPORTS; portid++) { /* skip ports that are not enabled */ if ((lsi_enabled_port_mask & (1 << portid)) == 0) continue; memset(&link, 0, sizeof(link)); - rte_eth_link_get_nowait((uint8_t)portid, &link); + rte_eth_link_get_nowait(portid, &link); printf("\nStatistics for port %u ------------------------------" "\nLink status: %25s" "\nLink speed: %26u" @@ -248,70 +160,28 @@ print_stats(void) printf("\n====================================================\n"); } -/* Send the packet on an output interface */ -static int -lsi_send_burst(struct lcore_queue_conf *qconf, unsigned n, uint8_t port) -{ - struct rte_mbuf **m_table; - unsigned ret; - unsigned queueid; - - queueid = (uint16_t) qconf->tx_queue_id; - m_table = (struct rte_mbuf **)qconf->tx_mbufs[port].m_table; - - ret = rte_eth_tx_burst(port, (uint16_t) queueid, m_table, (uint16_t) n); - port_statistics[port].tx += ret; - if (unlikely(ret < n)) { - port_statistics[port].dropped += (n - ret); - do { - rte_pktmbuf_free(m_table[ret]); - } while (++ret < n); - } - - return 0; -} - -/* Send the packet on an output interface */ -static int -lsi_send_packet(struct rte_mbuf *m, uint8_t port) -{ - unsigned lcore_id, len; - struct lcore_queue_conf *qconf; - - lcore_id = rte_lcore_id(); - - qconf = &lcore_queue_conf[lcore_id]; - len = qconf->tx_mbufs[port].len; - qconf->tx_mbufs[port].m_table[len] = m; - len++; - - /* enough pkts to be sent */ - if (unlikely(len == MAX_PKT_BURST)) { - lsi_send_burst(qconf, MAX_PKT_BURST, port); - len = 0; - } - - qconf->tx_mbufs[port].len = len; - return 0; -} - static void lsi_simple_forward(struct rte_mbuf *m, unsigned portid) { - struct ether_hdr *eth; + struct rte_ether_hdr *eth; void *tmp; unsigned dst_port = lsi_dst_ports[portid]; + int sent; + struct rte_eth_dev_tx_buffer *buffer; - eth = rte_pktmbuf_mtod(m, struct ether_hdr *); + eth = rte_pktmbuf_mtod(m, struct rte_ether_hdr *); - /* 00:09:c0:00:00:xx */ + /* 02:00:00:00:00:xx */ tmp = ð->d_addr.addr_bytes[0]; - *((uint64_t *)tmp) = 0x000000c00900 + (dst_port << 24); + *((uint64_t *)tmp) = 0x000000000002 + ((uint64_t)dst_port << 40); /* src addr */ ether_addr_copy(&lsi_ports_eth_addr[dst_port], ð->s_addr); - lsi_send_packet(m, (uint8_t) dst_port); + buffer = tx_buffer[dst_port]; + sent = rte_eth_tx_buffer(dst_port, 0, buffer, m); + if (sent) + port_statistics[dst_port].tx += sent; } /* main processing loop */ @@ -321,26 +191,30 @@ lsi_main_loop(void) struct rte_mbuf *pkts_burst[MAX_PKT_BURST]; struct rte_mbuf *m; unsigned lcore_id; - uint64_t prev_tsc = 0; - uint64_t diff_tsc, cur_tsc, timer_tsc; + unsigned sent; + uint64_t prev_tsc, diff_tsc, cur_tsc, timer_tsc; unsigned i, j, portid, nb_rx; struct lcore_queue_conf *qconf; + const uint64_t drain_tsc = (rte_get_tsc_hz() + US_PER_S - 1) / US_PER_S * + BURST_TX_DRAIN_US; + struct rte_eth_dev_tx_buffer *buffer; + prev_tsc = 0; timer_tsc = 0; lcore_id = rte_lcore_id(); qconf = &lcore_queue_conf[lcore_id]; - if (qconf->n_rx_queue == 0) { + if (qconf->n_rx_port == 0) { RTE_LOG(INFO, LSI, "lcore %u has nothing to do\n", lcore_id); - while(1); + return; } RTE_LOG(INFO, LSI, "entering main loop on lcore %u\n", lcore_id); - for (i = 0; i < qconf->n_rx_queue; i++) { + for (i = 0; i < qconf->n_rx_port; i++) { - portid = qconf->rx_queue_list[i]; + portid = qconf->rx_port_list[i]; RTE_LOG(INFO, LSI, " -- lcoreid=%u portid=%u\n", lcore_id, portid); } @@ -353,17 +227,17 @@ lsi_main_loop(void) * TX burst queue drain */ diff_tsc = cur_tsc - prev_tsc; - if (unlikely(diff_tsc > BURST_TX_DRAIN)) { - - /* this could be optimized (use queueid instead of - * portid), but it is not called so often */ - for (portid = 0; portid < LSI_MAX_PORTS; portid++) { - if (qconf->tx_mbufs[portid].len == 0) - continue; - lsi_send_burst(&lcore_queue_conf[lcore_id], - qconf->tx_mbufs[portid].len, - (uint8_t) portid); - qconf->tx_mbufs[portid].len = 0; + if (unlikely(diff_tsc > drain_tsc)) { + + for (i = 0; i < qconf->n_rx_port; i++) { + + portid = lsi_dst_ports[qconf->rx_port_list[i]]; + buffer = tx_buffer[portid]; + + sent = rte_eth_tx_buffer_flush(portid, 0, buffer); + if (sent) + port_statistics[portid].tx += sent; + } /* if timer is enabled */ @@ -390,9 +264,9 @@ lsi_main_loop(void) /* * Read packet from RX queues */ - for (i = 0; i < qconf->n_rx_queue; i++) { + for (i = 0; i < qconf->n_rx_port; i++) { - portid = qconf->rx_queue_list[i]; + portid = qconf->rx_port_list[i]; nb_rx = rte_eth_rx_burst((uint8_t) portid, 0, pkts_burst, MAX_PKT_BURST); @@ -539,7 +413,7 @@ lsi_parse_args(int argc, char **argv) argv[optind-1] = prgname; ret = optind-1; - optind = 0; /* reset getopt lib */ + optind = 1; /* reset getopt lib */ return ret; } @@ -557,18 +431,20 @@ lsi_parse_args(int argc, char **argv) * Pointer to(address of) the parameters. * * @return - * void. + * int. */ -static void -lsi_event_callback(uint8_t port_id, enum rte_eth_event_type type, void *param) +static int +lsi_event_callback(uint16_t port_id, enum rte_eth_event_type type, void *param, + void *ret_param) { struct rte_eth_link link; RTE_SET_USED(param); + RTE_SET_USED(ret_param); printf("\n\nIn registered callback...\n"); printf("Event type: %s\n", type == RTE_ETH_EVENT_INTR_LSC ? "LSC interrupt" : "unknown event"); - rte_eth_link_get(port_id, &link); + rte_eth_link_get_nowait(port_id, &link); if (link.link_status) { printf("Port %d Link Up - speed %u Mbps - %s\n\n", port_id, (unsigned)link.link_speed, @@ -576,19 +452,73 @@ lsi_event_callback(uint8_t port_id, enum rte_eth_event_type type, void *param) ("full-duplex") : ("half-duplex")); } else printf("Port %d Link Down\n\n", port_id); + + return 0; +} + +/* Check the link status of all ports in up to 9s, and print them finally */ +static void +check_all_ports_link_status(uint16_t port_num, uint32_t port_mask) +{ +#define CHECK_INTERVAL 100 /* 100ms */ +#define MAX_CHECK_TIME 90 /* 9s (90 * 100ms) in total */ + uint8_t count, all_ports_up, print_flag = 0; + uint16_t portid; + struct rte_eth_link link; + + printf("\nChecking link status"); + fflush(stdout); + for (count = 0; count <= MAX_CHECK_TIME; count++) { + all_ports_up = 1; + for (portid = 0; portid < port_num; portid++) { + if ((port_mask & (1 << portid)) == 0) + continue; + memset(&link, 0, sizeof(link)); + rte_eth_link_get_nowait(portid, &link); + /* print link status if flag set */ + if (print_flag == 1) { + if (link.link_status) + printf( + "Port%d Link Up. Speed %u Mbps - %s\n", + portid, link.link_speed, + (link.link_duplex == ETH_LINK_FULL_DUPLEX) ? + ("full-duplex") : ("half-duplex\n")); + else + printf("Port %d Link Down\n", portid); + continue; + } + /* clear all_ports_up flag if any link down */ + if (link.link_status == ETH_LINK_DOWN) { + all_ports_up = 0; + break; + } + } + /* after finally printing all link status, get out */ + if (print_flag == 1) + break; + + if (all_ports_up == 0) { + printf("."); + fflush(stdout); + rte_delay_ms(CHECK_INTERVAL); + } + + /* set the print_flag if all ports up or timeout */ + if (all_ports_up == 1 || count == (MAX_CHECK_TIME - 1)) { + print_flag = 1; + printf("done\n"); + } + } } int -MAIN(int argc, char **argv) +main(int argc, char **argv) { struct lcore_queue_conf *qconf; - struct rte_eth_dev_info dev_info; - struct rte_eth_link link; int ret; - unsigned int nb_ports, nb_lcores; - unsigned portid, portid_last = 0, queueid = 0; + uint16_t nb_ports; + uint16_t portid, portid_last = 0; unsigned lcore_id, rx_lcore_id; - unsigned n_tx_queue, max_tx_queues; unsigned nb_ports_in_mask = 0; /* init EAL */ @@ -605,36 +535,18 @@ MAIN(int argc, char **argv) /* create the mbuf pool */ lsi_pktmbuf_pool = - rte_mempool_create("mbuf_pool", NB_MBUF, - MBUF_SIZE, 32, - sizeof(struct rte_pktmbuf_pool_private), - rte_pktmbuf_pool_init, NULL, - rte_pktmbuf_init, NULL, - SOCKET0, 0); + rte_pktmbuf_pool_create("mbuf_pool", NB_MBUF, 32, 0, + RTE_MBUF_DEFAULT_BUF_SIZE, rte_socket_id()); if (lsi_pktmbuf_pool == NULL) rte_panic("Cannot init mbuf pool\n"); - /* init driver(s) */ - if (rte_pmd_init_all() < 0) - rte_panic("Cannot init pmd\n"); - - if (rte_eal_pci_probe() < 0) - rte_panic("Cannot probe PCI\n"); - - nb_ports = rte_eth_dev_count(); + nb_ports = rte_eth_dev_count_avail(); if (nb_ports == 0) rte_panic("No Ethernet port - bye\n"); - if (nb_ports > LSI_MAX_PORTS) - nb_ports = LSI_MAX_PORTS; - - nb_lcores = rte_lcore_count(); - /* * Each logical core is assigned a dedicated TX queue on each port. - * Compute the maximum number of TX queues that can be used. */ - max_tx_queues = nb_lcores; for (portid = 0; portid < nb_ports; portid++) { /* skip ports that are not enabled */ if ((lsi_enabled_port_mask & (1 << portid)) == 0) @@ -649,12 +561,7 @@ MAIN(int argc, char **argv) portid_last = portid; nb_ports_in_mask++; - - rte_eth_dev_info_get((uint8_t) portid, &dev_info); - if (max_tx_queues > dev_info.max_tx_queues) - max_tx_queues = dev_info.max_tx_queues; } - if (nb_ports_in_mask < 2 || nb_ports_in_mask % 2) rte_exit(EXIT_FAILURE, "Current enabled port number is %u, " "but it should be even and at least 2\n", @@ -662,8 +569,6 @@ MAIN(int argc, char **argv) rx_lcore_id = 0; qconf = &lcore_queue_conf[rx_lcore_id]; - qconf->tx_queue_id = 0; - n_tx_queue = 1; /* Initialize the port/queue configuration of each logical core */ for (portid = 0; portid < nb_ports; portid++) { @@ -673,97 +578,115 @@ MAIN(int argc, char **argv) /* get the lcore_id for this port */ while (rte_lcore_is_enabled(rx_lcore_id) == 0 || - lcore_queue_conf[rx_lcore_id].n_rx_queue == + lcore_queue_conf[rx_lcore_id].n_rx_port == lsi_rx_queue_per_lcore) { rx_lcore_id++; if (rx_lcore_id >= RTE_MAX_LCORE) rte_exit(EXIT_FAILURE, "Not enough cores\n"); - if (n_tx_queue == max_tx_queues) - rte_exit(EXIT_FAILURE, "Not enough TX queues\n"); } - if (qconf != &lcore_queue_conf[rx_lcore_id]) { + if (qconf != &lcore_queue_conf[rx_lcore_id]) /* Assigned a new logical core in the loop above. */ qconf = &lcore_queue_conf[rx_lcore_id]; - qconf->tx_queue_id = n_tx_queue; - n_tx_queue++; - } - qconf->rx_queue_list[qconf->n_rx_queue] = portid; - qconf->n_rx_queue++; - printf("Lcore %u: RX port %u TX queue %u\n", - rx_lcore_id, portid, qconf->tx_queue_id); + + qconf->rx_port_list[qconf->n_rx_port] = portid; + qconf->n_rx_port++; + printf("Lcore %u: RX port %u\n",rx_lcore_id, (unsigned) portid); } /* Initialise each port */ for (portid = 0; portid < nb_ports; portid++) { + struct rte_eth_rxconf rxq_conf; + struct rte_eth_txconf txq_conf; + struct rte_eth_conf local_port_conf = port_conf; + struct rte_eth_dev_info dev_info; /* skip ports that are not enabled */ if ((lsi_enabled_port_mask & (1 << portid)) == 0) { - printf("Skipping disabled port %u\n", portid); + printf("Skipping disabled port %u\n", (unsigned) portid); continue; } /* init port */ - printf("Initializing port %u... ", portid); + printf("Initializing port %u... ", (unsigned) portid); fflush(stdout); - ret = rte_eth_dev_configure((uint8_t) portid, 1, - (uint16_t) n_tx_queue, &port_conf); + rte_eth_dev_info_get(portid, &dev_info); + if (dev_info.tx_offload_capa & DEV_TX_OFFLOAD_MBUF_FAST_FREE) + local_port_conf.txmode.offloads |= + DEV_TX_OFFLOAD_MBUF_FAST_FREE; + ret = rte_eth_dev_configure(portid, 1, 1, &local_port_conf); if (ret < 0) rte_exit(EXIT_FAILURE, "Cannot configure device: err=%d, port=%u\n", - ret, portid); + ret, (unsigned) portid); + + ret = rte_eth_dev_adjust_nb_rx_tx_desc(portid, &nb_rxd, + &nb_txd); + if (ret < 0) + rte_exit(EXIT_FAILURE, + "rte_eth_dev_adjust_nb_rx_tx_desc: err=%d, port=%u\n", + ret, (unsigned) portid); /* register lsi interrupt callback, need to be after * rte_eth_dev_configure(). if (intr_conf.lsc == 0), no * lsc interrupt will be present, and below callback to * be registered will never be called. */ - rte_eth_dev_callback_register((uint8_t)portid, + rte_eth_dev_callback_register(portid, RTE_ETH_EVENT_INTR_LSC, lsi_event_callback, NULL); - rte_eth_macaddr_get((uint8_t) portid, + rte_eth_macaddr_get(portid, &lsi_ports_eth_addr[portid]); /* init one RX queue */ fflush(stdout); - ret = rte_eth_rx_queue_setup((uint8_t) portid, 0, nb_rxd, - SOCKET0, &rx_conf, + rxq_conf = dev_info.default_rxconf; + rxq_conf.offloads = local_port_conf.rxmode.offloads; + ret = rte_eth_rx_queue_setup(portid, 0, nb_rxd, + rte_eth_dev_socket_id(portid), + &rxq_conf, lsi_pktmbuf_pool); if (ret < 0) - rte_exit(EXIT_FAILURE, "rte_eth_tx_queue_setup: err=%d, port=%u\n", - ret, portid); + rte_exit(EXIT_FAILURE, "rte_eth_rx_queue_setup: err=%d, port=%u\n", + ret, (unsigned) portid); /* init one TX queue logical core on each port */ - for (queueid = 0; queueid < n_tx_queue; queueid++) { - fflush(stdout); - ret = rte_eth_tx_queue_setup((uint8_t) portid, - (uint16_t) queueid, nb_txd, - SOCKET0, &tx_conf); - if (ret < 0) - rte_exit(EXIT_FAILURE, "rte_eth_tx_queue_setup: err=%d, " - "port=%u queue=%u\n", - ret, portid, queueid); - } + fflush(stdout); + txq_conf = dev_info.default_txconf; + txq_conf.offloads = local_port_conf.txmode.offloads; + ret = rte_eth_tx_queue_setup(portid, 0, nb_txd, + rte_eth_dev_socket_id(portid), + &txq_conf); + if (ret < 0) + rte_exit(EXIT_FAILURE, "rte_eth_tx_queue_setup: err=%d,port=%u\n", + ret, (unsigned) portid); + + /* Initialize TX buffers */ + tx_buffer[portid] = rte_zmalloc_socket("tx_buffer", + RTE_ETH_TX_BUFFER_SIZE(MAX_PKT_BURST), 0, + rte_eth_dev_socket_id(portid)); + if (tx_buffer[portid] == NULL) + rte_exit(EXIT_FAILURE, "Cannot allocate buffer for tx on port %u\n", + (unsigned) portid); + + rte_eth_tx_buffer_init(tx_buffer[portid], MAX_PKT_BURST); + + ret = rte_eth_tx_buffer_set_err_callback(tx_buffer[portid], + rte_eth_tx_buffer_count_callback, + &port_statistics[portid].dropped); + if (ret < 0) + rte_exit(EXIT_FAILURE, "Cannot set error callback for " + "tx buffer on port %u\n", (unsigned) portid); /* Start device */ - ret = rte_eth_dev_start((uint8_t) portid); + ret = rte_eth_dev_start(portid); if (ret < 0) rte_exit(EXIT_FAILURE, "rte_eth_dev_start: err=%d, port=%u\n", - ret, portid); - - printf("done: "); - - /* get link status */ - rte_eth_link_get((uint8_t) portid, &link); - if (link.link_status) { - printf(" Link Up - speed %u Mbps - %s\n", - (unsigned) link.link_speed, - (link.link_duplex == ETH_LINK_FULL_DUPLEX) ? - ("full-duplex") : ("half-duplex\n")); - } else { - printf(" Link Down\n"); - } + ret, (unsigned) portid); + printf("done:\n"); + + rte_eth_promiscuous_enable(portid); printf("Port %u, MAC address: %02X:%02X:%02X:%02X:%02X:%02X\n\n", - portid, + (unsigned) portid, lsi_ports_eth_addr[portid].addr_bytes[0], lsi_ports_eth_addr[portid].addr_bytes[1], lsi_ports_eth_addr[portid].addr_bytes[2], @@ -775,6 +698,8 @@ MAIN(int argc, char **argv) memset(&port_statistics, 0, sizeof(port_statistics)); } + check_all_ports_link_status(nb_ports, lsi_enabled_port_mask); + /* launch per-lcore init on every lcore */ rte_eal_mp_remote_launch(lsi_launch_one_lcore, NULL, CALL_MASTER); RTE_LCORE_FOREACH_SLAVE(lcore_id) {