app/testpmd: fix running without Xen whereas enabled
[dpdk.git] / app / test-pmd / testpmd.c
index 386bf84..26a174c 100644 (file)
@@ -1,7 +1,7 @@
 /*-
  *   BSD LICENSE
  *
- *   Copyright(c) 2010-2015 Intel Corporation. All rights reserved.
+ *   Copyright(c) 2010-2016 Intel Corporation. All rights reserved.
  *   All rights reserved.
  *
  *   Redistribution and use in source and binary forms, with or without
@@ -173,6 +173,9 @@ uint16_t tx_pkt_seg_lengths[RTE_MAX_SEGS_PER_PKT] = {
 };
 uint8_t  tx_pkt_nb_segs = 1; /**< Number of segments in TXONLY packets */
 
+enum tx_pkt_split tx_pkt_split = TX_PKT_SPLIT_OFF;
+/**< Split policy for packets to TX. */
+
 uint16_t nb_pkt_per_burst = DEF_PKT_BURST; /**< Number of packets per burst. */
 uint16_t mb_mempool_cache = DEF_MBUF_CACHE; /**< Size of mbuf mempool cache. */
 
@@ -182,9 +185,6 @@ uint8_t dcb_config = 0;
 /* Whether the dcb is in testing status */
 uint8_t dcb_test = 0;
 
-/* DCB on and VT on mapping is default */
-enum dcb_queue_mapping_mode dcb_q_mapping = DCB_VT_Q_MAPPING;
-
 /*
  * Configurable number of RX/TX queues.
  */
@@ -298,6 +298,9 @@ struct rte_fdir_conf fdir_conf = {
                },
                .src_port_mask = 0xFFFF,
                .dst_port_mask = 0xFFFF,
+               .mac_addr_byte_mask = 0xFF,
+               .tunnel_type_mask = 1,
+               .tunnel_id_mask = 0xFFFFFFFF,
        },
        .drop_queue = 127,
 };
@@ -313,6 +316,8 @@ struct queue_stats_mappings *rx_queue_stats_mappings = rx_queue_stats_mappings_a
 uint16_t nb_tx_queue_stats_mappings = 0;
 uint16_t nb_rx_queue_stats_mappings = 0;
 
+unsigned max_socket = 0;
+
 /* Forward function declarations */
 static void map_port_queue_stats_mapping_registers(uint8_t pi, struct rte_port *port);
 static void check_all_ports_link_status(uint32_t port_mask);
@@ -345,10 +350,17 @@ set_default_fwd_lcores_config(void)
 {
        unsigned int i;
        unsigned int nb_lc;
+       unsigned int sock_num;
 
        nb_lc = 0;
        for (i = 0; i < RTE_MAX_LCORE; i++) {
-               if (! rte_lcore_is_enabled(i))
+               sock_num = rte_lcore_to_socket_id(i) + 1;
+               if (sock_num > max_socket) {
+                       if (sock_num > RTE_MAX_NUMA_NODES)
+                               rte_exit(EXIT_FAILURE, "Total sockets greater than %u\n", RTE_MAX_NUMA_NODES);
+                       max_socket = sock_num;
+               }
+               if (!rte_lcore_is_enabled(i))
                        continue;
                if (i == rte_get_master_lcore())
                        continue;
@@ -398,7 +410,7 @@ mbuf_pool_create(uint16_t mbuf_seg_size, unsigned nb_mbuf,
                 unsigned int socket_id)
 {
        char pool_name[RTE_MEMPOOL_NAMESIZE];
-       struct rte_mempool *rte_mp;
+       struct rte_mempool *rte_mp = NULL;
        uint32_t mb_size;
 
        mb_size = sizeof(struct rte_mbuf) + mbuf_seg_size;
@@ -411,24 +423,23 @@ mbuf_pool_create(uint16_t mbuf_seg_size, unsigned nb_mbuf,
                rte_pktmbuf_pool_init, NULL,
                rte_pktmbuf_init, NULL,
                socket_id, 0);
-
-
-
-#else
-       if (mp_anon != 0)
-               rte_mp = mempool_anon_create(pool_name, nb_mbuf, mb_size,
-                                   (unsigned) mb_mempool_cache,
-                                   sizeof(struct rte_pktmbuf_pool_private),
-                                   rte_pktmbuf_pool_init, NULL,
-                                   rte_pktmbuf_init, NULL,
-                                   socket_id, 0);
-       else
-               /* wrapper to rte_mempool_create() */
-               rte_mp = rte_pktmbuf_pool_create(pool_name, nb_mbuf,
-                       mb_mempool_cache, 0, mbuf_seg_size, socket_id);
-
 #endif
 
+       /* if the former XEN allocation failed fall back to normal allocation */
+       if (rte_mp == NULL) {
+               if (mp_anon != 0)
+                       rte_mp = mempool_anon_create(pool_name, nb_mbuf,
+                                       mb_size, (unsigned) mb_mempool_cache,
+                                       sizeof(struct rte_pktmbuf_pool_private),
+                                       rte_pktmbuf_pool_init, NULL,
+                                       rte_pktmbuf_init, NULL,
+                                       socket_id, 0);
+               else
+                       /* wrapper to rte_mempool_create() */
+                       rte_mp = rte_pktmbuf_pool_create(pool_name, nb_mbuf,
+                               mb_mempool_cache, 0, mbuf_seg_size, socket_id);
+       }
+
        if (rte_mp == NULL) {
                rte_exit(EXIT_FAILURE, "Creation of mbuf pool for socket %u "
                                                "failed\n", socket_id);
@@ -446,7 +457,7 @@ check_socket_id(const unsigned int socket_id)
 {
        static int warning_once = 0;
 
-       if (socket_id >= MAX_SOCKET) {
+       if (socket_id >= max_socket) {
                if (!warning_once && numa_support)
                        printf("Warning: NUMA should be configured manually by"
                               " using --port-numa-config and"
@@ -466,9 +477,9 @@ init_config(void)
        struct rte_mempool *mbp;
        unsigned int nb_mbuf_per_pool;
        lcoreid_t  lc_id;
-       uint8_t port_per_socket[MAX_SOCKET];
+       uint8_t port_per_socket[RTE_MAX_NUMA_NODES];
 
-       memset(port_per_socket,0,MAX_SOCKET);
+       memset(port_per_socket,0,RTE_MAX_NUMA_NODES);
        /* Configuration of logical cores. */
        fwd_lcores = rte_zmalloc("testpmd: fwd_lcores",
                                sizeof(struct fwd_lcore *) * nb_lcores,
@@ -545,7 +556,7 @@ init_config(void)
                if (param_total_num_mbufs)
                        nb_mbuf_per_pool = nb_mbuf_per_pool/nb_ports;
 
-               for (i = 0; i < MAX_SOCKET; i++) {
+               for (i = 0; i < max_socket; i++) {
                        nb_mbuf = (nb_mbuf_per_pool * RTE_MAX_ETHPORTS);
                        if (nb_mbuf)
                                mbuf_pool_create(mbuf_data_size,
@@ -596,6 +607,7 @@ init_fwd_streams(void)
        portid_t pid;
        struct rte_port *port;
        streamid_t sm_id, nb_fwd_streams_new;
+       queueid_t q;
 
        /* set socket id according to numa or not */
        FOREACH_PORT(pid, ports) {
@@ -631,7 +643,12 @@ init_fwd_streams(void)
                }
        }
 
-       nb_fwd_streams_new = (streamid_t)(nb_ports * nb_rxq);
+       q = RTE_MAX(nb_rxq, nb_txq);
+       if (q == 0) {
+               printf("Fail: Cannot allocate fwd streams as number of queues is 0\n");
+               return -1;
+       }
+       nb_fwd_streams_new = (streamid_t)(nb_ports * q);
        if (nb_fwd_streams_new == nb_fwd_streams)
                return 0;
        /* clear the old */
@@ -741,10 +758,8 @@ fwd_port_stats_display(portid_t port_id, struct rte_eth_stats *stats)
                if (cur_fwd_eng == &csum_fwd_engine)
                        printf("  Bad-ipcsum: %-14"PRIu64" Bad-l4csum: %-14"PRIu64" \n",
                               port->rx_bad_ip_csum, port->rx_bad_l4_csum);
-               if (((stats->ierrors - stats->imissed) + stats->rx_nombuf) > 0) {
-                       printf("  RX-badcrc:  %-14"PRIu64" RX-badlen:  %-14"PRIu64
-                              "RX-error: %-"PRIu64"\n",
-                              stats->ibadcrc, stats->ibadlen, stats->ierrors);
+               if ((stats->ierrors + stats->rx_nombuf) > 0) {
+                       printf("  RX-error: %-"PRIu64"\n",  stats->ierrors);
                        printf("  RX-nombufs: %-14"PRIu64"\n", stats->rx_nombuf);
                }
 
@@ -762,10 +777,8 @@ fwd_port_stats_display(portid_t port_id, struct rte_eth_stats *stats)
                if (cur_fwd_eng == &csum_fwd_engine)
                        printf("  Bad-ipcsum:%14"PRIu64"    Bad-l4csum:%14"PRIu64"\n",
                               port->rx_bad_ip_csum, port->rx_bad_l4_csum);
-               if (((stats->ierrors - stats->imissed) + stats->rx_nombuf) > 0) {
-                       printf("  RX-badcrc:              %14"PRIu64"    RX-badlen: %14"PRIu64
-                              "    RX-error:%"PRIu64"\n",
-                              stats->ibadcrc, stats->ibadlen, stats->ierrors);
+               if ((stats->ierrors + stats->rx_nombuf) > 0) {
+                       printf("  RX-error:%"PRIu64"\n", stats->ierrors);
                        printf("  RX-nombufs:             %14"PRIu64"\n",
                               stats->rx_nombuf);
                }
@@ -776,15 +789,6 @@ fwd_port_stats_display(portid_t port_id, struct rte_eth_stats *stats)
                       (uint64_t) (stats->opackets + port->tx_dropped));
        }
 
-       /* Display statistics of XON/XOFF pause frames, if any. */
-       if ((stats->tx_pause_xon  | stats->rx_pause_xon |
-            stats->tx_pause_xoff | stats->rx_pause_xoff) > 0) {
-               printf("  RX-XOFF:    %-14"PRIu64" RX-XON:     %-14"PRIu64"\n",
-                      stats->rx_pause_xoff, stats->rx_pause_xon);
-               printf("  TX-XOFF:    %-14"PRIu64" TX-XON:     %-14"PRIu64"\n",
-                      stats->tx_pause_xoff, stats->tx_pause_xon);
-       }
-
 #ifdef RTE_TEST_PMD_RECORD_BURST_STATS
        if (port->rx_stream)
                pkt_burst_stats_display("RX",
@@ -793,11 +797,6 @@ fwd_port_stats_display(portid_t port_id, struct rte_eth_stats *stats)
                pkt_burst_stats_display("TX",
                        &port->tx_stream->tx_burst_stats);
 #endif
-       /* stats fdir */
-       if (fdir_conf.mode != RTE_FDIR_MODE_NONE)
-               printf("  Fdirmiss:%14"PRIu64"    Fdirmatch:%14"PRIu64"\n",
-                      stats->fdirmiss,
-                      stats->fdirmatch);
 
        if (port->rx_queue_stats_mapping_enabled) {
                printf("\n");
@@ -961,6 +960,19 @@ start_packet_forwarding(int with_tx_first)
        portid_t   pt_id;
        streamid_t sm_id;
 
+       if (strcmp(cur_fwd_eng->fwd_mode_name, "rxonly") == 0 && !nb_rxq)
+               rte_exit(EXIT_FAILURE, "rxq are 0, cannot use rxonly fwd mode\n");
+
+       if (strcmp(cur_fwd_eng->fwd_mode_name, "txonly") == 0 && !nb_txq)
+               rte_exit(EXIT_FAILURE, "txq are 0, cannot use txonly fwd mode\n");
+
+       if ((strcmp(cur_fwd_eng->fwd_mode_name, "rxonly") != 0 &&
+               strcmp(cur_fwd_eng->fwd_mode_name, "txonly") != 0) &&
+               (!nb_rxq || !nb_txq))
+               rte_exit(EXIT_FAILURE,
+                       "Either rxq or txq are 0, cannot use %s fwd mode\n",
+                       cur_fwd_eng->fwd_mode_name);
+
        if (all_ports_started() == 0) {
                printf("Not all ports were started\n");
                return;
@@ -1141,10 +1153,6 @@ stop_packet_forwarding(void)
                port->stats.oerrors = 0;
                stats.rx_nombuf -= port->stats.rx_nombuf;
                port->stats.rx_nombuf = 0;
-               stats.fdirmatch -= port->stats.fdirmatch;
-               port->stats.rx_nombuf = 0;
-               stats.fdirmiss -= port->stats.fdirmiss;
-               port->stats.rx_nombuf = 0;
 
                total_recv += stats.ipackets;
                total_xmit += stats.opackets;
@@ -1522,7 +1530,7 @@ attach_port(char *identifier)
        nb_ports = rte_eth_dev_count();
 
        /* set_default_fwd_ports_config(); */
-       bzero(fwd_ports_ids, sizeof(fwd_ports_ids));
+       memset(fwd_ports_ids, 0, sizeof(fwd_ports_ids));
        i = 0;
        FOREACH_PORT(j, ports) {
                fwd_ports_ids[i] = j;
@@ -1557,7 +1565,7 @@ detach_port(uint8_t port_id)
        nb_ports = rte_eth_dev_count();
 
        /* set_default_fwd_ports_config(); */
-       bzero(fwd_ports_ids, sizeof(fwd_ports_ids));
+       memset(fwd_ports_ids, 0, sizeof(fwd_ports_ids));
        i = 0;
        FOREACH_PORT(pi, ports) {
                fwd_ports_ids[i] = pi;
@@ -1580,13 +1588,16 @@ pmd_test_exit(void)
        if (test_done == 0)
                stop_packet_forwarding();
 
-       FOREACH_PORT(pt_id, ports) {
-               printf("Stopping port %d...", pt_id);
-               fflush(stdout);
-               rte_eth_dev_close(pt_id);
-               printf("done\n");
+       if (ports != NULL) {
+               no_link_check = 1;
+               FOREACH_PORT(pt_id, ports) {
+                       printf("\nShutting down port %d...\n", pt_id);
+                       fflush(stdout);
+                       stop_port(pt_id);
+                       close_port(pt_id);
+               }
        }
-       printf("bye...\n");
+       printf("\nBye...\n");
 }
 
 typedef void (*cmd_func_t)(void);
@@ -1629,7 +1640,7 @@ check_all_ports_link_status(uint32_t port_mask)
                                continue;
                        }
                        /* clear all_ports_up flag if any link down */
-                       if (link.link_status == 0) {
+                       if (link.link_status == ETH_LINK_DOWN) {
                                all_ports_up = 0;
                                break;
                        }
@@ -1840,115 +1851,131 @@ const uint16_t vlan_tags[] = {
 };
 
 static  int
-get_eth_dcb_conf(struct rte_eth_conf *eth_conf, struct dcb_config *dcb_conf)
+get_eth_dcb_conf(struct rte_eth_conf *eth_conf,
+                enum dcb_mode_enable dcb_mode,
+                enum rte_eth_nb_tcs num_tcs,
+                uint8_t pfc_en)
 {
-        uint8_t i;
+       uint8_t i;
 
        /*
         * Builds up the correct configuration for dcb+vt based on the vlan tags array
         * given above, and the number of traffic classes available for use.
         */
-       if (dcb_conf->dcb_mode == DCB_VT_ENABLED) {
-               struct rte_eth_vmdq_dcb_conf vmdq_rx_conf;
-               struct rte_eth_vmdq_dcb_tx_conf vmdq_tx_conf;
+       if (dcb_mode == DCB_VT_ENABLED) {
+               struct rte_eth_vmdq_dcb_conf *vmdq_rx_conf =
+                               &eth_conf->rx_adv_conf.vmdq_dcb_conf;
+               struct rte_eth_vmdq_dcb_tx_conf *vmdq_tx_conf =
+                               &eth_conf->tx_adv_conf.vmdq_dcb_tx_conf;
 
                /* VMDQ+DCB RX and TX configrations */
-               vmdq_rx_conf.enable_default_pool = 0;
-               vmdq_rx_conf.default_pool = 0;
-               vmdq_rx_conf.nb_queue_pools =
-                       (dcb_conf->num_tcs ==  ETH_4_TCS ? ETH_32_POOLS : ETH_16_POOLS);
-               vmdq_tx_conf.nb_queue_pools =
-                       (dcb_conf->num_tcs ==  ETH_4_TCS ? ETH_32_POOLS : ETH_16_POOLS);
-
-               vmdq_rx_conf.nb_pool_maps = sizeof( vlan_tags )/sizeof( vlan_tags[ 0 ]);
-               for (i = 0; i < vmdq_rx_conf.nb_pool_maps; i++) {
-                       vmdq_rx_conf.pool_map[i].vlan_id = vlan_tags[ i ];
-                       vmdq_rx_conf.pool_map[i].pools = 1 << (i % vmdq_rx_conf.nb_queue_pools);
+               vmdq_rx_conf->enable_default_pool = 0;
+               vmdq_rx_conf->default_pool = 0;
+               vmdq_rx_conf->nb_queue_pools =
+                       (num_tcs ==  ETH_4_TCS ? ETH_32_POOLS : ETH_16_POOLS);
+               vmdq_tx_conf->nb_queue_pools =
+                       (num_tcs ==  ETH_4_TCS ? ETH_32_POOLS : ETH_16_POOLS);
+
+               vmdq_rx_conf->nb_pool_maps = vmdq_rx_conf->nb_queue_pools;
+               for (i = 0; i < vmdq_rx_conf->nb_pool_maps; i++) {
+                       vmdq_rx_conf->pool_map[i].vlan_id = vlan_tags[i];
+                       vmdq_rx_conf->pool_map[i].pools =
+                               1 << (i % vmdq_rx_conf->nb_queue_pools);
                }
                for (i = 0; i < ETH_DCB_NUM_USER_PRIORITIES; i++) {
-                       vmdq_rx_conf.dcb_queue[i] = i;
-                       vmdq_tx_conf.dcb_queue[i] = i;
+                       vmdq_rx_conf->dcb_tc[i] = i;
+                       vmdq_tx_conf->dcb_tc[i] = i;
                }
 
-               /*set DCB mode of RX and TX of multiple queues*/
+               /* set DCB mode of RX and TX of multiple queues */
                eth_conf->rxmode.mq_mode = ETH_MQ_RX_VMDQ_DCB;
                eth_conf->txmode.mq_mode = ETH_MQ_TX_VMDQ_DCB;
-               if (dcb_conf->pfc_en)
-                       eth_conf->dcb_capability_en = ETH_DCB_PG_SUPPORT|ETH_DCB_PFC_SUPPORT;
-               else
-                       eth_conf->dcb_capability_en = ETH_DCB_PG_SUPPORT;
-
-               (void)(rte_memcpy(&eth_conf->rx_adv_conf.vmdq_dcb_conf, &vmdq_rx_conf,
-                                sizeof(struct rte_eth_vmdq_dcb_conf)));
-               (void)(rte_memcpy(&eth_conf->tx_adv_conf.vmdq_dcb_tx_conf, &vmdq_tx_conf,
-                                sizeof(struct rte_eth_vmdq_dcb_tx_conf)));
-       }
-       else {
-               struct rte_eth_dcb_rx_conf rx_conf;
-               struct rte_eth_dcb_tx_conf tx_conf;
-
-               /* queue mapping configuration of DCB RX and TX */
-               if (dcb_conf->num_tcs == ETH_4_TCS)
-                       dcb_q_mapping = DCB_4_TCS_Q_MAPPING;
-               else
-                       dcb_q_mapping = DCB_8_TCS_Q_MAPPING;
-
-               rx_conf.nb_tcs = dcb_conf->num_tcs;
-               tx_conf.nb_tcs = dcb_conf->num_tcs;
-
-               for (i = 0; i < ETH_DCB_NUM_USER_PRIORITIES; i++){
-                       rx_conf.dcb_queue[i] = i;
-                       tx_conf.dcb_queue[i] = i;
+       } else {
+               struct rte_eth_dcb_rx_conf *rx_conf =
+                               &eth_conf->rx_adv_conf.dcb_rx_conf;
+               struct rte_eth_dcb_tx_conf *tx_conf =
+                               &eth_conf->tx_adv_conf.dcb_tx_conf;
+
+               rx_conf->nb_tcs = num_tcs;
+               tx_conf->nb_tcs = num_tcs;
+
+               for (i = 0; i < num_tcs; i++) {
+                       rx_conf->dcb_tc[i] = i;
+                       tx_conf->dcb_tc[i] = i;
                }
-               eth_conf->rxmode.mq_mode = ETH_MQ_RX_DCB;
+               eth_conf->rxmode.mq_mode = ETH_MQ_RX_DCB_RSS;
+               eth_conf->rx_adv_conf.rss_conf.rss_hf = rss_hf;
                eth_conf->txmode.mq_mode = ETH_MQ_TX_DCB;
-               if (dcb_conf->pfc_en)
-                       eth_conf->dcb_capability_en = ETH_DCB_PG_SUPPORT|ETH_DCB_PFC_SUPPORT;
-               else
-                       eth_conf->dcb_capability_en = ETH_DCB_PG_SUPPORT;
-
-               (void)(rte_memcpy(&eth_conf->rx_adv_conf.dcb_rx_conf, &rx_conf,
-                                sizeof(struct rte_eth_dcb_rx_conf)));
-               (void)(rte_memcpy(&eth_conf->tx_adv_conf.dcb_tx_conf, &tx_conf,
-                                sizeof(struct rte_eth_dcb_tx_conf)));
        }
 
+       if (pfc_en)
+               eth_conf->dcb_capability_en =
+                               ETH_DCB_PG_SUPPORT | ETH_DCB_PFC_SUPPORT;
+       else
+               eth_conf->dcb_capability_en = ETH_DCB_PG_SUPPORT;
+
        return 0;
 }
 
 int
-init_port_dcb_config(portid_t pid,struct dcb_config *dcb_conf)
+init_port_dcb_config(portid_t pid,
+                    enum dcb_mode_enable dcb_mode,
+                    enum rte_eth_nb_tcs num_tcs,
+                    uint8_t pfc_en)
 {
        struct rte_eth_conf port_conf;
+       struct rte_eth_dev_info dev_info;
        struct rte_port *rte_port;
        int retval;
-       uint16_t nb_vlan;
        uint16_t i;
 
-       /* rxq and txq configuration in dcb mode */
-       nb_rxq = 128;
-       nb_txq = 128;
+       rte_eth_dev_info_get(pid, &dev_info);
+
+       /* If dev_info.vmdq_pool_base is greater than 0,
+        * the queue id of vmdq pools is started after pf queues.
+        */
+       if (dcb_mode == DCB_VT_ENABLED && dev_info.vmdq_pool_base > 0) {
+               printf("VMDQ_DCB multi-queue mode is nonsensical"
+                       " for port %d.", pid);
+               return -1;
+       }
+
+       /* Assume the ports in testpmd have the same dcb capability
+        * and has the same number of rxq and txq in dcb mode
+        */
+       if (dcb_mode == DCB_VT_ENABLED) {
+               nb_rxq = dev_info.max_rx_queues;
+               nb_txq = dev_info.max_tx_queues;
+       } else {
+               /*if vt is disabled, use all pf queues */
+               if (dev_info.vmdq_pool_base == 0) {
+                       nb_rxq = dev_info.max_rx_queues;
+                       nb_txq = dev_info.max_tx_queues;
+               } else {
+                       nb_rxq = (queueid_t)num_tcs;
+                       nb_txq = (queueid_t)num_tcs;
+
+               }
+       }
        rx_free_thresh = 64;
 
-       memset(&port_conf,0,sizeof(struct rte_eth_conf));
+       memset(&port_conf, 0, sizeof(struct rte_eth_conf));
        /* Enter DCB configuration status */
        dcb_config = 1;
 
-       nb_vlan = sizeof( vlan_tags )/sizeof( vlan_tags[ 0 ]);
        /*set configuration of DCB in vt mode and DCB in non-vt mode*/
-       retval = get_eth_dcb_conf(&port_conf, dcb_conf);
+       retval = get_eth_dcb_conf(&port_conf, dcb_mode, num_tcs, pfc_en);
        if (retval < 0)
                return retval;
 
        rte_port = &ports[pid];
-       memcpy(&rte_port->dev_conf, &port_conf,sizeof(struct rte_eth_conf));
+       memcpy(&rte_port->dev_conf, &port_conf, sizeof(struct rte_eth_conf));
 
        rxtx_port_config(rte_port);
        /* VLAN filter */
        rte_port->dev_conf.rxmode.hw_vlan_filter = 1;
-       for (i = 0; i < nb_vlan; i++){
+       for (i = 0; i < RTE_DIM(vlan_tags); i++)
                rx_vft_set(pid, vlan_tags[i], 1);
-       }
 
        rte_eth_macaddr_get(pid, &rte_port->eth_addr);
        map_port_queue_stats_mapping_registers(pid, rte_port);
@@ -1978,12 +2005,35 @@ init_port(void)
                ports[pid].enabled = 1;
 }
 
+static void
+force_quit(void)
+{
+       pmd_test_exit();
+       prompt_exit();
+}
+
+static void
+signal_handler(int signum)
+{
+       if (signum == SIGINT || signum == SIGTERM) {
+               printf("\nSignal %d received, preparing to exit...\n",
+                               signum);
+               force_quit();
+               /* exit with the expected status */
+               signal(signum, SIG_DFL);
+               kill(getpid(), signum);
+       }
+}
+
 int
 main(int argc, char** argv)
 {
        int  diag;
        uint8_t port_id;
 
+       signal(SIGINT, signal_handler);
+       signal(SIGTERM, signal_handler);
+
        diag = rte_eal_init(argc, argv);
        if (diag < 0)
                rte_panic("Cannot init EAL\n");
@@ -2005,7 +2055,10 @@ main(int argc, char** argv)
        if (argc > 1)
                launch_args_parse(argc, argv);
 
-       if (nb_rxq > nb_txq)
+       if (!nb_rxq && !nb_txq)
+               printf("Warning: Either rx or tx queues should be non-zero\n");
+
+       if (nb_rxq > 1 && nb_rxq > nb_txq)
                printf("Warning: nb_rxq=%d enables RSS configuration, "
                       "but nb_txq=%d will prevent to fully test it.\n",
                       nb_rxq, nb_txq);
@@ -2035,6 +2088,7 @@ main(int argc, char** argv)
                start_packet_forwarding(0);
                printf("Press enter to exit\n");
                rc = read(0, &c, 1);
+               pmd_test_exit();
                if (rc < 0)
                        return 1;
        }