ethdev: add namespace
[dpdk.git] / drivers / net / null / rte_eth_null.c
index 2c08ebf..ca03469 100644 (file)
@@ -4,8 +4,8 @@
  */
 
 #include <rte_mbuf.h>
-#include <rte_ethdev_driver.h>
-#include <rte_ethdev_vdev.h>
+#include <ethdev_driver.h>
+#include <ethdev_vdev.h>
 #include <rte_malloc.h>
 #include <rte_memcpy.h>
 #include <rte_bus_vdev.h>
 
 #define ETH_NULL_PACKET_SIZE_ARG       "size"
 #define ETH_NULL_PACKET_COPY_ARG       "copy"
+#define ETH_NULL_PACKET_NO_RX_ARG      "no-rx"
 
 static unsigned int default_packet_size = 64;
 static unsigned int default_packet_copy;
+static unsigned int default_no_rx;
 
 static const char *valid_arguments[] = {
        ETH_NULL_PACKET_SIZE_ARG,
        ETH_NULL_PACKET_COPY_ARG,
+       ETH_NULL_PACKET_NO_RX_ARG,
        NULL
 };
 
@@ -39,11 +42,13 @@ struct null_queue {
 struct pmd_options {
        unsigned int packet_copy;
        unsigned int packet_size;
+       unsigned int no_rx;
 };
 
 struct pmd_internals {
        unsigned int packet_size;
        unsigned int packet_copy;
+       unsigned int no_rx;
        uint16_t port_id;
 
        struct null_queue rx_null_queues[RTE_MAX_QUEUES_PER_PORT];
@@ -56,19 +61,19 @@ struct pmd_internals {
        rte_spinlock_t rss_lock;
 
        uint16_t reta_size;
-       struct rte_eth_rss_reta_entry64 reta_conf[ETH_RSS_RETA_SIZE_128 /
-                       RTE_RETA_GROUP_SIZE];
+       struct rte_eth_rss_reta_entry64 reta_conf[RTE_ETH_RSS_RETA_SIZE_128 /
+                       RTE_ETH_RETA_GROUP_SIZE];
 
        uint8_t rss_key[40];                /**< 40-byte hash key. */
 };
 static struct rte_eth_link pmd_link = {
-       .link_speed = ETH_SPEED_NUM_10G,
-       .link_duplex = ETH_LINK_FULL_DUPLEX,
-       .link_status = ETH_LINK_DOWN,
-       .link_autoneg = ETH_LINK_FIXED,
+       .link_speed = RTE_ETH_SPEED_NUM_10G,
+       .link_duplex = RTE_ETH_LINK_FULL_DUPLEX,
+       .link_status = RTE_ETH_LINK_DOWN,
+       .link_autoneg = RTE_ETH_LINK_FIXED,
 };
 
-static int eth_null_logtype;
+RTE_LOG_REGISTER_DEFAULT(eth_null_logtype, NOTICE);
 
 #define PMD_LOG(level, fmt, args...) \
        rte_log(RTE_LOG_ ## level, eth_null_logtype, \
@@ -126,6 +131,13 @@ eth_null_copy_rx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs)
        return i;
 }
 
+static uint16_t
+eth_null_no_rx(void *q __rte_unused, struct rte_mbuf **bufs __rte_unused,
+               uint16_t nb_bufs __rte_unused)
+{
+       return 0;
+}
+
 static uint16_t
 eth_null_tx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs)
 {
@@ -177,17 +189,19 @@ eth_dev_start(struct rte_eth_dev *dev)
        if (dev == NULL)
                return -EINVAL;
 
-       dev->data->dev_link.link_status = ETH_LINK_UP;
+       dev->data->dev_link.link_status = RTE_ETH_LINK_UP;
        return 0;
 }
 
-static void
+static int
 eth_dev_stop(struct rte_eth_dev *dev)
 {
        if (dev == NULL)
-               return;
+               return 0;
+
+       dev->data->dev_link.link_status = RTE_ETH_LINK_DOWN;
 
-       dev->data->dev_link.link_status = ETH_LINK_DOWN;
+       return 0;
 }
 
 static int
@@ -339,14 +353,24 @@ eth_stats_reset(struct rte_eth_dev *dev)
 }
 
 static void
-eth_queue_release(void *q)
+eth_rx_queue_release(struct rte_eth_dev *dev, uint16_t qid)
 {
-       struct null_queue *nq;
+       struct null_queue *nq = dev->data->rx_queues[qid];
 
-       if (q == NULL)
+       if (nq == NULL)
+               return;
+
+       rte_free(nq->dummy_packet);
+}
+
+static void
+eth_tx_queue_release(struct rte_eth_dev *dev, uint16_t qid)
+{
+       struct null_queue *nq = dev->data->tx_queues[qid];
+
+       if (nq == NULL)
                return;
 
-       nq = q;
        rte_free(nq->dummy_packet);
 }
 
@@ -367,9 +391,9 @@ eth_rss_reta_update(struct rte_eth_dev *dev,
        rte_spinlock_lock(&internal->rss_lock);
 
        /* Copy RETA table */
-       for (i = 0; i < (internal->reta_size / RTE_RETA_GROUP_SIZE); i++) {
+       for (i = 0; i < (internal->reta_size / RTE_ETH_RETA_GROUP_SIZE); i++) {
                internal->reta_conf[i].mask = reta_conf[i].mask;
-               for (j = 0; j < RTE_RETA_GROUP_SIZE; j++)
+               for (j = 0; j < RTE_ETH_RETA_GROUP_SIZE; j++)
                        if ((reta_conf[i].mask >> j) & 0x01)
                                internal->reta_conf[i].reta[j] = reta_conf[i].reta[j];
        }
@@ -392,8 +416,8 @@ eth_rss_reta_query(struct rte_eth_dev *dev,
        rte_spinlock_lock(&internal->rss_lock);
 
        /* Copy RETA table */
-       for (i = 0; i < (internal->reta_size / RTE_RETA_GROUP_SIZE); i++) {
-               for (j = 0; j < RTE_RETA_GROUP_SIZE; j++)
+       for (i = 0; i < (internal->reta_size / RTE_ETH_RETA_GROUP_SIZE); i++) {
+               for (j = 0; j < RTE_ETH_RETA_GROUP_SIZE; j++)
                        if ((reta_conf[i].mask >> j) & 0x01)
                                reta_conf[i].reta[j] = internal->reta_conf[i].reta[j];
        }
@@ -446,15 +470,31 @@ eth_mac_address_set(__rte_unused struct rte_eth_dev *dev,
        return 0;
 }
 
+static int
+eth_dev_close(struct rte_eth_dev *dev)
+{
+       PMD_LOG(INFO, "Closing null ethdev on NUMA socket %u",
+                       rte_socket_id());
+
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return 0;
+
+       /* mac_addrs must not be freed alone because part of dev_private */
+       dev->data->mac_addrs = NULL;
+
+       return 0;
+}
+
 static const struct eth_dev_ops ops = {
+       .dev_close = eth_dev_close,
        .dev_start = eth_dev_start,
        .dev_stop = eth_dev_stop,
        .dev_configure = eth_dev_configure,
        .dev_infos_get = eth_dev_info,
        .rx_queue_setup = eth_rx_queue_setup,
        .tx_queue_setup = eth_tx_queue_setup,
-       .rx_queue_release = eth_queue_release,
-       .tx_queue_release = eth_queue_release,
+       .rx_queue_release = eth_rx_queue_release,
+       .tx_queue_release = eth_tx_queue_release,
        .mtu_set = eth_mtu_set,
        .link_update = eth_link_update,
        .mac_addr_set = eth_mac_address_set,
@@ -504,11 +544,12 @@ eth_dev_null_create(struct rte_vdev_device *dev, struct pmd_options *args)
        internals = eth_dev->data->dev_private;
        internals->packet_size = args->packet_size;
        internals->packet_copy = args->packet_copy;
+       internals->no_rx = args->no_rx;
        internals->port_id = eth_dev->data->port_id;
        rte_eth_random_addr(internals->eth_addr.addr_bytes);
 
-       internals->flow_type_rss_offloads =  ETH_RSS_PROTO_MASK;
-       internals->reta_size = RTE_DIM(internals->reta_conf) * RTE_RETA_GROUP_SIZE;
+       internals->flow_type_rss_offloads =  RTE_ETH_RSS_PROTO_MASK;
+       internals->reta_size = RTE_DIM(internals->reta_conf) * RTE_ETH_RETA_GROUP_SIZE;
 
        rte_memcpy(internals->rss_key, default_rss_key, 40);
 
@@ -519,6 +560,7 @@ eth_dev_null_create(struct rte_vdev_device *dev, struct pmd_options *args)
        data->mac_addrs = &internals->eth_addr;
        data->promiscuous = 1;
        data->all_multicast = 1;
+       data->dev_flags |= RTE_ETH_DEV_AUTOFILL_QUEUE_XSTATS;
 
        eth_dev->dev_ops = &ops;
 
@@ -526,6 +568,9 @@ eth_dev_null_create(struct rte_vdev_device *dev, struct pmd_options *args)
        if (internals->packet_copy) {
                eth_dev->rx_pkt_burst = eth_null_copy_rx;
                eth_dev->tx_pkt_burst = eth_null_copy_tx;
+       } else if (internals->no_rx) {
+               eth_dev->rx_pkt_burst = eth_null_no_rx;
+               eth_dev->tx_pkt_burst = eth_null_tx;
        } else {
                eth_dev->rx_pkt_burst = eth_null_rx;
                eth_dev->tx_pkt_burst = eth_null_tx;
@@ -569,6 +614,24 @@ get_packet_copy_arg(const char *key __rte_unused,
        return 0;
 }
 
+static int
+get_packet_no_rx_arg(const char *key __rte_unused,
+               const char *value, void *extra_args)
+{
+       const char *a = value;
+       unsigned int no_rx;
+
+       if (value == NULL || extra_args == NULL)
+               return -EINVAL;
+
+       no_rx = (unsigned int)strtoul(a, NULL, 0);
+       if (no_rx != 0 && no_rx != 1)
+               return -1;
+
+       *(unsigned int *)extra_args = no_rx;
+       return 0;
+}
+
 static int
 rte_pmd_null_probe(struct rte_vdev_device *dev)
 {
@@ -576,6 +639,7 @@ rte_pmd_null_probe(struct rte_vdev_device *dev)
        struct pmd_options args = {
                .packet_copy = default_packet_copy,
                .packet_size = default_packet_size,
+               .no_rx = default_no_rx,
        };
        struct rte_kvargs *kvlist = NULL;
        struct rte_eth_dev *eth_dev;
@@ -602,6 +666,9 @@ rte_pmd_null_probe(struct rte_vdev_device *dev)
                if (internals->packet_copy) {
                        eth_dev->rx_pkt_burst = eth_null_copy_rx;
                        eth_dev->tx_pkt_burst = eth_null_copy_tx;
+               } else if (internals->no_rx) {
+                       eth_dev->rx_pkt_burst = eth_null_no_rx;
+                       eth_dev->tx_pkt_burst = eth_null_tx;
                } else {
                        eth_dev->rx_pkt_burst = eth_null_rx;
                        eth_dev->tx_pkt_burst = eth_null_tx;
@@ -627,6 +694,20 @@ rte_pmd_null_probe(struct rte_vdev_device *dev)
                                &get_packet_copy_arg, &args.packet_copy);
                if (ret < 0)
                        goto free_kvlist;
+
+               ret = rte_kvargs_process(kvlist,
+                               ETH_NULL_PACKET_NO_RX_ARG,
+                               &get_packet_no_rx_arg, &args.no_rx);
+               if (ret < 0)
+                       goto free_kvlist;
+
+               if (args.no_rx && args.packet_copy) {
+                       PMD_LOG(ERR,
+                               "Both %s and %s arguments at the same time not supported",
+                               ETH_NULL_PACKET_COPY_ARG,
+                               ETH_NULL_PACKET_NO_RX_ARG);
+                       goto free_kvlist;
+               }
        }
 
        PMD_LOG(INFO, "Configure pmd_null: packet size is %d, "
@@ -649,18 +730,12 @@ rte_pmd_null_remove(struct rte_vdev_device *dev)
        if (!dev)
                return -EINVAL;
 
-       PMD_LOG(INFO, "Closing null ethdev on numa socket %u",
-                       rte_socket_id());
-
        /* find the ethdev entry */
        eth_dev = rte_eth_dev_allocated(rte_vdev_device_name(dev));
        if (eth_dev == NULL)
-               return -1;
-
-       if (rte_eal_process_type() == RTE_PROC_PRIMARY)
-               /* mac_addrs must not be freed alone because part of dev_private */
-               eth_dev->data->mac_addrs = NULL;
+               return 0; /* port already released */
 
+       eth_dev_close(eth_dev);
        rte_eth_dev_release_port(eth_dev);
 
        return 0;
@@ -675,11 +750,5 @@ RTE_PMD_REGISTER_VDEV(net_null, pmd_null_drv);
 RTE_PMD_REGISTER_ALIAS(net_null, eth_null);
 RTE_PMD_REGISTER_PARAM_STRING(net_null,
        "size=<int> "
-       "copy=<int>");
-
-RTE_INIT(eth_null_init_log)
-{
-       eth_null_logtype = rte_log_register("pmd.net.null");
-       if (eth_null_logtype >= 0)
-               rte_log_set_level(eth_null_logtype, RTE_LOG_NOTICE);
-}
+       "copy=<int> "
+       ETH_NULL_PACKET_NO_RX_ARG "=0|1");