X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=drivers%2Fnet%2Favp%2Favp_ethdev.c;h=7ac55584ff378e4fd5aba4a7a77076a8cb507f01;hb=cc91ec13eb831351e291d682d2e5c841e0dd6888;hp=1abe96ce5079ebd62d1548ea03915b5037f7a491;hpb=ec87d3b2c2451ce0fa71f53f64987c68a0b3423c;p=dpdk.git diff --git a/drivers/net/avp/avp_ethdev.c b/drivers/net/avp/avp_ethdev.c index 1abe96ce50..7ac55584ff 100644 --- a/drivers/net/avp/avp_ethdev.c +++ b/drivers/net/avp/avp_ethdev.c @@ -8,8 +8,8 @@ #include #include -#include -#include +#include +#include #include #include #include @@ -32,15 +32,13 @@ #include "avp_logs.h" -int avp_logtype_driver; - static int avp_dev_create(struct rte_pci_device *pci_dev, struct rte_eth_dev *eth_dev); static int avp_dev_configure(struct rte_eth_dev *dev); static int avp_dev_start(struct rte_eth_dev *dev); -static void avp_dev_stop(struct rte_eth_dev *dev); -static void avp_dev_close(struct rte_eth_dev *dev); +static int avp_dev_stop(struct rte_eth_dev *dev); +static int avp_dev_close(struct rte_eth_dev *dev); static int avp_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info); static int avp_vlan_offload_set(struct rte_eth_dev *dev, int mask); @@ -77,8 +75,8 @@ static uint16_t avp_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts); -static void avp_dev_rx_queue_release(void *rxq); -static void avp_dev_tx_queue_release(void *txq); +static void avp_dev_rx_queue_release(struct rte_eth_dev *dev, uint16_t qid); +static void avp_dev_tx_queue_release(struct rte_eth_dev *dev, uint16_t qid); static int avp_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats); @@ -269,7 +267,7 @@ avp_dev_process_request(struct avp_dev *avp, struct rte_avp_request *request) break; } - if ((count < 1) && (retry == 0)) { + if (retry == 0) { PMD_DRV_LOG(ERR, "Timeout while waiting for a response for %u\n", request->req_id); ret = -ETIME; @@ -713,7 +711,7 @@ avp_dev_interrupt_handler(void *data) status); /* re-enable UIO interrupt handling */ - ret = rte_intr_ack(&pci_dev->intr_handle); + ret = rte_intr_ack(pci_dev->intr_handle); if (ret < 0) { PMD_DRV_LOG(ERR, "Failed to re-enable UIO interrupts, ret=%d\n", ret); @@ -732,7 +730,7 @@ avp_dev_enable_interrupts(struct rte_eth_dev *eth_dev) return -EINVAL; /* enable UIO interrupt handling */ - ret = rte_intr_enable(&pci_dev->intr_handle); + ret = rte_intr_enable(pci_dev->intr_handle); if (ret < 0) { PMD_DRV_LOG(ERR, "Failed to enable UIO interrupts, ret=%d\n", ret); @@ -761,7 +759,7 @@ avp_dev_disable_interrupts(struct rte_eth_dev *eth_dev) RTE_PTR_ADD(registers, RTE_AVP_INTERRUPT_MASK_OFFSET)); /* enable UIO interrupt handling */ - ret = rte_intr_disable(&pci_dev->intr_handle); + ret = rte_intr_disable(pci_dev->intr_handle); if (ret < 0) { PMD_DRV_LOG(ERR, "Failed to disable UIO interrupts, ret=%d\n", ret); @@ -778,7 +776,7 @@ avp_dev_setup_interrupts(struct rte_eth_dev *eth_dev) int ret; /* register a callback handler with UIO for interrupt notifications */ - ret = rte_intr_callback_register(&pci_dev->intr_handle, + ret = rte_intr_callback_register(pci_dev->intr_handle, avp_dev_interrupt_handler, (void *)eth_dev); if (ret < 0) { @@ -959,8 +957,6 @@ eth_avp_dev_init(struct rte_eth_dev *eth_dev) eth_dev->dev_ops = &avp_eth_dev_ops; eth_dev->rx_pkt_burst = &avp_recv_pkts; eth_dev->tx_pkt_burst = &avp_xmit_pkts; - /* Let rte_eth_dev_close() release the port resources */ - eth_dev->data->dev_flags |= RTE_ETH_DEV_CLOSE_REMOVE; if (rte_eal_process_type() != RTE_PROC_PRIMARY) { /* @@ -978,6 +974,7 @@ eth_avp_dev_init(struct rte_eth_dev *eth_dev) } rte_eth_copy_pci_info(eth_dev, pci_dev); + eth_dev->data->dev_flags |= RTE_ETH_DEV_AUTOFILL_QUEUE_XSTATS; /* Check current migration status */ if (avp_dev_migration_pending(eth_dev)) { @@ -1062,17 +1059,18 @@ static int avp_dev_enable_scattered(struct rte_eth_dev *eth_dev, struct avp_dev *avp) { - unsigned int max_rx_pkt_len; + unsigned int max_rx_pktlen; - max_rx_pkt_len = eth_dev->data->dev_conf.rxmode.max_rx_pkt_len; + max_rx_pktlen = eth_dev->data->mtu + RTE_ETHER_HDR_LEN + + RTE_ETHER_CRC_LEN; - if ((max_rx_pkt_len > avp->guest_mbuf_size) || - (max_rx_pkt_len > avp->host_mbuf_size)) { + if (max_rx_pktlen > avp->guest_mbuf_size || + max_rx_pktlen > avp->host_mbuf_size) { /* * If the guest MTU is greater than either the host or guest * buffers then chained mbufs have to be enabled in the TX * direction. It is assumed that the application will not need - * to send packets larger than their max_rx_pkt_len (MRU). + * to send packets larger than their MTU. */ return 1; } @@ -1127,7 +1125,7 @@ avp_dev_rx_queue_setup(struct rte_eth_dev *eth_dev, PMD_DRV_LOG(DEBUG, "AVP max_rx_pkt_len=(%u,%u) mbuf_size=(%u,%u)\n", avp->max_rx_pkt_len, - eth_dev->data->dev_conf.rxmode.max_rx_pkt_len, + eth_dev->data->mtu + RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN, avp->host_mbuf_size, avp->guest_mbuf_size); @@ -1208,17 +1206,17 @@ _avp_mac_filter(struct avp_dev *avp, struct rte_mbuf *m) { struct rte_ether_hdr *eth = rte_pktmbuf_mtod(m, struct rte_ether_hdr *); - if (likely(_avp_cmp_ether_addr(&avp->ethaddr, ð->d_addr) == 0)) { + if (likely(_avp_cmp_ether_addr(&avp->ethaddr, ð->dst_addr) == 0)) { /* allow all packets destined to our address */ return 0; } - if (likely(rte_is_broadcast_ether_addr(ð->d_addr))) { + if (likely(rte_is_broadcast_ether_addr(ð->dst_addr))) { /* allow all broadcast packets */ return 0; } - if (likely(rte_is_multicast_ether_addr(ð->d_addr))) { + if (likely(rte_is_multicast_ether_addr(ð->dst_addr))) { /* allow all multicast packets */ return 0; } @@ -1313,7 +1311,7 @@ avp_dev_copy_from_buffers(struct avp_dev *avp, src_offset = 0; if (pkt_buf->ol_flags & RTE_AVP_RX_VLAN_PKT) { - ol_flags = PKT_RX_VLAN; + ol_flags = RTE_MBUF_F_RX_VLAN; vlan_tci = pkt_buf->vlan_tci; } else { ol_flags = 0; @@ -1571,7 +1569,7 @@ avp_recv_pkts(void *rx_queue, m->port = avp->port_id; if (pkt_buf->ol_flags & RTE_AVP_RX_VLAN_PKT) { - m->ol_flags = PKT_RX_VLAN; + m->ol_flags = RTE_MBUF_F_RX_VLAN; m->vlan_tci = pkt_buf->vlan_tci; } @@ -1677,7 +1675,7 @@ avp_dev_copy_to_buffers(struct avp_dev *avp, first_buf->nb_segs = count; first_buf->pkt_len = total_length; - if (mbuf->ol_flags & PKT_TX_VLAN_PKT) { + if (mbuf->ol_flags & RTE_MBUF_F_TX_VLAN) { first_buf->ol_flags |= RTE_AVP_TX_VLAN_PKT; first_buf->vlan_tci = mbuf->vlan_tci; } @@ -1892,8 +1890,8 @@ avp_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) * function; send it truncated to avoid the performance * hit of having to manage returning the already * allocated buffer to the free list. This should not - * happen since the application should have set the - * max_rx_pkt_len based on its MTU and it should be + * happen since the application should have not send + * packages larger than its MTU and it should be * policing its own packet sizes. */ txq->errors++; @@ -1908,7 +1906,7 @@ avp_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) pkt_buf->nb_segs = 1; pkt_buf->next = NULL; - if (m->ol_flags & PKT_TX_VLAN_PKT) { + if (m->ol_flags & RTE_MBUF_F_TX_VLAN) { pkt_buf->ol_flags |= RTE_AVP_TX_VLAN_PKT; pkt_buf->vlan_tci = m->vlan_tci; } @@ -1929,18 +1927,11 @@ avp_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) } static void -avp_dev_rx_queue_release(void *rx_queue) +avp_dev_rx_queue_release(struct rte_eth_dev *eth_dev, uint16_t rx_queue_id) { - struct avp_queue *rxq = (struct avp_queue *)rx_queue; - struct avp_dev *avp = rxq->avp; - struct rte_eth_dev_data *data = avp->dev_data; - unsigned int i; - - for (i = 0; i < avp->num_rx_queues; i++) { - if (data->rx_queues[i] == rxq) { - rte_free(data->rx_queues[i]); - data->rx_queues[i] = NULL; - } + if (eth_dev->data->rx_queues[rx_queue_id] != NULL) { + rte_free(eth_dev->data->rx_queues[rx_queue_id]); + eth_dev->data->rx_queues[rx_queue_id] = NULL; } } @@ -1960,18 +1951,11 @@ avp_dev_rx_queue_release_all(struct rte_eth_dev *eth_dev) } static void -avp_dev_tx_queue_release(void *tx_queue) +avp_dev_tx_queue_release(struct rte_eth_dev *eth_dev, uint16_t tx_queue_id) { - struct avp_queue *txq = (struct avp_queue *)tx_queue; - struct avp_dev *avp = txq->avp; - struct rte_eth_dev_data *data = avp->dev_data; - unsigned int i; - - for (i = 0; i < avp->num_tx_queues; i++) { - if (data->tx_queues[i] == txq) { - rte_free(data->tx_queues[i]); - data->tx_queues[i] = NULL; - } + if (eth_dev->data->tx_queues[tx_queue_id] != NULL) { + rte_free(eth_dev->data->tx_queues[tx_queue_id]); + eth_dev->data->tx_queues[tx_queue_id] = NULL; } } @@ -2014,9 +1998,9 @@ avp_dev_configure(struct rte_eth_dev *eth_dev) /* Setup required number of queues */ _avp_set_queue_counts(eth_dev); - mask = (ETH_VLAN_STRIP_MASK | - ETH_VLAN_FILTER_MASK | - ETH_VLAN_EXTEND_MASK); + mask = (RTE_ETH_VLAN_STRIP_MASK | + RTE_ETH_VLAN_FILTER_MASK | + RTE_ETH_VLAN_EXTEND_MASK); ret = avp_vlan_offload_set(eth_dev, mask); if (ret < 0) { PMD_DRV_LOG(ERR, "VLAN offload set failed by host, ret=%d\n", @@ -2079,7 +2063,7 @@ unlock: return ret; } -static void +static int avp_dev_stop(struct rte_eth_dev *eth_dev) { struct avp_dev *avp = AVP_DEV_PRIVATE_TO_HW(eth_dev->data->dev_private); @@ -2088,6 +2072,7 @@ avp_dev_stop(struct rte_eth_dev *eth_dev) rte_spinlock_lock(&avp->lock); if (avp->flags & AVP_F_DETACHED) { PMD_DRV_LOG(ERR, "Operation not supported during VM live migration\n"); + ret = -ENOTSUP; goto unlock; } @@ -2103,14 +2088,18 @@ avp_dev_stop(struct rte_eth_dev *eth_dev) unlock: rte_spinlock_unlock(&avp->lock); + return ret; } -static void +static int avp_dev_close(struct rte_eth_dev *eth_dev) { struct avp_dev *avp = AVP_DEV_PRIVATE_TO_HW(eth_dev->data->dev_private); int ret; + if (rte_eal_process_type() != RTE_PROC_PRIMARY) + return 0; + rte_spinlock_lock(&avp->lock); if (avp->flags & AVP_F_DETACHED) { PMD_DRV_LOG(ERR, "Operation not supported during VM live migration\n"); @@ -2141,6 +2130,7 @@ avp_dev_close(struct rte_eth_dev *eth_dev) unlock: rte_spinlock_unlock(&avp->lock); + return 0; } static int @@ -2150,8 +2140,8 @@ avp_dev_link_update(struct rte_eth_dev *eth_dev, struct avp_dev *avp = AVP_DEV_PRIVATE_TO_HW(eth_dev->data->dev_private); struct rte_eth_link *link = ð_dev->data->dev_link; - link->link_speed = ETH_SPEED_NUM_10G; - link->link_duplex = ETH_LINK_FULL_DUPLEX; + link->link_speed = RTE_ETH_SPEED_NUM_10G; + link->link_duplex = RTE_ETH_LINK_FULL_DUPLEX; link->link_status = !!(avp->flags & AVP_F_LINKUP); return -1; @@ -2201,8 +2191,8 @@ avp_dev_info_get(struct rte_eth_dev *eth_dev, dev_info->max_rx_pktlen = avp->max_rx_pkt_len; dev_info->max_mac_addrs = AVP_MAX_MAC_ADDRS; if (avp->host_features & RTE_AVP_FEATURE_VLAN_OFFLOAD) { - dev_info->rx_offload_capa = DEV_RX_OFFLOAD_VLAN_STRIP; - dev_info->tx_offload_capa = DEV_TX_OFFLOAD_VLAN_INSERT; + dev_info->rx_offload_capa = RTE_ETH_RX_OFFLOAD_VLAN_STRIP; + dev_info->tx_offload_capa = RTE_ETH_TX_OFFLOAD_VLAN_INSERT; } return 0; @@ -2215,9 +2205,9 @@ avp_vlan_offload_set(struct rte_eth_dev *eth_dev, int mask) struct rte_eth_conf *dev_conf = ð_dev->data->dev_conf; uint64_t offloads = dev_conf->rxmode.offloads; - if (mask & ETH_VLAN_STRIP_MASK) { + if (mask & RTE_ETH_VLAN_STRIP_MASK) { if (avp->host_features & RTE_AVP_FEATURE_VLAN_OFFLOAD) { - if (offloads & DEV_RX_OFFLOAD_VLAN_STRIP) + if (offloads & RTE_ETH_RX_OFFLOAD_VLAN_STRIP) avp->features |= RTE_AVP_FEATURE_VLAN_OFFLOAD; else avp->features &= ~RTE_AVP_FEATURE_VLAN_OFFLOAD; @@ -2226,13 +2216,13 @@ avp_vlan_offload_set(struct rte_eth_dev *eth_dev, int mask) } } - if (mask & ETH_VLAN_FILTER_MASK) { - if (offloads & DEV_RX_OFFLOAD_VLAN_FILTER) + if (mask & RTE_ETH_VLAN_FILTER_MASK) { + if (offloads & RTE_ETH_RX_OFFLOAD_VLAN_FILTER) PMD_DRV_LOG(ERR, "VLAN filter offload not supported\n"); } - if (mask & ETH_VLAN_EXTEND_MASK) { - if (offloads & DEV_RX_OFFLOAD_VLAN_EXTEND) + if (mask & RTE_ETH_VLAN_EXTEND_MASK) { + if (offloads & RTE_ETH_RX_OFFLOAD_VLAN_EXTEND) PMD_DRV_LOG(ERR, "VLAN extend offload not supported\n"); } @@ -2306,10 +2296,4 @@ avp_dev_stats_reset(struct rte_eth_dev *eth_dev) RTE_PMD_REGISTER_PCI(net_avp, rte_avp_pmd); RTE_PMD_REGISTER_PCI_TABLE(net_avp, pci_id_avp_map); - -RTE_INIT(avp_init_log) -{ - avp_logtype_driver = rte_log_register("pmd.net.avp.driver"); - if (avp_logtype_driver >= 0) - rte_log_set_level(avp_logtype_driver, RTE_LOG_NOTICE); -} +RTE_LOG_REGISTER_SUFFIX(avp_logtype_driver, driver, NOTICE);