eth_dev->dev_ops = &avp_eth_dev_ops;
eth_dev->rx_pkt_burst = &avp_recv_pkts;
eth_dev->tx_pkt_burst = &avp_xmit_pkts;
+ /* Let rte_eth_dev_close() release the port resources */
+ eth_dev->data->dev_flags |= RTE_ETH_DEV_CLOSE_REMOVE;
if (rte_eal_process_type() != RTE_PROC_PRIMARY) {
/*
static int
eth_avp_dev_uninit(struct rte_eth_dev *eth_dev)
{
- int ret;
-
if (rte_eal_process_type() != RTE_PROC_PRIMARY)
return -EPERM;
if (eth_dev->data == NULL)
return 0;
- ret = avp_dev_disable_interrupts(eth_dev);
- if (ret != 0) {
- PMD_DRV_LOG(ERR, "Failed to disable interrupts, ret=%d\n", ret);
- return ret;
- }
+ avp_dev_close(eth_dev);
return 0;
}
unsigned int i;
for (i = 0; i < avp->num_rx_queues; i++) {
- if (data->rx_queues[i] == rxq)
+ if (data->rx_queues[i] == rxq) {
+ rte_free(data->rx_queues[i]);
+ data->rx_queues[i] = NULL;
+ }
+ }
+}
+
+static void
+avp_dev_rx_queue_release_all(struct rte_eth_dev *eth_dev)
+{
+ struct avp_dev *avp = AVP_DEV_PRIVATE_TO_HW(eth_dev->data->dev_private);
+ struct rte_eth_dev_data *data = avp->dev_data;
+ unsigned int i;
+
+ for (i = 0; i < avp->num_rx_queues; i++) {
+ if (data->rx_queues[i]) {
+ rte_free(data->rx_queues[i]);
data->rx_queues[i] = NULL;
+ }
}
}
unsigned int i;
for (i = 0; i < avp->num_tx_queues; i++) {
- if (data->tx_queues[i] == txq)
+ if (data->tx_queues[i] == txq) {
+ rte_free(data->tx_queues[i]);
+ data->tx_queues[i] = NULL;
+ }
+ }
+}
+
+static void
+avp_dev_tx_queue_release_all(struct rte_eth_dev *eth_dev)
+{
+ struct avp_dev *avp = AVP_DEV_PRIVATE_TO_HW(eth_dev->data->dev_private);
+ struct rte_eth_dev_data *data = avp->dev_data;
+ unsigned int i;
+
+ for (i = 0; i < avp->num_tx_queues; i++) {
+ if (data->tx_queues[i]) {
+ rte_free(data->tx_queues[i]);
data->tx_queues[i] = NULL;
+ }
}
}
/* continue */
}
+ /* release dynamic storage for rx/tx queues */
+ avp_dev_rx_queue_release_all(eth_dev);
+ avp_dev_tx_queue_release_all(eth_dev);
+
unlock:
rte_spinlock_unlock(&avp->lock);
}