1 /* SPDX-License-Identifier: BSD-3-Clause
2 * Copyright (c) 2013-2015 Brocade Communications Systems, Inc.
3 * Copyright (c) 2015-2018 Cavium Inc.
9 #include "bnx2x_rxtx.h"
11 #include <rte_string_fns.h>
13 #include <rte_ethdev_pci.h>
14 #include <rte_alarm.h>
16 int bnx2x_logtype_init;
17 int bnx2x_logtype_driver;
20 * The set of PCI devices this driver supports
22 #define BROADCOM_PCI_VENDOR_ID 0x14E4
23 static const struct rte_pci_id pci_id_bnx2x_map[] = {
24 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57800) },
25 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57711) },
26 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57810) },
27 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57811) },
28 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_OBS) },
29 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_4_10) },
30 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_2_20) },
31 #ifdef RTE_LIBRTE_BNX2X_MF_SUPPORT
32 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57810_MF) },
33 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57811_MF) },
34 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_MF) },
39 static const struct rte_pci_id pci_id_bnx2xvf_map[] = {
40 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57800_VF) },
41 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57810_VF) },
42 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57811_VF) },
43 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_VF) },
47 struct rte_bnx2x_xstats_name_off {
48 char name[RTE_ETH_XSTATS_NAME_SIZE];
53 static const struct rte_bnx2x_xstats_name_off bnx2x_xstats_strings[] = {
55 offsetof(struct bnx2x_eth_stats, brb_drop_hi),
56 offsetof(struct bnx2x_eth_stats, brb_drop_lo)},
57 {"rx_buffer_truncates",
58 offsetof(struct bnx2x_eth_stats, brb_truncate_hi),
59 offsetof(struct bnx2x_eth_stats, brb_truncate_lo)},
60 {"rx_buffer_truncate_discard",
61 offsetof(struct bnx2x_eth_stats, brb_truncate_discard),
62 offsetof(struct bnx2x_eth_stats, brb_truncate_discard)},
63 {"mac_filter_discard",
64 offsetof(struct bnx2x_eth_stats, mac_filter_discard),
65 offsetof(struct bnx2x_eth_stats, mac_filter_discard)},
66 {"no_match_vlan_tag_discard",
67 offsetof(struct bnx2x_eth_stats, mf_tag_discard),
68 offsetof(struct bnx2x_eth_stats, mf_tag_discard)},
70 offsetof(struct bnx2x_eth_stats, pause_frames_sent_hi),
71 offsetof(struct bnx2x_eth_stats, pause_frames_sent_lo)},
73 offsetof(struct bnx2x_eth_stats, pause_frames_received_hi),
74 offsetof(struct bnx2x_eth_stats, pause_frames_received_lo)},
75 {"tx_priority_flow_control",
76 offsetof(struct bnx2x_eth_stats, pfc_frames_sent_hi),
77 offsetof(struct bnx2x_eth_stats, pfc_frames_sent_lo)},
78 {"rx_priority_flow_control",
79 offsetof(struct bnx2x_eth_stats, pfc_frames_received_hi),
80 offsetof(struct bnx2x_eth_stats, pfc_frames_received_lo)}
84 bnx2x_link_update(struct rte_eth_dev *dev)
86 struct bnx2x_softc *sc = dev->data->dev_private;
87 struct rte_eth_link link;
89 PMD_INIT_FUNC_TRACE(sc);
91 memset(&link, 0, sizeof(link));
93 link.link_speed = sc->link_vars.line_speed;
94 switch (sc->link_vars.duplex) {
96 link.link_duplex = ETH_LINK_FULL_DUPLEX;
99 link.link_duplex = ETH_LINK_HALF_DUPLEX;
102 link.link_autoneg = !(dev->data->dev_conf.link_speeds &
103 ETH_LINK_SPEED_FIXED);
104 link.link_status = sc->link_vars.link_up;
106 return rte_eth_linkstatus_set(dev, &link);
110 bnx2x_interrupt_action(struct rte_eth_dev *dev, int intr_cxt)
112 struct bnx2x_softc *sc = dev->data->dev_private;
113 uint32_t link_status;
115 bnx2x_intr_legacy(sc);
117 if ((atomic_load_acq_long(&sc->periodic_flags) == PERIODIC_GO) &&
119 bnx2x_periodic_callout(sc);
120 link_status = REG_RD(sc, sc->link_params.shmem_base +
121 offsetof(struct shmem_region,
122 port_mb[sc->link_params.port].link_status));
123 if ((link_status & LINK_STATUS_LINK_UP) != dev->data->dev_link.link_status)
124 bnx2x_link_update(dev);
128 bnx2x_interrupt_handler(void *param)
130 struct rte_eth_dev *dev = (struct rte_eth_dev *)param;
131 struct bnx2x_softc *sc = dev->data->dev_private;
133 PMD_DEBUG_PERIODIC_LOG(INFO, sc, "Interrupt handled");
135 bnx2x_interrupt_action(dev, 1);
136 rte_intr_enable(&sc->pci_dev->intr_handle);
139 static void bnx2x_periodic_start(void *param)
141 struct rte_eth_dev *dev = (struct rte_eth_dev *)param;
142 struct bnx2x_softc *sc = dev->data->dev_private;
145 atomic_store_rel_long(&sc->periodic_flags, PERIODIC_GO);
146 bnx2x_interrupt_action(dev, 0);
148 ret = rte_eal_alarm_set(BNX2X_SP_TIMER_PERIOD,
149 bnx2x_periodic_start, (void *)dev);
151 PMD_DRV_LOG(ERR, sc, "Unable to start periodic"
152 " timer rc %d", ret);
157 void bnx2x_periodic_stop(void *param)
159 struct rte_eth_dev *dev = (struct rte_eth_dev *)param;
160 struct bnx2x_softc *sc = dev->data->dev_private;
162 atomic_store_rel_long(&sc->periodic_flags, PERIODIC_STOP);
164 rte_eal_alarm_cancel(bnx2x_periodic_start, (void *)dev);
166 PMD_DRV_LOG(DEBUG, sc, "Periodic poll stopped");
170 * Devops - helper functions can be called from user application
174 bnx2x_dev_configure(struct rte_eth_dev *dev)
176 struct bnx2x_softc *sc = dev->data->dev_private;
177 struct rte_eth_rxmode *rxmode = &dev->data->dev_conf.rxmode;
179 int mp_ncpus = sysconf(_SC_NPROCESSORS_CONF);
181 PMD_INIT_FUNC_TRACE(sc);
183 if (rxmode->offloads & DEV_RX_OFFLOAD_JUMBO_FRAME) {
184 sc->mtu = dev->data->dev_conf.rxmode.max_rx_pkt_len;
185 dev->data->mtu = sc->mtu;
188 if (dev->data->nb_tx_queues > dev->data->nb_rx_queues) {
189 PMD_DRV_LOG(ERR, sc, "The number of TX queues is greater than number of RX queues");
193 sc->num_queues = MAX(dev->data->nb_rx_queues, dev->data->nb_tx_queues);
194 if (sc->num_queues > mp_ncpus) {
195 PMD_DRV_LOG(ERR, sc, "The number of queues is more than number of CPUs");
199 PMD_DRV_LOG(DEBUG, sc, "num_queues=%d, mtu=%d",
200 sc->num_queues, sc->mtu);
203 if (bnx2x_alloc_ilt_mem(sc) != 0) {
204 PMD_DRV_LOG(ERR, sc, "bnx2x_alloc_ilt_mem was failed");
208 bnx2x_dev_rxtx_init_dummy(dev);
213 bnx2x_dev_start(struct rte_eth_dev *dev)
215 struct bnx2x_softc *sc = dev->data->dev_private;
218 PMD_INIT_FUNC_TRACE(sc);
220 /* start the periodic callout */
221 if (atomic_load_acq_long(&sc->periodic_flags) == PERIODIC_STOP) {
222 bnx2x_periodic_start(dev);
223 PMD_DRV_LOG(DEBUG, sc, "Periodic poll re-started");
226 ret = bnx2x_init(sc);
228 PMD_DRV_LOG(DEBUG, sc, "bnx2x_init failed (%d)", ret);
233 rte_intr_callback_register(&sc->pci_dev->intr_handle,
234 bnx2x_interrupt_handler, (void *)dev);
236 if (rte_intr_enable(&sc->pci_dev->intr_handle))
237 PMD_DRV_LOG(ERR, sc, "rte_intr_enable failed");
240 bnx2x_dev_rxtx_init(dev);
242 bnx2x_print_device_info(sc);
248 bnx2x_dev_stop(struct rte_eth_dev *dev)
250 struct bnx2x_softc *sc = dev->data->dev_private;
253 PMD_INIT_FUNC_TRACE(sc);
255 bnx2x_dev_rxtx_init_dummy(dev);
258 rte_intr_disable(&sc->pci_dev->intr_handle);
259 rte_intr_callback_unregister(&sc->pci_dev->intr_handle,
260 bnx2x_interrupt_handler, (void *)dev);
263 /* stop the periodic callout */
264 bnx2x_periodic_stop(dev);
266 ret = bnx2x_nic_unload(sc, UNLOAD_NORMAL, FALSE);
268 PMD_DRV_LOG(DEBUG, sc, "bnx2x_nic_unload failed (%d)", ret);
276 bnx2x_dev_close(struct rte_eth_dev *dev)
278 struct bnx2x_softc *sc = dev->data->dev_private;
280 PMD_INIT_FUNC_TRACE(sc);
285 bnx2x_dev_clear_queues(dev);
286 memset(&(dev->data->dev_link), 0 , sizeof(struct rte_eth_link));
289 bnx2x_free_ilt_mem(sc);
293 bnx2x_promisc_enable(struct rte_eth_dev *dev)
295 struct bnx2x_softc *sc = dev->data->dev_private;
297 PMD_INIT_FUNC_TRACE(sc);
298 sc->rx_mode = BNX2X_RX_MODE_PROMISC;
299 if (rte_eth_allmulticast_get(dev->data->port_id) == 1)
300 sc->rx_mode = BNX2X_RX_MODE_ALLMULTI_PROMISC;
301 bnx2x_set_rx_mode(sc);
305 bnx2x_promisc_disable(struct rte_eth_dev *dev)
307 struct bnx2x_softc *sc = dev->data->dev_private;
309 PMD_INIT_FUNC_TRACE(sc);
310 sc->rx_mode = BNX2X_RX_MODE_NORMAL;
311 if (rte_eth_allmulticast_get(dev->data->port_id) == 1)
312 sc->rx_mode = BNX2X_RX_MODE_ALLMULTI;
313 bnx2x_set_rx_mode(sc);
317 bnx2x_dev_allmulticast_enable(struct rte_eth_dev *dev)
319 struct bnx2x_softc *sc = dev->data->dev_private;
321 PMD_INIT_FUNC_TRACE(sc);
322 sc->rx_mode = BNX2X_RX_MODE_ALLMULTI;
323 if (rte_eth_promiscuous_get(dev->data->port_id) == 1)
324 sc->rx_mode = BNX2X_RX_MODE_ALLMULTI_PROMISC;
325 bnx2x_set_rx_mode(sc);
329 bnx2x_dev_allmulticast_disable(struct rte_eth_dev *dev)
331 struct bnx2x_softc *sc = dev->data->dev_private;
333 PMD_INIT_FUNC_TRACE(sc);
334 sc->rx_mode = BNX2X_RX_MODE_NORMAL;
335 if (rte_eth_promiscuous_get(dev->data->port_id) == 1)
336 sc->rx_mode = BNX2X_RX_MODE_PROMISC;
337 bnx2x_set_rx_mode(sc);
341 bnx2x_dev_link_update(struct rte_eth_dev *dev, __rte_unused int wait_to_complete)
343 struct bnx2x_softc *sc = dev->data->dev_private;
345 PMD_INIT_FUNC_TRACE(sc);
347 return bnx2x_link_update(dev);
351 bnx2xvf_dev_link_update(struct rte_eth_dev *dev, __rte_unused int wait_to_complete)
353 struct bnx2x_softc *sc = dev->data->dev_private;
356 ret = bnx2x_link_update(dev);
358 bnx2x_check_bull(sc);
359 if (sc->old_bulletin.valid_bitmap & (1 << CHANNEL_DOWN)) {
360 PMD_DRV_LOG(ERR, sc, "PF indicated channel is down."
361 "VF device is no longer operational");
362 dev->data->dev_link.link_status = ETH_LINK_DOWN;
369 bnx2x_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats)
371 struct bnx2x_softc *sc = dev->data->dev_private;
372 uint32_t brb_truncate_discard;
374 uint64_t brb_truncates;
376 PMD_INIT_FUNC_TRACE(sc);
378 bnx2x_stats_handle(sc, STATS_EVENT_UPDATE);
380 memset(stats, 0, sizeof (struct rte_eth_stats));
383 HILO_U64(sc->eth_stats.total_unicast_packets_received_hi,
384 sc->eth_stats.total_unicast_packets_received_lo) +
385 HILO_U64(sc->eth_stats.total_multicast_packets_received_hi,
386 sc->eth_stats.total_multicast_packets_received_lo) +
387 HILO_U64(sc->eth_stats.total_broadcast_packets_received_hi,
388 sc->eth_stats.total_broadcast_packets_received_lo);
391 HILO_U64(sc->eth_stats.total_unicast_packets_transmitted_hi,
392 sc->eth_stats.total_unicast_packets_transmitted_lo) +
393 HILO_U64(sc->eth_stats.total_multicast_packets_transmitted_hi,
394 sc->eth_stats.total_multicast_packets_transmitted_lo) +
395 HILO_U64(sc->eth_stats.total_broadcast_packets_transmitted_hi,
396 sc->eth_stats.total_broadcast_packets_transmitted_lo);
399 HILO_U64(sc->eth_stats.total_bytes_received_hi,
400 sc->eth_stats.total_bytes_received_lo);
403 HILO_U64(sc->eth_stats.total_bytes_transmitted_hi,
404 sc->eth_stats.total_bytes_transmitted_lo);
407 HILO_U64(sc->eth_stats.error_bytes_received_hi,
408 sc->eth_stats.error_bytes_received_lo);
413 HILO_U64(sc->eth_stats.no_buff_discard_hi,
414 sc->eth_stats.no_buff_discard_lo);
417 HILO_U64(sc->eth_stats.brb_drop_hi,
418 sc->eth_stats.brb_drop_lo);
421 HILO_U64(sc->eth_stats.brb_truncate_hi,
422 sc->eth_stats.brb_truncate_lo);
424 brb_truncate_discard = sc->eth_stats.brb_truncate_discard;
426 stats->imissed = brb_drops + brb_truncates +
427 brb_truncate_discard + stats->rx_nombuf;
433 bnx2x_get_xstats_names(__rte_unused struct rte_eth_dev *dev,
434 struct rte_eth_xstat_name *xstats_names,
435 __rte_unused unsigned limit)
437 unsigned int i, stat_cnt = RTE_DIM(bnx2x_xstats_strings);
439 if (xstats_names != NULL)
440 for (i = 0; i < stat_cnt; i++)
441 strlcpy(xstats_names[i].name,
442 bnx2x_xstats_strings[i].name,
443 sizeof(xstats_names[i].name));
449 bnx2x_dev_xstats_get(struct rte_eth_dev *dev, struct rte_eth_xstat *xstats,
452 struct bnx2x_softc *sc = dev->data->dev_private;
453 unsigned int num = RTE_DIM(bnx2x_xstats_strings);
458 bnx2x_stats_handle(sc, STATS_EVENT_UPDATE);
460 for (num = 0; num < n; num++) {
461 if (bnx2x_xstats_strings[num].offset_hi !=
462 bnx2x_xstats_strings[num].offset_lo)
463 xstats[num].value = HILO_U64(
464 *(uint32_t *)((char *)&sc->eth_stats +
465 bnx2x_xstats_strings[num].offset_hi),
466 *(uint32_t *)((char *)&sc->eth_stats +
467 bnx2x_xstats_strings[num].offset_lo));
470 *(uint64_t *)((char *)&sc->eth_stats +
471 bnx2x_xstats_strings[num].offset_lo);
472 xstats[num].id = num;
479 bnx2x_dev_infos_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info)
481 struct bnx2x_softc *sc = dev->data->dev_private;
483 dev_info->max_rx_queues = sc->max_rx_queues;
484 dev_info->max_tx_queues = sc->max_tx_queues;
485 dev_info->min_rx_bufsize = BNX2X_MIN_RX_BUF_SIZE;
486 dev_info->max_rx_pktlen = BNX2X_MAX_RX_PKT_LEN;
487 dev_info->max_mac_addrs = BNX2X_MAX_MAC_ADDRS;
488 dev_info->speed_capa = ETH_LINK_SPEED_10G | ETH_LINK_SPEED_20G;
489 dev_info->rx_offload_capa = DEV_RX_OFFLOAD_JUMBO_FRAME;
491 dev_info->rx_desc_lim.nb_max = MAX_RX_AVAIL;
492 dev_info->rx_desc_lim.nb_min = MIN_RX_SIZE_NONTPA;
493 dev_info->tx_desc_lim.nb_max = MAX_TX_AVAIL;
497 bnx2x_mac_addr_add(struct rte_eth_dev *dev, struct rte_ether_addr *mac_addr,
498 uint32_t index, uint32_t pool)
500 struct bnx2x_softc *sc = dev->data->dev_private;
502 if (sc->mac_ops.mac_addr_add) {
503 sc->mac_ops.mac_addr_add(dev, mac_addr, index, pool);
510 bnx2x_mac_addr_remove(struct rte_eth_dev *dev, uint32_t index)
512 struct bnx2x_softc *sc = dev->data->dev_private;
514 if (sc->mac_ops.mac_addr_remove)
515 sc->mac_ops.mac_addr_remove(dev, index);
518 static const struct eth_dev_ops bnx2x_eth_dev_ops = {
519 .dev_configure = bnx2x_dev_configure,
520 .dev_start = bnx2x_dev_start,
521 .dev_stop = bnx2x_dev_stop,
522 .dev_close = bnx2x_dev_close,
523 .promiscuous_enable = bnx2x_promisc_enable,
524 .promiscuous_disable = bnx2x_promisc_disable,
525 .allmulticast_enable = bnx2x_dev_allmulticast_enable,
526 .allmulticast_disable = bnx2x_dev_allmulticast_disable,
527 .link_update = bnx2x_dev_link_update,
528 .stats_get = bnx2x_dev_stats_get,
529 .xstats_get = bnx2x_dev_xstats_get,
530 .xstats_get_names = bnx2x_get_xstats_names,
531 .dev_infos_get = bnx2x_dev_infos_get,
532 .rx_queue_setup = bnx2x_dev_rx_queue_setup,
533 .rx_queue_release = bnx2x_dev_rx_queue_release,
534 .tx_queue_setup = bnx2x_dev_tx_queue_setup,
535 .tx_queue_release = bnx2x_dev_tx_queue_release,
536 .mac_addr_add = bnx2x_mac_addr_add,
537 .mac_addr_remove = bnx2x_mac_addr_remove,
541 * dev_ops for virtual function
543 static const struct eth_dev_ops bnx2xvf_eth_dev_ops = {
544 .dev_configure = bnx2x_dev_configure,
545 .dev_start = bnx2x_dev_start,
546 .dev_stop = bnx2x_dev_stop,
547 .dev_close = bnx2x_dev_close,
548 .promiscuous_enable = bnx2x_promisc_enable,
549 .promiscuous_disable = bnx2x_promisc_disable,
550 .allmulticast_enable = bnx2x_dev_allmulticast_enable,
551 .allmulticast_disable = bnx2x_dev_allmulticast_disable,
552 .link_update = bnx2xvf_dev_link_update,
553 .stats_get = bnx2x_dev_stats_get,
554 .xstats_get = bnx2x_dev_xstats_get,
555 .xstats_get_names = bnx2x_get_xstats_names,
556 .dev_infos_get = bnx2x_dev_infos_get,
557 .rx_queue_setup = bnx2x_dev_rx_queue_setup,
558 .rx_queue_release = bnx2x_dev_rx_queue_release,
559 .tx_queue_setup = bnx2x_dev_tx_queue_setup,
560 .tx_queue_release = bnx2x_dev_tx_queue_release,
561 .mac_addr_add = bnx2x_mac_addr_add,
562 .mac_addr_remove = bnx2x_mac_addr_remove,
567 bnx2x_common_dev_init(struct rte_eth_dev *eth_dev, int is_vf)
570 struct rte_pci_device *pci_dev;
571 struct rte_pci_addr pci_addr;
572 struct bnx2x_softc *sc;
573 static bool adapter_info = true;
575 /* Extract key data structures */
576 sc = eth_dev->data->dev_private;
577 pci_dev = RTE_DEV_TO_PCI(eth_dev->device);
578 pci_addr = pci_dev->addr;
580 snprintf(sc->devinfo.name, NAME_SIZE, PCI_SHORT_PRI_FMT ":dpdk-port-%u",
581 pci_addr.bus, pci_addr.devid, pci_addr.function,
582 eth_dev->data->port_id);
584 PMD_INIT_FUNC_TRACE(sc);
586 eth_dev->dev_ops = is_vf ? &bnx2xvf_eth_dev_ops : &bnx2x_eth_dev_ops;
588 rte_eth_copy_pci_info(eth_dev, pci_dev);
590 sc->pcie_bus = pci_dev->addr.bus;
591 sc->pcie_device = pci_dev->addr.devid;
593 sc->devinfo.vendor_id = pci_dev->id.vendor_id;
594 sc->devinfo.device_id = pci_dev->id.device_id;
595 sc->devinfo.subvendor_id = pci_dev->id.subsystem_vendor_id;
596 sc->devinfo.subdevice_id = pci_dev->id.subsystem_device_id;
599 sc->flags = BNX2X_IS_VF_FLAG;
601 sc->pcie_func = pci_dev->addr.function;
602 sc->bar[BAR0].base_addr = (void *)pci_dev->mem_resource[0].addr;
604 sc->bar[BAR1].base_addr = (void *)
605 ((uintptr_t)pci_dev->mem_resource[0].addr + PXP_VF_ADDR_DB_START);
607 sc->bar[BAR1].base_addr = pci_dev->mem_resource[2].addr;
609 assert(sc->bar[BAR0].base_addr);
610 assert(sc->bar[BAR1].base_addr);
612 bnx2x_load_firmware(sc);
613 assert(sc->firmware);
615 if (eth_dev->data->dev_conf.rx_adv_conf.rss_conf.rss_hf & ETH_RSS_NONFRAG_IPV4_UDP)
618 sc->rx_budget = BNX2X_RX_BUDGET;
619 sc->hc_rx_ticks = BNX2X_RX_TICKS;
620 sc->hc_tx_ticks = BNX2X_TX_TICKS;
622 sc->interrupt_mode = INTR_MODE_SINGLE_MSIX;
623 sc->rx_mode = BNX2X_RX_MODE_NORMAL;
625 sc->pci_dev = pci_dev;
626 ret = bnx2x_attach(sc);
628 PMD_DRV_LOG(ERR, sc, "bnx2x_attach failed (%d)", ret);
632 /* Print important adapter info for the user. */
634 bnx2x_print_adapter_info(sc);
635 adapter_info = false;
638 /* schedule periodic poll for slowpath link events */
640 PMD_DRV_LOG(DEBUG, sc, "Scheduling periodic poll for slowpath link events");
641 ret = rte_eal_alarm_set(BNX2X_SP_TIMER_PERIOD,
642 bnx2x_periodic_start, (void *)eth_dev);
644 PMD_DRV_LOG(ERR, sc, "Unable to start periodic"
645 " timer rc %d", ret);
650 eth_dev->data->mac_addrs =
651 (struct rte_ether_addr *)sc->link_params.mac_addr;
654 rte_spinlock_init(&sc->vf2pf_lock);
656 ret = bnx2x_dma_alloc(sc, sizeof(struct bnx2x_vf_mbx_msg),
657 &sc->vf2pf_mbox_mapping, "vf2pf_mbox",
658 RTE_CACHE_LINE_SIZE);
662 sc->vf2pf_mbox = (struct bnx2x_vf_mbx_msg *)
663 sc->vf2pf_mbox_mapping.vaddr;
665 ret = bnx2x_dma_alloc(sc, sizeof(struct bnx2x_vf_bulletin),
666 &sc->pf2vf_bulletin_mapping, "vf2pf_bull",
667 RTE_CACHE_LINE_SIZE);
671 sc->pf2vf_bulletin = (struct bnx2x_vf_bulletin *)
672 sc->pf2vf_bulletin_mapping.vaddr;
674 ret = bnx2x_vf_get_resources(sc, sc->max_tx_queues,
683 bnx2x_periodic_stop(eth_dev);
688 eth_bnx2x_dev_init(struct rte_eth_dev *eth_dev)
690 struct bnx2x_softc *sc = eth_dev->data->dev_private;
691 PMD_INIT_FUNC_TRACE(sc);
692 return bnx2x_common_dev_init(eth_dev, 0);
696 eth_bnx2xvf_dev_init(struct rte_eth_dev *eth_dev)
698 struct bnx2x_softc *sc = eth_dev->data->dev_private;
699 PMD_INIT_FUNC_TRACE(sc);
700 return bnx2x_common_dev_init(eth_dev, 1);
703 static int eth_bnx2x_dev_uninit(struct rte_eth_dev *eth_dev)
705 /* mac_addrs must not be freed alone because part of dev_private */
706 eth_dev->data->mac_addrs = NULL;
710 static struct rte_pci_driver rte_bnx2x_pmd;
711 static struct rte_pci_driver rte_bnx2xvf_pmd;
713 static int eth_bnx2x_pci_probe(struct rte_pci_driver *pci_drv,
714 struct rte_pci_device *pci_dev)
716 if (pci_drv == &rte_bnx2x_pmd)
717 return rte_eth_dev_pci_generic_probe(pci_dev,
718 sizeof(struct bnx2x_softc), eth_bnx2x_dev_init);
719 else if (pci_drv == &rte_bnx2xvf_pmd)
720 return rte_eth_dev_pci_generic_probe(pci_dev,
721 sizeof(struct bnx2x_softc), eth_bnx2xvf_dev_init);
726 static int eth_bnx2x_pci_remove(struct rte_pci_device *pci_dev)
728 return rte_eth_dev_pci_generic_remove(pci_dev, eth_bnx2x_dev_uninit);
731 static struct rte_pci_driver rte_bnx2x_pmd = {
732 .id_table = pci_id_bnx2x_map,
733 .drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_INTR_LSC,
734 .probe = eth_bnx2x_pci_probe,
735 .remove = eth_bnx2x_pci_remove,
739 * virtual function driver struct
741 static struct rte_pci_driver rte_bnx2xvf_pmd = {
742 .id_table = pci_id_bnx2xvf_map,
743 .drv_flags = RTE_PCI_DRV_NEED_MAPPING,
744 .probe = eth_bnx2x_pci_probe,
745 .remove = eth_bnx2x_pci_remove,
748 RTE_PMD_REGISTER_PCI(net_bnx2x, rte_bnx2x_pmd);
749 RTE_PMD_REGISTER_PCI_TABLE(net_bnx2x, pci_id_bnx2x_map);
750 RTE_PMD_REGISTER_KMOD_DEP(net_bnx2x, "* igb_uio | uio_pci_generic | vfio-pci");
751 RTE_PMD_REGISTER_PCI(net_bnx2xvf, rte_bnx2xvf_pmd);
752 RTE_PMD_REGISTER_PCI_TABLE(net_bnx2xvf, pci_id_bnx2xvf_map);
753 RTE_PMD_REGISTER_KMOD_DEP(net_bnx2xvf, "* igb_uio | vfio-pci");
755 RTE_INIT(bnx2x_init_log)
757 bnx2x_logtype_init = rte_log_register("pmd.net.bnx2x.init");
758 if (bnx2x_logtype_init >= 0)
759 rte_log_set_level(bnx2x_logtype_init, RTE_LOG_NOTICE);
760 bnx2x_logtype_driver = rte_log_register("pmd.net.bnx2x.driver");
761 if (bnx2x_logtype_driver >= 0)
762 rte_log_set_level(bnx2x_logtype_driver, RTE_LOG_NOTICE);