X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=drivers%2Fnet%2Fenic%2Fenic.h;h=cd66348f2fa82eb41b946b5c01f1c51859e1bd62;hb=56175f74eaa565cdfde931d83be1db9faaa81667;hp=4ea4e4a31e38c06b7767346b200f8a565a89fffe;hpb=dfbd6a9cb50409dc6558f97cd5adda2590921542;p=dpdk.git diff --git a/drivers/net/enic/enic.h b/drivers/net/enic/enic.h index 4ea4e4a31e..cd66348f2f 100644 --- a/drivers/net/enic/enic.h +++ b/drivers/net/enic/enic.h @@ -1,42 +1,16 @@ -/* - * Copyright 2008-2014 Cisco Systems, Inc. All rights reserved. +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright 2008-2017 Cisco Systems, Inc. All rights reserved. * Copyright 2007 Nuova Systems, Inc. All rights reserved. - * - * Copyright (c) 2014, Cisco Systems, Inc. - * All rights reserved. - * - * Redistribution and use in source and binary forms, with or without - * modification, are permitted provided that the following conditions - * are met: - * - * 1. Redistributions of source code must retain the above copyright - * notice, this list of conditions and the following disclaimer. - * - * 2. Redistributions in binary form must reproduce the above copyright - * notice, this list of conditions and the following disclaimer in - * the documentation and/or other materials provided with the - * distribution. - * - * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS - * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT - * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS - * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE - * COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, - * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, - * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; - * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER - * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT - * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN - * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE - * POSSIBILITY OF SUCH DAMAGE. - * */ #ifndef _ENIC_H_ #define _ENIC_H_ +#include +#include #include "vnic_enet.h" #include "vnic_dev.h" +#include "vnic_flowman.h" #include "vnic_wq.h" #include "vnic_rq.h" #include "vnic_cq.h" @@ -46,6 +20,7 @@ #include "vnic_rss.h" #include "enic_res.h" #include "cq_enet_desc.h" +#include #include #include @@ -53,14 +28,6 @@ #define DRV_DESCRIPTION "Cisco VIC Ethernet NIC Poll-mode Driver" #define DRV_COPYRIGHT "Copyright 2008-2015 Cisco Systems, Inc" -#define ENIC_WQ_MAX 8 -/* With Rx scatter support, we use two RQs on VIC per RQ used by app. Both - * RQs use the same CQ. - */ -#define ENIC_RQ_MAX 16 -#define ENIC_CQ_MAX (ENIC_WQ_MAX + (ENIC_RQ_MAX / 2)) -#define ENIC_INTR_MAX (ENIC_CQ_MAX + 2) - #define VLAN_ETH_HLEN 18 #define ENICPMD_SETTING(enic, f) ((enic->config.flags & VENETF_##f) ? 1 : 0) @@ -78,30 +45,25 @@ #define PCI_DEVICE_ID_CISCO_VIC_ENET 0x0043 /* ethernet vnic */ #define PCI_DEVICE_ID_CISCO_VIC_ENET_VF 0x0071 /* enet SRIOV VF */ +/* enet SRIOV Standalone vNic VF */ +#define PCI_DEVICE_ID_CISCO_VIC_ENET_SN 0x02B7 +/* Special Filter id for non-specific packet flagging. Don't change value */ +#define ENIC_MAGIC_FILTER_ID 0xffff -#define ENICPMD_FDIR_MAX 64 - -struct enic_fdir_node { - struct rte_eth_fdir_filter filter; - u16 fltr_id; - u16 rq_index; -}; - -struct enic_fdir { - struct rte_eth_fdir_stats stats; - struct rte_hash *hash; - struct enic_fdir_node *nodes[ENICPMD_FDIR_MAX]; - u32 modes; - u32 types_mask; - void (*copy_fltr_fn)(struct filter_v2 *filt, - struct rte_eth_fdir_input *input, - struct rte_eth_fdir_masks *masks); -}; +/* + * Interrupt 0: LSC and errors + * Interrupt 1: rx queue 0 + * Interrupt 2: rx queue 1 + * ... + */ +#define ENICPMD_LSC_INTR_OFFSET 0 +#define ENICPMD_RXQ_INTR_OFFSET 1 struct enic_soft_stats { rte_atomic64_t rx_nombuf; rte_atomic64_t rx_packet_errors; + rte_atomic64_t tx_oversized; }; struct enic_memzone_entry { @@ -109,50 +71,86 @@ struct enic_memzone_entry { LIST_ENTRY(enic_memzone_entry) entries; }; +/* Defined in enic_fm_flow.c */ +struct enic_flowman; +struct enic_fm_flow; + +struct rte_flow { + LIST_ENTRY(rte_flow) next; + /* Data for filter API based flow (enic_flow.c) */ + uint16_t enic_filter_id; + struct filter_v2 enic_filter; + /* Data for flow manager based flow (enic_fm_flow.c) */ + struct enic_fm_flow *fm; + int internal; +}; + /* Per-instance private data structure */ struct enic { - struct enic *next; struct rte_pci_device *pdev; struct vnic_enet_config config; struct vnic_dev_bar bar0; struct vnic_dev *vdev; + /* + * mbuf_initializer contains 64 bits of mbuf rearm_data, used by + * the avx2 handler at this time. + */ + uint64_t mbuf_initializer; unsigned int port_id; + bool overlay_offload; struct rte_eth_dev *rte_dev; - struct enic_fdir fdir; + struct rte_eth_dev_data *dev_data; char bdf_name[ENICPMD_BDF_LENGTH]; int dev_fd; int iommu_group_fd; int iommu_groupid; int eventfd; - uint8_t mac_addr[ETH_ALEN]; + uint8_t mac_addr[RTE_ETHER_ADDR_LEN]; pthread_t err_intr_thread; int promisc; int allmulti; - u8 ig_vlan_strip_en; + uint8_t ig_vlan_strip_en; int link_status; - u8 hw_ip_checksum; - u16 max_mtu; - u16 adv_filters; + uint8_t hw_ip_checksum; + uint16_t max_mtu; + uint8_t adv_filters; + uint32_t flow_filter_mode; + uint8_t filter_actions; /* HW supported actions */ + uint64_t cq_entry_sizes; /* supported CQ entry sizes */ + bool vxlan; + bool cq64; /* actually using 64B CQ entry */ + bool cq64_request; /* devargs cq64=1 */ + bool disable_overlay; /* devargs disable_overlay=1 */ + uint8_t enable_avx2_rx; /* devargs enable-avx2-rx=1 */ + uint8_t geneve_opt_avail; /* Geneve with options offload available */ + uint8_t geneve_opt_enabled; /* Geneve with options offload enabled */ + uint8_t geneve_opt_request; /* devargs geneve-opt=1 */ + bool nic_cfg_chk; /* NIC_CFG_CHK available */ + bool udp_rss_weak; /* Bodega style UDP RSS */ + uint8_t ig_vlan_rewrite_mode; /* devargs ig-vlan-rewrite */ + uint16_t vxlan_port; /* current vxlan port pushed to NIC */ + int use_simple_tx_handler; + int use_noscatter_vec_rx_handler; unsigned int flags; unsigned int priv_flags; - /* work queue */ - struct vnic_wq wq[ENIC_WQ_MAX]; - unsigned int wq_count; + /* work queue (len = conf_wq_count) */ + struct vnic_wq *wq; + unsigned int wq_count; /* equals eth_dev nb_tx_queues */ - /* receive queue */ - struct vnic_rq rq[ENIC_RQ_MAX]; - unsigned int rq_count; + /* receive queue (len = conf_rq_count) */ + struct vnic_rq *rq; + unsigned int rq_count; /* equals eth_dev nb_rx_queues */ - /* completion queue */ - struct vnic_cq cq[ENIC_CQ_MAX]; - unsigned int cq_count; + /* completion queue (len = conf_cq_count) */ + struct vnic_cq *cq; + unsigned int cq_count; /* equals rq_count + wq_count */ - /* interrupt resource */ - struct vnic_intr intr; - unsigned int intr_count; + /* interrupt vectors (len = conf_intr_count) */ + struct vnic_intr *intr; + unsigned int intr_count; /* equals enabled interrupts (lsc + rxqs) */ /* software counters */ struct enic_soft_stats soft_stats; @@ -168,21 +166,109 @@ struct enic { rte_spinlock_t memzone_list_lock; rte_spinlock_t mtu_lock; + LIST_HEAD(enic_flows, rte_flow) flows; + + /* RSS */ + uint16_t reta_size; + uint8_t hash_key_size; + uint64_t flow_type_rss_offloads; /* 0 indicates RSS not supported */ + /* + * Keep a copy of current RSS config for queries, as we cannot retrieve + * it from the NIC. + */ + uint8_t rss_hash_type; /* NIC_CFG_RSS_HASH_TYPE flags */ + uint8_t rss_enable; + uint64_t rss_hf; /* ETH_RSS flags */ + union vnic_rss_key rss_key; + union vnic_rss_cpu rss_cpu; + + uint64_t rx_offload_capa; /* DEV_RX_OFFLOAD flags */ + uint64_t tx_offload_capa; /* DEV_TX_OFFLOAD flags */ + uint64_t tx_queue_offload_capa; /* DEV_TX_OFFLOAD flags */ + uint64_t tx_offload_mask; /* PKT_TX flags accepted */ + + /* Multicast MAC addresses added to the NIC */ + uint32_t mc_count; + struct rte_ether_addr mc_addrs[ENIC_MULTICAST_PERFECT_FILTERS]; + + /* Flow manager API */ + struct enic_flowman *fm; + uint64_t fm_vnic_handle; + uint32_t fm_vnic_uif; + /* switchdev */ + uint8_t switchdev_mode; + uint16_t switch_domain_id; + uint16_t max_vf_id; + /* Number of queues needed for VF representor paths */ + uint32_t vf_required_wq; + uint32_t vf_required_cq; + uint32_t vf_required_rq; + /* + * Lock to serialize devcmds from PF, VF representors as they all share + * the same PF devcmd instance in firmware. + */ + rte_spinlock_t devcmd_lock; +}; + +struct enic_vf_representor { + struct enic enic; + struct vnic_enet_config config; + struct rte_eth_dev *eth_dev; + struct rte_ether_addr mac_addr; + struct rte_pci_addr bdf; + struct enic *pf; + uint16_t switch_domain_id; + uint16_t vf_id; + int allmulti; + int promisc; + /* Representor path uses PF queues. These are reserved during init */ + uint16_t pf_wq_idx; /* WQ dedicated to VF rep */ + uint16_t pf_wq_cq_idx; /* CQ for WQ */ + uint16_t pf_rq_sop_idx; /* SOP RQ dedicated to VF rep */ + uint16_t pf_rq_data_idx; /* Data RQ */ + /* Representor flows managed by flowman */ + struct rte_flow *vf2rep_flow[2]; + struct rte_flow *rep2vf_flow[2]; }; -static inline unsigned int enic_rq_sop(unsigned int sop_rq) +#define VF_ENIC_TO_VF_REP(vf_enic) \ + container_of(vf_enic, struct enic_vf_representor, enic) + +static inline int enic_is_vf_rep(struct enic *enic) { - return sop_rq / 2; + return !!(enic->rte_dev->data->dev_flags & RTE_ETH_DEV_REPRESENTOR); } -static inline unsigned int enic_sop_rq(unsigned int rq) +/* Compute ethdev's max packet size from MTU */ +static inline uint32_t enic_mtu_to_max_rx_pktlen(uint32_t mtu) { - return rq * 2; + /* ethdev max size includes eth whereas NIC MTU does not */ + return mtu + RTE_ETHER_HDR_LEN; } -static inline unsigned int enic_data_rq(unsigned int rq) +/* Get the CQ index from a Start of Packet(SOP) RQ index */ +static inline unsigned int enic_sop_rq_idx_to_cq_idx(unsigned int sop_idx) { - return rq * 2 + 1; + return sop_idx; +} + +/* Get the RTE RQ index from a Start of Packet(SOP) RQ index */ +static inline unsigned int enic_sop_rq_idx_to_rte_idx(unsigned int sop_idx) +{ + return sop_idx; +} + +/* Get the Start of Packet(SOP) RQ index from a RTE RQ index */ +static inline unsigned int enic_rte_rq_idx_to_sop_idx(unsigned int rte_idx) +{ + return rte_idx; +} + +/* Get the Data RQ index from a RTE RQ index */ +static inline unsigned int enic_rte_rq_idx_to_data_idx(unsigned int rte_idx, + struct enic *enic) +{ + return enic->rq_count + rte_idx; } static inline unsigned int enic_vnic_rq_count(struct enic *enic) @@ -192,11 +278,7 @@ static inline unsigned int enic_vnic_rq_count(struct enic *enic) static inline unsigned int enic_cq_rq(__rte_unused struct enic *enic, unsigned int rq) { - /* Scatter rx uses two receive queues together with one - * completion queue, so the completion queue number is no - * longer the same as the rq number. - */ - return rq / 2; + return rq; } static inline unsigned int enic_cq_wq(struct enic *enic, unsigned int wq) @@ -204,14 +286,70 @@ static inline unsigned int enic_cq_wq(struct enic *enic, unsigned int wq) return enic->rq_count + wq; } -static inline unsigned int enic_msix_err_intr(__rte_unused struct enic *enic) +/* + * WQ, RQ, CQ allocation scheme. Firmware gives the driver an array of + * WQs, an array of RQs, and an array of CQs. Fow now, these are + * statically allocated between PF app send/receive queues and VF + * representor app send/receive queues. VF representor supports only 1 + * send and 1 receive queue. The number of PF app queue is not known + * until the queue setup time. + * + * R = number of receive queues for PF app + * S = number of send queues for PF app + * V = number of VF representors + * + * wI = WQ for PF app send queue I + * rI = SOP RQ for PF app receive queue I + * dI = Data RQ for rI + * cwI = CQ for wI + * crI = CQ for rI + * vwI = WQ for VF representor send queue I + * vrI = SOP RQ for VF representor receive queue I + * vdI = Data RQ for vrI + * vcwI = CQ for vwI + * vcrI = CQ for vrI + * + * WQ array: | w0 |..| wS-1 |..| vwV-1 |..| vw0 | + * ^ ^ ^ ^ + * index 0 S-1 W-V W-1 W=len(WQ array) + * + * RQ array: | r0 |..| rR-1 |d0 |..|dR-1| ..|vdV-1 |..| vd0 |vrV-1 |..|vr0 | + * ^ ^ ^ ^ ^ ^ ^ ^ + * index 0 R-1 R 2R-1 X-2V X-(V+1) X-V X-1 + * X=len(RQ array) + * + * CQ array: | cr0 |..| crR-1 |cw0|..|cwS-1|..|vcwV-1|..| vcw0|vcrV-1|..|vcr0|.. + * ^ ^ ^ ^ ^ ^ ^ ^ + * index 0 R-1 R R+S-1 X-2V X-(V+1) X-V X-1 + * X is not a typo. It really is len(RQ array) to accommodate enic_cq_rq() used + * throughout RX handlers. The current scheme requires + * len(CQ array) >= len(RQ array). + */ + +static inline unsigned int vf_wq_cq_idx(struct enic_vf_representor *vf) +{ + /* rq is not a typo. index(vcwI) coincides with index(vdI) */ + return vf->pf->conf_rq_count - (vf->pf->max_vf_id + vf->vf_id + 2); +} + +static inline unsigned int vf_wq_idx(struct enic_vf_representor *vf) +{ + return vf->pf->conf_wq_count - vf->vf_id - 1; +} + +static inline unsigned int vf_rq_sop_idx(struct enic_vf_representor *vf) +{ + return vf->pf->conf_rq_count - vf->vf_id - 1; +} + +static inline unsigned int vf_rq_data_idx(struct enic_vf_representor *vf) { - return 0; + return vf->pf->conf_rq_count - (vf->pf->max_vf_id + vf->vf_id + 2); } static inline struct enic *pmd_priv(struct rte_eth_dev *eth_dev) { - return (struct enic *)eth_dev->data->dev_private; + return eth_dev->data->dev_private; } static inline uint32_t @@ -238,62 +376,80 @@ enic_ring_incr(uint32_t n_descriptors, uint32_t idx) return idx; } -extern void enic_fdir_stats_get(struct enic *enic, - struct rte_eth_fdir_stats *stats); -extern int enic_fdir_add_fltr(struct enic *enic, - struct rte_eth_fdir_filter *params); -extern int enic_fdir_del_fltr(struct enic *enic, - struct rte_eth_fdir_filter *params); -extern void enic_free_wq(void *txq); -extern int enic_alloc_intr_resources(struct enic *enic); -extern int enic_setup_finish(struct enic *enic); -extern int enic_alloc_wq(struct enic *enic, uint16_t queue_idx, - unsigned int socket_id, uint16_t nb_desc); -extern void enic_start_wq(struct enic *enic, uint16_t queue_idx); -extern int enic_stop_wq(struct enic *enic, uint16_t queue_idx); -extern void enic_start_rq(struct enic *enic, uint16_t queue_idx); -extern int enic_stop_rq(struct enic *enic, uint16_t queue_idx); -extern void enic_free_rq(void *rxq); -extern int enic_alloc_rq(struct enic *enic, uint16_t queue_idx, - unsigned int socket_id, struct rte_mempool *mp, - uint16_t nb_desc); -extern int enic_set_rss_nic_cfg(struct enic *enic); -extern int enic_set_vnic_res(struct enic *enic); -extern void enic_set_hdr_split_size(struct enic *enic, u16 split_hdr_size); -extern int enic_enable(struct enic *enic); -extern int enic_disable(struct enic *enic); -extern void enic_remove(struct enic *enic); -extern int enic_get_link_status(struct enic *enic); -extern void enic_dev_stats_get(struct enic *enic, - struct rte_eth_stats *r_stats); -extern void enic_dev_stats_clear(struct enic *enic); -extern void enic_add_packet_filter(struct enic *enic); -extern void enic_set_mac_address(struct enic *enic, uint8_t *mac_addr); -extern void enic_del_mac_address(struct enic *enic); -extern unsigned int enic_cleanup_wq(struct enic *enic, struct vnic_wq *wq); -extern void enic_send_pkt(struct enic *enic, struct vnic_wq *wq, - struct rte_mbuf *tx_pkt, unsigned short len, - uint8_t sop, uint8_t eop, uint8_t cq_entry, - uint16_t ol_flags, uint16_t vlan_tag); - -extern void enic_post_wq_index(struct vnic_wq *wq); -extern int enic_probe(struct enic *enic); -extern int enic_clsf_init(struct enic *enic); -extern void enic_clsf_destroy(struct enic *enic); +int dev_is_enic(struct rte_eth_dev *dev); +void enic_free_wq(void *txq); +int enic_alloc_intr_resources(struct enic *enic); +int enic_setup_finish(struct enic *enic); +int enic_alloc_wq(struct enic *enic, uint16_t queue_idx, + unsigned int socket_id, uint16_t nb_desc); +void enic_start_wq(struct enic *enic, uint16_t queue_idx); +int enic_stop_wq(struct enic *enic, uint16_t queue_idx); +void enic_start_rq(struct enic *enic, uint16_t queue_idx); +int enic_stop_rq(struct enic *enic, uint16_t queue_idx); +void enic_free_rq(void *rxq); +int enic_alloc_rq(struct enic *enic, uint16_t queue_idx, + unsigned int socket_id, struct rte_mempool *mp, + uint16_t nb_desc, uint16_t free_thresh); +int enic_set_vnic_res(struct enic *enic); +int enic_init_rss_nic_cfg(struct enic *enic); +int enic_set_rss_conf(struct enic *enic, + struct rte_eth_rss_conf *rss_conf); +int enic_set_rss_reta(struct enic *enic, union vnic_rss_cpu *rss_cpu); +int enic_set_vlan_strip(struct enic *enic); +int enic_enable(struct enic *enic); +int enic_disable(struct enic *enic); +void enic_remove(struct enic *enic); +int enic_get_link_status(struct enic *enic); +int enic_dev_stats_get(struct enic *enic, + struct rte_eth_stats *r_stats); +int enic_dev_stats_clear(struct enic *enic); +int enic_add_packet_filter(struct enic *enic); +int enic_set_mac_address(struct enic *enic, uint8_t *mac_addr); +int enic_del_mac_address(struct enic *enic, int mac_index); +unsigned int enic_cleanup_wq(struct enic *enic, struct vnic_wq *wq); +void enic_send_pkt(struct enic *enic, struct vnic_wq *wq, + struct rte_mbuf *tx_pkt, unsigned short len, + uint8_t sop, uint8_t eop, uint8_t cq_entry, + uint16_t ol_flags, uint16_t vlan_tag); + +void enic_post_wq_index(struct vnic_wq *wq); +int enic_probe(struct enic *enic); +int enic_fm_init(struct enic *enic); +void enic_fm_destroy(struct enic *enic); +void *enic_alloc_consistent(void *priv, size_t size, dma_addr_t *dma_handle, + uint8_t *name); +void enic_free_consistent(void *priv, size_t size, void *vaddr, + dma_addr_t dma_handle); uint16_t enic_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, uint16_t nb_pkts); -uint16_t enic_dummy_recv_pkts(__rte_unused void *rx_queue, - __rte_unused struct rte_mbuf **rx_pkts, - __rte_unused uint16_t nb_pkts); +uint16_t enic_recv_pkts_64(void *rx_queue, struct rte_mbuf **rx_pkts, + uint16_t nb_pkts); +uint16_t enic_noscatter_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, + uint16_t nb_pkts); +uint16_t enic_dummy_recv_pkts(void *rx_queue, + struct rte_mbuf **rx_pkts, + uint16_t nb_pkts); uint16_t enic_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, + uint16_t nb_pkts); +uint16_t enic_simple_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts); +uint16_t enic_prep_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, + uint16_t nb_pkts); int enic_set_mtu(struct enic *enic, uint16_t new_mtu); -int enic_link_update(struct enic *enic); -void enic_fdir_info(struct enic *enic); -void enic_fdir_info_get(struct enic *enic, struct rte_eth_fdir_info *stats); -void copy_fltr_v1(struct filter_v2 *fltr, struct rte_eth_fdir_input *input, - struct rte_eth_fdir_masks *masks); -void copy_fltr_v2(__rte_unused struct filter_v2 *fltr, - __rte_unused struct rte_eth_fdir_input *input, - __rte_unused struct rte_eth_fdir_masks *masks); +int enic_link_update(struct rte_eth_dev *eth_dev); +bool enic_use_vector_rx_handler(struct rte_eth_dev *eth_dev); +void enic_pick_rx_handler(struct rte_eth_dev *eth_dev); +void enic_pick_tx_handler(struct rte_eth_dev *eth_dev); +int enic_vf_representor_init(struct rte_eth_dev *eth_dev, void *init_params); +int enic_vf_representor_uninit(struct rte_eth_dev *ethdev); +int enic_fm_allocate_switch_domain(struct enic *pf); +int enic_fm_add_rep2vf_flow(struct enic_vf_representor *vf); +int enic_fm_add_vf2rep_flow(struct enic_vf_representor *vf); +int enic_alloc_rx_queue_mbufs(struct enic *enic, struct vnic_rq *rq); +void enic_rxmbuf_queue_release(struct enic *enic, struct vnic_rq *rq); +void enic_free_wq_buf(struct rte_mbuf **buf); +void enic_free_rq_buf(struct rte_mbuf **mbuf); +extern const struct rte_flow_ops enic_flow_ops; +extern const struct rte_flow_ops enic_fm_flow_ops; + #endif /* _ENIC_H_ */