1 /* SPDX-License-Identifier: BSD-3-Clause
2 * Copyright(c) 2017 Intel Corporation
5 #ifndef _IAVF_ETHDEV_H_
6 #define _IAVF_ETHDEV_H_
10 #include <rte_kvargs.h>
11 #include <rte_tm_driver.h>
13 #include <iavf_prototype.h>
14 #include <iavf_adminq_cmd.h>
15 #include <iavf_type.h>
19 #define IAVF_AQ_LEN 32
20 #define IAVF_AQ_BUF_SZ 4096
21 #define IAVF_RESET_WAIT_CNT 50
22 #define IAVF_BUF_SIZE_MIN 1024
23 #define IAVF_FRAME_SIZE_MAX 9728
24 #define IAVF_QUEUE_BASE_ADDR_UNIT 128
26 #define IAVF_MAX_NUM_QUEUES_DFLT 16
27 #define IAVF_MAX_NUM_QUEUES_LV 256
28 #define IAVF_CFG_Q_NUM_PER_BUF 32
29 #define IAVF_IRQ_MAP_NUM_PER_BUF 128
30 #define IAVF_RXTX_QUEUE_CHUNKS_NUM 2
32 #define IAVF_NUM_MACADDR_MAX 64
34 #define IAVF_DEV_WATCHDOG_PERIOD 0
36 #define IAVF_DEFAULT_RX_PTHRESH 8
37 #define IAVF_DEFAULT_RX_HTHRESH 8
38 #define IAVF_DEFAULT_RX_WTHRESH 0
40 #define IAVF_DEFAULT_RX_FREE_THRESH 32
42 #define IAVF_DEFAULT_TX_PTHRESH 32
43 #define IAVF_DEFAULT_TX_HTHRESH 0
44 #define IAVF_DEFAULT_TX_WTHRESH 0
46 #define IAVF_DEFAULT_TX_FREE_THRESH 32
47 #define IAVF_DEFAULT_TX_RS_THRESH 32
49 #define IAVF_BASIC_OFFLOAD_CAPS ( \
50 VF_BASE_MODE_OFFLOADS | \
51 VIRTCHNL_VF_OFFLOAD_WB_ON_ITR | \
52 VIRTCHNL_VF_OFFLOAD_RX_POLLING)
54 #define IAVF_RSS_OFFLOAD_ALL ( \
56 RTE_ETH_RSS_FRAG_IPV4 | \
57 RTE_ETH_RSS_NONFRAG_IPV4_TCP | \
58 RTE_ETH_RSS_NONFRAG_IPV4_UDP | \
59 RTE_ETH_RSS_NONFRAG_IPV4_SCTP | \
60 RTE_ETH_RSS_NONFRAG_IPV4_OTHER | \
62 RTE_ETH_RSS_FRAG_IPV6 | \
63 RTE_ETH_RSS_NONFRAG_IPV6_TCP | \
64 RTE_ETH_RSS_NONFRAG_IPV6_UDP | \
65 RTE_ETH_RSS_NONFRAG_IPV6_SCTP | \
66 RTE_ETH_RSS_NONFRAG_IPV6_OTHER)
68 #define IAVF_MISC_VEC_ID RTE_INTR_VEC_ZERO_OFFSET
69 #define IAVF_RX_VEC_START RTE_INTR_VEC_RXTX_OFFSET
71 /* Default queue interrupt throttling time in microseconds */
72 #define IAVF_ITR_INDEX_DEFAULT 0
73 #define IAVF_QUEUE_ITR_INTERVAL_DEFAULT 32 /* 32 us */
74 #define IAVF_QUEUE_ITR_INTERVAL_MAX 8160 /* 8160 us */
76 #define IAVF_ALARM_INTERVAL 50000 /* us */
78 /* The overhead from MTU to max frame size.
79 * Considering QinQ packet, the VLAN tag needs to be counted twice.
81 #define IAVF_ETH_OVERHEAD \
82 (RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + RTE_VLAN_HLEN * 2)
83 #define IAVF_ETH_MAX_LEN (RTE_ETHER_MTU + IAVF_ETH_OVERHEAD)
85 #define IAVF_32_BIT_WIDTH (CHAR_BIT * 4)
86 #define IAVF_48_BIT_WIDTH (CHAR_BIT * 6)
87 #define IAVF_48_BIT_MASK RTE_LEN2MASK(IAVF_48_BIT_WIDTH, uint64_t)
89 #define IAVF_RX_DESC_EXT_STATUS_FLEXBH_MASK 0x03
90 #define IAVF_RX_DESC_EXT_STATUS_FLEXBH_FD_ID 0x01
92 #define IAVF_BITS_PER_BYTE 8
94 #define IAVF_VLAN_TAG_PCP_OFFSET 13
96 #define IAVF_L2TPV2_FLAGS_LEN 0x4000
100 struct iavf_tx_queue;
103 struct iavf_ipsec_crypto_stats {
109 uint64_t not_processed;
111 uint64_t ipsec_length;
116 struct iavf_eth_xstats {
117 struct virtchnl_eth_stats eth_stats;
118 struct iavf_ipsec_crypto_stats ips_stats;
121 /* Structure that defines a VSI, associated with a adapter. */
123 struct iavf_adapter *adapter; /* Backreference to associated adapter */
125 uint16_t nb_qps; /* Number of queue pairs VSI can occupy */
126 uint16_t nb_used_qps; /* Number of queue pairs VSI uses */
127 uint16_t max_macaddrs; /* Maximum number of MAC addresses */
128 uint16_t base_vector;
129 uint16_t msix_intr; /* The MSIX interrupt binds to VSI */
130 struct iavf_eth_xstats eth_stats_offset;
134 TAILQ_HEAD(iavf_flow_list, rte_flow);
136 struct iavf_flow_parser_node;
137 TAILQ_HEAD(iavf_parser_list, iavf_flow_parser_node);
139 struct iavf_fdir_conf {
140 struct virtchnl_fdir_add add_fltr;
141 struct virtchnl_fdir_del del_fltr;
147 struct iavf_fdir_info {
148 struct iavf_fdir_conf conf;
156 /* Message type read in admin queue from PF */
157 enum iavf_aq_result {
158 IAVF_MSG_ERR = -1, /* Meet error when accessing admin queue */
159 IAVF_MSG_NON, /* Read nothing from admin queue */
160 IAVF_MSG_SYS, /* Read system msg from admin queue */
161 IAVF_MSG_CMD, /* Read async command result */
164 /* Struct to store Traffic Manager node configuration. */
165 struct iavf_tm_node {
166 TAILQ_ENTRY(iavf_tm_node) node;
171 uint32_t reference_count;
172 struct iavf_tm_node *parent;
173 struct iavf_tm_shaper_profile *shaper_profile;
174 struct rte_tm_node_params params;
177 TAILQ_HEAD(iavf_tm_node_list, iavf_tm_node);
179 struct iavf_tm_shaper_profile {
180 TAILQ_ENTRY(iavf_tm_shaper_profile) node;
181 uint32_t shaper_profile_id;
182 uint32_t reference_count;
183 struct rte_tm_shaper_params profile;
186 TAILQ_HEAD(iavf_shaper_profile_list, iavf_tm_shaper_profile);
188 /* node type of Traffic Manager */
189 enum iavf_tm_node_type {
190 IAVF_TM_NODE_TYPE_PORT,
191 IAVF_TM_NODE_TYPE_TC,
192 IAVF_TM_NODE_TYPE_QUEUE,
193 IAVF_TM_NODE_TYPE_MAX,
196 /* Struct to store all the Traffic Manager configuration. */
197 struct iavf_tm_conf {
198 struct iavf_tm_node *root; /* root node - vf vsi */
199 struct iavf_tm_node_list tc_list; /* node list for all the TCs */
200 struct iavf_tm_node_list queue_list; /* node list for all the queues */
201 struct iavf_shaper_profile_list shaper_profile_list;
203 uint32_t nb_queue_node;
207 /* Struct to store queue TC mapping. Queue is continuous in one TC */
208 struct iavf_qtc_map {
210 uint16_t start_queue_id;
211 uint16_t queue_count;
214 /* Structure to store private data specific for VF instance. */
216 uint16_t num_queue_pairs;
217 uint16_t max_pkt_len; /* Maximum packet length */
218 uint16_t mac_num; /* Number of MAC addresses */
219 bool promisc_unicast_enabled;
220 bool promisc_multicast_enabled;
222 struct virtchnl_version_info virtchnl_version;
223 struct virtchnl_vf_resource *vf_res; /* VF resource */
224 struct virtchnl_vsi_resource *vsi_res; /* LAN VSI */
225 struct virtchnl_vlan_caps vlan_v2_caps;
226 uint64_t supported_rxdid;
227 uint8_t *proto_xtr; /* proto xtr type for all queues */
228 volatile enum virtchnl_ops pend_cmd; /* pending command not finished */
229 uint32_t pend_cmd_count;
230 int cmd_retval; /* return value of the cmd response from PF */
231 uint8_t *aq_resp; /* buffer to store the adminq response from PF */
233 /** iAVF watchdog enable */
234 bool watchdog_enabled;
241 /* Multicast addrs */
242 struct rte_ether_addr mc_addrs[IAVF_NUM_MACADDR_MAX];
243 uint16_t mc_addrs_num; /* Multicast mac addresses number */
246 bool vf_reset; /* true for VF reset pending, false for no VF reset */
252 uint16_t nb_msix; /* number of MSI-X interrupts on Rx */
253 uint16_t msix_base; /* msix vector base from */
254 uint16_t max_rss_qregion; /* max RSS queue region supported by PF */
255 struct iavf_qv_map *qv_map; /* queue vector mapping */
256 struct iavf_flow_list flow_list;
257 rte_spinlock_t flow_ops_lock;
258 struct iavf_parser_list rss_parser_list;
259 struct iavf_parser_list dist_parser_list;
260 struct iavf_parser_list ipsec_crypto_parser_list;
262 struct iavf_fdir_info fdir; /* flow director info */
263 /* indicate large VF support enabled or not */
266 struct virtchnl_qos_cap_list *qos_cap;
267 struct iavf_qtc_map *qtc_map;
268 struct iavf_tm_conf tm_conf;
270 struct rte_eth_dev *eth_dev;
273 rte_spinlock_t phc_time_aq_lock;
276 #define IAVF_MAX_PKT_TYPE 1024
278 #define IAVF_MAX_QUEUE_NUM 2048
280 enum iavf_proto_xtr_type {
285 IAVF_PROTO_XTR_IPV6_FLOW,
287 IAVF_PROTO_XTR_IP_OFFSET,
288 IAVF_PROTO_XTR_IPSEC_CRYPTO_SAID,
293 * Cache devargs parse result.
295 struct iavf_devargs {
296 uint8_t proto_xtr_dflt;
297 uint8_t proto_xtr[IAVF_MAX_QUEUE_NUM];
298 uint16_t quanta_size;
301 struct iavf_security_ctx;
303 /* Structure to store private data for each VF instance. */
304 struct iavf_adapter {
306 struct rte_eth_dev_data *dev_data;
308 struct iavf_security_ctx *security_ctx;
310 bool rx_bulk_alloc_allowed;
314 uint32_t ptype_tbl[IAVF_MAX_PKT_TYPE] __rte_cache_min_aligned;
317 uint16_t fdir_ref_cnt;
318 struct iavf_devargs devargs;
321 /* IAVF_DEV_PRIVATE_TO */
322 #define IAVF_DEV_PRIVATE_TO_ADAPTER(adapter) \
323 ((struct iavf_adapter *)adapter)
324 #define IAVF_DEV_PRIVATE_TO_VF(adapter) \
325 (&((struct iavf_adapter *)adapter)->vf)
326 #define IAVF_DEV_PRIVATE_TO_HW(adapter) \
327 (&((struct iavf_adapter *)adapter)->hw)
328 #define IAVF_DEV_PRIVATE_TO_IAVF_SECURITY_CTX(adapter) \
329 (((struct iavf_adapter *)adapter)->security_ctx)
332 #define IAVF_VSI_TO_HW(vsi) \
333 (&(((struct iavf_vsi *)vsi)->adapter->hw))
334 #define IAVF_VSI_TO_VF(vsi) \
335 (&(((struct iavf_vsi *)vsi)->adapter->vf))
338 iavf_init_adminq_parameter(struct iavf_hw *hw)
340 hw->aq.num_arq_entries = IAVF_AQ_LEN;
341 hw->aq.num_asq_entries = IAVF_AQ_LEN;
342 hw->aq.arq_buf_size = IAVF_AQ_BUF_SZ;
343 hw->aq.asq_buf_size = IAVF_AQ_BUF_SZ;
346 static inline uint16_t
347 iavf_calc_itr_interval(int16_t interval)
349 if (interval < 0 || interval > IAVF_QUEUE_ITR_INTERVAL_MAX)
350 interval = IAVF_QUEUE_ITR_INTERVAL_DEFAULT;
352 /* Convert to hardware count, as writing each 1 represents 2 us */
356 /* structure used for sending and checking response of virtchnl ops */
357 struct iavf_cmd_info {
358 enum virtchnl_ops ops;
359 uint8_t *in_args; /* buffer for sending */
360 uint32_t in_args_size; /* buffer size for sending */
361 uint8_t *out_buffer; /* buffer for response */
362 uint32_t out_size; /* buffer size for response */
365 /* notify current command done. Only call in case execute
366 * _atomic_set_cmd successfully.
369 _notify_cmd(struct iavf_info *vf, int msg_ret)
371 vf->cmd_retval = msg_ret;
373 vf->pend_cmd = VIRTCHNL_OP_UNKNOWN;
376 /* clear current command. Only call in case execute
377 * _atomic_set_cmd successfully.
380 _clear_cmd(struct iavf_info *vf)
383 vf->pend_cmd = VIRTCHNL_OP_UNKNOWN;
384 vf->cmd_retval = VIRTCHNL_STATUS_SUCCESS;
387 /* Check there is pending cmd in execution. If none, set new command. */
389 _atomic_set_cmd(struct iavf_info *vf, enum virtchnl_ops ops)
391 enum virtchnl_ops op_unk = VIRTCHNL_OP_UNKNOWN;
392 int ret = __atomic_compare_exchange(&vf->pend_cmd, &op_unk, &ops,
393 0, __ATOMIC_ACQUIRE, __ATOMIC_ACQUIRE);
396 PMD_DRV_LOG(ERR, "There is incomplete cmd %d", vf->pend_cmd);
398 __atomic_store_n(&vf->pend_cmd_count, 1, __ATOMIC_RELAXED);
403 /* Check there is pending cmd in execution. If none, set new command. */
405 _atomic_set_async_response_cmd(struct iavf_info *vf, enum virtchnl_ops ops)
407 enum virtchnl_ops op_unk = VIRTCHNL_OP_UNKNOWN;
408 int ret = __atomic_compare_exchange(&vf->pend_cmd, &op_unk, &ops,
409 0, __ATOMIC_ACQUIRE, __ATOMIC_ACQUIRE);
412 PMD_DRV_LOG(ERR, "There is incomplete cmd %d", vf->pend_cmd);
414 __atomic_store_n(&vf->pend_cmd_count, 2, __ATOMIC_RELAXED);
418 int iavf_check_api_version(struct iavf_adapter *adapter);
419 int iavf_get_vf_resource(struct iavf_adapter *adapter);
420 void iavf_handle_virtchnl_msg(struct rte_eth_dev *dev);
421 int iavf_enable_vlan_strip(struct iavf_adapter *adapter);
422 int iavf_disable_vlan_strip(struct iavf_adapter *adapter);
423 int iavf_switch_queue(struct iavf_adapter *adapter, uint16_t qid,
425 int iavf_switch_queue_lv(struct iavf_adapter *adapter, uint16_t qid,
427 int iavf_enable_queues(struct iavf_adapter *adapter);
428 int iavf_enable_queues_lv(struct iavf_adapter *adapter);
429 int iavf_disable_queues(struct iavf_adapter *adapter);
430 int iavf_disable_queues_lv(struct iavf_adapter *adapter);
431 int iavf_configure_rss_lut(struct iavf_adapter *adapter);
432 int iavf_configure_rss_key(struct iavf_adapter *adapter);
433 int iavf_configure_queues(struct iavf_adapter *adapter,
434 uint16_t num_queue_pairs, uint16_t index);
435 int iavf_get_supported_rxdid(struct iavf_adapter *adapter);
436 int iavf_config_vlan_strip_v2(struct iavf_adapter *adapter, bool enable);
437 int iavf_config_vlan_insert_v2(struct iavf_adapter *adapter, bool enable);
438 int iavf_add_del_vlan_v2(struct iavf_adapter *adapter, uint16_t vlanid,
440 int iavf_get_vlan_offload_caps_v2(struct iavf_adapter *adapter);
441 int iavf_config_irq_map(struct iavf_adapter *adapter);
442 int iavf_config_irq_map_lv(struct iavf_adapter *adapter, uint16_t num,
444 void iavf_add_del_all_mac_addr(struct iavf_adapter *adapter, bool add);
445 int iavf_dev_link_update(struct rte_eth_dev *dev,
446 __rte_unused int wait_to_complete);
447 void iavf_dev_alarm_handler(void *param);
448 int iavf_query_stats(struct iavf_adapter *adapter,
449 struct virtchnl_eth_stats **pstats);
450 int iavf_config_promisc(struct iavf_adapter *adapter, bool enable_unicast,
451 bool enable_multicast);
452 int iavf_add_del_eth_addr(struct iavf_adapter *adapter,
453 struct rte_ether_addr *addr, bool add, uint8_t type);
454 int iavf_add_del_vlan(struct iavf_adapter *adapter, uint16_t vlanid, bool add);
455 int iavf_fdir_add(struct iavf_adapter *adapter, struct iavf_fdir_conf *filter);
456 int iavf_fdir_del(struct iavf_adapter *adapter, struct iavf_fdir_conf *filter);
457 int iavf_fdir_check(struct iavf_adapter *adapter,
458 struct iavf_fdir_conf *filter);
459 int iavf_add_del_rss_cfg(struct iavf_adapter *adapter,
460 struct virtchnl_rss_cfg *rss_cfg, bool add);
461 int iavf_get_hena_caps(struct iavf_adapter *adapter, uint64_t *caps);
462 int iavf_set_hena(struct iavf_adapter *adapter, uint64_t hena);
463 int iavf_rss_hash_set(struct iavf_adapter *ad, uint64_t rss_hf, bool add);
464 int iavf_add_del_mc_addr_list(struct iavf_adapter *adapter,
465 struct rte_ether_addr *mc_addrs,
466 uint32_t mc_addrs_num, bool add);
467 int iavf_request_queues(struct rte_eth_dev *dev, uint16_t num);
468 int iavf_get_max_rss_queue_region(struct iavf_adapter *adapter);
469 int iavf_get_qos_cap(struct iavf_adapter *adapter);
470 int iavf_set_q_bw(struct rte_eth_dev *dev,
471 struct virtchnl_queues_bw_cfg *q_bw, uint16_t size);
472 int iavf_set_q_tc_map(struct rte_eth_dev *dev,
473 struct virtchnl_queue_tc_mapping *q_tc_mapping,
475 int iavf_set_vf_quanta_size(struct iavf_adapter *adapter, u16 start_queue_id,
477 void iavf_tm_conf_init(struct rte_eth_dev *dev);
478 void iavf_tm_conf_uninit(struct rte_eth_dev *dev);
479 int iavf_ipsec_crypto_request(struct iavf_adapter *adapter,
480 uint8_t *msg, size_t msg_len,
481 uint8_t *resp_msg, size_t resp_msg_len);
482 extern const struct rte_tm_ops iavf_tm_ops;
483 int iavf_get_ptp_cap(struct iavf_adapter *adapter);
484 int iavf_get_phc_time(struct iavf_rx_queue *rxq);
485 #endif /* _IAVF_ETHDEV_H_ */