rte_atomic32_t rndis_req_id;
uint8_t rndis_resp[256];
+ uint32_t rss_hash;
+ uint8_t rss_key[40];
+ uint16_t rss_ind[128];
+
struct rte_ether_addr mac_addr;
struct rte_eth_dev_owner owner;
return &rte_eth_devices[vf_port];
}
-void hn_vf_info_get(struct hn_data *hv,
+int hn_vf_info_get(struct hn_data *hv,
struct rte_eth_dev_info *info);
int hn_vf_add(struct rte_eth_dev *dev, struct hn_data *hv);
int hn_vf_configure(struct rte_eth_dev *dev,
void hn_vf_stop(struct rte_eth_dev *dev);
void hn_vf_close(struct rte_eth_dev *dev);
-void hn_vf_allmulticast_enable(struct rte_eth_dev *dev);
-void hn_vf_allmulticast_disable(struct rte_eth_dev *dev);
-void hn_vf_promiscuous_enable(struct rte_eth_dev *dev);
-void hn_vf_promiscuous_disable(struct rte_eth_dev *dev);
+int hn_vf_allmulticast_enable(struct rte_eth_dev *dev);
+int hn_vf_allmulticast_disable(struct rte_eth_dev *dev);
+int hn_vf_promiscuous_enable(struct rte_eth_dev *dev);
+int hn_vf_promiscuous_disable(struct rte_eth_dev *dev);
int hn_vf_mc_addr_list(struct rte_eth_dev *dev,
struct rte_ether_addr *mc_addr_set,
uint32_t nb_mc_addr);
void hn_vf_rx_queue_release(struct hn_data *hv, uint16_t queue_id);
int hn_vf_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats);
-void hn_vf_stats_reset(struct rte_eth_dev *dev);
+int hn_vf_stats_reset(struct rte_eth_dev *dev);
int hn_vf_xstats_get_names(struct rte_eth_dev *dev,
struct rte_eth_xstat_name *xstats_names,
unsigned int size);
int hn_vf_xstats_get(struct rte_eth_dev *dev,
struct rte_eth_xstat *xstats,
- unsigned int n);
-void hn_vf_xstats_reset(struct rte_eth_dev *dev);
+ unsigned int offset, unsigned int n);
+int hn_vf_xstats_reset(struct rte_eth_dev *dev);
+int hn_vf_rss_hash_update(struct rte_eth_dev *dev,
+ struct rte_eth_rss_conf *rss_conf);
+int hn_vf_reta_hash_update(struct rte_eth_dev *dev,
+ struct rte_eth_rss_reta_entry64 *reta_conf,
+ uint16_t reta_size);