X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;ds=sidebyside;f=drivers%2Fnet%2Fhns3%2Fhns3_rxtx.h;h=62efc854e4371d08bfb60e0818352e9b3cdabae0;hb=0a94d6bc5daf9b474ed05f3a36bf2d9dc2e3df82;hp=63bafc68b6dbf870b1934747e3f3be784454ba42;hpb=dc153889d53d146b08bb8bb6413cac4ea4d0585d;p=dpdk.git diff --git a/drivers/net/hns3/hns3_rxtx.h b/drivers/net/hns3/hns3_rxtx.h index 63bafc68b6..62efc854e4 100644 --- a/drivers/net/hns3/hns3_rxtx.h +++ b/drivers/net/hns3/hns3_rxtx.h @@ -318,7 +318,7 @@ struct hns3_rx_queue { * should not be transitted to the upper-layer application. For hardware * network engine whose vlan mode is HNS3_HW_SHIFT_AND_DISCARD_MODE, * such as kunpeng 930, PVID will not be reported to the BDs. So, PMD - * driver does not need to perform PVID-related operation in Rx. At this + * does not need to perform PVID-related operation in Rx. At this * point, the pvid_sw_discard_en will be false. */ uint8_t pvid_sw_discard_en:1; @@ -344,12 +344,11 @@ struct hns3_rx_queue { struct rte_mbuf fake_mbuf; /* fake mbuf used with vector rx */ - /* * The following fields are not accessed in the I/O path, so they are * placed at the end. */ - void *io_base; + void *io_base __rte_cache_aligned; struct hns3_adapter *hns; uint64_t rx_ring_phys_addr; /* RX ring DMA address */ const struct rte_memzone *mz; @@ -490,11 +489,13 @@ struct hns3_tx_queue { * PVID will overwrite the outer VLAN field of Tx BD. For the hardware * network engine whose vlan mode is HNS3_HW_SHIFT_AND_DISCARD_MODE, * such as kunpeng 930, if the PVID is set, the hardware will shift the - * VLAN field automatically. So, PMD driver does not need to do + * VLAN field automatically. So, PMD does not need to do * PVID-related operations in Tx. And pvid_sw_shift_en will be false at * this point. */ uint16_t pvid_sw_shift_en:1; + /* check whether the mbuf fast free offload is enabled */ + uint16_t mbuf_fast_free_en:1; /* * For better performance in tx datapath, releasing mbuf in batches is @@ -516,12 +517,11 @@ struct hns3_tx_queue { struct hns3_tx_basic_stats basic_stats; struct hns3_tx_dfx_stats dfx_stats; - /* * The following fields are not accessed in the I/O path, so they are * placed at the end. */ - void *io_base; + void *io_base __rte_cache_aligned; struct hns3_adapter *hns; uint64_t tx_ring_phys_addr; /* TX ring DMA address */ const struct rte_memzone *mz; @@ -609,7 +609,7 @@ hns3_handle_bdinfo(struct hns3_rx_queue *rxq, struct rte_mbuf *rxm, /* * If packet len bigger than mtu when recv with no-scattered algorithm, - * the first n bd will without FE bit, we need process this sisution. + * the first n bd will without FE bit, we need process this situation. * Note: we don't need add statistic counter because latest BD which * with FE bit will mark HNS3_RXD_L2E_B bit. */ @@ -729,9 +729,6 @@ void hns3_init_rx_ptype_tble(struct rte_eth_dev *dev); void hns3_set_rxtx_function(struct rte_eth_dev *eth_dev); eth_tx_burst_t hns3_get_tx_function(struct rte_eth_dev *dev, eth_tx_prep_t *prep); -uint16_t hns3_dummy_rxtx_burst(void *dpdk_txq __rte_unused, - struct rte_mbuf **pkts __rte_unused, - uint16_t pkts_n __rte_unused); uint32_t hns3_get_tqp_intr_reg_offset(uint16_t tqp_intr_id); void hns3_set_queue_intr_gl(struct hns3_hw *hw, uint16_t queue_id,