X-Git-Url: http://git.droids-corp.org/?a=blobdiff_plain;f=drivers%2Fnet%2Fi40e%2Fi40e_pf.c;h=9804ed425335a9f60e4cbafcb05d92ad5c5b9e6f;hb=da573c0e4205d818cd602eaa27c720896f3b6f1c;hp=91be45027ac53165419c11cd45d8fefc2c26af8a;hpb=b89f6deb867f61c8b486a8db79a34cd19a9d2a66;p=dpdk.git diff --git a/drivers/net/i40e/i40e_pf.c b/drivers/net/i40e/i40e_pf.c index 91be45027a..9804ed4253 100644 --- a/drivers/net/i40e/i40e_pf.c +++ b/drivers/net/i40e/i40e_pf.c @@ -14,7 +14,7 @@ #include #include #include -#include +#include #include #include @@ -333,6 +333,10 @@ i40e_pf_host_process_cmd_get_vf_resource(struct i40e_pf_vf *vf, uint8_t *msg, vf_res->vf_cap_flags = vf->request_caps & I40E_VIRTCHNL_OFFLOAD_CAPS; + + if (vf->request_caps & VIRTCHNL_VF_OFFLOAD_REQ_QUEUES) + vf_res->vf_cap_flags |= VIRTCHNL_VF_OFFLOAD_REQ_QUEUES; + /* For X722, it supports write back on ITR * without binding queue to interrupt vector. */ @@ -348,8 +352,8 @@ i40e_pf_host_process_cmd_get_vf_resource(struct i40e_pf_vf *vf, uint8_t *msg, vf_res->vsi_res[0].vsi_type = VIRTCHNL_VSI_SRIOV; vf_res->vsi_res[0].vsi_id = vf->vsi->vsi_id; vf_res->vsi_res[0].num_queue_pairs = vf->vsi->nb_qps; - ether_addr_copy(&vf->mac_addr, - (struct ether_addr *)vf_res->vsi_res[0].default_mac_addr); + rte_ether_addr_copy(&vf->mac_addr, + (struct rte_ether_addr *)vf_res->vsi_res[0].default_mac_addr); send_msg: i40e_pf_host_send_msg_to_vf(vf, VIRTCHNL_OP_GET_VF_RESOURCES, @@ -823,7 +827,7 @@ i40e_pf_host_process_cmd_add_ether_address(struct i40e_pf_vf *vf, (struct virtchnl_ether_addr_list *)msg; struct i40e_mac_filter_info filter; int i; - struct ether_addr *mac; + struct rte_ether_addr *mac; if (!b_op) { i40e_pf_host_send_msg_to_vf( @@ -842,10 +846,10 @@ i40e_pf_host_process_cmd_add_ether_address(struct i40e_pf_vf *vf, } for (i = 0; i < addr_list->num_elements; i++) { - mac = (struct ether_addr *)(addr_list->list[i].addr); - rte_memcpy(&filter.mac_addr, mac, ETHER_ADDR_LEN); - filter.filter_type = RTE_MACVLAN_PERFECT_MATCH; - if (is_zero_ether_addr(mac) || + mac = (struct rte_ether_addr *)(addr_list->list[i].addr); + rte_memcpy(&filter.mac_addr, mac, RTE_ETHER_ADDR_LEN); + filter.filter_type = I40E_MACVLAN_PERFECT_MATCH; + if (rte_is_zero_ether_addr(mac) || i40e_vsi_add_mac(vf->vsi, &filter)) { ret = I40E_ERR_INVALID_MAC_ADDR; goto send_msg; @@ -869,7 +873,7 @@ i40e_pf_host_process_cmd_del_ether_address(struct i40e_pf_vf *vf, struct virtchnl_ether_addr_list *addr_list = (struct virtchnl_ether_addr_list *)msg; int i; - struct ether_addr *mac; + struct rte_ether_addr *mac; if (!b_op) { i40e_pf_host_send_msg_to_vf( @@ -886,8 +890,8 @@ i40e_pf_host_process_cmd_del_ether_address(struct i40e_pf_vf *vf, } for (i = 0; i < addr_list->num_elements; i++) { - mac = (struct ether_addr *)(addr_list->list[i].addr); - if(is_zero_ether_addr(mac) || + mac = (struct rte_ether_addr *)(addr_list->list[i].addr); + if (rte_is_zero_ether_addr(mac) || i40e_vsi_delete_mac(vf->vsi, mac)) { ret = I40E_ERR_INVALID_MAC_ADDR; goto send_msg; @@ -1297,6 +1301,7 @@ i40e_pf_host_handle_vf_msg(struct rte_eth_dev *dev, /* AdminQ will pass absolute VF id, transfer to internal vf id */ uint16_t vf_id = abs_vf_id - hw->func_caps.vf_base_id; struct rte_pmd_i40e_mb_event_param ret_param; + uint64_t first_cycle, cur_cycle; bool b_op = TRUE; int ret; @@ -1306,11 +1311,18 @@ i40e_pf_host_handle_vf_msg(struct rte_eth_dev *dev, } vf = &pf->vfs[vf_id]; + + cur_cycle = rte_get_timer_cycles(); + + /* if the VF being blocked, ignore the message and return */ + if (cur_cycle < vf->ignore_end_cycle) + return; + if (!vf->vsi) { PMD_DRV_LOG(ERR, "NO VSI associated with VF found"); i40e_pf_host_send_msg_to_vf(vf, opcode, I40E_ERR_NO_AVAILABLE_VSI, NULL, 0); - return; + goto check; } /* perform basic checks on the msg */ @@ -1334,7 +1346,7 @@ i40e_pf_host_handle_vf_msg(struct rte_eth_dev *dev, vf_id, opcode, msglen); i40e_pf_host_send_msg_to_vf(vf, opcode, I40E_ERR_PARAM, NULL, 0); - return; + goto check; } /** @@ -1355,7 +1367,7 @@ i40e_pf_host_handle_vf_msg(struct rte_eth_dev *dev, * do nothing and send not_supported to VF. As PF must send a response * to VF and ACK/NACK is not defined. */ - _rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_VF_MBOX, &ret_param); + rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_VF_MBOX, &ret_param); if (ret_param.retval != RTE_PMD_I40E_MB_EVENT_PROCEED) { PMD_DRV_LOG(WARNING, "VF to PF message(%d) is not permitted!", opcode); @@ -1456,6 +1468,37 @@ i40e_pf_host_handle_vf_msg(struct rte_eth_dev *dev, NULL, 0); break; } + +check: + /* if message validation not enabled */ + if (!pf->vf_msg_cfg.max_msg) + return; + + /* store current cycle */ + vf->msg_timestamps[vf->msg_index++] = cur_cycle; + vf->msg_index %= pf->vf_msg_cfg.max_msg; + + /* read the timestamp of earliest message */ + first_cycle = vf->msg_timestamps[vf->msg_index]; + + /* + * If the time span from the arrival time of first message to + * the arrival time of current message smaller than `period`, + * that mean too much message in this statistic period. + */ + if (first_cycle && cur_cycle < first_cycle + + (uint64_t)pf->vf_msg_cfg.period * rte_get_timer_hz()) { + PMD_DRV_LOG(WARNING, "VF %u too much messages(%u in %u" + " seconds),\n\tany new message from which" + " will be ignored during next %u seconds!", + vf_id, pf->vf_msg_cfg.max_msg, + (uint32_t)((cur_cycle - first_cycle + + rte_get_timer_hz() - 1) / rte_get_timer_hz()), + pf->vf_msg_cfg.ignore_second); + vf->ignore_end_cycle = rte_get_timer_cycles() + + pf->vf_msg_cfg.ignore_second * + rte_get_timer_hz(); + } } int @@ -1463,6 +1506,7 @@ i40e_pf_host_init(struct rte_eth_dev *dev) { struct i40e_pf *pf = I40E_DEV_PRIVATE_TO_PF(dev->data->dev_private); struct i40e_hw *hw = I40E_PF_TO_HW(pf); + size_t size; int ret, i; uint32_t val; @@ -1489,10 +1533,24 @@ i40e_pf_host_init(struct rte_eth_dev *dev) I40E_WRITE_REG(hw, I40E_PFGEN_PORTMDIO_NUM, val); I40E_WRITE_FLUSH(hw); + /* calculate the memory size for storing timestamp of messages */ + size = pf->vf_msg_cfg.max_msg * sizeof(uint64_t); + for (i = 0; i < pf->vf_num; i++) { pf->vfs[i].pf = pf; pf->vfs[i].state = I40E_VF_INACTIVE; pf->vfs[i].vf_idx = i; + + if (size) { + /* allocate memory for store timestamp of messages */ + pf->vfs[i].msg_timestamps = + rte_zmalloc("i40e_pf_vf", size, 0); + if (pf->vfs[i].msg_timestamps == NULL) { + ret = -ENOMEM; + goto fail; + } + } + ret = i40e_pf_host_vf_reset(&pf->vfs[i], 0); if (ret != I40E_SUCCESS) goto fail; @@ -1505,6 +1563,8 @@ i40e_pf_host_init(struct rte_eth_dev *dev) return I40E_SUCCESS; fail: + for (; i >= 0; i--) + rte_free(pf->vfs[i].msg_timestamps); rte_free(pf->vfs); i40e_pf_enable_irq0(hw); @@ -1517,6 +1577,7 @@ i40e_pf_host_uninit(struct rte_eth_dev *dev) struct i40e_pf *pf = I40E_DEV_PRIVATE_TO_PF(dev->data->dev_private); struct i40e_hw *hw = I40E_PF_TO_HW(pf); uint32_t val; + int i; PMD_INIT_FUNC_TRACE(); @@ -1529,6 +1590,10 @@ i40e_pf_host_uninit(struct rte_eth_dev *dev) (pf->vf_nb_qps == 0)) return I40E_SUCCESS; + /* free memory for store timestamp of messages */ + for (i = 0; i < pf->vf_num; i++) + rte_free(pf->vfs[i].msg_timestamps); + /* free memory to store VF structure */ rte_free(pf->vfs); pf->vfs = NULL;