ethdev: add namespace
[dpdk.git] / drivers / net / ice / ice_dcf.c
1 /* SPDX-License-Identifier: BSD-3-Clause
2  * Copyright(c) 2020 Intel Corporation
3  */
4
5 #include <sys/queue.h>
6 #include <stdio.h>
7 #include <errno.h>
8 #include <stdint.h>
9 #include <string.h>
10 #include <unistd.h>
11 #include <stdarg.h>
12 #include <inttypes.h>
13 #include <rte_byteorder.h>
14 #include <rte_common.h>
15
16 #include <rte_pci.h>
17 #include <rte_atomic.h>
18 #include <rte_eal.h>
19 #include <rte_ether.h>
20 #include <ethdev_driver.h>
21 #include <ethdev_pci.h>
22 #include <rte_malloc.h>
23 #include <rte_memzone.h>
24 #include <rte_dev.h>
25
26 #include "ice_dcf.h"
27 #include "ice_rxtx.h"
28
29 #define ICE_DCF_AQ_LEN     32
30 #define ICE_DCF_AQ_BUF_SZ  4096
31
32 #define ICE_DCF_ARQ_MAX_RETRIES 200
33 #define ICE_DCF_ARQ_CHECK_TIME  2   /* msecs */
34
35 #define ICE_DCF_VF_RES_BUF_SZ   \
36         (sizeof(struct virtchnl_vf_resource) +  \
37                 IAVF_MAX_VF_VSI * sizeof(struct virtchnl_vsi_resource))
38
39 static __rte_always_inline int
40 ice_dcf_send_cmd_req_no_irq(struct ice_dcf_hw *hw, enum virtchnl_ops op,
41                             uint8_t *req_msg, uint16_t req_msglen)
42 {
43         return iavf_aq_send_msg_to_pf(&hw->avf, op, IAVF_SUCCESS,
44                                       req_msg, req_msglen, NULL);
45 }
46
47 static int
48 ice_dcf_recv_cmd_rsp_no_irq(struct ice_dcf_hw *hw, enum virtchnl_ops op,
49                             uint8_t *rsp_msgbuf, uint16_t rsp_buflen,
50                             uint16_t *rsp_msglen)
51 {
52         struct iavf_arq_event_info event;
53         enum virtchnl_ops v_op;
54         int i = 0;
55         int err;
56
57         event.buf_len = rsp_buflen;
58         event.msg_buf = rsp_msgbuf;
59
60         do {
61                 err = iavf_clean_arq_element(&hw->avf, &event, NULL);
62                 if (err != IAVF_SUCCESS)
63                         goto again;
64
65                 v_op = rte_le_to_cpu_32(event.desc.cookie_high);
66                 if (v_op != op)
67                         goto again;
68
69                 if (rsp_msglen != NULL)
70                         *rsp_msglen = event.msg_len;
71                 return rte_le_to_cpu_32(event.desc.cookie_low);
72
73 again:
74                 rte_delay_ms(ICE_DCF_ARQ_CHECK_TIME);
75         } while (i++ < ICE_DCF_ARQ_MAX_RETRIES);
76
77         return -EIO;
78 }
79
80 static __rte_always_inline void
81 ice_dcf_aq_cmd_clear(struct ice_dcf_hw *hw, struct dcf_virtchnl_cmd *cmd)
82 {
83         rte_spinlock_lock(&hw->vc_cmd_queue_lock);
84
85         TAILQ_REMOVE(&hw->vc_cmd_queue, cmd, next);
86
87         rte_spinlock_unlock(&hw->vc_cmd_queue_lock);
88 }
89
90 static __rte_always_inline void
91 ice_dcf_vc_cmd_set(struct ice_dcf_hw *hw, struct dcf_virtchnl_cmd *cmd)
92 {
93         cmd->v_ret = IAVF_ERR_NOT_READY;
94         cmd->rsp_msglen = 0;
95         cmd->pending = 1;
96
97         rte_spinlock_lock(&hw->vc_cmd_queue_lock);
98
99         TAILQ_INSERT_TAIL(&hw->vc_cmd_queue, cmd, next);
100
101         rte_spinlock_unlock(&hw->vc_cmd_queue_lock);
102 }
103
104 static __rte_always_inline int
105 ice_dcf_vc_cmd_send(struct ice_dcf_hw *hw, struct dcf_virtchnl_cmd *cmd)
106 {
107         return iavf_aq_send_msg_to_pf(&hw->avf,
108                                       cmd->v_op, IAVF_SUCCESS,
109                                       cmd->req_msg, cmd->req_msglen, NULL);
110 }
111
112 static __rte_always_inline void
113 ice_dcf_aq_cmd_handle(struct ice_dcf_hw *hw, struct iavf_arq_event_info *info)
114 {
115         struct dcf_virtchnl_cmd *cmd;
116         enum virtchnl_ops v_op;
117         enum iavf_status v_ret;
118         uint16_t aq_op;
119
120         aq_op = rte_le_to_cpu_16(info->desc.opcode);
121         if (unlikely(aq_op != iavf_aqc_opc_send_msg_to_vf)) {
122                 PMD_DRV_LOG(ERR,
123                             "Request %u is not supported yet", aq_op);
124                 return;
125         }
126
127         v_op = rte_le_to_cpu_32(info->desc.cookie_high);
128         if (v_op == VIRTCHNL_OP_EVENT) {
129                 if (hw->vc_event_msg_cb != NULL)
130                         hw->vc_event_msg_cb(hw,
131                                             info->msg_buf,
132                                             info->msg_len);
133                 return;
134         }
135
136         v_ret = rte_le_to_cpu_32(info->desc.cookie_low);
137
138         rte_spinlock_lock(&hw->vc_cmd_queue_lock);
139
140         TAILQ_FOREACH(cmd, &hw->vc_cmd_queue, next) {
141                 if (cmd->v_op == v_op && cmd->pending) {
142                         cmd->v_ret = v_ret;
143                         cmd->rsp_msglen = RTE_MIN(info->msg_len,
144                                                   cmd->rsp_buflen);
145                         if (likely(cmd->rsp_msglen != 0))
146                                 rte_memcpy(cmd->rsp_msgbuf, info->msg_buf,
147                                            cmd->rsp_msglen);
148
149                         /* prevent compiler reordering */
150                         rte_compiler_barrier();
151                         cmd->pending = 0;
152                         break;
153                 }
154         }
155
156         rte_spinlock_unlock(&hw->vc_cmd_queue_lock);
157 }
158
159 static void
160 ice_dcf_handle_virtchnl_msg(struct ice_dcf_hw *hw)
161 {
162         struct iavf_arq_event_info info;
163         uint16_t pending = 1;
164         int ret;
165
166         info.buf_len = ICE_DCF_AQ_BUF_SZ;
167         info.msg_buf = hw->arq_buf;
168
169         while (pending) {
170                 ret = iavf_clean_arq_element(&hw->avf, &info, &pending);
171                 if (ret != IAVF_SUCCESS)
172                         break;
173
174                 ice_dcf_aq_cmd_handle(hw, &info);
175         }
176 }
177
178 static int
179 ice_dcf_init_check_api_version(struct ice_dcf_hw *hw)
180 {
181 #define ICE_CPF_VIRTCHNL_VERSION_MAJOR_START    1
182 #define ICE_CPF_VIRTCHNL_VERSION_MINOR_START    1
183         struct virtchnl_version_info version, *pver;
184         int err;
185
186         version.major = VIRTCHNL_VERSION_MAJOR;
187         version.minor = VIRTCHNL_VERSION_MINOR;
188         err = ice_dcf_send_cmd_req_no_irq(hw, VIRTCHNL_OP_VERSION,
189                                           (uint8_t *)&version, sizeof(version));
190         if (err) {
191                 PMD_INIT_LOG(ERR, "Failed to send OP_VERSION");
192                 return err;
193         }
194
195         pver = &hw->virtchnl_version;
196         err = ice_dcf_recv_cmd_rsp_no_irq(hw, VIRTCHNL_OP_VERSION,
197                                           (uint8_t *)pver, sizeof(*pver), NULL);
198         if (err) {
199                 PMD_INIT_LOG(ERR, "Failed to get response of OP_VERSION");
200                 return -1;
201         }
202
203         PMD_INIT_LOG(DEBUG,
204                      "Peer PF API version: %u.%u", pver->major, pver->minor);
205
206         if (pver->major < ICE_CPF_VIRTCHNL_VERSION_MAJOR_START ||
207             (pver->major == ICE_CPF_VIRTCHNL_VERSION_MAJOR_START &&
208              pver->minor < ICE_CPF_VIRTCHNL_VERSION_MINOR_START)) {
209                 PMD_INIT_LOG(ERR,
210                              "VIRTCHNL API version should not be lower than (%u.%u)",
211                              ICE_CPF_VIRTCHNL_VERSION_MAJOR_START,
212                              ICE_CPF_VIRTCHNL_VERSION_MAJOR_START);
213                 return -1;
214         } else if (pver->major > VIRTCHNL_VERSION_MAJOR ||
215                    (pver->major == VIRTCHNL_VERSION_MAJOR &&
216                     pver->minor > VIRTCHNL_VERSION_MINOR)) {
217                 PMD_INIT_LOG(ERR,
218                              "PF/VF API version mismatch:(%u.%u)-(%u.%u)",
219                              pver->major, pver->minor,
220                              VIRTCHNL_VERSION_MAJOR, VIRTCHNL_VERSION_MINOR);
221                 return -1;
222         }
223
224         PMD_INIT_LOG(DEBUG, "Peer is supported PF host");
225
226         return 0;
227 }
228
229 static int
230 ice_dcf_get_vf_resource(struct ice_dcf_hw *hw)
231 {
232         uint32_t caps;
233         int err, i;
234
235         caps = VIRTCHNL_VF_OFFLOAD_WB_ON_ITR | VIRTCHNL_VF_OFFLOAD_RX_POLLING |
236                VIRTCHNL_VF_CAP_ADV_LINK_SPEED | VIRTCHNL_VF_CAP_DCF |
237                VIRTCHNL_VF_OFFLOAD_VLAN_V2 |
238                VF_BASE_MODE_OFFLOADS | VIRTCHNL_VF_OFFLOAD_RX_FLEX_DESC |
239                VIRTCHNL_VF_OFFLOAD_QOS;
240
241         err = ice_dcf_send_cmd_req_no_irq(hw, VIRTCHNL_OP_GET_VF_RESOURCES,
242                                           (uint8_t *)&caps, sizeof(caps));
243         if (err) {
244                 PMD_DRV_LOG(ERR, "Failed to send msg OP_GET_VF_RESOURCE");
245                 return err;
246         }
247
248         err = ice_dcf_recv_cmd_rsp_no_irq(hw, VIRTCHNL_OP_GET_VF_RESOURCES,
249                                           (uint8_t *)hw->vf_res,
250                                           ICE_DCF_VF_RES_BUF_SZ, NULL);
251         if (err) {
252                 PMD_DRV_LOG(ERR, "Failed to get response of OP_GET_VF_RESOURCE");
253                 return -1;
254         }
255
256         iavf_vf_parse_hw_config(&hw->avf, hw->vf_res);
257
258         hw->vsi_res = NULL;
259         for (i = 0; i < hw->vf_res->num_vsis; i++) {
260                 if (hw->vf_res->vsi_res[i].vsi_type == VIRTCHNL_VSI_SRIOV)
261                         hw->vsi_res = &hw->vf_res->vsi_res[i];
262         }
263
264         if (!hw->vsi_res) {
265                 PMD_DRV_LOG(ERR, "no LAN VSI found");
266                 return -1;
267         }
268
269         hw->vsi_id = hw->vsi_res->vsi_id;
270         PMD_DRV_LOG(DEBUG, "VSI ID is %u", hw->vsi_id);
271
272         return 0;
273 }
274
275 static int
276 ice_dcf_get_vf_vsi_map(struct ice_dcf_hw *hw)
277 {
278         struct virtchnl_dcf_vsi_map *vsi_map;
279         uint32_t valid_msg_len;
280         uint16_t len;
281         int err;
282
283         err = ice_dcf_send_cmd_req_no_irq(hw, VIRTCHNL_OP_DCF_GET_VSI_MAP,
284                                           NULL, 0);
285         if (err) {
286                 PMD_DRV_LOG(ERR, "Failed to send msg OP_DCF_GET_VSI_MAP");
287                 return err;
288         }
289
290         err = ice_dcf_recv_cmd_rsp_no_irq(hw, VIRTCHNL_OP_DCF_GET_VSI_MAP,
291                                           hw->arq_buf, ICE_DCF_AQ_BUF_SZ,
292                                           &len);
293         if (err) {
294                 PMD_DRV_LOG(ERR, "Failed to get response of OP_DCF_GET_VSI_MAP");
295                 return err;
296         }
297
298         vsi_map = (struct virtchnl_dcf_vsi_map *)hw->arq_buf;
299         valid_msg_len = (vsi_map->num_vfs - 1) * sizeof(vsi_map->vf_vsi[0]) +
300                         sizeof(*vsi_map);
301         if (len != valid_msg_len) {
302                 PMD_DRV_LOG(ERR, "invalid vf vsi map response with length %u",
303                             len);
304                 return -EINVAL;
305         }
306
307         if (hw->num_vfs != 0 && hw->num_vfs != vsi_map->num_vfs) {
308                 PMD_DRV_LOG(ERR, "The number VSI map (%u) doesn't match the number of VFs (%u)",
309                             vsi_map->num_vfs, hw->num_vfs);
310                 return -EINVAL;
311         }
312
313         len = vsi_map->num_vfs * sizeof(vsi_map->vf_vsi[0]);
314
315         if (!hw->vf_vsi_map) {
316                 hw->vf_vsi_map = rte_zmalloc("vf_vsi_ctx", len, 0);
317                 if (!hw->vf_vsi_map) {
318                         PMD_DRV_LOG(ERR, "Failed to alloc memory for VSI context");
319                         return -ENOMEM;
320                 }
321
322                 hw->num_vfs = vsi_map->num_vfs;
323                 hw->pf_vsi_id = vsi_map->pf_vsi;
324         }
325
326         if (!memcmp(hw->vf_vsi_map, vsi_map->vf_vsi, len)) {
327                 PMD_DRV_LOG(DEBUG, "VF VSI map doesn't change");
328                 return 1;
329         }
330
331         rte_memcpy(hw->vf_vsi_map, vsi_map->vf_vsi, len);
332         return 0;
333 }
334
335 static int
336 ice_dcf_mode_disable(struct ice_dcf_hw *hw)
337 {
338         int err;
339
340         if (hw->resetting)
341                 return 0;
342
343         err = ice_dcf_send_cmd_req_no_irq(hw, VIRTCHNL_OP_DCF_DISABLE,
344                                           NULL, 0);
345         if (err) {
346                 PMD_DRV_LOG(ERR, "Failed to send msg OP_DCF_DISABLE");
347                 return err;
348         }
349
350         err = ice_dcf_recv_cmd_rsp_no_irq(hw, VIRTCHNL_OP_DCF_DISABLE,
351                                           hw->arq_buf, ICE_DCF_AQ_BUF_SZ, NULL);
352         if (err) {
353                 PMD_DRV_LOG(ERR,
354                             "Failed to get response of OP_DCF_DISABLE %d",
355                             err);
356                 return -1;
357         }
358
359         return 0;
360 }
361
362 static int
363 ice_dcf_check_reset_done(struct ice_dcf_hw *hw)
364 {
365 #define ICE_DCF_RESET_WAIT_CNT       50
366         struct iavf_hw *avf = &hw->avf;
367         int i, reset;
368
369         for (i = 0; i < ICE_DCF_RESET_WAIT_CNT; i++) {
370                 reset = IAVF_READ_REG(avf, IAVF_VFGEN_RSTAT) &
371                                         IAVF_VFGEN_RSTAT_VFR_STATE_MASK;
372                 reset = reset >> IAVF_VFGEN_RSTAT_VFR_STATE_SHIFT;
373
374                 if (reset == VIRTCHNL_VFR_VFACTIVE ||
375                     reset == VIRTCHNL_VFR_COMPLETED)
376                         break;
377
378                 rte_delay_ms(20);
379         }
380
381         if (i >= ICE_DCF_RESET_WAIT_CNT)
382                 return -1;
383
384         return 0;
385 }
386
387 static inline void
388 ice_dcf_enable_irq0(struct ice_dcf_hw *hw)
389 {
390         struct iavf_hw *avf = &hw->avf;
391
392         /* Enable admin queue interrupt trigger */
393         IAVF_WRITE_REG(avf, IAVF_VFINT_ICR0_ENA1,
394                        IAVF_VFINT_ICR0_ENA1_ADMINQ_MASK);
395         IAVF_WRITE_REG(avf, IAVF_VFINT_DYN_CTL01,
396                        IAVF_VFINT_DYN_CTL01_INTENA_MASK |
397                        IAVF_VFINT_DYN_CTL01_CLEARPBA_MASK |
398                        IAVF_VFINT_DYN_CTL01_ITR_INDX_MASK);
399
400         IAVF_WRITE_FLUSH(avf);
401 }
402
403 static inline void
404 ice_dcf_disable_irq0(struct ice_dcf_hw *hw)
405 {
406         struct iavf_hw *avf = &hw->avf;
407
408         /* Disable all interrupt types */
409         IAVF_WRITE_REG(avf, IAVF_VFINT_ICR0_ENA1, 0);
410         IAVF_WRITE_REG(avf, IAVF_VFINT_DYN_CTL01,
411                        IAVF_VFINT_DYN_CTL01_ITR_INDX_MASK);
412
413         IAVF_WRITE_FLUSH(avf);
414 }
415
416 static void
417 ice_dcf_dev_interrupt_handler(void *param)
418 {
419         struct ice_dcf_hw *hw = param;
420
421         ice_dcf_disable_irq0(hw);
422
423         ice_dcf_handle_virtchnl_msg(hw);
424
425         ice_dcf_enable_irq0(hw);
426 }
427
428 int
429 ice_dcf_execute_virtchnl_cmd(struct ice_dcf_hw *hw,
430                              struct dcf_virtchnl_cmd *cmd)
431 {
432         int i = 0;
433         int err;
434
435         if ((cmd->req_msg && !cmd->req_msglen) ||
436             (!cmd->req_msg && cmd->req_msglen) ||
437             (cmd->rsp_msgbuf && !cmd->rsp_buflen) ||
438             (!cmd->rsp_msgbuf && cmd->rsp_buflen))
439                 return -EINVAL;
440
441         rte_spinlock_lock(&hw->vc_cmd_send_lock);
442         ice_dcf_vc_cmd_set(hw, cmd);
443
444         err = ice_dcf_vc_cmd_send(hw, cmd);
445         if (err) {
446                 PMD_DRV_LOG(ERR, "fail to send cmd %d", cmd->v_op);
447                 goto ret;
448         }
449
450         do {
451                 if (!cmd->pending)
452                         break;
453
454                 rte_delay_ms(ICE_DCF_ARQ_CHECK_TIME);
455         } while (i++ < ICE_DCF_ARQ_MAX_RETRIES);
456
457         if (cmd->v_ret != IAVF_SUCCESS) {
458                 err = -1;
459                 PMD_DRV_LOG(ERR,
460                             "No response (%d times) or return failure (%d) for cmd %d",
461                             i, cmd->v_ret, cmd->v_op);
462         }
463
464 ret:
465         ice_dcf_aq_cmd_clear(hw, cmd);
466         rte_spinlock_unlock(&hw->vc_cmd_send_lock);
467         return err;
468 }
469
470 int
471 ice_dcf_send_aq_cmd(void *dcf_hw, struct ice_aq_desc *desc,
472                     void *buf, uint16_t buf_size)
473 {
474         struct dcf_virtchnl_cmd desc_cmd, buff_cmd;
475         struct ice_dcf_hw *hw = dcf_hw;
476         int err = 0;
477         int i = 0;
478
479         if ((buf && !buf_size) || (!buf && buf_size) ||
480             buf_size > ICE_DCF_AQ_BUF_SZ)
481                 return -EINVAL;
482
483         desc_cmd.v_op = VIRTCHNL_OP_DCF_CMD_DESC;
484         desc_cmd.req_msglen = sizeof(*desc);
485         desc_cmd.req_msg = (uint8_t *)desc;
486         desc_cmd.rsp_buflen = sizeof(*desc);
487         desc_cmd.rsp_msgbuf = (uint8_t *)desc;
488
489         if (buf == NULL)
490                 return ice_dcf_execute_virtchnl_cmd(hw, &desc_cmd);
491
492         desc->flags |= rte_cpu_to_le_16(ICE_AQ_FLAG_BUF);
493
494         buff_cmd.v_op = VIRTCHNL_OP_DCF_CMD_BUFF;
495         buff_cmd.req_msglen = buf_size;
496         buff_cmd.req_msg = buf;
497         buff_cmd.rsp_buflen = buf_size;
498         buff_cmd.rsp_msgbuf = buf;
499
500         rte_spinlock_lock(&hw->vc_cmd_send_lock);
501         ice_dcf_vc_cmd_set(hw, &desc_cmd);
502         ice_dcf_vc_cmd_set(hw, &buff_cmd);
503
504         if (ice_dcf_vc_cmd_send(hw, &desc_cmd) ||
505             ice_dcf_vc_cmd_send(hw, &buff_cmd)) {
506                 err = -1;
507                 PMD_DRV_LOG(ERR, "fail to send OP_DCF_CMD_DESC/BUFF");
508                 goto ret;
509         }
510
511         do {
512                 if (!desc_cmd.pending && !buff_cmd.pending)
513                         break;
514
515                 rte_delay_ms(ICE_DCF_ARQ_CHECK_TIME);
516         } while (i++ < ICE_DCF_ARQ_MAX_RETRIES);
517
518         if (desc_cmd.v_ret != IAVF_SUCCESS || buff_cmd.v_ret != IAVF_SUCCESS) {
519                 err = -1;
520                 PMD_DRV_LOG(ERR,
521                             "No response (%d times) or return failure (desc: %d / buff: %d)",
522                             i, desc_cmd.v_ret, buff_cmd.v_ret);
523         }
524
525 ret:
526         ice_dcf_aq_cmd_clear(hw, &desc_cmd);
527         ice_dcf_aq_cmd_clear(hw, &buff_cmd);
528         rte_spinlock_unlock(&hw->vc_cmd_send_lock);
529
530         return err;
531 }
532
533 int
534 ice_dcf_handle_vsi_update_event(struct ice_dcf_hw *hw)
535 {
536         struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(hw->eth_dev);
537         int i = 0;
538         int err = -1;
539
540         rte_spinlock_lock(&hw->vc_cmd_send_lock);
541
542         rte_intr_disable(&pci_dev->intr_handle);
543         ice_dcf_disable_irq0(hw);
544
545         for (;;) {
546                 if (ice_dcf_get_vf_resource(hw) == 0 &&
547                     ice_dcf_get_vf_vsi_map(hw) >= 0) {
548                         err = 0;
549                         break;
550                 }
551
552                 if (++i >= ICE_DCF_ARQ_MAX_RETRIES)
553                         break;
554
555                 rte_delay_ms(ICE_DCF_ARQ_CHECK_TIME);
556         }
557
558         rte_intr_enable(&pci_dev->intr_handle);
559         ice_dcf_enable_irq0(hw);
560
561         rte_spinlock_unlock(&hw->vc_cmd_send_lock);
562
563         return err;
564 }
565
566 static int
567 ice_dcf_get_supported_rxdid(struct ice_dcf_hw *hw)
568 {
569         int err;
570
571         err = ice_dcf_send_cmd_req_no_irq(hw,
572                                           VIRTCHNL_OP_GET_SUPPORTED_RXDIDS,
573                                           NULL, 0);
574         if (err) {
575                 PMD_INIT_LOG(ERR, "Failed to send OP_GET_SUPPORTED_RXDIDS");
576                 return -1;
577         }
578
579         err = ice_dcf_recv_cmd_rsp_no_irq(hw, VIRTCHNL_OP_GET_SUPPORTED_RXDIDS,
580                                           (uint8_t *)&hw->supported_rxdid,
581                                           sizeof(uint64_t), NULL);
582         if (err) {
583                 PMD_INIT_LOG(ERR, "Failed to get response of OP_GET_SUPPORTED_RXDIDS");
584                 return -1;
585         }
586
587         return 0;
588 }
589
590 int
591 ice_dcf_init_hw(struct rte_eth_dev *eth_dev, struct ice_dcf_hw *hw)
592 {
593         struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
594         int ret, size;
595
596         hw->avf.hw_addr = pci_dev->mem_resource[0].addr;
597         hw->avf.back = hw;
598
599         hw->avf.bus.bus_id = pci_dev->addr.bus;
600         hw->avf.bus.device = pci_dev->addr.devid;
601         hw->avf.bus.func = pci_dev->addr.function;
602
603         hw->avf.device_id = pci_dev->id.device_id;
604         hw->avf.vendor_id = pci_dev->id.vendor_id;
605         hw->avf.subsystem_device_id = pci_dev->id.subsystem_device_id;
606         hw->avf.subsystem_vendor_id = pci_dev->id.subsystem_vendor_id;
607
608         hw->avf.aq.num_arq_entries = ICE_DCF_AQ_LEN;
609         hw->avf.aq.num_asq_entries = ICE_DCF_AQ_LEN;
610         hw->avf.aq.arq_buf_size = ICE_DCF_AQ_BUF_SZ;
611         hw->avf.aq.asq_buf_size = ICE_DCF_AQ_BUF_SZ;
612
613         rte_spinlock_init(&hw->vc_cmd_send_lock);
614         rte_spinlock_init(&hw->vc_cmd_queue_lock);
615         TAILQ_INIT(&hw->vc_cmd_queue);
616
617         hw->arq_buf = rte_zmalloc("arq_buf", ICE_DCF_AQ_BUF_SZ, 0);
618         if (hw->arq_buf == NULL) {
619                 PMD_INIT_LOG(ERR, "unable to allocate AdminQ buffer memory");
620                 goto err;
621         }
622
623         ret = iavf_set_mac_type(&hw->avf);
624         if (ret) {
625                 PMD_INIT_LOG(ERR, "set_mac_type failed: %d", ret);
626                 goto err;
627         }
628
629         ret = ice_dcf_check_reset_done(hw);
630         if (ret) {
631                 PMD_INIT_LOG(ERR, "VF is still resetting");
632                 goto err;
633         }
634
635         ret = iavf_init_adminq(&hw->avf);
636         if (ret) {
637                 PMD_INIT_LOG(ERR, "init_adminq failed: %d", ret);
638                 goto err;
639         }
640
641         if (ice_dcf_init_check_api_version(hw)) {
642                 PMD_INIT_LOG(ERR, "check_api version failed");
643                 goto err_api;
644         }
645
646         hw->vf_res = rte_zmalloc("vf_res", ICE_DCF_VF_RES_BUF_SZ, 0);
647         if (hw->vf_res == NULL) {
648                 PMD_INIT_LOG(ERR, "unable to allocate vf_res memory");
649                 goto err_api;
650         }
651
652         if (ice_dcf_get_vf_resource(hw)) {
653                 PMD_INIT_LOG(ERR, "Failed to get VF resource");
654                 goto err_alloc;
655         }
656
657         if (ice_dcf_get_vf_vsi_map(hw) < 0) {
658                 PMD_INIT_LOG(ERR, "Failed to get VF VSI map");
659                 ice_dcf_mode_disable(hw);
660                 goto err_alloc;
661         }
662
663         /* Allocate memory for RSS info */
664         if (hw->vf_res->vf_cap_flags & VIRTCHNL_VF_OFFLOAD_RSS_PF) {
665                 hw->rss_key = rte_zmalloc(NULL,
666                                           hw->vf_res->rss_key_size, 0);
667                 if (!hw->rss_key) {
668                         PMD_INIT_LOG(ERR, "unable to allocate rss_key memory");
669                         goto err_alloc;
670                 }
671                 hw->rss_lut = rte_zmalloc("rss_lut",
672                                           hw->vf_res->rss_lut_size, 0);
673                 if (!hw->rss_lut) {
674                         PMD_INIT_LOG(ERR, "unable to allocate rss_lut memory");
675                         goto err_rss;
676                 }
677         }
678
679         if (hw->vf_res->vf_cap_flags & VIRTCHNL_VF_OFFLOAD_RX_FLEX_DESC) {
680                 if (ice_dcf_get_supported_rxdid(hw) != 0) {
681                         PMD_INIT_LOG(ERR, "failed to do get supported rxdid");
682                         goto err_rss;
683                 }
684         }
685
686         if (hw->vf_res->vf_cap_flags & VIRTCHNL_VF_OFFLOAD_QOS) {
687                 ice_dcf_tm_conf_init(eth_dev);
688                 size = sizeof(struct virtchnl_dcf_bw_cfg_list *) * hw->num_vfs;
689                 hw->qos_bw_cfg = rte_zmalloc("qos_bw_cfg", size, 0);
690                 if (!hw->qos_bw_cfg) {
691                         PMD_INIT_LOG(ERR, "no memory for qos_bw_cfg");
692                         goto err_rss;
693                 }
694         }
695
696         hw->eth_dev = eth_dev;
697         rte_intr_callback_register(&pci_dev->intr_handle,
698                                    ice_dcf_dev_interrupt_handler, hw);
699         rte_intr_enable(&pci_dev->intr_handle);
700         ice_dcf_enable_irq0(hw);
701
702         return 0;
703
704 err_rss:
705         rte_free(hw->rss_key);
706         rte_free(hw->rss_lut);
707 err_alloc:
708         rte_free(hw->vf_res);
709 err_api:
710         iavf_shutdown_adminq(&hw->avf);
711 err:
712         rte_free(hw->arq_buf);
713
714         return -1;
715 }
716
717 void
718 ice_dcf_uninit_hw(struct rte_eth_dev *eth_dev, struct ice_dcf_hw *hw)
719 {
720         struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
721         struct rte_intr_handle *intr_handle = &pci_dev->intr_handle;
722
723         if (hw->vf_res->vf_cap_flags & VIRTCHNL_VF_OFFLOAD_QOS)
724                 if (hw->tm_conf.committed) {
725                         ice_dcf_clear_bw(hw);
726                         ice_dcf_tm_conf_uninit(eth_dev);
727                 }
728
729         ice_dcf_disable_irq0(hw);
730         rte_intr_disable(intr_handle);
731         rte_intr_callback_unregister(intr_handle,
732                                      ice_dcf_dev_interrupt_handler, hw);
733
734         ice_dcf_mode_disable(hw);
735         iavf_shutdown_adminq(&hw->avf);
736
737         rte_free(hw->arq_buf);
738         hw->arq_buf = NULL;
739
740         rte_free(hw->vf_vsi_map);
741         hw->vf_vsi_map = NULL;
742
743         rte_free(hw->vf_res);
744         hw->vf_res = NULL;
745
746         rte_free(hw->rss_lut);
747         hw->rss_lut = NULL;
748
749         rte_free(hw->rss_key);
750         hw->rss_key = NULL;
751
752         rte_free(hw->qos_bw_cfg);
753         hw->qos_bw_cfg = NULL;
754
755         rte_free(hw->ets_config);
756         hw->ets_config = NULL;
757 }
758
759 static int
760 ice_dcf_configure_rss_key(struct ice_dcf_hw *hw)
761 {
762         struct virtchnl_rss_key *rss_key;
763         struct dcf_virtchnl_cmd args;
764         int len, err;
765
766         len = sizeof(*rss_key) + hw->vf_res->rss_key_size - 1;
767         rss_key = rte_zmalloc("rss_key", len, 0);
768         if (!rss_key)
769                 return -ENOMEM;
770
771         rss_key->vsi_id = hw->vsi_res->vsi_id;
772         rss_key->key_len = hw->vf_res->rss_key_size;
773         rte_memcpy(rss_key->key, hw->rss_key, hw->vf_res->rss_key_size);
774
775         args.v_op = VIRTCHNL_OP_CONFIG_RSS_KEY;
776         args.req_msglen = len;
777         args.req_msg = (uint8_t *)rss_key;
778         args.rsp_msglen = 0;
779         args.rsp_buflen = 0;
780         args.rsp_msgbuf = NULL;
781         args.pending = 0;
782
783         err = ice_dcf_execute_virtchnl_cmd(hw, &args);
784         if (err)
785                 PMD_INIT_LOG(ERR, "Failed to execute OP_CONFIG_RSS_KEY");
786
787         rte_free(rss_key);
788         return err;
789 }
790
791 static int
792 ice_dcf_configure_rss_lut(struct ice_dcf_hw *hw)
793 {
794         struct virtchnl_rss_lut *rss_lut;
795         struct dcf_virtchnl_cmd args;
796         int len, err;
797
798         len = sizeof(*rss_lut) + hw->vf_res->rss_lut_size - 1;
799         rss_lut = rte_zmalloc("rss_lut", len, 0);
800         if (!rss_lut)
801                 return -ENOMEM;
802
803         rss_lut->vsi_id = hw->vsi_res->vsi_id;
804         rss_lut->lut_entries = hw->vf_res->rss_lut_size;
805         rte_memcpy(rss_lut->lut, hw->rss_lut, hw->vf_res->rss_lut_size);
806
807         args.v_op = VIRTCHNL_OP_CONFIG_RSS_LUT;
808         args.req_msglen = len;
809         args.req_msg = (uint8_t *)rss_lut;
810         args.rsp_msglen = 0;
811         args.rsp_buflen = 0;
812         args.rsp_msgbuf = NULL;
813         args.pending = 0;
814
815         err = ice_dcf_execute_virtchnl_cmd(hw, &args);
816         if (err)
817                 PMD_INIT_LOG(ERR, "Failed to execute OP_CONFIG_RSS_LUT");
818
819         rte_free(rss_lut);
820         return err;
821 }
822
823 int
824 ice_dcf_init_rss(struct ice_dcf_hw *hw)
825 {
826         struct rte_eth_dev *dev = hw->eth_dev;
827         struct rte_eth_rss_conf *rss_conf;
828         uint8_t i, j, nb_q;
829         int ret;
830
831         rss_conf = &dev->data->dev_conf.rx_adv_conf.rss_conf;
832         nb_q = dev->data->nb_rx_queues;
833
834         if (!(hw->vf_res->vf_cap_flags & VIRTCHNL_VF_OFFLOAD_RSS_PF)) {
835                 PMD_DRV_LOG(DEBUG, "RSS is not supported");
836                 return -ENOTSUP;
837         }
838         if (dev->data->dev_conf.rxmode.mq_mode != RTE_ETH_MQ_RX_RSS) {
839                 PMD_DRV_LOG(WARNING, "RSS is enabled by PF by default");
840                 /* set all lut items to default queue */
841                 memset(hw->rss_lut, 0, hw->vf_res->rss_lut_size);
842                 return ice_dcf_configure_rss_lut(hw);
843         }
844
845         /* In IAVF, RSS enablement is set by PF driver. It is not supported
846          * to set based on rss_conf->rss_hf.
847          */
848
849         /* configure RSS key */
850         if (!rss_conf->rss_key)
851                 /* Calculate the default hash key */
852                 for (i = 0; i < hw->vf_res->rss_key_size; i++)
853                         hw->rss_key[i] = (uint8_t)rte_rand();
854         else
855                 rte_memcpy(hw->rss_key, rss_conf->rss_key,
856                            RTE_MIN(rss_conf->rss_key_len,
857                                    hw->vf_res->rss_key_size));
858
859         /* init RSS LUT table */
860         for (i = 0, j = 0; i < hw->vf_res->rss_lut_size; i++, j++) {
861                 if (j >= nb_q)
862                         j = 0;
863                 hw->rss_lut[i] = j;
864         }
865         /* send virtchnnl ops to configure rss*/
866         ret = ice_dcf_configure_rss_lut(hw);
867         if (ret)
868                 return ret;
869         ret = ice_dcf_configure_rss_key(hw);
870         if (ret)
871                 return ret;
872
873         return 0;
874 }
875
876 #define IAVF_RXDID_LEGACY_0 0
877 #define IAVF_RXDID_LEGACY_1 1
878 #define IAVF_RXDID_COMMS_OVS_1 22
879
880 int
881 ice_dcf_configure_queues(struct ice_dcf_hw *hw)
882 {
883         struct ice_rx_queue **rxq =
884                 (struct ice_rx_queue **)hw->eth_dev->data->rx_queues;
885         struct ice_tx_queue **txq =
886                 (struct ice_tx_queue **)hw->eth_dev->data->tx_queues;
887         struct virtchnl_vsi_queue_config_info *vc_config;
888         struct virtchnl_queue_pair_info *vc_qp;
889         struct dcf_virtchnl_cmd args;
890         uint16_t i, size;
891         int err;
892
893         size = sizeof(*vc_config) +
894                sizeof(vc_config->qpair[0]) * hw->num_queue_pairs;
895         vc_config = rte_zmalloc("cfg_queue", size, 0);
896         if (!vc_config)
897                 return -ENOMEM;
898
899         vc_config->vsi_id = hw->vsi_res->vsi_id;
900         vc_config->num_queue_pairs = hw->num_queue_pairs;
901
902         for (i = 0, vc_qp = vc_config->qpair;
903              i < hw->num_queue_pairs;
904              i++, vc_qp++) {
905                 vc_qp->txq.vsi_id = hw->vsi_res->vsi_id;
906                 vc_qp->txq.queue_id = i;
907                 if (i < hw->eth_dev->data->nb_tx_queues) {
908                         vc_qp->txq.ring_len = txq[i]->nb_tx_desc;
909                         vc_qp->txq.dma_ring_addr = txq[i]->tx_ring_dma;
910                 }
911                 vc_qp->rxq.vsi_id = hw->vsi_res->vsi_id;
912                 vc_qp->rxq.queue_id = i;
913
914                 if (i >= hw->eth_dev->data->nb_rx_queues)
915                         continue;
916
917                 vc_qp->rxq.max_pkt_size = rxq[i]->max_pkt_len;
918                 vc_qp->rxq.ring_len = rxq[i]->nb_rx_desc;
919                 vc_qp->rxq.dma_ring_addr = rxq[i]->rx_ring_dma;
920                 vc_qp->rxq.databuffer_size = rxq[i]->rx_buf_len;
921
922 #ifndef RTE_LIBRTE_ICE_16BYTE_RX_DESC
923                 if (hw->vf_res->vf_cap_flags &
924                     VIRTCHNL_VF_OFFLOAD_RX_FLEX_DESC &&
925                     hw->supported_rxdid &
926                     BIT(IAVF_RXDID_COMMS_OVS_1)) {
927                         vc_qp->rxq.rxdid = IAVF_RXDID_COMMS_OVS_1;
928                         PMD_DRV_LOG(NOTICE, "request RXDID == %d in "
929                                     "Queue[%d]", vc_qp->rxq.rxdid, i);
930                 } else {
931                         PMD_DRV_LOG(ERR, "RXDID 16 is not supported");
932                         return -EINVAL;
933                 }
934 #else
935                 if (hw->vf_res->vf_cap_flags &
936                         VIRTCHNL_VF_OFFLOAD_RX_FLEX_DESC &&
937                         hw->supported_rxdid &
938                         BIT(IAVF_RXDID_LEGACY_0)) {
939                         vc_qp->rxq.rxdid = IAVF_RXDID_LEGACY_0;
940                         PMD_DRV_LOG(NOTICE, "request RXDID == %d in "
941                                         "Queue[%d]", vc_qp->rxq.rxdid, i);
942                 } else {
943                         PMD_DRV_LOG(ERR, "RXDID == 0 is not supported");
944                         return -EINVAL;
945                 }
946 #endif
947                 ice_select_rxd_to_pkt_fields_handler(rxq[i], vc_qp->rxq.rxdid);
948         }
949
950         memset(&args, 0, sizeof(args));
951         args.v_op = VIRTCHNL_OP_CONFIG_VSI_QUEUES;
952         args.req_msg = (uint8_t *)vc_config;
953         args.req_msglen = size;
954
955         err = ice_dcf_execute_virtchnl_cmd(hw, &args);
956         if (err)
957                 PMD_DRV_LOG(ERR, "Failed to execute command of"
958                             " VIRTCHNL_OP_CONFIG_VSI_QUEUES");
959
960         rte_free(vc_config);
961         return err;
962 }
963
964 int
965 ice_dcf_config_irq_map(struct ice_dcf_hw *hw)
966 {
967         struct virtchnl_irq_map_info *map_info;
968         struct virtchnl_vector_map *vecmap;
969         struct dcf_virtchnl_cmd args;
970         int len, i, err;
971
972         len = sizeof(struct virtchnl_irq_map_info) +
973               sizeof(struct virtchnl_vector_map) * hw->nb_msix;
974
975         map_info = rte_zmalloc("map_info", len, 0);
976         if (!map_info)
977                 return -ENOMEM;
978
979         map_info->num_vectors = hw->nb_msix;
980         for (i = 0; i < hw->nb_msix; i++) {
981                 vecmap = &map_info->vecmap[i];
982                 vecmap->vsi_id = hw->vsi_res->vsi_id;
983                 vecmap->rxitr_idx = 0;
984                 vecmap->vector_id = hw->msix_base + i;
985                 vecmap->txq_map = 0;
986                 vecmap->rxq_map = hw->rxq_map[hw->msix_base + i];
987         }
988
989         memset(&args, 0, sizeof(args));
990         args.v_op = VIRTCHNL_OP_CONFIG_IRQ_MAP;
991         args.req_msg = (u8 *)map_info;
992         args.req_msglen = len;
993
994         err = ice_dcf_execute_virtchnl_cmd(hw, &args);
995         if (err)
996                 PMD_DRV_LOG(ERR, "fail to execute command OP_CONFIG_IRQ_MAP");
997
998         rte_free(map_info);
999         return err;
1000 }
1001
1002 int
1003 ice_dcf_switch_queue(struct ice_dcf_hw *hw, uint16_t qid, bool rx, bool on)
1004 {
1005         struct virtchnl_queue_select queue_select;
1006         struct dcf_virtchnl_cmd args;
1007         int err;
1008
1009         memset(&queue_select, 0, sizeof(queue_select));
1010         queue_select.vsi_id = hw->vsi_res->vsi_id;
1011         if (rx)
1012                 queue_select.rx_queues |= 1 << qid;
1013         else
1014                 queue_select.tx_queues |= 1 << qid;
1015
1016         memset(&args, 0, sizeof(args));
1017         if (on)
1018                 args.v_op = VIRTCHNL_OP_ENABLE_QUEUES;
1019         else
1020                 args.v_op = VIRTCHNL_OP_DISABLE_QUEUES;
1021
1022         args.req_msg = (u8 *)&queue_select;
1023         args.req_msglen = sizeof(queue_select);
1024
1025         err = ice_dcf_execute_virtchnl_cmd(hw, &args);
1026         if (err)
1027                 PMD_DRV_LOG(ERR, "Failed to execute command of %s",
1028                             on ? "OP_ENABLE_QUEUES" : "OP_DISABLE_QUEUES");
1029
1030         return err;
1031 }
1032
1033 int
1034 ice_dcf_disable_queues(struct ice_dcf_hw *hw)
1035 {
1036         struct virtchnl_queue_select queue_select;
1037         struct dcf_virtchnl_cmd args;
1038         int err;
1039
1040         if (hw->resetting)
1041                 return 0;
1042
1043         memset(&queue_select, 0, sizeof(queue_select));
1044         queue_select.vsi_id = hw->vsi_res->vsi_id;
1045
1046         queue_select.rx_queues = BIT(hw->eth_dev->data->nb_rx_queues) - 1;
1047         queue_select.tx_queues = BIT(hw->eth_dev->data->nb_tx_queues) - 1;
1048
1049         memset(&args, 0, sizeof(args));
1050         args.v_op = VIRTCHNL_OP_DISABLE_QUEUES;
1051         args.req_msg = (u8 *)&queue_select;
1052         args.req_msglen = sizeof(queue_select);
1053
1054         err = ice_dcf_execute_virtchnl_cmd(hw, &args);
1055         if (err)
1056                 PMD_DRV_LOG(ERR,
1057                             "Failed to execute command of OP_DISABLE_QUEUES");
1058
1059         return err;
1060 }
1061
1062 int
1063 ice_dcf_query_stats(struct ice_dcf_hw *hw,
1064                                    struct virtchnl_eth_stats *pstats)
1065 {
1066         struct virtchnl_queue_select q_stats;
1067         struct dcf_virtchnl_cmd args;
1068         int err;
1069
1070         memset(&q_stats, 0, sizeof(q_stats));
1071         q_stats.vsi_id = hw->vsi_res->vsi_id;
1072
1073         args.v_op = VIRTCHNL_OP_GET_STATS;
1074         args.req_msg = (uint8_t *)&q_stats;
1075         args.req_msglen = sizeof(q_stats);
1076         args.rsp_msglen = sizeof(*pstats);
1077         args.rsp_msgbuf = (uint8_t *)pstats;
1078         args.rsp_buflen = sizeof(*pstats);
1079
1080         err = ice_dcf_execute_virtchnl_cmd(hw, &args);
1081         if (err) {
1082                 PMD_DRV_LOG(ERR, "fail to execute command OP_GET_STATS");
1083                 return err;
1084         }
1085
1086         return 0;
1087 }
1088
1089 int
1090 ice_dcf_add_del_all_mac_addr(struct ice_dcf_hw *hw, bool add)
1091 {
1092         struct virtchnl_ether_addr_list *list;
1093         struct rte_ether_addr *addr;
1094         struct dcf_virtchnl_cmd args;
1095         int len, err = 0;
1096
1097         if (hw->resetting) {
1098                 if (!add)
1099                         return 0;
1100
1101                 PMD_DRV_LOG(ERR, "fail to add all MACs for VF resetting");
1102                 return -EIO;
1103         }
1104
1105         len = sizeof(struct virtchnl_ether_addr_list);
1106         addr = hw->eth_dev->data->mac_addrs;
1107         len += sizeof(struct virtchnl_ether_addr);
1108
1109         list = rte_zmalloc(NULL, len, 0);
1110         if (!list) {
1111                 PMD_DRV_LOG(ERR, "fail to allocate memory");
1112                 return -ENOMEM;
1113         }
1114
1115         rte_memcpy(list->list[0].addr, addr->addr_bytes,
1116                         sizeof(addr->addr_bytes));
1117         PMD_DRV_LOG(DEBUG, "add/rm mac:" RTE_ETHER_ADDR_PRT_FMT,
1118                             RTE_ETHER_ADDR_BYTES(addr));
1119
1120         list->vsi_id = hw->vsi_res->vsi_id;
1121         list->num_elements = 1;
1122
1123         memset(&args, 0, sizeof(args));
1124         args.v_op = add ? VIRTCHNL_OP_ADD_ETH_ADDR :
1125                         VIRTCHNL_OP_DEL_ETH_ADDR;
1126         args.req_msg = (uint8_t *)list;
1127         args.req_msglen  = len;
1128         err = ice_dcf_execute_virtchnl_cmd(hw, &args);
1129         if (err)
1130                 PMD_DRV_LOG(ERR, "fail to execute command %s",
1131                             add ? "OP_ADD_ETHER_ADDRESS" :
1132                             "OP_DEL_ETHER_ADDRESS");
1133         rte_free(list);
1134         return err;
1135 }