net/ice: init RSS and supported RXDID in DCF
[dpdk.git] / drivers / net / ice / ice_dcf.c
1 /* SPDX-License-Identifier: BSD-3-Clause
2  * Copyright(c) 2020 Intel Corporation
3  */
4
5 #include <sys/queue.h>
6 #include <stdio.h>
7 #include <errno.h>
8 #include <stdint.h>
9 #include <string.h>
10 #include <unistd.h>
11 #include <stdarg.h>
12 #include <inttypes.h>
13 #include <rte_byteorder.h>
14 #include <rte_common.h>
15
16 #include <rte_pci.h>
17 #include <rte_atomic.h>
18 #include <rte_eal.h>
19 #include <rte_ether.h>
20 #include <rte_ethdev_driver.h>
21 #include <rte_ethdev_pci.h>
22 #include <rte_malloc.h>
23 #include <rte_memzone.h>
24 #include <rte_dev.h>
25
26 #include "ice_dcf.h"
27
28 #define ICE_DCF_AQ_LEN     32
29 #define ICE_DCF_AQ_BUF_SZ  4096
30
31 #define ICE_DCF_ARQ_MAX_RETRIES 200
32 #define ICE_DCF_ARQ_CHECK_TIME  2   /* msecs */
33
34 #define ICE_DCF_VF_RES_BUF_SZ   \
35         (sizeof(struct virtchnl_vf_resource) +  \
36                 IAVF_MAX_VF_VSI * sizeof(struct virtchnl_vsi_resource))
37
38 static __rte_always_inline int
39 ice_dcf_send_cmd_req_no_irq(struct ice_dcf_hw *hw, enum virtchnl_ops op,
40                             uint8_t *req_msg, uint16_t req_msglen)
41 {
42         return iavf_aq_send_msg_to_pf(&hw->avf, op, IAVF_SUCCESS,
43                                       req_msg, req_msglen, NULL);
44 }
45
46 static int
47 ice_dcf_recv_cmd_rsp_no_irq(struct ice_dcf_hw *hw, enum virtchnl_ops op,
48                             uint8_t *rsp_msgbuf, uint16_t rsp_buflen,
49                             uint16_t *rsp_msglen)
50 {
51         struct iavf_arq_event_info event;
52         enum virtchnl_ops v_op;
53         int i = 0;
54         int err;
55
56         event.buf_len = rsp_buflen;
57         event.msg_buf = rsp_msgbuf;
58
59         do {
60                 err = iavf_clean_arq_element(&hw->avf, &event, NULL);
61                 if (err != IAVF_SUCCESS)
62                         goto again;
63
64                 v_op = rte_le_to_cpu_32(event.desc.cookie_high);
65                 if (v_op != op)
66                         goto again;
67
68                 if (rsp_msglen != NULL)
69                         *rsp_msglen = event.msg_len;
70                 return rte_le_to_cpu_32(event.desc.cookie_low);
71
72 again:
73                 rte_delay_ms(ICE_DCF_ARQ_CHECK_TIME);
74         } while (i++ < ICE_DCF_ARQ_MAX_RETRIES);
75
76         return -EIO;
77 }
78
79 static __rte_always_inline void
80 ice_dcf_aq_cmd_clear(struct ice_dcf_hw *hw, struct dcf_virtchnl_cmd *cmd)
81 {
82         rte_spinlock_lock(&hw->vc_cmd_queue_lock);
83
84         TAILQ_REMOVE(&hw->vc_cmd_queue, cmd, next);
85
86         rte_spinlock_unlock(&hw->vc_cmd_queue_lock);
87 }
88
89 static __rte_always_inline void
90 ice_dcf_vc_cmd_set(struct ice_dcf_hw *hw, struct dcf_virtchnl_cmd *cmd)
91 {
92         cmd->v_ret = IAVF_ERR_NOT_READY;
93         cmd->rsp_msglen = 0;
94         cmd->pending = 1;
95
96         rte_spinlock_lock(&hw->vc_cmd_queue_lock);
97
98         TAILQ_INSERT_TAIL(&hw->vc_cmd_queue, cmd, next);
99
100         rte_spinlock_unlock(&hw->vc_cmd_queue_lock);
101 }
102
103 static __rte_always_inline int
104 ice_dcf_vc_cmd_send(struct ice_dcf_hw *hw, struct dcf_virtchnl_cmd *cmd)
105 {
106         return iavf_aq_send_msg_to_pf(&hw->avf,
107                                       cmd->v_op, IAVF_SUCCESS,
108                                       cmd->req_msg, cmd->req_msglen, NULL);
109 }
110
111 static __rte_always_inline void
112 ice_dcf_aq_cmd_handle(struct ice_dcf_hw *hw, struct iavf_arq_event_info *info)
113 {
114         struct dcf_virtchnl_cmd *cmd;
115         enum virtchnl_ops v_op;
116         enum iavf_status v_ret;
117         uint16_t aq_op;
118
119         aq_op = rte_le_to_cpu_16(info->desc.opcode);
120         if (unlikely(aq_op != iavf_aqc_opc_send_msg_to_vf)) {
121                 PMD_DRV_LOG(ERR,
122                             "Request %u is not supported yet", aq_op);
123                 return;
124         }
125
126         v_op = rte_le_to_cpu_32(info->desc.cookie_high);
127         if (v_op == VIRTCHNL_OP_EVENT) {
128                 if (hw->vc_event_msg_cb != NULL)
129                         hw->vc_event_msg_cb(hw,
130                                             info->msg_buf,
131                                             info->msg_len);
132                 return;
133         }
134
135         v_ret = rte_le_to_cpu_32(info->desc.cookie_low);
136
137         rte_spinlock_lock(&hw->vc_cmd_queue_lock);
138
139         TAILQ_FOREACH(cmd, &hw->vc_cmd_queue, next) {
140                 if (cmd->v_op == v_op && cmd->pending) {
141                         cmd->v_ret = v_ret;
142                         cmd->rsp_msglen = RTE_MIN(info->msg_len,
143                                                   cmd->rsp_buflen);
144                         if (likely(cmd->rsp_msglen != 0))
145                                 rte_memcpy(cmd->rsp_msgbuf, info->msg_buf,
146                                            cmd->rsp_msglen);
147
148                         /* prevent compiler reordering */
149                         rte_compiler_barrier();
150                         cmd->pending = 0;
151                         break;
152                 }
153         }
154
155         rte_spinlock_unlock(&hw->vc_cmd_queue_lock);
156 }
157
158 static void
159 ice_dcf_handle_virtchnl_msg(struct ice_dcf_hw *hw)
160 {
161         struct iavf_arq_event_info info;
162         uint16_t pending = 1;
163         int ret;
164
165         info.buf_len = ICE_DCF_AQ_BUF_SZ;
166         info.msg_buf = hw->arq_buf;
167
168         while (pending) {
169                 ret = iavf_clean_arq_element(&hw->avf, &info, &pending);
170                 if (ret != IAVF_SUCCESS)
171                         break;
172
173                 ice_dcf_aq_cmd_handle(hw, &info);
174         }
175 }
176
177 static int
178 ice_dcf_init_check_api_version(struct ice_dcf_hw *hw)
179 {
180 #define ICE_CPF_VIRTCHNL_VERSION_MAJOR_START    1
181 #define ICE_CPF_VIRTCHNL_VERSION_MINOR_START    1
182         struct virtchnl_version_info version, *pver;
183         int err;
184
185         version.major = VIRTCHNL_VERSION_MAJOR;
186         version.minor = VIRTCHNL_VERSION_MINOR;
187         err = ice_dcf_send_cmd_req_no_irq(hw, VIRTCHNL_OP_VERSION,
188                                           (uint8_t *)&version, sizeof(version));
189         if (err) {
190                 PMD_INIT_LOG(ERR, "Failed to send OP_VERSION");
191                 return err;
192         }
193
194         pver = &hw->virtchnl_version;
195         err = ice_dcf_recv_cmd_rsp_no_irq(hw, VIRTCHNL_OP_VERSION,
196                                           (uint8_t *)pver, sizeof(*pver), NULL);
197         if (err) {
198                 PMD_INIT_LOG(ERR, "Failed to get response of OP_VERSION");
199                 return -1;
200         }
201
202         PMD_INIT_LOG(DEBUG,
203                      "Peer PF API version: %u.%u", pver->major, pver->minor);
204
205         if (pver->major < ICE_CPF_VIRTCHNL_VERSION_MAJOR_START ||
206             (pver->major == ICE_CPF_VIRTCHNL_VERSION_MAJOR_START &&
207              pver->minor < ICE_CPF_VIRTCHNL_VERSION_MINOR_START)) {
208                 PMD_INIT_LOG(ERR,
209                              "VIRTCHNL API version should not be lower than (%u.%u)",
210                              ICE_CPF_VIRTCHNL_VERSION_MAJOR_START,
211                              ICE_CPF_VIRTCHNL_VERSION_MAJOR_START);
212                 return -1;
213         } else if (pver->major > VIRTCHNL_VERSION_MAJOR ||
214                    (pver->major == VIRTCHNL_VERSION_MAJOR &&
215                     pver->minor > VIRTCHNL_VERSION_MINOR)) {
216                 PMD_INIT_LOG(ERR,
217                              "PF/VF API version mismatch:(%u.%u)-(%u.%u)",
218                              pver->major, pver->minor,
219                              VIRTCHNL_VERSION_MAJOR, VIRTCHNL_VERSION_MINOR);
220                 return -1;
221         }
222
223         PMD_INIT_LOG(DEBUG, "Peer is supported PF host");
224
225         return 0;
226 }
227
228 static int
229 ice_dcf_get_vf_resource(struct ice_dcf_hw *hw)
230 {
231         uint32_t caps;
232         int err, i;
233
234         caps = VIRTCHNL_VF_OFFLOAD_WB_ON_ITR | VIRTCHNL_VF_OFFLOAD_RX_POLLING |
235                VIRTCHNL_VF_CAP_ADV_LINK_SPEED | VIRTCHNL_VF_CAP_DCF |
236                VF_BASE_MODE_OFFLOADS | VIRTCHNL_VF_OFFLOAD_RX_FLEX_DESC;
237
238         err = ice_dcf_send_cmd_req_no_irq(hw, VIRTCHNL_OP_GET_VF_RESOURCES,
239                                           (uint8_t *)&caps, sizeof(caps));
240         if (err) {
241                 PMD_DRV_LOG(ERR, "Failed to send msg OP_GET_VF_RESOURCE");
242                 return err;
243         }
244
245         err = ice_dcf_recv_cmd_rsp_no_irq(hw, VIRTCHNL_OP_GET_VF_RESOURCES,
246                                           (uint8_t *)hw->vf_res,
247                                           ICE_DCF_VF_RES_BUF_SZ, NULL);
248         if (err) {
249                 PMD_DRV_LOG(ERR, "Failed to get response of OP_GET_VF_RESOURCE");
250                 return -1;
251         }
252
253         iavf_vf_parse_hw_config(&hw->avf, hw->vf_res);
254
255         hw->vsi_res = NULL;
256         for (i = 0; i < hw->vf_res->num_vsis; i++) {
257                 if (hw->vf_res->vsi_res[i].vsi_type == VIRTCHNL_VSI_SRIOV)
258                         hw->vsi_res = &hw->vf_res->vsi_res[i];
259         }
260
261         if (!hw->vsi_res) {
262                 PMD_DRV_LOG(ERR, "no LAN VSI found");
263                 return -1;
264         }
265
266         hw->vsi_id = hw->vsi_res->vsi_id;
267         PMD_DRV_LOG(DEBUG, "VSI ID is %u", hw->vsi_id);
268
269         return 0;
270 }
271
272 static int
273 ice_dcf_get_vf_vsi_map(struct ice_dcf_hw *hw)
274 {
275         struct virtchnl_dcf_vsi_map *vsi_map;
276         uint32_t valid_msg_len;
277         uint16_t len;
278         int err;
279
280         err = ice_dcf_send_cmd_req_no_irq(hw, VIRTCHNL_OP_DCF_GET_VSI_MAP,
281                                           NULL, 0);
282         if (err) {
283                 PMD_DRV_LOG(ERR, "Failed to send msg OP_DCF_GET_VSI_MAP");
284                 return err;
285         }
286
287         err = ice_dcf_recv_cmd_rsp_no_irq(hw, VIRTCHNL_OP_DCF_GET_VSI_MAP,
288                                           hw->arq_buf, ICE_DCF_AQ_BUF_SZ,
289                                           &len);
290         if (err) {
291                 PMD_DRV_LOG(ERR, "Failed to get response of OP_DCF_GET_VSI_MAP");
292                 return err;
293         }
294
295         vsi_map = (struct virtchnl_dcf_vsi_map *)hw->arq_buf;
296         valid_msg_len = (vsi_map->num_vfs - 1) * sizeof(vsi_map->vf_vsi[0]) +
297                         sizeof(*vsi_map);
298         if (len != valid_msg_len) {
299                 PMD_DRV_LOG(ERR, "invalid vf vsi map response with length %u",
300                             len);
301                 return -EINVAL;
302         }
303
304         if (hw->num_vfs != 0 && hw->num_vfs != vsi_map->num_vfs) {
305                 PMD_DRV_LOG(ERR, "The number VSI map (%u) doesn't match the number of VFs (%u)",
306                             vsi_map->num_vfs, hw->num_vfs);
307                 return -EINVAL;
308         }
309
310         len = vsi_map->num_vfs * sizeof(vsi_map->vf_vsi[0]);
311
312         if (!hw->vf_vsi_map) {
313                 hw->vf_vsi_map = rte_zmalloc("vf_vsi_ctx", len, 0);
314                 if (!hw->vf_vsi_map) {
315                         PMD_DRV_LOG(ERR, "Failed to alloc memory for VSI context");
316                         return -ENOMEM;
317                 }
318
319                 hw->num_vfs = vsi_map->num_vfs;
320         }
321
322         if (!memcmp(hw->vf_vsi_map, vsi_map->vf_vsi, len)) {
323                 PMD_DRV_LOG(DEBUG, "VF VSI map doesn't change");
324                 return 1;
325         }
326
327         rte_memcpy(hw->vf_vsi_map, vsi_map->vf_vsi, len);
328         return 0;
329 }
330
331 static int
332 ice_dcf_mode_disable(struct ice_dcf_hw *hw)
333 {
334         int err;
335
336         err = ice_dcf_send_cmd_req_no_irq(hw, VIRTCHNL_OP_DCF_DISABLE,
337                                           NULL, 0);
338         if (err) {
339                 PMD_DRV_LOG(ERR, "Failed to send msg OP_DCF_DISABLE");
340                 return err;
341         }
342
343         err = ice_dcf_recv_cmd_rsp_no_irq(hw, VIRTCHNL_OP_DCF_DISABLE,
344                                           hw->arq_buf, ICE_DCF_AQ_BUF_SZ, NULL);
345         if (err) {
346                 PMD_DRV_LOG(ERR,
347                             "Failed to get response of OP_DCF_DISABLE %d",
348                             err);
349                 return -1;
350         }
351
352         return 0;
353 }
354
355 static int
356 ice_dcf_check_reset_done(struct ice_dcf_hw *hw)
357 {
358 #define ICE_DCF_RESET_WAIT_CNT       50
359         struct iavf_hw *avf = &hw->avf;
360         int i, reset;
361
362         for (i = 0; i < ICE_DCF_RESET_WAIT_CNT; i++) {
363                 reset = IAVF_READ_REG(avf, IAVF_VFGEN_RSTAT) &
364                                         IAVF_VFGEN_RSTAT_VFR_STATE_MASK;
365                 reset = reset >> IAVF_VFGEN_RSTAT_VFR_STATE_SHIFT;
366
367                 if (reset == VIRTCHNL_VFR_VFACTIVE ||
368                     reset == VIRTCHNL_VFR_COMPLETED)
369                         break;
370
371                 rte_delay_ms(20);
372         }
373
374         if (i >= ICE_DCF_RESET_WAIT_CNT)
375                 return -1;
376
377         return 0;
378 }
379
380 static inline void
381 ice_dcf_enable_irq0(struct ice_dcf_hw *hw)
382 {
383         struct iavf_hw *avf = &hw->avf;
384
385         /* Enable admin queue interrupt trigger */
386         IAVF_WRITE_REG(avf, IAVF_VFINT_ICR0_ENA1,
387                        IAVF_VFINT_ICR0_ENA1_ADMINQ_MASK);
388         IAVF_WRITE_REG(avf, IAVF_VFINT_DYN_CTL01,
389                        IAVF_VFINT_DYN_CTL01_INTENA_MASK |
390                        IAVF_VFINT_DYN_CTL01_CLEARPBA_MASK |
391                        IAVF_VFINT_DYN_CTL01_ITR_INDX_MASK);
392
393         IAVF_WRITE_FLUSH(avf);
394 }
395
396 static inline void
397 ice_dcf_disable_irq0(struct ice_dcf_hw *hw)
398 {
399         struct iavf_hw *avf = &hw->avf;
400
401         /* Disable all interrupt types */
402         IAVF_WRITE_REG(avf, IAVF_VFINT_ICR0_ENA1, 0);
403         IAVF_WRITE_REG(avf, IAVF_VFINT_DYN_CTL01,
404                        IAVF_VFINT_DYN_CTL01_ITR_INDX_MASK);
405
406         IAVF_WRITE_FLUSH(avf);
407 }
408
409 static void
410 ice_dcf_dev_interrupt_handler(void *param)
411 {
412         struct ice_dcf_hw *hw = param;
413
414         ice_dcf_disable_irq0(hw);
415
416         ice_dcf_handle_virtchnl_msg(hw);
417
418         ice_dcf_enable_irq0(hw);
419 }
420
421 int
422 ice_dcf_execute_virtchnl_cmd(struct ice_dcf_hw *hw,
423                              struct dcf_virtchnl_cmd *cmd)
424 {
425         int i = 0;
426         int err;
427
428         if ((cmd->req_msg && !cmd->req_msglen) ||
429             (!cmd->req_msg && cmd->req_msglen) ||
430             (cmd->rsp_msgbuf && !cmd->rsp_buflen) ||
431             (!cmd->rsp_msgbuf && cmd->rsp_buflen))
432                 return -EINVAL;
433
434         rte_spinlock_lock(&hw->vc_cmd_send_lock);
435         ice_dcf_vc_cmd_set(hw, cmd);
436
437         err = ice_dcf_vc_cmd_send(hw, cmd);
438         if (err) {
439                 PMD_DRV_LOG(ERR, "fail to send cmd %d", cmd->v_op);
440                 goto ret;
441         }
442
443         do {
444                 if (!cmd->pending)
445                         break;
446
447                 rte_delay_ms(ICE_DCF_ARQ_CHECK_TIME);
448         } while (i++ < ICE_DCF_ARQ_MAX_RETRIES);
449
450         if (cmd->v_ret != IAVF_SUCCESS) {
451                 err = -1;
452                 PMD_DRV_LOG(ERR,
453                             "No response (%d times) or return failure (%d) for cmd %d",
454                             i, cmd->v_ret, cmd->v_op);
455         }
456
457 ret:
458         ice_dcf_aq_cmd_clear(hw, cmd);
459         rte_spinlock_unlock(&hw->vc_cmd_send_lock);
460         return err;
461 }
462
463 int
464 ice_dcf_send_aq_cmd(void *dcf_hw, struct ice_aq_desc *desc,
465                     void *buf, uint16_t buf_size)
466 {
467         struct dcf_virtchnl_cmd desc_cmd, buff_cmd;
468         struct ice_dcf_hw *hw = dcf_hw;
469         int err = 0;
470         int i = 0;
471
472         if ((buf && !buf_size) || (!buf && buf_size) ||
473             buf_size > ICE_DCF_AQ_BUF_SZ)
474                 return -EINVAL;
475
476         desc_cmd.v_op = VIRTCHNL_OP_DCF_CMD_DESC;
477         desc_cmd.req_msglen = sizeof(*desc);
478         desc_cmd.req_msg = (uint8_t *)desc;
479         desc_cmd.rsp_buflen = sizeof(*desc);
480         desc_cmd.rsp_msgbuf = (uint8_t *)desc;
481
482         if (buf == NULL)
483                 return ice_dcf_execute_virtchnl_cmd(hw, &desc_cmd);
484
485         desc->flags |= rte_cpu_to_le_16(ICE_AQ_FLAG_BUF);
486
487         buff_cmd.v_op = VIRTCHNL_OP_DCF_CMD_BUFF;
488         buff_cmd.req_msglen = buf_size;
489         buff_cmd.req_msg = buf;
490         buff_cmd.rsp_buflen = buf_size;
491         buff_cmd.rsp_msgbuf = buf;
492
493         rte_spinlock_lock(&hw->vc_cmd_send_lock);
494         ice_dcf_vc_cmd_set(hw, &desc_cmd);
495         ice_dcf_vc_cmd_set(hw, &buff_cmd);
496
497         if (ice_dcf_vc_cmd_send(hw, &desc_cmd) ||
498             ice_dcf_vc_cmd_send(hw, &buff_cmd)) {
499                 err = -1;
500                 PMD_DRV_LOG(ERR, "fail to send OP_DCF_CMD_DESC/BUFF");
501                 goto ret;
502         }
503
504         do {
505                 if ((!desc_cmd.pending && !buff_cmd.pending) ||
506                     (!desc_cmd.pending && desc_cmd.v_ret != IAVF_SUCCESS) ||
507                     (!buff_cmd.pending && buff_cmd.v_ret != IAVF_SUCCESS))
508                         break;
509
510                 rte_delay_ms(ICE_DCF_ARQ_CHECK_TIME);
511         } while (i++ < ICE_DCF_ARQ_MAX_RETRIES);
512
513         if (desc_cmd.v_ret != IAVF_SUCCESS || buff_cmd.v_ret != IAVF_SUCCESS) {
514                 err = -1;
515                 PMD_DRV_LOG(ERR,
516                             "No response (%d times) or return failure (desc: %d / buff: %d)",
517                             i, desc_cmd.v_ret, buff_cmd.v_ret);
518         }
519
520 ret:
521         ice_dcf_aq_cmd_clear(hw, &desc_cmd);
522         ice_dcf_aq_cmd_clear(hw, &buff_cmd);
523         rte_spinlock_unlock(&hw->vc_cmd_send_lock);
524
525         return err;
526 }
527
528 int
529 ice_dcf_handle_vsi_update_event(struct ice_dcf_hw *hw)
530 {
531         struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(hw->eth_dev);
532         int err = 0;
533
534         rte_spinlock_lock(&hw->vc_cmd_send_lock);
535
536         rte_intr_disable(&pci_dev->intr_handle);
537         ice_dcf_disable_irq0(hw);
538
539         if (ice_dcf_get_vf_resource(hw) || ice_dcf_get_vf_vsi_map(hw) < 0)
540                 err = -1;
541
542         rte_intr_enable(&pci_dev->intr_handle);
543         ice_dcf_enable_irq0(hw);
544
545         rte_spinlock_unlock(&hw->vc_cmd_send_lock);
546
547         return err;
548 }
549
550 static int
551 ice_dcf_get_supported_rxdid(struct ice_dcf_hw *hw)
552 {
553         int err;
554
555         err = ice_dcf_send_cmd_req_no_irq(hw,
556                                           VIRTCHNL_OP_GET_SUPPORTED_RXDIDS,
557                                           NULL, 0);
558         if (err) {
559                 PMD_INIT_LOG(ERR, "Failed to send OP_GET_SUPPORTED_RXDIDS");
560                 return -1;
561         }
562
563         err = ice_dcf_recv_cmd_rsp_no_irq(hw, VIRTCHNL_OP_GET_SUPPORTED_RXDIDS,
564                                           (uint8_t *)&hw->supported_rxdid,
565                                           sizeof(uint64_t), NULL);
566         if (err) {
567                 PMD_INIT_LOG(ERR, "Failed to get response of OP_GET_SUPPORTED_RXDIDS");
568                 return -1;
569         }
570
571         return 0;
572 }
573
574 int
575 ice_dcf_init_hw(struct rte_eth_dev *eth_dev, struct ice_dcf_hw *hw)
576 {
577         struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
578         int ret;
579
580         hw->avf.hw_addr = pci_dev->mem_resource[0].addr;
581         hw->avf.back = hw;
582
583         hw->avf.bus.bus_id = pci_dev->addr.bus;
584         hw->avf.bus.device = pci_dev->addr.devid;
585         hw->avf.bus.func = pci_dev->addr.function;
586
587         hw->avf.device_id = pci_dev->id.device_id;
588         hw->avf.vendor_id = pci_dev->id.vendor_id;
589         hw->avf.subsystem_device_id = pci_dev->id.subsystem_device_id;
590         hw->avf.subsystem_vendor_id = pci_dev->id.subsystem_vendor_id;
591
592         hw->avf.aq.num_arq_entries = ICE_DCF_AQ_LEN;
593         hw->avf.aq.num_asq_entries = ICE_DCF_AQ_LEN;
594         hw->avf.aq.arq_buf_size = ICE_DCF_AQ_BUF_SZ;
595         hw->avf.aq.asq_buf_size = ICE_DCF_AQ_BUF_SZ;
596
597         rte_spinlock_init(&hw->vc_cmd_send_lock);
598         rte_spinlock_init(&hw->vc_cmd_queue_lock);
599         TAILQ_INIT(&hw->vc_cmd_queue);
600
601         hw->arq_buf = rte_zmalloc("arq_buf", ICE_DCF_AQ_BUF_SZ, 0);
602         if (hw->arq_buf == NULL) {
603                 PMD_INIT_LOG(ERR, "unable to allocate AdminQ buffer memory");
604                 goto err;
605         }
606
607         ret = iavf_set_mac_type(&hw->avf);
608         if (ret) {
609                 PMD_INIT_LOG(ERR, "set_mac_type failed: %d", ret);
610                 goto err;
611         }
612
613         ret = ice_dcf_check_reset_done(hw);
614         if (ret) {
615                 PMD_INIT_LOG(ERR, "VF is still resetting");
616                 goto err;
617         }
618
619         ret = iavf_init_adminq(&hw->avf);
620         if (ret) {
621                 PMD_INIT_LOG(ERR, "init_adminq failed: %d", ret);
622                 goto err;
623         }
624
625         if (ice_dcf_init_check_api_version(hw)) {
626                 PMD_INIT_LOG(ERR, "check_api version failed");
627                 goto err_api;
628         }
629
630         hw->vf_res = rte_zmalloc("vf_res", ICE_DCF_VF_RES_BUF_SZ, 0);
631         if (hw->vf_res == NULL) {
632                 PMD_INIT_LOG(ERR, "unable to allocate vf_res memory");
633                 goto err_api;
634         }
635
636         if (ice_dcf_get_vf_resource(hw)) {
637                 PMD_INIT_LOG(ERR, "Failed to get VF resource");
638                 goto err_alloc;
639         }
640
641         if (ice_dcf_get_vf_vsi_map(hw) < 0) {
642                 PMD_INIT_LOG(ERR, "Failed to get VF VSI map");
643                 ice_dcf_mode_disable(hw);
644                 goto err_alloc;
645         }
646
647         /* Allocate memory for RSS info */
648         if (hw->vf_res->vf_cap_flags & VIRTCHNL_VF_OFFLOAD_RSS_PF) {
649                 hw->rss_key = rte_zmalloc(NULL,
650                                           hw->vf_res->rss_key_size, 0);
651                 if (!hw->rss_key) {
652                         PMD_INIT_LOG(ERR, "unable to allocate rss_key memory");
653                         goto err_alloc;
654                 }
655                 hw->rss_lut = rte_zmalloc("rss_lut",
656                                           hw->vf_res->rss_lut_size, 0);
657                 if (!hw->rss_lut) {
658                         PMD_INIT_LOG(ERR, "unable to allocate rss_lut memory");
659                         goto err_rss;
660                 }
661         }
662
663         if (hw->vf_res->vf_cap_flags & VIRTCHNL_VF_OFFLOAD_RX_FLEX_DESC) {
664                 if (ice_dcf_get_supported_rxdid(hw) != 0) {
665                         PMD_INIT_LOG(ERR, "failed to do get supported rxdid");
666                         goto err_rss;
667                 }
668         }
669
670         hw->eth_dev = eth_dev;
671         rte_intr_callback_register(&pci_dev->intr_handle,
672                                    ice_dcf_dev_interrupt_handler, hw);
673         rte_intr_enable(&pci_dev->intr_handle);
674         ice_dcf_enable_irq0(hw);
675
676         return 0;
677
678 err_rss:
679         rte_free(hw->rss_key);
680         rte_free(hw->rss_lut);
681 err_alloc:
682         rte_free(hw->vf_res);
683 err_api:
684         iavf_shutdown_adminq(&hw->avf);
685 err:
686         rte_free(hw->arq_buf);
687
688         return -1;
689 }
690
691 void
692 ice_dcf_uninit_hw(struct rte_eth_dev *eth_dev, struct ice_dcf_hw *hw)
693 {
694         struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
695         struct rte_intr_handle *intr_handle = &pci_dev->intr_handle;
696
697         ice_dcf_disable_irq0(hw);
698         rte_intr_disable(intr_handle);
699         rte_intr_callback_unregister(intr_handle,
700                                      ice_dcf_dev_interrupt_handler, hw);
701
702         ice_dcf_mode_disable(hw);
703         iavf_shutdown_adminq(&hw->avf);
704
705         rte_free(hw->arq_buf);
706         rte_free(hw->vf_vsi_map);
707         rte_free(hw->vf_res);
708         rte_free(hw->rss_lut);
709         rte_free(hw->rss_key);
710 }