net/i40e: fix boundary check in RSS config
[dpdk.git] / drivers / net / octeontx / base / octeontx_pkovf.c
index 55332ad..bf28bc7 100644 (file)
@@ -1,34 +1,7 @@
-/*
- *   BSD LICENSE
- *
- *   Copyright (C) Cavium Inc. 2017. All rights reserved.
- *
- *   Redistribution and use in source and binary forms, with or without
- *   modification, are permitted provided that the following conditions
- *   are met:
- *
- *     * Redistributions of source code must retain the above copyright
- *       notice, this list of conditions and the following disclaimer.
- *     * Redistributions in binary form must reproduce the above copyright
- *       notice, this list of conditions and the following disclaimer in
- *       the documentation and/or other materials provided with the
- *       distribution.
- *     * Neither the name of Cavium networks nor the names of its
- *       contributors may be used to endorse or promote products derived
- *       from this software without specific prior written permission.
- *
- *   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
- *   "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
- *   LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
- *   A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
- *   OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
- *   SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
- *   LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
- *   DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
- *   THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
- *   (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
- *   OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2017 Cavium, Inc
  */
+
 #include <stdbool.h>
 #include <string.h>
 #include <stdio.h>
@@ -37,7 +10,7 @@
 #include <rte_cycles.h>
 #include <rte_malloc.h>
 #include <rte_memory.h>
-#include <rte_pci.h>
+#include <rte_bus_pci.h>
 #include <rte_spinlock.h>
 
 #include "../octeontx_logs.h"
 
 struct octeontx_pko_iomem {
        uint8_t         *va;
-       phys_addr_t     iova;
+       rte_iova_t      iova;
        size_t          size;
 };
 
 #define PKO_IOMEM_NULL (struct octeontx_pko_iomem){0, 0, 0}
+#define PKO_VALID      0x1
+#define PKO_INUSE      0x2
 
 struct octeontx_pko_fc_ctl_s {
        int64_t buf_cnt;
@@ -60,13 +35,14 @@ struct octeontx_pko_fc_ctl_s {
 struct octeontx_pkovf {
        uint8_t         *bar0;
        uint8_t         *bar2;
+       uint8_t         status;
        uint16_t        domain;
        uint16_t        vfid;
 };
 
 struct octeontx_pko_vf_ctl_s {
        rte_spinlock_t lock;
-
+       uint16_t global_domain;
        struct octeontx_pko_iomem fc_iomem;
        struct octeontx_pko_fc_ctl_s *fc_ctl;
        struct octeontx_pkovf pko[PKO_VF_MAX];
@@ -77,6 +53,494 @@ struct octeontx_pko_vf_ctl_s {
 
 static struct octeontx_pko_vf_ctl_s pko_vf_ctl;
 
+static void *
+octeontx_pko_dq_vf_bar0(uint16_t txq)
+{
+       int vf_ix;
+
+       vf_ix = txq / PKO_VF_NUM_DQ;
+       return pko_vf_ctl.pko[vf_ix].bar0;
+}
+
+static int
+octeontx_pko_dq_gdq(uint16_t txq)
+{
+       return txq % PKO_VF_NUM_DQ;
+}
+
+/**
+ * Open a PKO DQ.
+ */
+static inline
+int octeontx_pko_dq_open(uint16_t txq)
+{
+       unsigned int reg_off;
+       uint8_t *vf_bar0;
+       uint64_t rtn;
+       int gdq;
+
+       vf_bar0 = octeontx_pko_dq_vf_bar0(txq);
+       gdq = octeontx_pko_dq_gdq(txq);
+
+       if (unlikely(gdq < 0 || vf_bar0 == NULL))
+               return -EINVAL;
+       *(volatile int64_t*)(pko_vf_ctl.fc_ctl + txq) =
+               PKO_DQ_FC_DEPTH_PAGES - PKO_DQ_FC_SKID;
+
+       rte_wmb();
+
+       octeontx_write64(PKO_DQ_FC_DEPTH_PAGES,
+                        vf_bar0 + PKO_VF_DQ_FC_STATUS(gdq));
+
+       /* Set the register to return descriptor (packet) count as DEPTH */
+       /* KIND=1, NCB_QUERY_RSP=0 */
+       octeontx_write64(1ull << PKO_DQ_KIND_BIT,
+                               vf_bar0 + PKO_VF_DQ_WM_CTL(gdq));
+       reg_off = PKO_VF_DQ_OP_OPEN(gdq);
+
+       rtn = octeontx_reg_ldadd_u64(vf_bar0 + reg_off, 0);
+
+       /* PKO_DQOP_E::OPEN */
+       if (((rtn >> PKO_DQ_OP_BIT) & 0x3) != 0x1)
+               return -EIO;
+
+       switch (rtn >> PKO_DQ_STATUS_BIT) {
+       case 0xC:       /* DQALREADYCREATED */
+       case 0x0:       /* PASS */
+               break;
+       default:
+               return -EIO;
+       }
+
+       /* DRAIN=0, DRAIN_NULL_LINK=0, SW_XOFF=0 */
+       octeontx_write64(0, vf_bar0 + PKO_VF_DQ_SW_XOFF(gdq));
+
+       return rtn & ((1ull << PKO_DQ_OP_BIT) - 1);
+}
+
+/**
+ * Close a PKO DQ
+ * Flush all packets pending.
+ */
+static inline
+int octeontx_pko_dq_close(uint16_t txq)
+{
+       unsigned int reg_off;
+       uint8_t *vf_bar0;
+       uint64_t rtn;
+       int res;
+
+       vf_bar0 = octeontx_pko_dq_vf_bar0(txq);
+       res = octeontx_pko_dq_gdq(txq);
+
+       if (unlikely(res < 0 || vf_bar0 == NULL))
+               return -EINVAL;
+
+       reg_off = PKO_VF_DQ_OP_CLOSE(res);
+
+       rtn = octeontx_reg_ldadd_u64(vf_bar0 + reg_off, 0);
+
+       /* PKO_DQOP_E::CLOSE */
+       if (((rtn >> PKO_DQ_OP_BIT) & 0x3) != 0x2)
+               return -EIO;
+
+       switch (rtn >> PKO_DQ_STATUS_BIT) {
+       case 0xD:       /* DQNOTCREATED */
+       case 0x0:       /* PASS */
+               break;
+       default:
+               return -EIO;
+       }
+
+       res = rtn & ((1ull << PKO_DQ_OP_BIT) - 1); /* DEPTH */
+       return res;
+}
+
+/* Flush all packets pending on a DQ */
+static inline
+int octeontx_pko_dq_drain(uint16_t txq)
+{
+       unsigned int gdq;
+       uint8_t *vf_bar0;
+       uint64_t reg;
+       int res, timo = PKO_DQ_DRAIN_TO;
+
+       vf_bar0 = octeontx_pko_dq_vf_bar0(txq);
+       res = octeontx_pko_dq_gdq(txq);
+       gdq = res;
+
+        /* DRAIN=1, DRAIN_NULL_LINK=0, SW_XOFF=1 */
+        octeontx_write64(0x3, vf_bar0 + PKO_VF_DQ_SW_XOFF(gdq));
+       /* Wait until buffers leave DQs */
+       reg = octeontx_read64(vf_bar0 + PKO_VF_DQ_WM_CNT(gdq));
+       while (reg && timo > 0) {
+               rte_delay_us(100);
+               timo--;
+               reg = octeontx_read64(vf_bar0 + PKO_VF_DQ_WM_CNT(gdq));
+       }
+       /* DRAIN=0, DRAIN_NULL_LINK=0, SW_XOFF=0 */
+       octeontx_write64(0, vf_bar0 + PKO_VF_DQ_SW_XOFF(gdq));
+
+       return reg;
+}
+
+static inline int
+octeontx_pko_dq_range_lookup(struct octeontx_pko_vf_ctl_s *ctl, uint64_t chanid,
+                            unsigned int dq_num, unsigned int dq_from)
+{
+       unsigned int dq, dq_cnt;
+       unsigned int dq_base;
+
+       dq_cnt = 0;
+       dq = dq_from;
+       while (dq < RTE_DIM(ctl->dq_map)) {
+               dq_base = dq;
+               dq_cnt = 0;
+               while (ctl->dq_map[dq].chanid == ~chanid &&
+                       dq < RTE_DIM(ctl->dq_map)) {
+                       dq_cnt++;
+                       if (dq_cnt == dq_num)
+                               return dq_base;
+                       dq++;
+               }
+               dq++;
+       }
+       return -1;
+}
+
+static inline void
+octeontx_pko_dq_range_assign(struct octeontx_pko_vf_ctl_s *ctl, uint64_t chanid,
+                            unsigned int dq_base, unsigned int dq_num)
+{
+       unsigned int dq, dq_cnt;
+
+       dq_cnt = 0;
+       while (dq_cnt < dq_num) {
+               dq = dq_base + dq_cnt;
+
+               octeontx_log_dbg("DQ# %u assigned to CHAN# %" PRIx64 "", dq,
+                       chanid);
+
+               ctl->dq_map[dq].chanid = ~chanid;
+               dq_cnt++;
+       }
+}
+
+static inline int
+octeontx_pko_dq_claim(struct octeontx_pko_vf_ctl_s *ctl, unsigned int dq_base,
+                     unsigned int dq_num, uint64_t chanid)
+{
+       const uint64_t null_chanid = ~0ull;
+       int dq;
+
+       rte_spinlock_lock(&ctl->lock);
+
+       dq = octeontx_pko_dq_range_lookup(ctl, null_chanid, dq_num, dq_base);
+       if (dq < 0 || (unsigned int)dq != dq_base) {
+               rte_spinlock_unlock(&ctl->lock);
+               return -1;
+       }
+       octeontx_pko_dq_range_assign(ctl, chanid, dq_base, dq_num);
+
+       rte_spinlock_unlock(&ctl->lock);
+
+       return 0;
+}
+
+static inline int
+octeontx_pko_dq_free(struct octeontx_pko_vf_ctl_s *ctl, uint64_t chanid)
+{
+       const uint64_t null_chanid = ~0ull;
+       unsigned int dq = 0, dq_cnt = 0;
+
+       rte_spinlock_lock(&ctl->lock);
+       while (dq < RTE_DIM(ctl->dq_map)) {
+               if (ctl->dq_map[dq].chanid == ~chanid) {
+                       ctl->dq_map[dq].chanid = ~null_chanid;
+                       dq_cnt++;
+               }
+               dq++;
+       }
+       rte_spinlock_unlock(&ctl->lock);
+
+       return dq_cnt > 0 ? 0 : -EINVAL;
+}
+
+int
+octeontx_pko_channel_open(int dq_base, int dq_num, int chanid)
+{
+       struct octeontx_pko_vf_ctl_s *ctl = &pko_vf_ctl;
+       int res;
+
+       res = octeontx_pko_dq_claim(ctl, dq_base, dq_num, chanid);
+       if (res < 0)
+               return -1;
+
+       return 0;
+}
+
+int
+octeontx_pko_channel_close(int chanid)
+{
+       struct octeontx_pko_vf_ctl_s *ctl = &pko_vf_ctl;
+       int res;
+
+       res = octeontx_pko_dq_free(ctl, chanid);
+       if (res < 0)
+               return -1;
+
+       return 0;
+}
+
+static inline int
+octeontx_pko_chan_start(struct octeontx_pko_vf_ctl_s *ctl, uint64_t chanid)
+{
+       unsigned int dq_vf;
+       unsigned int dq, dq_cnt;
+
+       dq_cnt = 0;
+       dq = 0;
+       while (dq < RTE_DIM(ctl->dq_map)) {
+               dq_vf = dq / PKO_VF_NUM_DQ;
+
+               if (!ctl->pko[dq_vf].bar0) {
+                       dq += PKO_VF_NUM_DQ;
+                       continue;
+               }
+
+               if (ctl->dq_map[dq].chanid != ~chanid) {
+                       dq++;
+                       continue;
+               }
+
+               if (octeontx_pko_dq_open(dq) < 0)
+                       break;
+
+               dq_cnt++;
+               dq++;
+       }
+
+       return dq_cnt;
+}
+
+int
+octeontx_pko_channel_start(int chanid)
+{
+       struct octeontx_pko_vf_ctl_s *ctl = &pko_vf_ctl;
+       int dq_cnt;
+
+       dq_cnt = octeontx_pko_chan_start(ctl, chanid);
+       if (dq_cnt < 0)
+               return -1;
+
+       return dq_cnt;
+}
+
+static inline int
+octeontx_pko_chan_stop(struct octeontx_pko_vf_ctl_s *ctl, uint64_t chanid)
+{
+       unsigned int dq, dq_cnt, dq_vf;
+       int res;
+
+       dq_cnt = 0;
+       dq = 0;
+       while (dq < RTE_DIM(ctl->dq_map)) {
+               dq_vf = dq / PKO_VF_NUM_DQ;
+
+               if (!ctl->pko[dq_vf].bar0) {
+                       dq += PKO_VF_NUM_DQ;
+                       continue;
+               }
+
+               if (ctl->dq_map[dq].chanid != ~chanid) {
+                       dq++;
+                       continue;
+               }
+
+               res = octeontx_pko_dq_drain(dq);
+               if (res > 0)
+                       octeontx_log_err("draining DQ%d, buffers left: %x",
+                                        dq, res);
+
+               res = octeontx_pko_dq_close(dq);
+               if (res < 0)
+                       octeontx_log_err("closing DQ%d failed\n", dq);
+
+               dq_cnt++;
+               dq++;
+       }
+       return dq_cnt;
+}
+
+int
+octeontx_pko_channel_stop(int chanid)
+{
+       struct octeontx_pko_vf_ctl_s *ctl = &pko_vf_ctl;
+
+       octeontx_pko_chan_stop(ctl, chanid);
+       return 0;
+}
+
+static inline int
+octeontx_pko_channel_query(struct octeontx_pko_vf_ctl_s *ctl, uint64_t chanid,
+                          void *out, size_t out_elem_size,
+                          size_t dq_num, octeontx_pko_dq_getter_t getter)
+{
+       octeontx_dq_t curr;
+       unsigned int dq_vf;
+       unsigned int dq;
+
+       RTE_SET_USED(out_elem_size);
+       memset(&curr, 0, sizeof(octeontx_dq_t));
+
+       dq_vf = dq_num / PKO_VF_NUM_DQ;
+       dq = dq_num % PKO_VF_NUM_DQ;
+
+       if (!ctl->pko[dq_vf].bar0)
+               return -EINVAL;
+
+       if (ctl->dq_map[dq_num].chanid != ~chanid)
+               return -EINVAL;
+
+       uint8_t *iter = (uint8_t *)out;
+       curr.lmtline_va = ctl->pko[dq_vf].bar2;
+       curr.ioreg_va = (void *)((uintptr_t)ctl->pko[dq_vf].bar0
+               + PKO_VF_DQ_OP_SEND((dq), 0));
+       curr.fc_status_va = ctl->fc_ctl + dq_num;
+
+       octeontx_log_dbg("lmtline=%p ioreg_va=%p fc_status_va=%p",
+                        curr.lmtline_va, curr.ioreg_va,
+                        curr.fc_status_va);
+
+       getter(&curr, (void *)iter);
+       return 0;
+}
+
+int
+octeontx_pko_channel_query_dqs(int chanid, void *out, size_t out_elem_size,
+                               size_t dq_num, octeontx_pko_dq_getter_t getter)
+{
+       struct octeontx_pko_vf_ctl_s *ctl = &pko_vf_ctl;
+       int dq_cnt;
+
+       dq_cnt = octeontx_pko_channel_query(ctl, chanid, out, out_elem_size,
+                                               dq_num, getter);
+       if (dq_cnt < 0)
+               return -1;
+
+       return dq_cnt;
+}
+
+int
+octeontx_pko_vf_count(void)
+{
+       uint16_t global_domain = octeontx_get_global_domain();
+       int vf_cnt;
+
+       pko_vf_ctl.global_domain = global_domain;
+       vf_cnt = 0;
+       while (pko_vf_ctl.pko[vf_cnt].bar0)
+               vf_cnt++;
+
+       return vf_cnt;
+}
+
+size_t
+octeontx_pko_get_vfid(void)
+{
+       size_t vf_cnt = octeontx_pko_vf_count();
+       size_t vf_idx;
+
+
+       for (vf_idx = 0; vf_idx < vf_cnt; vf_idx++) {
+               if (!(pko_vf_ctl.pko[vf_idx].status & PKO_VALID))
+                       continue;
+               if (pko_vf_ctl.pko[vf_idx].status & PKO_INUSE)
+                       continue;
+
+               pko_vf_ctl.pko[vf_idx].status |= PKO_INUSE;
+               return pko_vf_ctl.pko[vf_idx].vfid;
+       }
+
+       return SIZE_MAX;
+}
+
+int
+octeontx_pko_send_mtu(int port, int mtu)
+{
+       struct octeontx_mbox_hdr hdr;
+       int res;
+       mbox_pko_mtu_cfg_t cfg;
+
+       cfg.mtu = mtu;
+
+       hdr.coproc = OCTEONTX_PKO_COPROC;
+       hdr.msg = MBOX_PKO_MTU_CONFIG;
+       hdr.vfid = port;
+
+       res = octeontx_mbox_send(&hdr, &cfg, sizeof(mbox_pko_mtu_cfg_t),
+                                NULL, 0);
+       if (res < 0)
+               return -EACCES;
+
+       return res;
+}
+
+int
+octeontx_pko_init_fc(const size_t pko_vf_count)
+{
+       int dq_ix;
+       uint64_t reg;
+       uint8_t *vf_bar0;
+       size_t vf_idx;
+       size_t fc_mem_size;
+
+       fc_mem_size = sizeof(struct octeontx_pko_fc_ctl_s) *
+                       pko_vf_count * PKO_VF_NUM_DQ;
+
+       pko_vf_ctl.fc_iomem.va = rte_malloc(NULL, fc_mem_size, 128);
+       if (unlikely(!pko_vf_ctl.fc_iomem.va)) {
+               octeontx_log_err("fc_iomem: not enough memory");
+               return -ENOMEM;
+       }
+
+       pko_vf_ctl.fc_iomem.iova = rte_malloc_virt2iova((void *)
+                                                       pko_vf_ctl.fc_iomem.va);
+       pko_vf_ctl.fc_iomem.size = fc_mem_size;
+
+       pko_vf_ctl.fc_ctl =
+               (struct octeontx_pko_fc_ctl_s *)pko_vf_ctl.fc_iomem.va;
+
+       /* Configure Flow-Control feature for all DQs of open VFs */
+       for (vf_idx = 0; vf_idx < pko_vf_count; vf_idx++) {
+               if (pko_vf_ctl.pko[vf_idx].domain != pko_vf_ctl.global_domain)
+                       continue;
+
+               dq_ix = pko_vf_ctl.pko[vf_idx].vfid * PKO_VF_NUM_DQ;
+               vf_bar0 = pko_vf_ctl.pko[vf_idx].bar0;
+
+               reg = (pko_vf_ctl.fc_iomem.iova +
+                       (sizeof(struct octeontx_pko_fc_ctl_s) * dq_ix)) & ~0x7F;
+               reg |=                  /* BASE */
+                   (0x2 << 3) |        /* HYST_BITS */
+                   (((PKO_DQ_FC_STRIDE == PKO_DQ_FC_STRIDE_16) ? 1 : 0) << 2) |
+                   (0x1 << 0);         /* ENABLE */
+
+               octeontx_write64(reg, vf_bar0 + PKO_VF_DQ_FC_CONFIG);
+               pko_vf_ctl.pko[vf_idx].status = PKO_VALID;
+
+               octeontx_log_dbg("PKO: bar0 %p VF_idx %d DQ_FC_CFG=%" PRIx64 "",
+                                vf_bar0, (int)vf_idx, reg);
+       }
+       return 0;
+}
+
+void
+octeontx_pko_fc_free(void)
+{
+       rte_free(pko_vf_ctl.fc_iomem.va);
+}
+
 static void
 octeontx_pkovf_setup(void)
 {
@@ -113,6 +577,7 @@ pkovf_probe(struct rte_pci_driver *pci_drv, struct rte_pci_device *pci_dev)
        uint16_t domain;
        uint8_t *bar0;
        uint8_t *bar2;
+       static uint8_t vf_cnt;
        struct octeontx_pkovf *res;
 
        RTE_SET_USED(pci_drv);
@@ -143,7 +608,7 @@ pkovf_probe(struct rte_pci_driver *pci_drv, struct rte_pci_device *pci_dev)
                return -EINVAL;
        }
 
-       res = &pko_vf_ctl.pko[vfid];
+       res = &pko_vf_ctl.pko[vf_cnt++];
        res->vfid = vfid;
        res->domain = domain;
        res->bar0 = bar0;