regex/octeontx2: introduce REE driver
authorGuy Kaneti <guyk@marvell.com>
Tue, 13 Oct 2020 10:10:10 +0000 (13:10 +0300)
committerThomas Monjalon <thomas@monjalon.net>
Wed, 14 Oct 2020 08:41:21 +0000 (10:41 +0200)
Add meson based build infrastructure along with the
OTX2 regexdev (REE) device functions.
Add Marvell OCTEON TX2 regex guide.

Signed-off-by: Guy Kaneti <guyk@marvell.com>
17 files changed:
MAINTAINERS
doc/guides/platform/octeontx2.rst
doc/guides/regexdevs/features/octeontx2.ini [new file with mode: 0644]
doc/guides/regexdevs/index.rst
doc/guides/regexdevs/octeontx2.rst [new file with mode: 0644]
doc/guides/rel_notes/release_20_11.rst
drivers/regex/meson.build
drivers/regex/octeontx2/meson.build [new file with mode: 0644]
drivers/regex/octeontx2/otx2_regexdev.c [new file with mode: 0644]
drivers/regex/octeontx2/otx2_regexdev.h [new file with mode: 0644]
drivers/regex/octeontx2/otx2_regexdev_compiler.c [new file with mode: 0644]
drivers/regex/octeontx2/otx2_regexdev_compiler.h [new file with mode: 0644]
drivers/regex/octeontx2/otx2_regexdev_hw_access.c [new file with mode: 0644]
drivers/regex/octeontx2/otx2_regexdev_hw_access.h [new file with mode: 0644]
drivers/regex/octeontx2/otx2_regexdev_mbox.c [new file with mode: 0644]
drivers/regex/octeontx2/otx2_regexdev_mbox.h [new file with mode: 0644]
drivers/regex/octeontx2/rte_pmd_octeontx2_regex_version.map [new file with mode: 0644]

index cdaf0b0..499c7d2 100644 (file)
@@ -1112,6 +1112,12 @@ F: doc/guides/compressdevs/features/zlib.ini
 RegEx Drivers
 -------------
 
+Marvell OCTEON TX2 regex
+M: Guy Kaneti <guyk@marvell.com>
+F: drivers/regex/octeontx2/
+F: doc/guides/regexdevs/octeontx2.rst
+F: doc/guides/regexdevs/features/octeontx2.ini
+
 Mellanox mlx5
 M: Ori Kam <orika@nvidia.com>
 F: drivers/regex/mlx5/
index 13255ee..c4d64ab 100644 (file)
@@ -67,6 +67,8 @@ DPDK subsystem.
    +---+-----+--------------------------------------------------------------+
    | 9 | SDP | rte_ethdev                                                   |
    +---+-----+--------------------------------------------------------------+
+   | 10| REE | rte_regexdev                                                 |
+   +---+-----+--------------------------------------------------------------+
 
 PF0 is called the administrative / admin function (AF) and has exclusive
 privileges to provision RVU functional block's LFs to each of the PF/VF.
@@ -156,6 +158,9 @@ This section lists dataplane H/W block(s) available in OCTEON TX2 SoC.
 #. **Crypto Device Driver**
    See :doc:`../cryptodevs/octeontx2` for CPT crypto device driver information.
 
+#. **Regex Device Driver**
+   See :doc:`../regexdevs/octeontx2` for REE regex device driver information.
+
 Procedure to Setup Platform
 ---------------------------
 
diff --git a/doc/guides/regexdevs/features/octeontx2.ini b/doc/guides/regexdevs/features/octeontx2.ini
new file mode 100644 (file)
index 0000000..c9b421a
--- /dev/null
@@ -0,0 +1,10 @@
+;
+; Supported features of the 'octeontx2' regex driver.
+;
+; Refer to default.ini for the full list of available driver features.
+;
+[Features]
+PCRE back reference         = Y
+PCRE word boundary          = Y
+Run time compilation        = Y
+Armv8                       = Y
index 49216a9..b1abc82 100644 (file)
@@ -13,3 +13,4 @@ which can be used from an application through RegEx API.
 
    features_overview
    mlx5
+   octeontx2
diff --git a/doc/guides/regexdevs/octeontx2.rst b/doc/guides/regexdevs/octeontx2.rst
new file mode 100644 (file)
index 0000000..5a0e53e
--- /dev/null
@@ -0,0 +1,38 @@
+..  SPDX-License-Identifier: BSD-3-Clause
+    Copyright(c) 2020 Marvell International Ltd.
+
+OCTEON TX2 REE Regexdev Driver
+==============================
+
+The OCTEON TX2 REE PMD (**librte_pmd_octeontx2_regex**) provides poll mode
+regexdev driver support for the inbuilt regex device found in the **Marvell OCTEON TX2**
+SoC family.
+
+More information about OCTEON TX2 SoC can be found at `Marvell Official Website
+<https://www.marvell.com/embedded-processors/infrastructure-processors/>`_.
+
+Features
+--------
+
+Features of the OCTEON TX2 REE PMD are:
+
+- 36 queues
+- Up to 254 matches for each regex operation
+
+Prerequisites and Compilation procedure
+---------------------------------------
+
+   See :doc:`../platform/octeontx2` for setup information.
+
+Debugging Options
+-----------------
+
+.. _table_octeontx2_regex_debug_options:
+
+.. table:: OCTEON TX2 regex device debug options
+
+   +---+------------+-------------------------------------------------------+
+   | # | Component  | EAL log command                                       |
+   +===+============+=======================================================+
+   | 1 | REE        | --log-level='pmd\.regex\.octeontx2,8'                 |
+   +---+------------+-------------------------------------------------------+
index 71665c1..683dd38 100644 (file)
@@ -146,6 +146,12 @@ New Features
     ``--portmask=N``
     where N represents the hexadecimal bitmask of ports used.
 
+* **Added Marvell OCTEON TX2 regex PMD.**
+
+  Added a new PMD driver for hardware regex offload block for OCTEON TX2 SoC.
+
+  See the :doc:`../regexdevs/octeontx2` for more details.
+
 * **Updated ioat rawdev driver**
 
   The ioat rawdev driver has been updated and enhanced. Changes include:
index 8edeba3..79bb5d5 100644 (file)
@@ -1,7 +1,7 @@
 # SPDX-License-Identifier: BSD-3-Clause
 # Copyright 2020 Mellanox Technologies, Ltd
 
-drivers = ['mlx5']
+drivers = ['mlx5', 'octeontx2']
 std_deps = ['ethdev', 'kvargs'] # 'ethdev' also pulls in mbuf, net, eal etc
 config_flag_fmt = 'RTE_LIBRTE_@0@_PMD'
 driver_name_fmt = 'rte_pmd_@0@'
diff --git a/drivers/regex/octeontx2/meson.build b/drivers/regex/octeontx2/meson.build
new file mode 100644 (file)
index 0000000..398a981
--- /dev/null
@@ -0,0 +1,44 @@
+# SPDX-License-Identifier: BSD-3-Clause
+# Copyright(C) 2020 Marvell International Ltd.
+#
+
+if not is_linux
+       build = false
+       reason = 'only supported on Linux'
+endif
+
+lib = cc.find_library('librxp_compiler', required: false)
+if lib.found()
+       ext_deps += lib
+       ext_deps += cc.find_library('libstdc++', required: true)
+       includes += include_directories(inc_dir)
+       cflags += ['-DREE_COMPILER_SDK']
+endif
+
+sources = files('otx2_regexdev.c',
+               'otx2_regexdev_hw_access.c',
+               'otx2_regexdev_mbox.c',
+               'otx2_regexdev_compiler.c'
+               )
+
+extra_flags = []
+# This integrated controller runs only on a arm64 machine, remove 32bit warnings
+if not dpdk_conf.get('RTE_ARCH_64')
+       extra_flags += ['-Wno-int-to-pointer-cast', '-Wno-pointer-to-int-cast']
+endif
+
+# for clang 32-bit compiles we need libatomic for 64-bit atomic ops
+if cc.get_id() == 'clang' and dpdk_conf.get('RTE_ARCH_64') == false
+       ext_deps += cc.find_library('atomic')
+endif
+
+foreach flag: extra_flags
+       if cc.has_argument(flag)
+               cflags += flag
+       endif
+endforeach
+
+name = 'octeontx2_regex'
+deps += ['bus_pci', 'common_octeontx2', 'regexdev']
+
+includes += include_directories('../../common/octeontx2')
diff --git a/drivers/regex/octeontx2/otx2_regexdev.c b/drivers/regex/octeontx2/otx2_regexdev.c
new file mode 100644 (file)
index 0000000..39eed7a
--- /dev/null
@@ -0,0 +1,1002 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright (C) 2020 Marvell International Ltd.
+ */
+
+#include <stdio.h>
+#include <unistd.h>
+
+#include <rte_malloc.h>
+#include <rte_memzone.h>
+#include <rte_regexdev.h>
+#include <rte_regexdev_core.h>
+#include <rte_regexdev_driver.h>
+
+
+/* REE common headers */
+#include "otx2_common.h"
+#include "otx2_dev.h"
+#include "otx2_regexdev.h"
+#include "otx2_regexdev_compiler.h"
+#include "otx2_regexdev_hw_access.h"
+#include "otx2_regexdev_mbox.h"
+
+
+/* HW matches are at offset 0x80 from RES_PTR_ADDR
+ * In op structure matches starts at W5 (0x28)
+ * There is a need to copy to 0x28 to 0x80 The matches that are at the tail
+ * Which are 88 B. Each match holds 8 B, so up to 11 matches can be copied
+ */
+#define REE_NUM_MATCHES_ALIGN  11
+/* The REE co-processor will write up to 254 job match structures
+ * (REE_MATCH_S) starting at address [RES_PTR_ADDR] + 0x80.
+ */
+#define REE_MATCH_OFFSET       0x80
+
+#define REE_MAX_RULES_PER_GROUP 0xFFFF
+#define REE_MAX_GROUPS 0xFFFF
+
+/* This is temporarily here */
+#define REE0_PF        19
+#define REE1_PF        20
+
+#define REE_RULE_DB_VERSION    2
+#define REE_RULE_DB_REVISION   0
+
+struct ree_rule_db_entry {
+       uint8_t         type;
+       uint32_t        addr;
+       uint64_t        value;
+};
+
+struct ree_rule_db {
+       uint32_t version;
+       uint32_t revision;
+       uint32_t number_of_entries;
+       struct ree_rule_db_entry entries[];
+} __rte_packed;
+
+static void
+qp_memzone_name_get(char *name, int size, int dev_id, int qp_id)
+{
+       snprintf(name, size, "otx2_ree_lf_mem_%u:%u", dev_id, qp_id);
+}
+
+static struct otx2_ree_qp *
+ree_qp_create(const struct rte_regexdev *dev, uint16_t qp_id)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       uint64_t pg_sz = sysconf(_SC_PAGESIZE);
+       struct otx2_ree_vf *vf = &data->vf;
+       const struct rte_memzone *lf_mem;
+       uint32_t len, iq_len, size_div2;
+       char name[RTE_MEMZONE_NAMESIZE];
+       uint64_t used_len, iova;
+       struct otx2_ree_qp *qp;
+       uint8_t *va;
+       int ret;
+
+       /* Allocate queue pair */
+       qp = rte_zmalloc("OCTEON TX2 Regex PMD Queue Pair", sizeof(*qp),
+                               OTX2_ALIGN);
+       if (qp == NULL) {
+               otx2_err("Could not allocate queue pair");
+               return NULL;
+       }
+
+       iq_len = OTX2_REE_IQ_LEN;
+
+       /*
+        * Queue size must be in units of 128B 2 * REE_INST_S (which is 64B),
+        * and a power of 2.
+        * effective queue size to software is (size - 1) * 128
+        */
+       size_div2 = iq_len >> 1;
+
+       /* For pending queue */
+       len = iq_len * RTE_ALIGN(sizeof(struct otx2_ree_rid), 8);
+
+       /* So that instruction queues start as pg size aligned */
+       len = RTE_ALIGN(len, pg_sz);
+
+       /* For instruction queues */
+       len += OTX2_REE_IQ_LEN * sizeof(union otx2_ree_inst);
+
+       /* Waste after instruction queues */
+       len = RTE_ALIGN(len, pg_sz);
+
+       qp_memzone_name_get(name, RTE_MEMZONE_NAMESIZE, dev->data->dev_id,
+                           qp_id);
+
+       lf_mem = rte_memzone_reserve_aligned(name, len, vf->otx2_dev.node,
+                       RTE_MEMZONE_SIZE_HINT_ONLY | RTE_MEMZONE_256MB,
+                       RTE_CACHE_LINE_SIZE);
+       if (lf_mem == NULL) {
+               otx2_err("Could not allocate reserved memzone");
+               goto qp_free;
+       }
+
+       va = lf_mem->addr;
+       iova = lf_mem->iova;
+
+       memset(va, 0, len);
+
+       /* Initialize pending queue */
+       qp->pend_q.rid_queue = (struct otx2_ree_rid *)va;
+       qp->pend_q.enq_tail = 0;
+       qp->pend_q.deq_head = 0;
+       qp->pend_q.pending_count = 0;
+
+       used_len = iq_len * RTE_ALIGN(sizeof(struct otx2_ree_rid), 8);
+       used_len = RTE_ALIGN(used_len, pg_sz);
+       iova += used_len;
+
+       qp->iq_dma_addr = iova;
+       qp->id = qp_id;
+       qp->base = OTX2_REE_LF_BAR2(vf, qp_id);
+       qp->otx2_regexdev_jobid = 0;
+       qp->write_offset = 0;
+
+       ret = otx2_ree_iq_enable(dev, qp, OTX2_REE_QUEUE_HI_PRIO, size_div2);
+       if (ret) {
+               otx2_err("Could not enable instruction queue");
+               goto qp_free;
+       }
+
+       return qp;
+
+qp_free:
+       rte_free(qp);
+       return NULL;
+}
+
+static int
+ree_qp_destroy(const struct rte_regexdev *dev, struct otx2_ree_qp *qp)
+{
+       const struct rte_memzone *lf_mem;
+       char name[RTE_MEMZONE_NAMESIZE];
+       int ret;
+
+       otx2_ree_iq_disable(qp);
+
+       qp_memzone_name_get(name, RTE_MEMZONE_NAMESIZE, dev->data->dev_id,
+                           qp->id);
+
+       lf_mem = rte_memzone_lookup(name);
+
+       ret = rte_memzone_free(lf_mem);
+       if (ret)
+               return ret;
+
+       rte_free(qp);
+
+       return 0;
+}
+
+static int
+ree_queue_pair_release(struct rte_regexdev *dev, uint16_t qp_id)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct otx2_ree_qp *qp = data->queue_pairs[qp_id];
+       int ret;
+
+       ree_func_trace("Queue=%d", qp_id);
+
+       if (qp == NULL)
+               return -EINVAL;
+
+       ret = ree_qp_destroy(dev, qp);
+       if (ret) {
+               otx2_err("Could not destroy queue pair %d", qp_id);
+               return ret;
+       }
+
+       data->queue_pairs[qp_id] = NULL;
+
+       return 0;
+}
+
+static struct rte_regexdev *
+ree_dev_register(const char *name)
+{
+       struct rte_regexdev *dev;
+
+       otx2_ree_dbg("Creating regexdev %s\n", name);
+
+       /* allocate device structure */
+       dev = rte_regexdev_register(name);
+       if (dev == NULL) {
+               otx2_err("Failed to allocate regex device for %s", name);
+               return NULL;
+       }
+
+       /* allocate private device structure */
+       if (rte_eal_process_type() == RTE_PROC_PRIMARY) {
+               dev->data->dev_private =
+                               rte_zmalloc_socket("regexdev device private",
+                                               sizeof(struct otx2_ree_data),
+                                               RTE_CACHE_LINE_SIZE,
+                                               rte_socket_id());
+
+               if (dev->data->dev_private == NULL) {
+                       otx2_err("Cannot allocate memory for dev %s private data",
+                                       name);
+
+                       rte_regexdev_unregister(dev);
+                       return NULL;
+               }
+       }
+
+       return dev;
+}
+
+static int
+ree_dev_unregister(struct rte_regexdev *dev)
+{
+       otx2_ree_dbg("Closing regex device %s", dev->device->name);
+
+       /* free regex device */
+       rte_regexdev_unregister(dev);
+
+       if (rte_eal_process_type() == RTE_PROC_PRIMARY)
+               rte_free(dev->data->dev_private);
+
+       return 0;
+}
+
+static int
+ree_dev_fini(struct rte_regexdev *dev)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct rte_pci_device *pci_dev;
+       int i, ret;
+
+       ree_func_trace();
+
+       for (i = 0; i < data->nb_queue_pairs; i++) {
+               ret = ree_queue_pair_release(dev, i);
+               if (ret)
+                       return ret;
+       }
+
+       ret = otx2_ree_queues_detach(dev);
+       if (ret)
+               otx2_err("Could not detach queues");
+
+       /* TEMP : should be in lib */
+       if (data->queue_pairs)
+               rte_free(data->queue_pairs);
+       if (data->rules)
+               rte_free(data->rules);
+
+       pci_dev = container_of(dev->device, struct rte_pci_device, device);
+       otx2_dev_fini(pci_dev, &(data->vf.otx2_dev));
+
+       ret = ree_dev_unregister(dev);
+       if (ret)
+               otx2_err("Could not destroy PMD");
+
+       return ret;
+}
+
+static inline int
+ree_enqueue(struct otx2_ree_qp *qp, struct rte_regex_ops *op,
+                struct otx2_ree_pending_queue *pend_q)
+{
+       union otx2_ree_inst inst;
+       union otx2_ree_res *res;
+       uint32_t offset;
+
+       if (unlikely(pend_q->pending_count >= OTX2_REE_DEFAULT_CMD_QLEN)) {
+               otx2_err("Pending count %" PRIu64 " is greater than Q size %d",
+               pend_q->pending_count, OTX2_REE_DEFAULT_CMD_QLEN);
+               return -EAGAIN;
+       }
+       if (unlikely(op->mbuf->data_len > OTX2_REE_MAX_PAYLOAD_SIZE ||
+                       op->mbuf->data_len == 0)) {
+               otx2_err("Packet length %d is greater than MAX payload %d",
+                               op->mbuf->data_len, OTX2_REE_MAX_PAYLOAD_SIZE);
+               return -EAGAIN;
+       }
+
+       /* W 0 */
+       inst.cn98xx.ooj = 1;
+       inst.cn98xx.dg = 0;
+       inst.cn98xx.doneint = 0;
+       /* W 1 */
+       inst.cn98xx.inp_ptr_addr = rte_pktmbuf_mtod(op->mbuf, uint64_t);
+       /* W 2 */
+       inst.cn98xx.inp_ptr_ctl = op->mbuf->data_len & 0x7FFF;
+       inst.cn98xx.inp_ptr_ctl = inst.cn98xx.inp_ptr_ctl << 32;
+
+       /* W 3 */
+       inst.cn98xx.res_ptr_addr = (uint64_t)op;
+       /* W 4 */
+       inst.cn98xx.wq_ptr = 0;
+       /* W 5 */
+       inst.cn98xx.ggrp = 0;
+       inst.cn98xx.tt = 0;
+       inst.cn98xx.tag = 0;
+       /* W 6 */
+       inst.cn98xx.ree_job_length = op->mbuf->data_len & 0x7FFF;
+       if (op->req_flags & RTE_REGEX_OPS_REQ_STOP_ON_MATCH_F)
+               inst.cn98xx.ree_job_ctrl = (0x2 << 8);
+       else if (op->req_flags & RTE_REGEX_OPS_REQ_MATCH_HIGH_PRIORITY_F)
+               inst.cn98xx.ree_job_ctrl = (0x1 << 8);
+       else
+               inst.cn98xx.ree_job_ctrl = 0;
+       inst.cn98xx.ree_job_id = qp->otx2_regexdev_jobid;
+       /* W 7 */
+       inst.cn98xx.ree_job_subset_id_0 = op->group_id0;
+       if (op->req_flags & RTE_REGEX_OPS_REQ_GROUP_ID1_VALID_F)
+               inst.cn98xx.ree_job_subset_id_1 = op->group_id1;
+       else
+               inst.cn98xx.ree_job_subset_id_1 = op->group_id0;
+       if (op->req_flags & RTE_REGEX_OPS_REQ_GROUP_ID2_VALID_F)
+               inst.cn98xx.ree_job_subset_id_2 = op->group_id2;
+       else
+               inst.cn98xx.ree_job_subset_id_2 = op->group_id0;
+       if (op->req_flags & RTE_REGEX_OPS_REQ_GROUP_ID3_VALID_F)
+               inst.cn98xx.ree_job_subset_id_3 = op->group_id3;
+       else
+               inst.cn98xx.ree_job_subset_id_3 = op->group_id0;
+
+       /* Copy REE command to Q */
+       offset = qp->write_offset * sizeof(inst);
+       memcpy((void *)(qp->iq_dma_addr + offset), &inst, sizeof(inst));
+
+       pend_q->rid_queue[pend_q->enq_tail].rid = (uintptr_t)op;
+       pend_q->rid_queue[pend_q->enq_tail].user_id = op->user_id;
+
+       /* Mark result as not done */
+       res = (union otx2_ree_res *)(op);
+       res->s.done = 0;
+       res->s.ree_err = 0;
+
+       /* We will use soft queue length here to limit requests */
+       REE_MOD_INC(pend_q->enq_tail, OTX2_REE_DEFAULT_CMD_QLEN);
+       pend_q->pending_count += 1;
+       REE_MOD_INC(qp->otx2_regexdev_jobid, 0xFFFFFF);
+       REE_MOD_INC(qp->write_offset, OTX2_REE_IQ_LEN);
+
+       return 0;
+}
+
+static uint16_t
+otx2_ree_enqueue_burst(struct rte_regexdev *dev, uint16_t qp_id,
+                      struct rte_regex_ops **ops, uint16_t nb_ops)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct otx2_ree_qp *qp = data->queue_pairs[qp_id];
+       struct otx2_ree_pending_queue *pend_q;
+       uint16_t nb_allowed, count = 0;
+       struct rte_regex_ops *op;
+       int ret;
+
+       pend_q = &qp->pend_q;
+
+       nb_allowed = OTX2_REE_DEFAULT_CMD_QLEN - pend_q->pending_count;
+       if (nb_ops > nb_allowed)
+               nb_ops = nb_allowed;
+
+       for (count = 0; count < nb_ops; count++) {
+               op = ops[count];
+               ret = ree_enqueue(qp, op, pend_q);
+
+               if (unlikely(ret))
+                       break;
+       }
+
+       /*
+        * Make sure all instructions are written before DOORBELL is activated
+        */
+       rte_io_wmb();
+
+       /* Update Doorbell */
+       otx2_write64(count, qp->base + OTX2_REE_LF_DOORBELL);
+
+       return count;
+}
+
+static inline void
+ree_dequeue_post_process(struct rte_regex_ops *ops)
+{
+       uint8_t ree_res_mcnt, ree_res_dmcnt;
+       int off = REE_MATCH_OFFSET;
+       struct ree_res_s_98 *res;
+       uint16_t ree_res_status;
+       uint64_t match;
+
+       res = (struct ree_res_s_98 *)ops;
+       /* store res values on stack since ops and res
+        * are using the same memory
+        */
+       ree_res_status = res->ree_res_status;
+       ree_res_mcnt = res->ree_res_mcnt;
+       ree_res_dmcnt = res->ree_res_dmcnt;
+       ops->rsp_flags = 0;
+       ops->nb_actual_matches = ree_res_dmcnt;
+       ops->nb_matches = ree_res_mcnt;
+       if (unlikely(res->ree_err)) {
+               ops->nb_actual_matches = 0;
+               ops->nb_matches = 0;
+       }
+
+       if (unlikely(ree_res_status != REE_TYPE_RESULT_DESC)) {
+               if (ree_res_status & OTX2_REE_STATUS_PMI_SOJ_BIT)
+                       ops->rsp_flags |= RTE_REGEX_OPS_RSP_PMI_SOJ_F;
+               if (ree_res_status & OTX2_REE_STATUS_PMI_EOJ_BIT)
+                       ops->rsp_flags |= RTE_REGEX_OPS_RSP_PMI_EOJ_F;
+               if (ree_res_status & OTX2_REE_STATUS_ML_CNT_DET_BIT)
+                       ops->rsp_flags |= RTE_REGEX_OPS_RSP_MAX_SCAN_TIMEOUT_F;
+               if (ree_res_status & OTX2_REE_STATUS_MM_CNT_DET_BIT)
+                       ops->rsp_flags |= RTE_REGEX_OPS_RSP_MAX_MATCH_F;
+               if (ree_res_status & OTX2_REE_STATUS_MP_CNT_DET_BIT)
+                       ops->rsp_flags |= RTE_REGEX_OPS_RSP_MAX_PREFIX_F;
+       }
+       if (ops->nb_matches > 0) {
+               /* Move the matches to the correct offset */
+               off = ((ops->nb_matches < REE_NUM_MATCHES_ALIGN) ?
+                       ops->nb_matches : REE_NUM_MATCHES_ALIGN);
+               match = (uint64_t)ops + REE_MATCH_OFFSET;
+               match += (ops->nb_matches - off) *
+                       sizeof(union otx2_ree_match);
+               memcpy((void *)ops->matches, (void *)match,
+                       off * sizeof(union otx2_ree_match));
+       }
+}
+
+static uint16_t
+otx2_ree_dequeue_burst(struct rte_regexdev *dev, uint16_t qp_id,
+                      struct rte_regex_ops **ops, uint16_t nb_ops)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct otx2_ree_qp *qp = data->queue_pairs[qp_id];
+       struct otx2_ree_pending_queue *pend_q;
+       int i, nb_pending, nb_completed = 0;
+       volatile struct ree_res_s_98 *res;
+       struct otx2_ree_rid *rid;
+
+       pend_q = &qp->pend_q;
+
+       nb_pending = pend_q->pending_count;
+
+       if (nb_ops > nb_pending)
+               nb_ops = nb_pending;
+
+       for (i = 0; i < nb_ops; i++) {
+               rid = &pend_q->rid_queue[pend_q->deq_head];
+               res = (volatile struct ree_res_s_98 *)(rid->rid);
+
+               /* Check response header done bit if completed */
+               if (unlikely(!res->done))
+                       break;
+
+               ops[i] = (struct rte_regex_ops *)(rid->rid);
+               ops[i]->user_id = rid->user_id;
+
+               REE_MOD_INC(pend_q->deq_head, OTX2_REE_DEFAULT_CMD_QLEN);
+               pend_q->pending_count -= 1;
+       }
+
+       nb_completed = i;
+
+       for (i = 0; i < nb_completed; i++)
+               ree_dequeue_post_process(ops[i]);
+
+       return nb_completed;
+}
+
+static int
+otx2_ree_dev_info_get(struct rte_regexdev *dev, struct rte_regexdev_info *info)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct otx2_ree_vf *vf = &data->vf;
+
+       ree_func_trace();
+
+       if (info == NULL)
+               return -EINVAL;
+
+       info->driver_name = dev->device->driver->name;
+       info->dev = dev->device;
+
+       info->max_queue_pairs = vf->max_queues;
+       info->max_matches = vf->max_matches;
+       info->max_payload_size = OTX2_REE_MAX_PAYLOAD_SIZE;
+       info->max_rules_per_group = data->max_rules_per_group;
+       info->max_groups = data->max_groups;
+       info->regexdev_capa = data->regexdev_capa;
+       info->rule_flags = data->rule_flags;
+
+       return 0;
+}
+
+static int
+otx2_ree_dev_config(struct rte_regexdev *dev,
+                   const struct rte_regexdev_config *cfg)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct otx2_ree_vf *vf = &data->vf;
+       const struct ree_rule_db *rule_db;
+       uint32_t rule_db_len;
+       int ret;
+
+       ree_func_trace();
+
+       if (cfg->nb_queue_pairs > vf->max_queues) {
+               otx2_err("Invalid number of queue pairs requested");
+               return -EINVAL;
+       }
+
+       if (cfg->nb_max_matches != vf->max_matches) {
+               otx2_err("Invalid number of max matches requested");
+               return -EINVAL;
+       }
+
+       if (cfg->dev_cfg_flags != 0) {
+               otx2_err("Invalid device configuration flags requested");
+               return -EINVAL;
+       }
+
+       /* Unregister error interrupts */
+       if (vf->err_intr_registered)
+               otx2_ree_err_intr_unregister(dev);
+
+       /* Detach queues */
+       if (vf->nb_queues) {
+               ret = otx2_ree_queues_detach(dev);
+               if (ret) {
+                       otx2_err("Could not detach REE queues");
+                       return ret;
+               }
+       }
+
+       /* TEMP : should be in lib */
+       if (data->queue_pairs == NULL) { /* first time configuration */
+               data->queue_pairs = rte_zmalloc("regexdev->queue_pairs",
+                               sizeof(data->queue_pairs[0]) *
+                               cfg->nb_queue_pairs, RTE_CACHE_LINE_SIZE);
+
+               if (data->queue_pairs == NULL) {
+                       data->nb_queue_pairs = 0;
+                       otx2_err("Failed to get memory for qp meta data, nb_queues %u",
+                                       cfg->nb_queue_pairs);
+                       return -ENOMEM;
+               }
+       } else { /* re-configure */
+               uint16_t old_nb_queues = data->nb_queue_pairs;
+               void **qp;
+               unsigned int i;
+
+               qp = data->queue_pairs;
+
+               for (i = cfg->nb_queue_pairs; i < old_nb_queues; i++) {
+                       ret = ree_queue_pair_release(dev, i);
+                       if (ret < 0)
+                               return ret;
+               }
+
+               qp = rte_realloc(qp, sizeof(qp[0]) * cfg->nb_queue_pairs,
+                               RTE_CACHE_LINE_SIZE);
+               if (qp == NULL) {
+                       otx2_err("Failed to realloc qp meta data, nb_queues %u",
+                                       cfg->nb_queue_pairs);
+                       return -ENOMEM;
+               }
+
+               if (cfg->nb_queue_pairs > old_nb_queues) {
+                       uint16_t new_qs = cfg->nb_queue_pairs - old_nb_queues;
+                       memset(qp + old_nb_queues, 0, sizeof(qp[0]) * new_qs);
+               }
+
+               data->queue_pairs = qp;
+       }
+       data->nb_queue_pairs = cfg->nb_queue_pairs;
+
+       /* Attach queues */
+       otx2_ree_dbg("Attach %d queues", cfg->nb_queue_pairs);
+       ret = otx2_ree_queues_attach(dev, cfg->nb_queue_pairs);
+       if (ret) {
+               otx2_err("Could not attach queues");
+               return -ENODEV;
+       }
+
+       ret = otx2_ree_msix_offsets_get(dev);
+       if (ret) {
+               otx2_err("Could not get MSI-X offsets");
+               goto queues_detach;
+       }
+
+       if (cfg->rule_db && cfg->rule_db_len) {
+               otx2_ree_dbg("rule_db length %d", cfg->rule_db_len);
+               rule_db = (const struct ree_rule_db *)cfg->rule_db;
+               rule_db_len = rule_db->number_of_entries *
+                               sizeof(struct ree_rule_db_entry);
+               otx2_ree_dbg("rule_db number of entries %d",
+                               rule_db->number_of_entries);
+               if (rule_db_len > cfg->rule_db_len) {
+                       otx2_err("Could not program rule db");
+                       ret = -EINVAL;
+                       goto queues_detach;
+               }
+               ret = otx2_ree_rule_db_prog(dev, (const char *)rule_db->entries,
+                               rule_db_len, NULL, OTX2_REE_NON_INC_PROG);
+               if (ret) {
+                       otx2_err("Could not program rule db");
+                       goto queues_detach;
+               }
+       }
+
+       dev->enqueue = otx2_ree_enqueue_burst;
+       dev->dequeue = otx2_ree_dequeue_burst;
+
+       rte_mb();
+       return 0;
+
+queues_detach:
+       otx2_ree_queues_detach(dev);
+       return ret;
+}
+
+static int
+otx2_ree_stop(struct rte_regexdev *dev)
+{
+       RTE_SET_USED(dev);
+
+       ree_func_trace();
+       return 0;
+}
+
+static int
+otx2_ree_start(struct rte_regexdev *dev)
+{
+       uint32_t rule_db_len = 0;
+       int ret;
+
+       ree_func_trace();
+
+       ret = otx2_ree_rule_db_len_get(dev, &rule_db_len, NULL);
+       if (ret)
+               return ret;
+       if (rule_db_len == 0) {
+               otx2_err("Rule db not programmed");
+               return -EFAULT;
+       }
+
+       return 0;
+}
+
+static int
+otx2_ree_close(struct rte_regexdev *dev)
+{
+       return ree_dev_fini(dev);
+}
+
+static int
+otx2_ree_queue_pair_setup(struct rte_regexdev *dev, uint16_t qp_id,
+               const struct rte_regexdev_qp_conf *qp_conf)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct otx2_ree_qp *qp;
+
+       ree_func_trace("Queue=%d", qp_id);
+
+       if (data->queue_pairs[qp_id] != NULL)
+               ree_queue_pair_release(dev, qp_id);
+
+       if (qp_conf->nb_desc > OTX2_REE_DEFAULT_CMD_QLEN) {
+               otx2_err("Could not setup queue pair for %u descriptors",
+                               qp_conf->nb_desc);
+               return -EINVAL;
+       }
+       if (qp_conf->qp_conf_flags != 0) {
+               otx2_err("Could not setup queue pair with configuration flags 0x%x",
+                               qp_conf->qp_conf_flags);
+               return -EINVAL;
+       }
+
+       qp = ree_qp_create(dev, qp_id);
+       if (qp == NULL) {
+               otx2_err("Could not create queue pair %d", qp_id);
+               return -ENOMEM;
+       }
+       qp->cb = qp_conf->cb;
+       data->queue_pairs[qp_id] = qp;
+
+       return 0;
+}
+
+static int
+otx2_ree_rule_db_compile_activate(struct rte_regexdev *dev)
+{
+       return otx2_ree_rule_db_compile_prog(dev);
+}
+
+static int
+otx2_ree_rule_db_update(struct rte_regexdev *dev,
+               const struct rte_regexdev_rule *rules, uint16_t nb_rules)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct rte_regexdev_rule *old_ptr;
+       uint32_t i, sum_nb_rules;
+
+       ree_func_trace("nb_rules=%d", nb_rules);
+
+       for (i = 0; i < nb_rules; i++) {
+               if (rules[i].op == RTE_REGEX_RULE_OP_REMOVE)
+                       break;
+               if (rules[i].group_id >= data->max_groups)
+                       break;
+               if (rules[i].rule_id >= data->max_rules_per_group)
+                       break;
+               /* logical implication
+                * p    q    p -> q
+                * 0    0      1
+                * 0    1      1
+                * 1    0      0
+                * 1    1      1
+                */
+               if ((~(rules[i].rule_flags) | data->rule_flags) == 0)
+                       break;
+       }
+       nb_rules = i;
+
+       if (data->nb_rules == 0) {
+
+               data->rules = rte_malloc("rte_regexdev_rules",
+                               nb_rules*sizeof(struct rte_regexdev_rule), 0);
+               if (data->rules == NULL)
+                       return -ENOMEM;
+
+               memcpy(data->rules, rules,
+                               nb_rules*sizeof(struct rte_regexdev_rule));
+               data->nb_rules = nb_rules;
+       } else {
+
+               old_ptr = data->rules;
+               sum_nb_rules = data->nb_rules + nb_rules;
+               data->rules = rte_realloc(data->rules,
+                               sum_nb_rules * sizeof(struct rte_regexdev_rule),
+                                                       0);
+               if (data->rules == NULL) {
+                       data->rules = old_ptr;
+                       return -ENOMEM;
+               }
+               memcpy(&data->rules[data->nb_rules], rules,
+                               nb_rules*sizeof(struct rte_regexdev_rule));
+               data->nb_rules = sum_nb_rules;
+       }
+       return nb_rules;
+}
+
+static int
+otx2_ree_rule_db_import(struct rte_regexdev *dev, const char *rule_db,
+               uint32_t rule_db_len)
+{
+
+       const struct ree_rule_db *ree_rule_db;
+       uint32_t ree_rule_db_len;
+       int ret;
+
+       ree_func_trace("rule_db_len=%d", rule_db_len);
+
+       ree_rule_db = (const struct ree_rule_db *)rule_db;
+       ree_rule_db_len = ree_rule_db->number_of_entries *
+                       sizeof(struct ree_rule_db_entry);
+       if (ree_rule_db_len > rule_db_len) {
+               otx2_err("Could not program rule db");
+               return -EINVAL;
+       }
+       ret = otx2_ree_rule_db_prog(dev, (const char *)ree_rule_db->entries,
+                       ree_rule_db_len, NULL, OTX2_REE_NON_INC_PROG);
+       if (ret) {
+               otx2_err("Could not program rule db");
+               return -ENOSPC;
+       }
+       return 0;
+}
+
+static int
+otx2_ree_rule_db_export(struct rte_regexdev *dev, char *rule_db)
+{
+       struct ree_rule_db *ree_rule_db;
+       uint32_t rule_dbi_len;
+       uint32_t rule_db_len;
+       int ret;
+
+       ree_func_trace();
+
+       ret = otx2_ree_rule_db_len_get(dev, &rule_db_len, &rule_dbi_len);
+       if (ret)
+               return ret;
+
+       if (rule_db == NULL) {
+               rule_db_len += sizeof(struct ree_rule_db);
+               return rule_db_len;
+       }
+
+       ree_rule_db = (struct ree_rule_db *)rule_db;
+       ret = otx2_ree_rule_db_get(dev, (char *)ree_rule_db->entries,
+                       rule_db_len, NULL, 0);
+       if (ret) {
+               otx2_err("Could not export rule db");
+               return -EFAULT;
+       }
+       ree_rule_db->number_of_entries =
+                       rule_db_len/sizeof(struct ree_rule_db_entry);
+       ree_rule_db->revision = REE_RULE_DB_REVISION;
+       ree_rule_db->version = REE_RULE_DB_VERSION;
+
+       return 0;
+}
+
+static int
+ree_get_blkaddr(struct otx2_dev *dev)
+{
+       int pf;
+
+       pf = otx2_get_pf(dev->pf_func);
+       if (pf == REE0_PF)
+               return RVU_BLOCK_ADDR_REE0;
+       else if (pf == REE1_PF)
+               return RVU_BLOCK_ADDR_REE1;
+       else
+               return 0;
+}
+
+static struct rte_regexdev_ops otx2_ree_ops = {
+               .dev_info_get = otx2_ree_dev_info_get,
+               .dev_configure = otx2_ree_dev_config,
+               .dev_qp_setup = otx2_ree_queue_pair_setup,
+               .dev_start = otx2_ree_start,
+               .dev_stop = otx2_ree_stop,
+               .dev_close = otx2_ree_close,
+               .dev_attr_get = NULL,
+               .dev_attr_set = NULL,
+               .dev_rule_db_update = otx2_ree_rule_db_update,
+               .dev_rule_db_compile_activate =
+                               otx2_ree_rule_db_compile_activate,
+               .dev_db_import = otx2_ree_rule_db_import,
+               .dev_db_export = otx2_ree_rule_db_export,
+               .dev_xstats_names_get = NULL,
+               .dev_xstats_get = NULL,
+               .dev_xstats_by_name_get = NULL,
+               .dev_xstats_reset = NULL,
+               .dev_selftest = NULL,
+               .dev_dump = NULL,
+};
+
+static int
+otx2_ree_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
+                  struct rte_pci_device *pci_dev)
+{
+       char name[RTE_REGEXDEV_NAME_MAX_LEN];
+       struct otx2_ree_data *data;
+       struct otx2_dev *otx2_dev;
+       struct rte_regexdev *dev;
+       uint8_t max_matches = 0;
+       struct otx2_ree_vf *vf;
+       uint16_t nb_queues = 0;
+       int ret;
+
+       rte_pci_device_name(&pci_dev->addr, name, sizeof(name));
+
+       dev = ree_dev_register(name);
+       if (dev == NULL) {
+               ret = -ENODEV;
+               goto exit;
+       }
+
+       dev->dev_ops = &otx2_ree_ops;
+       dev->device = &pci_dev->device;
+
+       /* Get private data space allocated */
+       data = dev->data->dev_private;
+       vf = &data->vf;
+
+       otx2_dev = &vf->otx2_dev;
+
+       /* Initialize the base otx2_dev object */
+       ret = otx2_dev_init(pci_dev, otx2_dev);
+       if (ret) {
+               otx2_err("Could not initialize otx2_dev");
+               goto dev_unregister;
+       }
+       /* Get REE block address */
+       vf->block_address = ree_get_blkaddr(otx2_dev);
+       if (!vf->block_address) {
+               otx2_err("Could not determine block PF number");
+               goto otx2_dev_fini;
+       }
+
+       /* Get number of queues available on the device */
+       ret = otx2_ree_available_queues_get(dev, &nb_queues);
+       if (ret) {
+               otx2_err("Could not determine the number of queues available");
+               goto otx2_dev_fini;
+       }
+
+       /* Don't exceed the limits set per VF */
+       nb_queues = RTE_MIN(nb_queues, OTX2_REE_MAX_QUEUES_PER_VF);
+
+       if (nb_queues == 0) {
+               otx2_err("No free queues available on the device");
+               goto otx2_dev_fini;
+       }
+
+       vf->max_queues = nb_queues;
+
+       otx2_ree_dbg("Max queues supported by device: %d", vf->max_queues);
+
+       /* Get number of maximum matches supported on the device */
+       ret = otx2_ree_max_matches_get(dev, &max_matches);
+       if (ret) {
+               otx2_err("Could not determine the maximum matches supported");
+               goto otx2_dev_fini;
+       }
+       /* Don't exceed the limits set per VF */
+       max_matches = RTE_MIN(max_matches, OTX2_REE_MAX_MATCHES_PER_VF);
+       if (max_matches == 0) {
+               otx2_err("Could not determine the maximum matches supported");
+               goto otx2_dev_fini;
+       }
+
+       vf->max_matches = max_matches;
+
+       otx2_ree_dbg("Max matches supported by device: %d", vf->max_matches);
+       data->rule_flags = RTE_REGEX_PCRE_RULE_ALLOW_EMPTY_F |
+                       RTE_REGEX_PCRE_RULE_ANCHORED_F;
+       data->regexdev_capa = 0;
+       data->max_groups = REE_MAX_GROUPS;
+       data->max_rules_per_group = REE_MAX_RULES_PER_GROUP;
+       data->nb_rules = 0;
+
+       dev->state = RTE_REGEXDEV_READY;
+       return 0;
+
+otx2_dev_fini:
+       otx2_dev_fini(pci_dev, otx2_dev);
+dev_unregister:
+       ree_dev_unregister(dev);
+exit:
+       otx2_err("Could not create device (vendor_id: 0x%x device_id: 0x%x)",
+                   pci_dev->id.vendor_id, pci_dev->id.device_id);
+       return ret;
+}
+
+static int
+otx2_ree_pci_remove(struct rte_pci_device *pci_dev)
+{
+       char name[RTE_REGEXDEV_NAME_MAX_LEN];
+       struct rte_regexdev *dev = NULL;
+
+       if (pci_dev == NULL)
+               return -EINVAL;
+
+       rte_pci_device_name(&pci_dev->addr, name, sizeof(name));
+
+       dev = rte_regexdev_get_device_by_name(name);
+
+       if (dev == NULL)
+               return -ENODEV;
+
+       return ree_dev_fini(dev);
+}
+
+static struct rte_pci_id pci_id_ree_table[] = {
+       {
+               RTE_PCI_DEVICE(PCI_VENDOR_ID_CAVIUM,
+                               PCI_DEVID_OCTEONTX2_RVU_REE_PF)
+       },
+};
+
+static struct rte_pci_driver otx2_regexdev_pmd = {
+       .id_table = pci_id_ree_table,
+       .drv_flags = RTE_PCI_DRV_NEED_MAPPING,
+       .probe = otx2_ree_pci_probe,
+       .remove = otx2_ree_pci_remove,
+};
+
+
+RTE_PMD_REGISTER_PCI(REGEXDEV_NAME_OCTEONTX2_PMD, otx2_regexdev_pmd);
+RTE_PMD_REGISTER_PCI_TABLE(REGEXDEV_NAME_OCTEONTX2_PMD, pci_id_ree_table);
diff --git a/drivers/regex/octeontx2/otx2_regexdev.h b/drivers/regex/octeontx2/otx2_regexdev.h
new file mode 100644 (file)
index 0000000..d710535
--- /dev/null
@@ -0,0 +1,109 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright (C) 2020 Marvell International Ltd.
+ */
+
+#ifndef _OTX2_REGEXDEV_H_
+#define _OTX2_REGEXDEV_H_
+
+#include <rte_common.h>
+#include <rte_regexdev.h>
+
+#include "otx2_dev.h"
+
+#define ree_func_trace otx2_ree_dbg
+
+/* Marvell OCTEON TX2 Regex PMD device name */
+#define REGEXDEV_NAME_OCTEONTX2_PMD    regex_octeontx2
+
+#define OTX2_REE_MAX_LFS               36
+#define OTX2_REE_MAX_QUEUES_PER_VF     36
+#define OTX2_REE_MAX_MATCHES_PER_VF    254
+
+#define OTX2_REE_MAX_PAYLOAD_SIZE      (1 << 14)
+
+#define OTX2_REE_NON_INC_PROG 0
+#define OTX2_REE_INC_PROG 1
+
+#define REE_MOD_INC(i, l)   ((i) == (l - 1) ? (i) = 0 : (i)++)
+
+
+/**
+ * Device vf data
+ */
+struct otx2_ree_vf {
+       struct otx2_dev otx2_dev;
+       /**< Base class */
+       uint16_t max_queues;
+       /**< Max queues supported */
+       uint8_t nb_queues;
+       /**< Number of regex queues attached */
+       uint16_t max_matches;
+       /**<  Max matches supported*/
+       uint16_t lf_msixoff[OTX2_REE_MAX_LFS];
+       /**< MSI-X offsets */
+       uint8_t block_address;
+       /**< REE Block Address */
+       uint8_t err_intr_registered:1;
+       /**< Are error interrupts registered? */
+};
+
+/**
+ * Device private data
+ */
+struct otx2_ree_data {
+       uint32_t regexdev_capa;
+       uint64_t rule_flags;
+       /**< Feature flags exposes HW/SW features for the given device */
+       uint16_t max_rules_per_group;
+       /**< Maximum rules supported per subset by this device */
+       uint16_t max_groups;
+       /**< Maximum subset supported by this device */
+       void **queue_pairs;
+       /**< Array of pointers to queue pairs. */
+       uint16_t nb_queue_pairs;
+       /**< Number of device queue pairs. */
+       struct otx2_ree_vf vf;
+       /**< vf data */
+       struct rte_regexdev_rule *rules;
+       /**< rules to be compiled */
+       uint16_t nb_rules;
+       /**< number of rules */
+} __rte_cache_aligned;
+
+struct otx2_ree_rid {
+       uintptr_t rid;
+       /** Request id of a ree operation */
+       uint64_t user_id;
+       /* Client data */
+       /**< IOVA address of the pattern to be matched. */
+};
+
+struct otx2_ree_pending_queue {
+       uint64_t pending_count;
+       /** Pending requests count */
+       struct otx2_ree_rid *rid_queue;
+       /** Array of pending requests */
+       uint16_t enq_tail;
+       /** Tail of queue to be used for enqueue */
+       uint16_t deq_head;
+       /** Head of queue to be used for dequeue */
+};
+
+struct otx2_ree_qp {
+       uint32_t id;
+       /**< Queue pair id */
+       uintptr_t base;
+       /**< Base address where BAR is mapped */
+       struct otx2_ree_pending_queue pend_q;
+       /**< Pending queue */
+       rte_iova_t iq_dma_addr;
+       /**< Instruction queue address */
+       uint32_t otx2_regexdev_jobid;
+       /**< Job ID */
+       uint32_t write_offset;
+       /**< write offset */
+       regexdev_stop_flush_t cb;
+       /**< Callback function called during rte_regex_dev_stop()*/
+};
+
+#endif /* _OTX2_REGEXDEV_H_ */
diff --git a/drivers/regex/octeontx2/otx2_regexdev_compiler.c b/drivers/regex/octeontx2/otx2_regexdev_compiler.c
new file mode 100644 (file)
index 0000000..785459f
--- /dev/null
@@ -0,0 +1,229 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright (C) 2020 Marvell International Ltd.
+ */
+
+#include <rte_malloc.h>
+#include <rte_regexdev.h>
+
+#include "otx2_regexdev.h"
+#include "otx2_regexdev_compiler.h"
+#include "otx2_regexdev_mbox.h"
+
+#ifdef REE_COMPILER_SDK
+#include <rxp-compiler.h>
+
+static int
+ree_rule_db_compile(const struct rte_regexdev_rule *rules,
+               uint16_t nb_rules, struct rxp_rof **rof, struct rxp_rof **rofi,
+               struct rxp_rof *rof_for_incremental_compile,
+               struct rxp_rof *rofi_for_incremental_compile)
+{
+       /*INPUT*/
+       struct rxp_prefix_selection_control_list *prefix_selection_control_list
+               = NULL;
+       struct rxp_blacklist_data_sample *blacklist_sample_data = NULL;
+       struct rxp_rule_ids_to_remove *rule_ids_to_remove = NULL;
+       struct rxp_roff *roff_for_incremental_compile = NULL;
+
+       /*OPTIONS - setting default values*/
+       enum rxp_virtual_prefix_mode virtual_prefix_mode =
+                       RXP_VIRTUAL_PREFIX_MODE_0;
+       enum rxp_prefix_capacity prefix_capacity = RXP_PREFIX_CAPACITY_32K;
+       /**< rxp_global_regex_options_flags*/
+       enum rxp_compiler_objective objective = RXP_COMPILER_OBJECTIVE_5;
+       enum rxp_tpe_data_width tpe_data_width = RXP_TPE_DATA_WIDTH_4;
+       uint32_t compiler_options = RXP_COMPILER_OPTIONS_FORCE;
+       /**< rxp_compiler_options_flags*/
+       enum rxp_verbose_level verbose = RXP_VERBOSE_LEVEL_3;
+       enum rxp_version set_rxp_version = RXP_VERSION_V5_8;
+       uint32_t compiler_output_flags = 0;
+       /**< rxp_compiler_output_flags*/
+       uint32_t global_regex_options = 0;
+       /**< rxp_global_regex_options_flags*/
+       float set_auto_blacklist = 0;
+       uint32_t max_rep_max = 65535;
+       uint32_t divide_ruleset = 1;
+       struct rxp_ruleset ruleset;
+       float ptpb_threshold = 0;
+       uint32_t set_max = 0;
+       uint32_t threads = 1;
+
+       /*OUTPUT*/
+       struct rxp_rule_direction_analysis *rule_direction_analysis = NULL;
+       struct rxp_compilation_statistics *compilation_statistics = NULL;
+       struct rxp_prefix_selection_control_list *generated_pscl = NULL;
+       struct rxp_uncompiled_rules_log *uncompiled_rules_log = NULL;
+       struct rxp_critical_rules_rank *critical_rules_rank = NULL;
+       struct rxp_compiled_rules_log *compiled_rules_log = NULL;
+       struct rxp_roff *roff = NULL;
+
+       uint16_t i;
+       int ret;
+
+       ruleset.number_of_entries = nb_rules;
+       ruleset.rules = rte_malloc("rxp_rule_entry",
+                       nb_rules*sizeof(struct rxp_rule_entry), 0);
+
+       if (ruleset.rules == NULL) {
+               otx2_err("Could not allocate memory for rule compilation\n");
+               return -EFAULT;
+       }
+       if (rof_for_incremental_compile)
+               compiler_options |= RXP_COMPILER_OPTIONS_INCREMENTAL;
+       if (rofi_for_incremental_compile)
+               compiler_options |= RXP_COMPILER_OPTIONS_CHECKSUM;
+
+       for (i = 0; i < nb_rules; i++) {
+               ruleset.rules[i].number_of_prefix_entries = 0;
+               ruleset.rules[i].prefix = NULL;
+               ruleset.rules[i].rule = rules[i].pcre_rule;
+               ruleset.rules[i].rule_id = rules[i].rule_id;
+               ruleset.rules[i].subset_id = rules[i].group_id;
+               ruleset.rules[i].rule_direction_type =
+                               RXP_RULE_DIRECTION_TYPE_NONE;
+       }
+
+       ret = rxp_compile_advanced(
+                       /*INPUT*/
+                       &ruleset,
+                       prefix_selection_control_list,
+                       rof_for_incremental_compile,
+                       roff_for_incremental_compile,
+                       rofi_for_incremental_compile,
+                       rule_ids_to_remove,
+                       blacklist_sample_data,
+
+                       /*OPTIONS*/
+                       compiler_options,
+                       prefix_capacity,
+                       global_regex_options,
+                       set_auto_blacklist,
+                       set_max,
+                       objective,
+                       ptpb_threshold,
+                       max_rep_max,
+                       threads,
+                       set_rxp_version,
+                       verbose,
+                       tpe_data_width,
+                       virtual_prefix_mode,
+                       compiler_output_flags,
+                       divide_ruleset,
+
+                       /*OUTPUT*/
+                       &compilation_statistics,
+                       &compiled_rules_log,
+                       &critical_rules_rank,
+                       &rule_direction_analysis,
+                       &uncompiled_rules_log,
+                       rof,
+                       &roff,
+                       rofi,
+                       &generated_pscl);
+       rte_free(ruleset.rules);
+
+       return ret;
+}
+
+int
+otx2_ree_rule_db_compile_prog(struct rte_regexdev *dev)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       char compiler_version[] = "20.5.2.eda0fa2";
+       char timestamp[] = "19700101_000001";
+       uint32_t rule_db_len, rule_dbi_len;
+       struct rxp_rof *rofi_inc_p = NULL;
+       struct rxp_rof_entry rule_dbi[6];
+       char *rofi_rof_entries = NULL;
+       struct rxp_rof *rofi = NULL;
+       struct rxp_rof *rof = NULL;
+       struct rxp_rof rofi_inc;
+       struct rxp_rof rof_inc;
+       char *rule_db = NULL;
+       int ret;
+
+       ree_func_trace();
+
+       ret = otx2_ree_rule_db_len_get(dev, &rule_db_len, &rule_dbi_len);
+       if (ret != 0) {
+               otx2_err("Could not get rule db length");
+               return ret;
+       }
+
+       if (rule_db_len > 0) {
+               otx2_ree_dbg("Incremental compile, rule db len %d rule dbi len %d",
+                               rule_db_len, rule_dbi_len);
+               rule_db = rte_malloc("ree_rule_db", rule_db_len, 0);
+               if (!rule_db) {
+                       otx2_err("Could not allocate memory for rule db");
+                       return -EFAULT;
+               }
+
+               ret = otx2_ree_rule_db_get(dev, rule_db, rule_db_len,
+                               (char *)rule_dbi, rule_dbi_len);
+               if (ret) {
+                       otx2_err("Could not read rule db");
+                       rte_free(rule_db);
+                       return -EFAULT;
+               }
+               rof_inc.rof_revision = 0;
+               rof_inc.rof_version = 2;
+               rof_inc.rof_entries = (struct rxp_rof_entry *)rule_db;
+               rof_inc.rxp_compiler_version = compiler_version;
+               rof_inc.timestamp = timestamp;
+               rof_inc.number_of_entries =
+                               (rule_db_len/sizeof(struct rxp_rof_entry));
+
+               if (rule_dbi_len > 0) {
+                       /* incremental compilation not the first time */
+                       rofi_inc.rof_revision = 0;
+                       rofi_inc.rof_version = 2;
+                       rofi_inc.rof_entries = rule_dbi;
+                       rofi_inc.rxp_compiler_version = compiler_version;
+                       rofi_inc.timestamp = timestamp;
+                       rofi_inc.number_of_entries =
+                               (rule_dbi_len/sizeof(struct rxp_rof_entry));
+                       rofi_inc_p = &rofi_inc;
+               }
+               ret = ree_rule_db_compile(data->rules, data->nb_rules, &rof,
+                               &rofi, &rof_inc, rofi_inc_p);
+               if (rofi->number_of_entries == 0) {
+                       otx2_ree_dbg("No change to rule db");
+                       ret = 0;
+                       goto free_structs;
+               }
+               rule_dbi_len = rofi->number_of_entries *
+                               sizeof(struct rxp_rof_entry);
+               rofi_rof_entries = (char *)rofi->rof_entries;
+       } else {
+               /* full compilation */
+               ret = ree_rule_db_compile(data->rules, data->nb_rules, &rof,
+                               &rofi, NULL, NULL);
+       }
+       if (ret != 0) {
+               otx2_err("Could not compile rule db");
+               goto free_structs;
+       }
+       rule_db_len = rof->number_of_entries * sizeof(struct rxp_rof_entry);
+       ret = otx2_ree_rule_db_prog(dev, (char *)rof->rof_entries, rule_db_len,
+                       rofi_rof_entries, rule_dbi_len);
+       if (ret)
+               otx2_err("Could not program rule db");
+
+free_structs:
+       rxp_free_structs(NULL, NULL, NULL, NULL, NULL, &rof, NULL, &rofi, NULL,
+                       1);
+
+       if (rule_db)
+               rte_free(rule_db);
+
+       return ret;
+}
+#else
+int
+otx2_ree_rule_db_compile_prog(struct rte_regexdev *dev)
+{
+       RTE_SET_USED(dev);
+       return -ENOTSUP;
+}
+#endif
diff --git a/drivers/regex/octeontx2/otx2_regexdev_compiler.h b/drivers/regex/octeontx2/otx2_regexdev_compiler.h
new file mode 100644 (file)
index 0000000..8d2625b
--- /dev/null
@@ -0,0 +1,11 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright (C) 2020 Marvell International Ltd.
+ */
+
+#ifndef _OTX2_REGEXDEV_COMPILER_H_
+#define _OTX2_REGEXDEV_COMPILER_H_
+
+int
+otx2_ree_rule_db_compile_prog(struct rte_regexdev *dev);
+
+#endif /* _OTX2_REGEXDEV_COMPILER_H_ */
diff --git a/drivers/regex/octeontx2/otx2_regexdev_hw_access.c b/drivers/regex/octeontx2/otx2_regexdev_hw_access.c
new file mode 100644 (file)
index 0000000..620d5c9
--- /dev/null
@@ -0,0 +1,167 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright (C) 2020 Marvell International Ltd.
+ */
+
+#include "otx2_common.h"
+#include "otx2_dev.h"
+#include "otx2_regexdev_hw_access.h"
+#include "otx2_regexdev_mbox.h"
+
+static void
+ree_lf_err_intr_handler(void *param)
+{
+       uintptr_t base = (uintptr_t)param;
+       uint8_t lf_id;
+       uint64_t intr;
+
+       lf_id = (base >> 12) & 0xFF;
+
+       intr = otx2_read64(base + OTX2_REE_LF_MISC_INT);
+       if (intr == 0)
+               return;
+
+       otx2_ree_dbg("LF %d MISC_INT: 0x%" PRIx64 "", lf_id, intr);
+
+       /* Clear interrupt */
+       otx2_write64(intr, base + OTX2_REE_LF_MISC_INT);
+}
+
+static void
+ree_lf_err_intr_unregister(const struct rte_regexdev *dev, uint16_t msix_off,
+                          uintptr_t base)
+{
+       struct rte_pci_device *pci_dev = RTE_DEV_TO_PCI(dev->device);
+       struct rte_intr_handle *handle = &pci_dev->intr_handle;
+
+       /* Disable error interrupts */
+       otx2_write64(~0ull, base + OTX2_REE_LF_MISC_INT_ENA_W1C);
+
+       otx2_unregister_irq(handle, ree_lf_err_intr_handler, (void *)base,
+                           msix_off);
+}
+
+void
+otx2_ree_err_intr_unregister(const struct rte_regexdev *dev)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct otx2_ree_vf *vf = &data->vf;
+       uintptr_t base;
+       uint32_t i;
+
+       for (i = 0; i < vf->nb_queues; i++) {
+               base = OTX2_REE_LF_BAR2(vf, i);
+               ree_lf_err_intr_unregister(dev, vf->lf_msixoff[i], base);
+       }
+
+       vf->err_intr_registered = 0;
+}
+
+static int
+ree_lf_err_intr_register(const struct rte_regexdev *dev, uint16_t msix_off,
+                        uintptr_t base)
+{
+       struct rte_pci_device *pci_dev = RTE_DEV_TO_PCI(dev->device);
+       struct rte_intr_handle *handle = &pci_dev->intr_handle;
+       int ret;
+
+       /* Disable error interrupts */
+       otx2_write64(~0ull, base + OTX2_REE_LF_MISC_INT_ENA_W1C);
+
+       /* Register error interrupt handler */
+       ret = otx2_register_irq(handle, ree_lf_err_intr_handler, (void *)base,
+                               msix_off);
+       if (ret)
+               return ret;
+
+       /* Enable error interrupts */
+       otx2_write64(~0ull, base + OTX2_REE_LF_MISC_INT_ENA_W1S);
+
+       return 0;
+}
+
+int
+otx2_ree_err_intr_register(const struct rte_regexdev *dev)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct otx2_ree_vf *vf = &data->vf;
+       uint32_t i, j, ret;
+       uintptr_t base;
+
+       for (i = 0; i < vf->nb_queues; i++) {
+               if (vf->lf_msixoff[i] == MSIX_VECTOR_INVALID) {
+                       otx2_err("Invalid REE LF MSI-X offset: 0x%x",
+                                   vf->lf_msixoff[i]);
+                       return -EINVAL;
+               }
+       }
+
+       for (i = 0; i < vf->nb_queues; i++) {
+               base = OTX2_REE_LF_BAR2(vf, i);
+               ret = ree_lf_err_intr_register(dev, vf->lf_msixoff[i], base);
+               if (ret)
+                       goto intr_unregister;
+       }
+
+       vf->err_intr_registered = 1;
+       return 0;
+
+intr_unregister:
+       /* Unregister the ones already registered */
+       for (j = 0; j < i; j++) {
+               base = OTX2_REE_LF_BAR2(vf, j);
+               ree_lf_err_intr_unregister(dev, vf->lf_msixoff[j], base);
+       }
+       return ret;
+}
+
+int
+otx2_ree_iq_enable(const struct rte_regexdev *dev, const struct otx2_ree_qp *qp,
+                  uint8_t pri, uint32_t size_div2)
+{
+       union otx2_ree_lf_sbuf_addr base;
+       union otx2_ree_lf_ena lf_ena;
+
+       /* Set instruction queue size and priority */
+       otx2_ree_config_lf(dev, qp->id, pri, size_div2);
+
+       /* Set instruction queue base address */
+       /* Should be written after SBUF_CTL and before LF_ENA */
+
+       base.u = otx2_read64(qp->base + OTX2_REE_LF_SBUF_ADDR);
+       base.s.ptr = qp->iq_dma_addr >> 7;
+       otx2_write64(base.u, qp->base + OTX2_REE_LF_SBUF_ADDR);
+
+       /* Enable instruction queue */
+
+       lf_ena.u = otx2_read64(qp->base + OTX2_REE_LF_ENA);
+       lf_ena.s.ena = 1;
+       otx2_write64(lf_ena.u, qp->base + OTX2_REE_LF_ENA);
+
+       return 0;
+}
+
+void
+otx2_ree_iq_disable(struct otx2_ree_qp *qp)
+{
+       union otx2_ree_lf_ena lf_ena;
+
+       /* Stop instruction execution */
+       lf_ena.u = otx2_read64(qp->base + OTX2_REE_LF_ENA);
+       lf_ena.s.ena = 0x0;
+       otx2_write64(lf_ena.u, qp->base + OTX2_REE_LF_ENA);
+}
+
+int
+otx2_ree_max_matches_get(const struct rte_regexdev *dev, uint8_t *max_matches)
+{
+       union otx2_ree_af_reexm_max_match reexm_max_match;
+       int ret;
+
+       ret = otx2_ree_af_reg_read(dev, REE_AF_REEXM_MAX_MATCH,
+                                  &reexm_max_match.u);
+       if (ret)
+               return ret;
+
+       *max_matches = reexm_max_match.s.max;
+       return 0;
+}
diff --git a/drivers/regex/octeontx2/otx2_regexdev_hw_access.h b/drivers/regex/octeontx2/otx2_regexdev_hw_access.h
new file mode 100644 (file)
index 0000000..dedf5f3
--- /dev/null
@@ -0,0 +1,202 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright (C) 2020 Marvell International Ltd.
+ */
+
+#ifndef _OTX2_REGEXDEV_HW_ACCESS_H_
+#define _OTX2_REGEXDEV_HW_ACCESS_H_
+
+#include <stdint.h>
+
+#include "otx2_regexdev.h"
+
+/* REE instruction queue length */
+#define OTX2_REE_IQ_LEN                        (1 << 13)
+
+#define OTX2_REE_DEFAULT_CMD_QLEN      OTX2_REE_IQ_LEN
+
+/* Status register bits */
+#define OTX2_REE_STATUS_PMI_EOJ_BIT            (1 << 14)
+#define OTX2_REE_STATUS_PMI_SOJ_BIT            (1 << 13)
+#define OTX2_REE_STATUS_MP_CNT_DET_BIT         (1 << 7)
+#define OTX2_REE_STATUS_MM_CNT_DET_BIT         (1 << 6)
+#define OTX2_REE_STATUS_ML_CNT_DET_BIT         (1 << 5)
+#define OTX2_REE_STATUS_MST_CNT_DET_BIT                (1 << 4)
+#define OTX2_REE_STATUS_MPT_CNT_DET_BIT                (1 << 3)
+
+/* Register offsets */
+/* REE LF registers */
+#define OTX2_REE_LF_DONE_INT           0x120ull
+#define OTX2_REE_LF_DONE_INT_W1S       0x130ull
+#define OTX2_REE_LF_DONE_INT_ENA_W1S   0x138ull
+#define OTX2_REE_LF_DONE_INT_ENA_W1C   0x140ull
+#define OTX2_REE_LF_MISC_INT           0x300ull
+#define OTX2_REE_LF_MISC_INT_W1S       0x310ull
+#define OTX2_REE_LF_MISC_INT_ENA_W1S   0x320ull
+#define OTX2_REE_LF_MISC_INT_ENA_W1C   0x330ull
+#define OTX2_REE_LF_ENA                        0x10ull
+#define OTX2_REE_LF_SBUF_ADDR          0x20ull
+#define OTX2_REE_LF_DONE               0x100ull
+#define OTX2_REE_LF_DONE_ACK           0x110ull
+#define OTX2_REE_LF_DONE_WAIT          0x148ull
+#define OTX2_REE_LF_DOORBELL           0x400ull
+#define OTX2_REE_LF_OUTSTAND_JOB       0x410ull
+
+/* BAR 0 */
+#define OTX2_REE_AF_QUE_SBUF_CTL(a)    (0x1200ull | (uint64_t)(a) << 3)
+#define OTX2_REE_PRIV_LF_CFG(a)                (0x41000ull | (uint64_t)(a) << 3)
+
+#define OTX2_REE_LF_BAR2(vf, q_id) \
+               ((vf)->otx2_dev.bar2 + \
+                (((vf)->block_address << 20) | ((q_id) << 12)))
+
+
+#define OTX2_REE_QUEUE_HI_PRIO 0x1
+
+enum ree_desc_type_e {
+       REE_TYPE_JOB_DESC    = 0x0,
+       REE_TYPE_RESULT_DESC = 0x1,
+       REE_TYPE_ENUM_LAST   = 0x2
+};
+
+union otx2_ree_priv_lf_cfg {
+       uint64_t u;
+       struct {
+               uint64_t slot                        : 8;
+               uint64_t pf_func                     : 16;
+               uint64_t reserved_24_62              : 39;
+               uint64_t ena                         : 1;
+       } s;
+};
+
+
+union otx2_ree_lf_sbuf_addr {
+       uint64_t u;
+       struct {
+               uint64_t off                         : 7;
+               uint64_t ptr                         : 46;
+               uint64_t reserved_53_63              : 11;
+       } s;
+};
+
+union otx2_ree_lf_ena {
+       uint64_t u;
+       struct {
+               uint64_t ena                         : 1;
+               uint64_t reserved_1_63               : 63;
+       } s;
+};
+
+union otx2_ree_af_reexm_max_match {
+       uint64_t u;
+       struct {
+               uint64_t max                         : 8;
+               uint64_t reserved_8_63               : 56;
+       } s;
+};
+
+union otx2_ree_lf_done {
+       uint64_t u;
+       struct {
+               uint64_t done                        : 20;
+               uint64_t reserved_20_63              : 44;
+       } s;
+};
+
+union otx2_ree_inst {
+       uint64_t u[8];
+       struct  {
+               uint64_t doneint                     :  1;
+               uint64_t reserved_1_3                :  3;
+               uint64_t dg                          :  1;
+               uint64_t reserved_5_7                :  3;
+               uint64_t ooj                         :  1;
+               uint64_t reserved_9_15               :  7;
+               uint64_t reserved_16_63              : 48;
+               uint64_t inp_ptr_addr                : 64;
+               uint64_t inp_ptr_ctl                 : 64;
+               uint64_t res_ptr_addr                : 64;
+               uint64_t wq_ptr                      : 64;
+               uint64_t tag                         : 32;
+               uint64_t tt                          :  2;
+               uint64_t ggrp                        : 10;
+               uint64_t reserved_364_383            : 20;
+               uint64_t reserved_384_391            :  8;
+               uint64_t ree_job_id                  : 24;
+               uint64_t ree_job_ctrl                : 16;
+               uint64_t ree_job_length              : 15;
+               uint64_t reserved_447_447            :  1;
+               uint64_t ree_job_subset_id_0         : 16;
+               uint64_t ree_job_subset_id_1         : 16;
+               uint64_t ree_job_subset_id_2         : 16;
+               uint64_t ree_job_subset_id_3         : 16;
+       } cn98xx;
+};
+
+union otx2_ree_res_status {
+       uint64_t u;
+       struct {
+               uint64_t job_type                    :  3;
+               uint64_t mpt_cnt_det                 :  1;
+               uint64_t mst_cnt_det                 :  1;
+               uint64_t ml_cnt_det                  :  1;
+               uint64_t mm_cnt_det                  :  1;
+               uint64_t mp_cnt_det                  :  1;
+               uint64_t mode                        :  2;
+               uint64_t reserved_10_11              :  2;
+               uint64_t reserved_12_12              :  1;
+               uint64_t pmi_soj                     :  1;
+               uint64_t pmi_eoj                     :  1;
+               uint64_t reserved_15_15              :  1;
+               uint64_t reserved_16_63              : 48;
+       } s;
+};
+
+union otx2_ree_res {
+       uint64_t u[8];
+       struct ree_res_s_98 {
+               uint64_t done                   :  1;
+               uint64_t hwjid                  :  7;
+               uint64_t ree_res_job_id         : 24;
+               uint64_t ree_res_status         : 16;
+               uint64_t ree_res_dmcnt          :  8;
+               uint64_t ree_res_mcnt           :  8;
+               uint64_t ree_meta_ptcnt         : 16;
+               uint64_t ree_meta_icnt          : 16;
+               uint64_t ree_meta_lcnt          : 16;
+               uint64_t ree_pmi_min_byte_ptr   : 16;
+               uint64_t ree_err                :  1;
+               uint64_t reserved_129_190       : 62;
+               uint64_t doneint                :  1;
+               uint64_t reserved_192_255       : 64;
+               uint64_t reserved_256_319       : 64;
+               uint64_t reserved_320_383       : 64;
+               uint64_t reserved_384_447       : 64;
+               uint64_t reserved_448_511       : 64;
+       } s;
+};
+
+union otx2_ree_match {
+       uint64_t u;
+       struct {
+               uint64_t ree_rule_id                 : 32;
+               uint64_t start_ptr                   : 14;
+               uint64_t reserved_46_47              :  2;
+               uint64_t match_length                : 15;
+               uint64_t reserved_63_63              :  1;
+       } s;
+};
+
+void otx2_ree_err_intr_unregister(const struct rte_regexdev *dev);
+
+int otx2_ree_err_intr_register(const struct rte_regexdev *dev);
+
+int otx2_ree_iq_enable(const struct rte_regexdev *dev,
+                      const struct otx2_ree_qp *qp,
+                      uint8_t pri, uint32_t size_div128);
+
+void otx2_ree_iq_disable(struct otx2_ree_qp *qp);
+
+int otx2_ree_max_matches_get(const struct rte_regexdev *dev,
+                            uint8_t *max_matches);
+
+#endif /* _OTX2_REGEXDEV_HW_ACCESS_H_ */
diff --git a/drivers/regex/octeontx2/otx2_regexdev_mbox.c b/drivers/regex/octeontx2/otx2_regexdev_mbox.c
new file mode 100644 (file)
index 0000000..6d58d36
--- /dev/null
@@ -0,0 +1,401 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright (C) 2020 Marvell International Ltd.
+ */
+
+#include "otx2_common.h"
+#include "otx2_dev.h"
+#include "otx2_regexdev_mbox.h"
+#include "otx2_regexdev.h"
+
+int
+otx2_ree_available_queues_get(const struct rte_regexdev *dev,
+                             uint16_t *nb_queues)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct otx2_ree_vf *vf = &data->vf;
+       struct free_rsrcs_rsp *rsp;
+       struct otx2_dev *otx2_dev;
+       int ret;
+
+       otx2_dev = &vf->otx2_dev;
+       otx2_mbox_alloc_msg_free_rsrc_cnt(otx2_dev->mbox);
+
+       ret = otx2_mbox_process_msg(otx2_dev->mbox, (void *)&rsp);
+       if (ret)
+               return -EIO;
+
+       if (vf->block_address == RVU_BLOCK_ADDR_REE0)
+               *nb_queues = rsp->ree0;
+       else
+               *nb_queues = rsp->ree1;
+       return 0;
+}
+
+int
+otx2_ree_queues_attach(const struct rte_regexdev *dev, uint8_t nb_queues)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct otx2_ree_vf *vf = &data->vf;
+       struct rsrc_attach_req *req;
+       struct otx2_mbox *mbox;
+
+       /* Ask AF to attach required LFs */
+       mbox = vf->otx2_dev.mbox;
+       req = otx2_mbox_alloc_msg_attach_resources(mbox);
+
+       /* 1 LF = 1 queue */
+       req->reelfs = nb_queues;
+       req->ree_blkaddr = vf->block_address;
+
+       if (otx2_mbox_process(mbox) < 0)
+               return -EIO;
+
+       /* Update number of attached queues */
+       vf->nb_queues = nb_queues;
+
+       return 0;
+}
+
+int
+otx2_ree_queues_detach(const struct rte_regexdev *dev)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct otx2_ree_vf *vf = &data->vf;
+       struct rsrc_detach_req *req;
+       struct otx2_mbox *mbox;
+
+       mbox = vf->otx2_dev.mbox;
+       req = otx2_mbox_alloc_msg_detach_resources(mbox);
+       req->reelfs = true;
+       req->partial = true;
+       if (otx2_mbox_process(mbox) < 0)
+               return -EIO;
+
+       /* Queues have been detached */
+       vf->nb_queues = 0;
+
+       return 0;
+}
+
+int
+otx2_ree_msix_offsets_get(const struct rte_regexdev *dev)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct otx2_ree_vf *vf = &data->vf;
+       struct msix_offset_rsp *rsp;
+       struct otx2_mbox *mbox;
+       uint32_t i, ret;
+
+       /* Get REE MSI-X vector offsets */
+       mbox = vf->otx2_dev.mbox;
+       otx2_mbox_alloc_msg_msix_offset(mbox);
+
+       ret = otx2_mbox_process_msg(mbox, (void *)&rsp);
+       if (ret)
+               return ret;
+
+       for (i = 0; i < vf->nb_queues; i++) {
+               if (vf->block_address == RVU_BLOCK_ADDR_REE0)
+                       vf->lf_msixoff[i] = rsp->ree0_lf_msixoff[i];
+               else
+                       vf->lf_msixoff[i] = rsp->ree1_lf_msixoff[i];
+               otx2_ree_dbg("lf_msixoff[%d]  0x%x", i, vf->lf_msixoff[i]);
+       }
+
+       return 0;
+}
+
+static int
+ree_send_mbox_msg(struct otx2_ree_vf *vf)
+{
+       struct otx2_mbox *mbox = vf->otx2_dev.mbox;
+       int ret;
+
+       otx2_mbox_msg_send(mbox, 0);
+
+       ret = otx2_mbox_wait_for_rsp(mbox, 0);
+       if (ret < 0) {
+               otx2_err("Could not get mailbox response");
+               return ret;
+       }
+
+       return 0;
+}
+
+int
+otx2_ree_config_lf(const struct rte_regexdev *dev, uint8_t lf, uint8_t pri,
+                  uint32_t size)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct otx2_ree_vf *vf = &data->vf;
+       struct ree_lf_req_msg *req;
+       struct otx2_mbox *mbox;
+       int ret;
+
+       mbox = vf->otx2_dev.mbox;
+       req = otx2_mbox_alloc_msg_ree_config_lf(mbox);
+
+       req->lf = lf;
+       req->pri =  pri ? 1 : 0;
+       req->size = size;
+       req->blkaddr = vf->block_address;
+
+       ret = otx2_mbox_process(mbox);
+       if (ret < 0) {
+               otx2_err("Could not get mailbox response");
+               return ret;
+       }
+       return 0;
+}
+
+int
+otx2_ree_af_reg_read(const struct rte_regexdev *dev, uint64_t reg,
+                    uint64_t *val)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct otx2_ree_vf *vf = &data->vf;
+       struct ree_rd_wr_reg_msg *msg;
+       struct otx2_mbox_dev *mdev;
+       struct otx2_mbox *mbox;
+       int ret, off;
+
+       mbox = vf->otx2_dev.mbox;
+       mdev = &mbox->dev[0];
+       msg = (struct ree_rd_wr_reg_msg *)otx2_mbox_alloc_msg_rsp(mbox, 0,
+                                               sizeof(*msg), sizeof(*msg));
+       if (msg == NULL) {
+               otx2_err("Could not allocate mailbox message");
+               return -EFAULT;
+       }
+
+       msg->hdr.id = MBOX_MSG_REE_RD_WR_REGISTER;
+       msg->hdr.sig = OTX2_MBOX_REQ_SIG;
+       msg->hdr.pcifunc = vf->otx2_dev.pf_func;
+       msg->is_write = 0;
+       msg->reg_offset = reg;
+       msg->ret_val = val;
+       msg->blkaddr = vf->block_address;
+
+       ret = ree_send_mbox_msg(vf);
+       if (ret < 0)
+               return ret;
+
+       off = mbox->rx_start +
+                       RTE_ALIGN(sizeof(struct mbox_hdr), MBOX_MSG_ALIGN);
+       msg = (struct ree_rd_wr_reg_msg *) ((uintptr_t)mdev->mbase + off);
+
+       *val = msg->val;
+
+       return 0;
+}
+
+int
+otx2_ree_af_reg_write(const struct rte_regexdev *dev, uint64_t reg,
+                     uint64_t val)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct otx2_ree_vf *vf = &data->vf;
+       struct ree_rd_wr_reg_msg *msg;
+       struct otx2_mbox *mbox;
+
+       mbox = vf->otx2_dev.mbox;
+       msg = (struct ree_rd_wr_reg_msg *)otx2_mbox_alloc_msg_rsp(mbox, 0,
+                                               sizeof(*msg), sizeof(*msg));
+       if (msg == NULL) {
+               otx2_err("Could not allocate mailbox message");
+               return -EFAULT;
+       }
+
+       msg->hdr.id = MBOX_MSG_REE_RD_WR_REGISTER;
+       msg->hdr.sig = OTX2_MBOX_REQ_SIG;
+       msg->hdr.pcifunc = vf->otx2_dev.pf_func;
+       msg->is_write = 1;
+       msg->reg_offset = reg;
+       msg->val = val;
+       msg->blkaddr = vf->block_address;
+
+       return ree_send_mbox_msg(vf);
+}
+
+int
+otx2_ree_rule_db_get(const struct rte_regexdev *dev, char *rule_db,
+               uint32_t rule_db_len, char *rule_dbi, uint32_t rule_dbi_len)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct ree_rule_db_get_req_msg *req;
+       struct ree_rule_db_get_rsp_msg *rsp;
+       char *rule_db_ptr = (char *)rule_db;
+       struct otx2_ree_vf *vf = &data->vf;
+       struct otx2_mbox *mbox;
+       int ret, last = 0;
+       uint32_t len = 0;
+
+       mbox = vf->otx2_dev.mbox;
+       if (!rule_db) {
+               otx2_err("Couldn't return rule db due to NULL pointer");
+               return -EFAULT;
+       }
+
+       while (!last) {
+               req = (struct ree_rule_db_get_req_msg *)
+                       otx2_mbox_alloc_msg_rsp(mbox, 0, sizeof(*req),
+                                               sizeof(*rsp));
+               if (!req) {
+                       otx2_err("Could not allocate mailbox message");
+                       return -EFAULT;
+               }
+
+               req->hdr.id = MBOX_MSG_REE_RULE_DB_GET;
+               req->hdr.sig = OTX2_MBOX_REQ_SIG;
+               req->hdr.pcifunc = vf->otx2_dev.pf_func;
+               req->blkaddr = vf->block_address;
+               req->is_dbi = 0;
+               req->offset = len;
+               ret = otx2_mbox_process_msg(mbox, (void *)&rsp);
+               if (ret)
+                       return ret;
+               if (rule_db_len < len + rsp->len) {
+                       otx2_err("Rule db size is too small");
+                       return -EFAULT;
+               }
+               otx2_mbox_memcpy(rule_db_ptr, rsp->rule_db, rsp->len);
+               len += rsp->len;
+               rule_db_ptr = rule_db_ptr + rsp->len;
+               last = rsp->is_last;
+       }
+
+       if (rule_dbi) {
+               req = (struct ree_rule_db_get_req_msg *)
+                       otx2_mbox_alloc_msg_rsp(mbox, 0, sizeof(*req),
+                                               sizeof(*rsp));
+               if (!req) {
+                       otx2_err("Could not allocate mailbox message");
+                       return -EFAULT;
+               }
+
+               req->hdr.id = MBOX_MSG_REE_RULE_DB_GET;
+               req->hdr.sig = OTX2_MBOX_REQ_SIG;
+               req->hdr.pcifunc = vf->otx2_dev.pf_func;
+               req->blkaddr = vf->block_address;
+               req->is_dbi = 1;
+               req->offset = 0;
+
+               ret = otx2_mbox_process_msg(mbox, (void *)&rsp);
+               if (ret)
+                       return ret;
+               if (rule_dbi_len < rsp->len) {
+                       otx2_err("Rule dbi size is too small");
+                       return -EFAULT;
+               }
+               otx2_mbox_memcpy(rule_dbi, rsp->rule_db, rsp->len);
+       }
+       return 0;
+}
+
+int
+otx2_ree_rule_db_len_get(const struct rte_regexdev *dev,
+               uint32_t *rule_db_len,
+               uint32_t *rule_dbi_len)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       struct ree_rule_db_len_rsp_msg *rsp;
+       struct otx2_ree_vf *vf = &data->vf;
+       struct ree_req_msg *req;
+       struct otx2_mbox *mbox;
+       int ret;
+
+       mbox = vf->otx2_dev.mbox;
+       req = (struct ree_req_msg *)
+               otx2_mbox_alloc_msg_rsp(mbox, 0, sizeof(*req), sizeof(*rsp));
+       if (!req) {
+               otx2_err("Could not allocate mailbox message");
+               return -EFAULT;
+       }
+
+       req->hdr.id = MBOX_MSG_REE_RULE_DB_LEN_GET;
+       req->hdr.sig = OTX2_MBOX_REQ_SIG;
+       req->hdr.pcifunc = vf->otx2_dev.pf_func;
+       req->blkaddr = vf->block_address;
+       ret = otx2_mbox_process_msg(mbox, (void *)&rsp);
+       if (ret)
+               return ret;
+       if (rule_db_len != NULL)
+               *rule_db_len = rsp->len;
+       if (rule_dbi_len != NULL)
+               *rule_dbi_len = rsp->inc_len;
+
+       return 0;
+}
+
+static int
+ree_db_msg(const struct rte_regexdev *dev, const char *db, uint32_t db_len,
+               int inc, int dbi)
+{
+       struct otx2_ree_data *data = dev->data->dev_private;
+       uint32_t len_left = db_len, offset = 0;
+       struct ree_rule_db_prog_req_msg *req;
+       struct otx2_ree_vf *vf = &data->vf;
+       const char *rule_db_ptr = db;
+       struct otx2_mbox *mbox;
+       struct msg_rsp *rsp;
+       int ret;
+
+       mbox = vf->otx2_dev.mbox;
+       while (len_left) {
+               req = (struct ree_rule_db_prog_req_msg *)
+                       otx2_mbox_alloc_msg_rsp(mbox, 0, sizeof(*req),
+                                               sizeof(*rsp));
+               if (!req) {
+                       otx2_err("Could not allocate mailbox message");
+                       return -EFAULT;
+               }
+               req->hdr.id = MBOX_MSG_REE_RULE_DB_PROG;
+               req->hdr.sig = OTX2_MBOX_REQ_SIG;
+               req->hdr.pcifunc = vf->otx2_dev.pf_func;
+               req->offset = offset;
+               req->total_len = db_len;
+               req->len = REE_RULE_DB_REQ_BLOCK_SIZE;
+               req->is_incremental = inc;
+               req->is_dbi = dbi;
+               req->blkaddr = vf->block_address;
+
+               if (len_left < REE_RULE_DB_REQ_BLOCK_SIZE) {
+                       req->is_last = true;
+                       req->len = len_left;
+               }
+               otx2_mbox_memcpy(req->rule_db, rule_db_ptr, req->len);
+               ret = otx2_mbox_process_msg(mbox, (void *)&rsp);
+               if (ret) {
+                       otx2_err("Programming mailbox processing failed");
+                       return ret;
+               }
+               len_left -= req->len;
+               offset += req->len;
+               rule_db_ptr = rule_db_ptr + req->len;
+       }
+       return 0;
+}
+
+int
+otx2_ree_rule_db_prog(const struct rte_regexdev *dev, const char *rule_db,
+               uint32_t rule_db_len, const char *rule_dbi,
+               uint32_t rule_dbi_len)
+{
+       int inc, ret;
+
+       if (rule_db_len == 0) {
+               otx2_err("Couldn't program empty rule db");
+               return -EFAULT;
+       }
+       inc = (rule_dbi_len != 0);
+       if ((rule_db == NULL) || (inc && (rule_dbi == NULL))) {
+               otx2_err("Couldn't program NULL rule db");
+               return -EFAULT;
+       }
+       if (inc) {
+               ret = ree_db_msg(dev, rule_dbi, rule_dbi_len, inc, 1);
+               if (ret)
+                       return ret;
+       }
+       return ree_db_msg(dev, rule_db, rule_db_len, inc, 0);
+}
diff --git a/drivers/regex/octeontx2/otx2_regexdev_mbox.h b/drivers/regex/octeontx2/otx2_regexdev_mbox.h
new file mode 100644 (file)
index 0000000..953efa6
--- /dev/null
@@ -0,0 +1,38 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright (C) 2020 Marvell International Ltd.
+ */
+
+#ifndef _OTX2_REGEXDEV_MBOX_H_
+#define _OTX2_REGEXDEV_MBOX_H_
+
+#include <rte_regexdev.h>
+
+int otx2_ree_available_queues_get(const struct rte_regexdev *dev,
+                                 uint16_t *nb_queues);
+
+int otx2_ree_queues_attach(const struct rte_regexdev *dev, uint8_t nb_queues);
+
+int otx2_ree_queues_detach(const struct rte_regexdev *dev);
+
+int otx2_ree_msix_offsets_get(const struct rte_regexdev *dev);
+
+int otx2_ree_config_lf(const struct rte_regexdev *dev, uint8_t lf, uint8_t pri,
+                      uint32_t size);
+
+int otx2_ree_af_reg_read(const struct rte_regexdev *dev, uint64_t reg,
+                        uint64_t *val);
+
+int otx2_ree_af_reg_write(const struct rte_regexdev *dev, uint64_t reg,
+                         uint64_t val);
+
+int otx2_ree_rule_db_get(const struct rte_regexdev *dev, char *rule_db,
+                uint32_t rule_db_len, char *rule_dbi, uint32_t rule_dbi_len);
+
+int otx2_ree_rule_db_len_get(const struct rte_regexdev *dev,
+                            uint32_t *rule_db_len, uint32_t *rule_dbi_len);
+
+int otx2_ree_rule_db_prog(const struct rte_regexdev *dev, const char *rule_db,
+               uint32_t rule_db_len, const char *rule_dbi,
+               uint32_t rule_dbi_len);
+
+#endif /* _OTX2_REGEXDEV_MBOX_H_ */
diff --git a/drivers/regex/octeontx2/rte_pmd_octeontx2_regex_version.map b/drivers/regex/octeontx2/rte_pmd_octeontx2_regex_version.map
new file mode 100644 (file)
index 0000000..4a76d1d
--- /dev/null
@@ -0,0 +1,3 @@
+DPDK_21 {
+       local: *;
+};