event/dpaa: support crypto adapter
authorAkhil Goyal <akhil.goyal@nxp.com>
Fri, 4 Oct 2019 11:26:59 +0000 (16:56 +0530)
committerAkhil Goyal <akhil.goyal@nxp.com>
Wed, 9 Oct 2019 11:14:19 +0000 (13:14 +0200)
event dpaa device support both ethernet as well as
crypto queues to be attached to it. eth_rx_adapter
provide infrastructure to attach ethernet queues and
crypto_adapter provide support for crypto queues.

This patch add support for dpaa_eventdev to attach
dpaa_sec queues.

Signed-off-by: Akhil Goyal <akhil.goyal@nxp.com>
drivers/event/dpaa/Makefile
drivers/event/dpaa/dpaa_eventdev.c
drivers/event/dpaa/dpaa_eventdev.h
drivers/event/dpaa/meson.build

index cf96264..1856fa4 100644 (file)
@@ -21,6 +21,9 @@ CFLAGS += -I$(RTE_SDK)/drivers/bus/dpaa/include/
 CFLAGS += -I$(RTE_SDK)/drivers/mempool/dpaa
 CFLAGS += -I$(RTE_SDK)/lib/librte_eal/common/include
 
+LDLIBS += -lrte_pmd_dpaa_sec
+CFLAGS += -I$(RTE_SDK)/drivers/crypto/dpaa_sec
+
 EXPORT_MAP := rte_pmd_dpaa_event_version.map
 
 LIBABIVER := 1
index 1e247e4..d02b869 100644 (file)
 #include <rte_eventdev_pmd_vdev.h>
 #include <rte_ethdev.h>
 #include <rte_event_eth_rx_adapter.h>
+#include <rte_cryptodev.h>
 #include <rte_dpaa_bus.h>
 #include <rte_dpaa_logs.h>
 #include <rte_cycles.h>
 #include <rte_kvargs.h>
 
 #include <dpaa_ethdev.h>
+#include <dpaa_sec_event.h>
 #include "dpaa_eventdev.h"
 #include <dpaa_mempool.h>
 
@@ -322,7 +324,7 @@ dpaa_event_dev_info_get(struct rte_eventdev *dev,
        EVENTDEV_INIT_FUNC_TRACE();
 
        RTE_SET_USED(dev);
-       dev_info->driver_name = "event_dpaa";
+       dev_info->driver_name = "event_dpaa1";
        dev_info->min_dequeue_timeout_ns =
                DPAA_EVENT_MIN_DEQUEUE_TIMEOUT;
        dev_info->max_dequeue_timeout_ns =
@@ -718,6 +720,149 @@ dpaa_event_eth_rx_adapter_stop(const struct rte_eventdev *dev,
        return 0;
 }
 
+static int
+dpaa_eventdev_crypto_caps_get(const struct rte_eventdev *dev,
+                           const struct rte_cryptodev *cdev,
+                           uint32_t *caps)
+{
+       const char *name = cdev->data->name;
+
+       EVENTDEV_INIT_FUNC_TRACE();
+
+       RTE_SET_USED(dev);
+
+       if (!strncmp(name, "dpaa_sec-", 9))
+               *caps = RTE_EVENT_CRYPTO_ADAPTER_DPAA_CAP;
+       else
+               return -1;
+
+       return 0;
+}
+
+static int
+dpaa_eventdev_crypto_queue_add_all(const struct rte_eventdev *dev,
+               const struct rte_cryptodev *cryptodev,
+               const struct rte_event *ev)
+{
+       struct dpaa_eventdev *priv = dev->data->dev_private;
+       uint8_t ev_qid = ev->queue_id;
+       u16 ch_id = priv->evq_info[ev_qid].ch_id;
+       int i, ret;
+
+       EVENTDEV_INIT_FUNC_TRACE();
+
+       for (i = 0; i < cryptodev->data->nb_queue_pairs; i++) {
+               ret = dpaa_sec_eventq_attach(cryptodev, i,
+                               ch_id, ev);
+               if (ret) {
+                       DPAA_EVENTDEV_ERR("dpaa_sec_eventq_attach failed: ret %d\n",
+                                   ret);
+                       goto fail;
+               }
+       }
+       return 0;
+fail:
+       for (i = (i - 1); i >= 0 ; i--)
+               dpaa_sec_eventq_detach(cryptodev, i);
+
+       return ret;
+}
+
+static int
+dpaa_eventdev_crypto_queue_add(const struct rte_eventdev *dev,
+               const struct rte_cryptodev *cryptodev,
+               int32_t rx_queue_id,
+               const struct rte_event *ev)
+{
+       struct dpaa_eventdev *priv = dev->data->dev_private;
+       uint8_t ev_qid = ev->queue_id;
+       u16 ch_id = priv->evq_info[ev_qid].ch_id;
+       int ret;
+
+       EVENTDEV_INIT_FUNC_TRACE();
+
+       if (rx_queue_id == -1)
+               return dpaa_eventdev_crypto_queue_add_all(dev,
+                               cryptodev, ev);
+
+       ret = dpaa_sec_eventq_attach(cryptodev, rx_queue_id,
+                       ch_id, ev);
+       if (ret) {
+               DPAA_EVENTDEV_ERR(
+                       "dpaa_sec_eventq_attach failed: ret: %d\n", ret);
+               return ret;
+       }
+       return 0;
+}
+
+static int
+dpaa_eventdev_crypto_queue_del_all(const struct rte_eventdev *dev,
+                            const struct rte_cryptodev *cdev)
+{
+       int i, ret;
+
+       EVENTDEV_INIT_FUNC_TRACE();
+
+       RTE_SET_USED(dev);
+
+       for (i = 0; i < cdev->data->nb_queue_pairs; i++) {
+               ret = dpaa_sec_eventq_detach(cdev, i);
+               if (ret) {
+                       DPAA_EVENTDEV_ERR(
+                               "dpaa_sec_eventq_detach failed:ret %d\n", ret);
+                       return ret;
+               }
+       }
+
+       return 0;
+}
+
+static int
+dpaa_eventdev_crypto_queue_del(const struct rte_eventdev *dev,
+                            const struct rte_cryptodev *cryptodev,
+                            int32_t rx_queue_id)
+{
+       int ret;
+
+       EVENTDEV_INIT_FUNC_TRACE();
+
+       if (rx_queue_id == -1)
+               return dpaa_eventdev_crypto_queue_del_all(dev, cryptodev);
+
+       ret = dpaa_sec_eventq_detach(cryptodev, rx_queue_id);
+       if (ret) {
+               DPAA_EVENTDEV_ERR(
+                       "dpaa_sec_eventq_detach failed: ret: %d\n", ret);
+               return ret;
+       }
+
+       return 0;
+}
+
+static int
+dpaa_eventdev_crypto_start(const struct rte_eventdev *dev,
+                          const struct rte_cryptodev *cryptodev)
+{
+       EVENTDEV_INIT_FUNC_TRACE();
+
+       RTE_SET_USED(dev);
+       RTE_SET_USED(cryptodev);
+
+       return 0;
+}
+
+static int
+dpaa_eventdev_crypto_stop(const struct rte_eventdev *dev,
+                         const struct rte_cryptodev *cryptodev)
+{
+       EVENTDEV_INIT_FUNC_TRACE();
+
+       RTE_SET_USED(dev);
+       RTE_SET_USED(cryptodev);
+
+       return 0;
+}
+
 static struct rte_eventdev_ops dpaa_eventdev_ops = {
        .dev_infos_get    = dpaa_event_dev_info_get,
        .dev_configure    = dpaa_event_dev_configure,
@@ -738,6 +883,11 @@ static struct rte_eventdev_ops dpaa_eventdev_ops = {
        .eth_rx_adapter_queue_del = dpaa_event_eth_rx_adapter_queue_del,
        .eth_rx_adapter_start = dpaa_event_eth_rx_adapter_start,
        .eth_rx_adapter_stop = dpaa_event_eth_rx_adapter_stop,
+       .crypto_adapter_caps_get        = dpaa_eventdev_crypto_caps_get,
+       .crypto_adapter_queue_pair_add  = dpaa_eventdev_crypto_queue_add,
+       .crypto_adapter_queue_pair_del  = dpaa_eventdev_crypto_queue_del,
+       .crypto_adapter_start           = dpaa_eventdev_crypto_start,
+       .crypto_adapter_stop            = dpaa_eventdev_crypto_stop,
 };
 
 static int flag_check_handler(__rte_unused const char *key,
@@ -806,6 +956,8 @@ dpaa_event_dev_create(const char *name, const char *params)
                eventdev->dequeue_burst = dpaa_event_dequeue_burst_intr;
        }
 
+       RTE_LOG(INFO, PMD, "%s eventdev added", name);
+
        /* For secondary processes, the primary has done all the work */
        if (rte_eal_process_type() != RTE_PROC_PRIMARY)
                return 0;
index 8134e6b..b8f247c 100644 (file)
@@ -40,6 +40,11 @@ do {                                         \
                RTE_EVENT_ETH_RX_ADAPTER_CAP_MULTI_EVENTQ | \
                RTE_EVENT_ETH_RX_ADAPTER_CAP_OVERRIDE_FLOW_ID)
 
+#define RTE_EVENT_CRYPTO_ADAPTER_DPAA_CAP \
+               (RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_OP_NEW | \
+               RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_QP_EV_BIND | \
+               RTE_EVENT_CRYPTO_ADAPTER_CAP_SESSION_PRIVATE_DATA)
+
 struct dpaa_eventq {
        /* Channel Id */
        uint16_t ch_id;
index c1e7254..b74f3bf 100644 (file)
@@ -5,7 +5,8 @@ if not is_linux
        build = false
        reason = 'only supported on linux'
 endif
-deps += ['pmd_dpaa']
+deps += ['pmd_dpaa', 'pmd_dpaa_sec']
 sources = files('dpaa_eventdev.c')
 
 allow_experimental_apis = true
+includes += include_directories('../../crypto/dpaa_sec/')