--- /dev/null
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright 2018 NXP
+ */
+
+#include <sys/queue.h>
+#include <stdio.h>
+#include <errno.h>
+#include <stdint.h>
+#include <string.h>
+#include <unistd.h>
+#include <stdarg.h>
+
+#include <rte_ethdev.h>
+#include <rte_log.h>
+#include <rte_eth_ctrl.h>
+#include <rte_malloc.h>
+#include <rte_flow_driver.h>
+#include <rte_tailq.h>
+
+#include <rte_fslmc.h>
+#include <fsl_dpdmux.h>
+#include <fsl_dpkg.h>
+
+#include <dpaa2_ethdev.h>
+#include <dpaa2_pmd_logs.h>
+
+struct dpaa2_dpdmux_dev {
+ TAILQ_ENTRY(dpaa2_dpdmux_dev) next;
+ /**< Pointer to Next device instance */
+ struct fsl_mc_io dpdmux; /** handle to DPDMUX portal object */
+ uint16_t token;
+ uint32_t dpdmux_id; /*HW ID for DPDMUX object */
+ uint8_t num_ifs; /* Number of interfaces in DPDMUX */
+};
+
+struct rte_flow {
+ struct dpdmux_rule_cfg rule;
+};
+
+TAILQ_HEAD(dpdmux_dev_list, dpaa2_dpdmux_dev);
+static struct dpdmux_dev_list dpdmux_dev_list =
+ TAILQ_HEAD_INITIALIZER(dpdmux_dev_list); /*!< DPDMUX device list */
+
+static struct dpaa2_dpdmux_dev *get_dpdmux_from_id(uint32_t dpdmux_id)
+{
+ struct dpaa2_dpdmux_dev *dpdmux_dev = NULL;
+
+ /* Get DPBP dev handle from list using index */
+ TAILQ_FOREACH(dpdmux_dev, &dpdmux_dev_list, next) {
+ if (dpdmux_dev->dpdmux_id == dpdmux_id)
+ break;
+ }
+
+ return dpdmux_dev;
+}
+
+struct rte_flow *
+rte_pmd_dpaa2_mux_flow_create(uint32_t dpdmux_id,
+ struct rte_flow_item *pattern[],
+ struct rte_flow_action *actions[])
+{
+ struct dpaa2_dpdmux_dev *dpdmux_dev;
+ struct dpkg_profile_cfg kg_cfg;
+ const struct rte_flow_item_ipv4 *spec;
+ const struct rte_flow_action_vf *vf_conf;
+ struct dpdmux_cls_action dpdmux_action;
+ struct rte_flow *flow = NULL;
+ void *key_iova, *mask_iova, *key_cfg_iova = NULL;
+ int ret;
+
+ if (pattern[0]->type != RTE_FLOW_ITEM_TYPE_IPV4) {
+ DPAA2_PMD_ERR("Not supported pattern type: %d",
+ pattern[0]->type);
+ return NULL;
+ }
+
+ /* Find the DPDMUX from dpdmux_id in our list */
+ dpdmux_dev = get_dpdmux_from_id(dpdmux_id);
+ if (!dpdmux_dev) {
+ DPAA2_PMD_ERR("Invalid dpdmux_id: %d", dpdmux_id);
+ return NULL;
+ }
+
+ key_cfg_iova = rte_zmalloc(NULL, DIST_PARAM_IOVA_SIZE,
+ RTE_CACHE_LINE_SIZE);
+ if (!key_cfg_iova) {
+ DPAA2_PMD_ERR("Unable to allocate flow-dist parameters");
+ return NULL;
+ }
+
+ /* Currently taking only IP protocol as an extract type.
+ * This can be exended to other fields using pattern->type.
+ */
+ memset(&kg_cfg, 0, sizeof(struct dpkg_profile_cfg));
+ kg_cfg.extracts[0].extract.from_hdr.prot = NET_PROT_IP;
+ kg_cfg.extracts[0].extract.from_hdr.field = NH_FLD_IP_PROTO;
+ kg_cfg.extracts[0].type = DPKG_EXTRACT_FROM_HDR;
+ kg_cfg.extracts[0].extract.from_hdr.type = DPKG_FULL_FIELD;
+ kg_cfg.num_extracts = 1;
+
+ ret = dpkg_prepare_key_cfg(&kg_cfg, key_cfg_iova);
+ if (ret) {
+ DPAA2_PMD_ERR("dpkg_prepare_key_cfg failed: err(%d)", ret);
+ goto creation_error;
+ }
+
+ ret = dpdmux_set_custom_key(&dpdmux_dev->dpdmux, CMD_PRI_LOW,
+ dpdmux_dev->token,
+ (uint64_t)(DPAA2_VADDR_TO_IOVA(key_cfg_iova)));
+ if (ret) {
+ DPAA2_PMD_ERR("dpdmux_set_custom_key failed: err(%d)", ret);
+ goto creation_error;
+ }
+
+ /* As now our key extract parameters are set, let us configure
+ * the rule.
+ */
+ flow = rte_zmalloc(NULL, sizeof(struct rte_flow) +
+ (2 * DIST_PARAM_IOVA_SIZE), RTE_CACHE_LINE_SIZE);
+ if (!flow) {
+ DPAA2_PMD_ERR(
+ "Memory allocation failure for rule configration\n");
+ goto creation_error;
+ }
+ key_iova = (void *)((size_t)flow + sizeof(struct rte_flow));
+ mask_iova = (void *)((size_t)key_iova + DIST_PARAM_IOVA_SIZE);
+
+ spec = (const struct rte_flow_item_ipv4 *)pattern[0]->spec;
+ memcpy(key_iova, (const void *)&spec->hdr.next_proto_id,
+ sizeof(uint8_t));
+ memcpy(mask_iova, pattern[0]->mask, sizeof(uint8_t));
+
+ flow->rule.key_iova = (uint64_t)(DPAA2_VADDR_TO_IOVA(key_iova));
+ flow->rule.mask_iova = (uint64_t)(DPAA2_VADDR_TO_IOVA(mask_iova));
+ flow->rule.key_size = sizeof(uint8_t);
+
+ vf_conf = (const struct rte_flow_action_vf *)(actions[0]->conf);
+ if (vf_conf->id == 0 || vf_conf->id > dpdmux_dev->num_ifs) {
+ DPAA2_PMD_ERR("Invalid destination id\n");
+ goto creation_error;
+ }
+ dpdmux_action.dest_if = vf_conf->id;
+
+ ret = dpdmux_add_custom_cls_entry(&dpdmux_dev->dpdmux, CMD_PRI_LOW,
+ dpdmux_dev->token, &flow->rule,
+ &dpdmux_action);
+ if (ret) {
+ DPAA2_PMD_ERR("dpdmux_add_custom_cls_entry failed: err(%d)",
+ ret);
+ goto creation_error;
+ }
+
+ return flow;
+
+creation_error:
+ rte_free((void *)key_cfg_iova);
+ rte_free((void *)flow);
+ return NULL;
+}
+
+static int
+dpaa2_create_dpdmux_device(int vdev_fd __rte_unused,
+ struct vfio_device_info *obj_info __rte_unused,
+ int dpdmux_id)
+{
+ struct dpaa2_dpdmux_dev *dpdmux_dev;
+ struct dpdmux_attr attr;
+ int ret;
+
+ PMD_INIT_FUNC_TRACE();
+
+ /* Allocate DPAA2 dpdmux handle */
+ dpdmux_dev = rte_malloc(NULL, sizeof(struct dpaa2_dpdmux_dev), 0);
+ if (!dpdmux_dev) {
+ DPAA2_PMD_ERR("Memory allocation failed for DPDMUX Device");
+ return -1;
+ }
+
+ /* Open the dpdmux object */
+ dpdmux_dev->dpdmux.regs = rte_mcp_ptr_list[MC_PORTAL_INDEX];
+ ret = dpdmux_open(&dpdmux_dev->dpdmux, CMD_PRI_LOW, dpdmux_id,
+ &dpdmux_dev->token);
+ if (ret) {
+ DPAA2_PMD_ERR("Unable to open dpdmux object: err(%d)", ret);
+ goto init_err;
+ }
+
+ ret = dpdmux_get_attributes(&dpdmux_dev->dpdmux, CMD_PRI_LOW,
+ dpdmux_dev->token, &attr);
+ if (ret) {
+ DPAA2_PMD_ERR("Unable to get dpdmux attr: err(%d)", ret);
+ goto init_err;
+ }
+
+ ret = dpdmux_if_set_default(&dpdmux_dev->dpdmux, CMD_PRI_LOW,
+ dpdmux_dev->token, 1);
+ if (ret) {
+ DPAA2_PMD_ERR("setting default interface failed in %s",
+ __func__);
+ goto init_err;
+ }
+
+ dpdmux_dev->dpdmux_id = dpdmux_id;
+ dpdmux_dev->num_ifs = attr.num_ifs;
+
+ TAILQ_INSERT_TAIL(&dpdmux_dev_list, dpdmux_dev, next);
+
+ return 0;
+
+init_err:
+ if (dpdmux_dev)
+ rte_free(dpdmux_dev);
+
+ return -1;
+}
+
+static struct rte_dpaa2_object rte_dpaa2_dpdmux_obj = {
+ .dev_type = DPAA2_MUX,
+ .create = dpaa2_create_dpdmux_device,
+};
+
+RTE_PMD_REGISTER_DPAA2_OBJECT(dpdmux, rte_dpaa2_dpdmux_obj);