1 /* SPDX-License-Identifier: BSD-3-Clause
3 * Copyright (c) 2015-2016 Freescale Semiconductor, Inc. All rights reserved.
4 * Copyright 2016-2019 NXP
10 #include <sys/types.h>
15 #include <sys/ioctl.h>
21 #include <sys/eventfd.h>
23 #include <eal_filesystem.h>
25 #include <rte_ethdev_driver.h>
26 #include <rte_malloc.h>
27 #include <rte_memcpy.h>
28 #include <rte_string_fns.h>
29 #include <rte_cycles.h>
30 #include <rte_kvargs.h>
33 #include <rte_eal_memconfig.h>
35 #include "rte_fslmc.h"
36 #include "fslmc_vfio.h"
37 #include "fslmc_logs.h"
38 #include <mc/fsl_dpmng.h>
40 #include "portal/dpaa2_hw_pvt.h"
41 #include "portal/dpaa2_hw_dpio.h"
43 /** Pathname of FSL-MC devices directory. */
44 #define SYSFS_FSL_MC_DEVICES "/sys/bus/fsl-mc/devices"
46 #define FSLMC_CONTAINER_MAX_LEN 8 /**< Of the format dprc.XX */
48 /* Number of VFIO containers & groups with in */
49 static struct fslmc_vfio_group vfio_group;
50 static struct fslmc_vfio_container vfio_container;
51 static int container_device_fd;
52 static char *fslmc_container;
53 static int fslmc_iommu_type;
54 static uint32_t *msi_intr_vaddr;
55 void *(*rte_mcp_ptr_list);
57 static struct rte_dpaa2_object_list dpaa2_obj_list =
58 TAILQ_HEAD_INITIALIZER(dpaa2_obj_list);
60 /*register a fslmc bus based dpaa2 driver */
62 rte_fslmc_object_register(struct rte_dpaa2_object *object)
66 TAILQ_INSERT_TAIL(&dpaa2_obj_list, object, next);
70 fslmc_get_container_group(int *groupid)
75 if (!fslmc_container) {
76 container = getenv("DPRC");
77 if (container == NULL) {
78 DPAA2_BUS_DEBUG("DPAA2: DPRC not available");
82 if (strlen(container) >= FSLMC_CONTAINER_MAX_LEN) {
83 DPAA2_BUS_ERR("Invalid container name: %s", container);
87 fslmc_container = strdup(container);
88 if (!fslmc_container) {
89 DPAA2_BUS_ERR("Mem alloc failure; Container name");
94 fslmc_iommu_type = (rte_vfio_noiommu_is_enabled() == 1) ?
95 RTE_VFIO_NOIOMMU : VFIO_TYPE1_IOMMU;
97 /* get group number */
98 ret = rte_vfio_get_group_num(SYSFS_FSL_MC_DEVICES,
99 fslmc_container, groupid);
101 DPAA2_BUS_ERR("Unable to find %s IOMMU group", fslmc_container);
105 DPAA2_BUS_DEBUG("Container: %s has VFIO iommu group id = %d",
106 fslmc_container, *groupid);
112 vfio_connect_container(void)
116 if (vfio_container.used) {
117 DPAA2_BUS_DEBUG("No container available");
121 /* Try connecting to vfio container if already created */
122 if (!ioctl(vfio_group.fd, VFIO_GROUP_SET_CONTAINER,
123 &vfio_container.fd)) {
125 "Container pre-exists with FD[0x%x] for this group",
127 vfio_group.container = &vfio_container;
131 /* Opens main vfio file descriptor which represents the "container" */
132 fd = rte_vfio_get_container_fd();
134 DPAA2_BUS_ERR("Failed to open VFIO container");
138 /* Check whether support for SMMU type IOMMU present or not */
139 if (ioctl(fd, VFIO_CHECK_EXTENSION, fslmc_iommu_type)) {
140 /* Connect group to container */
141 ret = ioctl(vfio_group.fd, VFIO_GROUP_SET_CONTAINER, &fd);
143 DPAA2_BUS_ERR("Failed to setup group container");
148 ret = ioctl(fd, VFIO_SET_IOMMU, fslmc_iommu_type);
150 DPAA2_BUS_ERR("Failed to setup VFIO iommu");
155 DPAA2_BUS_ERR("No supported IOMMU available");
160 vfio_container.used = 1;
161 vfio_container.fd = fd;
162 vfio_container.group = &vfio_group;
163 vfio_group.container = &vfio_container;
168 static int vfio_map_irq_region(struct fslmc_vfio_group *group)
171 unsigned long *vaddr = NULL;
172 struct vfio_iommu_type1_dma_map map = {
173 .argsz = sizeof(map),
174 .flags = VFIO_DMA_MAP_FLAG_READ | VFIO_DMA_MAP_FLAG_WRITE,
180 vaddr = (unsigned long *)mmap(NULL, 0x1000, PROT_WRITE |
181 PROT_READ, MAP_SHARED, container_device_fd, 0x6030000);
182 if (vaddr == MAP_FAILED) {
183 DPAA2_BUS_INFO("Unable to map region (errno = %d)", errno);
187 msi_intr_vaddr = (uint32_t *)((char *)(vaddr) + 64);
188 map.vaddr = (unsigned long)vaddr;
189 ret = ioctl(group->container->fd, VFIO_IOMMU_MAP_DMA, &map);
193 DPAA2_BUS_ERR("Unable to map DMA address (errno = %d)", errno);
197 static int fslmc_map_dma(uint64_t vaddr, rte_iova_t iovaddr, size_t len);
198 static int fslmc_unmap_dma(uint64_t vaddr, rte_iova_t iovaddr, size_t len);
201 fslmc_memevent_cb(enum rte_mem_event type, const void *addr, size_t len,
202 void *arg __rte_unused)
204 struct rte_memseg_list *msl;
205 struct rte_memseg *ms;
206 size_t cur_len = 0, map_len = 0;
208 rte_iova_t iova_addr;
211 msl = rte_mem_virt2memseg_list(addr);
213 while (cur_len < len) {
214 const void *va = RTE_PTR_ADD(addr, cur_len);
216 ms = rte_mem_virt2memseg(va, msl);
217 iova_addr = ms->iova;
218 virt_addr = ms->addr_64;
221 DPAA2_BUS_DEBUG("Request for %s, va=%p, "
222 "virt_addr=0x%" PRIx64 ", "
223 "iova=0x%" PRIx64 ", map_len=%zu",
224 type == RTE_MEM_EVENT_ALLOC ?
226 va, virt_addr, iova_addr, map_len);
228 /* iova_addr may be set to RTE_BAD_IOVA */
229 if (iova_addr == RTE_BAD_IOVA) {
230 DPAA2_BUS_DEBUG("Segment has invalid iova, skipping\n");
235 if (type == RTE_MEM_EVENT_ALLOC)
236 ret = fslmc_map_dma(virt_addr, iova_addr, map_len);
238 ret = fslmc_unmap_dma(virt_addr, iova_addr, map_len);
241 DPAA2_BUS_ERR("DMA Mapping/Unmapping failed. "
242 "Map=%d, addr=%p, len=%zu, err:(%d)",
243 type, va, map_len, ret);
250 if (type == RTE_MEM_EVENT_ALLOC)
251 DPAA2_BUS_DEBUG("Total Mapped: addr=%p, len=%zu",
254 DPAA2_BUS_DEBUG("Total Unmapped: addr=%p, len=%zu",
259 fslmc_map_dma(uint64_t vaddr, rte_iova_t iovaddr __rte_unused, size_t len)
261 struct fslmc_vfio_group *group;
262 struct vfio_iommu_type1_dma_map dma_map = {
263 .argsz = sizeof(struct vfio_iommu_type1_dma_map),
264 .flags = VFIO_DMA_MAP_FLAG_READ | VFIO_DMA_MAP_FLAG_WRITE,
268 if (fslmc_iommu_type == RTE_VFIO_NOIOMMU) {
269 DPAA2_BUS_DEBUG("Running in NOIOMMU mode");
274 dma_map.vaddr = vaddr;
276 #ifdef RTE_LIBRTE_DPAA2_USE_PHYS_IOVA
277 dma_map.iova = iovaddr;
279 dma_map.iova = dma_map.vaddr;
282 /* SET DMA MAP for IOMMU */
285 if (!group->container) {
286 DPAA2_BUS_ERR("Container is not connected ");
290 DPAA2_BUS_DEBUG("--> Map address: 0x%"PRIx64", size: %"PRIu64"",
291 (uint64_t)dma_map.vaddr, (uint64_t)dma_map.size);
292 ret = ioctl(group->container->fd, VFIO_IOMMU_MAP_DMA, &dma_map);
294 DPAA2_BUS_ERR("VFIO_IOMMU_MAP_DMA API(errno = %d)",
303 fslmc_unmap_dma(uint64_t vaddr, uint64_t iovaddr __rte_unused, size_t len)
305 struct fslmc_vfio_group *group;
306 struct vfio_iommu_type1_dma_unmap dma_unmap = {
307 .argsz = sizeof(struct vfio_iommu_type1_dma_unmap),
312 if (fslmc_iommu_type == RTE_VFIO_NOIOMMU) {
313 DPAA2_BUS_DEBUG("Running in NOIOMMU mode");
317 dma_unmap.size = len;
318 dma_unmap.iova = vaddr;
320 /* SET DMA MAP for IOMMU */
323 if (!group->container) {
324 DPAA2_BUS_ERR("Container is not connected ");
328 DPAA2_BUS_DEBUG("--> Unmap address: 0x%"PRIx64", size: %"PRIu64"",
329 (uint64_t)dma_unmap.iova, (uint64_t)dma_unmap.size);
330 ret = ioctl(group->container->fd, VFIO_IOMMU_UNMAP_DMA, &dma_unmap);
332 DPAA2_BUS_ERR("VFIO_IOMMU_UNMAP_DMA API(errno = %d)",
341 fslmc_dmamap_seg(const struct rte_memseg_list *msl __rte_unused,
342 const struct rte_memseg *ms, void *arg)
347 /* if IOVA address is invalid, skip */
348 if (ms->iova == RTE_BAD_IOVA)
351 ret = fslmc_map_dma(ms->addr_64, ms->iova, ms->len);
353 DPAA2_BUS_ERR("Unable to VFIO map (addr=%p, len=%zu)",
362 rte_fslmc_vfio_mem_dmamap(uint64_t vaddr, uint64_t iova, uint64_t size)
365 struct fslmc_vfio_group *group;
366 struct vfio_iommu_type1_dma_map dma_map = {
367 .argsz = sizeof(struct vfio_iommu_type1_dma_map),
368 .flags = VFIO_DMA_MAP_FLAG_READ | VFIO_DMA_MAP_FLAG_WRITE,
371 if (fslmc_iommu_type == RTE_VFIO_NOIOMMU) {
372 DPAA2_BUS_DEBUG("Running in NOIOMMU mode");
376 /* SET DMA MAP for IOMMU */
378 if (!group->container) {
379 DPAA2_BUS_ERR("Container is not connected");
384 dma_map.vaddr = vaddr;
387 DPAA2_BUS_DEBUG("VFIO dmamap 0x%llx:0x%llx, size 0x%llx\n",
388 dma_map.vaddr, dma_map.iova, dma_map.size);
389 ret = ioctl(group->container->fd, VFIO_IOMMU_MAP_DMA,
392 printf("Unable to map DMA address (errno = %d)\n",
400 int rte_fslmc_vfio_dmamap(void)
404 /* Lock before parsing and registering callback to memory subsystem */
405 rte_mcfg_mem_read_lock();
407 if (rte_memseg_walk(fslmc_dmamap_seg, &i) < 0) {
408 rte_mcfg_mem_read_unlock();
412 ret = rte_mem_event_callback_register("fslmc_memevent_clb",
413 fslmc_memevent_cb, NULL);
414 if (ret && rte_errno == ENOTSUP)
415 DPAA2_BUS_DEBUG("Memory event callbacks not supported");
417 DPAA2_BUS_DEBUG("Unable to install memory handler");
419 DPAA2_BUS_DEBUG("Installed memory callback handler");
421 DPAA2_BUS_DEBUG("Total %d segments found.", i);
423 /* TODO - This is a W.A. as VFIO currently does not add the mapping of
424 * the interrupt region to SMMU. This should be removed once the
425 * support is added in the Kernel.
427 vfio_map_irq_region(&vfio_group);
429 /* Existing segments have been mapped and memory callback for hotplug
430 * has been installed.
432 rte_mcfg_mem_read_unlock();
437 static int64_t vfio_map_mcp_obj(struct fslmc_vfio_group *group, char *mcp_obj)
439 intptr_t v_addr = (intptr_t)MAP_FAILED;
442 struct vfio_device_info d_info = { .argsz = sizeof(d_info) };
443 struct vfio_region_info reg_info = { .argsz = sizeof(reg_info) };
445 /* getting the mcp object's fd*/
446 mc_fd = ioctl(group->fd, VFIO_GROUP_GET_DEVICE_FD, mcp_obj);
448 DPAA2_BUS_ERR("Error in VFIO get dev %s fd from group %d",
453 /* getting device info*/
454 ret = ioctl(mc_fd, VFIO_DEVICE_GET_INFO, &d_info);
456 DPAA2_BUS_ERR("Error in VFIO getting DEVICE_INFO");
460 /* getting device region info*/
461 ret = ioctl(mc_fd, VFIO_DEVICE_GET_REGION_INFO, ®_info);
463 DPAA2_BUS_ERR("Error in VFIO getting REGION_INFO");
467 v_addr = (size_t)mmap(NULL, reg_info.size,
468 PROT_WRITE | PROT_READ, MAP_SHARED,
469 mc_fd, reg_info.offset);
477 #define IRQ_SET_BUF_LEN (sizeof(struct vfio_irq_set) + sizeof(int))
479 int rte_dpaa2_intr_enable(struct rte_intr_handle *intr_handle, int index)
482 char irq_set_buf[IRQ_SET_BUF_LEN];
483 struct vfio_irq_set *irq_set;
486 len = sizeof(irq_set_buf);
488 irq_set = (struct vfio_irq_set *)irq_set_buf;
489 irq_set->argsz = len;
492 VFIO_IRQ_SET_DATA_EVENTFD | VFIO_IRQ_SET_ACTION_TRIGGER;
493 irq_set->index = index;
495 fd_ptr = (int *)&irq_set->data;
496 *fd_ptr = intr_handle->fd;
498 ret = ioctl(intr_handle->vfio_dev_fd, VFIO_DEVICE_SET_IRQS, irq_set);
500 DPAA2_BUS_ERR("Error:dpaa2 SET IRQs fd=%d, err = %d(%s)",
501 intr_handle->fd, errno, strerror(errno));
508 int rte_dpaa2_intr_disable(struct rte_intr_handle *intr_handle, int index)
510 struct vfio_irq_set *irq_set;
511 char irq_set_buf[IRQ_SET_BUF_LEN];
514 len = sizeof(struct vfio_irq_set);
516 irq_set = (struct vfio_irq_set *)irq_set_buf;
517 irq_set->argsz = len;
518 irq_set->flags = VFIO_IRQ_SET_DATA_NONE | VFIO_IRQ_SET_ACTION_TRIGGER;
519 irq_set->index = index;
523 ret = ioctl(intr_handle->vfio_dev_fd, VFIO_DEVICE_SET_IRQS, irq_set);
526 "Error disabling dpaa2 interrupts for fd %d",
532 /* set up interrupt support (but not enable interrupts) */
534 rte_dpaa2_vfio_setup_intr(struct rte_intr_handle *intr_handle,
540 /* start from MSI-X interrupt type */
541 for (i = 0; i < num_irqs; i++) {
542 struct vfio_irq_info irq_info = { .argsz = sizeof(irq_info) };
547 ret = ioctl(vfio_dev_fd, VFIO_DEVICE_GET_IRQ_INFO, &irq_info);
549 DPAA2_BUS_ERR("Cannot get IRQ(%d) info, error %i (%s)",
550 i, errno, strerror(errno));
554 /* if this vector cannot be used with eventfd,
555 * fail if we explicitly
556 * specified interrupt type, otherwise continue
558 if ((irq_info.flags & VFIO_IRQ_INFO_EVENTFD) == 0)
561 /* set up an eventfd for interrupts */
562 fd = eventfd(0, EFD_NONBLOCK | EFD_CLOEXEC);
564 DPAA2_BUS_ERR("Cannot set up eventfd, error %i (%s)",
565 errno, strerror(errno));
569 intr_handle->fd = fd;
570 intr_handle->type = RTE_INTR_HANDLE_VFIO_MSI;
571 intr_handle->vfio_dev_fd = vfio_dev_fd;
576 /* if we're here, we haven't found a suitable interrupt vector */
581 * fslmc_process_iodevices for processing only IO (ETH, CRYPTO, and possibly
585 fslmc_process_iodevices(struct rte_dpaa2_device *dev)
588 struct vfio_device_info device_info = { .argsz = sizeof(device_info) };
589 struct rte_dpaa2_object *object = NULL;
591 dev_fd = ioctl(vfio_group.fd, VFIO_GROUP_GET_DEVICE_FD,
594 DPAA2_BUS_ERR("Unable to obtain device FD for device:%s",
599 if (ioctl(dev_fd, VFIO_DEVICE_GET_INFO, &device_info)) {
600 DPAA2_BUS_ERR("Unable to obtain information for device:%s",
605 switch (dev->dev_type) {
607 rte_dpaa2_vfio_setup_intr(&dev->intr_handle, dev_fd,
608 device_info.num_irqs);
615 TAILQ_FOREACH(object, &dpaa2_obj_list, next) {
616 if (dev->dev_type == object->dev_type)
617 object->create(dev_fd, &device_info,
627 DPAA2_BUS_LOG(DEBUG, "Device (%s) abstracted from VFIO",
633 fslmc_process_mcp(struct rte_dpaa2_device *dev)
637 char *dev_name = NULL;
638 struct fsl_mc_io dpmng = {0};
639 struct mc_version mc_ver_info = {0};
641 rte_mcp_ptr_list = malloc(sizeof(void *) * 1);
642 if (!rte_mcp_ptr_list) {
643 DPAA2_BUS_ERR("Unable to allocate MC portal memory");
648 dev_name = strdup(dev->device.name);
650 DPAA2_BUS_ERR("Unable to allocate MC device name memory");
655 v_addr = vfio_map_mcp_obj(&vfio_group, dev_name);
656 if (v_addr == (intptr_t)MAP_FAILED) {
657 DPAA2_BUS_ERR("Error mapping region (errno = %d)", errno);
662 /* check the MC version compatibility */
663 dpmng.regs = (void *)v_addr;
665 /* In case of secondary processes, MC version check is no longer
668 if (rte_eal_process_type() == RTE_PROC_SECONDARY) {
669 rte_mcp_ptr_list[0] = (void *)v_addr;
673 if (mc_get_version(&dpmng, CMD_PRI_LOW, &mc_ver_info)) {
674 DPAA2_BUS_ERR("Unable to obtain MC version");
679 if ((mc_ver_info.major != MC_VER_MAJOR) ||
680 (mc_ver_info.minor < MC_VER_MINOR)) {
681 DPAA2_BUS_ERR("DPAA2 MC version not compatible!"
682 " Expected %d.%d.x, Detected %d.%d.%d",
683 MC_VER_MAJOR, MC_VER_MINOR,
684 mc_ver_info.major, mc_ver_info.minor,
685 mc_ver_info.revision);
689 rte_mcp_ptr_list[0] = (void *)v_addr;
698 if (rte_mcp_ptr_list) {
699 free(rte_mcp_ptr_list);
700 rte_mcp_ptr_list = NULL;
707 fslmc_vfio_process_group(void)
710 int found_mportal = 0;
711 struct rte_dpaa2_device *dev, *dev_temp;
713 /* Search the MCP as that should be initialized first. */
714 TAILQ_FOREACH_SAFE(dev, &rte_fslmc_bus.device_list, next, dev_temp) {
715 if (dev->dev_type == DPAA2_MPORTAL) {
716 if (dev->device.devargs &&
717 dev->device.devargs->policy == RTE_DEV_BLACKLISTED) {
718 DPAA2_BUS_LOG(DEBUG, "%s Blacklisted, skipping",
720 TAILQ_REMOVE(&rte_fslmc_bus.device_list,
725 ret = fslmc_process_mcp(dev);
727 DPAA2_BUS_ERR("Unable to map MC Portal");
733 TAILQ_REMOVE(&rte_fslmc_bus.device_list, dev, next);
736 /* Ideally there is only a single dpmcp, but in case
737 * multiple exists, looping on remaining devices.
742 /* Cannot continue if there is not even a single mportal */
743 if (!found_mportal) {
744 DPAA2_BUS_ERR("No MC Portal device found. Not continuing");
748 TAILQ_FOREACH_SAFE(dev, &rte_fslmc_bus.device_list, next, dev_temp) {
749 if (dev->device.devargs &&
750 dev->device.devargs->policy == RTE_DEV_BLACKLISTED) {
751 DPAA2_BUS_LOG(DEBUG, "%s Blacklisted, skipping",
753 TAILQ_REMOVE(&rte_fslmc_bus.device_list, dev, next);
756 switch (dev->dev_type) {
760 ret = fslmc_process_iodevices(dev);
762 DPAA2_BUS_DEBUG("Dev (%s) init failed",
771 /* IN case of secondary processes, all control objects
772 * like dpbp, dpcon, dpci are not initialized/required
773 * - all of these are assumed to be initialized and made
774 * available by primary.
776 if (rte_eal_process_type() == RTE_PROC_SECONDARY)
779 /* Call the object creation routine and remove the
780 * device entry from device list
782 ret = fslmc_process_iodevices(dev);
784 DPAA2_BUS_DEBUG("Dev (%s) init failed",
791 ret = fslmc_process_iodevices(dev);
793 DPAA2_BUS_DEBUG("Dev (%s) init failed",
801 /* Unknown - ignore */
802 DPAA2_BUS_DEBUG("Found unknown device (%s)",
804 TAILQ_REMOVE(&rte_fslmc_bus.device_list, dev, next);
814 fslmc_vfio_setup_group(void)
818 struct vfio_group_status status = { .argsz = sizeof(status) };
820 /* if already done once */
821 if (container_device_fd)
824 ret = fslmc_get_container_group(&groupid);
828 /* In case this group was already opened, continue without any
831 if (vfio_group.groupid == groupid) {
832 DPAA2_BUS_ERR("groupid already exists %d", groupid);
836 /* Get the actual group fd */
837 ret = rte_vfio_get_group_fd(groupid);
842 /* Check group viability */
843 ret = ioctl(vfio_group.fd, VFIO_GROUP_GET_STATUS, &status);
845 DPAA2_BUS_ERR("VFIO error getting group status");
846 close(vfio_group.fd);
847 rte_vfio_clear_group(vfio_group.fd);
851 if (!(status.flags & VFIO_GROUP_FLAGS_VIABLE)) {
852 DPAA2_BUS_ERR("VFIO group not viable");
853 close(vfio_group.fd);
854 rte_vfio_clear_group(vfio_group.fd);
857 /* Since Group is VIABLE, Store the groupid */
858 vfio_group.groupid = groupid;
860 /* check if group does not have a container yet */
861 if (!(status.flags & VFIO_GROUP_FLAGS_CONTAINER_SET)) {
862 /* Now connect this IOMMU group to given container */
863 ret = vfio_connect_container();
866 "Error connecting container with groupid %d",
868 close(vfio_group.fd);
869 rte_vfio_clear_group(vfio_group.fd);
874 /* Get Device information */
875 ret = ioctl(vfio_group.fd, VFIO_GROUP_GET_DEVICE_FD, fslmc_container);
877 DPAA2_BUS_ERR("Error getting device %s fd from group %d",
878 fslmc_container, vfio_group.groupid);
879 close(vfio_group.fd);
880 rte_vfio_clear_group(vfio_group.fd);
883 container_device_fd = ret;
884 DPAA2_BUS_DEBUG("VFIO Container FD is [0x%X]",
885 container_device_fd);