1 /* SPDX-License-Identifier: BSD-3-Clause
14 #include <sys/types.h>
15 #include <sys/syscall.h>
17 #include <rte_byteorder.h>
18 #include <rte_common.h>
19 #include <rte_interrupts.h>
21 #include <rte_debug.h>
22 #include <rte_atomic.h>
23 #include <rte_branch_prediction.h>
24 #include <rte_memory.h>
25 #include <rte_tailq.h>
27 #include <rte_alarm.h>
28 #include <rte_ether.h>
29 #include <rte_ethdev_driver.h>
30 #include <rte_malloc.h>
33 #include <rte_mbuf_pool_ops.h>
35 #include <rte_dpaa_bus.h>
36 #include <rte_dpaa_logs.h>
37 #include <dpaax_iova_table.h>
46 int dpaa_logtype_mempool;
48 int dpaa_logtype_eventdev;
50 static struct rte_dpaa_bus rte_dpaa_bus;
51 struct netcfg_info *dpaa_netcfg;
53 /* define a variable to hold the portal_key, once created.*/
54 static pthread_key_t dpaa_portal_key;
56 unsigned int dpaa_svr_family;
58 #define FSL_DPAA_BUS_NAME dpaa_bus
60 RTE_DEFINE_PER_LCORE(bool, dpaa_io);
61 RTE_DEFINE_PER_LCORE(struct dpaa_portal_dqrr, held_bufs);
64 compare_dpaa_devices(struct rte_dpaa_device *dev1,
65 struct rte_dpaa_device *dev2)
69 /* Segragating ETH from SEC devices */
70 if (dev1->device_type > dev2->device_type)
72 else if (dev1->device_type < dev2->device_type)
77 if ((comp != 0) || (dev1->device_type != FSL_DPAA_ETH))
80 if (dev1->id.fman_id > dev2->id.fman_id) {
82 } else if (dev1->id.fman_id < dev2->id.fman_id) {
85 /* FMAN ids match, check for mac_id */
86 if (dev1->id.mac_id > dev2->id.mac_id)
88 else if (dev1->id.mac_id < dev2->id.mac_id)
98 dpaa_add_to_device_list(struct rte_dpaa_device *newdev)
100 int comp, inserted = 0;
101 struct rte_dpaa_device *dev = NULL;
102 struct rte_dpaa_device *tdev = NULL;
104 TAILQ_FOREACH_SAFE(dev, &rte_dpaa_bus.device_list, next, tdev) {
105 comp = compare_dpaa_devices(newdev, dev);
107 TAILQ_INSERT_BEFORE(dev, newdev, next);
114 TAILQ_INSERT_TAIL(&rte_dpaa_bus.device_list, newdev, next);
118 * Reads the SEC device from DTS
119 * Returns -1 if SEC devices not available, 0 otherwise
122 dpaa_sec_available(void)
124 const struct device_node *caam_node;
126 for_each_compatible_node(caam_node, NULL, "fsl,sec-v4.0") {
133 static void dpaa_clean_device_list(void);
135 static struct rte_devargs *
136 dpaa_devargs_lookup(struct rte_dpaa_device *dev)
138 struct rte_devargs *devargs;
141 RTE_EAL_DEVARGS_FOREACH("dpaa_bus", devargs) {
142 devargs->bus->parse(devargs->name, &dev_name);
143 if (strcmp(dev_name, dev->device.name) == 0) {
144 DPAA_BUS_INFO("**Devargs matched %s", dev_name);
152 dpaa_create_device_list(void)
156 struct rte_dpaa_device *dev;
157 struct fm_eth_port_cfg *cfg;
158 struct fman_if *fman_intf;
160 /* Creating Ethernet Devices */
161 for (i = 0; i < dpaa_netcfg->num_ethports; i++) {
162 dev = calloc(1, sizeof(struct rte_dpaa_device));
164 DPAA_BUS_LOG(ERR, "Failed to allocate ETH devices");
169 dev->device.bus = &rte_dpaa_bus.bus;
171 cfg = &dpaa_netcfg->port_cfg[i];
172 fman_intf = cfg->fman_if;
174 /* Device identifiers */
175 dev->id.fman_id = fman_intf->fman_idx + 1;
176 dev->id.mac_id = fman_intf->mac_idx;
177 dev->device_type = FSL_DPAA_ETH;
180 /* Create device name */
181 memset(dev->name, 0, RTE_ETH_NAME_MAX_LEN);
182 sprintf(dev->name, "fm%d-mac%d", (fman_intf->fman_idx + 1),
184 DPAA_BUS_LOG(INFO, "%s netdev added", dev->name);
185 dev->device.name = dev->name;
186 dev->device.devargs = dpaa_devargs_lookup(dev);
188 dpaa_add_to_device_list(dev);
191 rte_dpaa_bus.device_count = i;
193 /* Unlike case of ETH, RTE_LIBRTE_DPAA_MAX_CRYPTODEV SEC devices are
194 * constantly created only if "sec" property is found in the device
195 * tree. Logically there is no limit for number of devices (QI
196 * interfaces) that can be created.
199 if (dpaa_sec_available()) {
200 DPAA_BUS_LOG(INFO, "DPAA SEC devices are not available");
204 /* Creating SEC Devices */
205 for (i = 0; i < RTE_LIBRTE_DPAA_MAX_CRYPTODEV; i++) {
206 dev = calloc(1, sizeof(struct rte_dpaa_device));
208 DPAA_BUS_LOG(ERR, "Failed to allocate SEC devices");
213 dev->device_type = FSL_DPAA_CRYPTO;
214 dev->id.dev_id = rte_dpaa_bus.device_count + i;
216 /* Even though RTE_CRYPTODEV_NAME_MAX_LEN is valid length of
217 * crypto PMD, using RTE_ETH_NAME_MAX_LEN as that is the size
218 * allocated for dev->name/
220 memset(dev->name, 0, RTE_ETH_NAME_MAX_LEN);
221 sprintf(dev->name, "dpaa-sec%d", i);
222 DPAA_BUS_LOG(INFO, "%s cryptodev added", dev->name);
223 dev->device.name = dev->name;
224 dev->device.devargs = dpaa_devargs_lookup(dev);
226 dpaa_add_to_device_list(dev);
229 rte_dpaa_bus.device_count += i;
234 dpaa_clean_device_list();
239 dpaa_clean_device_list(void)
241 struct rte_dpaa_device *dev = NULL;
242 struct rte_dpaa_device *tdev = NULL;
244 TAILQ_FOREACH_SAFE(dev, &rte_dpaa_bus.device_list, next, tdev) {
245 TAILQ_REMOVE(&rte_dpaa_bus.device_list, dev, next);
251 int rte_dpaa_portal_init(void *arg)
254 unsigned int cpu, lcore = rte_lcore_id();
256 struct dpaa_portal *dpaa_io_portal;
258 BUS_INIT_FUNC_TRACE();
260 if ((size_t)arg == 1 || lcore == LCORE_ID_ANY)
261 lcore = rte_get_master_lcore();
263 if (lcore >= RTE_MAX_LCORE)
266 cpu = lcore_config[lcore].core_id;
268 /* Set CPU affinity for this thread.*/
270 ret = pthread_setaffinity_np(id, sizeof(cpu_set_t),
271 &lcore_config[lcore].cpuset);
273 DPAA_BUS_LOG(ERR, "pthread_setaffinity_np failed on core :%u"
274 " (lcore=%u) with ret: %d", cpu, lcore, ret);
278 /* Initialise bman thread portals */
279 ret = bman_thread_init();
281 DPAA_BUS_LOG(ERR, "bman_thread_init failed on core %u"
282 " (lcore=%u) with ret: %d", cpu, lcore, ret);
286 DPAA_BUS_LOG(DEBUG, "BMAN thread initialized - CPU=%d lcore=%d",
289 /* Initialise qman thread portals */
290 ret = qman_thread_init();
292 DPAA_BUS_LOG(ERR, "qman_thread_init failed on core %u"
293 " (lcore=%u) with ret: %d", cpu, lcore, ret);
294 bman_thread_finish();
298 DPAA_BUS_LOG(DEBUG, "QMAN thread initialized - CPU=%d lcore=%d",
301 dpaa_io_portal = rte_malloc(NULL, sizeof(struct dpaa_portal),
302 RTE_CACHE_LINE_SIZE);
303 if (!dpaa_io_portal) {
304 DPAA_BUS_LOG(ERR, "Unable to allocate memory");
305 bman_thread_finish();
306 qman_thread_finish();
310 dpaa_io_portal->qman_idx = qman_get_portal_index();
311 dpaa_io_portal->bman_idx = bman_get_portal_index();
312 dpaa_io_portal->tid = syscall(SYS_gettid);
314 ret = pthread_setspecific(dpaa_portal_key, (void *)dpaa_io_portal);
316 DPAA_BUS_LOG(ERR, "pthread_setspecific failed on core %u"
317 " (lcore=%u) with ret: %d", cpu, lcore, ret);
318 dpaa_portal_finish(NULL);
323 RTE_PER_LCORE(dpaa_io) = true;
325 DPAA_BUS_LOG(DEBUG, "QMAN thread initialized");
331 rte_dpaa_portal_fq_init(void *arg, struct qman_fq *fq)
333 /* Affine above created portal with channel*/
335 struct qman_portal *qp;
338 if (unlikely(!RTE_PER_LCORE(dpaa_io))) {
339 ret = rte_dpaa_portal_init(arg);
341 DPAA_BUS_LOG(ERR, "portal initialization failure");
346 /* Initialise qman specific portals */
347 qp = fsl_qman_portal_create();
349 DPAA_BUS_LOG(ERR, "Unable to alloc fq portal");
353 sdqcr = QM_SDQCR_CHANNELS_POOL_CONV(fq->ch_id);
354 qman_static_dequeue_add(sdqcr, qp);
359 int rte_dpaa_portal_fq_close(struct qman_fq *fq)
361 return fsl_qman_portal_destroy(fq->qp);
365 dpaa_portal_finish(void *arg)
367 struct dpaa_portal *dpaa_io_portal = (struct dpaa_portal *)arg;
369 if (!dpaa_io_portal) {
370 DPAA_BUS_LOG(DEBUG, "Portal already cleaned");
374 bman_thread_finish();
375 qman_thread_finish();
377 pthread_setspecific(dpaa_portal_key, NULL);
379 rte_free(dpaa_io_portal);
380 dpaa_io_portal = NULL;
382 RTE_PER_LCORE(dpaa_io) = false;
386 rte_dpaa_bus_parse(const char *name, void *out_name)
389 int max_fman = 2, max_macs = 16;
390 char *sep = strchr(name, ':');
392 if (strncmp(name, RTE_STR(FSL_DPAA_BUS_NAME),
393 strlen(RTE_STR(FSL_DPAA_BUS_NAME)))) {
398 DPAA_BUS_ERR("Incorrect device name observed");
402 sep = (char *) (sep + 1);
404 for (i = 0; i < max_fman; i++) {
405 for (j = 0; j < max_macs; j++) {
407 snprintf(fm_name, 16, "fm%d-mac%d", i, j);
408 if (strcmp(fm_name, sep) == 0) {
410 strcpy(out_name, sep);
416 for (i = 0; i < RTE_LIBRTE_DPAA_MAX_CRYPTODEV; i++) {
419 snprintf(sec_name, 16, "dpaa-sec%d", i);
420 if (strcmp(sec_name, sep) == 0) {
422 strcpy(out_name, sep);
430 #define DPAA_DEV_PATH1 "/sys/devices/platform/soc/soc:fsl,dpaa"
431 #define DPAA_DEV_PATH2 "/sys/devices/platform/fsl,dpaa"
434 rte_dpaa_bus_scan(void)
438 BUS_INIT_FUNC_TRACE();
440 if ((access(DPAA_DEV_PATH1, F_OK) != 0) &&
441 (access(DPAA_DEV_PATH2, F_OK) != 0)) {
442 RTE_LOG(DEBUG, EAL, "DPAA Bus not present. Skipping.\n");
445 /* detected DPAA devices */
446 rte_dpaa_bus.detected = 1;
448 /* create the key, supplying a function that'll be invoked
449 * when a portal affined thread will be deleted.
451 ret = pthread_key_create(&dpaa_portal_key, dpaa_portal_finish);
453 DPAA_BUS_LOG(DEBUG, "Unable to create pthread key. (%d)", ret);
454 dpaa_clean_device_list();
461 /* register a dpaa bus based dpaa driver */
463 rte_dpaa_driver_register(struct rte_dpaa_driver *driver)
467 BUS_INIT_FUNC_TRACE();
469 TAILQ_INSERT_TAIL(&rte_dpaa_bus.driver_list, driver, next);
470 /* Update Bus references */
471 driver->dpaa_bus = &rte_dpaa_bus;
474 /* un-register a dpaa bus based dpaa driver */
476 rte_dpaa_driver_unregister(struct rte_dpaa_driver *driver)
478 struct rte_dpaa_bus *dpaa_bus;
480 BUS_INIT_FUNC_TRACE();
482 dpaa_bus = driver->dpaa_bus;
484 TAILQ_REMOVE(&dpaa_bus->driver_list, driver, next);
485 /* Update Bus references */
486 driver->dpaa_bus = NULL;
490 rte_dpaa_device_match(struct rte_dpaa_driver *drv,
491 struct rte_dpaa_device *dev)
494 DPAA_BUS_DEBUG("Invalid drv or dev received.");
498 if (drv->drv_type == dev->device_type)
505 rte_dpaa_bus_dev_build(void)
509 /* Load the device-tree driver */
512 DPAA_BUS_LOG(ERR, "of_init failed with ret: %d", ret);
516 /* Get the interface configurations from device-tree */
517 dpaa_netcfg = netcfg_acquire();
519 DPAA_BUS_LOG(ERR, "netcfg_acquire failed");
523 RTE_LOG(NOTICE, EAL, "DPAA Bus Detected\n");
525 if (!dpaa_netcfg->num_ethports) {
526 DPAA_BUS_LOG(INFO, "no network interfaces available");
527 /* This is not an error */
531 #ifdef RTE_LIBRTE_DPAA_DEBUG_DRIVER
532 dump_netcfg(dpaa_netcfg);
535 DPAA_BUS_LOG(DEBUG, "Number of ethernet devices = %d",
536 dpaa_netcfg->num_ethports);
537 ret = dpaa_create_device_list();
539 DPAA_BUS_LOG(ERR, "Unable to create device list. (%d)", ret);
546 rte_dpaa_bus_probe(void)
549 struct rte_dpaa_device *dev;
550 struct rte_dpaa_driver *drv;
551 FILE *svr_file = NULL;
552 unsigned int svr_ver;
553 int probe_all = rte_dpaa_bus.bus.conf.scan_mode != RTE_BUS_SCAN_WHITELIST;
555 /* If DPAA bus is not present nothing needs to be done */
556 if (!rte_dpaa_bus.detected)
559 rte_dpaa_bus_dev_build();
561 /* If no device present on DPAA bus nothing needs to be done */
562 if (TAILQ_EMPTY(&rte_dpaa_bus.device_list))
565 svr_file = fopen(DPAA_SOC_ID_FILE, "r");
567 if (fscanf(svr_file, "svr:%x", &svr_ver) > 0)
568 dpaa_svr_family = svr_ver & SVR_MASK;
572 /* And initialize the PA->VA translation table */
573 dpaax_iova_table_populate();
575 /* For each registered driver, and device, call the driver->probe */
576 TAILQ_FOREACH(dev, &rte_dpaa_bus.device_list, next) {
577 TAILQ_FOREACH(drv, &rte_dpaa_bus.driver_list, next) {
578 ret = rte_dpaa_device_match(drv, dev);
582 if (rte_dev_is_probed(&dev->device))
586 (dev->device.devargs &&
587 dev->device.devargs->policy == RTE_DEV_BLACKLISTED))
591 (dev->device.devargs &&
592 dev->device.devargs->policy ==
593 RTE_DEV_WHITELISTED)) {
594 ret = drv->probe(drv, dev);
596 DPAA_BUS_ERR("Unable to probe.\n");
599 dev->device.driver = &drv->driver;
606 /* Register DPAA mempool ops only if any DPAA device has
609 rte_mbuf_set_platform_mempool_ops(DPAA_MEMPOOL_OPS_NAME);
614 static struct rte_device *
615 rte_dpaa_find_device(const struct rte_device *start, rte_dev_cmp_t cmp,
618 struct rte_dpaa_device *dev;
620 TAILQ_FOREACH(dev, &rte_dpaa_bus.device_list, next) {
621 if (start && &dev->device == start) {
622 start = NULL; /* starting point found */
626 if (cmp(&dev->device, data) == 0)
634 * Get iommu class of DPAA2 devices on the bus.
636 static enum rte_iova_mode
637 rte_dpaa_get_iommu_class(void)
639 if ((access(DPAA_DEV_PATH1, F_OK) != 0) &&
640 (access(DPAA_DEV_PATH2, F_OK) != 0)) {
646 static struct rte_dpaa_bus rte_dpaa_bus = {
648 .scan = rte_dpaa_bus_scan,
649 .probe = rte_dpaa_bus_probe,
650 .parse = rte_dpaa_bus_parse,
651 .find_device = rte_dpaa_find_device,
652 .get_iommu_class = rte_dpaa_get_iommu_class,
654 .device_list = TAILQ_HEAD_INITIALIZER(rte_dpaa_bus.device_list),
655 .driver_list = TAILQ_HEAD_INITIALIZER(rte_dpaa_bus.driver_list),
659 RTE_REGISTER_BUS(FSL_DPAA_BUS_NAME, rte_dpaa_bus.bus);
661 RTE_INIT(dpaa_init_log)
663 dpaa_logtype_bus = rte_log_register("bus.dpaa");
664 if (dpaa_logtype_bus >= 0)
665 rte_log_set_level(dpaa_logtype_bus, RTE_LOG_NOTICE);
667 dpaa_logtype_mempool = rte_log_register("mempool.dpaa");
668 if (dpaa_logtype_mempool >= 0)
669 rte_log_set_level(dpaa_logtype_mempool, RTE_LOG_NOTICE);
671 dpaa_logtype_pmd = rte_log_register("pmd.net.dpaa");
672 if (dpaa_logtype_pmd >= 0)
673 rte_log_set_level(dpaa_logtype_pmd, RTE_LOG_NOTICE);
675 dpaa_logtype_eventdev = rte_log_register("pmd.event.dpaa");
676 if (dpaa_logtype_eventdev >= 0)
677 rte_log_set_level(dpaa_logtype_eventdev, RTE_LOG_NOTICE);