5a018d4871fb9285b10fa4c1702b6dd33ea994db
[dpdk.git] / drivers / event / dpaa / dpaa_eventdev.c
1 /*   SPDX-License-Identifier:        BSD-3-Clause
2  *   Copyright 2017-2019 NXP
3  */
4
5 #include <assert.h>
6 #include <stdio.h>
7 #include <stdbool.h>
8 #include <errno.h>
9 #include <stdint.h>
10 #include <string.h>
11 #include <sys/epoll.h>
12
13 #include <rte_atomic.h>
14 #include <rte_byteorder.h>
15 #include <rte_common.h>
16 #include <rte_debug.h>
17 #include <rte_dev.h>
18 #include <rte_eal.h>
19 #include <rte_lcore.h>
20 #include <rte_log.h>
21 #include <rte_malloc.h>
22 #include <rte_memcpy.h>
23 #include <rte_memory.h>
24 #include <rte_memzone.h>
25 #include <rte_pci.h>
26 #include <rte_eventdev.h>
27 #include <rte_eventdev_pmd_vdev.h>
28 #include <rte_ethdev.h>
29 #include <rte_event_eth_rx_adapter.h>
30 #include <rte_event_eth_tx_adapter.h>
31 #include <rte_cryptodev.h>
32 #include <rte_dpaa_bus.h>
33 #include <rte_dpaa_logs.h>
34 #include <rte_cycles.h>
35 #include <rte_kvargs.h>
36
37 #include <dpaa_ethdev.h>
38 #include <dpaa_sec_event.h>
39 #include "dpaa_eventdev.h"
40 #include <dpaa_mempool.h>
41
42 /*
43  * Clarifications
44  * Evendev = Virtual Instance for SoC
45  * Eventport = Portal Instance
46  * Eventqueue = Channel Instance
47  * 1 Eventdev can have N Eventqueue
48  */
49 int dpaa_logtype_eventdev;
50
51 #define DISABLE_INTR_MODE "disable_intr"
52
53 static int
54 dpaa_event_dequeue_timeout_ticks(struct rte_eventdev *dev, uint64_t ns,
55                                  uint64_t *timeout_ticks)
56 {
57         EVENTDEV_INIT_FUNC_TRACE();
58
59         RTE_SET_USED(dev);
60
61         uint64_t cycles_per_second;
62
63         cycles_per_second = rte_get_timer_hz();
64         *timeout_ticks = (ns * cycles_per_second) / NS_PER_S;
65
66         return 0;
67 }
68
69 static int
70 dpaa_event_dequeue_timeout_ticks_intr(struct rte_eventdev *dev, uint64_t ns,
71                                  uint64_t *timeout_ticks)
72 {
73         RTE_SET_USED(dev);
74
75         *timeout_ticks = ns/1000;
76         return 0;
77 }
78
79 static void
80 dpaa_eventq_portal_add(u16 ch_id)
81 {
82         uint32_t sdqcr;
83
84         sdqcr = QM_SDQCR_CHANNELS_POOL_CONV(ch_id);
85         qman_static_dequeue_add(sdqcr, NULL);
86 }
87
88 static uint16_t
89 dpaa_event_enqueue_burst(void *port, const struct rte_event ev[],
90                          uint16_t nb_events)
91 {
92         uint16_t i;
93         struct rte_mbuf *mbuf;
94
95         RTE_SET_USED(port);
96         /*Release all the contexts saved previously*/
97         for (i = 0; i < nb_events; i++) {
98                 switch (ev[i].op) {
99                 case RTE_EVENT_OP_RELEASE:
100                         qman_dca_index(ev[i].impl_opaque, 0);
101                         mbuf = DPAA_PER_LCORE_DQRR_MBUF(i);
102                         mbuf->seqn = DPAA_INVALID_MBUF_SEQN;
103                         DPAA_PER_LCORE_DQRR_HELD &= ~(1 << i);
104                         DPAA_PER_LCORE_DQRR_SIZE--;
105                         break;
106                 default:
107                         break;
108                 }
109         }
110
111         return nb_events;
112 }
113
114 static uint16_t
115 dpaa_event_enqueue(void *port, const struct rte_event *ev)
116 {
117         return dpaa_event_enqueue_burst(port, ev, 1);
118 }
119
120 static void drain_4_bytes(int fd, fd_set *fdset)
121 {
122         if (FD_ISSET(fd, fdset)) {
123                 /* drain 4 bytes */
124                 uint32_t junk;
125                 ssize_t sjunk = read(qman_thread_fd(), &junk, sizeof(junk));
126                 if (sjunk != sizeof(junk))
127                         DPAA_EVENTDEV_ERR("UIO irq read error");
128         }
129 }
130
131 static inline int
132 dpaa_event_dequeue_wait(uint64_t timeout_ticks)
133 {
134         int fd_qman, nfds;
135         int ret;
136         fd_set readset;
137
138         /* Go into (and back out of) IRQ mode for each select,
139          * it simplifies exit-path considerations and other
140          * potential nastiness.
141          */
142         struct timeval tv = {
143                 .tv_sec = timeout_ticks / 1000000,
144                 .tv_usec = timeout_ticks % 1000000
145         };
146
147         fd_qman = qman_thread_fd();
148         nfds = fd_qman + 1;
149         FD_ZERO(&readset);
150         FD_SET(fd_qman, &readset);
151
152         qman_irqsource_add(QM_PIRQ_DQRI);
153
154         ret = select(nfds, &readset, NULL, NULL, &tv);
155         if (ret < 0)
156                 return ret;
157         /* Calling irqsource_remove() prior to thread_irq()
158          * means thread_irq() will not process whatever caused
159          * the interrupts, however it does ensure that, once
160          * thread_irq() re-enables interrupts, they won't fire
161          * again immediately.
162          */
163         qman_irqsource_remove(~0);
164         drain_4_bytes(fd_qman, &readset);
165         qman_thread_irq();
166
167         return ret;
168 }
169
170 static uint16_t
171 dpaa_event_dequeue_burst(void *port, struct rte_event ev[],
172                          uint16_t nb_events, uint64_t timeout_ticks)
173 {
174         int ret;
175         u16 ch_id;
176         void *buffers[8];
177         u32 num_frames, i, irq = 0;
178         uint64_t cur_ticks = 0, wait_time_ticks = 0;
179         struct dpaa_port *portal = (struct dpaa_port *)port;
180         struct rte_mbuf *mbuf;
181
182         if (unlikely(!RTE_PER_LCORE(dpaa_io))) {
183                 /* Affine current thread context to a qman portal */
184                 ret = rte_dpaa_portal_init((void *)0);
185                 if (ret) {
186                         DPAA_EVENTDEV_ERR("Unable to initialize portal");
187                         return ret;
188                 }
189         }
190
191         if (unlikely(!portal->is_port_linked)) {
192                 /*
193                  * Affine event queue for current thread context
194                  * to a qman portal.
195                  */
196                 for (i = 0; i < portal->num_linked_evq; i++) {
197                         ch_id = portal->evq_info[i].ch_id;
198                         dpaa_eventq_portal_add(ch_id);
199                 }
200                 portal->is_port_linked = true;
201         }
202
203         /* Check if there are atomic contexts to be released */
204         i = 0;
205         while (DPAA_PER_LCORE_DQRR_SIZE) {
206                 if (DPAA_PER_LCORE_DQRR_HELD & (1 << i)) {
207                         qman_dca_index(i, 0);
208                         mbuf = DPAA_PER_LCORE_DQRR_MBUF(i);
209                         mbuf->seqn = DPAA_INVALID_MBUF_SEQN;
210                         DPAA_PER_LCORE_DQRR_HELD &= ~(1 << i);
211                         DPAA_PER_LCORE_DQRR_SIZE--;
212                 }
213                 i++;
214         }
215         DPAA_PER_LCORE_DQRR_HELD = 0;
216
217         if (timeout_ticks)
218                 wait_time_ticks = timeout_ticks;
219         else
220                 wait_time_ticks = portal->timeout_us;
221
222         wait_time_ticks += rte_get_timer_cycles();
223         do {
224                 /* Lets dequeue the frames */
225                 num_frames = qman_portal_dequeue(ev, nb_events, buffers);
226                 if (irq)
227                         irq = 0;
228                 if (num_frames)
229                         break;
230                 cur_ticks = rte_get_timer_cycles();
231         } while (cur_ticks < wait_time_ticks);
232
233         return num_frames;
234 }
235
236 static uint16_t
237 dpaa_event_dequeue(void *port, struct rte_event *ev, uint64_t timeout_ticks)
238 {
239         return dpaa_event_dequeue_burst(port, ev, 1, timeout_ticks);
240 }
241
242 static uint16_t
243 dpaa_event_dequeue_burst_intr(void *port, struct rte_event ev[],
244                               uint16_t nb_events, uint64_t timeout_ticks)
245 {
246         int ret;
247         u16 ch_id;
248         void *buffers[8];
249         u32 num_frames, i, irq = 0;
250         uint64_t cur_ticks = 0, wait_time_ticks = 0;
251         struct dpaa_port *portal = (struct dpaa_port *)port;
252         struct rte_mbuf *mbuf;
253
254         if (unlikely(!RTE_PER_LCORE(dpaa_io))) {
255                 /* Affine current thread context to a qman portal */
256                 ret = rte_dpaa_portal_init((void *)0);
257                 if (ret) {
258                         DPAA_EVENTDEV_ERR("Unable to initialize portal");
259                         return ret;
260                 }
261         }
262
263         if (unlikely(!portal->is_port_linked)) {
264                 /*
265                  * Affine event queue for current thread context
266                  * to a qman portal.
267                  */
268                 for (i = 0; i < portal->num_linked_evq; i++) {
269                         ch_id = portal->evq_info[i].ch_id;
270                         dpaa_eventq_portal_add(ch_id);
271                 }
272                 portal->is_port_linked = true;
273         }
274
275         /* Check if there are atomic contexts to be released */
276         i = 0;
277         while (DPAA_PER_LCORE_DQRR_SIZE) {
278                 if (DPAA_PER_LCORE_DQRR_HELD & (1 << i)) {
279                         qman_dca_index(i, 0);
280                         mbuf = DPAA_PER_LCORE_DQRR_MBUF(i);
281                         mbuf->seqn = DPAA_INVALID_MBUF_SEQN;
282                         DPAA_PER_LCORE_DQRR_HELD &= ~(1 << i);
283                         DPAA_PER_LCORE_DQRR_SIZE--;
284                 }
285                 i++;
286         }
287         DPAA_PER_LCORE_DQRR_HELD = 0;
288
289         if (timeout_ticks)
290                 wait_time_ticks = timeout_ticks;
291         else
292                 wait_time_ticks = portal->timeout_us;
293
294         do {
295                 /* Lets dequeue the frames */
296                 num_frames = qman_portal_dequeue(ev, nb_events, buffers);
297                 if (irq)
298                         irq = 0;
299                 if (num_frames)
300                         break;
301                 if (wait_time_ticks) { /* wait for time */
302                         if (dpaa_event_dequeue_wait(wait_time_ticks) > 0) {
303                                 irq = 1;
304                                 continue;
305                         }
306                         break; /* no event after waiting */
307                 }
308                 cur_ticks = rte_get_timer_cycles();
309         } while (cur_ticks < wait_time_ticks);
310
311         return num_frames;
312 }
313
314 static uint16_t
315 dpaa_event_dequeue_intr(void *port,
316                         struct rte_event *ev,
317                         uint64_t timeout_ticks)
318 {
319         return dpaa_event_dequeue_burst_intr(port, ev, 1, timeout_ticks);
320 }
321
322 static void
323 dpaa_event_dev_info_get(struct rte_eventdev *dev,
324                         struct rte_event_dev_info *dev_info)
325 {
326         EVENTDEV_INIT_FUNC_TRACE();
327
328         RTE_SET_USED(dev);
329         dev_info->driver_name = "event_dpaa1";
330         dev_info->min_dequeue_timeout_ns =
331                 DPAA_EVENT_MIN_DEQUEUE_TIMEOUT;
332         dev_info->max_dequeue_timeout_ns =
333                 DPAA_EVENT_MAX_DEQUEUE_TIMEOUT;
334         dev_info->dequeue_timeout_ns =
335                 DPAA_EVENT_PORT_DEQUEUE_TIMEOUT_NS;
336         dev_info->max_event_queues =
337                 DPAA_EVENT_MAX_QUEUES;
338         dev_info->max_event_queue_flows =
339                 DPAA_EVENT_MAX_QUEUE_FLOWS;
340         dev_info->max_event_queue_priority_levels =
341                 DPAA_EVENT_MAX_QUEUE_PRIORITY_LEVELS;
342         dev_info->max_event_priority_levels =
343                 DPAA_EVENT_MAX_EVENT_PRIORITY_LEVELS;
344         dev_info->max_event_ports =
345                 DPAA_EVENT_MAX_EVENT_PORT;
346         dev_info->max_event_port_dequeue_depth =
347                 DPAA_EVENT_MAX_PORT_DEQUEUE_DEPTH;
348         dev_info->max_event_port_enqueue_depth =
349                 DPAA_EVENT_MAX_PORT_ENQUEUE_DEPTH;
350         /*
351          * TODO: Need to find out that how to fetch this info
352          * from kernel or somewhere else.
353          */
354         dev_info->max_num_events =
355                 DPAA_EVENT_MAX_NUM_EVENTS;
356         dev_info->event_dev_cap =
357                 RTE_EVENT_DEV_CAP_DISTRIBUTED_SCHED |
358                 RTE_EVENT_DEV_CAP_BURST_MODE |
359                 RTE_EVENT_DEV_CAP_MULTIPLE_QUEUE_PORT |
360                 RTE_EVENT_DEV_CAP_NONSEQ_MODE;
361 }
362
363 static int
364 dpaa_event_dev_configure(const struct rte_eventdev *dev)
365 {
366         struct dpaa_eventdev *priv = dev->data->dev_private;
367         struct rte_event_dev_config *conf = &dev->data->dev_conf;
368         int ret, i;
369         uint32_t *ch_id;
370
371         EVENTDEV_INIT_FUNC_TRACE();
372         priv->dequeue_timeout_ns = conf->dequeue_timeout_ns;
373         priv->nb_events_limit = conf->nb_events_limit;
374         priv->nb_event_queues = conf->nb_event_queues;
375         priv->nb_event_ports = conf->nb_event_ports;
376         priv->nb_event_queue_flows = conf->nb_event_queue_flows;
377         priv->nb_event_port_dequeue_depth = conf->nb_event_port_dequeue_depth;
378         priv->nb_event_port_enqueue_depth = conf->nb_event_port_enqueue_depth;
379         priv->event_dev_cfg = conf->event_dev_cfg;
380
381         ch_id = rte_malloc("dpaa-channels",
382                           sizeof(uint32_t) * priv->nb_event_queues,
383                           RTE_CACHE_LINE_SIZE);
384         if (ch_id == NULL) {
385                 DPAA_EVENTDEV_ERR("Fail to allocate memory for dpaa channels\n");
386                 return -ENOMEM;
387         }
388         /* Create requested event queues within the given event device */
389         ret = qman_alloc_pool_range(ch_id, priv->nb_event_queues, 1, 0);
390         if (ret < 0) {
391                 DPAA_EVENTDEV_ERR("qman_alloc_pool_range %u, err =%d\n",
392                                  priv->nb_event_queues, ret);
393                 rte_free(ch_id);
394                 return ret;
395         }
396         for (i = 0; i < priv->nb_event_queues; i++)
397                 priv->evq_info[i].ch_id = (u16)ch_id[i];
398
399         /* Lets prepare event ports */
400         memset(&priv->ports[0], 0,
401               sizeof(struct dpaa_port) * priv->nb_event_ports);
402
403         /* Check dequeue timeout method is per dequeue or global */
404         if (priv->event_dev_cfg & RTE_EVENT_DEV_CFG_PER_DEQUEUE_TIMEOUT) {
405                 /*
406                  * Use timeout value as given in dequeue operation.
407                  * So invalidating this timeout value.
408                  */
409                 priv->dequeue_timeout_ns = 0;
410
411         } else if (conf->dequeue_timeout_ns == 0) {
412                 priv->dequeue_timeout_ns = DPAA_EVENT_PORT_DEQUEUE_TIMEOUT_NS;
413         } else {
414                 priv->dequeue_timeout_ns = conf->dequeue_timeout_ns;
415         }
416
417         for (i = 0; i < priv->nb_event_ports; i++) {
418                 if (priv->intr_mode) {
419                         priv->ports[i].timeout_us =
420                                 priv->dequeue_timeout_ns/1000;
421                 } else {
422                         uint64_t cycles_per_second;
423
424                         cycles_per_second = rte_get_timer_hz();
425                         priv->ports[i].timeout_us =
426                                 (priv->dequeue_timeout_ns * cycles_per_second)
427                                         / NS_PER_S;
428                 }
429         }
430
431         /*
432          * TODO: Currently portals are affined with threads. Maximum threads
433          * can be created equals to number of lcore.
434          */
435         rte_free(ch_id);
436         DPAA_EVENTDEV_INFO("Configured eventdev devid=%d", dev->data->dev_id);
437
438         return 0;
439 }
440
441 static int
442 dpaa_event_dev_start(struct rte_eventdev *dev)
443 {
444         EVENTDEV_INIT_FUNC_TRACE();
445         RTE_SET_USED(dev);
446
447         return 0;
448 }
449
450 static void
451 dpaa_event_dev_stop(struct rte_eventdev *dev)
452 {
453         EVENTDEV_INIT_FUNC_TRACE();
454         RTE_SET_USED(dev);
455 }
456
457 static int
458 dpaa_event_dev_close(struct rte_eventdev *dev)
459 {
460         EVENTDEV_INIT_FUNC_TRACE();
461         RTE_SET_USED(dev);
462
463         return 0;
464 }
465
466 static void
467 dpaa_event_queue_def_conf(struct rte_eventdev *dev, uint8_t queue_id,
468                           struct rte_event_queue_conf *queue_conf)
469 {
470         EVENTDEV_INIT_FUNC_TRACE();
471
472         RTE_SET_USED(dev);
473         RTE_SET_USED(queue_id);
474
475         memset(queue_conf, 0, sizeof(struct rte_event_queue_conf));
476         queue_conf->nb_atomic_flows = DPAA_EVENT_QUEUE_ATOMIC_FLOWS;
477         queue_conf->schedule_type = RTE_SCHED_TYPE_PARALLEL;
478         queue_conf->priority = RTE_EVENT_DEV_PRIORITY_HIGHEST;
479 }
480
481 static int
482 dpaa_event_queue_setup(struct rte_eventdev *dev, uint8_t queue_id,
483                        const struct rte_event_queue_conf *queue_conf)
484 {
485         struct dpaa_eventdev *priv = dev->data->dev_private;
486         struct dpaa_eventq *evq_info = &priv->evq_info[queue_id];
487
488         EVENTDEV_INIT_FUNC_TRACE();
489
490         switch (queue_conf->schedule_type) {
491         case RTE_SCHED_TYPE_PARALLEL:
492         case RTE_SCHED_TYPE_ATOMIC:
493                 break;
494         case RTE_SCHED_TYPE_ORDERED:
495                 DPAA_EVENTDEV_ERR("Schedule type is not supported.");
496                 return -1;
497         }
498         evq_info->event_queue_cfg = queue_conf->event_queue_cfg;
499         evq_info->event_queue_id = queue_id;
500
501         return 0;
502 }
503
504 static void
505 dpaa_event_queue_release(struct rte_eventdev *dev, uint8_t queue_id)
506 {
507         EVENTDEV_INIT_FUNC_TRACE();
508
509         RTE_SET_USED(dev);
510         RTE_SET_USED(queue_id);
511 }
512
513 static void
514 dpaa_event_port_default_conf_get(struct rte_eventdev *dev, uint8_t port_id,
515                                  struct rte_event_port_conf *port_conf)
516 {
517         EVENTDEV_INIT_FUNC_TRACE();
518
519         RTE_SET_USED(dev);
520         RTE_SET_USED(port_id);
521
522         port_conf->new_event_threshold = DPAA_EVENT_MAX_NUM_EVENTS;
523         port_conf->dequeue_depth = DPAA_EVENT_MAX_PORT_DEQUEUE_DEPTH;
524         port_conf->enqueue_depth = DPAA_EVENT_MAX_PORT_ENQUEUE_DEPTH;
525 }
526
527 static int
528 dpaa_event_port_setup(struct rte_eventdev *dev, uint8_t port_id,
529                       const struct rte_event_port_conf *port_conf)
530 {
531         struct dpaa_eventdev *eventdev = dev->data->dev_private;
532
533         EVENTDEV_INIT_FUNC_TRACE();
534
535         RTE_SET_USED(port_conf);
536         dev->data->ports[port_id] = &eventdev->ports[port_id];
537
538         return 0;
539 }
540
541 static void
542 dpaa_event_port_release(void *port)
543 {
544         EVENTDEV_INIT_FUNC_TRACE();
545
546         RTE_SET_USED(port);
547 }
548
549 static int
550 dpaa_event_port_link(struct rte_eventdev *dev, void *port,
551                      const uint8_t queues[], const uint8_t priorities[],
552                      uint16_t nb_links)
553 {
554         struct dpaa_eventdev *priv = dev->data->dev_private;
555         struct dpaa_port *event_port = (struct dpaa_port *)port;
556         struct dpaa_eventq *event_queue;
557         uint8_t eventq_id;
558         int i;
559
560         RTE_SET_USED(dev);
561         RTE_SET_USED(priorities);
562
563         /* First check that input configuration are valid */
564         for (i = 0; i < nb_links; i++) {
565                 eventq_id = queues[i];
566                 event_queue = &priv->evq_info[eventq_id];
567                 if ((event_queue->event_queue_cfg
568                         & RTE_EVENT_QUEUE_CFG_SINGLE_LINK)
569                         && (event_queue->event_port)) {
570                         return -EINVAL;
571                 }
572         }
573
574         for (i = 0; i < nb_links; i++) {
575                 eventq_id = queues[i];
576                 event_queue = &priv->evq_info[eventq_id];
577                 event_port->evq_info[i].event_queue_id = eventq_id;
578                 event_port->evq_info[i].ch_id = event_queue->ch_id;
579                 event_queue->event_port = port;
580         }
581
582         event_port->num_linked_evq = event_port->num_linked_evq + i;
583
584         return (int)i;
585 }
586
587 static int
588 dpaa_event_port_unlink(struct rte_eventdev *dev, void *port,
589                        uint8_t queues[], uint16_t nb_links)
590 {
591         int i;
592         uint8_t eventq_id;
593         struct dpaa_eventq *event_queue;
594         struct dpaa_eventdev *priv = dev->data->dev_private;
595         struct dpaa_port *event_port = (struct dpaa_port *)port;
596
597         if (!event_port->num_linked_evq)
598                 return nb_links;
599
600         for (i = 0; i < nb_links; i++) {
601                 eventq_id = queues[i];
602                 event_port->evq_info[eventq_id].event_queue_id = -1;
603                 event_port->evq_info[eventq_id].ch_id = 0;
604                 event_queue = &priv->evq_info[eventq_id];
605                 event_queue->event_port = NULL;
606         }
607
608         if (event_port->num_linked_evq)
609                 event_port->num_linked_evq = event_port->num_linked_evq - i;
610
611         return (int)i;
612 }
613
614 static int
615 dpaa_event_eth_rx_adapter_caps_get(const struct rte_eventdev *dev,
616                                    const struct rte_eth_dev *eth_dev,
617                                    uint32_t *caps)
618 {
619         const char *ethdev_driver = eth_dev->device->driver->name;
620
621         EVENTDEV_INIT_FUNC_TRACE();
622
623         RTE_SET_USED(dev);
624
625         if (!strcmp(ethdev_driver, "net_dpaa"))
626                 *caps = RTE_EVENT_ETH_RX_ADAPTER_DPAA_CAP;
627         else
628                 *caps = RTE_EVENT_ETH_RX_ADAPTER_SW_CAP;
629
630         return 0;
631 }
632
633 static int
634 dpaa_event_eth_rx_adapter_queue_add(
635                 const struct rte_eventdev *dev,
636                 const struct rte_eth_dev *eth_dev,
637                 int32_t rx_queue_id,
638                 const struct rte_event_eth_rx_adapter_queue_conf *queue_conf)
639 {
640         struct dpaa_eventdev *eventdev = dev->data->dev_private;
641         uint8_t ev_qid = queue_conf->ev.queue_id;
642         u16 ch_id = eventdev->evq_info[ev_qid].ch_id;
643         struct dpaa_if *dpaa_intf = eth_dev->data->dev_private;
644         int ret, i;
645
646         EVENTDEV_INIT_FUNC_TRACE();
647
648         if (rx_queue_id == -1) {
649                 for (i = 0; i < dpaa_intf->nb_rx_queues; i++) {
650                         ret = dpaa_eth_eventq_attach(eth_dev, i, ch_id,
651                                                      queue_conf);
652                         if (ret) {
653                                 DPAA_EVENTDEV_ERR(
654                                         "Event Queue attach failed:%d\n", ret);
655                                 goto detach_configured_queues;
656                         }
657                 }
658                 return 0;
659         }
660
661         ret = dpaa_eth_eventq_attach(eth_dev, rx_queue_id, ch_id, queue_conf);
662         if (ret)
663                 DPAA_EVENTDEV_ERR("dpaa_eth_eventq_attach failed:%d\n", ret);
664         return ret;
665
666 detach_configured_queues:
667
668         for (i = (i - 1); i >= 0 ; i--)
669                 dpaa_eth_eventq_detach(eth_dev, i);
670
671         return ret;
672 }
673
674 static int
675 dpaa_event_eth_rx_adapter_queue_del(const struct rte_eventdev *dev,
676                                     const struct rte_eth_dev *eth_dev,
677                                     int32_t rx_queue_id)
678 {
679         int ret, i;
680         struct dpaa_if *dpaa_intf = eth_dev->data->dev_private;
681
682         EVENTDEV_INIT_FUNC_TRACE();
683
684         RTE_SET_USED(dev);
685         if (rx_queue_id == -1) {
686                 for (i = 0; i < dpaa_intf->nb_rx_queues; i++) {
687                         ret = dpaa_eth_eventq_detach(eth_dev, i);
688                         if (ret)
689                                 DPAA_EVENTDEV_ERR(
690                                         "Event Queue detach failed:%d\n", ret);
691                 }
692
693                 return 0;
694         }
695
696         ret = dpaa_eth_eventq_detach(eth_dev, rx_queue_id);
697         if (ret)
698                 DPAA_EVENTDEV_ERR("dpaa_eth_eventq_detach failed:%d\n", ret);
699         return ret;
700 }
701
702 static int
703 dpaa_event_eth_rx_adapter_start(const struct rte_eventdev *dev,
704                                 const struct rte_eth_dev *eth_dev)
705 {
706         EVENTDEV_INIT_FUNC_TRACE();
707
708         RTE_SET_USED(dev);
709         RTE_SET_USED(eth_dev);
710
711         return 0;
712 }
713
714 static int
715 dpaa_event_eth_rx_adapter_stop(const struct rte_eventdev *dev,
716                                const struct rte_eth_dev *eth_dev)
717 {
718         EVENTDEV_INIT_FUNC_TRACE();
719
720         RTE_SET_USED(dev);
721         RTE_SET_USED(eth_dev);
722
723         return 0;
724 }
725
726 static int
727 dpaa_eventdev_crypto_caps_get(const struct rte_eventdev *dev,
728                             const struct rte_cryptodev *cdev,
729                             uint32_t *caps)
730 {
731         const char *name = cdev->data->name;
732
733         EVENTDEV_INIT_FUNC_TRACE();
734
735         RTE_SET_USED(dev);
736
737         if (!strncmp(name, "dpaa_sec-", 9))
738                 *caps = RTE_EVENT_CRYPTO_ADAPTER_DPAA_CAP;
739         else
740                 return -1;
741
742         return 0;
743 }
744
745 static int
746 dpaa_eventdev_crypto_queue_add_all(const struct rte_eventdev *dev,
747                 const struct rte_cryptodev *cryptodev,
748                 const struct rte_event *ev)
749 {
750         struct dpaa_eventdev *priv = dev->data->dev_private;
751         uint8_t ev_qid = ev->queue_id;
752         u16 ch_id = priv->evq_info[ev_qid].ch_id;
753         int i, ret;
754
755         EVENTDEV_INIT_FUNC_TRACE();
756
757         for (i = 0; i < cryptodev->data->nb_queue_pairs; i++) {
758                 ret = dpaa_sec_eventq_attach(cryptodev, i,
759                                 ch_id, ev);
760                 if (ret) {
761                         DPAA_EVENTDEV_ERR("dpaa_sec_eventq_attach failed: ret %d\n",
762                                     ret);
763                         goto fail;
764                 }
765         }
766         return 0;
767 fail:
768         for (i = (i - 1); i >= 0 ; i--)
769                 dpaa_sec_eventq_detach(cryptodev, i);
770
771         return ret;
772 }
773
774 static int
775 dpaa_eventdev_crypto_queue_add(const struct rte_eventdev *dev,
776                 const struct rte_cryptodev *cryptodev,
777                 int32_t rx_queue_id,
778                 const struct rte_event *ev)
779 {
780         struct dpaa_eventdev *priv = dev->data->dev_private;
781         uint8_t ev_qid = ev->queue_id;
782         u16 ch_id = priv->evq_info[ev_qid].ch_id;
783         int ret;
784
785         EVENTDEV_INIT_FUNC_TRACE();
786
787         if (rx_queue_id == -1)
788                 return dpaa_eventdev_crypto_queue_add_all(dev,
789                                 cryptodev, ev);
790
791         ret = dpaa_sec_eventq_attach(cryptodev, rx_queue_id,
792                         ch_id, ev);
793         if (ret) {
794                 DPAA_EVENTDEV_ERR(
795                         "dpaa_sec_eventq_attach failed: ret: %d\n", ret);
796                 return ret;
797         }
798         return 0;
799 }
800
801 static int
802 dpaa_eventdev_crypto_queue_del_all(const struct rte_eventdev *dev,
803                              const struct rte_cryptodev *cdev)
804 {
805         int i, ret;
806
807         EVENTDEV_INIT_FUNC_TRACE();
808
809         RTE_SET_USED(dev);
810
811         for (i = 0; i < cdev->data->nb_queue_pairs; i++) {
812                 ret = dpaa_sec_eventq_detach(cdev, i);
813                 if (ret) {
814                         DPAA_EVENTDEV_ERR(
815                                 "dpaa_sec_eventq_detach failed:ret %d\n", ret);
816                         return ret;
817                 }
818         }
819
820         return 0;
821 }
822
823 static int
824 dpaa_eventdev_crypto_queue_del(const struct rte_eventdev *dev,
825                              const struct rte_cryptodev *cryptodev,
826                              int32_t rx_queue_id)
827 {
828         int ret;
829
830         EVENTDEV_INIT_FUNC_TRACE();
831
832         if (rx_queue_id == -1)
833                 return dpaa_eventdev_crypto_queue_del_all(dev, cryptodev);
834
835         ret = dpaa_sec_eventq_detach(cryptodev, rx_queue_id);
836         if (ret) {
837                 DPAA_EVENTDEV_ERR(
838                         "dpaa_sec_eventq_detach failed: ret: %d\n", ret);
839                 return ret;
840         }
841
842         return 0;
843 }
844
845 static int
846 dpaa_eventdev_crypto_start(const struct rte_eventdev *dev,
847                            const struct rte_cryptodev *cryptodev)
848 {
849         EVENTDEV_INIT_FUNC_TRACE();
850
851         RTE_SET_USED(dev);
852         RTE_SET_USED(cryptodev);
853
854         return 0;
855 }
856
857 static int
858 dpaa_eventdev_crypto_stop(const struct rte_eventdev *dev,
859                           const struct rte_cryptodev *cryptodev)
860 {
861         EVENTDEV_INIT_FUNC_TRACE();
862
863         RTE_SET_USED(dev);
864         RTE_SET_USED(cryptodev);
865
866         return 0;
867 }
868
869 static int
870 dpaa_eventdev_tx_adapter_create(uint8_t id,
871                                  const struct rte_eventdev *dev)
872 {
873         RTE_SET_USED(id);
874         RTE_SET_USED(dev);
875
876         /* Nothing to do. Simply return. */
877         return 0;
878 }
879
880 static int
881 dpaa_eventdev_tx_adapter_caps(const struct rte_eventdev *dev,
882                                const struct rte_eth_dev *eth_dev,
883                                uint32_t *caps)
884 {
885         RTE_SET_USED(dev);
886         RTE_SET_USED(eth_dev);
887
888         *caps = RTE_EVENT_ETH_TX_ADAPTER_CAP_INTERNAL_PORT;
889         return 0;
890 }
891
892 static uint16_t
893 dpaa_eventdev_txa_enqueue_same_dest(void *port,
894                                      struct rte_event ev[],
895                                      uint16_t nb_events)
896 {
897         struct rte_mbuf *m[DPAA_EVENT_MAX_PORT_ENQUEUE_DEPTH], *m0;
898         uint8_t qid, i;
899
900         RTE_SET_USED(port);
901
902         m0 = (struct rte_mbuf *)ev[0].mbuf;
903         qid = rte_event_eth_tx_adapter_txq_get(m0);
904
905         for (i = 0; i < nb_events; i++)
906                 m[i] = (struct rte_mbuf *)ev[i].mbuf;
907
908         return rte_eth_tx_burst(m0->port, qid, m, nb_events);
909 }
910
911 static uint16_t
912 dpaa_eventdev_txa_enqueue(void *port,
913                            struct rte_event ev[],
914                            uint16_t nb_events)
915 {
916         struct rte_mbuf *m = (struct rte_mbuf *)ev[0].mbuf;
917         uint8_t qid, i;
918
919         RTE_SET_USED(port);
920
921         for (i = 0; i < nb_events; i++) {
922                 qid = rte_event_eth_tx_adapter_txq_get(m);
923                 rte_eth_tx_burst(m->port, qid, &m, 1);
924         }
925
926         return nb_events;
927 }
928
929 static struct rte_eventdev_ops dpaa_eventdev_ops = {
930         .dev_infos_get    = dpaa_event_dev_info_get,
931         .dev_configure    = dpaa_event_dev_configure,
932         .dev_start        = dpaa_event_dev_start,
933         .dev_stop         = dpaa_event_dev_stop,
934         .dev_close        = dpaa_event_dev_close,
935         .queue_def_conf   = dpaa_event_queue_def_conf,
936         .queue_setup      = dpaa_event_queue_setup,
937         .queue_release    = dpaa_event_queue_release,
938         .port_def_conf    = dpaa_event_port_default_conf_get,
939         .port_setup       = dpaa_event_port_setup,
940         .port_release       = dpaa_event_port_release,
941         .port_link        = dpaa_event_port_link,
942         .port_unlink      = dpaa_event_port_unlink,
943         .timeout_ticks    = dpaa_event_dequeue_timeout_ticks,
944         .eth_rx_adapter_caps_get        = dpaa_event_eth_rx_adapter_caps_get,
945         .eth_rx_adapter_queue_add       = dpaa_event_eth_rx_adapter_queue_add,
946         .eth_rx_adapter_queue_del       = dpaa_event_eth_rx_adapter_queue_del,
947         .eth_rx_adapter_start           = dpaa_event_eth_rx_adapter_start,
948         .eth_rx_adapter_stop            = dpaa_event_eth_rx_adapter_stop,
949         .eth_tx_adapter_caps_get        = dpaa_eventdev_tx_adapter_caps,
950         .eth_tx_adapter_create          = dpaa_eventdev_tx_adapter_create,
951         .crypto_adapter_caps_get        = dpaa_eventdev_crypto_caps_get,
952         .crypto_adapter_queue_pair_add  = dpaa_eventdev_crypto_queue_add,
953         .crypto_adapter_queue_pair_del  = dpaa_eventdev_crypto_queue_del,
954         .crypto_adapter_start           = dpaa_eventdev_crypto_start,
955         .crypto_adapter_stop            = dpaa_eventdev_crypto_stop,
956 };
957
958 static int flag_check_handler(__rte_unused const char *key,
959                 const char *value, __rte_unused void *opaque)
960 {
961         if (strcmp(value, "1"))
962                 return -1;
963
964         return 0;
965 }
966
967 static int
968 dpaa_event_check_flags(const char *params)
969 {
970         struct rte_kvargs *kvlist;
971
972         if (params == NULL || params[0] == '\0')
973                 return 0;
974
975         kvlist = rte_kvargs_parse(params, NULL);
976         if (kvlist == NULL)
977                 return 0;
978
979         if (!rte_kvargs_count(kvlist, DISABLE_INTR_MODE)) {
980                 rte_kvargs_free(kvlist);
981                 return 0;
982         }
983         /* INTR MODE is disabled when there's key-value pair: disable_intr = 1*/
984         if (rte_kvargs_process(kvlist, DISABLE_INTR_MODE,
985                                 flag_check_handler, NULL) < 0) {
986                 rte_kvargs_free(kvlist);
987                 return 0;
988         }
989         rte_kvargs_free(kvlist);
990
991         return 1;
992 }
993
994 static int
995 dpaa_event_dev_create(const char *name, const char *params)
996 {
997         struct rte_eventdev *eventdev;
998         struct dpaa_eventdev *priv;
999
1000         eventdev = rte_event_pmd_vdev_init(name,
1001                                            sizeof(struct dpaa_eventdev),
1002                                            rte_socket_id());
1003         if (eventdev == NULL) {
1004                 DPAA_EVENTDEV_ERR("Failed to create eventdev vdev %s", name);
1005                 goto fail;
1006         }
1007         priv = eventdev->data->dev_private;
1008
1009         eventdev->dev_ops       = &dpaa_eventdev_ops;
1010         eventdev->enqueue       = dpaa_event_enqueue;
1011         eventdev->enqueue_burst = dpaa_event_enqueue_burst;
1012
1013         if (dpaa_event_check_flags(params)) {
1014                 eventdev->dequeue       = dpaa_event_dequeue;
1015                 eventdev->dequeue_burst = dpaa_event_dequeue_burst;
1016         } else {
1017                 priv->intr_mode = 1;
1018                 eventdev->dev_ops->timeout_ticks =
1019                                 dpaa_event_dequeue_timeout_ticks_intr;
1020                 eventdev->dequeue       = dpaa_event_dequeue_intr;
1021                 eventdev->dequeue_burst = dpaa_event_dequeue_burst_intr;
1022         }
1023         eventdev->txa_enqueue = dpaa_eventdev_txa_enqueue;
1024         eventdev->txa_enqueue_same_dest = dpaa_eventdev_txa_enqueue_same_dest;
1025
1026         RTE_LOG(INFO, PMD, "%s eventdev added", name);
1027
1028         /* For secondary processes, the primary has done all the work */
1029         if (rte_eal_process_type() != RTE_PROC_PRIMARY)
1030                 return 0;
1031
1032         priv->max_event_queues = DPAA_EVENT_MAX_QUEUES;
1033
1034         return 0;
1035 fail:
1036         return -EFAULT;
1037 }
1038
1039 static int
1040 dpaa_event_dev_probe(struct rte_vdev_device *vdev)
1041 {
1042         const char *name;
1043         const char *params;
1044
1045         name = rte_vdev_device_name(vdev);
1046         DPAA_EVENTDEV_INFO("Initializing %s", name);
1047
1048         params = rte_vdev_device_args(vdev);
1049
1050         return dpaa_event_dev_create(name, params);
1051 }
1052
1053 static int
1054 dpaa_event_dev_remove(struct rte_vdev_device *vdev)
1055 {
1056         const char *name;
1057
1058         name = rte_vdev_device_name(vdev);
1059         DPAA_EVENTDEV_INFO("Closing %s", name);
1060
1061         return rte_event_pmd_vdev_uninit(name);
1062 }
1063
1064 static struct rte_vdev_driver vdev_eventdev_dpaa_pmd = {
1065         .probe = dpaa_event_dev_probe,
1066         .remove = dpaa_event_dev_remove
1067 };
1068
1069 RTE_PMD_REGISTER_VDEV(EVENTDEV_NAME_DPAA_PMD, vdev_eventdev_dpaa_pmd);
1070 RTE_PMD_REGISTER_PARAM_STRING(EVENTDEV_NAME_DPAA_PMD,
1071                 DISABLE_INTR_MODE "=<int>");
1072 RTE_INIT(dpaa_event_init_log)
1073 {
1074         dpaa_logtype_eventdev = rte_log_register("pmd.event.dpaa");
1075         if (dpaa_logtype_eventdev >= 0)
1076                 rte_log_set_level(dpaa_logtype_eventdev, RTE_LOG_NOTICE);
1077 }