git.droids-corp.org
/
dpdk.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
eventdev: make driver interface as internal
[dpdk.git]
/
drivers
/
event
/
dpaa
/
dpaa_eventdev.c
diff --git
a/drivers/event/dpaa/dpaa_eventdev.c
b/drivers/event/dpaa/dpaa_eventdev.c
index
e78728b
..
9f14390
100644
(file)
--- a/
drivers/event/dpaa/dpaa_eventdev.c
+++ b/
drivers/event/dpaa/dpaa_eventdev.c
@@
-24,7
+24,7
@@
#include <rte_memzone.h>
#include <rte_pci.h>
#include <rte_eventdev.h>
#include <rte_memzone.h>
#include <rte_pci.h>
#include <rte_eventdev.h>
-#include <
rte_
eventdev_pmd_vdev.h>
+#include <eventdev_pmd_vdev.h>
#include <rte_ethdev.h>
#include <rte_event_eth_rx_adapter.h>
#include <rte_event_eth_tx_adapter.h>
#include <rte_ethdev.h>
#include <rte_event_eth_rx_adapter.h>
#include <rte_event_eth_tx_adapter.h>
@@
-46,7
+46,7
@@
* Eventqueue = Channel Instance
* 1 Eventdev can have N Eventqueue
*/
* Eventqueue = Channel Instance
* 1 Eventdev can have N Eventqueue
*/
-RTE_LOG_REGISTER
(dpaa_logtype_eventdev, pmd.event.dpaa
, NOTICE);
+RTE_LOG_REGISTER
_DEFAULT(dpaa_logtype_eventdev
, NOTICE);
#define DISABLE_INTR_MODE "disable_intr"
#define DISABLE_INTR_MODE "disable_intr"
@@
-99,7
+99,7
@@
dpaa_event_enqueue_burst(void *port, const struct rte_event ev[],
case RTE_EVENT_OP_RELEASE:
qman_dca_index(ev[i].impl_opaque, 0);
mbuf = DPAA_PER_LCORE_DQRR_MBUF(i);
case RTE_EVENT_OP_RELEASE:
qman_dca_index(ev[i].impl_opaque, 0);
mbuf = DPAA_PER_LCORE_DQRR_MBUF(i);
-
mbuf->seqn
= DPAA_INVALID_MBUF_SEQN;
+
*dpaa_seqn(mbuf)
= DPAA_INVALID_MBUF_SEQN;
DPAA_PER_LCORE_DQRR_HELD &= ~(1 << i);
DPAA_PER_LCORE_DQRR_SIZE--;
break;
DPAA_PER_LCORE_DQRR_HELD &= ~(1 << i);
DPAA_PER_LCORE_DQRR_SIZE--;
break;
@@
-174,12
+174,12
@@
dpaa_event_dequeue_burst(void *port, struct rte_event ev[],
int ret;
u16 ch_id;
void *buffers[8];
int ret;
u16 ch_id;
void *buffers[8];
- u32 num_frames, i
, irq = 0
;
+ u32 num_frames, i;
uint64_t cur_ticks = 0, wait_time_ticks = 0;
struct dpaa_port *portal = (struct dpaa_port *)port;
struct rte_mbuf *mbuf;
uint64_t cur_ticks = 0, wait_time_ticks = 0;
struct dpaa_port *portal = (struct dpaa_port *)port;
struct rte_mbuf *mbuf;
- if (unlikely(!
RTE_PER_LCORE(dpaa_io)
)) {
+ if (unlikely(!
DPAA_PER_LCORE_PORTAL
)) {
/* Affine current thread context to a qman portal */
ret = rte_dpaa_portal_init((void *)0);
if (ret) {
/* Affine current thread context to a qman portal */
ret = rte_dpaa_portal_init((void *)0);
if (ret) {
@@
-206,7
+206,7
@@
dpaa_event_dequeue_burst(void *port, struct rte_event ev[],
if (DPAA_PER_LCORE_DQRR_HELD & (1 << i)) {
qman_dca_index(i, 0);
mbuf = DPAA_PER_LCORE_DQRR_MBUF(i);
if (DPAA_PER_LCORE_DQRR_HELD & (1 << i)) {
qman_dca_index(i, 0);
mbuf = DPAA_PER_LCORE_DQRR_MBUF(i);
-
mbuf->seqn
= DPAA_INVALID_MBUF_SEQN;
+
*dpaa_seqn(mbuf)
= DPAA_INVALID_MBUF_SEQN;
DPAA_PER_LCORE_DQRR_HELD &= ~(1 << i);
DPAA_PER_LCORE_DQRR_SIZE--;
}
DPAA_PER_LCORE_DQRR_HELD &= ~(1 << i);
DPAA_PER_LCORE_DQRR_SIZE--;
}
@@
-223,8
+223,6
@@
dpaa_event_dequeue_burst(void *port, struct rte_event ev[],
do {
/* Lets dequeue the frames */
num_frames = qman_portal_dequeue(ev, nb_events, buffers);
do {
/* Lets dequeue the frames */
num_frames = qman_portal_dequeue(ev, nb_events, buffers);
- if (irq)
- irq = 0;
if (num_frames)
break;
cur_ticks = rte_get_timer_cycles();
if (num_frames)
break;
cur_ticks = rte_get_timer_cycles();
@@
-251,7
+249,7
@@
dpaa_event_dequeue_burst_intr(void *port, struct rte_event ev[],
struct dpaa_port *portal = (struct dpaa_port *)port;
struct rte_mbuf *mbuf;
struct dpaa_port *portal = (struct dpaa_port *)port;
struct rte_mbuf *mbuf;
- if (unlikely(!
RTE_PER_LCORE(dpaa_io)
)) {
+ if (unlikely(!
DPAA_PER_LCORE_PORTAL
)) {
/* Affine current thread context to a qman portal */
ret = rte_dpaa_portal_init((void *)0);
if (ret) {
/* Affine current thread context to a qman portal */
ret = rte_dpaa_portal_init((void *)0);
if (ret) {
@@
-278,7
+276,7
@@
dpaa_event_dequeue_burst_intr(void *port, struct rte_event ev[],
if (DPAA_PER_LCORE_DQRR_HELD & (1 << i)) {
qman_dca_index(i, 0);
mbuf = DPAA_PER_LCORE_DQRR_MBUF(i);
if (DPAA_PER_LCORE_DQRR_HELD & (1 << i)) {
qman_dca_index(i, 0);
mbuf = DPAA_PER_LCORE_DQRR_MBUF(i);
-
mbuf->seqn
= DPAA_INVALID_MBUF_SEQN;
+
*dpaa_seqn(mbuf)
= DPAA_INVALID_MBUF_SEQN;
DPAA_PER_LCORE_DQRR_HELD &= ~(1 << i);
DPAA_PER_LCORE_DQRR_SIZE--;
}
DPAA_PER_LCORE_DQRR_HELD &= ~(1 << i);
DPAA_PER_LCORE_DQRR_SIZE--;
}
@@
-357,7
+355,8
@@
dpaa_event_dev_info_get(struct rte_eventdev *dev,
RTE_EVENT_DEV_CAP_DISTRIBUTED_SCHED |
RTE_EVENT_DEV_CAP_BURST_MODE |
RTE_EVENT_DEV_CAP_MULTIPLE_QUEUE_PORT |
RTE_EVENT_DEV_CAP_DISTRIBUTED_SCHED |
RTE_EVENT_DEV_CAP_BURST_MODE |
RTE_EVENT_DEV_CAP_MULTIPLE_QUEUE_PORT |
- RTE_EVENT_DEV_CAP_NONSEQ_MODE;
+ RTE_EVENT_DEV_CAP_NONSEQ_MODE |
+ RTE_EVENT_DEV_CAP_CARRY_FLOW_ID;
}
static int
}
static int
@@
-926,7
+925,7
@@
dpaa_eventdev_txa_enqueue(void *port,
return nb_events;
}
return nb_events;
}
-static struct
rte_
eventdev_ops dpaa_eventdev_ops = {
+static struct eventdev_ops dpaa_eventdev_ops = {
.dev_infos_get = dpaa_event_dev_info_get,
.dev_configure = dpaa_event_dev_configure,
.dev_start = dpaa_event_dev_start,
.dev_infos_get = dpaa_event_dev_info_get,
.dev_configure = dpaa_event_dev_configure,
.dev_start = dpaa_event_dev_start,