1 /* SPDX-License-Identifier: BSD-3-Clause
2 * Copyright(C) 2019 Marvell International Ltd.
5 #ifndef __OTX2_WORKER_DUAL_H__
6 #define __OTX2_WORKER_DUAL_H__
8 #include <rte_branch_prediction.h>
9 #include <rte_common.h>
11 #include <otx2_common.h>
12 #include "otx2_evdev.h"
15 static __rte_always_inline uint16_t
16 otx2_ssogws_dual_get_work(struct otx2_ssogws_state *ws,
17 struct otx2_ssogws_state *ws_pair,
18 struct rte_event *ev, const uint32_t flags,
19 const void * const lookup_mem,
20 struct otx2_timesync_info * const tstamp)
22 const uint64_t set_gw = BIT_ULL(16) | 1;
23 union otx2_sso_event event;
28 if (flags & NIX_RX_OFFLOAD_PTYPE_F)
29 rte_prefetch_non_temporal(lookup_mem);
33 " ldr %[tag], [%[tag_loc]] \n"
34 " ldr %[wqp], [%[wqp_loc]] \n"
35 " tbnz %[tag], 63, rty%= \n"
36 "done%=: str %[gw], [%[pong]] \n"
38 " prfm pldl1keep, [%[wqp], #8]\n"
39 " sub %[mbuf], %[wqp], #0x80 \n"
40 " prfm pldl1keep, [%[mbuf]] \n"
41 : [tag] "=&r" (event.get_work0),
42 [wqp] "=&r" (get_work1),
44 : [tag_loc] "r" (ws->tag_op),
45 [wqp_loc] "r" (ws->wqp_op),
47 [pong] "r" (ws_pair->getwrk_op)
50 event.get_work0 = otx2_read64(ws->tag_op);
51 while ((BIT_ULL(63)) & event.get_work0)
52 event.get_work0 = otx2_read64(ws->tag_op);
53 get_work1 = otx2_read64(ws->wqp_op);
54 otx2_write64(set_gw, ws_pair->getwrk_op);
56 rte_prefetch0((const void *)get_work1);
57 mbuf = (uint64_t)((char *)get_work1 - sizeof(struct rte_mbuf));
58 rte_prefetch0((const void *)mbuf);
60 event.get_work0 = (event.get_work0 & (0x3ull << 32)) << 6 |
61 (event.get_work0 & (0x3FFull << 36)) << 4 |
62 (event.get_work0 & 0xffffffff);
63 ws->cur_tt = event.sched_type;
64 ws->cur_grp = event.queue_id;
66 if (event.sched_type != SSO_TT_EMPTY &&
67 event.event_type == RTE_EVENT_TYPE_ETHDEV) {
68 uint8_t port = event.sub_event_type;
70 event.sub_event_type = 0;
71 otx2_wqe_to_mbuf(get_work1, mbuf, port,
72 event.flow_id, flags, lookup_mem);
73 /* Extracting tstamp, if PTP enabled. CGX will prepend the
74 * timestamp at starting of packet data and it can be derieved
75 * from WQE 9 dword which corresponds to SG iova.
76 * rte_pktmbuf_mtod_offset can be used for this purpose but it
77 * brings down the performance as it reads mbuf->buf_addr which
78 * is not part of cache in general fast path.
80 tstamp_ptr = *(uint64_t *)(((struct nix_wqe_hdr_s *)get_work1)
81 + OTX2_SSO_WQE_SG_PTR);
82 otx2_nix_mbuf_to_tstamp((struct rte_mbuf *)mbuf, tstamp, flags,
83 (uint64_t *)tstamp_ptr);
87 ev->event = event.get_work0;
93 static __rte_always_inline void
94 otx2_ssogws_dual_add_work(struct otx2_ssogws_dual *ws, const uint64_t event_ptr,
95 const uint32_t tag, const uint8_t new_tt,
100 add_work0 = tag | ((uint64_t)(new_tt) << 32);
101 otx2_store_pair(add_work0, event_ptr, ws->grps_base[grp]);