net/ice/base: remove unnecessary braces
[dpdk.git] / drivers / net / ice / base / ice_flow.h
1 /* SPDX-License-Identifier: BSD-3-Clause
2  * Copyright(c) 2001-2020 Intel Corporation
3  */
4
5 #ifndef _ICE_FLOW_H_
6 #define _ICE_FLOW_H_
7
8 #include "ice_flex_type.h"
9 #include "ice_acl.h"
10
11 #define ICE_IPV4_MAKE_PREFIX_MASK(prefix) ((u32)(~0) << (32 - (prefix)))
12 #define ICE_FLOW_PROF_ID_INVAL          0xfffffffffffffffful
13 #define ICE_FLOW_PROF_ID_BYPASS         0
14 #define ICE_FLOW_PROF_ID_DEFAULT        1
15 #define ICE_FLOW_ENTRY_HANDLE_INVAL     0
16 #define ICE_FLOW_VSI_INVAL              0xffff
17 #define ICE_FLOW_FLD_OFF_INVAL          0xffff
18
19 /* Generate flow hash field from flow field type(s) */
20 #define ICE_FLOW_HASH_ETH       \
21         (BIT_ULL(ICE_FLOW_FIELD_IDX_ETH_DA) | \
22          BIT_ULL(ICE_FLOW_FIELD_IDX_ETH_SA))
23 #define ICE_FLOW_HASH_IPV4      \
24         (BIT_ULL(ICE_FLOW_FIELD_IDX_IPV4_SA) | \
25          BIT_ULL(ICE_FLOW_FIELD_IDX_IPV4_DA))
26 #define ICE_FLOW_HASH_IPV6      \
27         (BIT_ULL(ICE_FLOW_FIELD_IDX_IPV6_SA) | \
28          BIT_ULL(ICE_FLOW_FIELD_IDX_IPV6_DA))
29 #define ICE_FLOW_HASH_TCP_PORT  \
30         (BIT_ULL(ICE_FLOW_FIELD_IDX_TCP_SRC_PORT) | \
31          BIT_ULL(ICE_FLOW_FIELD_IDX_TCP_DST_PORT))
32 #define ICE_FLOW_HASH_UDP_PORT  \
33         (BIT_ULL(ICE_FLOW_FIELD_IDX_UDP_SRC_PORT) | \
34          BIT_ULL(ICE_FLOW_FIELD_IDX_UDP_DST_PORT))
35 #define ICE_FLOW_HASH_SCTP_PORT \
36         (BIT_ULL(ICE_FLOW_FIELD_IDX_SCTP_SRC_PORT) | \
37          BIT_ULL(ICE_FLOW_FIELD_IDX_SCTP_DST_PORT))
38
39 #define ICE_HASH_INVALID        0
40 #define ICE_HASH_TCP_IPV4       (ICE_FLOW_HASH_IPV4 | ICE_FLOW_HASH_TCP_PORT)
41 #define ICE_HASH_TCP_IPV6       (ICE_FLOW_HASH_IPV6 | ICE_FLOW_HASH_TCP_PORT)
42 #define ICE_HASH_UDP_IPV4       (ICE_FLOW_HASH_IPV4 | ICE_FLOW_HASH_UDP_PORT)
43 #define ICE_HASH_UDP_IPV6       (ICE_FLOW_HASH_IPV6 | ICE_FLOW_HASH_UDP_PORT)
44 #define ICE_HASH_SCTP_IPV4      (ICE_FLOW_HASH_IPV4 | ICE_FLOW_HASH_SCTP_PORT)
45 #define ICE_HASH_SCTP_IPV6      (ICE_FLOW_HASH_IPV6 | ICE_FLOW_HASH_SCTP_PORT)
46
47 #define ICE_FLOW_HASH_GTP_TEID \
48         (BIT_ULL(ICE_FLOW_FIELD_IDX_GTPC_TEID))
49
50 #define ICE_FLOW_HASH_GTP_IPV4_TEID \
51         (ICE_FLOW_HASH_IPV4 | ICE_FLOW_HASH_GTP_TEID)
52 #define ICE_FLOW_HASH_GTP_IPV6_TEID \
53         (ICE_FLOW_HASH_IPV6 | ICE_FLOW_HASH_GTP_TEID)
54
55 #define ICE_FLOW_HASH_GTP_U_TEID \
56         (BIT_ULL(ICE_FLOW_FIELD_IDX_GTPU_IP_TEID))
57
58 #define ICE_FLOW_HASH_GTP_U_IPV4_TEID \
59         (ICE_FLOW_HASH_IPV4 | ICE_FLOW_HASH_GTP_U_TEID)
60 #define ICE_FLOW_HASH_GTP_U_IPV6_TEID \
61         (ICE_FLOW_HASH_IPV6 | ICE_FLOW_HASH_GTP_U_TEID)
62
63 #define ICE_FLOW_HASH_GTP_U_EH_TEID \
64         (BIT_ULL(ICE_FLOW_FIELD_IDX_GTPU_EH_TEID))
65
66 #define ICE_FLOW_HASH_GTP_U_EH_QFI \
67         (BIT_ULL(ICE_FLOW_FIELD_IDX_GTPU_EH_QFI))
68
69 #define ICE_FLOW_HASH_GTP_U_IPV4_EH \
70         (ICE_FLOW_HASH_IPV4 | ICE_FLOW_HASH_GTP_U_EH_TEID | \
71          ICE_FLOW_HASH_GTP_U_EH_QFI)
72 #define ICE_FLOW_HASH_GTP_U_IPV6_EH \
73         (ICE_FLOW_HASH_IPV6 | ICE_FLOW_HASH_GTP_U_EH_TEID | \
74          ICE_FLOW_HASH_GTP_U_EH_QFI)
75
76 #define ICE_FLOW_HASH_PPPOE_SESS_ID \
77         (BIT_ULL(ICE_FLOW_FIELD_IDX_PPPOE_SESS_ID))
78
79 #define ICE_FLOW_HASH_PPPOE_SESS_ID_ETH \
80         (ICE_FLOW_HASH_ETH | ICE_FLOW_HASH_PPPOE_SESS_ID)
81 #define ICE_FLOW_HASH_PPPOE_TCP_ID \
82         (ICE_FLOW_HASH_TCP_PORT | ICE_FLOW_HASH_PPPOE_SESS_ID)
83 #define ICE_FLOW_HASH_PPPOE_UDP_ID \
84         (ICE_FLOW_HASH_UDP_PORT | ICE_FLOW_HASH_PPPOE_SESS_ID)
85
86 #define ICE_FLOW_HASH_PFCP_SEID \
87         (BIT_ULL(ICE_FLOW_FIELD_IDX_PFCP_SEID))
88 #define ICE_FLOW_HASH_PFCP_IPV4_SEID \
89         (ICE_FLOW_HASH_IPV4 | ICE_FLOW_HASH_PFCP_SEID)
90 #define ICE_FLOW_HASH_PFCP_IPV6_SEID \
91         (ICE_FLOW_HASH_IPV6 | ICE_FLOW_HASH_PFCP_SEID)
92
93 #define ICE_FLOW_HASH_L2TPV3_SESS_ID \
94         (BIT_ULL(ICE_FLOW_FIELD_IDX_L2TPV3_SESS_ID))
95 #define ICE_FLOW_HASH_L2TPV3_IPV4_SESS_ID \
96         (ICE_FLOW_HASH_IPV4 | ICE_FLOW_HASH_L2TPV3_SESS_ID)
97 #define ICE_FLOW_HASH_L2TPV3_IPV6_SESS_ID \
98         (ICE_FLOW_HASH_IPV6 | ICE_FLOW_HASH_L2TPV3_SESS_ID)
99
100 #define ICE_FLOW_HASH_ESP_SPI \
101         (BIT_ULL(ICE_FLOW_FIELD_IDX_ESP_SPI))
102 #define ICE_FLOW_HASH_ESP_IPV4_SPI \
103         (ICE_FLOW_HASH_IPV4 | ICE_FLOW_HASH_ESP_SPI)
104 #define ICE_FLOW_HASH_ESP_IPV6_SPI \
105         (ICE_FLOW_HASH_IPV6 | ICE_FLOW_HASH_ESP_SPI)
106
107 #define ICE_FLOW_HASH_AH_SPI \
108         (BIT_ULL(ICE_FLOW_FIELD_IDX_AH_SPI))
109 #define ICE_FLOW_HASH_AH_IPV4_SPI \
110         (ICE_FLOW_HASH_IPV4 | ICE_FLOW_HASH_AH_SPI)
111 #define ICE_FLOW_HASH_AH_IPV6_SPI \
112         (ICE_FLOW_HASH_IPV6 | ICE_FLOW_HASH_AH_SPI)
113
114 #define ICE_FLOW_HASH_NAT_T_ESP_SPI \
115         (BIT_ULL(ICE_FLOW_FIELD_IDX_NAT_T_ESP_SPI))
116 #define ICE_FLOW_HASH_NAT_T_ESP_IPV4_SPI \
117         (ICE_FLOW_HASH_IPV4 | ICE_FLOW_HASH_NAT_T_ESP_SPI)
118 #define ICE_FLOW_HASH_NAT_T_ESP_IPV6_SPI \
119         (ICE_FLOW_HASH_IPV6 | ICE_FLOW_HASH_NAT_T_ESP_SPI)
120
121 /* Protocol header fields within a packet segment. A segment consists of one or
122  * more protocol headers that make up a logical group of protocol headers. Each
123  * logical group of protocol headers encapsulates or is encapsulated using/by
124  * tunneling or encapsulation protocols for network virtualization such as GRE,
125  * VxLAN, etc.
126  */
127 enum ice_flow_seg_hdr {
128         ICE_FLOW_SEG_HDR_NONE           = 0x00000000,
129         ICE_FLOW_SEG_HDR_ETH            = 0x00000001,
130         ICE_FLOW_SEG_HDR_VLAN           = 0x00000002,
131         ICE_FLOW_SEG_HDR_IPV4           = 0x00000004,
132         ICE_FLOW_SEG_HDR_IPV6           = 0x00000008,
133         ICE_FLOW_SEG_HDR_ARP            = 0x00000010,
134         ICE_FLOW_SEG_HDR_ICMP           = 0x00000020,
135         ICE_FLOW_SEG_HDR_TCP            = 0x00000040,
136         ICE_FLOW_SEG_HDR_UDP            = 0x00000080,
137         ICE_FLOW_SEG_HDR_SCTP           = 0x00000100,
138         ICE_FLOW_SEG_HDR_GRE            = 0x00000200,
139         ICE_FLOW_SEG_HDR_GTPC           = 0x00000400,
140         ICE_FLOW_SEG_HDR_GTPC_TEID      = 0x00000800,
141         ICE_FLOW_SEG_HDR_GTPU_IP        = 0x00001000,
142         ICE_FLOW_SEG_HDR_GTPU_EH        = 0x00002000,
143         ICE_FLOW_SEG_HDR_GTPU_DWN       = 0x00004000,
144         ICE_FLOW_SEG_HDR_GTPU_UP        = 0x00008000,
145         ICE_FLOW_SEG_HDR_PPPOE          = 0x00010000,
146         ICE_FLOW_SEG_HDR_PFCP_NODE      = 0x00020000,
147         ICE_FLOW_SEG_HDR_PFCP_SESSION   = 0x00040000,
148         ICE_FLOW_SEG_HDR_L2TPV3         = 0x00080000,
149         ICE_FLOW_SEG_HDR_ESP            = 0x00100000,
150         ICE_FLOW_SEG_HDR_AH             = 0x00200000,
151         ICE_FLOW_SEG_HDR_NAT_T_ESP      = 0x00400000,
152         ICE_FLOW_SEG_HDR_ETH_NON_IP     = 0x00800000,
153 };
154
155 /* These segements all have the same PTYPES, but are otherwise distinguished by
156  * the value of the gtp_eh_pdu and gtp_eh_pdu_link flags:
157  *
158  *                                gtp_eh_pdu     gtp_eh_pdu_link
159  * ICE_FLOW_SEG_HDR_GTPU_IP           0              0
160  * ICE_FLOW_SEG_HDR_GTPU_EH           1              don't care
161  * ICE_FLOW_SEG_HDR_GTPU_DWN          1              0
162  * ICE_FLOW_SEG_HDR_GTPU_UP           1              1
163  */
164 #define ICE_FLOW_SEG_HDR_GTPU (ICE_FLOW_SEG_HDR_GTPU_IP | \
165                                ICE_FLOW_SEG_HDR_GTPU_EH | \
166                                ICE_FLOW_SEG_HDR_GTPU_DWN | \
167                                ICE_FLOW_SEG_HDR_GTPU_UP)
168 #define ICE_FLOW_SEG_HDR_PFCP (ICE_FLOW_SEG_HDR_PFCP_NODE | \
169                                ICE_FLOW_SEG_HDR_PFCP_SESSION)
170
171 enum ice_flow_field {
172         /* L2 */
173         ICE_FLOW_FIELD_IDX_ETH_DA,
174         ICE_FLOW_FIELD_IDX_ETH_SA,
175         ICE_FLOW_FIELD_IDX_S_VLAN,
176         ICE_FLOW_FIELD_IDX_C_VLAN,
177         ICE_FLOW_FIELD_IDX_ETH_TYPE,
178         /* L3 */
179         ICE_FLOW_FIELD_IDX_IPV4_DSCP,
180         ICE_FLOW_FIELD_IDX_IPV6_DSCP,
181         ICE_FLOW_FIELD_IDX_IPV4_TTL,
182         ICE_FLOW_FIELD_IDX_IPV4_PROT,
183         ICE_FLOW_FIELD_IDX_IPV6_TTL,
184         ICE_FLOW_FIELD_IDX_IPV6_PROT,
185         ICE_FLOW_FIELD_IDX_IPV4_SA,
186         ICE_FLOW_FIELD_IDX_IPV4_DA,
187         ICE_FLOW_FIELD_IDX_IPV6_SA,
188         ICE_FLOW_FIELD_IDX_IPV6_DA,
189         /* L4 */
190         ICE_FLOW_FIELD_IDX_TCP_SRC_PORT,
191         ICE_FLOW_FIELD_IDX_TCP_DST_PORT,
192         ICE_FLOW_FIELD_IDX_UDP_SRC_PORT,
193         ICE_FLOW_FIELD_IDX_UDP_DST_PORT,
194         ICE_FLOW_FIELD_IDX_SCTP_SRC_PORT,
195         ICE_FLOW_FIELD_IDX_SCTP_DST_PORT,
196         ICE_FLOW_FIELD_IDX_TCP_FLAGS,
197         /* ARP */
198         ICE_FLOW_FIELD_IDX_ARP_SIP,
199         ICE_FLOW_FIELD_IDX_ARP_DIP,
200         ICE_FLOW_FIELD_IDX_ARP_SHA,
201         ICE_FLOW_FIELD_IDX_ARP_DHA,
202         ICE_FLOW_FIELD_IDX_ARP_OP,
203         /* ICMP */
204         ICE_FLOW_FIELD_IDX_ICMP_TYPE,
205         ICE_FLOW_FIELD_IDX_ICMP_CODE,
206         /* GRE */
207         ICE_FLOW_FIELD_IDX_GRE_KEYID,
208         /* GTPC_TEID */
209         ICE_FLOW_FIELD_IDX_GTPC_TEID,
210         /* GTPU_IP */
211         ICE_FLOW_FIELD_IDX_GTPU_IP_TEID,
212         /* GTPU_EH */
213         ICE_FLOW_FIELD_IDX_GTPU_EH_TEID,
214         ICE_FLOW_FIELD_IDX_GTPU_EH_QFI,
215         /* GTPU_UP */
216         ICE_FLOW_FIELD_IDX_GTPU_UP_TEID,
217         /* GTPU_DWN */
218         ICE_FLOW_FIELD_IDX_GTPU_DWN_TEID,
219         /* PPPOE */
220         ICE_FLOW_FIELD_IDX_PPPOE_SESS_ID,
221         /* PFCP */
222         ICE_FLOW_FIELD_IDX_PFCP_SEID,
223         /* L2TPV3 */
224         ICE_FLOW_FIELD_IDX_L2TPV3_SESS_ID,
225         /* ESP */
226         ICE_FLOW_FIELD_IDX_ESP_SPI,
227         /* AH */
228         ICE_FLOW_FIELD_IDX_AH_SPI,
229         /* NAT_T ESP */
230         ICE_FLOW_FIELD_IDX_NAT_T_ESP_SPI,
231          /* The total number of enums must not exceed 64 */
232         ICE_FLOW_FIELD_IDX_MAX
233 };
234
235 /* Flow headers and fields for AVF support */
236 enum ice_flow_avf_hdr_field {
237         /* Values 0 - 28 are reserved for future use */
238         ICE_AVF_FLOW_FIELD_INVALID              = 0,
239         ICE_AVF_FLOW_FIELD_UNICAST_IPV4_UDP     = 29,
240         ICE_AVF_FLOW_FIELD_MULTICAST_IPV4_UDP,
241         ICE_AVF_FLOW_FIELD_IPV4_UDP,
242         ICE_AVF_FLOW_FIELD_IPV4_TCP_SYN_NO_ACK,
243         ICE_AVF_FLOW_FIELD_IPV4_TCP,
244         ICE_AVF_FLOW_FIELD_IPV4_SCTP,
245         ICE_AVF_FLOW_FIELD_IPV4_OTHER,
246         ICE_AVF_FLOW_FIELD_FRAG_IPV4,
247         /* Values 37-38 are reserved */
248         ICE_AVF_FLOW_FIELD_UNICAST_IPV6_UDP     = 39,
249         ICE_AVF_FLOW_FIELD_MULTICAST_IPV6_UDP,
250         ICE_AVF_FLOW_FIELD_IPV6_UDP,
251         ICE_AVF_FLOW_FIELD_IPV6_TCP_SYN_NO_ACK,
252         ICE_AVF_FLOW_FIELD_IPV6_TCP,
253         ICE_AVF_FLOW_FIELD_IPV6_SCTP,
254         ICE_AVF_FLOW_FIELD_IPV6_OTHER,
255         ICE_AVF_FLOW_FIELD_FRAG_IPV6,
256         ICE_AVF_FLOW_FIELD_RSVD47,
257         ICE_AVF_FLOW_FIELD_FCOE_OX,
258         ICE_AVF_FLOW_FIELD_FCOE_RX,
259         ICE_AVF_FLOW_FIELD_FCOE_OTHER,
260         /* Values 51-62 are reserved */
261         ICE_AVF_FLOW_FIELD_L2_PAYLOAD           = 63,
262         ICE_AVF_FLOW_FIELD_MAX
263 };
264
265 /* Supported RSS offloads  This macro is defined to support
266  * VIRTCHNL_OP_GET_RSS_HENA_CAPS ops. PF driver sends the RSS hardware
267  * capabilities to the caller of this ops.
268  */
269 #define ICE_DEFAULT_RSS_HENA ( \
270         BIT_ULL(ICE_AVF_FLOW_FIELD_IPV4_UDP) | \
271         BIT_ULL(ICE_AVF_FLOW_FIELD_IPV4_SCTP) | \
272         BIT_ULL(ICE_AVF_FLOW_FIELD_IPV4_TCP) | \
273         BIT_ULL(ICE_AVF_FLOW_FIELD_IPV4_OTHER) | \
274         BIT_ULL(ICE_AVF_FLOW_FIELD_FRAG_IPV4) | \
275         BIT_ULL(ICE_AVF_FLOW_FIELD_IPV6_UDP) | \
276         BIT_ULL(ICE_AVF_FLOW_FIELD_IPV6_TCP) | \
277         BIT_ULL(ICE_AVF_FLOW_FIELD_IPV6_SCTP) | \
278         BIT_ULL(ICE_AVF_FLOW_FIELD_IPV6_OTHER) | \
279         BIT_ULL(ICE_AVF_FLOW_FIELD_FRAG_IPV6) | \
280         BIT_ULL(ICE_AVF_FLOW_FIELD_IPV4_TCP_SYN_NO_ACK) | \
281         BIT_ULL(ICE_AVF_FLOW_FIELD_UNICAST_IPV4_UDP) | \
282         BIT_ULL(ICE_AVF_FLOW_FIELD_MULTICAST_IPV4_UDP) | \
283         BIT_ULL(ICE_AVF_FLOW_FIELD_IPV6_TCP_SYN_NO_ACK) | \
284         BIT_ULL(ICE_AVF_FLOW_FIELD_UNICAST_IPV6_UDP) | \
285         BIT_ULL(ICE_AVF_FLOW_FIELD_MULTICAST_IPV6_UDP))
286
287 enum ice_flow_dir {
288         ICE_FLOW_DIR_UNDEFINED  = 0,
289         ICE_FLOW_TX             = 0x01,
290         ICE_FLOW_RX             = 0x02,
291         ICE_FLOW_TX_RX          = ICE_FLOW_RX | ICE_FLOW_TX
292 };
293
294 enum ice_flow_priority {
295         ICE_FLOW_PRIO_LOW,
296         ICE_FLOW_PRIO_NORMAL,
297         ICE_FLOW_PRIO_HIGH
298 };
299
300 #define ICE_FLOW_SEG_MAX                2
301 #define ICE_FLOW_SEG_RAW_FLD_MAX        2
302 #define ICE_FLOW_PROFILE_MAX            1024
303 #define ICE_FLOW_SW_FIELD_VECTOR_MAX    48
304 #define ICE_FLOW_ACL_FIELD_VECTOR_MAX   32
305 #define ICE_FLOW_FV_EXTRACT_SZ          2
306
307 #define ICE_FLOW_SET_HDRS(seg, val)     ((seg)->hdrs |= (u32)(val))
308
309 struct ice_flow_seg_xtrct {
310         u8 prot_id;     /* Protocol ID of extracted header field */
311         u16 off;        /* Starting offset of the field in header in bytes */
312         u8 idx;         /* Index of FV entry used */
313         u8 disp;        /* Displacement of field in bits fr. FV entry's start */
314         u16 mask;       /* Mask for field */
315 };
316
317 enum ice_flow_fld_match_type {
318         ICE_FLOW_FLD_TYPE_REG,          /* Value, mask */
319         ICE_FLOW_FLD_TYPE_RANGE,        /* Value, mask, last (upper bound) */
320         ICE_FLOW_FLD_TYPE_PREFIX,       /* IP address, prefix, size of prefix */
321         ICE_FLOW_FLD_TYPE_SIZE,         /* Value, mask, size of match */
322 };
323
324 struct ice_flow_fld_loc {
325         /* Describe offsets of field information relative to the beginning of
326          * input buffer provided when adding flow entries.
327          */
328         u16 val;        /* Offset where the value is located */
329         u16 mask;       /* Offset where the mask/prefix value is located */
330         u16 last;       /* Length or offset where the upper value is located */
331 };
332
333 struct ice_flow_fld_info {
334         enum ice_flow_fld_match_type type;
335         /* Location where to retrieve data from an input buffer */
336         struct ice_flow_fld_loc src;
337         /* Location where to put the data into the final entry buffer */
338         struct ice_flow_fld_loc entry;
339         struct ice_flow_seg_xtrct xtrct;
340 };
341
342 struct ice_flow_seg_fld_raw {
343         struct ice_flow_fld_info info;
344         u16 off;        /* Offset from the start of the segment */
345 };
346
347 struct ice_flow_seg_info {
348         u32 hdrs;       /* Bitmask indicating protocol headers present */
349         u64 match;      /* Bitmask indicating header fields to be matched */
350         u64 range;      /* Bitmask indicating header fields matched as ranges */
351
352         struct ice_flow_fld_info fields[ICE_FLOW_FIELD_IDX_MAX];
353
354         u8 raws_cnt;    /* Number of raw fields to be matched */
355         struct ice_flow_seg_fld_raw raws[ICE_FLOW_SEG_RAW_FLD_MAX];
356 };
357
358 /* This structure describes a flow entry, and is tracked only in this file */
359 struct ice_flow_entry {
360         struct LIST_ENTRY_TYPE l_entry;
361
362         u64 id;
363         struct ice_flow_prof *prof;
364         /* Action list */
365         struct ice_flow_action *acts;
366         /* Flow entry's content */
367         void *entry;
368         /* Range buffer (For ACL only) */
369         struct ice_aqc_acl_profile_ranges *range_buf;
370         enum ice_flow_priority priority;
371         u16 vsi_handle;
372         u16 entry_sz;
373         /* Entry index in the ACL's scenario */
374         u16 scen_entry_idx;
375 #define ICE_FLOW_ACL_MAX_NUM_ACT        2
376         u8 acts_cnt;
377 };
378
379 #define ICE_FLOW_ENTRY_HNDL(e)  ((unsigned long)e)
380 #define ICE_FLOW_ENTRY_PTR(h)   ((struct ice_flow_entry *)(h))
381
382 struct ice_flow_prof {
383         struct LIST_ENTRY_TYPE l_entry;
384
385         u64 id;
386         enum ice_flow_dir dir;
387         u8 segs_cnt;
388         u8 acts_cnt;
389
390         /* Keep track of flow entries associated with this flow profile */
391         struct ice_lock entries_lock;
392         struct LIST_HEAD_TYPE entries;
393
394         struct ice_flow_seg_info segs[ICE_FLOW_SEG_MAX];
395
396         /* software VSI handles referenced by this flow profile */
397         ice_declare_bitmap(vsis, ICE_MAX_VSI);
398
399         union {
400                 /* struct sw_recipe */
401                 struct ice_acl_scen *scen;
402                 /* struct fd */
403                 u32 data;
404                 /* Symmetric Hash for RSS */
405                 bool symm;
406         } cfg;
407
408         /* Default actions */
409         struct ice_flow_action *acts;
410 };
411
412 struct ice_rss_cfg {
413         struct LIST_ENTRY_TYPE l_entry;
414         /* bitmap of VSIs added to the RSS entry */
415         ice_declare_bitmap(vsis, ICE_MAX_VSI);
416         u64 hashed_flds;
417         u32 packet_hdr;
418         bool symm;
419 };
420
421 enum ice_flow_action_type {
422         ICE_FLOW_ACT_NOP,
423         ICE_FLOW_ACT_ALLOW,
424         ICE_FLOW_ACT_DROP,
425         ICE_FLOW_ACT_CNTR_PKT,
426         ICE_FLOW_ACT_FWD_VSI,
427         ICE_FLOW_ACT_FWD_VSI_LIST,      /* Should be abstracted away */
428         ICE_FLOW_ACT_FWD_QUEUE,         /* Can Queues be abstracted away? */
429         ICE_FLOW_ACT_FWD_QUEUE_GROUP,   /* Can Queues be abstracted away? */
430         ICE_FLOW_ACT_PUSH,
431         ICE_FLOW_ACT_POP,
432         ICE_FLOW_ACT_MODIFY,
433         ICE_FLOW_ACT_CNTR_BYTES,
434         ICE_FLOW_ACT_CNTR_PKT_BYTES,
435         ICE_FLOW_ACT_GENERIC_0,
436         ICE_FLOW_ACT_GENERIC_1,
437         ICE_FLOW_ACT_GENERIC_2,
438         ICE_FLOW_ACT_GENERIC_3,
439         ICE_FLOW_ACT_GENERIC_4,
440         ICE_FLOW_ACT_RPT_FLOW_ID,
441         ICE_FLOW_ACT_BUILD_PROF_IDX,
442 };
443
444 struct ice_flow_action {
445         enum ice_flow_action_type type;
446         union {
447                 struct ice_acl_act_entry acl_act;
448                 u32 dummy;
449         } data;
450 };
451
452 u64
453 ice_flow_find_prof(struct ice_hw *hw, enum ice_block blk, enum ice_flow_dir dir,
454                    struct ice_flow_seg_info *segs, u8 segs_cnt);
455 enum ice_status
456 ice_flow_add_prof(struct ice_hw *hw, enum ice_block blk, enum ice_flow_dir dir,
457                   u64 prof_id, struct ice_flow_seg_info *segs, u8 segs_cnt,
458                   struct ice_flow_action *acts, u8 acts_cnt,
459                   struct ice_flow_prof **prof);
460 enum ice_status
461 ice_flow_rem_prof(struct ice_hw *hw, enum ice_block blk, u64 prof_id);
462 enum ice_status
463 ice_flow_assoc_vsig_vsi(struct ice_hw *hw, enum ice_block blk, u16 vsi_handle,
464                         u16 vsig);
465 enum ice_status
466 ice_flow_get_hw_prof(struct ice_hw *hw, enum ice_block blk, u64 prof_id,
467                      u8 *hw_prof);
468
469 u64 ice_flow_find_entry(struct ice_hw *hw, enum ice_block blk, u64 entry_id);
470 enum ice_status
471 ice_flow_add_entry(struct ice_hw *hw, enum ice_block blk, u64 prof_id,
472                    u64 entry_id, u16 vsi, enum ice_flow_priority prio,
473                    void *data, struct ice_flow_action *acts, u8 acts_cnt,
474                    u64 *entry_h);
475 enum ice_status
476 ice_flow_rem_entry(struct ice_hw *hw, enum ice_block blk, u64 entry_h);
477 void
478 ice_flow_set_fld(struct ice_flow_seg_info *seg, enum ice_flow_field fld,
479                  u16 val_loc, u16 mask_loc, u16 last_loc, bool range);
480 void
481 ice_flow_set_fld_prefix(struct ice_flow_seg_info *seg, enum ice_flow_field fld,
482                         u16 val_loc, u16 prefix_loc, u8 prefix_sz);
483 void
484 ice_flow_add_fld_raw(struct ice_flow_seg_info *seg, u16 off, u8 len,
485                      u16 val_loc, u16 mask_loc);
486 void ice_rem_vsi_rss_list(struct ice_hw *hw, u16 vsi_handle);
487 enum ice_status ice_replay_rss_cfg(struct ice_hw *hw, u16 vsi_handle);
488 enum ice_status
489 ice_add_avf_rss_cfg(struct ice_hw *hw, u16 vsi_handle, u64 hashed_flds);
490 enum ice_status ice_rem_vsi_rss_cfg(struct ice_hw *hw, u16 vsi_handle);
491 enum ice_status
492 ice_add_rss_cfg(struct ice_hw *hw, u16 vsi_handle, u64 hashed_flds,
493                 u32 addl_hdrs, bool symm);
494 enum ice_status
495 ice_rem_rss_cfg(struct ice_hw *hw, u16 vsi_handle, u64 hashed_flds,
496                 u32 addl_hdrs);
497 u64 ice_get_rss_cfg(struct ice_hw *hw, u16 vsi_handle, u32 hdrs);
498 #endif /* _ICE_FLOW_H_ */