git.droids-corp.org
/
dpdk.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
mbuf: fix VLAN flags documentation
[dpdk.git]
/
lib
/
librte_mbuf
/
rte_mbuf.h
diff --git
a/lib/librte_mbuf/rte_mbuf.h
b/lib/librte_mbuf/rte_mbuf.h
index
6d91f7d
..
719d04d
100644
(file)
--- a/
lib/librte_mbuf/rte_mbuf.h
+++ b/
lib/librte_mbuf/rte_mbuf.h
@@
-62,6
+62,7
@@
#include <stdint.h>
#include <rte_common.h>
#include <stdint.h>
#include <rte_common.h>
+#include <rte_config.h>
#include <rte_mempool.h>
#include <rte_memory.h>
#include <rte_atomic.h>
#include <rte_mempool.h>
#include <rte_memory.h>
#include <rte_atomic.h>
@@
-208,6
+209,13
@@
extern "C" {
/* add new TX flags here */
/* add new TX flags here */
+/**
+ * UDP Fragmentation Offload flag. This flag is used for enabling UDP
+ * fragmentation in SW or in HW. When use UFO, mbuf->tso_segsz is used
+ * to store the MSS of UDP fragments.
+ */
+#define PKT_TX_UDP_SEG (1ULL << 42)
+
/**
* Request security offload processing on the TX packet.
*/
/**
* Request security offload processing on the TX packet.
*/
@@
-236,7
+244,9
@@
extern "C" {
/**
* Second VLAN insertion (QinQ) flag.
*/
/**
* Second VLAN insertion (QinQ) flag.
*/
-#define PKT_TX_QINQ_PKT (1ULL << 49) /**< TX packet with double VLAN inserted. */
+#define PKT_TX_QINQ (1ULL << 49) /**< TX packet with double VLAN inserted. */
+/* this old name is deprecated */
+#define PKT_TX_QINQ_PKT PKT_TX_QINQ
/**
* TCP segmentation offload. To enable this offload feature for a
/**
* TCP segmentation offload. To enable this offload feature for a
@@
-297,7
+307,12
@@
extern "C" {
*/
#define PKT_TX_IPV6 (1ULL << 56)
*/
#define PKT_TX_IPV6 (1ULL << 56)
-#define PKT_TX_VLAN_PKT (1ULL << 57) /**< TX packet is a 802.1q VLAN packet. */
+/**
+ * TX packet is a 802.1q VLAN packet.
+ */
+#define PKT_TX_VLAN (1ULL << 57)
+/* this old name is deprecated */
+#define PKT_TX_VLAN_PKT PKT_TX_VLAN
/**
* Offload the IP checksum of an external header in the hardware. The
/**
* Offload the IP checksum of an external header in the hardware. The
@@
-502,7
+517,7
@@
struct rte_mbuf {
uint32_t pkt_len; /**< Total pkt len: sum of all segments. */
uint16_t data_len; /**< Amount of data in segment buffer. */
uint32_t pkt_len; /**< Total pkt len: sum of all segments. */
uint16_t data_len; /**< Amount of data in segment buffer. */
- /** VLAN TCI (CPU order), valid if PKT_RX_VLAN
_STRIPPED
is set. */
+ /** VLAN TCI (CPU order), valid if PKT_RX_VLAN is set. */
uint16_t vlan_tci;
union {
uint16_t vlan_tci;
union {
@@
-528,7
+543,7
@@
struct rte_mbuf {
uint32_t usr; /**< User defined tags. See rte_distributor_process() */
} hash; /**< hash information */
uint32_t usr; /**< User defined tags. See rte_distributor_process() */
} hash; /**< hash information */
- /** Outer VLAN TCI (CPU order), valid if PKT_RX_QINQ
_STRIPPED
is set. */
+ /** Outer VLAN TCI (CPU order), valid if PKT_RX_QINQ is set. */
uint16_t vlan_tci_outer;
uint16_t buf_len; /**< Length of segment buffer. */
uint16_t vlan_tci_outer;
uint16_t buf_len; /**< Length of segment buffer. */
@@
-584,6
+599,9
@@
struct rte_mbuf {
} __rte_cache_aligned;
} __rte_cache_aligned;
+/**< Maximum number of nb_segs allowed. */
+#define RTE_MBUF_MAX_NB_SEGS UINT16_MAX
+
/**
* Prefetch the first part of the mbuf
*
/**
* Prefetch the first part of the mbuf
*
@@
-761,6
+779,13
@@
rte_mbuf_refcnt_set(struct rte_mbuf *m, uint16_t new_value)
rte_atomic16_set(&m->refcnt_atomic, new_value);
}
rte_atomic16_set(&m->refcnt_atomic, new_value);
}
+/* internal */
+static inline uint16_t
+__rte_mbuf_refcnt_update(struct rte_mbuf *m, int16_t value)
+{
+ return (uint16_t)(rte_atomic16_add_return(&m->refcnt_atomic, value));
+}
+
/**
* Adds given value to an mbuf's refcnt and returns its new value.
* @param m
/**
* Adds given value to an mbuf's refcnt and returns its new value.
* @param m
@@
-785,19
+810,26
@@
rte_mbuf_refcnt_update(struct rte_mbuf *m, int16_t value)
return 1 + value;
}
return 1 + value;
}
- return
(uint16_t)(rte_atomic16_add_return(&m->refcnt_atomic, value)
);
+ return
__rte_mbuf_refcnt_update(m, value
);
}
#else /* ! RTE_MBUF_REFCNT_ATOMIC */
}
#else /* ! RTE_MBUF_REFCNT_ATOMIC */
+/* internal */
+static inline uint16_t
+__rte_mbuf_refcnt_update(struct rte_mbuf *m, int16_t value)
+{
+ m->refcnt = (uint16_t)(m->refcnt + value);
+ return m->refcnt;
+}
+
/**
* Adds given value to an mbuf's refcnt and returns its new value.
*/
static inline uint16_t
rte_mbuf_refcnt_update(struct rte_mbuf *m, int16_t value)
{
/**
* Adds given value to an mbuf's refcnt and returns its new value.
*/
static inline uint16_t
rte_mbuf_refcnt_update(struct rte_mbuf *m, int16_t value)
{
- m->refcnt = (uint16_t)(m->refcnt + value);
- return m->refcnt;
+ return __rte_mbuf_refcnt_update(m, value);
}
/**
}
/**
@@
-850,10
+882,10
@@
rte_mbuf_sanity_check(const struct rte_mbuf *m, int is_header);
} while (0)
/**
} while (0)
/**
- * Allocate an unitialized mbuf from mempool *mp*.
+ * Allocate an uni
ni
tialized mbuf from mempool *mp*.
*
* This function can be used by PMDs (especially in RX functions) to
*
* This function can be used by PMDs (especially in RX functions) to
- * allocate an unitialized mbuf. The driver is responsible of
+ * allocate an uni
ni
tialized mbuf. The driver is responsible of
* initializing all the required fields. See rte_pktmbuf_reset().
* For standard needs, prefer rte_pktmbuf_alloc().
*
* initializing all the required fields. See rte_pktmbuf_reset().
* For standard needs, prefer rte_pktmbuf_alloc().
*
@@
-871,11
+903,9
@@
rte_mbuf_sanity_check(const struct rte_mbuf *m, int is_header);
static inline struct rte_mbuf *rte_mbuf_raw_alloc(struct rte_mempool *mp)
{
struct rte_mbuf *m;
static inline struct rte_mbuf *rte_mbuf_raw_alloc(struct rte_mempool *mp)
{
struct rte_mbuf *m;
- void *mb = NULL;
- if (rte_mempool_get(mp,
&mb
) < 0)
+ if (rte_mempool_get(mp,
(void **)&m
) < 0)
return NULL;
return NULL;
- m = (struct rte_mbuf *)mb;
MBUF_RAW_ALLOC_CHECK(m);
return m;
}
MBUF_RAW_ALLOC_CHECK(m);
return m;
}
@@
-1361,8
+1391,7
@@
rte_pktmbuf_prefree_seg(struct rte_mbuf *m)
return m;
return m;
- } else if (rte_atomic16_add_return(&m->refcnt_atomic, -1) == 0) {
-
+ } else if (__rte_mbuf_refcnt_update(m, -1) == 0) {
if (RTE_MBUF_INDIRECT(m))
rte_pktmbuf_detach(m);
if (RTE_MBUF_INDIRECT(m))
rte_pktmbuf_detach(m);
@@
-1410,13
+1439,14
@@
rte_pktmbuf_free_seg(struct rte_mbuf *m)
* segment is added back into its original mempool.
*
* @param m
* segment is added back into its original mempool.
*
* @param m
- * The packet mbuf to be freed.
+ * The packet mbuf to be freed.
If NULL, the function does nothing.
*/
static inline void rte_pktmbuf_free(struct rte_mbuf *m)
{
struct rte_mbuf *m_next;
*/
static inline void rte_pktmbuf_free(struct rte_mbuf *m)
{
struct rte_mbuf *m_next;
- __rte_mbuf_sanity_check(m, 1);
+ if (m != NULL)
+ __rte_mbuf_sanity_check(m, 1);
while (m != NULL) {
m_next = m->next;
while (m != NULL) {
m_next = m->next;
@@
-1447,7
+1477,7
@@
static inline struct rte_mbuf *rte_pktmbuf_clone(struct rte_mbuf *md,
{
struct rte_mbuf *mc, *mi, **prev;
uint32_t pktlen;
{
struct rte_mbuf *mc, *mi, **prev;
uint32_t pktlen;
- uint
8
_t nseg;
+ uint
16
_t nseg;
if (unlikely ((mc = rte_pktmbuf_alloc(mp)) == NULL))
return NULL;
if (unlikely ((mc = rte_pktmbuf_alloc(mp)) == NULL))
return NULL;
@@
-1538,12
+1568,10
@@
static inline uint16_t rte_pktmbuf_tailroom(const struct rte_mbuf *m)
*/
static inline struct rte_mbuf *rte_pktmbuf_lastseg(struct rte_mbuf *m)
{
*/
static inline struct rte_mbuf *rte_pktmbuf_lastseg(struct rte_mbuf *m)
{
- struct rte_mbuf *m2 = (struct rte_mbuf *)m;
-
__rte_mbuf_sanity_check(m, 1);
__rte_mbuf_sanity_check(m, 1);
- while (m
2
->next != NULL)
- m
2 = m2
->next;
- return m
2
;
+ while (m->next != NULL)
+ m
= m
->next;
+ return m;
}
/**
}
/**
@@
-1778,7
+1806,7
@@
const void *__rte_pktmbuf_read(const struct rte_mbuf *m, uint32_t off,
* @param len
* The amount of bytes to read.
* @param buf
* @param len
* The amount of bytes to read.
* @param buf
- * The buffer where data is copied if it is not contigous in mbuf
+ * The buffer where data is copied if it is not contig
u
ous in mbuf
* data. Its length should be at least equal to the len parameter.
* @return
* The pointer to the data, either in the mbuf if it is contiguous,
* data. Its length should be at least equal to the len parameter.
* @return
* The pointer to the data, either in the mbuf if it is contiguous,
@@
-1807,14
+1835,14
@@
static inline const void *rte_pktmbuf_read(const struct rte_mbuf *m,
*
* @return
* - 0, on success.
*
* @return
* - 0, on success.
- * - -EOVERFLOW, if the chain
is full (256 entries)
+ * - -EOVERFLOW, if the chain
segment limit exceeded
*/
static inline int rte_pktmbuf_chain(struct rte_mbuf *head, struct rte_mbuf *tail)
{
struct rte_mbuf *cur_tail;
/* Check for number-of-segments-overflow */
*/
static inline int rte_pktmbuf_chain(struct rte_mbuf *head, struct rte_mbuf *tail)
{
struct rte_mbuf *cur_tail;
/* Check for number-of-segments-overflow */
- if (head->nb_segs + tail->nb_segs >
= 1 << (sizeof(head->nb_segs) * 8)
)
+ if (head->nb_segs + tail->nb_segs >
RTE_MBUF_MAX_NB_SEGS
)
return -EOVERFLOW;
/* Chain 'tail' onto the old tail */
return -EOVERFLOW;
/* Chain 'tail' onto the old tail */
@@
-1822,7
+1850,7
@@
static inline int rte_pktmbuf_chain(struct rte_mbuf *head, struct rte_mbuf *tail
cur_tail->next = tail;
/* accumulate number of segments and total length. */
cur_tail->next = tail;
/* accumulate number of segments and total length. */
- head->nb_segs
= (uint8_t)(head->nb_segs + tail->nb_segs)
;
+ head->nb_segs
+= tail->nb_segs
;
head->pkt_len += tail->pkt_len;
/* pkt_len is only set in the head */
head->pkt_len += tail->pkt_len;
/* pkt_len is only set in the head */