*/
#define MEMPOOL_HEADER_SIZE(mp, pgn) (sizeof(*(mp)) + \
RTE_ALIGN_CEIL(((pgn) - RTE_DIM((mp)->elt_pa)) * \
- sizeof ((mp)->elt_pa[0]), CACHE_LINE_SIZE))
+ sizeof ((mp)->elt_pa[0]), RTE_CACHE_LINE_SIZE))
/**
* Returns TRUE if whole mempool is allocated in one contiguous block of memory.
unsigned n, int is_mc)
{
int ret;
-#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
- unsigned n_orig = n;
-#endif
#if RTE_MEMPOOL_CACHE_MAX_SIZE > 0
struct rte_mempool_cache *cache;
uint32_t index, len;
cache->len -= n;
- __MEMPOOL_STAT_ADD(mp, get_success, n_orig);
+ __MEMPOOL_STAT_ADD(mp, get_success, n);
return 0;
ret = rte_ring_sc_dequeue_bulk(mp->ring, obj_table, n);
if (ret < 0)
- __MEMPOOL_STAT_ADD(mp, get_fail, n_orig);
+ __MEMPOOL_STAT_ADD(mp, get_fail, n);
else
- __MEMPOOL_STAT_ADD(mp, get_success, n_orig);
+ __MEMPOOL_STAT_ADD(mp, get_success, n);
return ret;
}