mempool: fold memory size calculation helper
authorAndrew Rybchenko <arybchenko@solarflare.com>
Fri, 27 Jul 2018 13:46:05 +0000 (14:46 +0100)
committerThomas Monjalon <thomas@monjalon.net>
Sun, 5 Aug 2018 21:41:57 +0000 (23:41 +0200)
rte_mempool_calc_mem_size_helper() was introduced to avoid
code duplication and used in deprecated rte_mempool_mem_size() and
rte_mempool_op_calc_mem_size_default(). Now the first one is removed
and it is better to fold the helper into the second one to make it
more readable.

Signed-off-by: Andrew Rybchenko <arybchenko@solarflare.com>
lib/librte_mempool/rte_mempool.c
lib/librte_mempool/rte_mempool.h
lib/librte_mempool/rte_mempool_ops_default.c

index a840da4..03e6b5f 100644 (file)
@@ -225,31 +225,6 @@ rte_mempool_calc_obj_size(uint32_t elt_size, uint32_t flags,
        return sz->total_size;
 }
 
-
-/*
- * Internal function to calculate required memory chunk size.
- */
-size_t
-rte_mempool_calc_mem_size_helper(uint32_t elt_num, size_t total_elt_sz,
-                                uint32_t pg_shift)
-{
-       size_t obj_per_page, pg_num, pg_sz;
-
-       if (total_elt_sz == 0)
-               return 0;
-
-       if (pg_shift == 0)
-               return total_elt_sz * elt_num;
-
-       pg_sz = (size_t)1 << pg_shift;
-       obj_per_page = pg_sz / total_elt_sz;
-       if (obj_per_page == 0)
-               return RTE_ALIGN_CEIL(total_elt_sz, pg_sz) * elt_num;
-
-       pg_num = (elt_num + obj_per_page - 1) / obj_per_page;
-       return pg_num << pg_shift;
-}
-
 /* free a memchunk allocated with rte_memzone_reserve() */
 static void
 rte_mempool_memchunk_mz_free(__rte_unused struct rte_mempool_memhdr *memhdr,
index 5d16025..7c9cd9a 100644 (file)
@@ -487,28 +487,6 @@ ssize_t rte_mempool_op_calc_mem_size_default(const struct rte_mempool *mp,
                uint32_t obj_num, uint32_t pg_shift,
                size_t *min_chunk_size, size_t *align);
 
-/**
- * @internal Helper function to calculate memory size required to store
- * specified number of objects in assumption that the memory buffer will
- * be aligned at page boundary.
- *
- * Note that if object size is bigger than page size, then it assumes
- * that pages are grouped in subsets of physically continuous pages big
- * enough to store at least one object.
- *
- * @param elt_num
- *   Number of elements.
- * @param total_elt_sz
- *   The size of each element, including header and trailer, as returned
- *   by rte_mempool_calc_obj_size().
- * @param pg_shift
- *   LOG2 of the physical pages size. If set to 0, ignore page boundaries.
- * @return
- *   Required memory size aligned at page boundary.
- */
-size_t rte_mempool_calc_mem_size_helper(uint32_t elt_num, size_t total_elt_sz,
-               uint32_t pg_shift);
-
 /**
  * Function to be called for each populated object.
  *
index fd63ca1..4e2bfc8 100644 (file)
@@ -12,12 +12,31 @@ rte_mempool_op_calc_mem_size_default(const struct rte_mempool *mp,
                                     size_t *min_chunk_size, size_t *align)
 {
        size_t total_elt_sz;
+       size_t obj_per_page, pg_num, pg_sz;
        size_t mem_size;
 
        total_elt_sz = mp->header_size + mp->elt_size + mp->trailer_size;
-
-       mem_size = rte_mempool_calc_mem_size_helper(obj_num, total_elt_sz,
-                                                   pg_shift);
+       if (total_elt_sz == 0) {
+               mem_size = 0;
+       } else if (pg_shift == 0) {
+               mem_size = total_elt_sz * obj_num;
+       } else {
+               pg_sz = (size_t)1 << pg_shift;
+               obj_per_page = pg_sz / total_elt_sz;
+               if (obj_per_page == 0) {
+                       /*
+                        * Note that if object size is bigger than page size,
+                        * then it is assumed that pages are grouped in subsets
+                        * of physically continuous pages big enough to store
+                        * at least one object.
+                        */
+                       mem_size =
+                               RTE_ALIGN_CEIL(total_elt_sz, pg_sz) * obj_num;
+               } else {
+                       pg_num = (obj_num + obj_per_page - 1) / obj_per_page;
+                       mem_size = pg_num << pg_shift;
+               }
+       }
 
        *min_chunk_size = RTE_MAX((size_t)1 << pg_shift, total_elt_sz);