1 /* SPDX-License-Identifier: BSD-3-Clause
2 * Copyright 2018 6WIND S.A.
3 * Copyright 2018 Mellanox Technologies, Ltd
6 #ifndef RTE_PMD_MLX5_MR_H_
7 #define RTE_PMD_MLX5_MR_H_
11 #include <sys/queue.h>
14 /* ISO C doesn't support unnamed structs/unions, disabling -pedantic. */
16 #pragma GCC diagnostic ignored "-Wpedantic"
18 #include <infiniband/verbs.h>
19 #include <infiniband/mlx5dv.h>
21 #pragma GCC diagnostic error "-Wpedantic"
24 #include <rte_ethdev.h>
25 #include <rte_rwlock.h>
26 #include <rte_bitmap.h>
28 /* Memory Region object. */
30 LIST_ENTRY(mlx5_mr) mr; /**< Pointer to the prev/next entry. */
31 struct ibv_mr *ibv_mr; /* Verbs Memory Region. */
32 const struct rte_memseg_list *msl;
33 int ms_base_idx; /* Start index of msl->memseg_arr[]. */
34 int ms_n; /* Number of memsegs in use. */
35 uint32_t ms_bmp_n; /* Number of bits in memsegs bit-mask. */
36 struct rte_bitmap *ms_bmp; /* Bit-mask of memsegs belonged to MR. */
39 /* Cache entry for Memory Region. */
40 struct mlx5_mr_cache {
41 uintptr_t start; /* Start address of MR. */
42 uintptr_t end; /* End address of MR. */
43 uint32_t lkey; /* rte_cpu_to_be_32(ibv_mr->lkey). */
46 /* MR Cache table for Binary search. */
47 struct mlx5_mr_btree {
48 uint16_t len; /* Number of entries. */
49 uint16_t size; /* Total number of entries. */
50 int overflow; /* Mark failure of table expansion. */
51 struct mlx5_mr_cache (*table)[];
54 /* Per-queue MR control descriptor. */
56 uint32_t *dev_gen_ptr; /* Generation number of device to poll. */
57 uint32_t cur_gen; /* Generation number saved to flush caches. */
58 uint16_t mru; /* Index of last hit entry in top-half cache. */
59 uint16_t head; /* Index of the oldest entry in top-half cache. */
60 struct mlx5_mr_cache cache[MLX5_MR_CACHE_N]; /* Cache for top-half. */
61 struct mlx5_mr_btree cache_bh; /* Cache for bottom-half. */
64 struct mlx5_ibv_shared;
65 extern struct mlx5_dev_list mlx5_mem_event_cb_list;
66 extern rte_rwlock_t mlx5_mem_event_rwlock;
68 /* First entry must be NULL for comparison. */
69 #define mlx5_mr_btree_len(bt) ((bt)->len - 1)
71 int mlx5_mr_btree_init(struct mlx5_mr_btree *bt, int n, int socket);
72 void mlx5_mr_btree_free(struct mlx5_mr_btree *bt);
73 uint32_t mlx5_mr_create_primary(struct rte_eth_dev *dev,
74 struct mlx5_mr_cache *entry, uintptr_t addr);
75 void mlx5_mr_mem_event_cb(enum rte_mem_event event_type, const void *addr,
76 size_t len, void *arg);
77 int mlx5_mr_update_mp(struct rte_eth_dev *dev, struct mlx5_mr_ctrl *mr_ctrl,
78 struct rte_mempool *mp);
79 void mlx5_mr_release(struct mlx5_ibv_shared *sh);
81 /* Debug purpose functions. */
82 void mlx5_mr_btree_dump(struct mlx5_mr_btree *bt);
83 void mlx5_mr_dump_dev(struct mlx5_ibv_shared *sh);
86 * Look up LKey from given lookup table by linear search. Firstly look up the
87 * last-hit entry. If miss, the entire array is searched. If found, update the
88 * last-hit index and return LKey.
91 * Pointer to lookup table.
92 * @param[in,out] cached_idx
93 * Pointer to last-hit index.
95 * Size of lookup table.
100 * Searched LKey on success, UINT32_MAX on no match.
102 static __rte_always_inline uint32_t
103 mlx5_mr_lookup_cache(struct mlx5_mr_cache *lkp_tbl, uint16_t *cached_idx,
104 uint16_t n, uintptr_t addr)
108 if (likely(addr >= lkp_tbl[*cached_idx].start &&
109 addr < lkp_tbl[*cached_idx].end))
110 return lkp_tbl[*cached_idx].lkey;
111 for (idx = 0; idx < n && lkp_tbl[idx].start != 0; ++idx) {
112 if (addr >= lkp_tbl[idx].start &&
113 addr < lkp_tbl[idx].end) {
116 return lkp_tbl[idx].lkey;
122 #endif /* RTE_PMD_MLX5_MR_H_ */