From: Intel Date: Wed, 19 Dec 2012 23:00:00 +0000 (+0100) Subject: lpm: rework lpm lookup + add bulk function X-Git-Tag: spdx-start~11394 X-Git-Url: http://git.droids-corp.org/?a=commitdiff_plain;h=3569fcf1c4be7ac6f0179b40fcfba4c89b685e1b;p=dpdk.git lpm: rework lpm lookup + add bulk function Signed-off-by: Intel --- diff --git a/lib/librte_lpm/rte_lpm.c b/lib/librte_lpm/rte_lpm.c index f1c9892407..b1bc6b301b 100644 --- a/lib/librte_lpm/rte_lpm.c +++ b/lib/librte_lpm/rte_lpm.c @@ -376,10 +376,10 @@ add_depth_small(struct rte_lpm *lpm, uint32_t ip, uint8_t depth, lpm->tbl24[i].depth <= depth)) { struct rte_lpm_tbl24_entry new_tbl24_entry = { + { .next_hop = next_hop, }, .valid = VALID, .ext_entry = 0, .depth = depth, - { .next_hop = next_hop, } }; /* Setting tbl24 entry in one go to avoid race @@ -458,10 +458,10 @@ add_depth_big(struct rte_lpm *lpm, uint32_t ip_masked, uint8_t depth, */ struct rte_lpm_tbl24_entry new_tbl24_entry = { + { .tbl8_gindex = (uint8_t)tbl8_group_index, }, .valid = VALID, .ext_entry = 1, .depth = 0, - { .tbl8_gindex = (uint8_t)tbl8_group_index, } }; lpm->tbl24[tbl24_index] = new_tbl24_entry; @@ -509,10 +509,10 @@ add_depth_big(struct rte_lpm *lpm, uint32_t ip_masked, uint8_t depth, */ struct rte_lpm_tbl24_entry new_tbl24_entry = { + { .tbl8_gindex = (uint8_t)tbl8_group_index, }, .valid = VALID, .ext_entry = 1, .depth = 0, - { .tbl8_gindex = (uint8_t)tbl8_group_index, } }; lpm->tbl24[tbl24_index] = new_tbl24_entry; @@ -670,10 +670,10 @@ delete_depth_small(struct rte_lpm *lpm, uint32_t ip_masked, lpm->max_rules_per_depth); struct rte_lpm_tbl24_entry new_tbl24_entry = { + {.next_hop = lpm->rules_tbl[sub_rule_index].next_hop,}, .valid = VALID, .ext_entry = 0, .depth = new_depth, - {.next_hop = lpm->rules_tbl[sub_rule_index].next_hop,} }; struct rte_lpm_tbl8_entry new_tbl8_entry = { @@ -835,10 +835,10 @@ delete_depth_big(struct rte_lpm *lpm, uint32_t ip_masked, else if (tbl8_recycle_index > -1) { /* Update tbl24 entry. */ struct rte_lpm_tbl24_entry new_tbl24_entry = { + { .next_hop = lpm->tbl8[tbl8_recycle_index].next_hop, }, .valid = VALID, .ext_entry = 0, .depth = lpm->tbl8[tbl8_recycle_index].depth, - { .next_hop = lpm->tbl8[tbl8_recycle_index].next_hop, } }; /* Set tbl24 before freeing tbl8 to avoid race condition. */ diff --git a/lib/librte_lpm/rte_lpm.h b/lib/librte_lpm/rte_lpm.h index 56eff4e338..b9e03b13a6 100644 --- a/lib/librte_lpm/rte_lpm.h +++ b/lib/librte_lpm/rte_lpm.h @@ -97,24 +97,24 @@ extern "C" { /** @internal Tbl24 entry structure. */ struct rte_lpm_tbl24_entry { - /* Using single uint8_t to store 3 values. */ - uint8_t valid :1; /**< Validation flag. */ - uint8_t ext_entry :1; /**< External entry. */ - uint8_t depth :6; /**< Rule depth. */ /* Stores Next hop or group index (i.e. gindex)into tbl8. */ union { uint8_t next_hop; uint8_t tbl8_gindex; }; + /* Using single uint8_t to store 3 values. */ + uint8_t valid :1; /**< Validation flag. */ + uint8_t ext_entry :1; /**< External entry. */ + uint8_t depth :6; /**< Rule depth. */ }; /** @internal Tbl8 entry structure. */ struct rte_lpm_tbl8_entry { + uint8_t next_hop; /**< next hop. */ /* Using single uint8_t to store 3 values. */ uint8_t valid :1; /**< Validation flag. */ uint8_t valid_group :1; /**< Group validation flag. */ uint8_t depth :6; /**< Rule depth. */ - uint8_t next_hop; /**< next hop. */ }; /** @internal Rule structure. */ @@ -247,45 +247,83 @@ rte_lpm_delete_all(struct rte_lpm *lpm); static inline int rte_lpm_lookup(struct rte_lpm *lpm, uint32_t ip, uint8_t *next_hop) { - uint32_t tbl24_index, tbl8_group_index, tbl8_index; + unsigned tbl24_index = (ip >> 8); + uint16_t tbl_entry; /* DEBUG: Check user input arguments. */ RTE_LPM_RETURN_IF_TRUE(((lpm == NULL) || (next_hop == NULL)), -EINVAL); - /* Calculate index into tbl24. */ - tbl24_index = (ip >> 8); + /* Copy tbl24 entry */ + tbl_entry = *(const uint16_t *)&lpm->tbl24[tbl24_index]; + + /* Copy tbl8 entry (only if needed) */ + if (unlikely((tbl_entry & RTE_LPM_VALID_EXT_ENTRY_BITMASK) == + RTE_LPM_VALID_EXT_ENTRY_BITMASK)) { + + unsigned tbl8_index = (uint8_t)ip + + ((uint8_t)tbl_entry * RTE_LPM_TBL8_GROUP_NUM_ENTRIES); - /* - * Use the tbl24_index to access the required tbl24 entry then check if - * the tbl24 entry is INVALID, if so return -ENOENT. - */ - if (!lpm->tbl24[tbl24_index].valid){ - return -ENOENT; /* Lookup miss. */ + tbl_entry = *(const uint16_t *)&lpm->tbl8[tbl8_index]; } - /* - * If tbl24 entry is valid check if it is NOT extended (i.e. it does - * not use a tbl8 extension) if so return the next hop. - */ - if (likely(lpm->tbl24[tbl24_index].ext_entry == 0)) { - *next_hop = lpm->tbl24[tbl24_index].next_hop; - return 0; /* Lookup hit. */ + + *next_hop = (uint8_t)tbl_entry; + return (tbl_entry & RTE_LPM_LOOKUP_SUCCESS) ? 0 : -ENOENT; +} + +/** + * Lookup multiple IP addresses in an LPM table. This may be implemented as a + * macro, so the address of the function should not be used. + * + * @param lpm + * LPM object handle + * @param ips + * Array of IPs to be looked up in the LPM table + * @param next_hops + * Next hop of the most specific rule found for IP (valid on lookup hit only). + * This is an array of two byte values. The most significant byte in each + * value says whether the lookup was successful (bitmask + * RTE_LPM_LOOKUP_SUCCESS is set). The least significant byte is the + * actual next hop. + * @param n + * Number of elements in ips (and next_hops) array to lookup. This should be a + * compile time constant, and divisible by 8 for best performance. + * @return + * -EINVAL for incorrect arguments, otherwise 0 + */ +#define rte_lpm_lookup_bulk(lpm, ips, next_hops, n) \ + rte_lpm_lookup_bulk_func(lpm, ips, next_hops, n) + +static inline int +rte_lpm_lookup_bulk_func(const struct rte_lpm *lpm, const uint32_t * ips, + uint16_t * next_hops, const unsigned n) +{ + unsigned i; + unsigned tbl24_indexes[n]; + + /* DEBUG: Check user input arguments. */ + RTE_LPM_RETURN_IF_TRUE(((lpm == NULL) || (ips == NULL) || + (next_hops == NULL)), -EINVAL); + + for (i = 0; i < n; i++) { + tbl24_indexes[i] = ips[i] >> 8; } - /* - * If tbl24 entry is valid and extended calculate the index into the - * tbl8 entry. - */ - tbl8_group_index = lpm->tbl24[tbl24_index].tbl8_gindex; - tbl8_index = (tbl8_group_index * RTE_LPM_TBL8_GROUP_NUM_ENTRIES) + - (ip & 0xFF); - - /* Check if the tbl8 entry is invalid and if so return -ENOENT. */ - if (!lpm->tbl8[tbl8_index].valid) - return -ENOENT;/* Lookup miss. */ - - /* If the tbl8 entry is valid return return the next_hop. */ - *next_hop = lpm->tbl8[tbl8_index].next_hop; - return 0; /* Lookup hit. */ + for (i = 0; i < n; i++) { + /* Simply copy tbl24 entry to output */ + next_hops[i] = *(const uint16_t *)&lpm->tbl24[tbl24_indexes[i]]; + + /* Overwrite output with tbl8 entry if needed */ + if (unlikely((next_hops[i] & RTE_LPM_VALID_EXT_ENTRY_BITMASK) == + RTE_LPM_VALID_EXT_ENTRY_BITMASK)) { + + unsigned tbl8_index = (uint8_t)ips[i] + + ((uint8_t)next_hops[i] * + RTE_LPM_TBL8_GROUP_NUM_ENTRIES); + + next_hops[i] = *(const uint16_t *)&lpm->tbl8[tbl8_index]; + } + } + return 0; } #ifdef __cplusplus