| /dpdk/app/test/ |
| H A D | test_lpm6.c | 561 depth = 25; in test11() 568 depth = 33; in test11() 575 depth = 41; in test11() 656 depth = 1; in test13() 660 depth = 2; in test13() 664 depth = 3; in test13() 668 depth = 2; in test13() 672 depth = 3; in test13() 834 for (depth = 1; depth <= 16; depth++) { in test17() 851 for (depth = 16; depth >= 1; depth--) { in test17() [all …]
|
| H A D | test_lpm.c | 395 for (depth = 1; depth <= 32; depth++) { in test8() 419 for (depth = 32; depth >= 1; depth--) { in test8() 478 depth = 24; in test9() 500 depth = 23; in test9() 509 depth = 24; in test9() 518 depth = 24; in test9() 523 depth = 23; in test9() 536 depth = 32; in test9() 546 depth = 32; in test9() 562 depth = 32; in test9() [all …]
|
| H A D | test_lpm6_data.h | 12 uint8_t depth; member 1045 const uint8_t *ip_in, uint8_t depth) in mask_ip6_prefix() argument 1051 if (depth >= 8) in mask_ip6_prefix() 1053 else if (depth > 0) { in mask_ip6_prefix() 1061 depth -= 8; in mask_ip6_prefix() 1069 const uint8_t *ip_rule, uint8_t depth) in check_lpm6_rule() argument 1075 if (depth >= 8) { in check_lpm6_rule() 1078 } else if (depth > 0) { in check_lpm6_rule() 1087 depth -= 8; in check_lpm6_rule() 1107 if (rule[i].depth >= max_depth) { in get_next_hop() [all …]
|
| H A D | test_fib_perf.c | 46 uint8_t depth; member 211 rule_num = rule_count.a[depth - 1]; in generate_random_rule_prefix() 215 rule_num = rule_count.b[depth - 1]; in generate_random_rule_prefix() 219 rule_num = rule_count.c[depth - 1]; in generate_random_rule_prefix() 228 class_depth = depth - fixed_bit_num; in generate_random_rule_prefix() 254 ptr_rule->depth = depth; in generate_random_rule_prefix() 277 large_route_table[pos].depth = depth; in insert_rule_in_random_pos() 285 uint8_t depth; in generate_large_route_rule_table() local 291 for (depth = 1; depth <= RTE_FIB_MAX_DEPTH; depth++) in generate_large_route_rule_table() 321 if (table[j].depth == (uint8_t) i) in print_route_distribution() [all …]
|
| H A D | test_lpm_perf.c | 61 uint8_t depth; member 232 rule_num = rule_count.a[depth - 1]; in generate_random_rule_prefix() 236 rule_num = rule_count.b[depth - 1]; in generate_random_rule_prefix() 240 rule_num = rule_count.c[depth - 1]; in generate_random_rule_prefix() 276 ptr_rule->depth = depth; in generate_random_rule_prefix() 280 if (depth > 24) { in generate_random_rule_prefix() 282 ptr_ldepth_rule->depth = ptr_rule->depth; in generate_random_rule_prefix() 308 large_route_table[pos].depth = depth; in insert_rule_in_random_pos() 316 uint8_t depth; in generate_large_route_rule_table() local 323 for (depth = 1; depth <= RTE_LPM_MAX_DEPTH; depth++) { in generate_large_route_rule_table() [all …]
|
| H A D | test_rib.c | 139 uint8_t depth = 24; in test_insert_invalid() local 145 node = rte_rib_insert(NULL, ip, depth); in test_insert_invalid() 159 node = rte_rib_insert(rib, ip, depth); in test_insert_invalid() 161 node1 = rte_rib_insert(rib, ip, depth); in test_insert_invalid() 186 uint8_t depth = 24; in test_get_fn() local 196 node = rte_rib_insert(rib, ip, depth); in test_get_fn() 265 uint8_t depth = 24; in test_basic() local 274 node = rte_rib_insert(rib, ip, depth); in test_basic() 296 rte_rib_remove(rib, ip, depth); in test_basic() 319 uint8_t depth = 30; in test_tree_traversal() local [all …]
|
| H A D | test_rib6.c | 140 uint8_t depth = 24; in test_insert_invalid() local 146 node = rte_rib6_insert(NULL, ip, depth); in test_insert_invalid() 160 node = rte_rib6_insert(rib, ip, depth); in test_insert_invalid() 162 node1 = rte_rib6_insert(rib, ip, depth); in test_insert_invalid() 188 uint8_t depth = 24; in test_get_fn() local 198 node = rte_rib6_insert(rib, ip, depth); in test_get_fn() 268 uint8_t depth = 24; in test_basic() local 277 node = rte_rib6_insert(rib, ip, depth); in test_basic() 299 rte_rib6_remove(rib, ip, depth); in test_basic() 326 uint8_t depth = 126; in test_tree_traversal() local [all …]
|
| /dpdk/lib/lpm/ |
| H A D | rte_lpm.c | 590 i_lpm->lpm.tbl24[i].depth <= depth)) { in add_depth_small() 596 .depth = depth, in add_depth_small() 619 i_lpm->lpm.tbl8[j].depth <= depth) { in add_depth_small() 624 .depth = depth, in add_depth_small() 675 .depth = depth, in add_depth_big() 734 .depth = depth, in add_depth_big() 772 i_lpm->lpm.tbl8[i].depth <= depth) { in add_depth_big() 775 .depth = depth, in add_depth_big() 923 i_lpm->lpm.tbl24[i].depth <= depth) { in delete_depth_small() 1080 if (i_lpm->lpm.tbl8[i].depth <= depth) in delete_depth_big() [all …]
|
| H A D | rte_lpm6.c | 217 key->depth = depth; in rule_key_init() 677 tbl[i].depth <= depth)) { in add_step() 681 .depth = depth, in add_step() 697 expand_rule(lpm, tbl8_gindex, depth, depth, in add_step() 764 .depth = tbl[entry_ind].depth, in add_step() 1163 rule_key.depth = depth; in rule_find_less_specific() 1168 rule->depth = depth; in rule_find_less_specific() 1262 .depth = lsp_rule->depth, in remove_tbl() 1346 depth, lsp_rule->depth, in rte_lpm6_delete() 1356 } else if (from->depth == depth) { in rte_lpm6_delete() [all …]
|
| /dpdk/lib/rib/ |
| H A D | rte_rib.c | 35 uint8_t depth; member 136 if ((cur->ip == ip) && (cur->depth == depth) && in __rib_lookup_exact() 139 if ((cur->depth > depth) || in __rib_lookup_exact() 177 while ((tmp) && (tmp->depth < depth)) in rte_rib_get_nxt() 186 (tmp->depth > depth))) in rte_rib_get_nxt() 194 (tmp->depth > depth))) { in rte_rib_get_nxt() 269 new_node->depth = depth; in rte_rib_insert() 287 if ((ip == (*tmp)->ip) && (depth == (*tmp)->depth)) { in rte_rib_insert() 293 d = (*tmp)->depth; in rte_rib_insert() 300 common_depth = RTE_MIN(depth, (*tmp)->depth); in rte_rib_insert() [all …]
|
| H A D | rte_rib6.c | 35 uint8_t depth; member 185 (cur->depth == depth) && in rte_rib6_lookup_exact() 190 (cur->depth >= depth)) in rte_rib6_lookup_exact() 223 while ((tmp) && (tmp->depth < depth)) in rte_rib6_get_nxt() 232 (tmp->depth > depth))) in rte_rib6_get_nxt() 240 (tmp->depth > depth))) { in rte_rib6_get_nxt() 322 new_node->depth = depth; in rte_rib6_insert() 341 (depth == (*tmp)->depth)) { in rte_rib6_insert() 349 ((*tmp)->depth >= depth)) { in rte_rib6_insert() 359 common_depth = RTE_MIN(depth, (*tmp)->depth); in rte_rib6_insert() [all …]
|
| H A D | rte_rib6.h | 104 get_msk_part(uint8_t depth, int byte) { in get_msk_part() argument 108 depth = RTE_MIN(depth, 128); in get_msk_part() 109 part = RTE_MAX((int16_t)depth - (byte * 8), 0); in get_msk_part() 157 const uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE], uint8_t depth); 185 uint8_t depth, struct rte_rib6_node *last, int flag); 199 const uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE], uint8_t depth); 216 const uint8_t ip[RTE_RIB6_IPV6_ADDR_SIZE], uint8_t depth); 245 rte_rib6_get_depth(const struct rte_rib6_node *node, uint8_t *depth);
|
| H A D | rte_rib.h | 60 rte_rib_depth_to_mask(uint8_t depth) in rte_rib_depth_to_mask() argument 62 return (uint32_t)(UINT64_MAX << (32 - depth)); in rte_rib_depth_to_mask() 106 rte_rib_lookup_exact(struct rte_rib *rib, uint32_t ip, uint8_t depth); 132 rte_rib_get_nxt(struct rte_rib *rib, uint32_t ip, uint8_t depth, 146 rte_rib_remove(struct rte_rib *rib, uint32_t ip, uint8_t depth); 162 rte_rib_insert(struct rte_rib *rib, uint32_t ip, uint8_t depth); 190 rte_rib_get_depth(const struct rte_rib_node *node, uint8_t *depth);
|
| /dpdk/app/test-fib/ |
| H A D | main.c | 120 uint8_t depth; member 174 uint8_t depth; in complete_distrib() local 182 for (depth = 0; depth <= depth_lim; depth++) { in complete_distrib() 202 for (depth = 0; depth <= depth_lim; depth++) { in complete_distrib() 287 tmp.depth = rt[i].depth; in shuffle_rt_4() 291 rt[i].depth = rt[j].depth; in shuffle_rt_4() 295 rt[j].depth = tmp.depth; in shuffle_rt_4() 309 tmp.depth = rt[i].depth; in shuffle_rt_6() 313 rt[i].depth = rt[j].depth; in shuffle_rt_6() 317 rt[j].depth = tmp.depth; in shuffle_rt_6() [all …]
|
| /dpdk/examples/ipsec-secgw/ |
| H A D | sp6.c | 355 (depth > 32) ? 32 : depth; in parse_sp6_tokens() 356 depth = (depth > 32) ? (depth - 32) : 0; in parse_sp6_tokens() 363 (depth > 32) ? 32 : depth; in parse_sp6_tokens() 364 depth = (depth > 32) ? (depth - 32) : 0; in parse_sp6_tokens() 371 (depth > 32) ? 32 : depth; in parse_sp6_tokens() 372 depth = (depth > 32) ? (depth - 32) : 0; in parse_sp6_tokens() 379 (depth > 32) ? 32 : depth; in parse_sp6_tokens() 409 (depth > 32) ? 32 : depth; in parse_sp6_tokens() 410 depth = (depth > 32) ? (depth - 32) : 0; in parse_sp6_tokens() 418 depth = (depth > 32) ? (depth - 32) : 0; in parse_sp6_tokens() [all …]
|
| H A D | rt.c | 22 uint8_t depth; member 28 uint8_t depth; member 78 uint32_t depth = 0; in parse_rt_tokens() local 81 &ip, &depth) == 0, status, in parse_rt_tokens() 89 route_ipv4->depth = (uint8_t)depth; in parse_rt_tokens() 92 uint32_t depth; in parse_rt_tokens() local 95 &ip, &depth) == 0, status, in parse_rt_tokens() 102 route_ipv6->depth = (uint8_t)depth; in parse_rt_tokens() 163 ret = rte_lpm_add(lpm, rt_ip4[i].ip, rt_ip4[i].depth, in rt_init() 171 a, b, c, d, rt_ip4[i].depth, in rt_init() [all …]
|
| H A D | flow.c | 54 uint32_t depth; in ipv4_addr_cpy() local 56 APP_CHECK(parse_ipv4_addr(token, &ip, &depth) == 0, status, in ipv4_addr_cpy() 61 if (depth > 32) in ipv4_addr_cpy() 67 if (depth < 32) in ipv4_addr_cpy() 68 *mask = *mask << (32-depth); in ipv4_addr_cpy() 106 uint32_t depth, i; in ipv6_addr_cpy() local 108 APP_CHECK(parse_ipv6_addr(token, &ip, &depth) == 0, status, in ipv6_addr_cpy() 116 for (i = 0; i < depth && (i%8 <= sizeof(struct in6_addr)); i++) in ipv6_addr_cpy()
|
| /dpdk/lib/fib/ |
| H A D | dir24_8.c | 363 uint8_t depth, uint64_t next_hop) in modify_fib() argument 372 tmp = rte_rib_get_nxt(rib, ip, depth, tmp, in modify_fib() 376 if (tmp_depth == depth) in modify_fib() 427 ip &= rte_rib_depth_to_mask(depth); in dir24_8_modify() 429 node = rte_rib_lookup_exact(rib, ip, depth); in dir24_8_modify() 441 if (depth > 24) { in dir24_8_modify() 449 node = rte_rib_insert(rib, ip, depth); in dir24_8_modify() 461 rte_rib_remove(rib, ip, depth); in dir24_8_modify() 464 if ((depth > 24) && (tmp == NULL)) in dir24_8_modify() 480 rte_rib_remove(rib, ip, depth); in dir24_8_modify() [all …]
|
| H A D | trie.c | 434 get_nxt_net(uint8_t *ip, uint8_t depth) in get_nxt_net() argument 457 uint8_t depth, uint64_t next_hop) in modify_dp() argument 474 if (tmp_depth == depth) in modify_dp() 489 get_nxt_net(redge, depth); in modify_dp() 504 uint8_t depth, uint64_t next_hop, int op) in trie_modify() argument 527 if (depth > 24) { in trie_modify() 529 RTE_ALIGN_FLOOR(depth, 8), NULL, in trie_modify() 537 depth_diff = RTE_ALIGN_CEIL(depth, 8) - in trie_modify() 555 if ((depth > 24) && (dp->rsvd_tbl8s >= in trie_modify() 571 rte_rib6_remove(rib, ip_masked, depth); in trie_modify() [all …]
|
| H A D | rte_fib.c | 66 dummy_modify(struct rte_fib *fib, uint32_t ip, uint8_t depth, in dummy_modify() argument 70 if ((fib == NULL) || (depth > RTE_FIB_MAXDEPTH)) in dummy_modify() 73 node = rte_rib_lookup_exact(fib->rib, ip, depth); in dummy_modify() 78 node = rte_rib_insert(fib->rib, ip, depth); in dummy_modify() 85 rte_rib_remove(fib->rib, ip, depth); in dummy_modify() 119 rte_fib_add(struct rte_fib *fib, uint32_t ip, uint8_t depth, uint64_t next_hop) in rte_fib_add() argument 122 (depth > RTE_FIB_MAXDEPTH)) in rte_fib_add() 124 return fib->modify(fib, ip, depth, next_hop, RTE_FIB_ADD); in rte_fib_add() 128 rte_fib_delete(struct rte_fib *fib, uint32_t ip, uint8_t depth) in rte_fib_delete() argument 131 (depth > RTE_FIB_MAXDEPTH)) in rte_fib_delete() [all …]
|
| H A D | rte_fib6.c | 67 uint8_t depth, uint64_t next_hop, int op) in dummy_modify() argument 70 if ((fib == NULL) || (depth > RTE_FIB6_MAXDEPTH)) in dummy_modify() 73 node = rte_rib6_lookup_exact(fib->rib, ip, depth); in dummy_modify() 78 node = rte_rib6_insert(fib->rib, ip, depth); in dummy_modify() 85 rte_rib6_remove(fib->rib, ip, depth); in dummy_modify() 119 uint8_t depth, uint64_t next_hop) in rte_fib6_add() argument 122 (depth > RTE_FIB6_MAXDEPTH)) in rte_fib6_add() 124 return fib->modify(fib, ip, depth, next_hop, RTE_FIB6_ADD); in rte_fib6_add() 129 uint8_t depth) in rte_fib6_delete() argument 132 (depth > RTE_FIB6_MAXDEPTH)) in rte_fib6_delete() [all …]
|
| /dpdk/drivers/net/nfp/nfpcore/ |
| H A D | nfp_mutex.c | 31 uint16_t depth; member 162 mutex->depth = 0; in nfp_cpp_mutex_alloc() 270 printf("\tdepth:%hd]\n", mutex->depth); in nfp_cpp_mutex_lock() 298 if (mutex->depth > 1) { in nfp_cpp_mutex_unlock() 299 mutex->depth--; in nfp_cpp_mutex_unlock() 325 mutex->depth = 0; in nfp_cpp_mutex_unlock() 353 if (mutex->depth > 0) { in nfp_cpp_mutex_trylock() 354 if (mutex->depth == MUTEX_DEPTH_MAX) in nfp_cpp_mutex_trylock() 357 mutex->depth++; in nfp_cpp_mutex_trylock() 410 mutex->depth = 1; in nfp_cpp_mutex_trylock() [all …]
|
| /dpdk/lib/table/ |
| H A D | rte_table_lpm_ipv6.c | 200 if ((ip_prefix->depth == 0) || (ip_prefix->depth > 128)) { in rte_table_lpm_ipv6_entry_add() 202 ip_prefix->depth); in rte_table_lpm_ipv6_entry_add() 208 ip_prefix->depth, &nht_pos0); in rte_table_lpm_ipv6_entry_add() 225 if (rte_lpm6_add(lpm->lpm, ip_prefix->ip, ip_prefix->depth, in rte_table_lpm_ipv6_entry_add() 263 if ((ip_prefix->depth == 0) || (ip_prefix->depth > 128)) { in rte_table_lpm_ipv6_entry_delete() 265 ip_prefix->depth); in rte_table_lpm_ipv6_entry_delete() 271 ip_prefix->depth, &nht_pos); in rte_table_lpm_ipv6_entry_delete() 283 status = rte_lpm6_delete(lpm->lpm, ip_prefix->ip, ip_prefix->depth); in rte_table_lpm_ipv6_entry_delete()
|
| H A D | rte_table_lpm.c | 203 if ((ip_prefix->depth == 0) || (ip_prefix->depth > 32)) { in rte_table_lpm_entry_add() 205 __func__, ip_prefix->depth); in rte_table_lpm_entry_add() 211 ip_prefix->depth, &nht_pos0); in rte_table_lpm_entry_add() 228 if (rte_lpm_add(lpm->lpm, ip_prefix->ip, ip_prefix->depth, nht_pos) < 0) { in rte_table_lpm_entry_add() 264 if ((ip_prefix->depth == 0) || (ip_prefix->depth > 32)) { in rte_table_lpm_entry_delete() 266 ip_prefix->depth); in rte_table_lpm_entry_delete() 272 ip_prefix->depth, &nht_pos); in rte_table_lpm_entry_delete() 283 status = rte_lpm_delete(lpm->lpm, ip_prefix->ip, ip_prefix->depth); in rte_table_lpm_entry_delete()
|
| /dpdk/drivers/raw/skeleton/ |
| H A D | skeleton_rawdev_test.c | 195 RTE_TEST_ASSERT_EQUAL(q.depth, SKELETON_QUEUE_DEF_DEPTH, in test_rawdev_queue_default_conf_get() 197 q.depth); in test_rawdev_queue_default_conf_get() 242 qset.depth = 15; in test_rawdev_queue_setup() 251 RTE_TEST_ASSERT_EQUAL(qset.depth, qget.depth, in test_rawdev_queue_setup() 253 qset.depth, qget.depth); in test_rawdev_queue_setup() 275 RTE_TEST_ASSERT_EQUAL(qget.depth, SKELETON_QUEUE_DEF_DEPTH, in test_rawdev_queue_release()
|