| /f-stack/dpdk/drivers/event/opdl/ |
| H A D | opdl_ring.c | 295 num_entries = num_to_process(s, num_entries, block); in opdl_ring_input_singlethread() 304 return num_entries; in opdl_ring_input_singlethread() 513 num_entries = s->nb_instance * num_entries; in opdl_stage_claim_singlethread() 515 num_entries = num_to_process(s, num_entries, block); in opdl_stage_claim_singlethread() 524 num_entries = num_to_process(s, num_entries, block); in opdl_stage_claim_singlethread() 573 num_entries = offset + (s->nb_instance * num_entries); in opdl_stage_claim_multithread() 594 num_entries = num_to_process(s, num_entries, block); in opdl_stage_claim_copy_singlethread() 653 num_entries = num_to_process(s, num_entries, block); in opdl_ring_copy_from_burst() 673 num_entries = num_to_process(s, num_entries, block); in opdl_ring_copy_to_burst() 701 return (avail <= num_entries) ? avail : num_entries; in opdl_stage_find_num_available() [all …]
|
| H A D | opdl_ring.h | 228 opdl_ring_input(struct opdl_ring *t, const void *entries, uint32_t num_entries, 255 const void *entries, uint32_t num_entries, bool block); 281 void *entries, uint32_t num_entries, bool block); 312 uint32_t num_entries, uint32_t *seq, bool block, bool atomic); 336 uint32_t num_entries, void *arg, bool block); 368 uint32_t num_entries, uint32_t *seq, bool block, 404 uint32_t num_entries, uint32_t *seq, bool block); 433 opdl_stage_disclaim(struct opdl_stage *s, uint32_t num_entries, 463 opdl_stage_disclaim_n(struct opdl_stage *s, uint32_t num_entries, 505 opdl_stage_find_num_available(struct opdl_stage *s, uint32_t num_entries);
|
| /f-stack/dpdk/drivers/net/bnxt/tf_core/ |
| H A D | tf_em_common.c | 77 uint32_t num_entries, in tf_create_tbl_pool_external() argument 86 parms.nitems = num_entries; in tf_create_tbl_pool_external() 269 uint32_t mask = num_entries - 1; in tf_em_get_key_mask() 377 uint32_t num_entries, in tf_em_size_page_tbl_lvl() argument 423 uint32_t num_entries; in tf_em_size_table() local 442 num_entries = tbl->num_entries / tbl->entry_size; in tf_em_size_table() 448 num_entries = cnt; in tf_em_size_table() 451 num_entries = tbl->num_entries; in tf_em_size_table() 456 tbl->num_entries, in tf_em_size_table() 524 while (num_entries > cnt && in tf_em_validate_num_entries() [all …]
|
| H A D | tf_em_internal.c | 56 uint32_t num_entries, in tf_create_em_pool() argument 65 parms.nitems = num_entries / TF_SESSION_EM_ENTRY_SIZE; in tf_create_em_pool() 81 rc = stack_init(num_entries / TF_SESSION_EM_ENTRY_SIZE, in tf_create_em_pool() 95 j = start + num_entries - TF_SESSION_EM_ENTRY_SIZE; in tf_create_em_pool() 97 for (i = 0; i < (num_entries / TF_SESSION_EM_ENTRY_SIZE); i++) { in tf_create_em_pool()
|
| H A D | tf_shadow_tbl.c | 129 uint16_t num_entries; member 198 return ctxt->shadow_ctxt.num_entries; in tf_shadow_tbl_sh_num_entries_get() 241 uint16_t num_entries, in tf_shadow_tbl_ctxt_create() argument 250 if (num_entries > TF_SHADOW_ENTRIES_MAX) { in tf_shadow_tbl_ctxt_create() 252 num_entries, in tf_shadow_tbl_ctxt_create() 257 while (hash_size < num_entries) in tf_shadow_tbl_ctxt_create() 275 cparms.nitems = num_entries; in tf_shadow_tbl_ctxt_create() 284 cparms.nitems = num_entries; in tf_shadow_tbl_ctxt_create() 292 ctxt->shadow_ctxt.num_entries = num_entries; in tf_shadow_tbl_ctxt_create()
|
| H A D | tf_em_common.h | 45 uint32_t num_entries, 66 uint32_t tf_em_get_key_mask(int num_entries);
|
| H A D | tf_shadow_tcam.c | 134 uint16_t num_entries; member 174 return ctxt->shadow_ctxt.num_entries; in tf_shadow_tcam_sh_num_entries_get() 250 uint16_t num_entries, in tf_shadow_tcam_ctxt_create() argument 259 if (num_entries > TF_SHADOW_TCAM_ENTRIES_MAX) { in tf_shadow_tcam_ctxt_create() 261 num_entries, in tf_shadow_tcam_ctxt_create() 266 while (hash_size < num_entries) in tf_shadow_tcam_ctxt_create() 284 cparms.nitems = num_entries; in tf_shadow_tcam_ctxt_create() 293 cparms.nitems = num_entries; in tf_shadow_tcam_ctxt_create() 301 ctxt->shadow_ctxt.num_entries = num_entries; in tf_shadow_tcam_ctxt_create()
|
| H A D | tf_shadow_identifier.c | 35 uint16_t num_entries; member 70 shadow_db->num_entries = parms->num_elements; in tf_shadow_ident_create_db() 115 for (i = 0; i < shadow_db->num_entries; i++) in tf_shadow_ident_free_db()
|
| H A D | stack.c | 18 stack_init(int num_entries, uint32_t *items, struct stack *st) in stack_init() argument 23 st->max = num_entries; in stack_init()
|
| H A D | tf_rm.c | 59 uint16_t num_entries; member 284 cparms.nitems = rm_db->num_entries; in tf_rm_check_residuals() 296 for (i = 0, found = 0; i < rm_db->num_entries; i++) { in tf_rm_check_residuals() 328 for (i = 0, f = 0; i < rm_db->num_entries; i++) { in tf_rm_check_residuals() 352 rm_db->num_entries, in tf_rm_check_residuals() 594 rm_db->num_entries = parms->num_elements; in tf_rm_create_db() 689 for (i = 0; i < rm_db->num_entries; i++) in tf_rm_free_db() 975 parms->starting_index + parms->num_entries > base_index + stride) in tf_rm_check_indexes_in_range()
|
| H A D | tf_msg.h | 331 uint32_t num_entries, 498 uint16_t num_entries,
|
| H A D | tf_em_host.c | 306 if (tbl->num_entries != 0 && tbl->entry_size != 0) { in tf_em_ctx_unreg() 342 if (tbl->num_entries && tbl->entry_size) { in tf_em_ctx_reg() 437 em_tables[TF_KEY0_TABLE].num_entries, in tf_em_ext_alloc() 470 em_tables[TF_RECORD_TABLE].num_entries, in tf_em_ext_alloc()
|
| H A D | tf_tbl.c | 107 shadow_cfg.num_entries = parms->num_elements; in tf_tbl_bind() 593 cparms.num_entries = parms->num_entries; in tf_tbl_bulk_get() 602 parms->num_entries, in tf_tbl_bulk_get() 625 parms->num_entries, in tf_tbl_bulk_get()
|
| H A D | stack.h | 35 int stack_init(int num_entries,
|
| H A D | tf_shadow_identifier.h | 37 int num_entries; member
|
| H A D | tf_shadow_tbl.h | 36 int num_entries; member
|
| H A D | tf_shadow_tcam.h | 21 int num_entries; member
|
| /f-stack/freebsd/contrib/octeon-sdk/ |
| H A D | cvmx-tim.h | 105 volatile uint32_t num_entries; /**< Zeroed by HW after traversing list */ member 245 work_bucket_ptr->num_entries++; in cvmx_tim_add_entry() 260 if (work_bucket_ptr->num_entries) in cvmx_tim_add_entry() 265 work_bucket_ptr->num_entries++; in cvmx_tim_add_entry() 271 work_bucket_ptr->num_entries = 1; in cvmx_tim_add_entry()
|
| H A D | cvmx-tim.c | 294 while (bucket_ptr->num_entries) in cvmx_tim_shutdown() 304 if (bucket_ptr->num_entries > entries_per_chunk) in cvmx_tim_shutdown() 305 bucket_ptr->num_entries -= entries_per_chunk; in cvmx_tim_shutdown() 307 bucket_ptr->num_entries = 0; in cvmx_tim_shutdown()
|
| /f-stack/dpdk/lib/librte_cfgfile/ |
| H A D | rte_cfgfile.c | 18 int num_entries; member 107 if (section->num_entries == section->allocated_entries) { in _add_entry() 122 §ion->entries[section->num_entries]; in _add_entry() 126 section->num_entries++; in _add_entry() 303 cfg->sections[i].num_entries = 0; in rte_cfgfile_create() 361 cfg->sections[cfg->num_sections].num_entries = 0; in rte_cfgfile_add_section() 408 for (i = 0; i < curr_section->num_entries; i++) in rte_cfgfile_set_entry() 434 for (j = 0; j < cfg->sections[i].num_entries; j++) { in rte_cfgfile_save() 504 return s->num_entries; in rte_cfgfile_section_num_entries() 517 return sect->num_entries; in rte_cfgfile_section_num_entries_by_index() [all …]
|
| /f-stack/dpdk/drivers/net/bnxt/tf_ulp/ |
| H A D | ulp_fc_mgr.c | 247 parms.num_entries = dparms->flow_count_db_entries / 2; /* direction */ in ulp_bulk_get_flow_stats() 271 for (i = 0; i < parms.num_entries; i++) { in ulp_bulk_get_flow_stats() 355 uint32_t dev_id, hw_cntr_id = 0, num_entries = 0; in ulp_fc_mgr_alarm_cb() local 385 if (!ulp_fc_info->num_entries) { in ulp_fc_mgr_alarm_cb() 404 num_entries = dparms->flow_count_db_entries / 2; in ulp_fc_mgr_alarm_cb() 406 for (j = 0; j < num_entries; j++) { in ulp_fc_mgr_alarm_cb() 510 ulp_fc_info->num_entries++; in ulp_fc_mgr_cntr_set() 544 ulp_fc_info->num_entries--; in ulp_fc_mgr_cntr_reset()
|
| H A D | ulp_template_db_tbl.c | 301 .num_entries = 16384 305 .num_entries = 16384 309 .num_entries = 16384 313 .num_entries = 16384
|
| H A D | ulp_template_struct.h | 113 uint32_t num_entries; member 258 uint16_t num_entries; member
|
| H A D | ulp_fc_mgr.h | 49 uint32_t num_entries; member
|
| /f-stack/dpdk/lib/librte_member/ |
| H A D | rte_member_ht.c | 81 uint32_t num_entries = rte_align32pow2(params->num_keys); in rte_member_create_ht() local 83 if ((num_entries > RTE_MEMBER_ENTRIES_MAX) || in rte_member_create_ht() 85 num_entries < RTE_MEMBER_BUCKET_ENTRIES) { in rte_member_create_ht() 92 uint32_t num_buckets = num_entries / RTE_MEMBER_BUCKET_ENTRIES; in rte_member_create_ht() 126 num_entries, num_buckets); in rte_member_create_ht()
|