| /dpdk/drivers/bus/fslmc/qbman/ |
| H A D | qbman_debug.c | 52 if (!r) { in qbman_bp_query() 64 r->rslt); in qbman_bp_query() 125 return r->vbpid; in qbman_bp_get_vbpid() 170 return r->fill; in qbman_bp_num_free_bufs() 213 if (!r) { in qbman_fq_query() 239 return r->cgid; in qbman_fq_attr_get_cgrid() 339 *r = *var; in qbman_fq_query_state() 409 if (!r) { in qbman_cgr_query() 484 if (!r) { in qbman_cgr_wred_query() 576 if (!r) { in qbman_cgr_statistics_query() [all …]
|
| /dpdk/lib/ring/ |
| H A D | rte_ring.h | 495 rte_ring_reset(struct rte_ring *r); 508 uint32_t prod_tail = r->prod.tail; in rte_ring_count() 509 uint32_t cons_tail = r->cons.tail; in rte_ring_count() 511 return (count > r->capacity) ? r->capacity : count; in rte_ring_count() 525 return r->capacity - rte_ring_count(r); in rte_ring_free_count() 555 uint32_t prod_tail = r->prod.tail; in rte_ring_empty() 556 uint32_t cons_tail = r->cons.tail; in rte_ring_empty() 573 return r->size; in rte_ring_get_size() 587 return r->capacity; in rte_ring_get_capacity() 601 return r->prod.sync_type; in rte_ring_get_prod_sync_type() [all …]
|
| H A D | rte_ring_elem.h | 135 return __rte_ring_do_enqueue_elem(r, obj_table, esize, n, in rte_ring_mp_enqueue_bulk_elem() 164 return __rte_ring_do_enqueue_elem(r, obj_table, esize, n, in rte_ring_sp_enqueue_bulk_elem() 198 switch (r->prod.sync_type) { in rte_ring_enqueue_bulk_elem() 321 return __rte_ring_do_dequeue_elem(r, obj_table, esize, n, in rte_ring_mc_dequeue_bulk_elem() 349 return __rte_ring_do_dequeue_elem(r, obj_table, esize, n, in rte_ring_sc_dequeue_bulk_elem() 380 switch (r->cons.sync_type) { in rte_ring_dequeue_bulk_elem() 422 rte_ring_mc_dequeue_elem(struct rte_ring *r, void *obj_p, in rte_ring_mc_dequeue_elem() argument 446 rte_ring_sc_dequeue_elem(struct rte_ring *r, void *obj_p, in rte_ring_sc_dequeue_elem() argument 506 return __rte_ring_do_enqueue_elem(r, obj_table, esize, n, in rte_ring_mp_enqueue_burst_elem() 566 switch (r->prod.sync_type) { in rte_ring_enqueue_burst_elem() [all …]
|
| H A D | rte_ring_peek.h | 172 switch (r->prod.sync_type) { in rte_ring_enqueue_elem_finish() 174 n = __rte_ring_st_get_tail(&r->prod, &tail, n); in rte_ring_enqueue_elem_finish() 177 __rte_ring_st_set_head_tail(&r->prod, tail, n, 1); in rte_ring_enqueue_elem_finish() 180 n = __rte_ring_hts_get_tail(&r->hts_prod, &tail, n); in rte_ring_enqueue_elem_finish() 183 __rte_ring_hts_set_head_tail(&r->hts_prod, tail, n, 1); in rte_ring_enqueue_elem_finish() 316 return rte_ring_dequeue_burst_elem_start(r, obj_table, in rte_ring_dequeue_burst_start() 335 switch (r->cons.sync_type) { in rte_ring_dequeue_elem_finish() 337 n = __rte_ring_st_get_tail(&r->cons, &tail, n); in rte_ring_dequeue_elem_finish() 338 __rte_ring_st_set_head_tail(&r->cons, tail, n, 0); in rte_ring_dequeue_elem_finish() 341 n = __rte_ring_hts_get_tail(&r->hts_cons, &tail, n); in rte_ring_dequeue_elem_finish() [all …]
|
| H A D | rte_ring_peek_zc.h | 102 uint32_t *ring = (uint32_t *)&r[1]; in __rte_ring_get_elem_addr() 106 idx = head & r->mask; in __rte_ring_get_elem_addr() 112 if (idx + num > r->size) { in __rte_ring_get_elem_addr() 113 *n1 = r->size - idx; in __rte_ring_get_elem_addr() 130 switch (r->prod.sync_type) { in __rte_ring_do_enqueue_zc_elem_start() 293 switch (r->prod.sync_type) { in rte_ring_enqueue_zc_elem_finish() 295 n = __rte_ring_st_get_tail(&r->prod, &tail, n); in rte_ring_enqueue_zc_elem_finish() 323 rte_ring_enqueue_zc_elem_finish(r, n); in rte_ring_enqueue_zc_finish() 337 switch (r->cons.sync_type) { in __rte_ring_do_dequeue_zc_elem_start() 499 switch (r->cons.sync_type) { in rte_ring_dequeue_zc_elem_finish() [all …]
|
| H A D | rte_ring_rts.h | 189 return rte_ring_mp_rts_enqueue_bulk_elem(r, obj_table, in rte_ring_mp_rts_enqueue_bulk() 274 rte_ring_get_prod_htd_max(const struct rte_ring *r) in rte_ring_get_prod_htd_max() argument 276 if (r->prod.sync_type == RTE_RING_SYNC_MT_RTS) in rte_ring_get_prod_htd_max() 277 return r->rts_prod.htd_max; in rte_ring_get_prod_htd_max() 295 if (r->prod.sync_type != RTE_RING_SYNC_MT_RTS) in rte_ring_set_prod_htd_max() 298 r->rts_prod.htd_max = v; in rte_ring_set_prod_htd_max() 312 rte_ring_get_cons_htd_max(const struct rte_ring *r) in rte_ring_get_cons_htd_max() argument 314 if (r->cons.sync_type == RTE_RING_SYNC_MT_RTS) in rte_ring_get_cons_htd_max() 315 return r->rts_cons.htd_max; in rte_ring_get_cons_htd_max() 333 if (r->cons.sync_type != RTE_RING_SYNC_MT_RTS) in rte_ring_set_cons_htd_max() [all …]
|
| H A D | rte_ring.c | 204 memset(r, 0, sizeof(*r)); in rte_ring_init() 205 ret = strlcpy(r->name, name, sizeof(r->name)); in rte_ring_init() 215 r->mask = r->size - 1; in rte_ring_init() 226 r->capacity = r->mask; in rte_ring_init() 231 rte_ring_set_prod_htd_max(r, r->capacity / HTD_MAX_DEF); in rte_ring_init() 233 rte_ring_set_cons_htd_max(r, r->capacity / HTD_MAX_DEF); in rte_ring_init() 287 r = mz->addr; in rte_ring_create_elem() 297 r = NULL; in rte_ring_create_elem() 303 return r; in rte_ring_create_elem() 365 fprintf(f, "ring <%s>@%p\n", r->name, r); in rte_ring_dump() [all …]
|
| H A D | rte_ring_hts.h | 56 return __rte_ring_do_hts_enqueue_elem(r, obj_table, esize, n, in rte_ring_mp_hts_enqueue_bulk_elem() 80 rte_ring_mc_hts_dequeue_bulk_elem(struct rte_ring *r, void *obj_table, in rte_ring_mc_hts_dequeue_bulk_elem() argument 83 return __rte_ring_do_hts_dequeue_elem(r, obj_table, esize, n, in rte_ring_mc_hts_dequeue_bulk_elem() 110 return __rte_ring_do_hts_enqueue_elem(r, obj_table, esize, n, in rte_ring_mp_hts_enqueue_burst_elem() 139 return __rte_ring_do_hts_dequeue_elem(r, obj_table, esize, n, in rte_ring_mc_hts_dequeue_burst_elem() 162 return rte_ring_mp_hts_enqueue_bulk_elem(r, obj_table, in rte_ring_mp_hts_enqueue_bulk() 182 rte_ring_mc_hts_dequeue_bulk(struct rte_ring *r, void **obj_table, in rte_ring_mc_hts_dequeue_bulk() argument 185 return rte_ring_mc_hts_dequeue_bulk_elem(r, obj_table, in rte_ring_mc_hts_dequeue_bulk() 208 return rte_ring_mp_hts_enqueue_burst_elem(r, obj_table, in rte_ring_mp_hts_enqueue_burst() 230 rte_ring_mc_hts_dequeue_burst(struct rte_ring *r, void **obj_table, in rte_ring_mc_hts_dequeue_burst() argument [all …]
|
| H A D | rte_ring_elem_pvt.h | 18 uint32_t *ring = (uint32_t *)&r[1]; in __rte_ring_enqueue_elems_32() 61 const uint32_t size = r->size; in __rte_ring_enqueue_elems_64() 62 uint32_t idx = prod_head & r->mask; in __rte_ring_enqueue_elems_64() 63 uint64_t *ring = (uint64_t *)&r[1]; in __rte_ring_enqueue_elems_64() 94 const uint32_t size = r->size; in __rte_ring_enqueue_elems_128() 139 idx = prod_head & r->mask; in __rte_ring_enqueue_elems() 141 nr_size = r->size * scale; in __rte_ring_enqueue_elems() 195 const uint32_t size = r->size; in __rte_ring_dequeue_elems_64() 228 const uint32_t size = r->size; in __rte_ring_dequeue_elems_128() 269 idx = cons_head & r->mask; in __rte_ring_dequeue_elems() [all …]
|
| H A D | rte_ring_rts_elem_pvt.h | 77 const uint32_t capacity = r->capacity; in __rte_ring_rts_move_prod_head() 90 __rte_ring_rts_head_wait(&r->rts_prod, &oh); in __rte_ring_rts_move_prod_head() 98 *free_entries = capacity + r->cons.tail - oh.val.pos; in __rte_ring_rts_move_prod_head() 116 } while (__atomic_compare_exchange_n(&r->rts_prod.head.raw, in __rte_ring_rts_move_prod_head() 147 __rte_ring_rts_head_wait(&r->rts_cons, &oh); in __rte_ring_rts_move_cons_head() 154 *entries = r->prod.tail - oh.val.pos; in __rte_ring_rts_move_cons_head() 171 } while (__atomic_compare_exchange_n(&r->rts_cons.head.raw, in __rte_ring_rts_move_cons_head() 211 __rte_ring_enqueue_elems(r, head, obj_table, esize, n); in __rte_ring_do_rts_enqueue_elem() 212 __rte_ring_rts_update_tail(&r->rts_prod); in __rte_ring_do_rts_enqueue_elem() 252 __rte_ring_dequeue_elems(r, head, obj_table, esize, n); in __rte_ring_do_rts_dequeue_elem() [all …]
|
| H A D | rte_ring_hts_elem_pvt.h | 62 const uint32_t capacity = r->capacity; in __rte_ring_hts_move_prod_head() 75 __rte_ring_hts_head_wait(&r->hts_prod, &op); in __rte_ring_hts_move_prod_head() 83 *free_entries = capacity + r->cons.tail - op.pos.head; in __rte_ring_hts_move_prod_head() 101 } while (__atomic_compare_exchange_n(&r->hts_prod.ht.raw, in __rte_ring_hts_move_prod_head() 132 __rte_ring_hts_head_wait(&r->hts_cons, &op); in __rte_ring_hts_move_cons_head() 139 *entries = r->prod.tail - op.pos.head; in __rte_ring_hts_move_cons_head() 156 } while (__atomic_compare_exchange_n(&r->hts_cons.ht.raw, in __rte_ring_hts_move_cons_head() 196 __rte_ring_enqueue_elems(r, head, obj_table, esize, n); in __rte_ring_do_hts_enqueue_elem() 197 __rte_ring_hts_update_tail(&r->hts_prod, head, n, 1); in __rte_ring_do_hts_enqueue_elem() 237 __rte_ring_dequeue_elems(r, head, obj_table, esize, n); in __rte_ring_do_hts_dequeue_elem() [all …]
|
| H A D | rte_ring_generic_pvt.h | 55 __rte_ring_move_prod_head(struct rte_ring *r, unsigned int is_sp, in __rte_ring_move_prod_head() argument 60 const uint32_t capacity = r->capacity; in __rte_ring_move_prod_head() 68 *old_head = r->prod.head; in __rte_ring_move_prod_head() 81 *free_entries = (capacity + r->cons.tail - *old_head); in __rte_ring_move_prod_head() 93 r->prod.head = *new_head, success = 1; in __rte_ring_move_prod_head() 95 success = rte_atomic32_cmpset(&r->prod.head, in __rte_ring_move_prod_head() 125 __rte_ring_move_cons_head(struct rte_ring *r, unsigned int is_sc, in __rte_ring_move_cons_head() argument 138 *old_head = r->cons.head; in __rte_ring_move_cons_head() 150 *entries = (r->prod.tail - *old_head); in __rte_ring_move_cons_head() 161 r->cons.head = *new_head; in __rte_ring_move_cons_head() [all …]
|
| H A D | rte_ring_c11_pvt.h | 54 __rte_ring_move_prod_head(struct rte_ring *r, unsigned int is_sp, in __rte_ring_move_prod_head() argument 59 const uint32_t capacity = r->capacity; in __rte_ring_move_prod_head() 64 *old_head = __atomic_load_n(&r->prod.head, __ATOMIC_RELAXED); in __rte_ring_move_prod_head() 75 cons_tail = __atomic_load_n(&r->cons.tail, in __rte_ring_move_prod_head() 95 r->prod.head = *new_head, success = 1; in __rte_ring_move_prod_head() 98 success = __atomic_compare_exchange_n(&r->prod.head, in __rte_ring_move_prod_head() 130 __rte_ring_move_cons_head(struct rte_ring *r, int is_sc, in __rte_ring_move_cons_head() argument 140 *old_head = __atomic_load_n(&r->cons.head, __ATOMIC_RELAXED); in __rte_ring_move_cons_head() 151 prod_tail = __atomic_load_n(&r->prod.tail, in __rte_ring_move_cons_head() 170 r->cons.head = *new_head, success = 1; in __rte_ring_move_cons_head() [all …]
|
| /dpdk/drivers/bus/fslmc/qbman/include/ |
| H A D | fsl_qbman_debug.h | 40 struct qbman_bp_query_rslt *r); 41 int qbman_bp_get_bdi(struct qbman_bp_query_rslt *r); 42 int qbman_bp_get_va(struct qbman_bp_query_rslt *r); 43 int qbman_bp_get_wae(struct qbman_bp_query_rslt *r); 52 int qbman_bp_get_pl(struct qbman_bp_query_rslt *r); 59 int qbman_bp_is_surplus(struct qbman_bp_query_rslt *r); 89 struct qbman_fq_query_rslt *r); 140 struct qbman_fq_query_np_rslt *r); 177 struct qbman_cgr_query_rslt *r); 200 struct qbman_wred_query_rslt *r); [all …]
|
| /dpdk/app/test/ |
| H A D | resource.c | 18 return r->end - r->begin; in resource_size() 23 struct resource *r; in resource_find() local 26 RTE_VERIFY(r->name); in resource_find() 29 return r; in resource_find() 105 struct archive *r; in resource_untar() local 119 if (r == NULL) { in resource_untar() 159 ret = do_copy(r, w); in resource_untar() 169 archive_read_free(r); in resource_untar() 175 archive_read_free(r); in resource_untar() 183 struct archive *r; in resource_rm_by_tar() local [all …]
|
| H A D | test_resource.c | 20 const struct resource *r; in test_resource_dpdk() local 22 r = resource_find("test_resource_dpdk"); in test_resource_dpdk() 26 r->name); in test_resource_dpdk() 38 const struct resource *r; in test_resource_c() local 41 r = resource_find("test_resource_c"); in test_resource_c() 45 r->name); in test_resource_c() 64 const struct resource *r; in test_resource_tar() local 67 r = resource_find("test_resource_tar"); in test_resource_tar() 71 r->name); in test_resource_tar() 73 TEST_ASSERT_SUCCESS(resource_untar(r), in test_resource_tar() [all …]
|
| H A D | test_ring.h | 108 return rte_ring_enqueue(r, *obj); in test_ring_enqueue() 110 return rte_ring_sp_enqueue(r, *obj); in test_ring_enqueue() 112 return rte_ring_mp_enqueue(r, *obj); in test_ring_enqueue() 114 return rte_ring_enqueue_bulk(r, obj, n, NULL); in test_ring_enqueue() 120 return rte_ring_enqueue_burst(r, obj, n, NULL); in test_ring_enqueue() 132 return rte_ring_enqueue_elem(r, obj, esize); in test_ring_enqueue() 169 return rte_ring_dequeue(r, obj); in test_ring_dequeue() 171 return rte_ring_sc_dequeue(r, obj); in test_ring_dequeue() 173 return rte_ring_mc_dequeue(r, obj); in test_ring_dequeue() 175 return rte_ring_dequeue_bulk(r, obj, n, NULL); in test_ring_dequeue() [all …]
|
| H A D | test_ring.c | 640 TEST_RING_VERIFY(rte_ring_full(r), r, goto fail); in test_ring_burst_bulk_tests1() 641 TEST_RING_VERIFY(rte_ring_empty(r) == 0, r, goto fail); in test_ring_burst_bulk_tests1() 649 TEST_RING_VERIFY(rte_ring_count(r) == 0, r, goto fail); in test_ring_burst_bulk_tests1() 650 TEST_RING_VERIFY(rte_ring_full(r) == 0, r, goto fail); in test_ring_burst_bulk_tests1() 651 TEST_RING_VERIFY(rte_ring_empty(r), r, goto fail); in test_ring_burst_bulk_tests1() 665 r = NULL; in test_ring_burst_bulk_tests1() 753 r = NULL; in test_ring_burst_bulk_tests2() 824 r = NULL; in test_ring_burst_bulk_tests3() 901 TEST_RING_VERIFY(rte_ring_full(r) == 1, r, goto fail); in test_ring_burst_bulk_tests4() 934 TEST_RING_VERIFY(rte_ring_empty(r) == 1, r, goto fail); in test_ring_burst_bulk_tests4() [all …]
|
| H A D | test_ring_perf.c | 166 struct rte_ring *r; member 183 struct rte_ring *r = p->r; in enqueue_dequeue_bulk_helper() local 299 param1.r = param2.r = r; in run_on_core_pair() 397 param.r = r; in run_on_all_cores() 498 struct rte_ring *r = NULL; in test_ring_perf_esize() local 505 if (r == NULL) in test_ring_perf_esize() 533 test_empty_dequeue(r, esize, in test_ring_perf_esize() 535 test_empty_dequeue(r, esize, in test_ring_perf_esize() 556 if (run_on_all_cores(r, esize) < 0) in test_ring_perf_esize() 559 rte_ring_free(r); in test_ring_perf_esize() [all …]
|
| H A D | test_ring_mt_peek_stress.c | 9 _st_ring_dequeue_bulk(struct rte_ring *r, void **obj, uint32_t n, in _st_ring_dequeue_bulk() argument 14 m = rte_ring_dequeue_bulk_start(r, obj, n, avail); in _st_ring_dequeue_bulk() 16 rte_ring_dequeue_finish(r, n); in _st_ring_dequeue_bulk() 21 _st_ring_enqueue_bulk(struct rte_ring *r, void * const *obj, uint32_t n, in _st_ring_enqueue_bulk() argument 26 m = rte_ring_enqueue_bulk_start(r, n, free); in _st_ring_enqueue_bulk() 28 rte_ring_enqueue_finish(r, obj, n); in _st_ring_enqueue_bulk() 33 _st_ring_init(struct rte_ring *r, const char *name, uint32_t num) in _st_ring_init() argument 35 return rte_ring_init(r, name, num, in _st_ring_init()
|
| /dpdk/lib/eventdev/ |
| H A D | rte_event_ring.h | 38 struct rte_ring r; member 50 rte_event_ring_count(const struct rte_event_ring *r) in rte_event_ring_count() argument 52 return rte_ring_count(&r->r); in rte_event_ring_count() 67 return rte_ring_free_count(&r->r); in rte_event_ring_free_count() 91 rte_event_ring_enqueue_burst(struct rte_event_ring *r, in rte_event_ring_enqueue_burst() argument 98 num = rte_ring_enqueue_burst_elem(&r->r, events, in rte_event_ring_enqueue_burst() 127 rte_event_ring_dequeue_burst(struct rte_event_ring *r, in rte_event_ring_dequeue_burst() argument 134 num = rte_ring_dequeue_burst_elem(&r->r, events, in rte_event_ring_dequeue_burst() 240 rte_event_ring_free(struct rte_event_ring *r); 255 return rte_ring_get_size(&r->r); in rte_event_ring_get_size() [all …]
|
| /dpdk/drivers/event/sw/ |
| H A D | event_ring.h | 54 rte_free(r); in rob_ring_free() 60 return r->write_idx - r->read_idx; in rob_ring_count() 66 return r->size - rob_ring_count(r); in rob_ring_free_count() 72 const uint32_t size = r->size; in rob_ring_enqueue() 73 const uint32_t mask = r->mask; in rob_ring_enqueue() 75 uint32_t write = r->write_idx; in rob_ring_enqueue() 79 r->ring[write & mask] = re; in rob_ring_enqueue() 80 r->write_idx++; in rob_ring_enqueue() 88 uint32_t read = r->read_idx; in rob_ring_dequeue() 93 *re = r->ring[read & mask]; in rob_ring_dequeue() [all …]
|
| /dpdk/drivers/net/bnxt/tf_core/ |
| H A D | bitalloc.c | 123 int r; in ba_alloc_helper() local 148 return r; in ba_alloc_helper() 196 int r; in ba_alloc_reverse_helper() local 221 return r; in ba_alloc_reverse_helper() 270 return r; in ba_alloc_index_helper() 311 if (r == 1) in ba_inuse_helper() 314 return r; in ba_inuse_helper() 358 return r; in ba_free_helper() 422 r = -1; in ba_find_next_helper() 425 r = -1; in ba_find_next_helper() [all …]
|
| /dpdk/devtools/ |
| H A D | update_version_map_abi.py | 20 func_line_regex = re.compile(r"\s*" 23 r"\s*" 24 r";" 25 r"\s*" 27 r")" 28 r"\s*" 29 r"$") 35 r"{" 42 r"}" 46 r";" [all …]
|
| /dpdk/lib/eal/linux/ |
| H A D | eal_vfio_mp_sync.c | 43 r->req = SOCKET_REQ_GROUP; in vfio_mp_primary() 44 r->group_num = m->group_num; in vfio_mp_primary() 47 r->result = SOCKET_ERR; in vfio_mp_primary() 50 r->result = SOCKET_NO_FD; in vfio_mp_primary() 53 r->result = SOCKET_OK; in vfio_mp_primary() 62 r->result = SOCKET_ERR; in vfio_mp_primary() 64 r->result = SOCKET_OK; in vfio_mp_primary() 73 r->result = SOCKET_ERR; in vfio_mp_primary() 75 r->result = SOCKET_OK; in vfio_mp_primary() 89 r->result = SOCKET_ERR; in vfio_mp_primary() [all …]
|