| /f-stack/dpdk/app/test/ |
| H A D | test_reorder.c | 168 bufs[i] = NULL; in test_reorder_insert() 183 bufs[4] = NULL; in test_reorder_insert() 194 bufs[5] = NULL; in test_reorder_insert() 205 bufs[6] = NULL; in test_reorder_insert() 211 if (bufs[i] != NULL) in test_reorder_insert() 262 bufs[1] = NULL; in test_reorder_drain() 280 bufs[2] = NULL; in test_reorder_drain() 281 bufs[3] = NULL; in test_reorder_drain() 288 bufs[4] = NULL; in test_reorder_drain() 295 bufs[7] = NULL; in test_reorder_drain() [all …]
|
| H A D | test_distributor.c | 112 struct rte_mbuf *bufs[BURST]; in sanity_test() local 128 bufs[i]->hash.usr = 0; in sanity_test() 190 bufs[i]->hash.usr = i+1; in sanity_test() 326 struct rte_mbuf *bufs[BURST]; in sanity_test_with_mbuf_alloc() local 337 bufs[j]->hash.usr = (i+j) << 1; in sanity_test_with_mbuf_alloc() 433 struct rte_mbuf *bufs[BURST]; in sanity_test_with_worker_shutdown() local 453 bufs[i]->hash.usr = 1; in sanity_test_with_worker_shutdown() 517 struct rte_mbuf *bufs[BURST]; in test_flush_with_worker_shutdown() local 533 bufs[i]->hash.usr = 0; in test_flush_with_worker_shutdown() 654 &bufs[i * burst + processed], in sanity_mark_test() [all …]
|
| H A D | test_distributor_perf.c | 142 struct rte_mbuf *bufs[BURST]; in perf_test() local 145 if (rte_mempool_get_bulk(p, (void *)bufs, BURST) != 0) { in perf_test() 151 bufs[i]->hash.usr = i; in perf_test() 155 rte_distributor_process(d, bufs, BURST); in perf_test() 168 rte_mempool_put_bulk(p, (void *)bufs, BURST); in perf_test() 186 struct rte_mbuf *bufs[RTE_MAX_LCORE]; in quit_workers() local 188 rte_mempool_get_bulk(p, (void *)bufs, num_workers); in quit_workers() 192 bufs[i]->hash.usr = i << 1; in quit_workers() 193 rte_distributor_process(d, bufs, num_workers); in quit_workers() 195 rte_mempool_put_bulk(p, (void *)bufs, num_workers); in quit_workers()
|
| H A D | virtual_pmd.c | 325 struct rte_mbuf **bufs, in virtual_ethdev_rx_burst_success() argument 338 rx_count = rte_ring_dequeue_burst(dev_private->rx_queue, (void **) bufs, in virtual_ethdev_rx_burst_success() 346 dev_private->eth_stats.ibytes += rte_pktmbuf_pkt_len(bufs[i]); in virtual_ethdev_rx_burst_success() 353 struct rte_mbuf **bufs __rte_unused, in virtual_ethdev_rx_burst_fail() 360 virtual_ethdev_tx_burst_success(void *queue, struct rte_mbuf **bufs, in virtual_ethdev_tx_burst_success() argument 376 nb_pkts = rte_ring_enqueue_burst(dev_private->tx_queue, (void **)bufs, in virtual_ethdev_tx_burst_success() 384 dev_private->eth_stats.obytes += rte_pktmbuf_pkt_len(bufs[i]); in virtual_ethdev_tx_burst_success() 390 virtual_ethdev_tx_burst_fail(void *queue, struct rte_mbuf **bufs, in virtual_ethdev_tx_burst_fail() argument 412 if (bufs[i] != NULL) in virtual_ethdev_tx_burst_fail() 413 rte_pktmbuf_free(bufs[i]); in virtual_ethdev_tx_burst_fail() [all …]
|
| /f-stack/dpdk/drivers/net/dpaa2/ |
| H A D | dpaa2_rxtx.c | 1077 if (*dpaa2_seqn(*bufs)) { in dpaa2_dev_tx() 1088 mp = (*bufs)->pool; in dpaa2_dev_tx() 1106 bufs++; in dpaa2_dev_tx() 1154 eth_mbuf_to_fd(*bufs, in dpaa2_dev_tx() 1161 bufs++; in dpaa2_dev_tx() 1345 (*bufs), in dpaa2_dev_tx_ordered() 1355 mp = (*bufs)->pool; in dpaa2_dev_tx_ordered() 1372 bufs++; in dpaa2_dev_tx_ordered() 1411 eth_mbuf_to_fd(*bufs, in dpaa2_dev_tx_ordered() 1415 bufs++; in dpaa2_dev_tx_ordered() [all …]
|
| H A D | dpaa2_ethdev.h | 200 uint16_t dpaa2_dev_rx(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts); 202 uint16_t dpaa2_dev_loopback_rx(void *queue, struct rte_mbuf **bufs, 205 uint16_t dpaa2_dev_prefetch_rx(void *queue, struct rte_mbuf **bufs, 222 uint16_t dpaa2_dev_tx(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts); 223 uint16_t dpaa2_dev_tx_ordered(void *queue, struct rte_mbuf **bufs, 225 uint16_t dummy_dev_tx(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts);
|
| /f-stack/dpdk/drivers/mempool/dpaa2/ |
| H A D | dpaa2_hw_mempool.c | 195 uint64_t bufs[DPAA2_MBUF_MAX_ACQ_REL]; in rte_dpaa2_mbuf_release() local 221 bufs[i] = (uint64_t)rte_mempool_virt2iova(obj_table[i]) in rte_dpaa2_mbuf_release() 224 bufs[i] = (uint64_t)obj_table[i] + meta_data_size; in rte_dpaa2_mbuf_release() 245 bufs[i] = (uint64_t) in rte_dpaa2_mbuf_release() 304 size_t bufs[DPAA2_MBUF_MAX_ACQ_REL]; in rte_dpaa2_mbuf_alloc_bulk() local 334 ret = qbman_swp_acquire(swp, bpid, (void *)bufs, in rte_dpaa2_mbuf_alloc_bulk() 337 ret = qbman_swp_acquire(swp, bpid, (void *)bufs, in rte_dpaa2_mbuf_alloc_bulk() 353 for (i = 0; (i < ret) && bufs[i]; i++) { in rte_dpaa2_mbuf_alloc_bulk() 354 DPAA2_MODIFY_IOVA_TO_VADDR(bufs[i], size_t); in rte_dpaa2_mbuf_alloc_bulk() 356 (bufs[i] - bp_info->meta_data_size); in rte_dpaa2_mbuf_alloc_bulk() [all …]
|
| /f-stack/dpdk/app/test-eventdev/ |
| H A D | test_perf_common.h | 88 void *bufs[16] __rte_cache_aligned;\ 89 int const sz = RTE_DIM(bufs);\ 97 void *bufs[], int const buf_sz, uint8_t count) in perf_process_last_stage() argument 99 bufs[count++] = ev->event_ptr; in perf_process_last_stage() 105 rte_mempool_put_bulk(pool, bufs, buf_sz); in perf_process_last_stage() 113 void *bufs[], int const buf_sz, uint8_t count) in perf_process_last_stage_latency() argument 118 bufs[count++] = ev->event_ptr; in perf_process_last_stage_latency() 124 rte_mempool_put_bulk(pool, bufs, buf_sz); in perf_process_last_stage_latency()
|
| /f-stack/dpdk/drivers/net/null/ |
| H A D | rte_eth_null.c | 89 if ((q == NULL) || (bufs == NULL)) in eth_null_rx() 97 bufs[i]->data_len = (uint16_t)packet_size; in eth_null_rx() 98 bufs[i]->pkt_len = packet_size; in eth_null_rx() 99 bufs[i]->port = h->internals->port_id; in eth_null_rx() 114 if ((q == NULL) || (bufs == NULL)) in eth_null_copy_rx() 125 bufs[i]->pkt_len = packet_size; in eth_null_copy_rx() 126 bufs[i]->port = h->internals->port_id; in eth_null_copy_rx() 147 if ((q == NULL) || (bufs == NULL)) in eth_null_tx() 151 rte_pktmbuf_free(bufs[i]); in eth_null_tx() 165 if ((q == NULL) || (bufs == NULL)) in eth_null_copy_tx() [all …]
|
| /f-stack/dpdk/lib/librte_distributor/ |
| H A D | rte_distributor_single.c | 33 union rte_distributor_buffer_single *buf = &d->bufs[worker_id]; in EAL_REGISTER_TAILQ() 48 union rte_distributor_buffer_single *buf = &d->bufs[worker_id]; in rte_distributor_poll_pkt_single() 74 union rte_distributor_buffer_single *buf = &d->bufs[worker_id]; in rte_distributor_return_pkt_single() 179 __atomic_store_n(&(d->bufs[wkr].bufptr64), in process_returns() 184 __atomic_store_n(&(d->bufs[wkr].bufptr64), in process_returns() 224 int64_t data = __atomic_load_n(&(d->bufs[wkr].bufptr64), in rte_distributor_process_single() 272 __atomic_store_n(&(d->bufs[wkr].bufptr64), in rte_distributor_process_single() 278 __atomic_store_n(&(d->bufs[wkr].bufptr64), in rte_distributor_process_single() 302 (__atomic_load_n(&(d->bufs[wkr].bufptr64), in rte_distributor_process_single() 305 int64_t oldbuf = d->bufs[wkr].bufptr64 >> in rte_distributor_process_single() [all …]
|
| H A D | rte_distributor.c | 39 struct rte_distributor_buffer *buf = &(d->bufs[worker_id]); in EAL_REGISTER_TAILQ() 90 struct rte_distributor_buffer *buf = &d->bufs[worker_id]; in rte_distributor_poll_pkt() 291 struct rte_distributor_buffer *buf = &(d->bufs[wkr]); in handle_worker_shutdown() 344 struct rte_distributor_buffer *buf = &(d->bufs[wkr]); in handle_returns() 400 struct rte_distributor_buffer *buf = &(d->bufs[wkr]); in release() 419 d->bufs[wkr].bufptr64[i] = d->backlog[wkr].pkts[i] | in release() 467 if (__atomic_load_n(&(d->bufs[wid].bufptr64[0]), in rte_distributor_process() 469 d->bufs[wid].count = 0; in rte_distributor_process() 602 if ((__atomic_load_n(&(d->bufs[wid].bufptr64[0]), in rte_distributor_process() 604 d->bufs[wid].count = 0; in rte_distributor_process() [all …]
|
| /f-stack/dpdk/examples/distributor/ |
| H A D | main.c | 230 struct rte_mbuf *bufs[BURST_SIZE*2]; in lcore_rx() local 270 bufs, BURST_SIZE*2); in lcore_rx() 281 (void *)bufs, nb_ret, NULL); in lcore_rx() 292 (void *)bufs, nb_ret, NULL); in lcore_rx() 301 rte_pktmbuf_free(bufs[sent++]); in lcore_rx() 354 struct rte_mbuf *bufs[BURST_SIZE * 4]; in lcore_distributor() local 360 (void *)bufs, BURST_SIZE*1, NULL); in lcore_distributor() 369 bufs, BURST_SIZE*2); in lcore_distributor() 376 (void *)bufs, nb_ret, NULL); in lcore_distributor() 384 rte_pktmbuf_free(bufs[sent++]); in lcore_distributor() [all …]
|
| /f-stack/dpdk/drivers/net/dpaa/ |
| H A D | dpaa_rxtx.h | 272 uint16_t dpaa_eth_queue_rx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs); 274 uint16_t dpaa_eth_queue_tx_slow(void *q, struct rte_mbuf **bufs, 276 uint16_t dpaa_eth_queue_tx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs); 279 struct rte_mbuf **bufs __rte_unused, 290 struct qm_dqrr_entry **dqrr, void **bufs, int num_bufs); 292 void dpaa_rx_cb_prepare(struct qm_dqrr_entry *dq, void **bufs); 295 struct qm_dqrr_entry **dqrr, void **bufs, int num_bufs);
|
| H A D | dpaa_rxtx.c | 503 void **bufs, int num_bufs) in dpaa_rx_cb_no_prefetch() argument 542 mbuf = bufs[i]; in dpaa_rx_cb_no_prefetch() 559 void **bufs, int num_bufs) in dpaa_rx_cb() argument 582 mbuf = bufs[i]; in dpaa_rx_cb() 614 struct rte_mbuf **bufs, in dpaa_eth_queue_portal_rx() argument 636 void **bufs) in dpaa_rx_cb_parallel() argument 653 *bufs = mbuf; in dpaa_rx_cb_parallel() 663 void **bufs) in dpaa_rx_cb_atomic() argument 687 *bufs = mbuf; in dpaa_rx_cb_atomic() 737 struct rte_mbuf **bufs, in dpaa_eth_queue_rx() argument [all …]
|
| /f-stack/freebsd/contrib/ncsw/Peripherals/BM/ |
| H A D | bm_pool.c | 80 struct bm_buffer *bufs, in BmPoolRelease() argument 111 bufs, in BmPoolRelease() 134 struct bm_buffer *bufs, uint8_t num, uint32_t flags) in BmPoolAcquire() argument 142 return BmPortalAcquire(h_BmPortal, p_BmPool->bpid, bufs, num); in BmPoolAcquire() 156 memcpy(bufs, in BmPoolAcquire() 405 struct bm_buffer bufs[1]; in BM_POOL_GetBuf() local 432 physAddr = (uint64_t)bufs[0].lo; in BM_POOL_GetBuf() 445 struct bm_buffer bufs[1]; in BM_POOL_PutBuf() local 453 bufs[0].bpid = p_BmPool->bpid; in BM_POOL_PutBuf() 454 bufs[0].hi = (uint8_t)((physAddress & 0x000000ff00000000LL) >> 32); in BM_POOL_PutBuf() [all …]
|
| H A D | bm_portal.c | 370 struct bm_buffer *bufs, in BmPortalRelease() argument 392 r->bufs[i].hi = bufs[i].hi; in BmPortalRelease() 393 r->bufs[i].lo = bufs[i].lo; in BmPortalRelease() 406 struct bm_buffer *bufs, in BmPortalAcquire() argument 425 bufs[num].bpid = bpid; in BmPortalAcquire() 426 bufs[num].hi = mcr->acquire.bufs[num].hi; in BmPortalAcquire() 427 bufs[num].lo = mcr->acquire.bufs[num].lo; in BmPortalAcquire()
|
| /f-stack/dpdk/drivers/mempool/dpaa/ |
| H A D | dpaa_mempool.c | 46 struct bm_buffer bufs[8]; in dpaa_mbuf_create_pool() local 78 ret = bman_acquire(bp, bufs, 8, 0); in dpaa_mbuf_create_pool() 80 ret = bman_acquire(bp, bufs, 1, 0); in dpaa_mbuf_create_pool() 210 struct bm_buffer bufs[DPAA_MBUF_MAX_ACQ_REL]; in dpaa_mbuf_alloc_bulk() local 241 ret = bman_acquire(bp_info->bp, bufs, in dpaa_mbuf_alloc_bulk() 244 ret = bman_acquire(bp_info->bp, bufs, count - n, 0); in dpaa_mbuf_alloc_bulk() 259 for (i = 0; (i < ret) && bufs[i].addr; i++) { in dpaa_mbuf_alloc_bulk() 264 bufaddr = DPAA_MEMPOOL_PTOV(bp_info, bufs[i].addr); in dpaa_mbuf_alloc_bulk()
|
| /f-stack/dpdk/drivers/bus/dpaa/base/qbman/ |
| H A D | bman.c | 250 int bman_release(struct bman_pool *pool, const struct bm_buffer *bufs, u8 num, in bman_release() argument 277 r->bufs[0].opaque = in bman_release() 279 (bufs[0].opaque & BMAN_BUF_MASK)); in bman_release() 282 r->bufs[i].opaque = in bman_release() 283 cpu_to_be64(bufs[i].opaque & BMAN_BUF_MASK); in bman_release() 292 int bman_acquire(struct bman_pool *pool, struct bm_buffer *bufs, u8 num, in bman_acquire() argument 314 if (bufs) { in bman_acquire() 316 bufs[i].opaque = in bman_acquire() 317 be64_to_cpu(mcr->acquire.bufs[i].opaque); in bman_acquire()
|
| /f-stack/dpdk/drivers/net/ark/ |
| H A D | ark_ethdev_tx.c | 27 struct rte_mbuf **bufs; member 142 queue->bufs[idx] = mbuf; in eth_ark_xmit_pkts() 194 queue->bufs[idx] = mbuf; in eth_ark_tx_jumbo() 254 queue->bufs = in eth_ark_tx_queue_setup() 260 if (queue->meta_q == 0 || queue->bufs == 0) { in eth_ark_tx_queue_setup() 264 rte_free(queue->bufs); in eth_ark_tx_queue_setup() 276 rte_free(queue->bufs); in eth_ark_tx_queue_setup() 346 rte_free(queue->bufs); in eth_ark_tx_queue_release() 400 mbuf = queue->bufs[queue->free_index & queue->queue_mask]; in free_completed_tx()
|
| /f-stack/dpdk/drivers/net/enic/base/ |
| H A D | vnic_wq.c | 35 wq->bufs = (struct rte_mbuf **)rte_zmalloc_socket("wq->bufs", in vnic_wq_alloc_bufs() 40 if (wq->bufs == NULL) in vnic_wq_alloc_bufs() 53 rte_free(wq->bufs); in vnic_wq_free() 154 buf = &wq->bufs[to_clean]; in vnic_wq_clean() 161 buf = &wq->bufs[to_clean]; in vnic_wq_clean()
|
| /f-stack/dpdk/app/test-bbdev/ |
| H A D | test_bbdev_perf.c | 2750 bufs->hard_outputs, bufs->soft_outputs, in throughput_intr_lcore_ldpc_dec() 2751 bufs->harq_inputs, bufs->harq_outputs, ref_op); in throughput_intr_lcore_ldpc_dec() 2847 bufs->hard_outputs, bufs->soft_outputs, in throughput_intr_lcore_dec() 3027 bufs->inputs, bufs->hard_outputs, in throughput_intr_lcore_ldpc_enc() 3113 bufs->hard_outputs, bufs->soft_outputs, ref_op); in throughput_pmd_lcore_dec() 3225 bufs->hard_outputs, bufs->soft_outputs, in bler_pmd_lcore_ldpc_dec() 3226 bufs->harq_inputs, bufs->harq_outputs, ref_op); in bler_pmd_lcore_ldpc_dec() 3355 bufs->hard_outputs, bufs->soft_outputs, in throughput_pmd_lcore_ldpc_dec() 3356 bufs->harq_inputs, bufs->harq_outputs, ref_op); in throughput_pmd_lcore_ldpc_dec() 3966 bufs->inputs, in latency_test_dec() [all …]
|
| /f-stack/app/nginx-1.16.1/src/http/ |
| H A D | ngx_http_copy_filter_module.c | 14 ngx_bufs_t bufs; member 45 offsetof(ngx_http_copy_filter_conf_t, bufs), 121 ctx->bufs = conf->bufs; in ngx_http_copy_filter() 354 conf->bufs.num = 0; in ngx_http_copy_filter_create_conf() 366 ngx_conf_merge_bufs_value(conf->bufs, prev->bufs, 2, 32768); in ngx_http_copy_filter_merge_conf()
|
| /f-stack/dpdk/drivers/net/bonding/ |
| H A D | rte_eth_bond_pmd.c | 82 bufs + num_rx_total, nb_pkts); in bond_ethdev_rx_burst() 106 bd_rx_q->queue_id, bufs, nb_pkts); in bond_ethdev_rx_burst_active_backup() 342 bufs[j])) || in rx_burst_8023ad() 353 internals, slaves[idx], bufs[j]); in rx_burst_8023ad() 355 rte_pktmbuf_free(bufs[j]); in rx_burst_8023ad() 360 memmove(&bufs[j], &bufs[j + 1], sizeof(bufs[0]) * in rx_burst_8023ad() 643 bufs, nb_pkts); in bond_ethdev_tx_burst_active_backup() 947 ether_hdr = rte_pktmbuf_mtod(bufs[j], in bond_ethdev_tx_burst_tlb() 1032 bufs[i]; in bond_ethdev_tx_burst_alb() 1117 bufs[nb_pkts - 1 - num_not_send - j] = in bond_ethdev_tx_burst_alb() [all …]
|
| /f-stack/dpdk/examples/l2fwd-cat/ |
| H A D | l2fwd-cat.c | 129 struct rte_mbuf *bufs[BURST_SIZE]; in lcore_main() local 131 bufs, BURST_SIZE); in lcore_main() 138 bufs, nb_rx); in lcore_main() 144 rte_pktmbuf_free(bufs[buf]); in lcore_main()
|
| /f-stack/dpdk/examples/skeleton/ |
| H A D | basicfwd.c | 144 struct rte_mbuf *bufs[BURST_SIZE]; in lcore_main() local 146 bufs, BURST_SIZE); in lcore_main() 153 bufs, nb_rx); in lcore_main() 159 rte_pktmbuf_free(bufs[buf]); in lcore_main()
|