Home
last modified time | relevance | path

Searched refs:bufs (Results 1 – 25 of 104) sorted by relevance

12345

/f-stack/dpdk/app/test/
H A Dtest_reorder.c168 bufs[i] = NULL; in test_reorder_insert()
183 bufs[4] = NULL; in test_reorder_insert()
194 bufs[5] = NULL; in test_reorder_insert()
205 bufs[6] = NULL; in test_reorder_insert()
211 if (bufs[i] != NULL) in test_reorder_insert()
262 bufs[1] = NULL; in test_reorder_drain()
280 bufs[2] = NULL; in test_reorder_drain()
281 bufs[3] = NULL; in test_reorder_drain()
288 bufs[4] = NULL; in test_reorder_drain()
295 bufs[7] = NULL; in test_reorder_drain()
[all …]
H A Dtest_distributor.c112 struct rte_mbuf *bufs[BURST]; in sanity_test() local
128 bufs[i]->hash.usr = 0; in sanity_test()
190 bufs[i]->hash.usr = i+1; in sanity_test()
326 struct rte_mbuf *bufs[BURST]; in sanity_test_with_mbuf_alloc() local
337 bufs[j]->hash.usr = (i+j) << 1; in sanity_test_with_mbuf_alloc()
433 struct rte_mbuf *bufs[BURST]; in sanity_test_with_worker_shutdown() local
453 bufs[i]->hash.usr = 1; in sanity_test_with_worker_shutdown()
517 struct rte_mbuf *bufs[BURST]; in test_flush_with_worker_shutdown() local
533 bufs[i]->hash.usr = 0; in test_flush_with_worker_shutdown()
654 &bufs[i * burst + processed], in sanity_mark_test()
[all …]
H A Dtest_distributor_perf.c142 struct rte_mbuf *bufs[BURST]; in perf_test() local
145 if (rte_mempool_get_bulk(p, (void *)bufs, BURST) != 0) { in perf_test()
151 bufs[i]->hash.usr = i; in perf_test()
155 rte_distributor_process(d, bufs, BURST); in perf_test()
168 rte_mempool_put_bulk(p, (void *)bufs, BURST); in perf_test()
186 struct rte_mbuf *bufs[RTE_MAX_LCORE]; in quit_workers() local
188 rte_mempool_get_bulk(p, (void *)bufs, num_workers); in quit_workers()
192 bufs[i]->hash.usr = i << 1; in quit_workers()
193 rte_distributor_process(d, bufs, num_workers); in quit_workers()
195 rte_mempool_put_bulk(p, (void *)bufs, num_workers); in quit_workers()
H A Dvirtual_pmd.c325 struct rte_mbuf **bufs, in virtual_ethdev_rx_burst_success() argument
338 rx_count = rte_ring_dequeue_burst(dev_private->rx_queue, (void **) bufs, in virtual_ethdev_rx_burst_success()
346 dev_private->eth_stats.ibytes += rte_pktmbuf_pkt_len(bufs[i]); in virtual_ethdev_rx_burst_success()
353 struct rte_mbuf **bufs __rte_unused, in virtual_ethdev_rx_burst_fail()
360 virtual_ethdev_tx_burst_success(void *queue, struct rte_mbuf **bufs, in virtual_ethdev_tx_burst_success() argument
376 nb_pkts = rte_ring_enqueue_burst(dev_private->tx_queue, (void **)bufs, in virtual_ethdev_tx_burst_success()
384 dev_private->eth_stats.obytes += rte_pktmbuf_pkt_len(bufs[i]); in virtual_ethdev_tx_burst_success()
390 virtual_ethdev_tx_burst_fail(void *queue, struct rte_mbuf **bufs, in virtual_ethdev_tx_burst_fail() argument
412 if (bufs[i] != NULL) in virtual_ethdev_tx_burst_fail()
413 rte_pktmbuf_free(bufs[i]); in virtual_ethdev_tx_burst_fail()
[all …]
/f-stack/dpdk/drivers/net/dpaa2/
H A Ddpaa2_rxtx.c1077 if (*dpaa2_seqn(*bufs)) { in dpaa2_dev_tx()
1088 mp = (*bufs)->pool; in dpaa2_dev_tx()
1106 bufs++; in dpaa2_dev_tx()
1154 eth_mbuf_to_fd(*bufs, in dpaa2_dev_tx()
1161 bufs++; in dpaa2_dev_tx()
1345 (*bufs), in dpaa2_dev_tx_ordered()
1355 mp = (*bufs)->pool; in dpaa2_dev_tx_ordered()
1372 bufs++; in dpaa2_dev_tx_ordered()
1411 eth_mbuf_to_fd(*bufs, in dpaa2_dev_tx_ordered()
1415 bufs++; in dpaa2_dev_tx_ordered()
[all …]
H A Ddpaa2_ethdev.h200 uint16_t dpaa2_dev_rx(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts);
202 uint16_t dpaa2_dev_loopback_rx(void *queue, struct rte_mbuf **bufs,
205 uint16_t dpaa2_dev_prefetch_rx(void *queue, struct rte_mbuf **bufs,
222 uint16_t dpaa2_dev_tx(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts);
223 uint16_t dpaa2_dev_tx_ordered(void *queue, struct rte_mbuf **bufs,
225 uint16_t dummy_dev_tx(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts);
/f-stack/dpdk/drivers/mempool/dpaa2/
H A Ddpaa2_hw_mempool.c195 uint64_t bufs[DPAA2_MBUF_MAX_ACQ_REL]; in rte_dpaa2_mbuf_release() local
221 bufs[i] = (uint64_t)rte_mempool_virt2iova(obj_table[i]) in rte_dpaa2_mbuf_release()
224 bufs[i] = (uint64_t)obj_table[i] + meta_data_size; in rte_dpaa2_mbuf_release()
245 bufs[i] = (uint64_t) in rte_dpaa2_mbuf_release()
304 size_t bufs[DPAA2_MBUF_MAX_ACQ_REL]; in rte_dpaa2_mbuf_alloc_bulk() local
334 ret = qbman_swp_acquire(swp, bpid, (void *)bufs, in rte_dpaa2_mbuf_alloc_bulk()
337 ret = qbman_swp_acquire(swp, bpid, (void *)bufs, in rte_dpaa2_mbuf_alloc_bulk()
353 for (i = 0; (i < ret) && bufs[i]; i++) { in rte_dpaa2_mbuf_alloc_bulk()
354 DPAA2_MODIFY_IOVA_TO_VADDR(bufs[i], size_t); in rte_dpaa2_mbuf_alloc_bulk()
356 (bufs[i] - bp_info->meta_data_size); in rte_dpaa2_mbuf_alloc_bulk()
[all …]
/f-stack/dpdk/app/test-eventdev/
H A Dtest_perf_common.h88 void *bufs[16] __rte_cache_aligned;\
89 int const sz = RTE_DIM(bufs);\
97 void *bufs[], int const buf_sz, uint8_t count) in perf_process_last_stage() argument
99 bufs[count++] = ev->event_ptr; in perf_process_last_stage()
105 rte_mempool_put_bulk(pool, bufs, buf_sz); in perf_process_last_stage()
113 void *bufs[], int const buf_sz, uint8_t count) in perf_process_last_stage_latency() argument
118 bufs[count++] = ev->event_ptr; in perf_process_last_stage_latency()
124 rte_mempool_put_bulk(pool, bufs, buf_sz); in perf_process_last_stage_latency()
/f-stack/dpdk/drivers/net/null/
H A Drte_eth_null.c89 if ((q == NULL) || (bufs == NULL)) in eth_null_rx()
97 bufs[i]->data_len = (uint16_t)packet_size; in eth_null_rx()
98 bufs[i]->pkt_len = packet_size; in eth_null_rx()
99 bufs[i]->port = h->internals->port_id; in eth_null_rx()
114 if ((q == NULL) || (bufs == NULL)) in eth_null_copy_rx()
125 bufs[i]->pkt_len = packet_size; in eth_null_copy_rx()
126 bufs[i]->port = h->internals->port_id; in eth_null_copy_rx()
147 if ((q == NULL) || (bufs == NULL)) in eth_null_tx()
151 rte_pktmbuf_free(bufs[i]); in eth_null_tx()
165 if ((q == NULL) || (bufs == NULL)) in eth_null_copy_tx()
[all …]
/f-stack/dpdk/lib/librte_distributor/
H A Drte_distributor_single.c33 union rte_distributor_buffer_single *buf = &d->bufs[worker_id]; in EAL_REGISTER_TAILQ()
48 union rte_distributor_buffer_single *buf = &d->bufs[worker_id]; in rte_distributor_poll_pkt_single()
74 union rte_distributor_buffer_single *buf = &d->bufs[worker_id]; in rte_distributor_return_pkt_single()
179 __atomic_store_n(&(d->bufs[wkr].bufptr64), in process_returns()
184 __atomic_store_n(&(d->bufs[wkr].bufptr64), in process_returns()
224 int64_t data = __atomic_load_n(&(d->bufs[wkr].bufptr64), in rte_distributor_process_single()
272 __atomic_store_n(&(d->bufs[wkr].bufptr64), in rte_distributor_process_single()
278 __atomic_store_n(&(d->bufs[wkr].bufptr64), in rte_distributor_process_single()
302 (__atomic_load_n(&(d->bufs[wkr].bufptr64), in rte_distributor_process_single()
305 int64_t oldbuf = d->bufs[wkr].bufptr64 >> in rte_distributor_process_single()
[all …]
H A Drte_distributor.c39 struct rte_distributor_buffer *buf = &(d->bufs[worker_id]); in EAL_REGISTER_TAILQ()
90 struct rte_distributor_buffer *buf = &d->bufs[worker_id]; in rte_distributor_poll_pkt()
291 struct rte_distributor_buffer *buf = &(d->bufs[wkr]); in handle_worker_shutdown()
344 struct rte_distributor_buffer *buf = &(d->bufs[wkr]); in handle_returns()
400 struct rte_distributor_buffer *buf = &(d->bufs[wkr]); in release()
419 d->bufs[wkr].bufptr64[i] = d->backlog[wkr].pkts[i] | in release()
467 if (__atomic_load_n(&(d->bufs[wid].bufptr64[0]), in rte_distributor_process()
469 d->bufs[wid].count = 0; in rte_distributor_process()
602 if ((__atomic_load_n(&(d->bufs[wid].bufptr64[0]), in rte_distributor_process()
604 d->bufs[wid].count = 0; in rte_distributor_process()
[all …]
/f-stack/dpdk/examples/distributor/
H A Dmain.c230 struct rte_mbuf *bufs[BURST_SIZE*2]; in lcore_rx() local
270 bufs, BURST_SIZE*2); in lcore_rx()
281 (void *)bufs, nb_ret, NULL); in lcore_rx()
292 (void *)bufs, nb_ret, NULL); in lcore_rx()
301 rte_pktmbuf_free(bufs[sent++]); in lcore_rx()
354 struct rte_mbuf *bufs[BURST_SIZE * 4]; in lcore_distributor() local
360 (void *)bufs, BURST_SIZE*1, NULL); in lcore_distributor()
369 bufs, BURST_SIZE*2); in lcore_distributor()
376 (void *)bufs, nb_ret, NULL); in lcore_distributor()
384 rte_pktmbuf_free(bufs[sent++]); in lcore_distributor()
[all …]
/f-stack/dpdk/drivers/net/dpaa/
H A Ddpaa_rxtx.h272 uint16_t dpaa_eth_queue_rx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs);
274 uint16_t dpaa_eth_queue_tx_slow(void *q, struct rte_mbuf **bufs,
276 uint16_t dpaa_eth_queue_tx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs);
279 struct rte_mbuf **bufs __rte_unused,
290 struct qm_dqrr_entry **dqrr, void **bufs, int num_bufs);
292 void dpaa_rx_cb_prepare(struct qm_dqrr_entry *dq, void **bufs);
295 struct qm_dqrr_entry **dqrr, void **bufs, int num_bufs);
H A Ddpaa_rxtx.c503 void **bufs, int num_bufs) in dpaa_rx_cb_no_prefetch() argument
542 mbuf = bufs[i]; in dpaa_rx_cb_no_prefetch()
559 void **bufs, int num_bufs) in dpaa_rx_cb() argument
582 mbuf = bufs[i]; in dpaa_rx_cb()
614 struct rte_mbuf **bufs, in dpaa_eth_queue_portal_rx() argument
636 void **bufs) in dpaa_rx_cb_parallel() argument
653 *bufs = mbuf; in dpaa_rx_cb_parallel()
663 void **bufs) in dpaa_rx_cb_atomic() argument
687 *bufs = mbuf; in dpaa_rx_cb_atomic()
737 struct rte_mbuf **bufs, in dpaa_eth_queue_rx() argument
[all …]
/f-stack/freebsd/contrib/ncsw/Peripherals/BM/
H A Dbm_pool.c80 struct bm_buffer *bufs, in BmPoolRelease() argument
111 bufs, in BmPoolRelease()
134 struct bm_buffer *bufs, uint8_t num, uint32_t flags) in BmPoolAcquire() argument
142 return BmPortalAcquire(h_BmPortal, p_BmPool->bpid, bufs, num); in BmPoolAcquire()
156 memcpy(bufs, in BmPoolAcquire()
405 struct bm_buffer bufs[1]; in BM_POOL_GetBuf() local
432 physAddr = (uint64_t)bufs[0].lo; in BM_POOL_GetBuf()
445 struct bm_buffer bufs[1]; in BM_POOL_PutBuf() local
453 bufs[0].bpid = p_BmPool->bpid; in BM_POOL_PutBuf()
454 bufs[0].hi = (uint8_t)((physAddress & 0x000000ff00000000LL) >> 32); in BM_POOL_PutBuf()
[all …]
H A Dbm_portal.c370 struct bm_buffer *bufs, in BmPortalRelease() argument
392 r->bufs[i].hi = bufs[i].hi; in BmPortalRelease()
393 r->bufs[i].lo = bufs[i].lo; in BmPortalRelease()
406 struct bm_buffer *bufs, in BmPortalAcquire() argument
425 bufs[num].bpid = bpid; in BmPortalAcquire()
426 bufs[num].hi = mcr->acquire.bufs[num].hi; in BmPortalAcquire()
427 bufs[num].lo = mcr->acquire.bufs[num].lo; in BmPortalAcquire()
/f-stack/dpdk/drivers/mempool/dpaa/
H A Ddpaa_mempool.c46 struct bm_buffer bufs[8]; in dpaa_mbuf_create_pool() local
78 ret = bman_acquire(bp, bufs, 8, 0); in dpaa_mbuf_create_pool()
80 ret = bman_acquire(bp, bufs, 1, 0); in dpaa_mbuf_create_pool()
210 struct bm_buffer bufs[DPAA_MBUF_MAX_ACQ_REL]; in dpaa_mbuf_alloc_bulk() local
241 ret = bman_acquire(bp_info->bp, bufs, in dpaa_mbuf_alloc_bulk()
244 ret = bman_acquire(bp_info->bp, bufs, count - n, 0); in dpaa_mbuf_alloc_bulk()
259 for (i = 0; (i < ret) && bufs[i].addr; i++) { in dpaa_mbuf_alloc_bulk()
264 bufaddr = DPAA_MEMPOOL_PTOV(bp_info, bufs[i].addr); in dpaa_mbuf_alloc_bulk()
/f-stack/dpdk/drivers/bus/dpaa/base/qbman/
H A Dbman.c250 int bman_release(struct bman_pool *pool, const struct bm_buffer *bufs, u8 num, in bman_release() argument
277 r->bufs[0].opaque = in bman_release()
279 (bufs[0].opaque & BMAN_BUF_MASK)); in bman_release()
282 r->bufs[i].opaque = in bman_release()
283 cpu_to_be64(bufs[i].opaque & BMAN_BUF_MASK); in bman_release()
292 int bman_acquire(struct bman_pool *pool, struct bm_buffer *bufs, u8 num, in bman_acquire() argument
314 if (bufs) { in bman_acquire()
316 bufs[i].opaque = in bman_acquire()
317 be64_to_cpu(mcr->acquire.bufs[i].opaque); in bman_acquire()
/f-stack/dpdk/drivers/net/ark/
H A Dark_ethdev_tx.c27 struct rte_mbuf **bufs; member
142 queue->bufs[idx] = mbuf; in eth_ark_xmit_pkts()
194 queue->bufs[idx] = mbuf; in eth_ark_tx_jumbo()
254 queue->bufs = in eth_ark_tx_queue_setup()
260 if (queue->meta_q == 0 || queue->bufs == 0) { in eth_ark_tx_queue_setup()
264 rte_free(queue->bufs); in eth_ark_tx_queue_setup()
276 rte_free(queue->bufs); in eth_ark_tx_queue_setup()
346 rte_free(queue->bufs); in eth_ark_tx_queue_release()
400 mbuf = queue->bufs[queue->free_index & queue->queue_mask]; in free_completed_tx()
/f-stack/dpdk/drivers/net/enic/base/
H A Dvnic_wq.c35 wq->bufs = (struct rte_mbuf **)rte_zmalloc_socket("wq->bufs", in vnic_wq_alloc_bufs()
40 if (wq->bufs == NULL) in vnic_wq_alloc_bufs()
53 rte_free(wq->bufs); in vnic_wq_free()
154 buf = &wq->bufs[to_clean]; in vnic_wq_clean()
161 buf = &wq->bufs[to_clean]; in vnic_wq_clean()
/f-stack/dpdk/app/test-bbdev/
H A Dtest_bbdev_perf.c2750 bufs->hard_outputs, bufs->soft_outputs, in throughput_intr_lcore_ldpc_dec()
2751 bufs->harq_inputs, bufs->harq_outputs, ref_op); in throughput_intr_lcore_ldpc_dec()
2847 bufs->hard_outputs, bufs->soft_outputs, in throughput_intr_lcore_dec()
3027 bufs->inputs, bufs->hard_outputs, in throughput_intr_lcore_ldpc_enc()
3113 bufs->hard_outputs, bufs->soft_outputs, ref_op); in throughput_pmd_lcore_dec()
3225 bufs->hard_outputs, bufs->soft_outputs, in bler_pmd_lcore_ldpc_dec()
3226 bufs->harq_inputs, bufs->harq_outputs, ref_op); in bler_pmd_lcore_ldpc_dec()
3355 bufs->hard_outputs, bufs->soft_outputs, in throughput_pmd_lcore_ldpc_dec()
3356 bufs->harq_inputs, bufs->harq_outputs, ref_op); in throughput_pmd_lcore_ldpc_dec()
3966 bufs->inputs, in latency_test_dec()
[all …]
/f-stack/app/nginx-1.16.1/src/http/
H A Dngx_http_copy_filter_module.c14 ngx_bufs_t bufs; member
45 offsetof(ngx_http_copy_filter_conf_t, bufs),
121 ctx->bufs = conf->bufs; in ngx_http_copy_filter()
354 conf->bufs.num = 0; in ngx_http_copy_filter_create_conf()
366 ngx_conf_merge_bufs_value(conf->bufs, prev->bufs, 2, 32768); in ngx_http_copy_filter_merge_conf()
/f-stack/dpdk/drivers/net/bonding/
H A Drte_eth_bond_pmd.c82 bufs + num_rx_total, nb_pkts); in bond_ethdev_rx_burst()
106 bd_rx_q->queue_id, bufs, nb_pkts); in bond_ethdev_rx_burst_active_backup()
342 bufs[j])) || in rx_burst_8023ad()
353 internals, slaves[idx], bufs[j]); in rx_burst_8023ad()
355 rte_pktmbuf_free(bufs[j]); in rx_burst_8023ad()
360 memmove(&bufs[j], &bufs[j + 1], sizeof(bufs[0]) * in rx_burst_8023ad()
643 bufs, nb_pkts); in bond_ethdev_tx_burst_active_backup()
947 ether_hdr = rte_pktmbuf_mtod(bufs[j], in bond_ethdev_tx_burst_tlb()
1032 bufs[i]; in bond_ethdev_tx_burst_alb()
1117 bufs[nb_pkts - 1 - num_not_send - j] = in bond_ethdev_tx_burst_alb()
[all …]
/f-stack/dpdk/examples/l2fwd-cat/
H A Dl2fwd-cat.c129 struct rte_mbuf *bufs[BURST_SIZE]; in lcore_main() local
131 bufs, BURST_SIZE); in lcore_main()
138 bufs, nb_rx); in lcore_main()
144 rte_pktmbuf_free(bufs[buf]); in lcore_main()
/f-stack/dpdk/examples/skeleton/
H A Dbasicfwd.c144 struct rte_mbuf *bufs[BURST_SIZE]; in lcore_main() local
146 bufs, BURST_SIZE); in lcore_main()
153 bufs, nb_rx); in lcore_main()
159 rte_pktmbuf_free(bufs[buf]); in lcore_main()

12345