Home
last modified time | relevance | path

Searched refs:processed_pkts (Results 1 – 23 of 23) sorted by relevance

/dpdk/app/test-eventdev/
H A Dtest_pipeline_queue.c37 w->processed_pkts++; in pipeline_queue_worker_single_stage_tx()
68 w->processed_pkts++; in pipeline_queue_worker_single_stage_fwd()
94 w->processed_pkts++; in pipeline_queue_worker_single_stage_burst_tx()
131 w->processed_pkts += nb_rx; in pipeline_queue_worker_single_stage_burst_fwd()
190 w->processed_pkts += vector_sz; in pipeline_queue_worker_single_stage_fwd_vector()
285 w->processed_pkts++; in pipeline_queue_worker_multi_stage_tx()
322 w->processed_pkts++; in pipeline_queue_worker_multi_stage_fwd()
356 w->processed_pkts++; in pipeline_queue_worker_multi_stage_burst_tx()
380 uint16_t processed_pkts = 0; in pipeline_queue_worker_multi_stage_burst_fwd() local
397 processed_pkts++; in pipeline_queue_worker_multi_stage_burst_fwd()
[all …]
H A Dtest_pipeline_atq.c35 w->processed_pkts++; in pipeline_atq_worker_single_stage_tx()
60 w->processed_pkts++; in pipeline_atq_worker_single_stage_fwd()
87 w->processed_pkts += nb_tx; in pipeline_atq_worker_single_stage_burst_tx()
117 w->processed_pkts += nb_tx; in pipeline_atq_worker_single_stage_burst_fwd()
140 w->processed_pkts += vector_sz; in pipeline_atq_worker_single_stage_tx_vector()
168 w->processed_pkts += vector_sz; in pipeline_atq_worker_single_stage_fwd_vector()
196 w->processed_pkts += vector_sz; in pipeline_atq_worker_single_stage_burst_tx_vector()
254 w->processed_pkts++; in pipeline_atq_worker_multi_stage_tx()
287 w->processed_pkts++; in pipeline_atq_worker_multi_stage_fwd()
321 w->processed_pkts++; in pipeline_atq_worker_multi_stage_burst_tx()
[all …]
H A Dtest_perf_common.h34 uint64_t processed_pkts; member
122 w->processed_pkts++; in perf_process_last_stage()
146 w->processed_pkts++; in perf_process_last_stage_latency()
H A Dtest_pipeline_common.c18 total += t->worker[i].processed_pkts; in pipeline_test_result()
22 t->worker[i].processed_pkts, in pipeline_test_result()
23 (((double)t->worker[i].processed_pkts)/total) in pipeline_test_result()
48 processed_pkts(struct test_pipeline *t) in processed_pkts() function
54 total += t->worker[i].processed_pkts; in processed_pkts()
93 const uint64_t curr_pkts = processed_pkts(t); in pipeline_launch_lcores()
306 w->processed_pkts = 0; in pipeline_event_port_setup()
H A Dtest_perf_common.c21 total += t->worker[i].processed_pkts; in perf_test_result()
25 t->worker[i].processed_pkts, in perf_test_result()
26 (((double)t->worker[i].processed_pkts)/total) in perf_test_result()
426 processed_pkts(struct test_perf *t) in processed_pkts() function
432 total += t->worker[i].processed_pkts; in processed_pkts()
500 int64_t remaining = t->outstand_pkts - processed_pkts(t); in perf_launch_lcores()
507 const uint64_t pkts = processed_pkts(t); in perf_launch_lcores()
543 remaining = t->outstand_pkts - processed_pkts(t); in perf_launch_lcores()
749 w->processed_pkts = 0; in perf_event_dev_port_setup()
H A Dtest_pipeline_common.h33 uint64_t processed_pkts; member
/dpdk/drivers/baseband/null/
H A Dbbdev_null.c50 struct rte_ring *processed_pkts; /* Ring for processed packets */ member
96 rte_ring_free(q->processed_pkts); in q_release()
124 q->processed_pkts = rte_ring_create(ring_name, queue_conf->queue_size, in q_setup()
126 if (q->processed_pkts == NULL) { in q_setup()
152 uint16_t nb_enqueued = rte_ring_enqueue_burst(q->processed_pkts, in enqueue_dec_ops()
167 uint16_t nb_enqueued = rte_ring_enqueue_burst(q->processed_pkts, in enqueue_enc_ops()
182 uint16_t nb_dequeued = rte_ring_dequeue_burst(q->processed_pkts, in dequeue_dec_ops()
195 uint16_t nb_dequeued = rte_ring_dequeue_burst(q->processed_pkts, in dequeue_enc_ops()
/dpdk/drivers/compress/zlib/
H A Dzlib_pmd_ops.c132 rte_ring_free(qp->processed_pkts); in zlib_pmd_qp_release()
159 struct rte_ring *r = qp->processed_pkts; in zlib_pmd_qp_create_processed_pkts_ring()
200 qp->processed_pkts = zlib_pmd_qp_create_processed_pkts_ring(qp, in zlib_pmd_qp_setup()
202 if (qp->processed_pkts == NULL) in zlib_pmd_qp_setup()
H A Dzlib_pmd_private.h34 struct rte_ring *processed_pkts; member
H A Dzlib_pmd.c215 return rte_ring_enqueue(qp->processed_pkts, (void *)op); in process_zlib_op()
347 nb_dequeued = rte_ring_dequeue_burst(qp->processed_pkts, in zlib_pmd_dequeue_burst()
/dpdk/drivers/compress/isal/
H A Disal_compress_pmd_ops.c184 rte_ring_free(qp->processed_pkts); in isal_comp_pmd_qp_release()
290 qp->processed_pkts = isal_comp_pmd_qp_create_processed_pkts_ring(qp, in isal_comp_pmd_qp_setup()
292 if (qp->processed_pkts == NULL) { in isal_comp_pmd_qp_setup()
298 qp->num_free_elements = rte_ring_free_count(qp->processed_pkts); in isal_comp_pmd_qp_setup()
H A Disal_compress_pmd_private.h28 struct rte_ring *processed_pkts; member
H A Disal_compress_pmd.c655 retval = rte_ring_enqueue_burst(qp->processed_pkts, (void *)ops, in isal_comp_pmd_enqueue_burst()
671 nb_dequeued = rte_ring_dequeue_burst(qp->processed_pkts, (void **)ops, in isal_comp_pmd_dequeue_burst()
/dpdk/drivers/compress/octeontx/
H A Dotx_zip_pmd.c342 rte_ring_free(qp->processed_pkts); in zip_pmd_qp_release()
413 qp->processed_pkts = zip_pmd_qp_create_processed_pkts_ring(qp, in zip_pmd_qp_setup()
415 if (qp->processed_pkts == NULL) in zip_pmd_qp_setup()
431 rte_ring_free(qp->processed_pkts); in zip_pmd_qp_setup()
511 ret = rte_ring_enqueue(qp->processed_pkts, (void *)op); in zip_pmd_enqueue_burst_sync()
532 nb_dequeued = rte_ring_dequeue_burst(qp->processed_pkts, in zip_pmd_dequeue_burst_sync()
H A Dotx_zip.h133 struct rte_ring *processed_pkts; member
/dpdk/drivers/crypto/null/
H A Dnull_crypto_pmd_ops.c138 rte_ring_free(qp->processed_pkts); in null_crypto_pmd_qp_release()
227 qp->processed_pkts = null_crypto_pmd_qp_create_processed_pkts_ring(qp, in null_crypto_pmd_qp_setup()
229 if (qp->processed_pkts == NULL) { in null_crypto_pmd_qp_setup()
H A Dnull_crypto_pmd_private.h30 struct rte_ring *processed_pkts; member
H A Dnull_crypto_pmd.c69 return rte_ring_enqueue(qp->processed_pkts, (void *)op); in process_op()
150 nb_dequeued = rte_ring_dequeue_burst(qp->processed_pkts, in null_crypto_pmd_dequeue_burst()
/dpdk/drivers/baseband/turbo_sw/
H A Dbbdev_turbo_software.c78 struct rte_ring *processed_pkts; member
268 rte_ring_free(q->processed_pkts); in q_release()
457 q->processed_pkts = rte_ring_create(name, queue_conf->queue_size, in q_setup()
459 if (q->processed_pkts == NULL) { in q_setup()
472 rte_ring_free(q->processed_pkts); in q_setup()
1175 return rte_ring_enqueue_burst(q->processed_pkts, (void **)ops, nb_ops, in enqueue_enc_all_ops()
1192 return rte_ring_enqueue_burst(q->processed_pkts, (void **)ops, nb_ops, in enqueue_ldpc_enc_all_ops()
1753 return rte_ring_enqueue_burst(q->processed_pkts, (void **)ops, nb_ops, in enqueue_dec_all_ops()
1770 return rte_ring_enqueue_burst(q->processed_pkts, (void **)ops, nb_ops, in enqueue_ldpc_dec_all_ops()
1850 uint16_t nb_dequeued = rte_ring_dequeue_burst(q->processed_pkts, in dequeue_dec_ops()
[all …]
/dpdk/drivers/crypto/ccp/
H A Dccp_pmd_private.h77 struct rte_ring *processed_pkts; member
H A Dccp_pmd_ops.c639 rte_ring_free(qp->processed_pkts); in ccp_pmd_qp_release()
722 qp->processed_pkts = ccp_pmd_qp_create_batch_info_ring(qp, in ccp_pmd_qp_setup()
724 if (qp->processed_pkts == NULL) { in ccp_pmd_qp_setup()
H A Drte_ccp_pmd.c105 if (unlikely(rte_ring_full(qp->processed_pkts) != 0)) in ccp_pmd_enqueue_burst()
H A Dccp_crypto.c2845 rte_ring_enqueue(qp->processed_pkts, (void *)b_info); in process_ops_to_enqueue()
2985 } else if (rte_ring_dequeue(qp->processed_pkts, in process_ops_to_dequeue()