| /dpdk/app/test-eventdev/ |
| H A D | test_pipeline_queue.c | 37 w->processed_pkts++; in pipeline_queue_worker_single_stage_tx() 68 w->processed_pkts++; in pipeline_queue_worker_single_stage_fwd() 94 w->processed_pkts++; in pipeline_queue_worker_single_stage_burst_tx() 131 w->processed_pkts += nb_rx; in pipeline_queue_worker_single_stage_burst_fwd() 190 w->processed_pkts += vector_sz; in pipeline_queue_worker_single_stage_fwd_vector() 285 w->processed_pkts++; in pipeline_queue_worker_multi_stage_tx() 322 w->processed_pkts++; in pipeline_queue_worker_multi_stage_fwd() 356 w->processed_pkts++; in pipeline_queue_worker_multi_stage_burst_tx() 380 uint16_t processed_pkts = 0; in pipeline_queue_worker_multi_stage_burst_fwd() local 397 processed_pkts++; in pipeline_queue_worker_multi_stage_burst_fwd() [all …]
|
| H A D | test_pipeline_atq.c | 35 w->processed_pkts++; in pipeline_atq_worker_single_stage_tx() 60 w->processed_pkts++; in pipeline_atq_worker_single_stage_fwd() 87 w->processed_pkts += nb_tx; in pipeline_atq_worker_single_stage_burst_tx() 117 w->processed_pkts += nb_tx; in pipeline_atq_worker_single_stage_burst_fwd() 140 w->processed_pkts += vector_sz; in pipeline_atq_worker_single_stage_tx_vector() 168 w->processed_pkts += vector_sz; in pipeline_atq_worker_single_stage_fwd_vector() 196 w->processed_pkts += vector_sz; in pipeline_atq_worker_single_stage_burst_tx_vector() 254 w->processed_pkts++; in pipeline_atq_worker_multi_stage_tx() 287 w->processed_pkts++; in pipeline_atq_worker_multi_stage_fwd() 321 w->processed_pkts++; in pipeline_atq_worker_multi_stage_burst_tx() [all …]
|
| H A D | test_perf_common.h | 34 uint64_t processed_pkts; member 122 w->processed_pkts++; in perf_process_last_stage() 146 w->processed_pkts++; in perf_process_last_stage_latency()
|
| H A D | test_pipeline_common.c | 18 total += t->worker[i].processed_pkts; in pipeline_test_result() 22 t->worker[i].processed_pkts, in pipeline_test_result() 23 (((double)t->worker[i].processed_pkts)/total) in pipeline_test_result() 48 processed_pkts(struct test_pipeline *t) in processed_pkts() function 54 total += t->worker[i].processed_pkts; in processed_pkts() 93 const uint64_t curr_pkts = processed_pkts(t); in pipeline_launch_lcores() 306 w->processed_pkts = 0; in pipeline_event_port_setup()
|
| H A D | test_perf_common.c | 21 total += t->worker[i].processed_pkts; in perf_test_result() 25 t->worker[i].processed_pkts, in perf_test_result() 26 (((double)t->worker[i].processed_pkts)/total) in perf_test_result() 426 processed_pkts(struct test_perf *t) in processed_pkts() function 432 total += t->worker[i].processed_pkts; in processed_pkts() 500 int64_t remaining = t->outstand_pkts - processed_pkts(t); in perf_launch_lcores() 507 const uint64_t pkts = processed_pkts(t); in perf_launch_lcores() 543 remaining = t->outstand_pkts - processed_pkts(t); in perf_launch_lcores() 749 w->processed_pkts = 0; in perf_event_dev_port_setup()
|
| H A D | test_pipeline_common.h | 33 uint64_t processed_pkts; member
|
| /dpdk/drivers/baseband/null/ |
| H A D | bbdev_null.c | 50 struct rte_ring *processed_pkts; /* Ring for processed packets */ member 96 rte_ring_free(q->processed_pkts); in q_release() 124 q->processed_pkts = rte_ring_create(ring_name, queue_conf->queue_size, in q_setup() 126 if (q->processed_pkts == NULL) { in q_setup() 152 uint16_t nb_enqueued = rte_ring_enqueue_burst(q->processed_pkts, in enqueue_dec_ops() 167 uint16_t nb_enqueued = rte_ring_enqueue_burst(q->processed_pkts, in enqueue_enc_ops() 182 uint16_t nb_dequeued = rte_ring_dequeue_burst(q->processed_pkts, in dequeue_dec_ops() 195 uint16_t nb_dequeued = rte_ring_dequeue_burst(q->processed_pkts, in dequeue_enc_ops()
|
| /dpdk/drivers/compress/zlib/ |
| H A D | zlib_pmd_ops.c | 132 rte_ring_free(qp->processed_pkts); in zlib_pmd_qp_release() 159 struct rte_ring *r = qp->processed_pkts; in zlib_pmd_qp_create_processed_pkts_ring() 200 qp->processed_pkts = zlib_pmd_qp_create_processed_pkts_ring(qp, in zlib_pmd_qp_setup() 202 if (qp->processed_pkts == NULL) in zlib_pmd_qp_setup()
|
| H A D | zlib_pmd_private.h | 34 struct rte_ring *processed_pkts; member
|
| H A D | zlib_pmd.c | 215 return rte_ring_enqueue(qp->processed_pkts, (void *)op); in process_zlib_op() 347 nb_dequeued = rte_ring_dequeue_burst(qp->processed_pkts, in zlib_pmd_dequeue_burst()
|
| /dpdk/drivers/compress/isal/ |
| H A D | isal_compress_pmd_ops.c | 184 rte_ring_free(qp->processed_pkts); in isal_comp_pmd_qp_release() 290 qp->processed_pkts = isal_comp_pmd_qp_create_processed_pkts_ring(qp, in isal_comp_pmd_qp_setup() 292 if (qp->processed_pkts == NULL) { in isal_comp_pmd_qp_setup() 298 qp->num_free_elements = rte_ring_free_count(qp->processed_pkts); in isal_comp_pmd_qp_setup()
|
| H A D | isal_compress_pmd_private.h | 28 struct rte_ring *processed_pkts; member
|
| H A D | isal_compress_pmd.c | 655 retval = rte_ring_enqueue_burst(qp->processed_pkts, (void *)ops, in isal_comp_pmd_enqueue_burst() 671 nb_dequeued = rte_ring_dequeue_burst(qp->processed_pkts, (void **)ops, in isal_comp_pmd_dequeue_burst()
|
| /dpdk/drivers/compress/octeontx/ |
| H A D | otx_zip_pmd.c | 342 rte_ring_free(qp->processed_pkts); in zip_pmd_qp_release() 413 qp->processed_pkts = zip_pmd_qp_create_processed_pkts_ring(qp, in zip_pmd_qp_setup() 415 if (qp->processed_pkts == NULL) in zip_pmd_qp_setup() 431 rte_ring_free(qp->processed_pkts); in zip_pmd_qp_setup() 511 ret = rte_ring_enqueue(qp->processed_pkts, (void *)op); in zip_pmd_enqueue_burst_sync() 532 nb_dequeued = rte_ring_dequeue_burst(qp->processed_pkts, in zip_pmd_dequeue_burst_sync()
|
| H A D | otx_zip.h | 133 struct rte_ring *processed_pkts; member
|
| /dpdk/drivers/crypto/null/ |
| H A D | null_crypto_pmd_ops.c | 138 rte_ring_free(qp->processed_pkts); in null_crypto_pmd_qp_release() 227 qp->processed_pkts = null_crypto_pmd_qp_create_processed_pkts_ring(qp, in null_crypto_pmd_qp_setup() 229 if (qp->processed_pkts == NULL) { in null_crypto_pmd_qp_setup()
|
| H A D | null_crypto_pmd_private.h | 30 struct rte_ring *processed_pkts; member
|
| H A D | null_crypto_pmd.c | 69 return rte_ring_enqueue(qp->processed_pkts, (void *)op); in process_op() 150 nb_dequeued = rte_ring_dequeue_burst(qp->processed_pkts, in null_crypto_pmd_dequeue_burst()
|
| /dpdk/drivers/baseband/turbo_sw/ |
| H A D | bbdev_turbo_software.c | 78 struct rte_ring *processed_pkts; member 268 rte_ring_free(q->processed_pkts); in q_release() 457 q->processed_pkts = rte_ring_create(name, queue_conf->queue_size, in q_setup() 459 if (q->processed_pkts == NULL) { in q_setup() 472 rte_ring_free(q->processed_pkts); in q_setup() 1175 return rte_ring_enqueue_burst(q->processed_pkts, (void **)ops, nb_ops, in enqueue_enc_all_ops() 1192 return rte_ring_enqueue_burst(q->processed_pkts, (void **)ops, nb_ops, in enqueue_ldpc_enc_all_ops() 1753 return rte_ring_enqueue_burst(q->processed_pkts, (void **)ops, nb_ops, in enqueue_dec_all_ops() 1770 return rte_ring_enqueue_burst(q->processed_pkts, (void **)ops, nb_ops, in enqueue_ldpc_dec_all_ops() 1850 uint16_t nb_dequeued = rte_ring_dequeue_burst(q->processed_pkts, in dequeue_dec_ops() [all …]
|
| /dpdk/drivers/crypto/ccp/ |
| H A D | ccp_pmd_private.h | 77 struct rte_ring *processed_pkts; member
|
| H A D | ccp_pmd_ops.c | 639 rte_ring_free(qp->processed_pkts); in ccp_pmd_qp_release() 722 qp->processed_pkts = ccp_pmd_qp_create_batch_info_ring(qp, in ccp_pmd_qp_setup() 724 if (qp->processed_pkts == NULL) { in ccp_pmd_qp_setup()
|
| H A D | rte_ccp_pmd.c | 105 if (unlikely(rte_ring_full(qp->processed_pkts) != 0)) in ccp_pmd_enqueue_burst()
|
| H A D | ccp_crypto.c | 2845 rte_ring_enqueue(qp->processed_pkts, (void *)b_info); in process_ops_to_enqueue() 2985 } else if (rte_ring_dequeue(qp->processed_pkts, in process_ops_to_dequeue()
|