Lines Matching refs:ops_enq
3120 struct rte_bbdev_dec_op *ops_enq[num_ops]; in throughput_pmd_lcore_dec() local
3141 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops_enq, num_ops); in throughput_pmd_lcore_dec()
3145 copy_reference_dec_op(ops_enq, num_ops, 0, bufs->inputs, in throughput_pmd_lcore_dec()
3150 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in throughput_pmd_lcore_dec()
3155 mbuf_reset(ops_enq[j]->turbo_dec.hard_output.data); in throughput_pmd_lcore_dec()
3166 queue_id, &ops_enq[enq], num_to_enq); in throughput_pmd_lcore_dec()
3184 tp->iter_count = RTE_MAX(ops_enq[i]->turbo_dec.iter_count, in throughput_pmd_lcore_dec()
3194 rte_bbdev_dec_op_free_bulk(ops_enq, num_ops); in throughput_pmd_lcore_dec()
3216 struct rte_bbdev_dec_op *ops_enq[num_ops]; in bler_pmd_lcore_ldpc_dec() local
3244 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops_enq, num_ops); in bler_pmd_lcore_ldpc_dec()
3256 copy_reference_ldpc_dec_op(ops_enq, num_ops, 0, bufs->inputs, in bler_pmd_lcore_ldpc_dec()
3263 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in bler_pmd_lcore_ldpc_dec()
3269 ops_enq[j]->ldpc_dec.hard_output.data); in bler_pmd_lcore_ldpc_dec()
3272 ops_enq[j]->ldpc_dec.harq_combined_output.data); in bler_pmd_lcore_ldpc_dec()
3275 preload_harq_ddr(tp->dev_id, queue_id, ops_enq, in bler_pmd_lcore_ldpc_dec()
3286 queue_id, &ops_enq[enq], num_to_enq); in bler_pmd_lcore_ldpc_dec()
3305 tp->iter_count = RTE_MAX(ops_enq[i]->ldpc_dec.iter_count, in bler_pmd_lcore_ldpc_dec()
3307 tp->iter_average += (double) ops_enq[i]->ldpc_dec.iter_count; in bler_pmd_lcore_ldpc_dec()
3308 if (ops_enq[i]->status & (1 << RTE_BBDEV_SYNDROME_ERROR)) in bler_pmd_lcore_ldpc_dec()
3325 rte_bbdev_dec_op_free_bulk(ops_enq, num_ops); in bler_pmd_lcore_ldpc_dec()
3346 struct rte_bbdev_dec_op *ops_enq[num_ops]; in throughput_pmd_lcore_ldpc_dec() local
3373 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops_enq, num_ops); in throughput_pmd_lcore_ldpc_dec()
3385 copy_reference_ldpc_dec_op(ops_enq, num_ops, 0, bufs->inputs, in throughput_pmd_lcore_ldpc_dec()
3391 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in throughput_pmd_lcore_ldpc_dec()
3397 ops_enq[j]->ldpc_dec.hard_output.data); in throughput_pmd_lcore_ldpc_dec()
3400 ops_enq[j]->ldpc_dec.harq_combined_output.data); in throughput_pmd_lcore_ldpc_dec()
3403 preload_harq_ddr(tp->dev_id, queue_id, ops_enq, in throughput_pmd_lcore_ldpc_dec()
3414 queue_id, &ops_enq[enq], num_to_enq); in throughput_pmd_lcore_ldpc_dec()
3432 tp->iter_count = RTE_MAX(ops_enq[i]->ldpc_dec.iter_count, in throughput_pmd_lcore_ldpc_dec()
3437 retrieve_harq_ddr(tp->dev_id, queue_id, ops_enq, num_ops); in throughput_pmd_lcore_ldpc_dec()
3446 rte_bbdev_dec_op_free_bulk(ops_enq, num_ops); in throughput_pmd_lcore_ldpc_dec()
3468 struct rte_bbdev_enc_op *ops_enq[num_ops]; in throughput_pmd_lcore_enc() local
3489 ret = rte_bbdev_enc_op_alloc_bulk(tp->op_params->mp, ops_enq, in throughput_pmd_lcore_enc()
3494 copy_reference_enc_op(ops_enq, num_ops, 0, bufs->inputs, in throughput_pmd_lcore_enc()
3499 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in throughput_pmd_lcore_enc()
3505 mbuf_reset(ops_enq[j]->turbo_enc.output.data); in throughput_pmd_lcore_enc()
3516 queue_id, &ops_enq[enq], num_to_enq); in throughput_pmd_lcore_enc()
3536 rte_bbdev_enc_op_free_bulk(ops_enq, num_ops); in throughput_pmd_lcore_enc()
3558 struct rte_bbdev_enc_op *ops_enq[num_ops]; in throughput_pmd_lcore_ldpc_enc() local
3579 ret = rte_bbdev_enc_op_alloc_bulk(tp->op_params->mp, ops_enq, in throughput_pmd_lcore_ldpc_enc()
3584 copy_reference_ldpc_enc_op(ops_enq, num_ops, 0, bufs->inputs, in throughput_pmd_lcore_ldpc_enc()
3589 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in throughput_pmd_lcore_ldpc_enc()
3595 mbuf_reset(ops_enq[j]->turbo_enc.output.data); in throughput_pmd_lcore_ldpc_enc()
3606 queue_id, &ops_enq[enq], num_to_enq); in throughput_pmd_lcore_ldpc_enc()
3626 rte_bbdev_enc_op_free_bulk(ops_enq, num_ops); in throughput_pmd_lcore_ldpc_enc()
3979 struct rte_bbdev_dec_op *ops_enq[MAX_BURST], *ops_deq[MAX_BURST]; in latency_test_dec() local
3990 ret = rte_bbdev_dec_op_alloc_bulk(mempool, ops_enq, burst_sz); in latency_test_dec()
3994 copy_reference_dec_op(ops_enq, burst_sz, dequeued, in latency_test_dec()
4002 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in latency_test_dec()
4006 enq = rte_bbdev_enqueue_dec_ops(dev_id, queue_id, &ops_enq[enq], in latency_test_dec()
4032 rte_bbdev_dec_op_free_bulk(ops_enq, deq); in latency_test_dec()
4050 struct rte_bbdev_dec_op *ops_enq[MAX_BURST], *ops_deq[MAX_BURST]; in latency_test_ldpc_dec() local
4063 ret = rte_bbdev_dec_op_alloc_bulk(mempool, ops_enq, burst_sz); in latency_test_ldpc_dec()
4076 copy_reference_ldpc_dec_op(ops_enq, burst_sz, dequeued, in latency_test_ldpc_dec()
4085 preload_harq_ddr(dev_id, queue_id, ops_enq, in latency_test_ldpc_dec()
4090 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in latency_test_ldpc_dec()
4095 &ops_enq[enq], burst_sz); in latency_test_ldpc_dec()
4115 retrieve_harq_ddr(dev_id, queue_id, ops_enq, burst_sz); in latency_test_ldpc_dec()
4123 rte_bbdev_dec_op_free_bulk(ops_enq, deq); in latency_test_ldpc_dec()
4138 struct rte_bbdev_enc_op *ops_enq[MAX_BURST], *ops_deq[MAX_BURST]; in latency_test_enc() local
4149 ret = rte_bbdev_enc_op_alloc_bulk(mempool, ops_enq, burst_sz); in latency_test_enc()
4153 copy_reference_enc_op(ops_enq, burst_sz, dequeued, in latency_test_enc()
4160 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in latency_test_enc()
4164 enq = rte_bbdev_enqueue_enc_ops(dev_id, queue_id, &ops_enq[enq], in latency_test_enc()
4189 rte_bbdev_enc_op_free_bulk(ops_enq, deq); in latency_test_enc()
4205 struct rte_bbdev_enc_op *ops_enq[MAX_BURST], *ops_deq[MAX_BURST]; in latency_test_ldpc_enc() local
4216 ret = rte_bbdev_enc_op_alloc_bulk(mempool, ops_enq, burst_sz); in latency_test_ldpc_enc()
4220 copy_reference_ldpc_enc_op(ops_enq, burst_sz, dequeued, in latency_test_ldpc_enc()
4227 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in latency_test_ldpc_enc()
4232 &ops_enq[enq], burst_sz); in latency_test_ldpc_enc()
4256 rte_bbdev_enc_op_free_bulk(ops_enq, deq); in latency_test_ldpc_enc()
4379 struct rte_bbdev_dec_op *ops_enq[MAX_BURST], *ops_deq[MAX_BURST]; in offload_latency_test_dec() local
4390 rte_bbdev_dec_op_alloc_bulk(mempool, ops_enq, burst_sz); in offload_latency_test_dec()
4392 copy_reference_dec_op(ops_enq, burst_sz, dequeued, in offload_latency_test_dec()
4402 &ops_enq[enq], burst_sz - enq); in offload_latency_test_dec()
4447 rte_bbdev_dec_op_free_bulk(ops_enq, deq); in offload_latency_test_dec()
4462 struct rte_bbdev_dec_op *ops_enq[MAX_BURST], *ops_deq[MAX_BURST]; in offload_latency_test_ldpc_dec() local
4475 rte_bbdev_dec_op_alloc_bulk(mempool, ops_enq, burst_sz); in offload_latency_test_ldpc_dec()
4477 copy_reference_ldpc_dec_op(ops_enq, burst_sz, dequeued, in offload_latency_test_ldpc_dec()
4486 preload_harq_ddr(dev_id, queue_id, ops_enq, in offload_latency_test_ldpc_dec()
4493 &ops_enq[enq], burst_sz - enq); in offload_latency_test_ldpc_dec()
4540 retrieve_harq_ddr(dev_id, queue_id, ops_enq, burst_sz); in offload_latency_test_ldpc_dec()
4543 rte_bbdev_dec_op_free_bulk(ops_enq, deq); in offload_latency_test_ldpc_dec()
4557 struct rte_bbdev_enc_op *ops_enq[MAX_BURST], *ops_deq[MAX_BURST]; in offload_latency_test_enc() local
4568 ret = rte_bbdev_enc_op_alloc_bulk(mempool, ops_enq, burst_sz); in offload_latency_test_enc()
4572 copy_reference_enc_op(ops_enq, burst_sz, dequeued, in offload_latency_test_enc()
4581 &ops_enq[enq], burst_sz - enq); in offload_latency_test_enc()
4625 rte_bbdev_enc_op_free_bulk(ops_enq, deq); in offload_latency_test_enc()
4640 struct rte_bbdev_enc_op *ops_enq[MAX_BURST], *ops_deq[MAX_BURST]; in offload_latency_test_ldpc_enc() local
4651 ret = rte_bbdev_enc_op_alloc_bulk(mempool, ops_enq, burst_sz); in offload_latency_test_ldpc_enc()
4655 copy_reference_ldpc_enc_op(ops_enq, burst_sz, dequeued, in offload_latency_test_ldpc_enc()
4664 &ops_enq[enq], burst_sz - enq); in offload_latency_test_ldpc_enc()
4708 rte_bbdev_enc_op_free_bulk(ops_enq, deq); in offload_latency_test_ldpc_enc()