Lines Matching refs:ops_enq

3086 	struct rte_bbdev_dec_op *ops_enq[num_ops];  in throughput_pmd_lcore_dec()  local
3108 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops_enq, num_ops); in throughput_pmd_lcore_dec()
3112 copy_reference_dec_op(ops_enq, num_ops, 0, bufs->inputs, in throughput_pmd_lcore_dec()
3117 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in throughput_pmd_lcore_dec()
3122 mbuf_reset(ops_enq[j]->turbo_dec.hard_output.data); in throughput_pmd_lcore_dec()
3133 queue_id, &ops_enq[enq], num_to_enq); in throughput_pmd_lcore_dec()
3151 tp->iter_count = RTE_MAX(ops_enq[i]->turbo_dec.iter_count, in throughput_pmd_lcore_dec()
3161 rte_bbdev_dec_op_free_bulk(ops_enq, num_ops); in throughput_pmd_lcore_dec()
3183 struct rte_bbdev_dec_op *ops_enq[num_ops]; in bler_pmd_lcore_ldpc_dec() local
3212 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops_enq, num_ops); in bler_pmd_lcore_ldpc_dec()
3224 copy_reference_ldpc_dec_op(ops_enq, num_ops, 0, bufs->inputs, in bler_pmd_lcore_ldpc_dec()
3231 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in bler_pmd_lcore_ldpc_dec()
3237 ops_enq[j]->ldpc_dec.hard_output.data); in bler_pmd_lcore_ldpc_dec()
3240 ops_enq[j]->ldpc_dec.harq_combined_output.data); in bler_pmd_lcore_ldpc_dec()
3243 preload_harq_ddr(tp->dev_id, queue_id, ops_enq, in bler_pmd_lcore_ldpc_dec()
3254 queue_id, &ops_enq[enq], num_to_enq); in bler_pmd_lcore_ldpc_dec()
3273 tp->iter_count = RTE_MAX(ops_enq[i]->ldpc_dec.iter_count, in bler_pmd_lcore_ldpc_dec()
3275 tp->iter_average += (double) ops_enq[i]->ldpc_dec.iter_count; in bler_pmd_lcore_ldpc_dec()
3276 if (ops_enq[i]->status & (1 << RTE_BBDEV_SYNDROME_ERROR)) in bler_pmd_lcore_ldpc_dec()
3293 rte_bbdev_dec_op_free_bulk(ops_enq, num_ops); in bler_pmd_lcore_ldpc_dec()
3314 struct rte_bbdev_dec_op *ops_enq[num_ops]; in throughput_pmd_lcore_ldpc_dec() local
3342 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops_enq, num_ops); in throughput_pmd_lcore_ldpc_dec()
3354 copy_reference_ldpc_dec_op(ops_enq, num_ops, 0, bufs->inputs, in throughput_pmd_lcore_ldpc_dec()
3360 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in throughput_pmd_lcore_ldpc_dec()
3366 ops_enq[j]->ldpc_dec.hard_output.data); in throughput_pmd_lcore_ldpc_dec()
3369 ops_enq[j]->ldpc_dec.harq_combined_output.data); in throughput_pmd_lcore_ldpc_dec()
3372 preload_harq_ddr(tp->dev_id, queue_id, ops_enq, in throughput_pmd_lcore_ldpc_dec()
3383 queue_id, &ops_enq[enq], num_to_enq); in throughput_pmd_lcore_ldpc_dec()
3401 tp->iter_count = RTE_MAX(ops_enq[i]->ldpc_dec.iter_count, in throughput_pmd_lcore_ldpc_dec()
3406 retrieve_harq_ddr(tp->dev_id, queue_id, ops_enq, num_ops); in throughput_pmd_lcore_ldpc_dec()
3415 rte_bbdev_dec_op_free_bulk(ops_enq, num_ops); in throughput_pmd_lcore_ldpc_dec()
3437 struct rte_bbdev_enc_op *ops_enq[num_ops]; in throughput_pmd_lcore_enc() local
3459 ret = rte_bbdev_enc_op_alloc_bulk(tp->op_params->mp, ops_enq, in throughput_pmd_lcore_enc()
3464 copy_reference_enc_op(ops_enq, num_ops, 0, bufs->inputs, in throughput_pmd_lcore_enc()
3469 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in throughput_pmd_lcore_enc()
3475 mbuf_reset(ops_enq[j]->turbo_enc.output.data); in throughput_pmd_lcore_enc()
3486 queue_id, &ops_enq[enq], num_to_enq); in throughput_pmd_lcore_enc()
3506 rte_bbdev_enc_op_free_bulk(ops_enq, num_ops); in throughput_pmd_lcore_enc()
3528 struct rte_bbdev_enc_op *ops_enq[num_ops]; in throughput_pmd_lcore_ldpc_enc() local
3550 ret = rte_bbdev_enc_op_alloc_bulk(tp->op_params->mp, ops_enq, in throughput_pmd_lcore_ldpc_enc()
3555 copy_reference_ldpc_enc_op(ops_enq, num_ops, 0, bufs->inputs, in throughput_pmd_lcore_ldpc_enc()
3560 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in throughput_pmd_lcore_ldpc_enc()
3566 mbuf_reset(ops_enq[j]->turbo_enc.output.data); in throughput_pmd_lcore_ldpc_enc()
3577 queue_id, &ops_enq[enq], num_to_enq); in throughput_pmd_lcore_ldpc_enc()
3597 rte_bbdev_enc_op_free_bulk(ops_enq, num_ops); in throughput_pmd_lcore_ldpc_enc()
3950 struct rte_bbdev_dec_op *ops_enq[MAX_BURST], *ops_deq[MAX_BURST]; in latency_test_dec() local
3961 ret = rte_bbdev_dec_op_alloc_bulk(mempool, ops_enq, burst_sz); in latency_test_dec()
3965 copy_reference_dec_op(ops_enq, burst_sz, dequeued, in latency_test_dec()
3973 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in latency_test_dec()
3977 enq = rte_bbdev_enqueue_dec_ops(dev_id, queue_id, &ops_enq[enq], in latency_test_dec()
4003 rte_bbdev_dec_op_free_bulk(ops_enq, deq); in latency_test_dec()
4021 struct rte_bbdev_dec_op *ops_enq[MAX_BURST], *ops_deq[MAX_BURST]; in latency_test_ldpc_dec() local
4034 ret = rte_bbdev_dec_op_alloc_bulk(mempool, ops_enq, burst_sz); in latency_test_ldpc_dec()
4047 copy_reference_ldpc_dec_op(ops_enq, burst_sz, dequeued, in latency_test_ldpc_dec()
4056 preload_harq_ddr(dev_id, queue_id, ops_enq, in latency_test_ldpc_dec()
4061 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in latency_test_ldpc_dec()
4066 &ops_enq[enq], burst_sz); in latency_test_ldpc_dec()
4086 retrieve_harq_ddr(dev_id, queue_id, ops_enq, burst_sz); in latency_test_ldpc_dec()
4094 rte_bbdev_dec_op_free_bulk(ops_enq, deq); in latency_test_ldpc_dec()
4109 struct rte_bbdev_enc_op *ops_enq[MAX_BURST], *ops_deq[MAX_BURST]; in latency_test_enc() local
4120 ret = rte_bbdev_enc_op_alloc_bulk(mempool, ops_enq, burst_sz); in latency_test_enc()
4124 copy_reference_enc_op(ops_enq, burst_sz, dequeued, in latency_test_enc()
4131 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in latency_test_enc()
4135 enq = rte_bbdev_enqueue_enc_ops(dev_id, queue_id, &ops_enq[enq], in latency_test_enc()
4160 rte_bbdev_enc_op_free_bulk(ops_enq, deq); in latency_test_enc()
4176 struct rte_bbdev_enc_op *ops_enq[MAX_BURST], *ops_deq[MAX_BURST]; in latency_test_ldpc_enc() local
4187 ret = rte_bbdev_enc_op_alloc_bulk(mempool, ops_enq, burst_sz); in latency_test_ldpc_enc()
4191 copy_reference_ldpc_enc_op(ops_enq, burst_sz, dequeued, in latency_test_ldpc_enc()
4198 ops_enq[j]->opaque_data = (void *)(uintptr_t)j; in latency_test_ldpc_enc()
4203 &ops_enq[enq], burst_sz); in latency_test_ldpc_enc()
4227 rte_bbdev_enc_op_free_bulk(ops_enq, deq); in latency_test_ldpc_enc()
4350 struct rte_bbdev_dec_op *ops_enq[MAX_BURST], *ops_deq[MAX_BURST]; in offload_latency_test_dec() local
4361 rte_bbdev_dec_op_alloc_bulk(mempool, ops_enq, burst_sz); in offload_latency_test_dec()
4363 copy_reference_dec_op(ops_enq, burst_sz, dequeued, in offload_latency_test_dec()
4373 &ops_enq[enq], burst_sz - enq); in offload_latency_test_dec()
4418 rte_bbdev_dec_op_free_bulk(ops_enq, deq); in offload_latency_test_dec()
4433 struct rte_bbdev_dec_op *ops_enq[MAX_BURST], *ops_deq[MAX_BURST]; in offload_latency_test_ldpc_dec() local
4446 rte_bbdev_dec_op_alloc_bulk(mempool, ops_enq, burst_sz); in offload_latency_test_ldpc_dec()
4448 copy_reference_ldpc_dec_op(ops_enq, burst_sz, dequeued, in offload_latency_test_ldpc_dec()
4457 preload_harq_ddr(dev_id, queue_id, ops_enq, in offload_latency_test_ldpc_dec()
4464 &ops_enq[enq], burst_sz - enq); in offload_latency_test_ldpc_dec()
4511 retrieve_harq_ddr(dev_id, queue_id, ops_enq, burst_sz); in offload_latency_test_ldpc_dec()
4514 rte_bbdev_dec_op_free_bulk(ops_enq, deq); in offload_latency_test_ldpc_dec()
4528 struct rte_bbdev_enc_op *ops_enq[MAX_BURST], *ops_deq[MAX_BURST]; in offload_latency_test_enc() local
4539 ret = rte_bbdev_enc_op_alloc_bulk(mempool, ops_enq, burst_sz); in offload_latency_test_enc()
4543 copy_reference_enc_op(ops_enq, burst_sz, dequeued, in offload_latency_test_enc()
4552 &ops_enq[enq], burst_sz - enq); in offload_latency_test_enc()
4596 rte_bbdev_enc_op_free_bulk(ops_enq, deq); in offload_latency_test_enc()
4611 struct rte_bbdev_enc_op *ops_enq[MAX_BURST], *ops_deq[MAX_BURST]; in offload_latency_test_ldpc_enc() local
4622 ret = rte_bbdev_enc_op_alloc_bulk(mempool, ops_enq, burst_sz); in offload_latency_test_ldpc_enc()
4626 copy_reference_ldpc_enc_op(ops_enq, burst_sz, dequeued, in offload_latency_test_ldpc_enc()
4635 &ops_enq[enq], burst_sz - enq); in offload_latency_test_ldpc_enc()
4679 rte_bbdev_enc_op_free_bulk(ops_enq, deq); in offload_latency_test_ldpc_enc()