Lines Matching refs:num_ops

442 		enum rte_bbdev_op_type org_op_type, uint16_t num_ops)  in create_mempools()  argument
463 RTE_MAX((unsigned int)(ad->nb_queues * num_ops + 1), in create_mempools()
2576 uint16_t deq, burst_sz, num_ops; in dequeue_event_callback() local
2601 num_ops = tp->op_params->num_to_process; in dequeue_event_callback()
2632 if (rte_atomic16_read(&tp->nb_dequeued) + deq < num_ops) { in dequeue_event_callback()
2645 ret = validate_dec_op(tp->dec_ops, num_ops, ref_op, in dequeue_event_callback()
2648 for (i = 0; i < num_ops; ++i) in dequeue_event_callback()
2655 ret = validate_enc_op(tp->enc_ops, num_ops, ref_op); in dequeue_event_callback()
2659 ret = validate_ldpc_enc_op(tp->enc_ops, num_ops, ref_op); in dequeue_event_callback()
2663 ret = validate_ldpc_dec_op(tp->dec_ops, num_ops, ref_op, in dequeue_event_callback()
2695 tp->ops_per_sec += ((double)num_ops) / in dequeue_event_callback()
2697 tp->mbps += (((double)(num_ops * tb_len_bits)) / 1000000.0) / in dequeue_event_callback()
3085 const uint16_t num_ops = tp->op_params->num_to_process; in throughput_pmd_lcore_dec() local
3086 struct rte_bbdev_dec_op *ops_enq[num_ops]; in throughput_pmd_lcore_dec()
3087 struct rte_bbdev_dec_op *ops_deq[num_ops]; in throughput_pmd_lcore_dec()
3099 TEST_ASSERT_SUCCESS((num_ops > info.drv.queue_size_lim), in throughput_pmd_lcore_dec()
3108 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops_enq, num_ops); in throughput_pmd_lcore_dec()
3109 TEST_ASSERT_SUCCESS(ret, "Allocation failed for %d ops", num_ops); in throughput_pmd_lcore_dec()
3112 copy_reference_dec_op(ops_enq, num_ops, 0, bufs->inputs, in throughput_pmd_lcore_dec()
3116 for (j = 0; j < num_ops; ++j) in throughput_pmd_lcore_dec()
3121 for (j = 0; j < num_ops; ++j) in throughput_pmd_lcore_dec()
3126 for (enq = 0, deq = 0; enq < num_ops;) { in throughput_pmd_lcore_dec()
3129 if (unlikely(num_ops - enq < num_to_enq)) in throughput_pmd_lcore_dec()
3130 num_to_enq = num_ops - enq; in throughput_pmd_lcore_dec()
3150 for (i = 0; i < num_ops; ++i) { in throughput_pmd_lcore_dec()
3156 ret = validate_dec_op(ops_deq, num_ops, ref_op, in throughput_pmd_lcore_dec()
3161 rte_bbdev_dec_op_free_bulk(ops_enq, num_ops); in throughput_pmd_lcore_dec()
3165 tp->ops_per_sec = ((double)num_ops * TEST_REPETITIONS) / in throughput_pmd_lcore_dec()
3167 tp->mbps = (((double)(num_ops * TEST_REPETITIONS * tb_len_bits)) / in throughput_pmd_lcore_dec()
3182 const uint16_t num_ops = tp->op_params->num_to_process; in bler_pmd_lcore_ldpc_dec() local
3183 struct rte_bbdev_dec_op *ops_enq[num_ops]; in bler_pmd_lcore_ldpc_dec()
3184 struct rte_bbdev_dec_op *ops_deq[num_ops]; in bler_pmd_lcore_ldpc_dec()
3203 TEST_ASSERT_SUCCESS((num_ops > info.drv.queue_size_lim), in bler_pmd_lcore_ldpc_dec()
3212 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops_enq, num_ops); in bler_pmd_lcore_ldpc_dec()
3213 TEST_ASSERT_SUCCESS(ret, "Allocation failed for %d ops", num_ops); in bler_pmd_lcore_ldpc_dec()
3224 copy_reference_ldpc_dec_op(ops_enq, num_ops, 0, bufs->inputs, in bler_pmd_lcore_ldpc_dec()
3227 generate_llr_input(num_ops, bufs->inputs, ref_op); in bler_pmd_lcore_ldpc_dec()
3230 for (j = 0; j < num_ops; ++j) in bler_pmd_lcore_ldpc_dec()
3234 for (j = 0; j < num_ops; ++j) { in bler_pmd_lcore_ldpc_dec()
3244 num_ops, true); in bler_pmd_lcore_ldpc_dec()
3247 for (enq = 0, deq = 0; enq < num_ops;) { in bler_pmd_lcore_ldpc_dec()
3250 if (unlikely(num_ops - enq < num_to_enq)) in bler_pmd_lcore_ldpc_dec()
3251 num_to_enq = num_ops - enq; in bler_pmd_lcore_ldpc_dec()
3272 for (i = 0; i < num_ops; ++i) { in bler_pmd_lcore_ldpc_dec()
3280 parity_bler /= num_ops; /* This one is based on SYND */ in bler_pmd_lcore_ldpc_dec()
3281 tp->iter_average /= num_ops; in bler_pmd_lcore_ldpc_dec()
3282 tp->bler = (double) validate_ldpc_bler(ops_deq, num_ops) / num_ops; in bler_pmd_lcore_ldpc_dec()
3288 ret = validate_ldpc_dec_op(ops_deq, num_ops, ref_op, in bler_pmd_lcore_ldpc_dec()
3293 rte_bbdev_dec_op_free_bulk(ops_enq, num_ops); in bler_pmd_lcore_ldpc_dec()
3296 tp->ops_per_sec = ((double)num_ops * 1) / in bler_pmd_lcore_ldpc_dec()
3298 tp->mbps = (((double)(num_ops * 1 * tb_len_bits)) / in bler_pmd_lcore_ldpc_dec()
3313 const uint16_t num_ops = tp->op_params->num_to_process; in throughput_pmd_lcore_ldpc_dec() local
3314 struct rte_bbdev_dec_op *ops_enq[num_ops]; in throughput_pmd_lcore_ldpc_dec()
3315 struct rte_bbdev_dec_op *ops_deq[num_ops]; in throughput_pmd_lcore_ldpc_dec()
3333 TEST_ASSERT_SUCCESS((num_ops > info.drv.queue_size_lim), in throughput_pmd_lcore_ldpc_dec()
3342 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops_enq, num_ops); in throughput_pmd_lcore_ldpc_dec()
3343 TEST_ASSERT_SUCCESS(ret, "Allocation failed for %d ops", num_ops); in throughput_pmd_lcore_ldpc_dec()
3354 copy_reference_ldpc_dec_op(ops_enq, num_ops, 0, bufs->inputs, in throughput_pmd_lcore_ldpc_dec()
3359 for (j = 0; j < num_ops; ++j) in throughput_pmd_lcore_ldpc_dec()
3363 for (j = 0; j < num_ops; ++j) { in throughput_pmd_lcore_ldpc_dec()
3373 num_ops, true); in throughput_pmd_lcore_ldpc_dec()
3376 for (enq = 0, deq = 0; enq < num_ops;) { in throughput_pmd_lcore_ldpc_dec()
3379 if (unlikely(num_ops - enq < num_to_enq)) in throughput_pmd_lcore_ldpc_dec()
3380 num_to_enq = num_ops - enq; in throughput_pmd_lcore_ldpc_dec()
3400 for (i = 0; i < num_ops; ++i) { in throughput_pmd_lcore_ldpc_dec()
3406 retrieve_harq_ddr(tp->dev_id, queue_id, ops_enq, num_ops); in throughput_pmd_lcore_ldpc_dec()
3410 ret = validate_ldpc_dec_op(ops_deq, num_ops, ref_op, in throughput_pmd_lcore_ldpc_dec()
3415 rte_bbdev_dec_op_free_bulk(ops_enq, num_ops); in throughput_pmd_lcore_ldpc_dec()
3419 tp->ops_per_sec = ((double)num_ops * TEST_REPETITIONS) / in throughput_pmd_lcore_ldpc_dec()
3421 tp->mbps = (((double)(num_ops * TEST_REPETITIONS * tb_len_bits)) / in throughput_pmd_lcore_ldpc_dec()
3436 const uint16_t num_ops = tp->op_params->num_to_process; in throughput_pmd_lcore_enc() local
3437 struct rte_bbdev_enc_op *ops_enq[num_ops]; in throughput_pmd_lcore_enc()
3438 struct rte_bbdev_enc_op *ops_deq[num_ops]; in throughput_pmd_lcore_enc()
3450 TEST_ASSERT_SUCCESS((num_ops > info.drv.queue_size_lim), in throughput_pmd_lcore_enc()
3460 num_ops); in throughput_pmd_lcore_enc()
3462 num_ops); in throughput_pmd_lcore_enc()
3464 copy_reference_enc_op(ops_enq, num_ops, 0, bufs->inputs, in throughput_pmd_lcore_enc()
3468 for (j = 0; j < num_ops; ++j) in throughput_pmd_lcore_enc()
3474 for (j = 0; j < num_ops; ++j) in throughput_pmd_lcore_enc()
3479 for (enq = 0, deq = 0; enq < num_ops;) { in throughput_pmd_lcore_enc()
3482 if (unlikely(num_ops - enq < num_to_enq)) in throughput_pmd_lcore_enc()
3483 num_to_enq = num_ops - enq; in throughput_pmd_lcore_enc()
3502 ret = validate_enc_op(ops_deq, num_ops, ref_op); in throughput_pmd_lcore_enc()
3506 rte_bbdev_enc_op_free_bulk(ops_enq, num_ops); in throughput_pmd_lcore_enc()
3510 tp->ops_per_sec = ((double)num_ops * TEST_REPETITIONS) / in throughput_pmd_lcore_enc()
3512 tp->mbps = (((double)(num_ops * TEST_REPETITIONS * tb_len_bits)) in throughput_pmd_lcore_enc()
3527 const uint16_t num_ops = tp->op_params->num_to_process; in throughput_pmd_lcore_ldpc_enc() local
3528 struct rte_bbdev_enc_op *ops_enq[num_ops]; in throughput_pmd_lcore_ldpc_enc()
3529 struct rte_bbdev_enc_op *ops_deq[num_ops]; in throughput_pmd_lcore_ldpc_enc()
3541 TEST_ASSERT_SUCCESS((num_ops > info.drv.queue_size_lim), in throughput_pmd_lcore_ldpc_enc()
3551 num_ops); in throughput_pmd_lcore_ldpc_enc()
3553 num_ops); in throughput_pmd_lcore_ldpc_enc()
3555 copy_reference_ldpc_enc_op(ops_enq, num_ops, 0, bufs->inputs, in throughput_pmd_lcore_ldpc_enc()
3559 for (j = 0; j < num_ops; ++j) in throughput_pmd_lcore_ldpc_enc()
3565 for (j = 0; j < num_ops; ++j) in throughput_pmd_lcore_ldpc_enc()
3570 for (enq = 0, deq = 0; enq < num_ops;) { in throughput_pmd_lcore_ldpc_enc()
3573 if (unlikely(num_ops - enq < num_to_enq)) in throughput_pmd_lcore_ldpc_enc()
3574 num_to_enq = num_ops - enq; in throughput_pmd_lcore_ldpc_enc()
3593 ret = validate_ldpc_enc_op(ops_deq, num_ops, ref_op); in throughput_pmd_lcore_ldpc_enc()
3597 rte_bbdev_enc_op_free_bulk(ops_enq, num_ops); in throughput_pmd_lcore_ldpc_enc()
3601 tp->ops_per_sec = ((double)num_ops * TEST_REPETITIONS) / in throughput_pmd_lcore_ldpc_enc()
3603 tp->mbps = (((double)(num_ops * TEST_REPETITIONS * tb_len_bits)) in throughput_pmd_lcore_ldpc_enc()