Lines Matching refs:num_ops
485 enum rte_bbdev_op_type org_op_type, uint16_t num_ops) in create_mempools() argument
506 RTE_MAX((unsigned int)(ad->nb_queues * num_ops + 1), in create_mempools()
2622 uint16_t deq, burst_sz, num_ops; in dequeue_event_callback() local
2647 num_ops = tp->op_params->num_to_process; in dequeue_event_callback()
2678 if (__atomic_load_n(&tp->nb_dequeued, __ATOMIC_RELAXED) + deq < num_ops) { in dequeue_event_callback()
2691 ret = validate_dec_op(tp->dec_ops, num_ops, ref_op, in dequeue_event_callback()
2694 for (i = 0; i < num_ops; ++i) in dequeue_event_callback()
2701 ret = validate_enc_op(tp->enc_ops, num_ops, ref_op); in dequeue_event_callback()
2705 ret = validate_ldpc_enc_op(tp->enc_ops, num_ops, ref_op); in dequeue_event_callback()
2709 ret = validate_ldpc_dec_op(tp->dec_ops, num_ops, ref_op, in dequeue_event_callback()
2741 tp->ops_per_sec += ((double)num_ops) / in dequeue_event_callback()
2743 tp->mbps += (((double)(num_ops * tb_len_bits)) / 1000000.0) / in dequeue_event_callback()
3119 const uint16_t num_ops = tp->op_params->num_to_process; in throughput_pmd_lcore_dec() local
3120 struct rte_bbdev_dec_op *ops_enq[num_ops]; in throughput_pmd_lcore_dec()
3121 struct rte_bbdev_dec_op *ops_deq[num_ops]; in throughput_pmd_lcore_dec()
3133 TEST_ASSERT_SUCCESS((num_ops > info.drv.queue_size_lim), in throughput_pmd_lcore_dec()
3141 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops_enq, num_ops); in throughput_pmd_lcore_dec()
3142 TEST_ASSERT_SUCCESS(ret, "Allocation failed for %d ops", num_ops); in throughput_pmd_lcore_dec()
3145 copy_reference_dec_op(ops_enq, num_ops, 0, bufs->inputs, in throughput_pmd_lcore_dec()
3149 for (j = 0; j < num_ops; ++j) in throughput_pmd_lcore_dec()
3154 for (j = 0; j < num_ops; ++j) in throughput_pmd_lcore_dec()
3159 for (enq = 0, deq = 0; enq < num_ops;) { in throughput_pmd_lcore_dec()
3162 if (unlikely(num_ops - enq < num_to_enq)) in throughput_pmd_lcore_dec()
3163 num_to_enq = num_ops - enq; in throughput_pmd_lcore_dec()
3183 for (i = 0; i < num_ops; ++i) { in throughput_pmd_lcore_dec()
3189 ret = validate_dec_op(ops_deq, num_ops, ref_op, in throughput_pmd_lcore_dec()
3194 rte_bbdev_dec_op_free_bulk(ops_enq, num_ops); in throughput_pmd_lcore_dec()
3198 tp->ops_per_sec = ((double)num_ops * TEST_REPETITIONS) / in throughput_pmd_lcore_dec()
3200 tp->mbps = (((double)(num_ops * TEST_REPETITIONS * tb_len_bits)) / in throughput_pmd_lcore_dec()
3215 const uint16_t num_ops = tp->op_params->num_to_process; in bler_pmd_lcore_ldpc_dec() local
3216 struct rte_bbdev_dec_op *ops_enq[num_ops]; in bler_pmd_lcore_ldpc_dec()
3217 struct rte_bbdev_dec_op *ops_deq[num_ops]; in bler_pmd_lcore_ldpc_dec()
3236 TEST_ASSERT_SUCCESS((num_ops > info.drv.queue_size_lim), in bler_pmd_lcore_ldpc_dec()
3244 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops_enq, num_ops); in bler_pmd_lcore_ldpc_dec()
3245 TEST_ASSERT_SUCCESS(ret, "Allocation failed for %d ops", num_ops); in bler_pmd_lcore_ldpc_dec()
3256 copy_reference_ldpc_dec_op(ops_enq, num_ops, 0, bufs->inputs, in bler_pmd_lcore_ldpc_dec()
3259 generate_llr_input(num_ops, bufs->inputs, ref_op); in bler_pmd_lcore_ldpc_dec()
3262 for (j = 0; j < num_ops; ++j) in bler_pmd_lcore_ldpc_dec()
3266 for (j = 0; j < num_ops; ++j) { in bler_pmd_lcore_ldpc_dec()
3276 num_ops, true); in bler_pmd_lcore_ldpc_dec()
3279 for (enq = 0, deq = 0; enq < num_ops;) { in bler_pmd_lcore_ldpc_dec()
3282 if (unlikely(num_ops - enq < num_to_enq)) in bler_pmd_lcore_ldpc_dec()
3283 num_to_enq = num_ops - enq; in bler_pmd_lcore_ldpc_dec()
3304 for (i = 0; i < num_ops; ++i) { in bler_pmd_lcore_ldpc_dec()
3312 parity_bler /= num_ops; /* This one is based on SYND */ in bler_pmd_lcore_ldpc_dec()
3313 tp->iter_average /= num_ops; in bler_pmd_lcore_ldpc_dec()
3314 tp->bler = (double) validate_ldpc_bler(ops_deq, num_ops) / num_ops; in bler_pmd_lcore_ldpc_dec()
3320 ret = validate_ldpc_dec_op(ops_deq, num_ops, ref_op, in bler_pmd_lcore_ldpc_dec()
3325 rte_bbdev_dec_op_free_bulk(ops_enq, num_ops); in bler_pmd_lcore_ldpc_dec()
3328 tp->ops_per_sec = ((double)num_ops * 1) / in bler_pmd_lcore_ldpc_dec()
3330 tp->mbps = (((double)(num_ops * 1 * tb_len_bits)) / in bler_pmd_lcore_ldpc_dec()
3345 const uint16_t num_ops = tp->op_params->num_to_process; in throughput_pmd_lcore_ldpc_dec() local
3346 struct rte_bbdev_dec_op *ops_enq[num_ops]; in throughput_pmd_lcore_ldpc_dec()
3347 struct rte_bbdev_dec_op *ops_deq[num_ops]; in throughput_pmd_lcore_ldpc_dec()
3365 TEST_ASSERT_SUCCESS((num_ops > info.drv.queue_size_lim), in throughput_pmd_lcore_ldpc_dec()
3373 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops_enq, num_ops); in throughput_pmd_lcore_ldpc_dec()
3374 TEST_ASSERT_SUCCESS(ret, "Allocation failed for %d ops", num_ops); in throughput_pmd_lcore_ldpc_dec()
3385 copy_reference_ldpc_dec_op(ops_enq, num_ops, 0, bufs->inputs, in throughput_pmd_lcore_ldpc_dec()
3390 for (j = 0; j < num_ops; ++j) in throughput_pmd_lcore_ldpc_dec()
3394 for (j = 0; j < num_ops; ++j) { in throughput_pmd_lcore_ldpc_dec()
3404 num_ops, true); in throughput_pmd_lcore_ldpc_dec()
3407 for (enq = 0, deq = 0; enq < num_ops;) { in throughput_pmd_lcore_ldpc_dec()
3410 if (unlikely(num_ops - enq < num_to_enq)) in throughput_pmd_lcore_ldpc_dec()
3411 num_to_enq = num_ops - enq; in throughput_pmd_lcore_ldpc_dec()
3431 for (i = 0; i < num_ops; ++i) { in throughput_pmd_lcore_ldpc_dec()
3437 retrieve_harq_ddr(tp->dev_id, queue_id, ops_enq, num_ops); in throughput_pmd_lcore_ldpc_dec()
3441 ret = validate_ldpc_dec_op(ops_deq, num_ops, ref_op, in throughput_pmd_lcore_ldpc_dec()
3446 rte_bbdev_dec_op_free_bulk(ops_enq, num_ops); in throughput_pmd_lcore_ldpc_dec()
3450 tp->ops_per_sec = ((double)num_ops * TEST_REPETITIONS) / in throughput_pmd_lcore_ldpc_dec()
3452 tp->mbps = (((double)(num_ops * TEST_REPETITIONS * tb_len_bits)) / in throughput_pmd_lcore_ldpc_dec()
3467 const uint16_t num_ops = tp->op_params->num_to_process; in throughput_pmd_lcore_enc() local
3468 struct rte_bbdev_enc_op *ops_enq[num_ops]; in throughput_pmd_lcore_enc()
3469 struct rte_bbdev_enc_op *ops_deq[num_ops]; in throughput_pmd_lcore_enc()
3481 TEST_ASSERT_SUCCESS((num_ops > info.drv.queue_size_lim), in throughput_pmd_lcore_enc()
3490 num_ops); in throughput_pmd_lcore_enc()
3492 num_ops); in throughput_pmd_lcore_enc()
3494 copy_reference_enc_op(ops_enq, num_ops, 0, bufs->inputs, in throughput_pmd_lcore_enc()
3498 for (j = 0; j < num_ops; ++j) in throughput_pmd_lcore_enc()
3504 for (j = 0; j < num_ops; ++j) in throughput_pmd_lcore_enc()
3509 for (enq = 0, deq = 0; enq < num_ops;) { in throughput_pmd_lcore_enc()
3512 if (unlikely(num_ops - enq < num_to_enq)) in throughput_pmd_lcore_enc()
3513 num_to_enq = num_ops - enq; in throughput_pmd_lcore_enc()
3532 ret = validate_enc_op(ops_deq, num_ops, ref_op); in throughput_pmd_lcore_enc()
3536 rte_bbdev_enc_op_free_bulk(ops_enq, num_ops); in throughput_pmd_lcore_enc()
3540 tp->ops_per_sec = ((double)num_ops * TEST_REPETITIONS) / in throughput_pmd_lcore_enc()
3542 tp->mbps = (((double)(num_ops * TEST_REPETITIONS * tb_len_bits)) in throughput_pmd_lcore_enc()
3557 const uint16_t num_ops = tp->op_params->num_to_process; in throughput_pmd_lcore_ldpc_enc() local
3558 struct rte_bbdev_enc_op *ops_enq[num_ops]; in throughput_pmd_lcore_ldpc_enc()
3559 struct rte_bbdev_enc_op *ops_deq[num_ops]; in throughput_pmd_lcore_ldpc_enc()
3571 TEST_ASSERT_SUCCESS((num_ops > info.drv.queue_size_lim), in throughput_pmd_lcore_ldpc_enc()
3580 num_ops); in throughput_pmd_lcore_ldpc_enc()
3582 num_ops); in throughput_pmd_lcore_ldpc_enc()
3584 copy_reference_ldpc_enc_op(ops_enq, num_ops, 0, bufs->inputs, in throughput_pmd_lcore_ldpc_enc()
3588 for (j = 0; j < num_ops; ++j) in throughput_pmd_lcore_ldpc_enc()
3594 for (j = 0; j < num_ops; ++j) in throughput_pmd_lcore_ldpc_enc()
3599 for (enq = 0, deq = 0; enq < num_ops;) { in throughput_pmd_lcore_ldpc_enc()
3602 if (unlikely(num_ops - enq < num_to_enq)) in throughput_pmd_lcore_ldpc_enc()
3603 num_to_enq = num_ops - enq; in throughput_pmd_lcore_ldpc_enc()
3622 ret = validate_ldpc_enc_op(ops_deq, num_ops, ref_op); in throughput_pmd_lcore_ldpc_enc()
3626 rte_bbdev_enc_op_free_bulk(ops_enq, num_ops); in throughput_pmd_lcore_ldpc_enc()
3630 tp->ops_per_sec = ((double)num_ops * TEST_REPETITIONS) / in throughput_pmd_lcore_ldpc_enc()
3632 tp->mbps = (((double)(num_ops * TEST_REPETITIONS * tb_len_bits)) in throughput_pmd_lcore_ldpc_enc()