Lines Matching refs:tp
2626 struct thread_params *tp = cb_arg; in dequeue_event_callback() local
2629 for (i = 0; i < MAX_QUEUES; ++i, ++tp) in dequeue_event_callback()
2630 if (tp->queue_id == queue_id) in dequeue_event_callback()
2640 __atomic_store_n(&tp->processing_status, TEST_FAILED, __ATOMIC_RELAXED); in dequeue_event_callback()
2646 burst_sz = __atomic_load_n(&tp->burst_sz, __ATOMIC_RELAXED); in dequeue_event_callback()
2647 num_ops = tp->op_params->num_to_process; in dequeue_event_callback()
2651 &tp->dec_ops[ in dequeue_event_callback()
2652 __atomic_load_n(&tp->nb_dequeued, __ATOMIC_RELAXED)], in dequeue_event_callback()
2656 &tp->dec_ops[ in dequeue_event_callback()
2657 __atomic_load_n(&tp->nb_dequeued, __ATOMIC_RELAXED)], in dequeue_event_callback()
2661 &tp->enc_ops[ in dequeue_event_callback()
2662 __atomic_load_n(&tp->nb_dequeued, __ATOMIC_RELAXED)], in dequeue_event_callback()
2666 &tp->enc_ops[ in dequeue_event_callback()
2667 __atomic_load_n(&tp->nb_dequeued, __ATOMIC_RELAXED)], in dequeue_event_callback()
2674 __atomic_store_n(&tp->processing_status, TEST_FAILED, __ATOMIC_RELAXED); in dequeue_event_callback()
2678 if (__atomic_load_n(&tp->nb_dequeued, __ATOMIC_RELAXED) + deq < num_ops) { in dequeue_event_callback()
2679 __atomic_fetch_add(&tp->nb_dequeued, deq, __ATOMIC_RELAXED); in dequeue_event_callback()
2683 total_time = rte_rdtsc_precise() - tp->start_time; in dequeue_event_callback()
2690 struct rte_bbdev_dec_op *ref_op = tp->op_params->ref_dec_op; in dequeue_event_callback()
2691 ret = validate_dec_op(tp->dec_ops, num_ops, ref_op, in dequeue_event_callback()
2692 tp->op_params->vector_mask); in dequeue_event_callback()
2695 tp->iter_count = RTE_MAX( in dequeue_event_callback()
2696 tp->dec_ops[i]->turbo_dec.iter_count, in dequeue_event_callback()
2697 tp->iter_count); in dequeue_event_callback()
2698 rte_bbdev_dec_op_free_bulk(tp->dec_ops, deq); in dequeue_event_callback()
2700 struct rte_bbdev_enc_op *ref_op = tp->op_params->ref_enc_op; in dequeue_event_callback()
2701 ret = validate_enc_op(tp->enc_ops, num_ops, ref_op); in dequeue_event_callback()
2702 rte_bbdev_enc_op_free_bulk(tp->enc_ops, deq); in dequeue_event_callback()
2704 struct rte_bbdev_enc_op *ref_op = tp->op_params->ref_enc_op; in dequeue_event_callback()
2705 ret = validate_ldpc_enc_op(tp->enc_ops, num_ops, ref_op); in dequeue_event_callback()
2706 rte_bbdev_enc_op_free_bulk(tp->enc_ops, deq); in dequeue_event_callback()
2708 struct rte_bbdev_dec_op *ref_op = tp->op_params->ref_dec_op; in dequeue_event_callback()
2709 ret = validate_ldpc_dec_op(tp->dec_ops, num_ops, ref_op, in dequeue_event_callback()
2710 tp->op_params->vector_mask); in dequeue_event_callback()
2711 rte_bbdev_dec_op_free_bulk(tp->dec_ops, deq); in dequeue_event_callback()
2716 __atomic_store_n(&tp->processing_status, TEST_FAILED, __ATOMIC_RELAXED); in dequeue_event_callback()
2721 tb_len_bits = calc_dec_TB_size(tp->op_params->ref_dec_op); in dequeue_event_callback()
2724 tb_len_bits = calc_enc_TB_size(tp->op_params->ref_enc_op); in dequeue_event_callback()
2727 tb_len_bits = calc_ldpc_dec_TB_size(tp->op_params->ref_dec_op); in dequeue_event_callback()
2730 tb_len_bits = calc_ldpc_enc_TB_size(tp->op_params->ref_enc_op); in dequeue_event_callback()
2737 __atomic_store_n(&tp->processing_status, TEST_FAILED, __ATOMIC_RELAXED); in dequeue_event_callback()
2741 tp->ops_per_sec += ((double)num_ops) / in dequeue_event_callback()
2743 tp->mbps += (((double)(num_ops * tb_len_bits)) / 1000000.0) / in dequeue_event_callback()
2746 __atomic_fetch_add(&tp->nb_dequeued, deq, __ATOMIC_RELAXED); in dequeue_event_callback()
2752 struct thread_params *tp = arg; in throughput_intr_lcore_ldpc_dec() local
2754 const uint16_t queue_id = tp->queue_id; in throughput_intr_lcore_ldpc_dec()
2755 const uint16_t burst_sz = tp->op_params->burst_sz; in throughput_intr_lcore_ldpc_dec()
2756 const uint16_t num_to_process = tp->op_params->num_to_process; in throughput_intr_lcore_ldpc_dec()
2761 struct rte_bbdev_dec_op *ref_op = tp->op_params->ref_dec_op; in throughput_intr_lcore_ldpc_dec()
2772 TEST_ASSERT_SUCCESS(rte_bbdev_queue_intr_enable(tp->dev_id, queue_id), in throughput_intr_lcore_ldpc_dec()
2774 tp->dev_id, queue_id); in throughput_intr_lcore_ldpc_dec()
2776 rte_bbdev_info_get(tp->dev_id, &info); in throughput_intr_lcore_ldpc_dec()
2782 bufs = &tp->op_params->q_bufs[GET_SOCKET(info.socket_id)][queue_id]; in throughput_intr_lcore_ldpc_dec()
2784 __atomic_store_n(&tp->processing_status, 0, __ATOMIC_RELAXED); in throughput_intr_lcore_ldpc_dec()
2785 __atomic_store_n(&tp->nb_dequeued, 0, __ATOMIC_RELAXED); in throughput_intr_lcore_ldpc_dec()
2787 rte_wait_until_equal_16(&tp->op_params->sync, SYNC_START, __ATOMIC_RELAXED); in throughput_intr_lcore_ldpc_dec()
2789 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops, in throughput_intr_lcore_ldpc_dec()
2812 tp->start_time = rte_rdtsc_precise(); in throughput_intr_lcore_ldpc_dec()
2822 tp->dev_id, in throughput_intr_lcore_ldpc_dec()
2835 __atomic_store_n(&tp->burst_sz, num_to_enq, __ATOMIC_RELAXED); in throughput_intr_lcore_ldpc_dec()
2840 rte_wait_until_equal_16(&tp->nb_dequeued, enqueued, __ATOMIC_RELAXED); in throughput_intr_lcore_ldpc_dec()
2843 __atomic_store_n(&tp->nb_dequeued, 0, __ATOMIC_RELAXED); in throughput_intr_lcore_ldpc_dec()
2852 struct thread_params *tp = arg; in throughput_intr_lcore_dec() local
2854 const uint16_t queue_id = tp->queue_id; in throughput_intr_lcore_dec()
2855 const uint16_t burst_sz = tp->op_params->burst_sz; in throughput_intr_lcore_dec()
2856 const uint16_t num_to_process = tp->op_params->num_to_process; in throughput_intr_lcore_dec()
2866 TEST_ASSERT_SUCCESS(rte_bbdev_queue_intr_enable(tp->dev_id, queue_id), in throughput_intr_lcore_dec()
2868 tp->dev_id, queue_id); in throughput_intr_lcore_dec()
2870 rte_bbdev_info_get(tp->dev_id, &info); in throughput_intr_lcore_dec()
2876 bufs = &tp->op_params->q_bufs[GET_SOCKET(info.socket_id)][queue_id]; in throughput_intr_lcore_dec()
2878 __atomic_store_n(&tp->processing_status, 0, __ATOMIC_RELAXED); in throughput_intr_lcore_dec()
2879 __atomic_store_n(&tp->nb_dequeued, 0, __ATOMIC_RELAXED); in throughput_intr_lcore_dec()
2881 rte_wait_until_equal_16(&tp->op_params->sync, SYNC_START, __ATOMIC_RELAXED); in throughput_intr_lcore_dec()
2883 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops, in throughput_intr_lcore_dec()
2890 tp->op_params->ref_dec_op); in throughput_intr_lcore_dec()
2900 tp->start_time = rte_rdtsc_precise(); in throughput_intr_lcore_dec()
2909 enq += rte_bbdev_enqueue_dec_ops(tp->dev_id, in throughput_intr_lcore_dec()
2922 __atomic_store_n(&tp->burst_sz, num_to_enq, __ATOMIC_RELAXED); in throughput_intr_lcore_dec()
2927 rte_wait_until_equal_16(&tp->nb_dequeued, enqueued, __ATOMIC_RELAXED); in throughput_intr_lcore_dec()
2930 __atomic_store_n(&tp->nb_dequeued, 0, __ATOMIC_RELAXED); in throughput_intr_lcore_dec()
2939 struct thread_params *tp = arg; in throughput_intr_lcore_enc() local
2941 const uint16_t queue_id = tp->queue_id; in throughput_intr_lcore_enc()
2942 const uint16_t burst_sz = tp->op_params->burst_sz; in throughput_intr_lcore_enc()
2943 const uint16_t num_to_process = tp->op_params->num_to_process; in throughput_intr_lcore_enc()
2953 TEST_ASSERT_SUCCESS(rte_bbdev_queue_intr_enable(tp->dev_id, queue_id), in throughput_intr_lcore_enc()
2955 tp->dev_id, queue_id); in throughput_intr_lcore_enc()
2957 rte_bbdev_info_get(tp->dev_id, &info); in throughput_intr_lcore_enc()
2963 bufs = &tp->op_params->q_bufs[GET_SOCKET(info.socket_id)][queue_id]; in throughput_intr_lcore_enc()
2965 __atomic_store_n(&tp->processing_status, 0, __ATOMIC_RELAXED); in throughput_intr_lcore_enc()
2966 __atomic_store_n(&tp->nb_dequeued, 0, __ATOMIC_RELAXED); in throughput_intr_lcore_enc()
2968 rte_wait_until_equal_16(&tp->op_params->sync, SYNC_START, __ATOMIC_RELAXED); in throughput_intr_lcore_enc()
2970 ret = rte_bbdev_enc_op_alloc_bulk(tp->op_params->mp, ops, in throughput_intr_lcore_enc()
2976 bufs->hard_outputs, tp->op_params->ref_enc_op); in throughput_intr_lcore_enc()
2986 tp->start_time = rte_rdtsc_precise(); in throughput_intr_lcore_enc()
2995 enq += rte_bbdev_enqueue_enc_ops(tp->dev_id, in throughput_intr_lcore_enc()
3008 __atomic_store_n(&tp->burst_sz, num_to_enq, __ATOMIC_RELAXED); in throughput_intr_lcore_enc()
3013 rte_wait_until_equal_16(&tp->nb_dequeued, enqueued, __ATOMIC_RELAXED); in throughput_intr_lcore_enc()
3016 __atomic_store_n(&tp->nb_dequeued, 0, __ATOMIC_RELAXED); in throughput_intr_lcore_enc()
3026 struct thread_params *tp = arg; in throughput_intr_lcore_ldpc_enc() local
3028 const uint16_t queue_id = tp->queue_id; in throughput_intr_lcore_ldpc_enc()
3029 const uint16_t burst_sz = tp->op_params->burst_sz; in throughput_intr_lcore_ldpc_enc()
3030 const uint16_t num_to_process = tp->op_params->num_to_process; in throughput_intr_lcore_ldpc_enc()
3040 TEST_ASSERT_SUCCESS(rte_bbdev_queue_intr_enable(tp->dev_id, queue_id), in throughput_intr_lcore_ldpc_enc()
3042 tp->dev_id, queue_id); in throughput_intr_lcore_ldpc_enc()
3044 rte_bbdev_info_get(tp->dev_id, &info); in throughput_intr_lcore_ldpc_enc()
3050 bufs = &tp->op_params->q_bufs[GET_SOCKET(info.socket_id)][queue_id]; in throughput_intr_lcore_ldpc_enc()
3052 __atomic_store_n(&tp->processing_status, 0, __ATOMIC_RELAXED); in throughput_intr_lcore_ldpc_enc()
3053 __atomic_store_n(&tp->nb_dequeued, 0, __ATOMIC_RELAXED); in throughput_intr_lcore_ldpc_enc()
3055 rte_wait_until_equal_16(&tp->op_params->sync, SYNC_START, __ATOMIC_RELAXED); in throughput_intr_lcore_ldpc_enc()
3057 ret = rte_bbdev_enc_op_alloc_bulk(tp->op_params->mp, ops, in throughput_intr_lcore_ldpc_enc()
3064 tp->op_params->ref_enc_op); in throughput_intr_lcore_ldpc_enc()
3074 tp->start_time = rte_rdtsc_precise(); in throughput_intr_lcore_ldpc_enc()
3084 tp->dev_id, in throughput_intr_lcore_ldpc_enc()
3097 __atomic_store_n(&tp->burst_sz, num_to_enq, __ATOMIC_RELAXED); in throughput_intr_lcore_ldpc_enc()
3102 rte_wait_until_equal_16(&tp->nb_dequeued, enqueued, __ATOMIC_RELAXED); in throughput_intr_lcore_ldpc_enc()
3105 __atomic_store_n(&tp->nb_dequeued, 0, __ATOMIC_RELAXED); in throughput_intr_lcore_ldpc_enc()
3114 struct thread_params *tp = arg; in throughput_pmd_lcore_dec() local
3117 const uint16_t queue_id = tp->queue_id; in throughput_pmd_lcore_dec()
3118 const uint16_t burst_sz = tp->op_params->burst_sz; in throughput_pmd_lcore_dec()
3119 const uint16_t num_ops = tp->op_params->num_to_process; in throughput_pmd_lcore_dec()
3122 struct rte_bbdev_dec_op *ref_op = tp->op_params->ref_dec_op; in throughput_pmd_lcore_dec()
3131 rte_bbdev_info_get(tp->dev_id, &info); in throughput_pmd_lcore_dec()
3137 bufs = &tp->op_params->q_bufs[GET_SOCKET(info.socket_id)][queue_id]; in throughput_pmd_lcore_dec()
3139 rte_wait_until_equal_16(&tp->op_params->sync, SYNC_START, __ATOMIC_RELAXED); in throughput_pmd_lcore_dec()
3141 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops_enq, num_ops); in throughput_pmd_lcore_dec()
3165 enq += rte_bbdev_enqueue_dec_ops(tp->dev_id, in throughput_pmd_lcore_dec()
3168 deq += rte_bbdev_dequeue_dec_ops(tp->dev_id, in throughput_pmd_lcore_dec()
3174 deq += rte_bbdev_dequeue_dec_ops(tp->dev_id, in throughput_pmd_lcore_dec()
3181 tp->iter_count = 0; in throughput_pmd_lcore_dec()
3184 tp->iter_count = RTE_MAX(ops_enq[i]->turbo_dec.iter_count, in throughput_pmd_lcore_dec()
3185 tp->iter_count); in throughput_pmd_lcore_dec()
3190 tp->op_params->vector_mask); in throughput_pmd_lcore_dec()
3198 tp->ops_per_sec = ((double)num_ops * TEST_REPETITIONS) / in throughput_pmd_lcore_dec()
3200 tp->mbps = (((double)(num_ops * TEST_REPETITIONS * tb_len_bits)) / in throughput_pmd_lcore_dec()
3210 struct thread_params *tp = arg; in bler_pmd_lcore_ldpc_dec() local
3213 const uint16_t queue_id = tp->queue_id; in bler_pmd_lcore_ldpc_dec()
3214 const uint16_t burst_sz = tp->op_params->burst_sz; in bler_pmd_lcore_ldpc_dec()
3215 const uint16_t num_ops = tp->op_params->num_to_process; in bler_pmd_lcore_ldpc_dec()
3218 struct rte_bbdev_dec_op *ref_op = tp->op_params->ref_dec_op; in bler_pmd_lcore_ldpc_dec()
3234 rte_bbdev_info_get(tp->dev_id, &info); in bler_pmd_lcore_ldpc_dec()
3240 bufs = &tp->op_params->q_bufs[GET_SOCKET(info.socket_id)][queue_id]; in bler_pmd_lcore_ldpc_dec()
3242 rte_wait_until_equal_16(&tp->op_params->sync, SYNC_START, __ATOMIC_RELAXED); in bler_pmd_lcore_ldpc_dec()
3244 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops_enq, num_ops); in bler_pmd_lcore_ldpc_dec()
3275 preload_harq_ddr(tp->dev_id, queue_id, ops_enq, in bler_pmd_lcore_ldpc_dec()
3285 enq += rte_bbdev_enqueue_ldpc_dec_ops(tp->dev_id, in bler_pmd_lcore_ldpc_dec()
3288 deq += rte_bbdev_dequeue_ldpc_dec_ops(tp->dev_id, in bler_pmd_lcore_ldpc_dec()
3294 deq += rte_bbdev_dequeue_ldpc_dec_ops(tp->dev_id, in bler_pmd_lcore_ldpc_dec()
3301 tp->iter_count = 0; in bler_pmd_lcore_ldpc_dec()
3302 tp->iter_average = 0; in bler_pmd_lcore_ldpc_dec()
3305 tp->iter_count = RTE_MAX(ops_enq[i]->ldpc_dec.iter_count, in bler_pmd_lcore_ldpc_dec()
3306 tp->iter_count); in bler_pmd_lcore_ldpc_dec()
3307 tp->iter_average += (double) ops_enq[i]->ldpc_dec.iter_count; in bler_pmd_lcore_ldpc_dec()
3313 tp->iter_average /= num_ops; in bler_pmd_lcore_ldpc_dec()
3314 tp->bler = (double) validate_ldpc_bler(ops_deq, num_ops) / num_ops; in bler_pmd_lcore_ldpc_dec()
3317 && tp->bler == 0 in bler_pmd_lcore_ldpc_dec()
3321 tp->op_params->vector_mask); in bler_pmd_lcore_ldpc_dec()
3328 tp->ops_per_sec = ((double)num_ops * 1) / in bler_pmd_lcore_ldpc_dec()
3330 tp->mbps = (((double)(num_ops * 1 * tb_len_bits)) / in bler_pmd_lcore_ldpc_dec()
3340 struct thread_params *tp = arg; in throughput_pmd_lcore_ldpc_dec() local
3343 const uint16_t queue_id = tp->queue_id; in throughput_pmd_lcore_ldpc_dec()
3344 const uint16_t burst_sz = tp->op_params->burst_sz; in throughput_pmd_lcore_ldpc_dec()
3345 const uint16_t num_ops = tp->op_params->num_to_process; in throughput_pmd_lcore_ldpc_dec()
3348 struct rte_bbdev_dec_op *ref_op = tp->op_params->ref_dec_op; in throughput_pmd_lcore_ldpc_dec()
3363 rte_bbdev_info_get(tp->dev_id, &info); in throughput_pmd_lcore_ldpc_dec()
3369 bufs = &tp->op_params->q_bufs[GET_SOCKET(info.socket_id)][queue_id]; in throughput_pmd_lcore_ldpc_dec()
3371 rte_wait_until_equal_16(&tp->op_params->sync, SYNC_START, __ATOMIC_RELAXED); in throughput_pmd_lcore_ldpc_dec()
3373 ret = rte_bbdev_dec_op_alloc_bulk(tp->op_params->mp, ops_enq, num_ops); in throughput_pmd_lcore_ldpc_dec()
3403 preload_harq_ddr(tp->dev_id, queue_id, ops_enq, in throughput_pmd_lcore_ldpc_dec()
3413 enq += rte_bbdev_enqueue_ldpc_dec_ops(tp->dev_id, in throughput_pmd_lcore_ldpc_dec()
3416 deq += rte_bbdev_dequeue_ldpc_dec_ops(tp->dev_id, in throughput_pmd_lcore_ldpc_dec()
3422 deq += rte_bbdev_dequeue_ldpc_dec_ops(tp->dev_id, in throughput_pmd_lcore_ldpc_dec()
3429 tp->iter_count = 0; in throughput_pmd_lcore_ldpc_dec()
3432 tp->iter_count = RTE_MAX(ops_enq[i]->ldpc_dec.iter_count, in throughput_pmd_lcore_ldpc_dec()
3433 tp->iter_count); in throughput_pmd_lcore_ldpc_dec()
3437 retrieve_harq_ddr(tp->dev_id, queue_id, ops_enq, num_ops); in throughput_pmd_lcore_ldpc_dec()
3442 tp->op_params->vector_mask); in throughput_pmd_lcore_ldpc_dec()
3450 tp->ops_per_sec = ((double)num_ops * TEST_REPETITIONS) / in throughput_pmd_lcore_ldpc_dec()
3452 tp->mbps = (((double)(num_ops * TEST_REPETITIONS * tb_len_bits)) / in throughput_pmd_lcore_ldpc_dec()
3462 struct thread_params *tp = arg; in throughput_pmd_lcore_enc() local
3465 const uint16_t queue_id = tp->queue_id; in throughput_pmd_lcore_enc()
3466 const uint16_t burst_sz = tp->op_params->burst_sz; in throughput_pmd_lcore_enc()
3467 const uint16_t num_ops = tp->op_params->num_to_process; in throughput_pmd_lcore_enc()
3470 struct rte_bbdev_enc_op *ref_op = tp->op_params->ref_enc_op; in throughput_pmd_lcore_enc()
3479 rte_bbdev_info_get(tp->dev_id, &info); in throughput_pmd_lcore_enc()
3485 bufs = &tp->op_params->q_bufs[GET_SOCKET(info.socket_id)][queue_id]; in throughput_pmd_lcore_enc()
3487 rte_wait_until_equal_16(&tp->op_params->sync, SYNC_START, __ATOMIC_RELAXED); in throughput_pmd_lcore_enc()
3489 ret = rte_bbdev_enc_op_alloc_bulk(tp->op_params->mp, ops_enq, in throughput_pmd_lcore_enc()
3515 enq += rte_bbdev_enqueue_enc_ops(tp->dev_id, in throughput_pmd_lcore_enc()
3518 deq += rte_bbdev_dequeue_enc_ops(tp->dev_id, in throughput_pmd_lcore_enc()
3524 deq += rte_bbdev_dequeue_enc_ops(tp->dev_id, in throughput_pmd_lcore_enc()
3540 tp->ops_per_sec = ((double)num_ops * TEST_REPETITIONS) / in throughput_pmd_lcore_enc()
3542 tp->mbps = (((double)(num_ops * TEST_REPETITIONS * tb_len_bits)) in throughput_pmd_lcore_enc()
3552 struct thread_params *tp = arg; in throughput_pmd_lcore_ldpc_enc() local
3555 const uint16_t queue_id = tp->queue_id; in throughput_pmd_lcore_ldpc_enc()
3556 const uint16_t burst_sz = tp->op_params->burst_sz; in throughput_pmd_lcore_ldpc_enc()
3557 const uint16_t num_ops = tp->op_params->num_to_process; in throughput_pmd_lcore_ldpc_enc()
3560 struct rte_bbdev_enc_op *ref_op = tp->op_params->ref_enc_op; in throughput_pmd_lcore_ldpc_enc()
3569 rte_bbdev_info_get(tp->dev_id, &info); in throughput_pmd_lcore_ldpc_enc()
3575 bufs = &tp->op_params->q_bufs[GET_SOCKET(info.socket_id)][queue_id]; in throughput_pmd_lcore_ldpc_enc()
3577 rte_wait_until_equal_16(&tp->op_params->sync, SYNC_START, __ATOMIC_RELAXED); in throughput_pmd_lcore_ldpc_enc()
3579 ret = rte_bbdev_enc_op_alloc_bulk(tp->op_params->mp, ops_enq, in throughput_pmd_lcore_ldpc_enc()
3605 enq += rte_bbdev_enqueue_ldpc_enc_ops(tp->dev_id, in throughput_pmd_lcore_ldpc_enc()
3608 deq += rte_bbdev_dequeue_ldpc_enc_ops(tp->dev_id, in throughput_pmd_lcore_ldpc_enc()
3614 deq += rte_bbdev_dequeue_ldpc_enc_ops(tp->dev_id, in throughput_pmd_lcore_ldpc_enc()
3630 tp->ops_per_sec = ((double)num_ops * TEST_REPETITIONS) / in throughput_pmd_lcore_ldpc_enc()
3632 tp->mbps = (((double)(num_ops * TEST_REPETITIONS * tb_len_bits)) in throughput_pmd_lcore_ldpc_enc()
3813 struct thread_params *t_params, *tp; in throughput_test() local
3930 tp = &t_params[0]; in throughput_test()
3931 while ((__atomic_load_n(&tp->nb_dequeued, __ATOMIC_RELAXED) < in throughput_test()
3933 (__atomic_load_n(&tp->processing_status, __ATOMIC_RELAXED) != in throughput_test()
3937 tp->ops_per_sec /= TEST_REPETITIONS; in throughput_test()
3938 tp->mbps /= TEST_REPETITIONS; in throughput_test()
3939 ret |= (int)__atomic_load_n(&tp->processing_status, __ATOMIC_RELAXED); in throughput_test()
3943 tp = &t_params[used_cores]; in throughput_test()
3945 while ((__atomic_load_n(&tp->nb_dequeued, __ATOMIC_RELAXED) < in throughput_test()
3947 (__atomic_load_n(&tp->processing_status, __ATOMIC_RELAXED) != in throughput_test()
3951 tp->ops_per_sec /= TEST_REPETITIONS; in throughput_test()
3952 tp->mbps /= TEST_REPETITIONS; in throughput_test()
3953 ret |= (int)__atomic_load_n(&tp->processing_status, __ATOMIC_RELAXED); in throughput_test()