Searched refs:ops_needed (Results 1 – 5 of 5) sorted by relevance
73 uint16_t ops_needed; in cperf_cyclecount_op_setup() local87 ops_needed = num_ops; in cperf_cyclecount_op_setup()90 if (ops_needed && rte_mempool_get_bulk( in cperf_cyclecount_op_setup()93 ops_needed) != 0) { in cperf_cyclecount_op_setup()100 for (i = 0; i < ops_needed; i++) { in cperf_cyclecount_op_setup()146 rte_mempool_put_bulk(mem->op_pool, (void **)ops, ops_needed); in cperf_cyclecount_op_setup()262 uint16_t ops_needed = num_ops - ops_unused; in main_loop() local276 if (ops_needed && rte_mempool_get_bulk( in main_loop()279 ops_needed) != 0) { in main_loop()285 allocated += ops_needed; in main_loop()[all …]
137 uint16_t ops_needed = num_ops - ops_unused; in main_loop() local151 if (ops_needed && !rte_comp_op_bulk_alloc( in main_loop()154 ops_needed)) { in main_loop()160 allocated += ops_needed; in main_loop()162 for (i = 0; i < ops_needed; i++) { in main_loop()
142 uint16_t ops_needed = num_ops - ops_unused; in main_loop() local156 if (ops_needed && !rte_comp_op_bulk_alloc( in main_loop()159 ops_needed)) { in main_loop()165 allocated += ops_needed; in main_loop()167 for (i = 0; i < ops_needed; i++) { in main_loop()
163 uint16_t ops_needed = burst_size - ops_unused; in cperf_throughput_test_runner() local167 ops_needed) != 0) { in cperf_throughput_test_runner()179 ops_needed, ctx->sess, in cperf_throughput_test_runner()188 if (unlikely(ops_enqd > ops_needed)) { in cperf_throughput_test_runner()192 memmove(&ops[ops_needed], &ops[ops_enqd], in cperf_throughput_test_runner()
285 uint16_t ops_needed = burst_size - ops_unused; in cperf_verify_test_runner() local289 ops_needed) != 0) { in cperf_verify_test_runner()301 ops_needed, ctx->sess, ctx->options, in cperf_verify_test_runner()306 for (i = 0; i < ops_needed; i++) in cperf_verify_test_runner()