| /f-stack/dpdk/lib/librte_eal/common/ |
| H A D | eal_common_dynmem.c | 154 max_segs_per_type = RTE_MIN(max_segs_per_type, in eal_dynmem_memseg_lists_init() 157 max_segs_per_list = RTE_MIN(max_segs_per_type, in eal_dynmem_memseg_lists_init() 161 max_mem_per_list = RTE_MIN(max_segs_per_list * pagesz, in eal_dynmem_memseg_lists_init() 165 n_segs = RTE_MIN(max_segs_per_list, max_mem_per_list / pagesz); in eal_dynmem_memseg_lists_init() 168 n_seglists = RTE_MIN(max_segs_per_type / n_segs, in eal_dynmem_memseg_lists_init() 172 n_seglists = RTE_MIN(n_seglists, max_seglists_per_type); in eal_dynmem_memseg_lists_init() 259 hpi->num_pages[i] = RTE_MIN(hpi->num_pages[i], in eal_dynmem_hugepage_init() 406 default_size = RTE_MIN( in eal_dynmem_calc_num_pages_per_socket() 421 default_size = RTE_MIN( in eal_dynmem_calc_num_pages_per_socket() 458 hp_used[i].num_pages[socket] = RTE_MIN( in eal_dynmem_calc_num_pages_per_socket() [all …]
|
| /f-stack/dpdk/drivers/net/mlx5/ |
| H A D | mlx5_rxtx_vec.c | 161 n = RTE_MIN(n, rxq->rq_repl_thresh); in mlx5_rx_mprq_replenish_bulk_mbuf() 302 pkts_n = RTE_MIN(pkts_n, MLX5_VPMD_RX_MAX_BURST); in rxq_burst_v() 307 rcvd_pkt = RTE_MIN(rcvd_pkt, pkts_n); in rxq_burst_v() 319 pkts_n = RTE_MIN(pkts_n, q_n - elts_idx); in rxq_burst_v() 320 pkts_n = RTE_MIN(pkts_n, q_n - cq_idx); in rxq_burst_v() 349 n = RTE_MIN(n, pkts_n - nocmp_n); in rxq_burst_v() 442 pkts_n = RTE_MIN(pkts_n, MLX5_VPMD_RX_MAX_BURST); in rxq_burst_mprq_v() 447 rcvd_pkt = RTE_MIN(rcvd_pkt, pkts_n); in rxq_burst_mprq_v() 457 pkts_n = RTE_MIN(pkts_n, elts_n - elts_idx); in rxq_burst_mprq_v() 458 pkts_n = RTE_MIN(pkts_n, q_n - cq_idx); in rxq_burst_mprq_v() [all …]
|
| /f-stack/dpdk/lib/librte_eal/freebsd/ |
| H A D | eal_memory.c | 114 n_pages = RTE_MIN(mem_needed / page_sz, max_pages); in rte_eal_hugepage_init() 312 max_mem = RTE_MIN((uint64_t)RTE_MAX_MEM_MB_PER_LIST << 20, max_mem); in get_mem_amount() 314 area_sz = RTE_MIN(page_sz * max_pages, max_mem); in get_mem_amount() 379 max_type_mem = RTE_MIN(max_mem - total_mem, in memseg_primary_init() 396 max_type_mem = RTE_MIN(avail_mem, max_type_mem); in memseg_primary_init() 397 max_segs = RTE_MIN(avail_segs, max_segs); in memseg_primary_init()
|
| /f-stack/dpdk/app/test-crypto-perf/ |
| H A D | cperf_test_pmd_cyclecount.c | 156 RTE_MIN(state->opts->nb_descriptors, iter_ops_left); in pmd_cyclecount_bench_ops() 162 uint32_t burst_size = RTE_MIN(iter_ops_needed - cur_iter_op, in pmd_cyclecount_bench_ops() 213 uint32_t burst_size = RTE_MIN( in pmd_cyclecount_build_ops() 248 uint32_t burst_size = RTE_MIN(iter_ops_needed - cur_iter_op, in pmd_cyclecount_bench_enq() 278 uint32_t burst_size = RTE_MIN(iter_ops_needed - cur_iter_op, in pmd_cyclecount_bench_deq() 343 uint32_t iter_ops_needed = RTE_MIN( in pmd_cyclecount_bench_burst_sz()
|
| H A D | cperf_test_latency.c | 255 deqd_min = RTE_MIN(ops_deqd, deqd_min); in cperf_latency_test_runner() 260 enqd_min = RTE_MIN(ops_enqd, enqd_min); in cperf_latency_test_runner() 285 deqd_min = RTE_MIN(ops_deqd, deqd_min); in cperf_latency_test_runner() 292 tsc_min = RTE_MIN(tsc_val, tsc_min); in cperf_latency_test_runner()
|
| /f-stack/dpdk/examples/vhost/ |
| H A D | virtio_net.c | 165 cpy_len = RTE_MIN(desc_chunck_len, mbuf_avail); in enqueue_pkt() 197 count = RTE_MIN(count, free_entries); in vs_enqueue_pkts() 198 count = RTE_MIN(count, (uint32_t)MAX_PKT_BURST); in vs_enqueue_pkts() 289 cpy_len = RTE_MIN(desc_chunck_len, mbuf_avail); in dequeue_pkt() 388 count = RTE_MIN(count, MAX_PKT_BURST); in vs_dequeue_pkts() 389 count = RTE_MIN(count, free_entries); in vs_dequeue_pkts()
|
| /f-stack/dpdk/drivers/net/bnxt/ |
| H A D | bnxt_rxtx_vec_sse.c | 169 nb_pkts = RTE_MIN(nb_pkts, RTE_BNXT_MAX_RX_BURST); in bnxt_recv_pkts_vec() 179 nb_pkts = RTE_MIN(nb_pkts, RTE_MIN(rx_ring_size - mbcons, in bnxt_recv_pkts_vec() 366 nb_pkts = RTE_MIN(nb_pkts, bnxt_tx_avail(txq)); in bnxt_xmit_fixed_burst_vec() 432 num = RTE_MIN(nb_pkts, RTE_BNXT_MAX_TX_BURST); in bnxt_xmit_pkts_vec() 438 num = RTE_MIN(num, in bnxt_xmit_pkts_vec()
|
| H A D | bnxt_rxtx_vec_neon.c | 177 nb_pkts = RTE_MIN(nb_pkts, RTE_BNXT_MAX_RX_BURST); in bnxt_recv_pkts_vec() 187 nb_pkts = RTE_MIN(nb_pkts, RTE_MIN(rx_ring_size - mbcons, in bnxt_recv_pkts_vec() 362 nb_pkts = RTE_MIN(nb_pkts, bnxt_tx_avail(txq)); in bnxt_xmit_fixed_burst_vec() 420 num = RTE_MIN(nb_pkts, RTE_BNXT_MAX_TX_BURST); in bnxt_xmit_pkts_vec()
|
| H A D | bnxt_rxq.c | 84 max_pools = RTE_MIN(bp->max_vnics, in bnxt_mq_rx_configure() 85 RTE_MIN(bp->max_l2_ctx, in bnxt_mq_rx_configure() 86 RTE_MIN(bp->max_rsscos_ctx, in bnxt_mq_rx_configure() 107 pools = RTE_MIN(pools, bp->rx_cp_nr_rings); in bnxt_mq_rx_configure() 341 RTE_MIN(rte_align32pow2(nb_desc) / 4, RTE_BNXT_MAX_RX_BURST); in bnxt_rx_queue_setup_op()
|
| /f-stack/dpdk/drivers/net/enic/ |
| H A D | enic_res.c | 71 enic->rte_dev->data->mtu = RTE_MIN(enic->max_mtu, in enic_get_vnic_config() 103 c->wq_desc_count = RTE_MIN((uint32_t)ENIC_MAX_WQ_DESCS, in enic_get_vnic_config() 107 c->rq_desc_count = RTE_MIN((uint32_t)ENIC_MAX_RQ_DESCS, in enic_get_vnic_config() 111 c->intr_timer_usec = RTE_MIN(c->intr_timer_usec, in enic_get_vnic_config()
|
| H A D | enic_rxtx.c | 74 max_rx = RTE_MIN(nb_pkts, cq->ring.desc_count - cq_idx); in enic_recv_pkts() 251 max_rx = RTE_MIN(nb_pkts, rq->num_free_mbufs); in enic_noscatter_recv_pkts() 252 max_rx = RTE_MIN(max_rx, cq->ring.desc_count - cq_idx); in enic_noscatter_recv_pkts() 466 nb_pkts = RTE_MIN(nb_pkts, ENIC_TX_XMIT_MAX); in enic_xmit_pkts() 652 nb_pkts = RTE_MIN(nb_pkts, wq->ring.desc_avail); in enic_simple_xmit_pkts() 661 n = RTE_MIN(nb_pkts, n); in enic_simple_xmit_pkts()
|
| /f-stack/dpdk/drivers/net/thunderx/ |
| H A D | nicvf_ethdev.h | 117 RTE_MIN(tmp, dev->data->nb_tx_queues - 1) : 0; in nicvf_tx_range() 131 RTE_MIN(tmp, dev->data->nb_rx_queues - 1) : 0; in nicvf_rx_range()
|
| /f-stack/dpdk/drivers/net/ionic/ |
| H A D | ionic_main.c | 283 nwords = RTE_MIN(drv_size, cmd_size); in ionic_identify() 290 nwords = RTE_MIN(dev_size, cmd_size); in ionic_identify() 336 nwords = RTE_MIN(port_words, cmd_words); in ionic_port_identify() 401 nwords = RTE_MIN(config_words, cmd_words); in ionic_port_init()
|
| /f-stack/dpdk/drivers/net/ena/ |
| H A D | ena_ethdev.c | 895 max_tx_queue_size = RTE_MIN(max_tx_queue_size, in ena_calc_io_queue_size() 898 max_tx_queue_size = RTE_MIN(max_tx_queue_size, in ena_calc_io_queue_size() 902 ctx->max_rx_sgl_size = RTE_MIN(ENA_PKT_MAX_BUFS, in ena_calc_io_queue_size() 904 ctx->max_tx_sgl_size = RTE_MIN(ENA_PKT_MAX_BUFS, in ena_calc_io_queue_size() 915 max_tx_queue_size = RTE_MIN(max_tx_queue_size, in ena_calc_io_queue_size() 918 max_tx_queue_size = RTE_MIN(max_tx_queue_size, in ena_calc_io_queue_size() 922 ctx->max_rx_sgl_size = RTE_MIN(ENA_PKT_MAX_BUFS, in ena_calc_io_queue_size() 924 ctx->max_tx_sgl_size = RTE_MIN(ENA_PKT_MAX_BUFS, in ena_calc_io_queue_size() 1007 max_rings_stats = RTE_MIN(dev->data->nb_rx_queues, in ena_stats_get() 1741 io_rx_num = RTE_MIN(io_tx_sq_num, io_tx_cq_num); in ena_calc_max_io_queue_num() [all …]
|
| /f-stack/dpdk/drivers/net/netvsc/ |
| H A D | hn_vf.c | 159 lim->nb_max = RTE_MIN(vf_lim->nb_max, lim->nb_max); in hn_vf_merge_desc_lim() 162 lim->nb_seg_max = RTE_MIN(vf_lim->nb_seg_max, lim->nb_seg_max); in hn_vf_merge_desc_lim() 163 lim->nb_mtu_seg_max = RTE_MIN(vf_lim->nb_seg_max, lim->nb_seg_max); in hn_vf_merge_desc_lim() 185 info->max_rx_queues = RTE_MIN(vf_info.max_rx_queues, in hn_vf_info_merge() 191 info->max_tx_queues = RTE_MIN(vf_info.max_tx_queues, in hn_vf_info_merge()
|
| /f-stack/dpdk/lib/librte_fib/ |
| H A D | dir24_8.h | 121 RTE_MIN((unsigned int)bulk_prefetch, n); \ 154 uint32_t prefetch_offset = RTE_MIN(15U, n); in dir24_8_lookup_bulk() 219 uint32_t prefetch_offset = RTE_MIN(15U, n); in dir24_8_lookup_bulk_uni()
|
| /f-stack/dpdk/drivers/net/failsafe/ |
| H A D | failsafe_ops.c | 1081 to->nb_max = RTE_MIN(to->nb_max, from->nb_max); in fs_dev_merge_desc_lim() 1085 to->nb_seg_max = RTE_MIN(to->nb_seg_max, from->nb_seg_max); in fs_dev_merge_desc_lim() 1086 to->nb_mtu_seg_max = RTE_MIN(to->nb_mtu_seg_max, from->nb_mtu_seg_max); in fs_dev_merge_desc_lim() 1098 info->max_rx_pktlen = RTE_MIN(info->max_rx_pktlen, sinfo->max_rx_pktlen); in fs_dev_merge_info() 1099 info->max_rx_queues = RTE_MIN(info->max_rx_queues, sinfo->max_rx_queues); in fs_dev_merge_info() 1100 info->max_tx_queues = RTE_MIN(info->max_tx_queues, sinfo->max_tx_queues); in fs_dev_merge_info() 1101 info->max_mac_addrs = RTE_MIN(info->max_mac_addrs, sinfo->max_mac_addrs); in fs_dev_merge_info() 1102 info->max_hash_mac_addrs = RTE_MIN(info->max_hash_mac_addrs, in fs_dev_merge_info() 1105 info->max_vfs = RTE_MIN(info->max_vfs, sinfo->max_vfs); in fs_dev_merge_info() 1120 info->reta_size = RTE_MIN(info->reta_size, sinfo->reta_size); in fs_dev_merge_info() [all …]
|
| /f-stack/dpdk/drivers/net/bnx2x/ |
| H A D | bnx2x_ethdev.h | 51 #define min(a,b) RTE_MIN(a,b)
|
| /f-stack/dpdk/drivers/net/avp/ |
| H A D | avp_ethdev.c | 1334 copy_length = RTE_MIN((avp->guest_mbuf_size - in avp_dev_copy_from_buffers() 1419 count = RTE_MIN(count, avail); in avp_recv_scattered_pkts() 1420 count = RTE_MIN(count, nb_pkts); in avp_recv_scattered_pkts() 1421 count = RTE_MIN(count, (unsigned int)AVP_MAX_RX_BURST); in avp_recv_scattered_pkts() 1516 count = RTE_MIN(count, avail); in avp_recv_pkts() 1517 count = RTE_MIN(count, nb_pkts); in avp_recv_pkts() 1648 copy_length = RTE_MIN((avp->host_mbuf_size - in avp_dev_copy_to_buffers() 1733 nb_pkts = RTE_MIN(count, nb_pkts); in avp_xmit_scattered_pkts() 1849 count = RTE_MIN(count, avail); in avp_xmit_pkts() 1850 count = RTE_MIN(count, nb_pkts); in avp_xmit_pkts() [all …]
|
| /f-stack/dpdk/app/test-compress-perf/ |
| H A D | comp_perf_test_throughput.c | 135 uint16_t num_ops = RTE_MIN(remaining_ops, in main_loop() 233 RTE_MIN(remaining_data, in main_loop() 287 RTE_MIN(remaining_data, in main_loop()
|
| H A D | comp_perf_test_cyclecount.c | 85 uint16_t num_ops = RTE_MIN(remaining_ops, in cperf_cyclecount_op_setup() 260 uint16_t num_ops = RTE_MIN(remaining_ops, in main_loop() 376 RTE_MIN(remaining_data, in main_loop() 434 RTE_MIN(remaining_data, in main_loop()
|
| H A D | comp_perf_test_verify.c | 140 uint16_t num_ops = RTE_MIN(remaining_ops, in main_loop() 269 RTE_MIN(remaining_data, in main_loop() 353 RTE_MIN(remaining_data, in main_loop()
|
| /f-stack/dpdk/examples/vm_power_manager/guest_cli/ |
| H A D | parse.c | 70 for (idx = RTE_MIN(min, max); in parse_set()
|
| /f-stack/dpdk/drivers/net/hns3/ |
| H A D | hns3_rxtx_vec.c | 37 new_burst = RTE_MIN(nb_pkts, txq->tx_rs_thresh); in hns3_xmit_pkts_vec() 110 nb_pkts = RTE_MIN(nb_pkts, HNS3_DEFAULT_RX_BURST); in hns3_recv_pkts_vec()
|
| /f-stack/dpdk/drivers/compress/isal/ |
| H A D | isal_compress_pmd.c | 247 qp->stream->avail_in = RTE_MIN(src->data_len - remaining_offset, in chained_mbuf_compression() 290 RTE_MIN(remaining_data, src->data_len); in chained_mbuf_compression() 337 qp->state->avail_in = RTE_MIN(src->data_len - src_remaining_offset, in chained_mbuf_decompression() 371 RTE_MIN(remaining_data, src->data_len); in chained_mbuf_decompression() 635 int16_t num_enq = RTE_MIN(qp->num_free_elements, nb_ops); in isal_comp_pmd_enqueue_burst()
|