| /dpdk/drivers/net/virtio/ |
| H A D | virtio_ring.h | 85 unsigned int num; member 92 unsigned int num; member 137 size = num * sizeof(struct vring_packed_desc); in vring_size() 144 size = num * sizeof(struct vring_desc); in vring_size() 148 (num * sizeof(struct vring_used_elem)); in vring_size() 153 unsigned int num) in vring_init_split() argument 155 vr->num = num; in vring_init_split() 158 num * sizeof(struct vring_desc)); in vring_init_split() 165 unsigned int num) in vring_init_packed() argument 167 vr->num = num; in vring_init_packed() [all …]
|
| H A D | virtio_rxtx_packed.c | 86 uint16_t num, nb_rx = 0; in virtio_recv_pkts_packed_vec() local 93 num = RTE_MIN(VIRTIO_MBUF_BURST_SZ, nb_pkts); in virtio_recv_pkts_packed_vec() 94 if (likely(num > PACKED_BATCH_SIZE)) in virtio_recv_pkts_packed_vec() 95 num = num - ((vq->vq_used_cons_idx + num) % PACKED_BATCH_SIZE); in virtio_recv_pkts_packed_vec() 97 while (num) { in virtio_recv_pkts_packed_vec() 98 if (num >= PACKED_BATCH_SIZE) { in virtio_recv_pkts_packed_vec() 102 num -= PACKED_BATCH_SIZE; in virtio_recv_pkts_packed_vec() 109 num--; in virtio_recv_pkts_packed_vec() 115 PMD_RX_LOG(DEBUG, "dequeue:%d", num); in virtio_recv_pkts_packed_vec()
|
| H A D | virtio_rxtx.c | 207 uint16_t num) in virtqueue_dequeue_rx_inorder() argument 244 uint16_t num) in virtqueue_enqueue_refill_inorder() argument 283 uint16_t num) in virtqueue_enqueue_recv_refill() argument 444 uint16_t num) in virtqueue_enqueue_xmit_inorder() argument 1007 num = num - ((vq->vq_used_cons_idx + num) % DESC_PER_CACHELINE); in virtio_recv_pkts() 1009 num = virtqueue_dequeue_burst_rx(vq, rcv_pkts, len, num); in virtio_recv_pkts() 1110 num = num - ((vq->vq_used_cons_idx + num) % DESC_PER_CACHELINE); in virtio_recv_pkts_packed() 1405 num = num - ((vq->vq_used_cons_idx + num) % in virtio_recv_mergeable_pkts() 1409 num = virtqueue_dequeue_burst_rx(vq, rcv_pkts, len, num); in virtio_recv_mergeable_pkts() 1576 num = nb_pkts; in virtio_recv_mergeable_pkts_packed() [all …]
|
| /dpdk/drivers/common/cnxk/ |
| H A D | roc_npa.h | 290 unsigned int num) in roc_npa_aura_batch_alloc_extract() argument 352 while (num) { in roc_npa_aura_op_batch_alloc() 355 num; in roc_npa_aura_op_batch_alloc() 366 num -= num_alloc; in roc_npa_aura_op_batch_alloc() 413 for (i = 0; i < num; i++) in roc_npa_aura_batch_free() 449 while (num) { in roc_npa_aura_batch_free_burst() 479 num -= bnum; in roc_npa_aura_op_batch_free() 489 if (num) in roc_npa_aura_op_batch_free() 509 switch (num) { in roc_npa_aura_bulk_alloc() 674 while (num) { in roc_npa_aura_op_bulk_alloc() [all …]
|
| /dpdk/lib/acl/ |
| H A D | acl_run_avx512.c | 119 while (num > max_iter) { in rte_acl_classify_avx512x16() 123 num -= max_iter; in rte_acl_classify_avx512x16() 127 if (num >= MAX_SEARCHES_AVX16) in rte_acl_classify_avx512x16() 129 if (num >= MAX_SEARCHES_SSE8) in rte_acl_classify_avx512x16() 131 if (num >= MAX_SEARCHES_SSE4) in rte_acl_classify_avx512x16() 146 while (num > max_iter) { in rte_acl_classify_avx512x32() 150 num -= max_iter; in rte_acl_classify_avx512x32() 154 if (num >= 2 * MAX_SEARCHES_AVX16) in rte_acl_classify_avx512x32() 156 if (num >= MAX_SEARCHES_AVX16) in rte_acl_classify_avx512x32() 158 if (num >= MAX_SEARCHES_SSE8) in rte_acl_classify_avx512x32() [all …]
|
| H A D | acl_run_avx2.c | 14 uint32_t *results, uint32_t num, uint32_t categories) in rte_acl_classify_avx2() argument 16 if (likely(num >= MAX_SEARCHES_AVX16)) in rte_acl_classify_avx2() 17 return search_avx2x16(ctx, data, results, num, categories); in rte_acl_classify_avx2() 18 else if (num >= MAX_SEARCHES_SSE8) in rte_acl_classify_avx2() 19 return search_sse_8(ctx, data, results, num, categories); in rte_acl_classify_avx2() 20 else if (num >= MAX_SEARCHES_SSE4) in rte_acl_classify_avx2() 21 return search_sse_4(ctx, data, results, num, categories); in rte_acl_classify_avx2() 23 return rte_acl_classify_scalar(ctx, data, results, num, in rte_acl_classify_avx2()
|
| H A D | acl_run_neon.c | 9 uint32_t *results, uint32_t num, uint32_t categories) in rte_acl_classify_neon() argument 11 if (likely(num >= 8)) in rte_acl_classify_neon() 12 return search_neon_8(ctx, data, results, num, categories); in rte_acl_classify_neon() 13 else if (num >= 4) in rte_acl_classify_neon() 14 return search_neon_4(ctx, data, results, num, categories); in rte_acl_classify_neon() 16 return rte_acl_classify_scalar(ctx, data, results, num, in rte_acl_classify_neon()
|
| H A D | acl_run_sse.c | 9 uint32_t *results, uint32_t num, uint32_t categories) in rte_acl_classify_sse() argument 11 if (likely(num >= MAX_SEARCHES_SSE8)) in rte_acl_classify_sse() 12 return search_sse_8(ctx, data, results, num, categories); in rte_acl_classify_sse() 13 else if (num >= MAX_SEARCHES_SSE4) in rte_acl_classify_sse() 14 return search_sse_4(ctx, data, results, num, categories); in rte_acl_classify_sse() 16 return rte_acl_classify_scalar(ctx, data, results, num, in rte_acl_classify_sse()
|
| H A D | acl_run_altivec.c | 10 uint32_t *results, uint32_t num, uint32_t categories) in rte_acl_classify_altivec() argument 12 if (likely(num >= MAX_SEARCHES_ALTIVEC8)) in rte_acl_classify_altivec() 13 return search_altivec_8(ctx, data, results, num, categories); in rte_acl_classify_altivec() 14 else if (num >= MAX_SEARCHES_ALTIVEC4) in rte_acl_classify_altivec() 15 return search_altivec_4(ctx, data, results, num, categories); in rte_acl_classify_altivec() 17 return rte_acl_classify_scalar(ctx, data, results, num, in rte_acl_classify_altivec()
|
| /dpdk/lib/ring/ |
| H A D | rte_ring_peek_elem_pvt.h | 32 uint32_t num) in __rte_ring_st_get_tail() argument 40 RTE_ASSERT(n >= num); in __rte_ring_st_get_tail() 41 num = (n >= num) ? num : 0; in __rte_ring_st_get_tail() 44 return num; in __rte_ring_st_get_tail() 54 uint32_t num, uint32_t enqueue) in __rte_ring_st_set_head_tail() argument 60 pos = tail + num; in __rte_ring_st_set_head_tail() 76 uint32_t num) in __rte_ring_hts_get_tail() argument 84 RTE_ASSERT(n >= num); in __rte_ring_hts_get_tail() 85 num = (n >= num) ? num : 0; in __rte_ring_hts_get_tail() 88 return num; in __rte_ring_hts_get_tail() [all …]
|
| /dpdk/drivers/crypto/virtio/ |
| H A D | virtio_ring.h | 61 unsigned int num; member 97 #define vring_used_event(vr) ((vr)->avail->ring[(vr)->num]) 98 #define vring_avail_event(vr) (*(uint16_t *)&(vr)->used->ring[(vr)->num]) 101 vring_size(unsigned int num, unsigned long align) in vring_size() argument 105 size = num * sizeof(struct vring_desc); in vring_size() 106 size += sizeof(struct vring_avail) + (num * sizeof(uint16_t)); in vring_size() 109 (num * sizeof(struct vring_used_elem)); in vring_size() 114 vring_init(struct vring *vr, unsigned int num, uint8_t *p, in vring_init() argument 117 vr->num = num; in vring_init() 120 num * sizeof(struct vring_desc)); in vring_init() [all …]
|
| /dpdk/lib/ipsec/ |
| H A D | esp_outb.c | 306 n = num; in esp_outb_tun_prepare() 308 if (n != num) in esp_outb_tun_prepare() 454 n = num; in esp_outb_trs_prepare() 456 if (n != num) in esp_outb_trs_prepare() 550 n = num; in cpu_outb_pkt_prepare() 656 move_bad_mbufs(mb, dr, num, num - k); in esp_outb_sqh_process() 745 num = esn_outb_nb_valid_packets(num, n_sqn, nb_segs); in inline_outb_tun_pkt_process() 773 move_bad_mbufs(mb, dr, num, num - k); in inline_outb_tun_pkt_process() 813 num = esn_outb_nb_valid_packets(num, n_sqn, nb_segs); in inline_outb_trs_pkt_process() 841 move_bad_mbufs(mb, dr, num, num - k); in inline_outb_trs_pkt_process() [all …]
|
| H A D | sa.h | 158 struct rte_mbuf *mb[], uint16_t num); 162 struct rte_mbuf *mb[], uint16_t num); 166 struct rte_mbuf *mb[], uint16_t num); 170 struct rte_mbuf *mb[], uint16_t num); 174 struct rte_mbuf *mb[], uint16_t num); 188 uint16_t num); 192 struct rte_mbuf *mb[], uint16_t num); 196 struct rte_mbuf *mb[], uint16_t num); 200 struct rte_mbuf *mb[], uint16_t num); 204 struct rte_mbuf *mb[], uint16_t num); [all …]
|
| H A D | esp_inb.c | 373 uint32_t dr[num]; in esp_inb_pkt_prepare() 398 move_bad_mbufs(mb, dr, num, num - k); in esp_inb_pkt_prepare() 579 uint32_t hl[num], to[num]; in tun_process() 643 uint32_t hl[num], to[num]; in trs_process() 703 return num; in esp_inb_rsn_update() 727 uint32_t sqn[num]; in esp_inb_pkt_process() 728 uint32_t dr[num]; in esp_inb_pkt_process() 735 move_bad_mbufs(mb, dr, num, num - k); in esp_inb_pkt_process() 744 if (n != num) in esp_inb_pkt_process() 766 uint32_t dr[num]; in cpu_inb_pkt_prepare() [all …]
|
| H A D | rte_ipsec.h | 40 uint16_t num); 43 uint16_t num); 47 uint16_t num); 119 struct rte_mbuf *mb[], struct rte_crypto_op *cop[], uint16_t num) in rte_ipsec_pkt_crypto_prepare() argument 121 return ss->pkt_func.prepare.async(ss, mb, cop, num); in rte_ipsec_pkt_crypto_prepare() 126 struct rte_mbuf *mb[], uint16_t num) in rte_ipsec_pkt_cpu_prepare() argument 128 return ss->pkt_func.prepare.sync(ss, mb, num); in rte_ipsec_pkt_cpu_prepare() 156 uint16_t num) in rte_ipsec_pkt_process() argument 158 return ss->pkt_func.process(ss, mb, num); in rte_ipsec_pkt_process()
|
| /dpdk/drivers/net/bnxt/ |
| H A D | bnxt_rxtx_vec_common.h | 107 uint32_t ring_mask, ring_size, num; in bnxt_tx_cmp_vec_fast() local 114 num = RTE_MIN(nr_pkts, ring_size - cons); in bnxt_tx_cmp_vec_fast() 119 raw_cons += num; in bnxt_tx_cmp_vec_fast() 120 num = nr_pkts - num; in bnxt_tx_cmp_vec_fast() 121 if (num) { in bnxt_tx_cmp_vec_fast() 124 num); in bnxt_tx_cmp_vec_fast() 126 num * sizeof(struct rte_mbuf *)); in bnxt_tx_cmp_vec_fast() 127 raw_cons += num; in bnxt_tx_cmp_vec_fast() 138 uint32_t ring_mask, ring_size, num, blk; in bnxt_tx_cmp_vec() local 148 num = RTE_MIN(nr_pkts, ring_size - cons); in bnxt_tx_cmp_vec() [all …]
|
| /dpdk/drivers/net/mvneta/ |
| H A D | mvneta_rxtx.c | 80 uint16_t nb_desc = *num; in mvneta_buffs_refill() 85 *num = 0; in mvneta_buffs_refill() 97 *num = 0; in mvneta_buffs_refill() 125 nb_desc = *num; in mvneta_buffs_alloc() 141 *num = sent; in mvneta_buffs_alloc() 420 num = nb_pkts; in mvneta_tx_pkt_burst() 469 pkts.num = 0; in mvneta_tx_sg_pkt_burst() 516 pkts.num++; in mvneta_tx_sg_pkt_burst() 548 num = total_descs; in mvneta_tx_sg_pkt_burst() 568 nb_pkts = pkts.num; in mvneta_tx_sg_pkt_burst() [all …]
|
| /dpdk/kernel/linux/kni/ |
| H A D | kni_net.c | 111 uint32_t num; in kni_net_process_request() local 131 if (num < 1) { in kni_net_process_request() 487 num = min(num, num_aq); in kni_net_rx_lo_fifo() 488 num = min(num, num_fq); in kni_net_rx_lo_fifo() 489 num = min_t(uint32_t, num, MBUF_BURST_SZ); in kni_net_rx_lo_fifo() 492 if (num == 0) in kni_net_rx_lo_fifo() 503 num = ret; in kni_net_rx_lo_fifo() 539 if (ret != num) in kni_net_rx_lo_fifo() 573 num = min_t(uint32_t, num, MBUF_BURST_SZ); in kni_net_rx_lo_fifo_skb() 576 if (num == 0) in kni_net_rx_lo_fifo_skb() [all …]
|
| /dpdk/examples/vm_power_manager/guest_cli/ |
| H A D | parse.c | 20 parse_set(const char *input, uint16_t set[], unsigned int num) in parse_set() argument 27 memset(set, 0, num * sizeof(uint16_t)); in parse_set() 41 min = num; in parse_set() 53 if (errno || end == NULL || idx >= num) in parse_set() 60 if (min == num) in parse_set() 67 if (min == num) in parse_set() 74 min = num; in parse_set()
|
| /dpdk/examples/ipsec-secgw/ |
| H A D | ipsec_process.c | 44 if (num >= RTE_DIM(cqp->buf) * 3 / 4 && len == 0) { in enqueue_cop_bulk() 47 free_cops(cop + n, num - n); in enqueue_cop_bulk() 55 n = RTE_MIN(num - k, n); in enqueue_cop_bulk() 74 } while (k != num); in enqueue_cop_bulk() 113 struct rte_ipsec_group grp[], uint32_t num) in sa_group() argument 121 for (i = 0, n = 0; i != num; i++) { in sa_group() 146 uint32_t num) in copy_to_trf() argument 171 for (j = 0, s = out->num; j != num; j++) { in copy_to_trf() 177 out->num += num; in copy_to_trf() 404 trf->ip4.num = 0; in ipsec_cqp_process() [all …]
|
| /dpdk/app/test/ |
| H A D | test_ring_stress_impl.h | 155 for (i = 0; i != num; i++) { in check_updt_elem() 160 __func__, rte_lcore_id(), num, i, elm[i]); in check_updt_elem() 188 uint32_t lc, n, num; in test_worker() local 222 rc = check_ring_op(num, n, lc, fname, in test_worker() 240 rc = check_ring_op(num, n, lc, fname, in test_worker() 287 sz = num * sizeof(*elm); in mt1_init() 291 __func__, sz, num); in mt1_init() 298 nr = 2 * num; in mt1_init() 316 for (i = 0; i != num; i++) { in mt1_init() 323 if (i != num) { in mt1_init() [all …]
|
| H A D | test_distributor.c | 92 unsigned int num; in handle_work() local 100 buf, buf, num); in handle_work() 312 unsigned int num; in handle_work_with_free_mbufs() local 319 for (i = 0; i < num; i++) in handle_work_with_free_mbufs() 380 unsigned int num; in handle_work_for_shutdown_test() local 388 if (num > 0) { in handle_work_for_shutdown_test() 402 if (num > 0) { in handle_work_for_shutdown_test() 427 num, __ATOMIC_RELAXED); in handle_work_for_shutdown_test() 597 unsigned int num, i; in handle_and_mark_work() local 603 for (i = 0; i < num; i++) in handle_and_mark_work() [all …]
|
| /dpdk/drivers/event/opdl/ |
| H A D | opdl_evdev_init.c | 21 uint16_t num, in enqueue_check() argument 64 if (num > 0 && in enqueue_check() 71 return num; in enqueue_check() 77 uint16_t num, in update_on_dequeue() argument 97 if (num > 0) in update_on_dequeue() 113 uint16_t num) in opdl_rx_error_enqueue() argument 135 uint16_t num) in opdl_rx_enqueue() argument 141 num, in opdl_rx_enqueue() 193 num, in opdl_tx_dequeue_single_thread() 217 num, in opdl_tx_dequeue_multi_inst() [all …]
|
| /dpdk/drivers/bus/dpaa/base/qbman/ |
| H A D | qman_priv.h | 35 static inline int qman_cgrs_get(struct qman_cgrs *c, int num) in qman_cgrs_get() argument 37 return QM_MCR_QUERYCONGESTION(&c->q, num); in qman_cgrs_get() 40 static inline void qman_cgrs_set(struct qman_cgrs *c, int num) in qman_cgrs_set() argument 42 c->q.state[__CGR_WORD(num)] |= (0x80000000 >> __CGR_SHIFT(num)); in qman_cgrs_set() 45 static inline void qman_cgrs_unset(struct qman_cgrs *c, int num) in qman_cgrs_unset() argument 47 c->q.state[__CGR_WORD(num)] &= ~(0x80000000 >> __CGR_SHIFT(num)); in qman_cgrs_unset() 50 static inline int qman_cgrs_next(struct qman_cgrs *c, int num) in qman_cgrs_next() argument 52 while ((++num < (int)__CGR_NUM) && !qman_cgrs_get(c, num)) in qman_cgrs_next() 54 return num; in qman_cgrs_next()
|
| /dpdk/lib/bpf/ |
| H A D | bpf_pkt.c | 165 struct rte_mbuf *dr[num]; in apply_filter() 195 void *dp[num]; in pkt_filter_vm() 196 uint64_t rc[num]; in pkt_filter_vm() 198 for (i = 0; i != num; i++) in pkt_filter_vm() 211 uint64_t rc[num]; in pkt_filter_jit() 221 num = apply_filter(mb, rc, num, drop); in pkt_filter_jit() 223 return num; in pkt_filter_jit() 230 uint64_t rc[num]; in pkt_filter_mb_vm() 241 uint64_t rc[num]; in pkt_filter_mb_jit() 250 num = apply_filter(mb, rc, num, drop); in pkt_filter_mb_jit() [all …]
|