| /f-stack/dpdk/lib/librte_gso/ |
| H A D | gso_common.c | 18 hdr_segment->nb_segs = 1; in hdr_segment_init() 53 uint16_t pyld_len, nb_segs; in gso_do_segment() local 57 nb_segs = 0; in gso_do_segment() 62 if (unlikely(nb_segs >= nb_pkts_out)) { in gso_do_segment() 63 free_gso_segment(pkts_out, nb_segs); in gso_do_segment() 70 free_gso_segment(pkts_out, nb_segs); in gso_do_segment() 85 free_gso_segment(pkts_out, nb_segs); in gso_do_segment() 104 hdr_segment->nb_segs++; in gso_do_segment() 121 pkts_out[nb_segs++] = hdr_segment; in gso_do_segment() 123 return nb_segs; in gso_do_segment()
|
| H A D | gso_udp4.c | 12 uint16_t nb_segs) in update_ipv4_udp_headers() argument 17 uint16_t tail_idx = nb_segs - 1, length, i; in update_ipv4_udp_headers() 24 for (i = 0; i < nb_segs; i++) { in update_ipv4_udp_headers()
|
| H A D | gso_tcp4.c | 10 struct rte_mbuf **segs, uint16_t nb_segs) in update_ipv4_tcp_headers() argument 24 tail_idx = nb_segs - 1; in update_ipv4_tcp_headers() 26 for (i = 0; i < nb_segs; i++) { in update_ipv4_tcp_headers()
|
| H A D | gso_tunnel_tcp4.c | 10 struct rte_mbuf **segs, uint16_t nb_segs) in update_tunnel_ipv4_tcp_headers() argument 37 tail_idx = nb_segs - 1; in update_tunnel_ipv4_tcp_headers() 42 for (i = 0; i < nb_segs; i++) { in update_tunnel_ipv4_tcp_headers()
|
| /f-stack/dpdk/kernel/linux/kni/ |
| H A D | kni_net.c | 202 int nb_segs; in kni_fifo_trans_pa2va() local 220 kva_nb_segs = kva->nb_segs; in kni_fifo_trans_pa2va() 221 for (nb_segs = 0; nb_segs < kva_nb_segs; nb_segs++) { in kni_fifo_trans_pa2va() 389 if (kva->nb_segs == 1) { in kni_net_rx_normal() 392 int nb_segs; in kni_net_rx_normal() local 393 int kva_nb_segs = kva->nb_segs; in kni_net_rx_normal() 395 for (nb_segs = 0; nb_segs < kva_nb_segs; nb_segs++) { in kni_net_rx_normal() 575 if (kva->nb_segs == 1) { in kni_net_rx_lo_fifo_skb() 578 int nb_segs; in kni_net_rx_lo_fifo_skb() local 579 int kva_nb_segs = kva->nb_segs; in kni_net_rx_lo_fifo_skb() [all …]
|
| /f-stack/dpdk/drivers/event/octeontx/ |
| H A D | ssovf_worker.h | 64 uint8_t nb_segs; in ssovf_octeontx_wqe_xtract_mseg() local 67 nb_segs = wqe->s.w0.bufs; in ssovf_octeontx_wqe_xtract_mseg() 72 while (--nb_segs) { in ssovf_octeontx_wqe_xtract_mseg() 81 if (nb_segs == 1) in ssovf_octeontx_wqe_xtract_mseg() 113 mbuf->nb_segs = wqe->s.w0.bufs; in ssovf_octeontx_wqe_to_pkt() 117 mbuf->nb_segs = 1; in ssovf_octeontx_wqe_to_pkt() 140 uint8_t nb_segs = wqe->s.w0.bufs; in ssovf_octeontx_wqe_free() local 149 while (--nb_segs) { in ssovf_octeontx_wqe_free()
|
| /f-stack/dpdk/app/test-crypto-perf/ |
| H A D | cperf_test_verify.c | 101 uint16_t nb_segs; in cperf_verify_op() local 114 nb_segs = m->nb_segs; in cperf_verify_op() 116 while (m && nb_segs != 0) { in cperf_verify_op() 119 nb_segs--; in cperf_verify_op() 130 nb_segs = m->nb_segs; in cperf_verify_op() 132 while (m && nb_segs != 0) { in cperf_verify_op() 137 nb_segs--; in cperf_verify_op()
|
| /f-stack/dpdk/lib/librte_mbuf/ |
| H A D | rte_mbuf.c | 106 m->nb_segs = 1; in rte_pktmbuf_init() 134 m->nb_segs = 1; in rte_pktmbuf_free_pinned_extmem() 209 m->nb_segs = 1; in __rte_pktmbuf_init_extmem() 422 nb_segs = m->nb_segs; in rte_mbuf_check() 434 nb_segs -= 1; in rte_mbuf_check() 438 if (nb_segs) { in rte_mbuf_check() 542 mc->nb_segs = nseg; in rte_pktmbuf_clone() 587 mbuf->nb_segs = 1; in __rte_pktmbuf_linearize() 638 ++mc->nb_segs; in rte_pktmbuf_copy() 688 nb_segs = m->nb_segs; in rte_pktmbuf_dump() [all …]
|
| H A D | rte_mbuf.h | 560 RTE_ASSERT(m->nb_segs == 1); in __rte_mbuf_raw_sanity_check() 868 m->nb_segs = 1; in rte_pktmbuf_reset() 1173 mi->nb_segs = 1; in rte_pktmbuf_attach() 1213 md->nb_segs = 1; in __rte_pktmbuf_free_direct() 1345 m->nb_segs = 1; in rte_pktmbuf_prefree_seg() 1362 m->nb_segs = 1; in rte_pktmbuf_prefree_seg() 1693 return m->nb_segs == 1; in rte_pktmbuf_is_contiguous() 1752 if (head->nb_segs + tail->nb_segs > RTE_MBUF_MAX_NB_SEGS) in rte_pktmbuf_chain() 1763 head->nb_segs = (uint16_t)(head->nb_segs + tail->nb_segs); in rte_pktmbuf_chain()
|
| /f-stack/dpdk/app/test/ |
| H A D | test_cryptodev.h | 161 int nb_segs, uint8_t pattern) { in create_segmented_mbuf() argument 174 if (nb_segs < 1) { in create_segmented_mbuf() 176 nb_segs); in create_segmented_mbuf() 180 t_len = pkt_len >= nb_segs ? pkt_len / nb_segs : 1; in create_segmented_mbuf()
|
| /f-stack/dpdk/drivers/net/mlx4/ |
| H A D | mlx4_rxtx.c | 432 buf->nb_segs; in mlx4_tx_burst_tso_get_params() 471 int nb_segs = buf->nb_segs; in mlx4_tx_burst_fill_tso_dsegs() local 747 int nb_segs = buf->nb_segs; in mlx4_tx_burst_segs() local 795 nb_segs--; in mlx4_tx_burst_segs() 798 switch (nb_segs) { in mlx4_tx_burst_segs() 813 nb_segs--; in mlx4_tx_burst_segs() 829 nb_segs--; in mlx4_tx_burst_segs() 843 nb_segs--; in mlx4_tx_burst_segs() 844 if (nb_segs) { in mlx4_tx_burst_segs() 1298 rep->nb_segs = 1; in mlx4_rx_burst() [all …]
|
| /f-stack/dpdk/drivers/net/thunderx/ |
| H A D | nicvf_rxtx.c | 39 sqe.hdr.subdesc_cnt = pkt->nb_segs; in fill_sq_desc_header() 195 uint16_t nb_segs; in nicvf_xmit_pkts_multiseg() local 206 nb_segs = pkt->nb_segs; in nicvf_xmit_pkts_multiseg() 208 next_used_desc = used_desc + nb_segs + 1; in nicvf_xmit_pkts_multiseg() 212 used_bufs += nb_segs; in nicvf_xmit_pkts_multiseg() 223 for (k = 1; k < nb_segs; k++) { in nicvf_xmit_pkts_multiseg() 529 uint16_t *rb_sz, nb_segs, seg_idx; in nicvf_process_cq_mseg_entry() local 537 nb_segs = cqe_rx_w0.rb_cnt; in nicvf_process_cq_mseg_entry() 572 return nb_segs; in nicvf_process_cq_mseg_entry() 587 uint16_t nb_segs; in nicvf_recv_pkts_multiseg() local [all …]
|
| H A D | nicvf_rxtx.h | 74 uint16_t apad, uint16_t nb_segs) in nicvf_mbuff_init_mseg_update() argument 82 init.fields.nb_segs = nb_segs; in nicvf_mbuff_init_mseg_update()
|
| /f-stack/dpdk/drivers/net/octeontx2/ |
| H A D | otx2_rx.h | 62 uint16_t nb_segs; member 182 uint8_t nb_segs; in nix_cqe_xtract_mseg() local 186 nb_segs = (sg >> 48) & 0x3; in nix_cqe_xtract_mseg() 187 mbuf->nb_segs = nb_segs; in nix_cqe_xtract_mseg() 194 nb_segs--; in nix_cqe_xtract_mseg() 199 while (nb_segs) { in nix_cqe_xtract_mseg() 208 nb_segs--; in nix_cqe_xtract_mseg() 211 if (!nb_segs && (iova_list + 1 < eol)) { in nix_cqe_xtract_mseg() 213 nb_segs = (sg >> 48) & 0x3; in nix_cqe_xtract_mseg() 214 head->nb_segs += nb_segs; in nix_cqe_xtract_mseg()
|
| H A D | otx2_tx.h | 108 m->nb_segs = 1; in otx2_pktmbuf_detach() 118 md->nb_segs = 1; in otx2_pktmbuf_detach() 133 m->nb_segs = 1; in otx2_nix_prefree_seg() 141 m->nb_segs = 1; in otx2_nix_prefree_seg() 415 uint64_t nb_segs; in otx2_nix_prepare_mseg() local 435 nb_segs = m->nb_segs; in otx2_nix_prepare_mseg() 456 nb_segs--; in otx2_nix_prepare_mseg() 457 if (i > 2 && nb_segs) { in otx2_nix_prepare_mseg() 468 } while (nb_segs); in otx2_nix_prepare_mseg()
|
| /f-stack/dpdk/app/test-pmd/ |
| H A D | txonly.c | 195 uint32_t nb_segs, pkt_len; in pkt_burst_prepare() local 199 nb_segs = rte_rand() % tx_pkt_nb_segs + 1; in pkt_burst_prepare() 201 nb_segs = tx_pkt_nb_segs; in pkt_burst_prepare() 203 if (nb_segs > 1) { in pkt_burst_prepare() 204 if (rte_mempool_get_bulk(mbp, (void **)pkt_segs, nb_segs - 1)) in pkt_burst_prepare() 219 for (i = 1; i < nb_segs; i++) { in pkt_burst_prepare() 311 pkt->nb_segs = nb_segs; in pkt_burst_prepare()
|
| H A D | parameters.c | 906 unsigned int nb_segs, i; in launch_args_parse() local 910 if (nb_segs <= 0) in launch_args_parse() 913 for (i = 0; i < nb_segs; i++) { in launch_args_parse() 920 mbuf_data_size_n = nb_segs; in launch_args_parse() 1309 unsigned int nb_segs; in launch_args_parse() local 1311 nb_segs = parse_item_list in launch_args_parse() 1315 if (nb_segs > 0) in launch_args_parse() 1316 set_rx_pkt_segments(seg_len, nb_segs); in launch_args_parse() 1322 unsigned int nb_segs; in launch_args_parse() local 1326 if (nb_segs > 0) in launch_args_parse() [all …]
|
| /f-stack/dpdk/drivers/net/octeontx/ |
| H A D | octeontx_rxtx.h | 188 m->nb_segs = 1; in octeontx_pktmbuf_detach() 198 md->nb_segs = 1; in octeontx_pktmbuf_detach() 213 m->nb_segs = 1; in octeontx_prefree_seg() 221 m->nb_segs = 1; in octeontx_prefree_seg() 356 uint16_t nb_segs, nb_desc = 0; in __octeontx_xmit_mseg_prepare() local 360 nb_segs = tx_pkt->nb_segs; in __octeontx_xmit_mseg_prepare() 404 nb_segs--; in __octeontx_xmit_mseg_prepare() 407 } while (nb_segs); in __octeontx_xmit_mseg_prepare()
|
| /f-stack/dpdk/drivers/net/dpaa/ |
| H A D | dpaa_rxtx.c | 366 first_seg->nb_segs = 1; in dpaa_eth_sg_to_mbuf() 379 first_seg->nb_segs += 1; in dpaa_eth_sg_to_mbuf() 428 mbuf->nb_segs = 1; in dpaa_eth_fd_to_mbuf() 469 first_seg->nb_segs = 1; in dpaa_free_mbuf() 478 first_seg->nb_segs += 1; in dpaa_free_mbuf() 548 mbuf->nb_segs = 1; in dpaa_rx_cb_no_prefetch() 588 mbuf->nb_segs = 1; in dpaa_rx_cb() 949 if (mbuf->nb_segs == 1) { in tx_on_dpaa_pool() 1024 new_mbufs[0]->nb_segs++; in reallocate_mbuf() 1032 new_mbufs[0]->nb_segs++; in reallocate_mbuf() [all …]
|
| /f-stack/dpdk/drivers/net/mvneta/ |
| H A D | mvneta_rxtx.c | 464 uint16_t nb_segs, total_descs = 0; in mvneta_tx_sg_pkt_burst() local 488 nb_segs = mbuf->nb_segs; in mvneta_tx_sg_pkt_burst() 489 total_descs += nb_segs; in mvneta_tx_sg_pkt_burst() 496 total_descs -= nb_segs; in mvneta_tx_sg_pkt_burst() 508 if (unlikely(nb_segs > NETA_PPIO_DESC_NUM_FRAGS)) { in mvneta_tx_sg_pkt_burst() 509 total_descs -= nb_segs; in mvneta_tx_sg_pkt_burst() 515 pkts.frags[pkts.num] = nb_segs; in mvneta_tx_sg_pkt_burst() 520 for (j = 0; j < nb_segs - 1; j++) { in mvneta_tx_sg_pkt_burst()
|
| /f-stack/dpdk/drivers/net/fm10k/ |
| H A D | fm10k_rxtx.c | 272 first_seg->nb_segs++; in fm10k_recv_scattered_pkts() 586 last_id = q->next_free + mb->nb_segs - 1; in tx_xmit_pkt() 592 if ((q->nb_used + mb->nb_segs) >= q->rs_thresh) { in tx_xmit_pkt() 597 q->nb_used = q->nb_used + mb->nb_segs; in tx_xmit_pkt() 600 q->nb_free -= mb->nb_segs; in tx_xmit_pkt() 673 if (q->nb_free < mb->nb_segs) in fm10k_xmit_pkts() 677 if ((mb->nb_segs == 0) || in fm10k_xmit_pkts() 678 ((mb->nb_segs > 1) && (mb->next == NULL))) in fm10k_xmit_pkts()
|
| /f-stack/dpdk/drivers/net/hinic/ |
| H A D | hinic_pmd_tx.c | 324 u16 nb_segs = sqe_info->sge_cnt - sqe_info->cpy_mbuf_cnt; in hinic_mbuf_dma_map_sge() local 325 u16 real_nb_segs = mbuf->nb_segs; in hinic_mbuf_dma_map_sge() 351 around_sges = nb_segs - i; in hinic_mbuf_dma_map_sge() 354 for (; i < nb_segs; i++) { in hinic_mbuf_dma_map_sge() 375 for (i = 0; i < nb_segs; i++) { in hinic_mbuf_dma_map_sge() 398 real_nb_segs - nb_segs); in hinic_mbuf_dma_map_sge() 458 left_sges = mbuf->nb_segs; in hinic_is_tso_sge_valid() 496 sqe_info->sge_cnt = mbuf_head->nb_segs - in hinic_is_tso_sge_valid() 509 sqe_info->sge_cnt = mbuf_head->nb_segs; in hinic_is_tso_sge_valid() 640 if (likely(mbuf->nb_segs == 1)) { in hinic_xmit_mbuf_cleanup() [all …]
|
| /f-stack/dpdk/drivers/net/i40e/ |
| H A D | i40e_rxtx_vec_common.h | 29 start->nb_segs++; in reassemble_packets() 46 start->nb_segs--; in reassemble_packets() 180 mb_def.nb_segs = 1; in i40e_rxq_vec_setup_default()
|
| /f-stack/dpdk/drivers/net/iavf/ |
| H A D | iavf_rxtx_vec_common.h | 29 start->nb_segs++; in reassemble_packets() 46 start->nb_segs--; in reassemble_packets() 200 mb_def.nb_segs = 1; in iavf_rxq_vec_setup_default()
|
| /f-stack/lib/ |
| H A D | ff_memory.c | 412 p_head->nb_segs = 0; in ff_bsd_to_rte() 429 p_head->nb_segs++; in ff_bsd_to_rte() 477 int ff_enq_tx_bsdmbuf(uint8_t portid, void *p_mbuf, int nb_segs) in ff_enq_tx_bsdmbuf() argument 479 return ff_txring_enqueue(&nic_tx_ring[portid], p_mbuf, nb_segs); in ff_enq_tx_bsdmbuf()
|