| /f-stack/dpdk/lib/librte_gro/ |
| H A D | rte_gro.c | 246 ret = gro_vxlan_tcp4_reassemble(pkts[i], in rte_gro_reassemble_burst() 255 ret = gro_vxlan_udp4_reassemble(pkts[i], in rte_gro_reassemble_burst() 279 unprocess_pkts[unprocess_num++] = pkts[i]; in rte_gro_reassemble_burst() 288 0, pkts, nb_pkts); in rte_gro_reassemble_burst() 293 0, &pkts[i], nb_pkts - i); in rte_gro_reassemble_burst() 299 &pkts[i], nb_pkts - i); in rte_gro_reassemble_burst() 304 &pkts[i], nb_pkts - i); in rte_gro_reassemble_burst() 308 memcpy(&pkts[i], unprocess_pkts, in rte_gro_reassemble_burst() 319 rte_gro_reassemble(struct rte_mbuf **pkts, in rte_gro_reassemble() argument 365 if (gro_tcp4_reassemble(pkts[i], tcp_tbl, in rte_gro_reassemble() [all …]
|
| /f-stack/dpdk/lib/librte_node/ |
| H A D | ethdev_rx.c | 74 struct rte_mbuf **pkts; in eth_pkt_parse_cb() local 81 pkts = mbufs; in eth_pkt_parse_cb() 86 rte_prefetch0(pkts[8]); in eth_pkt_parse_cb() 87 rte_prefetch0(pkts[9]); in eth_pkt_parse_cb() 101 mbuf0 = pkts[0]; in eth_pkt_parse_cb() 102 mbuf1 = pkts[1]; in eth_pkt_parse_cb() 103 mbuf2 = pkts[2]; in eth_pkt_parse_cb() 104 mbuf3 = pkts[3]; in eth_pkt_parse_cb() 105 pkts += 4; in eth_pkt_parse_cb() 130 mbuf0 = pkts[0]; in eth_pkt_parse_cb() [all …]
|
| H A D | ip4_lookup_sse.h | 48 rte_prefetch0(pkts[8]); in ip4_lookup_node_process_vec() 49 rte_prefetch0(pkts[9]); in ip4_lookup_node_process_vec() 50 rte_prefetch0(pkts[10]); in ip4_lookup_node_process_vec() 51 rte_prefetch0(pkts[11]); in ip4_lookup_node_process_vec() 66 mbuf0 = pkts[0]; in ip4_lookup_node_process_vec() 67 mbuf1 = pkts[1]; in ip4_lookup_node_process_vec() 68 mbuf2 = pkts[2]; in ip4_lookup_node_process_vec() 69 mbuf3 = pkts[3]; in ip4_lookup_node_process_vec() 71 pkts += 4; in ip4_lookup_node_process_vec() 194 mbuf0 = pkts[0]; in ip4_lookup_node_process_vec() [all …]
|
| H A D | ip4_lookup_neon.h | 52 rte_prefetch0(pkts[8]); in ip4_lookup_node_process_vec() 53 rte_prefetch0(pkts[9]); in ip4_lookup_node_process_vec() 54 rte_prefetch0(pkts[10]); in ip4_lookup_node_process_vec() 55 rte_prefetch0(pkts[11]); in ip4_lookup_node_process_vec() 70 mbuf0 = pkts[0]; in ip4_lookup_node_process_vec() 71 mbuf1 = pkts[1]; in ip4_lookup_node_process_vec() 72 mbuf2 = pkts[2]; in ip4_lookup_node_process_vec() 73 mbuf3 = pkts[3]; in ip4_lookup_node_process_vec() 75 pkts += 4; in ip4_lookup_node_process_vec() 189 mbuf0 = pkts[0]; in ip4_lookup_node_process_vec() [all …]
|
| H A D | pkt_cls.c | 54 rte_prefetch0(pkts[i]); in pkt_cls_node_process() 67 rte_prefetch0(pkts[4]); in pkt_cls_node_process() 68 rte_prefetch0(pkts[5]); in pkt_cls_node_process() 69 rte_prefetch0(pkts[6]); in pkt_cls_node_process() 74 mbuf0 = pkts[0]; in pkt_cls_node_process() 75 mbuf1 = pkts[1]; in pkt_cls_node_process() 76 mbuf2 = pkts[2]; in pkt_cls_node_process() 77 mbuf3 = pkts[3]; in pkt_cls_node_process() 78 pkts += 4; in pkt_cls_node_process() 173 mbuf0 = pkts[0]; in pkt_cls_node_process() [all …]
|
| H A D | ip4_rewrite.c | 62 rte_prefetch0(pkts[i]); in ip4_rewrite_node_process() 72 rte_prefetch0(pkts[4]); in ip4_rewrite_node_process() 73 rte_prefetch0(pkts[5]); in ip4_rewrite_node_process() 74 rte_prefetch0(pkts[6]); in ip4_rewrite_node_process() 78 mbuf0 = pkts[0]; in ip4_rewrite_node_process() 79 mbuf1 = pkts[1]; in ip4_rewrite_node_process() 80 mbuf2 = pkts[2]; in ip4_rewrite_node_process() 81 mbuf3 = pkts[3]; in ip4_rewrite_node_process() 83 pkts += 4; in ip4_rewrite_node_process() 214 mbuf0 = pkts[0]; in ip4_rewrite_node_process() [all …]
|
| /f-stack/dpdk/app/test-pmd/ |
| H A D | macswap_sse.h | 11 do_macswap(struct rte_mbuf *pkts[], uint16_t nb, in do_macswap() argument 31 vlan_qinq_set(pkts, nb, ol_flags, in do_macswap() 39 rte_prefetch0(rte_pktmbuf_mtod(pkts[i + 4], void *)); in do_macswap() 40 rte_prefetch0(rte_pktmbuf_mtod(pkts[i + 5], void *)); in do_macswap() 41 rte_prefetch0(rte_pktmbuf_mtod(pkts[i + 6], void *)); in do_macswap() 42 rte_prefetch0(rte_pktmbuf_mtod(pkts[i + 7], void *)); in do_macswap() 45 mb[0] = pkts[i++]; in do_macswap() 49 mb[1] = pkts[i++]; in do_macswap() 54 mb[2] = pkts[i++]; in do_macswap() 58 mb[3] = pkts[i++]; in do_macswap() [all …]
|
| H A D | macswap_neon.h | 16 do_macswap(struct rte_mbuf *pkts[], uint16_t nb, in do_macswap() argument 34 vlan_qinq_set(pkts, nb, ol_flags, in do_macswap() 42 rte_prefetch0(rte_pktmbuf_mtod(pkts[i + 4], void *)); in do_macswap() 43 rte_prefetch0(rte_pktmbuf_mtod(pkts[i + 5], void *)); in do_macswap() 44 rte_prefetch0(rte_pktmbuf_mtod(pkts[i + 6], void *)); in do_macswap() 48 mb[0] = pkts[i++]; in do_macswap() 51 mb[1] = pkts[i++]; in do_macswap() 54 mb[2] = pkts[i++]; in do_macswap() 57 mb[3] = pkts[i++]; in do_macswap() 84 rte_prefetch0(rte_pktmbuf_mtod(pkts[i+1], void *)); in do_macswap() [all …]
|
| H A D | macswap.h | 11 do_macswap(struct rte_mbuf *pkts[], uint16_t nb, in do_macswap() argument 21 vlan_qinq_set(pkts, nb, ol_flags, in do_macswap() 26 rte_prefetch0(rte_pktmbuf_mtod(pkts[i+1], void *)); in do_macswap() 27 mb = pkts[i]; in do_macswap()
|
| /f-stack/dpdk/drivers/net/mlx5/ |
| H A D | mlx5_rxtx_vec_sse.h | 51 _mm_storeu_si128((__m128i *)&pkts[pos], mbp); in rxq_copy_mbuf_v() 54 pkts[pos] = elts[pos]; in rxq_copy_mbuf_v() 380 __m128i op_err, struct rte_mbuf **pkts) in rxq_cq_to_ptype_oflags_v() argument 746 mlx5_timestamp_set(pkts[pos], offset, in rxq_cq_process_v() 749 mlx5_timestamp_set(pkts[pos + 1], offset, in rxq_cq_process_v() 752 mlx5_timestamp_set(pkts[pos + 2], offset, in rxq_cq_process_v() 755 mlx5_timestamp_set(pkts[pos + 3], offset, in rxq_cq_process_v() 758 mlx5_timestamp_set(pkts[pos], offset, in rxq_cq_process_v() 760 mlx5_timestamp_set(pkts[pos + 1], offset, in rxq_cq_process_v() 762 mlx5_timestamp_set(pkts[pos + 2], offset, in rxq_cq_process_v() [all …]
|
| H A D | mlx5_rxtx_vec.c | 59 struct rte_mbuf *pkt = pkts[i]; in rxq_handle_pending_error() 67 pkts[n++] = pkt; in rxq_handle_pending_error() 189 struct rte_mbuf **pkts, uint16_t pkts_n) in rxq_copy_mprq_mbuf_v() argument 245 pkts[copied++] = elts[i]; in rxq_copy_mprq_mbuf_v() 309 pkts, rcvd_pkt); in rxq_burst_v() 312 pkts += rcvd_pkt; in rxq_burst_v() 351 &pkts[nocmp_n], n); in rxq_burst_v() 386 nb_rx = rxq_burst_v(rxq, pkts + tn, pkts_n - tn, in mlx5_rx_burst_vec() 448 cp_pkt = rxq_copy_mprq_mbuf_v(rxq, pkts, rcvd_pkt); in rxq_burst_mprq_v() 450 pkts += cp_pkt; in rxq_burst_mprq_v() [all …]
|
| H A D | mlx5_rxtx_vec_altivec.h | 54 *(vector unsigned char *)&pkts[pos] = mbp; in rxq_copy_mbuf_v() 57 pkts[pos] = elts[pos]; in rxq_copy_mbuf_v() 528 struct rte_mbuf **pkts) in rxq_cq_to_ptype_oflags_v() argument 954 *(vector unsigned char *)&pkts[pos] = mbp1; in rxq_cq_process_v() 1012 &pkts[pos + 3]->pkt_len = pkt_mb3; in rxq_cq_process_v() 1014 &pkts[pos + 2]->pkt_len = pkt_mb2; in rxq_cq_process_v() 1081 &pkts[pos + 1]->pkt_len = pkt_mb1; in rxq_cq_process_v() 1083 &pkts[pos]->pkt_len = pkt_mb0; in rxq_cq_process_v() 1199 mlx5_timestamp_set(pkts[pos], offset, in rxq_cq_process_v() 1202 mlx5_timestamp_set(pkts[pos + 1], offset, in rxq_cq_process_v() [all …]
|
| /f-stack/dpdk/lib/librte_port/ |
| H A D | rte_port_source_sink.c | 42 uint8_t **pkts; member 103 port->pkts = rte_zmalloc_socket("PCAP", in pcap_source_load() 105 if (port->pkts == NULL) { in pcap_source_load() 149 port->pkts[i] = buff; in pcap_source_load() 170 if (port->pkts) in pcap_source_load() 171 rte_free(port->pkts); in pcap_source_load() 247 if (p->pkts) in rte_port_source_free() 248 rte_free(p->pkts); in rte_port_source_free() 274 pkts[i]->pkt_len = pkts[i]->data_len; in rte_port_source_rx() 519 PCAP_SINK_WRITE_PKT(p, pkts[i]); in rte_port_sink_tx_bulk() [all …]
|
| H A D | rte_swx_port_ethdev.c | 41 struct rte_mbuf **pkts; member 72 if (!p->pkts) { in reader_create() 96 p->pkts, in reader_pkt_rx() 112 m = p->pkts[p->pos++]; in reader_pkt_rx() 146 struct rte_mbuf *pkt = p->pkts[i]; in reader_free() 151 free(p->pkts); in reader_free() 174 struct rte_mbuf **pkts; member 204 if (!p->pkts) { in writer_create() 225 p->pkts + n_pkts, in __writer_flush() 262 p->pkts[p->n_pkts++] = m; in writer_pkt_tx() [all …]
|
| H A D | rte_port_ring.c | 269 struct rte_mbuf **pkts, in rte_port_ring_writer_tx_bulk_internal() argument 295 (void **)pkts, n_pkts, NULL); in rte_port_ring_writer_tx_bulk_internal() 298 (void **)pkts, n_pkts, NULL); in rte_port_ring_writer_tx_bulk_internal() 302 struct rte_mbuf *pkt = pkts[n_pkts_ok]; in rte_port_ring_writer_tx_bulk_internal() 331 struct rte_mbuf **pkts, in rte_port_ring_writer_tx_bulk() argument 339 struct rte_mbuf **pkts, in rte_port_ring_multi_writer_tx_bulk() argument 585 struct rte_mbuf **pkts, in rte_port_ring_writer_nodrop_tx_bulk_internal() argument 612 (void **)pkts, n_pkts, NULL); in rte_port_ring_writer_nodrop_tx_bulk_internal() 616 (void **)pkts, n_pkts, NULL); in rte_port_ring_writer_nodrop_tx_bulk_internal() 659 struct rte_mbuf **pkts, in rte_port_ring_writer_nodrop_tx_bulk() argument [all …]
|
| H A D | rte_port_frag.c | 48 struct rte_mbuf *pkts[RTE_PORT_IN_BURST_SIZE_MAX]; member 129 struct rte_mbuf **pkts, in rte_port_ring_reader_frag_rx() argument 140 memcpy(pkts, &p->frags[p->pos_frags], n_pkts * sizeof(void *)); in rte_port_ring_reader_frag_rx() 147 memcpy(pkts, &p->frags[p->pos_frags], p->n_frags * sizeof(void *)); in rte_port_ring_reader_frag_rx() 160 (void **) p->pkts, RTE_PORT_IN_BURST_SIZE_MAX, in rte_port_ring_reader_frag_rx() 169 pkt = p->pkts[p->pos_pkts++]; in rte_port_ring_reader_frag_rx() 174 pkts[n_pkts_out++] = pkt; in rte_port_ring_reader_frag_rx() 218 memcpy(&pkts[n_pkts_out], p->frags, in rte_port_ring_reader_frag_rx() 226 memcpy(&pkts[n_pkts_out], p->frags, in rte_port_ring_reader_frag_rx()
|
| /f-stack/dpdk/app/test/ |
| H A D | test_link_bonding_mode4.c | 198 if (pkts[i] != NULL) in free_pkts() 199 rte_pktmbuf_free(pkts[i]); in free_pkts() 866 RTE_DIM(pkts)); in test_mode4_rx() 887 free_pkts(pkts, retval); in test_mode4_rx() 914 RTE_DIM(pkts)); in test_mode4_rx() 928 free_pkts(pkts, retval); in test_mode4_rx() 967 retval = bond_rx(pkts, RTE_DIM(pkts)); in test_mode4_rx() 1049 retval = slave_get_pkts(slave, pkts, RTE_DIM(pkts)); in test_mode4_tx_burst() 1123 retval = slave_get_pkts(slave, pkts, RTE_DIM(pkts)); in test_mode4_tx_burst() 1225 RTE_DIM(pkts)); in test_mode4_marker() [all …]
|
| H A D | test_table_pipeline.c | 16 (struct rte_mbuf **pkts, uint32_t n, uint64_t *pkts_mask, void *arg); 18 (struct rte_mbuf **pkts, uint32_t n, uint64_t *pkts_mask, void *arg); 20 (struct rte_mbuf **pkts, uint32_t n, uint64_t *pkts_mask, void *arg); 28 RTE_SET_USED(pkts); 41 RTE_SET_USED(pkts); 54 RTE_SET_USED(pkts); 65 table_action_0x00(struct rte_pipeline *p, struct rte_mbuf **pkts, 69 table_action_stub_hit(struct rte_pipeline *p, struct rte_mbuf **pkts, 78 __rte_unused struct rte_mbuf **pkts, in table_action_0x00() argument 91 __rte_unused struct rte_mbuf **pkts, in table_action_stub_hit() argument [all …]
|
| /f-stack/dpdk/examples/ipsec-secgw/ |
| H A D | ipsec.c | 534 free_pkts(&pkts[i], 1); in ipsec_enqueue() 539 rte_prefetch0(pkts[i]); in ipsec_enqueue() 541 priv = get_priv(pkts[i]); in ipsec_enqueue() 555 free_pkts(&pkts[i], 1); in ipsec_enqueue() 569 free_pkts(&pkts[i], 1); in ipsec_enqueue() 581 free_pkts(&pkts[i], 1); in ipsec_enqueue() 590 free_pkts(&pkts[i], 1); in ipsec_enqueue() 601 pkts[i], NULL); in ipsec_enqueue() 623 pkts[i], NULL); in ipsec_enqueue() 652 pkts[nb_pkts++] = pkt; in ipsec_inline_dequeue() [all …]
|
| H A D | ipsec-secgw.c | 648 m = ip->pkts[i]; in inbound_sp_sa() 651 ip->pkts[j++] = m; in inbound_sp_sa() 670 ip->pkts[j++] = m; in inbound_sp_sa() 691 trf->ip4.pkts[n4] = m; in split46_traffic() 696 trf->ip6.pkts[n6] = m; in split46_traffic() 751 m = ip->pkts[i]; in outbound_sp() 756 ip->pkts[j++] = m; in outbound_sp() 859 traffic->ipsec.pkts[n] = traffic->ip4.pkts[i]; in process_pkts_outbound_nosp() 864 traffic->ipsec.pkts[n] = traffic->ip6.pkts[i]; in process_pkts_outbound_nosp() 884 traffic->ip4.pkts[i] = traffic->ipsec.pkts[i]; in process_pkts_outbound_nosp() [all …]
|
| /f-stack/dpdk/drivers/net/octeontx2/ |
| H A D | otx2_rx.c | 16 const uint16_t pkts, const uint32_t qmask) in nix_rx_nb_pkts() argument 21 if (unlikely(available < pkts)) { in nix_rx_nb_pkts() 41 return RTE_MIN(pkts, available); in nix_rx_nb_pkts() 46 uint16_t pkts, const uint16_t flags) in nix_recv_pkts() argument 115 uint16_t pkts, const uint16_t flags) in nix_recv_pkts_vector() argument 135 pkts = nix_rx_nb_pkts(rxq, wdata, pkts, qmask); in nix_recv_pkts_vector() 139 pkts = RTE_ALIGN_FLOOR(pkts, NIX_DESCS_PER_LOOP); in nix_recv_pkts_vector() 141 while (packets < pkts) { in nix_recv_pkts_vector() 145 pkts_left += (pkts - packets); in nix_recv_pkts_vector() 321 uint16_t pkts, const uint16_t flags) in nix_recv_pkts_vector() argument [all …]
|
| H A D | otx2_tx.c | 37 for (i = 0; i < pkts; i++) in nix_xmit_pkts() 47 for (i = 0; i < pkts; i++) { in nix_xmit_pkts() 56 txq->fc_cache_pkts -= pkts; in nix_xmit_pkts() 58 return pkts; in nix_xmit_pkts() 76 for (i = 0; i < pkts; i++) in nix_xmit_pkts_mseg() 96 txq->fc_cache_pkts -= pkts; in nix_xmit_pkts_mseg() 98 return pkts; in nix_xmit_pkts_mseg() 131 pkts = RTE_ALIGN_FLOOR(pkts, NIX_DESCS_PER_LOOP); in nix_xmit_pkts_vector() 134 txq->fc_cache_pkts -= pkts; in nix_xmit_pkts_vector() 945 return pkts; in nix_xmit_pkts_vector() [all …]
|
| /f-stack/dpdk/lib/librte_distributor/ |
| H A D | rte_distributor.c | 98 return (pkts[0]) ? 1 : 0; in rte_distributor_poll_pkt() 141 return (pkts[0]) ? 1 : 0; in rte_distributor_get_pkt() 319 pkts[pkts_count++] = (void *)((uintptr_t) in handle_worker_shutdown() 482 unsigned int pkts; in rte_distributor_process() local 485 pkts = num_mbufs - next_idx; in rte_distributor_process() 487 pkts = RTE_DIST_BURST_SIZE; in rte_distributor_process() 489 for (i = 0; i < pkts; i++) { in rte_distributor_process() 501 for (j = 0; j < pkts; j++) { in rte_distributor_process() 560 bl->pkts[idx] = next_value; in rte_distributor_process() 584 bl->pkts[idx] = next_value; in rte_distributor_process() [all …]
|
| /f-stack/dpdk/lib/librte_vhost/ |
| H A D | virtio_net.c | 1323 pkts[i]->pkt_len); in virtio_dev_rx_batch_packed() 1378 &pkts[pkt_idx])) { in virtio_dev_rx_packed() 1617 &pkts[n_free_slot], in virtio_dev_rx_async_submit_split() 2274 if (!pkts[i]) in vhost_reserve_avail_batch_packed() 2279 buf_lens[i] = pkts[i]->buf_len - pkts[i]->data_off; in vhost_reserve_avail_batch_packed() 2288 pkts[i]->data_len = pkts[i]->pkt_len; in vhost_reserve_avail_batch_packed() 2330 pkts[i]->pkt_len); in virtio_dev_tx_batch_packed() 2433 &pkts[pkt_idx])) { in virtio_dev_tx_packed() 2441 &pkts[pkt_idx])) in virtio_dev_tx_packed() 2550 memmove(&pkts[1], pkts, count * sizeof(struct rte_mbuf *)); in rte_vhost_dequeue_burst() [all …]
|
| /f-stack/dpdk/lib/librte_table/ |
| H A D | rte_table_hash_lru.c | 401 struct rte_mbuf **pkts, in rte_table_hash_lru_lookup_unoptimized() argument 423 pkt = pkts[pkt_index]; in rte_table_hash_lru_lookup_unoptimized() 616 mbuf00 = pkts[pkt00_index]; \ 621 mbuf01 = pkts[pkt01_index]; \ 637 mbuf00 = pkts[pkt00_index]; \ 645 mbuf01 = pkts[pkt01_index]; \ 664 mbuf10 = pkts[pkt10_index]; \ 670 mbuf11 = pkts[pkt11_index]; \ 746 mbuf30 = pkts[pkt30_index]; \ 759 mbuf31 = pkts[pkt31_index]; \ [all …]
|