Home
last modified time | relevance | path

Searched refs:sw_ring (Results 1 – 25 of 68) sorted by relevance

123

/f-stack/dpdk/drivers/net/hns3/
H A Dhns3_rxtx_vec_neon.h52 tx_entry = &txq->sw_ring[next_to_use]; in hns3_xmit_fixed_burst_vec()
69 tx_entry = &txq->sw_ring[next_to_use]; in hns3_xmit_fixed_burst_vec()
88 struct hns3_entry *sw_ring, in hns3_desc_parse_field() argument
100 pkt = sw_ring[i].mbuf; in hns3_desc_parse_field()
131 struct hns3_entry *sw_ring = &rxq->sw_ring[rx_id]; in hns3_recv_burst_vec() local
172 mbp1 = vld1q_u64((uint64_t *)&sw_ring[pos]); in hns3_recv_burst_vec()
182 mbp2 = vld1q_u64((uint64_t *)&sw_ring[pos + 2]); in hns3_recv_burst_vec()
269 rte_prefetch0(sw_ring[pos + in hns3_recv_burst_vec()
271 rte_prefetch0(sw_ring[pos + in hns3_recv_burst_vec()
273 rte_prefetch0(sw_ring[pos + in hns3_recv_burst_vec()
[all …]
H A Dhns3_rxtx_vec.c53 struct hns3_entry *rxep = &rxq->sw_ring[rxq->rx_rearm_start]; in hns3_rxq_rearm_mbuf()
122 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 0].mbuf); in hns3_recv_pkts_vec()
123 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 1].mbuf); in hns3_recv_pkts_vec()
124 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 2].mbuf); in hns3_recv_pkts_vec()
125 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 3].mbuf); in hns3_recv_pkts_vec()
155 struct hns3_entry *sw_ring = &rxq->sw_ring[rxq->nb_rx_desc]; in hns3_rxq_vec_setup() local
163 sw_ring[i].mbuf = &rxq->fake_mbuf; in hns3_rxq_vec_setup()
H A Dhns3_rxtx_vec_sve.c67 hns3_rx_prefetch_mbuf_sve(struct hns3_entry *sw_ring) in hns3_rx_prefetch_mbuf_sve() argument
69 svuint64_t prf1st = svld1_u64(PG64_256BIT, (uint64_t *)&sw_ring[0]); in hns3_rx_prefetch_mbuf_sve()
70 svuint64_t prf2st = svld1_u64(PG64_256BIT, (uint64_t *)&sw_ring[4]); in hns3_rx_prefetch_mbuf_sve()
85 struct hns3_entry *sw_ring = &rxq->sw_ring[rx_id]; in hns3_recv_burst_vec_sve() local
138 mbp1st = svld1_u64(PG64_256BIT, (uint64_t *)&sw_ring[pos]); in hns3_recv_burst_vec_sve()
149 mbp2st = svld1_u64(PG64_256BIT, (uint64_t *)&sw_ring[pos + 4]); in hns3_recv_burst_vec_sve()
219 hns3_rx_prefetch_mbuf_sve(&sw_ring[pos + in hns3_recv_burst_vec_sve()
239 struct hns3_entry *rxep = &rxq->sw_ring[rxq->rx_rearm_start]; in hns3_rxq_rearm_mbuf_sve()
305 hns3_rx_prefetch_mbuf_sve(&rxq->sw_ring[rxq->next_to_use]); in hns3_recv_pkts_vec_sve()
362 struct hns3_entry *tx_entry = &txq->sw_ring[txq->next_to_use]; in hns3_tx_fill_hw_ring_sve()
/f-stack/dpdk/drivers/net/virtio/
H A Dvirtio_rxtx_simple_neon.c49 struct rte_mbuf **sw_ring; in virtio_recv_pkts_vec() local
98 sw_ring = &vq->sw_ring[desc_idx]; in virtio_recv_pkts_vec()
99 sw_ring_end = &vq->sw_ring[vq->vq_nentries]; in virtio_recv_pkts_vec()
117 mbp[0] = vld1q_u64((uint64_t *)(sw_ring + 0)); in virtio_recv_pkts_vec()
121 mbp[1] = vld1q_u64((uint64_t *)(sw_ring + 2)); in virtio_recv_pkts_vec()
125 mbp[2] = vld1q_u64((uint64_t *)(sw_ring + 4)); in virtio_recv_pkts_vec()
129 mbp[3] = vld1q_u64((uint64_t *)(sw_ring + 6)); in virtio_recv_pkts_vec()
186 if (sw_ring + nb_used <= sw_ring_end) in virtio_recv_pkts_vec()
189 nb_pkts_received += sw_ring_end - sw_ring; in virtio_recv_pkts_vec()
194 nb_pkts_received += sw_ring_end - sw_ring; in virtio_recv_pkts_vec()
[all …]
H A Dvirtio_rxtx_simple_sse.c49 struct rte_mbuf **sw_ring; in virtio_recv_pkts_vec() local
98 sw_ring = &vq->sw_ring[desc_idx]; in virtio_recv_pkts_vec()
99 sw_ring_end = &vq->sw_ring[vq->vq_nentries]; in virtio_recv_pkts_vec()
117 mbp[0] = _mm_loadu_si128((__m128i *)(sw_ring + 0)); in virtio_recv_pkts_vec()
121 mbp[1] = _mm_loadu_si128((__m128i *)(sw_ring + 2)); in virtio_recv_pkts_vec()
125 mbp[2] = _mm_loadu_si128((__m128i *)(sw_ring + 4)); in virtio_recv_pkts_vec()
129 mbp[3] = _mm_loadu_si128((__m128i *)(sw_ring + 6)); in virtio_recv_pkts_vec()
170 if (sw_ring + nb_used <= sw_ring_end) in virtio_recv_pkts_vec()
173 nb_pkts_received += sw_ring_end - sw_ring; in virtio_recv_pkts_vec()
178 nb_pkts_received += sw_ring_end - sw_ring; in virtio_recv_pkts_vec()
[all …]
H A Dvirtio_rxtx_simple_altivec.c49 struct rte_mbuf **sw_ring; in virtio_recv_pkts_vec() local
100 sw_ring = &vq->sw_ring[desc_idx]; in virtio_recv_pkts_vec()
101 sw_ring_end = &vq->sw_ring[vq->vq_nentries]; in virtio_recv_pkts_vec()
119 mbp[0] = vec_vsx_ld(0, (unsigned char const *)(sw_ring + 0)); in virtio_recv_pkts_vec()
123 mbp[1] = vec_vsx_ld(0, (unsigned char const *)(sw_ring + 2)); in virtio_recv_pkts_vec()
127 mbp[2] = vec_vsx_ld(0, (unsigned char const *)(sw_ring + 4)); in virtio_recv_pkts_vec()
180 if (sw_ring + nb_used <= sw_ring_end) in virtio_recv_pkts_vec()
183 nb_pkts_received += sw_ring_end - sw_ring; in virtio_recv_pkts_vec()
186 if (unlikely(sw_ring + RTE_VIRTIO_DESC_PER_LOOP >= in virtio_recv_pkts_vec()
188 nb_pkts_received += sw_ring_end - sw_ring; in virtio_recv_pkts_vec()
[all …]
H A Dvirtio_rxtx_simple.h23 struct rte_mbuf **sw_ring; in virtio_rxq_rearm_vec() local
29 sw_ring = &vq->sw_ring[desc_idx]; in virtio_rxq_rearm_vec()
32 ret = rte_mempool_get_bulk(rxvq->mpool, (void **)sw_ring, in virtio_rxq_rearm_vec()
43 p = (uintptr_t)&sw_ring[i]->rearm_data; in virtio_rxq_rearm_vec()
47 VIRTIO_MBUF_ADDR(sw_ring[i], vq) + in virtio_rxq_rearm_vec()
49 start_dp[i].len = sw_ring[i]->buf_len - in virtio_rxq_rearm_vec()
/f-stack/dpdk/drivers/net/bnx2x/
H A Dbnx2x_rxtx.c23 struct rte_mbuf **sw_ring; in bnx2x_rx_queue_release() local
27 sw_ring = rx_queue->sw_ring; in bnx2x_rx_queue_release()
28 if (NULL != sw_ring) { in bnx2x_rx_queue_release()
30 if (NULL != sw_ring[i]) in bnx2x_rx_queue_release()
33 rte_free(sw_ring); in bnx2x_rx_queue_release()
129 rxq->sw_ring[idx] = mbuf; in bnx2x_dev_rx_queue_setup()
168 struct rte_mbuf **sw_ring; in bnx2x_tx_queue_release() local
172 sw_ring = tx_queue->sw_ring; in bnx2x_tx_queue_release()
173 if (NULL != sw_ring) { in bnx2x_tx_queue_release()
175 if (NULL != sw_ring[i]) in bnx2x_tx_queue_release()
[all …]
/f-stack/dpdk/drivers/net/ice/
H A Dice_rxtx_vec_common.h148 if (unlikely(!rxq->sw_ring)) { in _ice_rx_queue_release_mbufs_vec()
159 if (rxq->sw_ring[i].mbuf) in _ice_rx_queue_release_mbufs_vec()
160 rte_pktmbuf_free_seg(rxq->sw_ring[i].mbuf); in _ice_rx_queue_release_mbufs_vec()
166 if (rxq->sw_ring[i].mbuf) in _ice_rx_queue_release_mbufs_vec()
167 rte_pktmbuf_free_seg(rxq->sw_ring[i].mbuf); in _ice_rx_queue_release_mbufs_vec()
174 memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); in _ice_rx_queue_release_mbufs_vec()
182 if (unlikely(!txq || !txq->sw_ring)) { in _ice_tx_queue_release_mbufs_vec()
215 rte_pktmbuf_free_seg(txq->sw_ring[i].mbuf); in _ice_tx_queue_release_mbufs_vec()
216 txq->sw_ring[i].mbuf = NULL; in _ice_tx_queue_release_mbufs_vec()
221 rte_pktmbuf_free_seg(txq->sw_ring[i].mbuf); in _ice_tx_queue_release_mbufs_vec()
[all …]
/f-stack/dpdk/drivers/net/sfc/
H A Dsfc_tso.c20 sfc_efx_tso_alloc_tsoh_objs(struct sfc_efx_tx_sw_desc *sw_ring, in sfc_efx_tso_alloc_tsoh_objs() argument
26 sw_ring[i].tsoh = rte_malloc_socket("sfc-efx-txq-tsoh-obj", in sfc_efx_tso_alloc_tsoh_objs()
30 if (sw_ring[i].tsoh == NULL) in sfc_efx_tso_alloc_tsoh_objs()
38 rte_free(sw_ring[--i].tsoh); in sfc_efx_tso_alloc_tsoh_objs()
44 sfc_efx_tso_free_tsoh_objs(struct sfc_efx_tx_sw_desc *sw_ring, in sfc_efx_tso_free_tsoh_objs() argument
50 rte_free(sw_ring[i].tsoh); in sfc_efx_tso_free_tsoh_objs()
51 sw_ring[i].tsoh = NULL; in sfc_efx_tso_free_tsoh_objs()
128 tsoh = txq->sw_ring[idx & txq->ptr_mask].tsoh; in sfc_efx_tso_do()
H A Dsfc_ef10_rx.c70 struct sfc_ef10_rx_sw_desc *sw_ring; member
154 rxd = &rxq->sw_ring[id]; in sfc_ef10_rx_qrefill()
184 rte_prefetch0(&rxq->sw_ring[next_id]); in sfc_ef10_rx_prefetch_next()
191 next_mbuf = rxq->sw_ring[next_id].mbuf; in sfc_ef10_rx_prefetch_next()
297 rxd = &rxq->sw_ring[pending++ & ptr_mask]; in sfc_ef10_rx_process_event()
319 rxd = &rxq->sw_ring[pending++ & ptr_mask]; in sfc_ef10_rx_process_event()
369 rxd = &rxq->sw_ring[pending++ & ptr_mask]; in sfc_ef10_rx_process_event()
665 sizeof(*rxq->sw_ring), in sfc_ef10_rx_qcreate()
667 if (rxq->sw_ring == NULL) in sfc_ef10_rx_qcreate()
709 rte_free(rxq->sw_ring); in sfc_ef10_rx_qdestroy()
[all …]
H A Dsfc_tx.h68 struct sfc_efx_tx_sw_desc *sw_ring; member
125 int sfc_efx_tso_alloc_tsoh_objs(struct sfc_efx_tx_sw_desc *sw_ring,
128 void sfc_efx_tso_free_tsoh_objs(struct sfc_efx_tx_sw_desc *sw_ring,
/f-stack/dpdk/drivers/net/iavf/
H A Diavf_rxtx_vec_common.h98 txep = &txq->sw_ring[txq->next_dd - (n - 1)]; in iavf_tx_free_bufs()
151 if (!rxq->sw_ring || rxq->rxrearm_nb >= rxq->nb_rx_desc) in _iavf_rx_queue_release_mbufs_vec()
157 if (rxq->sw_ring[i]) in _iavf_rx_queue_release_mbufs_vec()
158 rte_pktmbuf_free_seg(rxq->sw_ring[i]); in _iavf_rx_queue_release_mbufs_vec()
164 if (rxq->sw_ring[i]) in _iavf_rx_queue_release_mbufs_vec()
165 rte_pktmbuf_free_seg(rxq->sw_ring[i]); in _iavf_rx_queue_release_mbufs_vec()
172 memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); in _iavf_rx_queue_release_mbufs_vec()
181 if (!txq->sw_ring || txq->nb_free == max_desc) in _iavf_tx_queue_release_mbufs_vec()
187 rte_pktmbuf_free_seg(txq->sw_ring[i].mbuf); in _iavf_tx_queue_release_mbufs_vec()
188 txq->sw_ring[i].mbuf = NULL; in _iavf_tx_queue_release_mbufs_vec()
H A Diavf_rxtx.c213 txe = txq->sw_ring; in reset_tx_queue()
281 if (!rxq->sw_ring) in release_rxq_mbufs()
285 if (rxq->sw_ring[i]) { in release_rxq_mbufs()
563 rxq->sw_ring = in iavf_dev_rx_queue_setup()
568 if (!rxq->sw_ring) { in iavf_dev_rx_queue_setup()
678 txq->sw_ring = in iavf_dev_tx_queue_setup()
683 if (!txq->sw_ring) { in iavf_dev_tx_queue_setup()
867 rte_free(q->sw_ring); in iavf_dev_rx_queue_release()
881 rte_free(q->sw_ring); in iavf_dev_tx_queue_release()
1928 struct iavf_tx_entry *sw_ring = txq->sw_ring; in iavf_xmit_cleanup() local
[all …]
/f-stack/dpdk/drivers/net/axgbe/
H A Daxgbe_rxtx.c19 struct rte_mbuf **sw_ring; in axgbe_rx_queue_release() local
22 sw_ring = rx_queue->sw_ring; in axgbe_rx_queue_release()
23 if (sw_ring) { in axgbe_rx_queue_release()
25 if (sw_ring[i]) in axgbe_rx_queue_release()
28 rte_free(sw_ring); in axgbe_rx_queue_release()
107 if (!rxq->sw_ring) { in axgbe_dev_rx_queue_setup()
468 sw_ring = tx_queue->sw_ring; in axgbe_tx_queue_release()
469 if (sw_ring) { in axgbe_tx_queue_release()
471 if (sw_ring[i]) in axgbe_tx_queue_release()
474 rte_free(sw_ring); in axgbe_tx_queue_release()
[all …]
/f-stack/dpdk/drivers/net/igc/
H A Digc_txrx.c351 struct igc_rx_entry * const sw_ring = rxq->sw_ring; in igc_recv_pkts() local
422 rxe = &sw_ring[rx_id]; in igc_recv_pkts()
495 struct igc_rx_entry * const sw_ring = rxq->sw_ring; in igc_recv_scattered_pkts() local
566 rxe = &sw_ring[rx_id]; in igc_recv_scattered_pkts()
700 if (rxq->sw_ring != NULL) { in igc_rx_queue_release_mbufs()
714 rte_free(rxq->sw_ring); in igc_rx_queue_release()
1398 if (rxq->sw_ring == NULL) { in eth_igc_rx_queue_setup()
1624 struct igc_tx_entry * const sw_ring = txq->sw_ring; in igc_xmit_pkts() local
1645 txe = &sw_ring[tx_id]; in igc_xmit_pkts()
1893 rte_free(txq->sw_ring); in igc_tx_queue_release()
[all …]
/f-stack/dpdk/drivers/net/atlantic/
H A Datl_rxtx.c56 struct atl_rx_entry *sw_ring; member
74 struct atl_tx_entry *sw_ring; member
158 if (rxq->sw_ring == NULL) { in atl_rx_queue_setup()
179 rte_free(rxq->sw_ring); in atl_rx_queue_setup()
206 tx_entry = txq->sw_ring; in atl_reset_tx_queue()
273 if (txq->sw_ring == NULL) { in atl_tx_queue_setup()
293 rte_free(txq->sw_ring); in atl_tx_queue_setup()
509 rte_free(rxq->sw_ring); in atl_rx_queue_release()
580 rte_free(txq->sw_ring); in atl_tx_queue_release()
926 struct atl_rx_entry *sw_ring = rxq->sw_ring; in atl_recv_pkts() local
[all …]
/f-stack/dpdk/drivers/net/e1000/
H A Dem_rxtx.c298 struct em_tx_entry *sw_ring = txq->sw_ring; in em_xmit_cleanup() local
369 struct em_tx_entry *sw_ring; in eth_em_xmit_pkts() local
390 sw_ring = txq->sw_ring; in eth_em_xmit_pkts()
393 txe = &sw_ring[tx_id]; in eth_em_xmit_pkts()
701 sw_ring = rxq->sw_ring; in eth_em_recv_pkts()
760 rxe = &sw_ring[rx_id]; in eth_em_recv_pkts()
877 sw_ring = rxq->sw_ring; in eth_em_recv_scattered_pkts()
939 rxe = &sw_ring[rx_id]; in eth_em_recv_scattered_pkts()
1103 if (txq->sw_ring != NULL) { in em_tx_queue_release_mbufs()
1118 rte_free(txq->sw_ring); in em_tx_queue_release()
[all …]
H A Digb_rxtx.c404 sw_ring = txq->sw_ring; in eth_igb_xmit_pkts()
407 txe = &sw_ring[tx_id]; in eth_igb_xmit_pkts()
829 sw_ring = rxq->sw_ring; in eth_igb_recv_pkts()
887 rxe = &sw_ring[rx_id]; in eth_igb_recv_pkts()
1016 sw_ring = rxq->sw_ring; in eth_igb_recv_scattered_pkts()
1078 rxe = &sw_ring[rx_id]; in eth_igb_recv_scattered_pkts()
1263 if (txq->sw_ring != NULL) { in igb_tx_queue_release_mbufs()
1278 rte_free(txq->sw_ring); in igb_tx_queue_release()
1303 sw_ring = txq->sw_ring; in igb_tx_done_cleanup()
1346 sw_ring[tx_id].mbuf); in igb_tx_done_cleanup()
[all …]
/f-stack/dpdk/drivers/net/bnxt/
H A Dbnxt_txq.c27 struct bnxt_sw_tx_bd *sw_ring; in bnxt_tx_queue_release_mbufs() local
33 sw_ring = txq->tx_ring->tx_buf_ring; in bnxt_tx_queue_release_mbufs()
34 if (sw_ring) { in bnxt_tx_queue_release_mbufs()
36 if (sw_ring[i].mbuf) { in bnxt_tx_queue_release_mbufs()
37 rte_pktmbuf_free_seg(sw_ring[i].mbuf); in bnxt_tx_queue_release_mbufs()
38 sw_ring[i].mbuf = NULL; in bnxt_tx_queue_release_mbufs()
H A Dbnxt_rxq.c206 struct rte_mbuf **sw_ring; in bnxt_rx_queue_release_mbufs() local
213 sw_ring = rxq->rx_ring->rx_buf_ring; in bnxt_rx_queue_release_mbufs()
214 if (sw_ring) { in bnxt_rx_queue_release_mbufs()
217 if (sw_ring[i]) { in bnxt_rx_queue_release_mbufs()
218 if (sw_ring[i] != &rxq->fake_mbuf) in bnxt_rx_queue_release_mbufs()
220 sw_ring[i] = NULL; in bnxt_rx_queue_release_mbufs()
225 sw_ring = rxq->rx_ring->ag_buf_ring; in bnxt_rx_queue_release_mbufs()
226 if (sw_ring) { in bnxt_rx_queue_release_mbufs()
229 if (sw_ring[i]) { in bnxt_rx_queue_release_mbufs()
230 rte_pktmbuf_free_seg(sw_ring[i]); in bnxt_rx_queue_release_mbufs()
[all …]
/f-stack/dpdk/drivers/net/ixgbe/
H A Dixgbe_rxtx_vec_common.h149 if (txq->sw_ring == NULL || txq->nb_tx_free == max_desc) in _ixgbe_tx_queue_release_mbufs_vec()
174 if (rxq->sw_ring == NULL || rxq->rxrearm_nb >= rxq->nb_rx_desc) in _ixgbe_rx_queue_release_mbufs_vec()
180 if (rxq->sw_ring[i].mbuf != NULL) in _ixgbe_rx_queue_release_mbufs_vec()
181 rte_pktmbuf_free_seg(rxq->sw_ring[i].mbuf); in _ixgbe_rx_queue_release_mbufs_vec()
187 if (rxq->sw_ring[i].mbuf != NULL) in _ixgbe_rx_queue_release_mbufs_vec()
188 rte_pktmbuf_free_seg(rxq->sw_ring[i].mbuf); in _ixgbe_rx_queue_release_mbufs_vec()
195 memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); in _ixgbe_rx_queue_release_mbufs_vec()
204 if (txq->sw_ring != NULL) { in _ixgbe_tx_free_swring_vec()
/f-stack/dpdk/drivers/net/i40e/
H A Di40e_rxtx_vec_common.h97 txep = &txq->sw_ring[txq->tx_next_dd - (n - 1)]; in i40e_tx_free_bufs()
150 if (rxq->sw_ring == NULL || rxq->rxrearm_nb >= rxq->nb_rx_desc) in _i40e_rx_queue_release_mbufs_vec()
156 if (rxq->sw_ring[i].mbuf != NULL) in _i40e_rx_queue_release_mbufs_vec()
157 rte_pktmbuf_free_seg(rxq->sw_ring[i].mbuf); in _i40e_rx_queue_release_mbufs_vec()
163 if (rxq->sw_ring[i].mbuf != NULL) in _i40e_rx_queue_release_mbufs_vec()
164 rte_pktmbuf_free_seg(rxq->sw_ring[i].mbuf); in _i40e_rx_queue_release_mbufs_vec()
171 memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); in _i40e_rx_queue_release_mbufs_vec()
/f-stack/dpdk/drivers/raw/ntb/
H A Dntb.c277 if (!q || !q->sw_ring) { in ntb_rxq_release_mbufs()
300 rte_free(rxq->sw_ring); in ntb_rxq_release()
339 rxq->sw_ring = in ntb_rxq_setup()
345 if (!rxq->sw_ring) { in ntb_rxq_setup()
362 if (!q || !q->sw_ring) { in ntb_txq_release_mbufs()
385 rte_free(txq->sw_ring); in ntb_txq_release()
420 txq->sw_ring = in ntb_txq_setup()
426 if (!txq->sw_ring) { in ntb_txq_setup()
581 struct ntb_tx_entry *sw_ring = txq->sw_ring; in ntb_enqueue_cleanup() local
607 struct ntb_tx_entry *sw_ring = txq->sw_ring; in ntb_enqueue_bufs() local
[all …]
/f-stack/dpdk/drivers/net/fm10k/
H A Dfm10k_rxtx.c116 mbuf = q->sw_ring[next_dd]; in fm10k_recv_pkts()
153 rte_prefetch0(q->sw_ring[next_dd]); in fm10k_recv_pkts()
162 rte_prefetch0(&q->sw_ring[next_dd]); in fm10k_recv_pkts()
170 (void **)&q->sw_ring[q->next_alloc], in fm10k_recv_pkts()
187 mbuf = q->sw_ring[q->next_alloc]; in fm10k_recv_pkts()
230 mbuf = q->sw_ring[next_dd]; in fm10k_recv_scattered_pkts()
242 rte_prefetch0(q->sw_ring[next_dd]); in fm10k_recv_scattered_pkts()
251 rte_prefetch0(&q->sw_ring[next_dd]); in fm10k_recv_scattered_pkts()
345 mbuf = q->sw_ring[q->next_alloc]; in fm10k_recv_scattered_pkts()
617 q->sw_ring[q->next_free] = mb; in tx_xmit_pkt()
[all …]

123