| /f-stack/dpdk/drivers/net/hns3/ |
| H A D | hns3_rxtx_vec_neon.h | 52 tx_entry = &txq->sw_ring[next_to_use]; in hns3_xmit_fixed_burst_vec() 69 tx_entry = &txq->sw_ring[next_to_use]; in hns3_xmit_fixed_burst_vec() 88 struct hns3_entry *sw_ring, in hns3_desc_parse_field() argument 100 pkt = sw_ring[i].mbuf; in hns3_desc_parse_field() 131 struct hns3_entry *sw_ring = &rxq->sw_ring[rx_id]; in hns3_recv_burst_vec() local 172 mbp1 = vld1q_u64((uint64_t *)&sw_ring[pos]); in hns3_recv_burst_vec() 182 mbp2 = vld1q_u64((uint64_t *)&sw_ring[pos + 2]); in hns3_recv_burst_vec() 269 rte_prefetch0(sw_ring[pos + in hns3_recv_burst_vec() 271 rte_prefetch0(sw_ring[pos + in hns3_recv_burst_vec() 273 rte_prefetch0(sw_ring[pos + in hns3_recv_burst_vec() [all …]
|
| H A D | hns3_rxtx_vec.c | 53 struct hns3_entry *rxep = &rxq->sw_ring[rxq->rx_rearm_start]; in hns3_rxq_rearm_mbuf() 122 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 0].mbuf); in hns3_recv_pkts_vec() 123 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 1].mbuf); in hns3_recv_pkts_vec() 124 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 2].mbuf); in hns3_recv_pkts_vec() 125 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 3].mbuf); in hns3_recv_pkts_vec() 155 struct hns3_entry *sw_ring = &rxq->sw_ring[rxq->nb_rx_desc]; in hns3_rxq_vec_setup() local 163 sw_ring[i].mbuf = &rxq->fake_mbuf; in hns3_rxq_vec_setup()
|
| H A D | hns3_rxtx_vec_sve.c | 67 hns3_rx_prefetch_mbuf_sve(struct hns3_entry *sw_ring) in hns3_rx_prefetch_mbuf_sve() argument 69 svuint64_t prf1st = svld1_u64(PG64_256BIT, (uint64_t *)&sw_ring[0]); in hns3_rx_prefetch_mbuf_sve() 70 svuint64_t prf2st = svld1_u64(PG64_256BIT, (uint64_t *)&sw_ring[4]); in hns3_rx_prefetch_mbuf_sve() 85 struct hns3_entry *sw_ring = &rxq->sw_ring[rx_id]; in hns3_recv_burst_vec_sve() local 138 mbp1st = svld1_u64(PG64_256BIT, (uint64_t *)&sw_ring[pos]); in hns3_recv_burst_vec_sve() 149 mbp2st = svld1_u64(PG64_256BIT, (uint64_t *)&sw_ring[pos + 4]); in hns3_recv_burst_vec_sve() 219 hns3_rx_prefetch_mbuf_sve(&sw_ring[pos + in hns3_recv_burst_vec_sve() 239 struct hns3_entry *rxep = &rxq->sw_ring[rxq->rx_rearm_start]; in hns3_rxq_rearm_mbuf_sve() 305 hns3_rx_prefetch_mbuf_sve(&rxq->sw_ring[rxq->next_to_use]); in hns3_recv_pkts_vec_sve() 362 struct hns3_entry *tx_entry = &txq->sw_ring[txq->next_to_use]; in hns3_tx_fill_hw_ring_sve()
|
| /f-stack/dpdk/drivers/net/virtio/ |
| H A D | virtio_rxtx_simple_neon.c | 49 struct rte_mbuf **sw_ring; in virtio_recv_pkts_vec() local 98 sw_ring = &vq->sw_ring[desc_idx]; in virtio_recv_pkts_vec() 99 sw_ring_end = &vq->sw_ring[vq->vq_nentries]; in virtio_recv_pkts_vec() 117 mbp[0] = vld1q_u64((uint64_t *)(sw_ring + 0)); in virtio_recv_pkts_vec() 121 mbp[1] = vld1q_u64((uint64_t *)(sw_ring + 2)); in virtio_recv_pkts_vec() 125 mbp[2] = vld1q_u64((uint64_t *)(sw_ring + 4)); in virtio_recv_pkts_vec() 129 mbp[3] = vld1q_u64((uint64_t *)(sw_ring + 6)); in virtio_recv_pkts_vec() 186 if (sw_ring + nb_used <= sw_ring_end) in virtio_recv_pkts_vec() 189 nb_pkts_received += sw_ring_end - sw_ring; in virtio_recv_pkts_vec() 194 nb_pkts_received += sw_ring_end - sw_ring; in virtio_recv_pkts_vec() [all …]
|
| H A D | virtio_rxtx_simple_sse.c | 49 struct rte_mbuf **sw_ring; in virtio_recv_pkts_vec() local 98 sw_ring = &vq->sw_ring[desc_idx]; in virtio_recv_pkts_vec() 99 sw_ring_end = &vq->sw_ring[vq->vq_nentries]; in virtio_recv_pkts_vec() 117 mbp[0] = _mm_loadu_si128((__m128i *)(sw_ring + 0)); in virtio_recv_pkts_vec() 121 mbp[1] = _mm_loadu_si128((__m128i *)(sw_ring + 2)); in virtio_recv_pkts_vec() 125 mbp[2] = _mm_loadu_si128((__m128i *)(sw_ring + 4)); in virtio_recv_pkts_vec() 129 mbp[3] = _mm_loadu_si128((__m128i *)(sw_ring + 6)); in virtio_recv_pkts_vec() 170 if (sw_ring + nb_used <= sw_ring_end) in virtio_recv_pkts_vec() 173 nb_pkts_received += sw_ring_end - sw_ring; in virtio_recv_pkts_vec() 178 nb_pkts_received += sw_ring_end - sw_ring; in virtio_recv_pkts_vec() [all …]
|
| H A D | virtio_rxtx_simple_altivec.c | 49 struct rte_mbuf **sw_ring; in virtio_recv_pkts_vec() local 100 sw_ring = &vq->sw_ring[desc_idx]; in virtio_recv_pkts_vec() 101 sw_ring_end = &vq->sw_ring[vq->vq_nentries]; in virtio_recv_pkts_vec() 119 mbp[0] = vec_vsx_ld(0, (unsigned char const *)(sw_ring + 0)); in virtio_recv_pkts_vec() 123 mbp[1] = vec_vsx_ld(0, (unsigned char const *)(sw_ring + 2)); in virtio_recv_pkts_vec() 127 mbp[2] = vec_vsx_ld(0, (unsigned char const *)(sw_ring + 4)); in virtio_recv_pkts_vec() 180 if (sw_ring + nb_used <= sw_ring_end) in virtio_recv_pkts_vec() 183 nb_pkts_received += sw_ring_end - sw_ring; in virtio_recv_pkts_vec() 186 if (unlikely(sw_ring + RTE_VIRTIO_DESC_PER_LOOP >= in virtio_recv_pkts_vec() 188 nb_pkts_received += sw_ring_end - sw_ring; in virtio_recv_pkts_vec() [all …]
|
| H A D | virtio_rxtx_simple.h | 23 struct rte_mbuf **sw_ring; in virtio_rxq_rearm_vec() local 29 sw_ring = &vq->sw_ring[desc_idx]; in virtio_rxq_rearm_vec() 32 ret = rte_mempool_get_bulk(rxvq->mpool, (void **)sw_ring, in virtio_rxq_rearm_vec() 43 p = (uintptr_t)&sw_ring[i]->rearm_data; in virtio_rxq_rearm_vec() 47 VIRTIO_MBUF_ADDR(sw_ring[i], vq) + in virtio_rxq_rearm_vec() 49 start_dp[i].len = sw_ring[i]->buf_len - in virtio_rxq_rearm_vec()
|
| /f-stack/dpdk/drivers/net/bnx2x/ |
| H A D | bnx2x_rxtx.c | 23 struct rte_mbuf **sw_ring; in bnx2x_rx_queue_release() local 27 sw_ring = rx_queue->sw_ring; in bnx2x_rx_queue_release() 28 if (NULL != sw_ring) { in bnx2x_rx_queue_release() 30 if (NULL != sw_ring[i]) in bnx2x_rx_queue_release() 33 rte_free(sw_ring); in bnx2x_rx_queue_release() 129 rxq->sw_ring[idx] = mbuf; in bnx2x_dev_rx_queue_setup() 168 struct rte_mbuf **sw_ring; in bnx2x_tx_queue_release() local 172 sw_ring = tx_queue->sw_ring; in bnx2x_tx_queue_release() 173 if (NULL != sw_ring) { in bnx2x_tx_queue_release() 175 if (NULL != sw_ring[i]) in bnx2x_tx_queue_release() [all …]
|
| /f-stack/dpdk/drivers/net/ice/ |
| H A D | ice_rxtx_vec_common.h | 148 if (unlikely(!rxq->sw_ring)) { in _ice_rx_queue_release_mbufs_vec() 159 if (rxq->sw_ring[i].mbuf) in _ice_rx_queue_release_mbufs_vec() 160 rte_pktmbuf_free_seg(rxq->sw_ring[i].mbuf); in _ice_rx_queue_release_mbufs_vec() 166 if (rxq->sw_ring[i].mbuf) in _ice_rx_queue_release_mbufs_vec() 167 rte_pktmbuf_free_seg(rxq->sw_ring[i].mbuf); in _ice_rx_queue_release_mbufs_vec() 174 memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); in _ice_rx_queue_release_mbufs_vec() 182 if (unlikely(!txq || !txq->sw_ring)) { in _ice_tx_queue_release_mbufs_vec() 215 rte_pktmbuf_free_seg(txq->sw_ring[i].mbuf); in _ice_tx_queue_release_mbufs_vec() 216 txq->sw_ring[i].mbuf = NULL; in _ice_tx_queue_release_mbufs_vec() 221 rte_pktmbuf_free_seg(txq->sw_ring[i].mbuf); in _ice_tx_queue_release_mbufs_vec() [all …]
|
| /f-stack/dpdk/drivers/net/sfc/ |
| H A D | sfc_tso.c | 20 sfc_efx_tso_alloc_tsoh_objs(struct sfc_efx_tx_sw_desc *sw_ring, in sfc_efx_tso_alloc_tsoh_objs() argument 26 sw_ring[i].tsoh = rte_malloc_socket("sfc-efx-txq-tsoh-obj", in sfc_efx_tso_alloc_tsoh_objs() 30 if (sw_ring[i].tsoh == NULL) in sfc_efx_tso_alloc_tsoh_objs() 38 rte_free(sw_ring[--i].tsoh); in sfc_efx_tso_alloc_tsoh_objs() 44 sfc_efx_tso_free_tsoh_objs(struct sfc_efx_tx_sw_desc *sw_ring, in sfc_efx_tso_free_tsoh_objs() argument 50 rte_free(sw_ring[i].tsoh); in sfc_efx_tso_free_tsoh_objs() 51 sw_ring[i].tsoh = NULL; in sfc_efx_tso_free_tsoh_objs() 128 tsoh = txq->sw_ring[idx & txq->ptr_mask].tsoh; in sfc_efx_tso_do()
|
| H A D | sfc_ef10_rx.c | 70 struct sfc_ef10_rx_sw_desc *sw_ring; member 154 rxd = &rxq->sw_ring[id]; in sfc_ef10_rx_qrefill() 184 rte_prefetch0(&rxq->sw_ring[next_id]); in sfc_ef10_rx_prefetch_next() 191 next_mbuf = rxq->sw_ring[next_id].mbuf; in sfc_ef10_rx_prefetch_next() 297 rxd = &rxq->sw_ring[pending++ & ptr_mask]; in sfc_ef10_rx_process_event() 319 rxd = &rxq->sw_ring[pending++ & ptr_mask]; in sfc_ef10_rx_process_event() 369 rxd = &rxq->sw_ring[pending++ & ptr_mask]; in sfc_ef10_rx_process_event() 665 sizeof(*rxq->sw_ring), in sfc_ef10_rx_qcreate() 667 if (rxq->sw_ring == NULL) in sfc_ef10_rx_qcreate() 709 rte_free(rxq->sw_ring); in sfc_ef10_rx_qdestroy() [all …]
|
| H A D | sfc_tx.h | 68 struct sfc_efx_tx_sw_desc *sw_ring; member 125 int sfc_efx_tso_alloc_tsoh_objs(struct sfc_efx_tx_sw_desc *sw_ring, 128 void sfc_efx_tso_free_tsoh_objs(struct sfc_efx_tx_sw_desc *sw_ring,
|
| /f-stack/dpdk/drivers/net/iavf/ |
| H A D | iavf_rxtx_vec_common.h | 98 txep = &txq->sw_ring[txq->next_dd - (n - 1)]; in iavf_tx_free_bufs() 151 if (!rxq->sw_ring || rxq->rxrearm_nb >= rxq->nb_rx_desc) in _iavf_rx_queue_release_mbufs_vec() 157 if (rxq->sw_ring[i]) in _iavf_rx_queue_release_mbufs_vec() 158 rte_pktmbuf_free_seg(rxq->sw_ring[i]); in _iavf_rx_queue_release_mbufs_vec() 164 if (rxq->sw_ring[i]) in _iavf_rx_queue_release_mbufs_vec() 165 rte_pktmbuf_free_seg(rxq->sw_ring[i]); in _iavf_rx_queue_release_mbufs_vec() 172 memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); in _iavf_rx_queue_release_mbufs_vec() 181 if (!txq->sw_ring || txq->nb_free == max_desc) in _iavf_tx_queue_release_mbufs_vec() 187 rte_pktmbuf_free_seg(txq->sw_ring[i].mbuf); in _iavf_tx_queue_release_mbufs_vec() 188 txq->sw_ring[i].mbuf = NULL; in _iavf_tx_queue_release_mbufs_vec()
|
| H A D | iavf_rxtx.c | 213 txe = txq->sw_ring; in reset_tx_queue() 281 if (!rxq->sw_ring) in release_rxq_mbufs() 285 if (rxq->sw_ring[i]) { in release_rxq_mbufs() 563 rxq->sw_ring = in iavf_dev_rx_queue_setup() 568 if (!rxq->sw_ring) { in iavf_dev_rx_queue_setup() 678 txq->sw_ring = in iavf_dev_tx_queue_setup() 683 if (!txq->sw_ring) { in iavf_dev_tx_queue_setup() 867 rte_free(q->sw_ring); in iavf_dev_rx_queue_release() 881 rte_free(q->sw_ring); in iavf_dev_tx_queue_release() 1928 struct iavf_tx_entry *sw_ring = txq->sw_ring; in iavf_xmit_cleanup() local [all …]
|
| /f-stack/dpdk/drivers/net/axgbe/ |
| H A D | axgbe_rxtx.c | 19 struct rte_mbuf **sw_ring; in axgbe_rx_queue_release() local 22 sw_ring = rx_queue->sw_ring; in axgbe_rx_queue_release() 23 if (sw_ring) { in axgbe_rx_queue_release() 25 if (sw_ring[i]) in axgbe_rx_queue_release() 28 rte_free(sw_ring); in axgbe_rx_queue_release() 107 if (!rxq->sw_ring) { in axgbe_dev_rx_queue_setup() 468 sw_ring = tx_queue->sw_ring; in axgbe_tx_queue_release() 469 if (sw_ring) { in axgbe_tx_queue_release() 471 if (sw_ring[i]) in axgbe_tx_queue_release() 474 rte_free(sw_ring); in axgbe_tx_queue_release() [all …]
|
| /f-stack/dpdk/drivers/net/igc/ |
| H A D | igc_txrx.c | 351 struct igc_rx_entry * const sw_ring = rxq->sw_ring; in igc_recv_pkts() local 422 rxe = &sw_ring[rx_id]; in igc_recv_pkts() 495 struct igc_rx_entry * const sw_ring = rxq->sw_ring; in igc_recv_scattered_pkts() local 566 rxe = &sw_ring[rx_id]; in igc_recv_scattered_pkts() 700 if (rxq->sw_ring != NULL) { in igc_rx_queue_release_mbufs() 714 rte_free(rxq->sw_ring); in igc_rx_queue_release() 1398 if (rxq->sw_ring == NULL) { in eth_igc_rx_queue_setup() 1624 struct igc_tx_entry * const sw_ring = txq->sw_ring; in igc_xmit_pkts() local 1645 txe = &sw_ring[tx_id]; in igc_xmit_pkts() 1893 rte_free(txq->sw_ring); in igc_tx_queue_release() [all …]
|
| /f-stack/dpdk/drivers/net/atlantic/ |
| H A D | atl_rxtx.c | 56 struct atl_rx_entry *sw_ring; member 74 struct atl_tx_entry *sw_ring; member 158 if (rxq->sw_ring == NULL) { in atl_rx_queue_setup() 179 rte_free(rxq->sw_ring); in atl_rx_queue_setup() 206 tx_entry = txq->sw_ring; in atl_reset_tx_queue() 273 if (txq->sw_ring == NULL) { in atl_tx_queue_setup() 293 rte_free(txq->sw_ring); in atl_tx_queue_setup() 509 rte_free(rxq->sw_ring); in atl_rx_queue_release() 580 rte_free(txq->sw_ring); in atl_tx_queue_release() 926 struct atl_rx_entry *sw_ring = rxq->sw_ring; in atl_recv_pkts() local [all …]
|
| /f-stack/dpdk/drivers/net/e1000/ |
| H A D | em_rxtx.c | 298 struct em_tx_entry *sw_ring = txq->sw_ring; in em_xmit_cleanup() local 369 struct em_tx_entry *sw_ring; in eth_em_xmit_pkts() local 390 sw_ring = txq->sw_ring; in eth_em_xmit_pkts() 393 txe = &sw_ring[tx_id]; in eth_em_xmit_pkts() 701 sw_ring = rxq->sw_ring; in eth_em_recv_pkts() 760 rxe = &sw_ring[rx_id]; in eth_em_recv_pkts() 877 sw_ring = rxq->sw_ring; in eth_em_recv_scattered_pkts() 939 rxe = &sw_ring[rx_id]; in eth_em_recv_scattered_pkts() 1103 if (txq->sw_ring != NULL) { in em_tx_queue_release_mbufs() 1118 rte_free(txq->sw_ring); in em_tx_queue_release() [all …]
|
| H A D | igb_rxtx.c | 404 sw_ring = txq->sw_ring; in eth_igb_xmit_pkts() 407 txe = &sw_ring[tx_id]; in eth_igb_xmit_pkts() 829 sw_ring = rxq->sw_ring; in eth_igb_recv_pkts() 887 rxe = &sw_ring[rx_id]; in eth_igb_recv_pkts() 1016 sw_ring = rxq->sw_ring; in eth_igb_recv_scattered_pkts() 1078 rxe = &sw_ring[rx_id]; in eth_igb_recv_scattered_pkts() 1263 if (txq->sw_ring != NULL) { in igb_tx_queue_release_mbufs() 1278 rte_free(txq->sw_ring); in igb_tx_queue_release() 1303 sw_ring = txq->sw_ring; in igb_tx_done_cleanup() 1346 sw_ring[tx_id].mbuf); in igb_tx_done_cleanup() [all …]
|
| /f-stack/dpdk/drivers/net/bnxt/ |
| H A D | bnxt_txq.c | 27 struct bnxt_sw_tx_bd *sw_ring; in bnxt_tx_queue_release_mbufs() local 33 sw_ring = txq->tx_ring->tx_buf_ring; in bnxt_tx_queue_release_mbufs() 34 if (sw_ring) { in bnxt_tx_queue_release_mbufs() 36 if (sw_ring[i].mbuf) { in bnxt_tx_queue_release_mbufs() 37 rte_pktmbuf_free_seg(sw_ring[i].mbuf); in bnxt_tx_queue_release_mbufs() 38 sw_ring[i].mbuf = NULL; in bnxt_tx_queue_release_mbufs()
|
| H A D | bnxt_rxq.c | 206 struct rte_mbuf **sw_ring; in bnxt_rx_queue_release_mbufs() local 213 sw_ring = rxq->rx_ring->rx_buf_ring; in bnxt_rx_queue_release_mbufs() 214 if (sw_ring) { in bnxt_rx_queue_release_mbufs() 217 if (sw_ring[i]) { in bnxt_rx_queue_release_mbufs() 218 if (sw_ring[i] != &rxq->fake_mbuf) in bnxt_rx_queue_release_mbufs() 220 sw_ring[i] = NULL; in bnxt_rx_queue_release_mbufs() 225 sw_ring = rxq->rx_ring->ag_buf_ring; in bnxt_rx_queue_release_mbufs() 226 if (sw_ring) { in bnxt_rx_queue_release_mbufs() 229 if (sw_ring[i]) { in bnxt_rx_queue_release_mbufs() 230 rte_pktmbuf_free_seg(sw_ring[i]); in bnxt_rx_queue_release_mbufs() [all …]
|
| /f-stack/dpdk/drivers/net/ixgbe/ |
| H A D | ixgbe_rxtx_vec_common.h | 149 if (txq->sw_ring == NULL || txq->nb_tx_free == max_desc) in _ixgbe_tx_queue_release_mbufs_vec() 174 if (rxq->sw_ring == NULL || rxq->rxrearm_nb >= rxq->nb_rx_desc) in _ixgbe_rx_queue_release_mbufs_vec() 180 if (rxq->sw_ring[i].mbuf != NULL) in _ixgbe_rx_queue_release_mbufs_vec() 181 rte_pktmbuf_free_seg(rxq->sw_ring[i].mbuf); in _ixgbe_rx_queue_release_mbufs_vec() 187 if (rxq->sw_ring[i].mbuf != NULL) in _ixgbe_rx_queue_release_mbufs_vec() 188 rte_pktmbuf_free_seg(rxq->sw_ring[i].mbuf); in _ixgbe_rx_queue_release_mbufs_vec() 195 memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); in _ixgbe_rx_queue_release_mbufs_vec() 204 if (txq->sw_ring != NULL) { in _ixgbe_tx_free_swring_vec()
|
| /f-stack/dpdk/drivers/net/i40e/ |
| H A D | i40e_rxtx_vec_common.h | 97 txep = &txq->sw_ring[txq->tx_next_dd - (n - 1)]; in i40e_tx_free_bufs() 150 if (rxq->sw_ring == NULL || rxq->rxrearm_nb >= rxq->nb_rx_desc) in _i40e_rx_queue_release_mbufs_vec() 156 if (rxq->sw_ring[i].mbuf != NULL) in _i40e_rx_queue_release_mbufs_vec() 157 rte_pktmbuf_free_seg(rxq->sw_ring[i].mbuf); in _i40e_rx_queue_release_mbufs_vec() 163 if (rxq->sw_ring[i].mbuf != NULL) in _i40e_rx_queue_release_mbufs_vec() 164 rte_pktmbuf_free_seg(rxq->sw_ring[i].mbuf); in _i40e_rx_queue_release_mbufs_vec() 171 memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); in _i40e_rx_queue_release_mbufs_vec()
|
| /f-stack/dpdk/drivers/raw/ntb/ |
| H A D | ntb.c | 277 if (!q || !q->sw_ring) { in ntb_rxq_release_mbufs() 300 rte_free(rxq->sw_ring); in ntb_rxq_release() 339 rxq->sw_ring = in ntb_rxq_setup() 345 if (!rxq->sw_ring) { in ntb_rxq_setup() 362 if (!q || !q->sw_ring) { in ntb_txq_release_mbufs() 385 rte_free(txq->sw_ring); in ntb_txq_release() 420 txq->sw_ring = in ntb_txq_setup() 426 if (!txq->sw_ring) { in ntb_txq_setup() 581 struct ntb_tx_entry *sw_ring = txq->sw_ring; in ntb_enqueue_cleanup() local 607 struct ntb_tx_entry *sw_ring = txq->sw_ring; in ntb_enqueue_bufs() local [all …]
|
| /f-stack/dpdk/drivers/net/fm10k/ |
| H A D | fm10k_rxtx.c | 116 mbuf = q->sw_ring[next_dd]; in fm10k_recv_pkts() 153 rte_prefetch0(q->sw_ring[next_dd]); in fm10k_recv_pkts() 162 rte_prefetch0(&q->sw_ring[next_dd]); in fm10k_recv_pkts() 170 (void **)&q->sw_ring[q->next_alloc], in fm10k_recv_pkts() 187 mbuf = q->sw_ring[q->next_alloc]; in fm10k_recv_pkts() 230 mbuf = q->sw_ring[next_dd]; in fm10k_recv_scattered_pkts() 242 rte_prefetch0(q->sw_ring[next_dd]); in fm10k_recv_scattered_pkts() 251 rte_prefetch0(&q->sw_ring[next_dd]); in fm10k_recv_scattered_pkts() 345 mbuf = q->sw_ring[q->next_alloc]; in fm10k_recv_scattered_pkts() 617 q->sw_ring[q->next_free] = mb; in tx_xmit_pkt() [all …]
|