Home
last modified time | relevance | path

Searched refs:rxep (Results 1 – 16 of 16) sorted by relevance

/f-stack/dpdk/drivers/net/hns3/
H A Dhns3_rxtx_vec.c53 struct hns3_entry *rxep = &rxq->sw_ring[rxq->rx_rearm_start]; in hns3_rxq_rearm_mbuf() local
58 if (unlikely(rte_mempool_get_bulk(rxq->mb_pool, (void *)rxep, in hns3_rxq_rearm_mbuf()
65 rxep += REARM_LOOP_STEP_NUM, rxdp += REARM_LOOP_STEP_NUM) { in hns3_rxq_rearm_mbuf()
68 rte_prefetch_non_temporal(rxep[4].mbuf); in hns3_rxq_rearm_mbuf()
69 rte_prefetch_non_temporal(rxep[5].mbuf); in hns3_rxq_rearm_mbuf()
70 rte_prefetch_non_temporal(rxep[6].mbuf); in hns3_rxq_rearm_mbuf()
71 rte_prefetch_non_temporal(rxep[7].mbuf); in hns3_rxq_rearm_mbuf()
74 dma_addr = rte_mbuf_data_iova_default(rxep[0].mbuf); in hns3_rxq_rearm_mbuf()
78 dma_addr = rte_mbuf_data_iova_default(rxep[1].mbuf); in hns3_rxq_rearm_mbuf()
82 dma_addr = rte_mbuf_data_iova_default(rxep[2].mbuf); in hns3_rxq_rearm_mbuf()
[all …]
H A Dhns3_rxtx_vec_sve.c239 struct hns3_entry *rxep = &rxq->sw_ring[rxq->rx_rearm_start]; in hns3_rxq_rearm_mbuf_sve() local
241 struct hns3_entry *rxep_tmp = rxep; in hns3_rxq_rearm_mbuf_sve()
244 if (unlikely(rte_mempool_get_bulk(rxq->mb_pool, (void *)rxep, in hns3_rxq_rearm_mbuf_sve()
257 rxep += REARM_LOOP_STEP_NUM, rxdp += REARM_LOOP_STEP_NUM) { in hns3_rxq_rearm_mbuf_sve()
259 iova[0] = rxep[0].mbuf->buf_iova; in hns3_rxq_rearm_mbuf_sve()
260 iova[1] = rxep[1].mbuf->buf_iova; in hns3_rxq_rearm_mbuf_sve()
261 iova[2] = rxep[2].mbuf->buf_iova; in hns3_rxq_rearm_mbuf_sve()
262 iova[3] = rxep[3].mbuf->buf_iova; in hns3_rxq_rearm_mbuf_sve()
/f-stack/dpdk/drivers/net/ice/
H A Dice_rxtx_vec_avx2.c19 struct ice_rx_entry *rxep = &rxq->sw_ring[rxq->rxrearm_start]; in ice_rxq_rearm() local
25 (void *)rxep, in ice_rxq_rearm()
33 rxep[i].mbuf = &rxq->fake_mbuf; in ice_rxq_rearm()
49 for (i = 0; i < ICE_RXQ_REARM_THRESH; i += 2, rxep += 2) { in ice_rxq_rearm()
52 mb0 = rxep[0].mbuf; in ice_rxq_rearm()
53 mb1 = rxep[1].mbuf; in ice_rxq_rearm()
79 i += 4, rxep += 4, rxdp += 4) { in ice_rxq_rearm()
83 mb0 = rxep[0].mbuf; in ice_rxq_rearm()
84 mb1 = rxep[1].mbuf; in ice_rxq_rearm()
85 mb2 = rxep[2].mbuf; in ice_rxq_rearm()
[all …]
H A Dice_rxtx_vec_sse.c38 struct ice_rx_entry *rxep = &rxq->sw_ring[rxq->rxrearm_start]; in ice_rxq_rearm() local
48 (void *)rxep, in ice_rxq_rearm()
54 rxep[i].mbuf = &rxq->fake_mbuf; in ice_rxq_rearm()
65 for (i = 0; i < ICE_RXQ_REARM_THRESH; i += 2, rxep += 2) { in ice_rxq_rearm()
68 mb0 = rxep[0].mbuf; in ice_rxq_rearm()
69 mb1 = rxep[1].mbuf; in ice_rxq_rearm()
H A Dice_rxtx_vec_avx512.c21 struct ice_rx_entry *rxep = &rxq->sw_ring[rxq->rxrearm_start]; in ice_rxq_rearm() local
43 rxep[i].mbuf = &rxq->fake_mbuf; in ice_rxq_rearm()
72 _mm512_store_si512(rxep, mbuf_ptrs); in ice_rxq_rearm()
115 rxep += 8, rxdp += 8, cache->len -= 8; in ice_rxq_rearm()
H A Dice_rxtx.c1497 struct ice_rx_entry *rxep; in ice_rx_scan_hw_ring() local
1507 rxep = &rxq->sw_ring[rxq->rx_tail]; in ice_rx_scan_hw_ring()
1520 rxdp += ICE_LOOK_AHEAD, rxep += ICE_LOOK_AHEAD) { in ice_rx_scan_hw_ring()
1535 mb = rxep[j].mbuf; in ice_rx_scan_hw_ring()
1552 rxq->rx_stage[i + j] = rxep[j].mbuf; in ice_rx_scan_hw_ring()
1592 struct ice_rx_entry *rxep; in ice_rx_alloc_bufs() local
1601 rxep = &rxq->sw_ring[alloc_idx]; in ice_rx_alloc_bufs()
1602 diag = rte_mempool_get_bulk(rxq->mp, (void *)rxep, in ice_rx_alloc_bufs()
1613 rte_prefetch0(rxep[i + 1].mbuf); in ice_rx_alloc_bufs()
1615 mb = rxep[i].mbuf; in ice_rx_alloc_bufs()
/f-stack/dpdk/drivers/net/i40e/
H A Di40e_rxtx_vec_avx2.c27 struct i40e_rx_entry *rxep = &rxq->sw_ring[rxq->rxrearm_start]; in i40e_rxq_rearm() local
33 (void *)rxep, in i40e_rxq_rearm()
40 rxep[i].mbuf = &rxq->fake_mbuf; in i40e_rxq_rearm()
56 for (i = 0; i < RTE_I40E_RXQ_REARM_THRESH; i += 2, rxep += 2) { in i40e_rxq_rearm()
59 mb0 = rxep[0].mbuf; in i40e_rxq_rearm()
60 mb1 = rxep[1].mbuf; in i40e_rxq_rearm()
86 i += 4, rxep += 4, rxdp += 4) { in i40e_rxq_rearm()
90 mb0 = rxep[0].mbuf; in i40e_rxq_rearm()
91 mb1 = rxep[1].mbuf; in i40e_rxq_rearm()
92 mb2 = rxep[2].mbuf; in i40e_rxq_rearm()
[all …]
H A Di40e_rxtx_vec_altivec.c27 struct i40e_rx_entry *rxep = &rxq->sw_ring[rxq->rxrearm_start]; in i40e_rxq_rearm() local
39 (void *)rxep, in i40e_rxq_rearm()
45 rxep[i].mbuf = &rxq->fake_mbuf; in i40e_rxq_rearm()
56 for (i = 0; i < RTE_I40E_RXQ_REARM_THRESH; i += 2, rxep += 2) { in i40e_rxq_rearm()
60 mb0 = rxep[0].mbuf; in i40e_rxq_rearm()
61 mb1 = rxep[1].mbuf; in i40e_rxq_rearm()
H A Di40e_rxtx_vec_neon.c26 struct i40e_rx_entry *rxep = &rxq->sw_ring[rxq->rxrearm_start]; in i40e_rxq_rearm() local
36 (void *)rxep, in i40e_rxq_rearm()
41 rxep[i].mbuf = &rxq->fake_mbuf; in i40e_rxq_rearm()
51 for (i = 0; i < RTE_I40E_RXQ_REARM_THRESH; i += 2, rxep += 2) { in i40e_rxq_rearm()
52 mb0 = rxep[0].mbuf; in i40e_rxq_rearm()
53 mb1 = rxep[1].mbuf; in i40e_rxq_rearm()
H A Di40e_rxtx_vec_sse.c27 struct i40e_rx_entry *rxep = &rxq->sw_ring[rxq->rxrearm_start]; in i40e_rxq_rearm() local
37 (void *)rxep, in i40e_rxq_rearm()
43 rxep[i].mbuf = &rxq->fake_mbuf; in i40e_rxq_rearm()
54 for (i = 0; i < RTE_I40E_RXQ_REARM_THRESH; i += 2, rxep += 2) { in i40e_rxq_rearm()
57 mb0 = rxep[0].mbuf; in i40e_rxq_rearm()
58 mb1 = rxep[1].mbuf; in i40e_rxq_rearm()
H A Di40e_rxtx.c416 struct i40e_rx_entry *rxep; in i40e_rx_scan_hw_ring() local
427 rxep = &rxq->sw_ring[rxq->rx_tail]; in i40e_rx_scan_hw_ring()
442 rxdp += I40E_LOOK_AHEAD, rxep += I40E_LOOK_AHEAD) { in i40e_rx_scan_hw_ring()
461 mb = rxep[j].mbuf; in i40e_rx_scan_hw_ring()
490 rxq->rx_stage[i + j] = rxep[j].mbuf; in i40e_rx_scan_hw_ring()
526 struct i40e_rx_entry *rxep; in i40e_rx_alloc_bufs() local
535 rxep = &(rxq->sw_ring[alloc_idx]); in i40e_rx_alloc_bufs()
536 diag = rte_mempool_get_bulk(rxq->mp, (void *)rxep, in i40e_rx_alloc_bufs()
547 rte_prefetch0(rxep[i + 1].mbuf); in i40e_rx_alloc_bufs()
549 mb = rxep[i].mbuf; in i40e_rx_alloc_bufs()
/f-stack/dpdk/drivers/net/ixgbe/
H A Dixgbe_rxtx_vec_neon.c22 struct ixgbe_rx_entry *rxep = &rxq->sw_ring[rxq->rxrearm_start]; in ixgbe_rxq_rearm() local
33 (void *)rxep, in ixgbe_rxq_rearm()
38 rxep[i].mbuf = &rxq->fake_mbuf; in ixgbe_rxq_rearm()
51 for (i = 0; i < RTE_IXGBE_RXQ_REARM_THRESH; i += 2, rxep += 2) { in ixgbe_rxq_rearm()
52 mb0 = rxep[0].mbuf; in ixgbe_rxq_rearm()
53 mb1 = rxep[1].mbuf; in ixgbe_rxq_rearm()
H A Dixgbe_rxtx_vec_sse.c25 struct ixgbe_rx_entry *rxep = &rxq->sw_ring[rxq->rxrearm_start]; in ixgbe_rxq_rearm() local
37 (void *)rxep, in ixgbe_rxq_rearm()
43 rxep[i].mbuf = &rxq->fake_mbuf; in ixgbe_rxq_rearm()
54 for (i = 0; i < RTE_IXGBE_RXQ_REARM_THRESH; i += 2, rxep += 2) { in ixgbe_rxq_rearm()
57 mb0 = rxep[0].mbuf; in ixgbe_rxq_rearm()
58 mb1 = rxep[1].mbuf; in ixgbe_rxq_rearm()
H A Dixgbe_rxtx.c1492 struct ixgbe_rx_entry *rxep; in ixgbe_rx_scan_hw_ring() local
1505 rxep = &rxq->sw_ring[rxq->rx_tail]; in ixgbe_rx_scan_hw_ring()
1517 i += LOOK_AHEAD, rxdp += LOOK_AHEAD, rxep += LOOK_AHEAD) { in ixgbe_rx_scan_hw_ring()
1537 mb = rxep[j].mbuf; in ixgbe_rx_scan_hw_ring()
1569 rxq->rx_stage[i + j] = rxep[j].mbuf; in ixgbe_rx_scan_hw_ring()
1590 struct ixgbe_rx_entry *rxep; in ixgbe_rx_alloc_bufs() local
1598 rxep = &rxq->sw_ring[alloc_idx]; in ixgbe_rx_alloc_bufs()
1599 diag = rte_mempool_get_bulk(rxq->mb_pool, (void *)rxep, in ixgbe_rx_alloc_bufs()
1607 mb = rxep[i].mbuf; in ixgbe_rx_alloc_bufs()
/f-stack/dpdk/drivers/net/iavf/
H A Diavf_rxtx.c1621 struct rte_mbuf **rxep; in iavf_rx_scan_hw_ring_flex_rxd() local
1631 rxep = &rxq->sw_ring[rxq->rx_tail]; in iavf_rx_scan_hw_ring_flex_rxd()
1662 mb = rxep[j]; in iavf_rx_scan_hw_ring_flex_rxd()
1680 rxq->rx_stage[i + j] = rxep[j]; in iavf_rx_scan_hw_ring_flex_rxd()
1697 struct rte_mbuf **rxep; in iavf_rx_scan_hw_ring() local
1744 mb = rxep[j]; in iavf_rx_scan_hw_ring()
1770 rxq->rx_stage[i + j] = rxep[j]; in iavf_rx_scan_hw_ring()
1806 struct rte_mbuf **rxep; in iavf_rx_alloc_bufs() local
1815 rxep = &rxq->sw_ring[alloc_idx]; in iavf_rx_alloc_bufs()
1827 rte_prefetch0(rxep[i + 1]); in iavf_rx_alloc_bufs()
[all …]
/f-stack/dpdk/drivers/net/txgbe/
H A Dtxgbe_rxtx.c1119 struct txgbe_rx_entry *rxep; in txgbe_rx_scan_hw_ring() local
1131 rxep = &rxq->sw_ring[rxq->rx_tail]; in txgbe_rx_scan_hw_ring()
1143 i += LOOK_AHEAD, rxdp += LOOK_AHEAD, rxep += LOOK_AHEAD) { in txgbe_rx_scan_hw_ring()
1162 mb = rxep[j].mbuf; in txgbe_rx_scan_hw_ring()
1194 rxq->rx_stage[i + j] = rxep[j].mbuf; in txgbe_rx_scan_hw_ring()
1212 struct txgbe_rx_entry *rxep; in txgbe_rx_alloc_bufs() local
1220 rxep = &rxq->sw_ring[alloc_idx]; in txgbe_rx_alloc_bufs()
1221 diag = rte_mempool_get_bulk(rxq->mb_pool, (void *)rxep, in txgbe_rx_alloc_bufs()
1229 mb = rxep[i].mbuf; in txgbe_rx_alloc_bufs()