| /dpdk/drivers/net/bnxt/ |
| H A D | bnxt_rxtx_vec_common.h | 106 uint16_t cons, raw_cons = txr->tx_raw_cons; in bnxt_tx_cmp_vec_fast() 110 ring_mask = txr->tx_ring_struct->ring_mask; in bnxt_tx_cmp_vec_fast() 111 ring_size = txr->tx_ring_struct->ring_size; in bnxt_tx_cmp_vec_fast() 115 pool = txr->tx_buf_ring[cons]->pool; in bnxt_tx_cmp_vec_fast() 125 memset(&txr->tx_buf_ring[cons], 0, in bnxt_tx_cmp_vec_fast() 130 txr->tx_raw_cons = raw_cons; in bnxt_tx_cmp_vec_fast() 149 pool = txr->tx_buf_ring[cons]->pool; in bnxt_tx_cmp_vec() 153 mbuf = txr->tx_buf_ring[cons + blk]; in bnxt_tx_cmp_vec() 162 (void **)&txr->tx_buf_ring[cons], in bnxt_tx_cmp_vec() 164 memset(&txr->tx_buf_ring[cons], 0, in bnxt_tx_cmp_vec() [all …]
|
| H A D | bnxt_txr.c | 63 struct bnxt_tx_ring_info *txr; in bnxt_init_tx_ring_struct() local 69 if (txr == NULL) in bnxt_init_tx_ring_struct() 71 txq->tx_ring = txr; in bnxt_init_tx_ring_struct() 78 txr->tx_ring_struct = ring; in bnxt_init_tx_ring_struct() 185 tx_buf = &txr->tx_buf_ring[prod]; in bnxt_start_xmit() 188 txbd = &txr->tx_desc_ring[prod]; in bnxt_start_xmit() 226 txr->tx_raw_prod = RING_NEXT(txr->tx_raw_prod); in bnxt_start_xmit() 334 txr->tx_raw_prod = RING_NEXT(txr->tx_raw_prod); in bnxt_start_xmit() 350 txr->tx_raw_prod = RING_NEXT(txr->tx_raw_prod); in bnxt_start_xmit() 388 txr->tx_raw_cons = raw_cons; in bnxt_tx_cmp_fast() [all …]
|
| H A D | bnxt_rxtx_vec_sse.c | 365 struct bnxt_tx_ring_info *txr = txq->tx_ring; in bnxt_xmit_fixed_burst_vec() local 366 uint16_t tx_prod, tx_raw_prod = txr->tx_raw_prod; in bnxt_xmit_fixed_burst_vec() 371 tx_prod = RING_IDX(txr->tx_ring_struct, tx_raw_prod); in bnxt_xmit_fixed_burst_vec() 372 txbd = &txr->tx_desc_ring[tx_prod]; in bnxt_xmit_fixed_burst_vec() 373 tx_buf = &txr->tx_buf_ring[tx_prod]; in bnxt_xmit_fixed_burst_vec() 412 bnxt_db_write(&txr->tx_db, tx_raw_prod); in bnxt_xmit_fixed_burst_vec() 414 txr->tx_raw_prod = tx_raw_prod; in bnxt_xmit_fixed_burst_vec() 425 struct bnxt_tx_ring_info *txr = txq->tx_ring; in bnxt_xmit_pkts_vec() local 426 uint16_t ring_size = txr->tx_ring_struct->ring_size; in bnxt_xmit_pkts_vec() 452 (txr->tx_raw_prod & (ring_size - 1))); in bnxt_xmit_pkts_vec()
|
| H A D | bnxt_rxtx_vec_avx2.c | 439 struct bnxt_tx_ring_info *txr = txq->tx_ring; in bnxt_xmit_fixed_burst_vec() local 440 uint16_t tx_prod, tx_raw_prod = txr->tx_raw_prod; in bnxt_xmit_fixed_burst_vec() 445 tx_prod = RING_IDX(txr->tx_ring_struct, tx_raw_prod); in bnxt_xmit_fixed_burst_vec() 446 txbd = &txr->tx_desc_ring[tx_prod]; in bnxt_xmit_fixed_burst_vec() 447 tx_buf = &txr->tx_buf_ring[tx_prod]; in bnxt_xmit_fixed_burst_vec() 538 bnxt_db_write(&txr->tx_db, tx_raw_prod); in bnxt_xmit_fixed_burst_vec() 540 txr->tx_raw_prod = tx_raw_prod; in bnxt_xmit_fixed_burst_vec() 551 struct bnxt_tx_ring_info *txr = txq->tx_ring; in bnxt_xmit_pkts_vec_avx2() local 552 uint16_t ring_size = txr->tx_ring_struct->ring_size; in bnxt_xmit_pkts_vec_avx2() 578 (txr->tx_raw_prod & (ring_size - 1))); in bnxt_xmit_pkts_vec_avx2()
|
| H A D | bnxt_rxtx_vec_neon.c | 378 struct bnxt_tx_ring_info *txr = txq->tx_ring; in bnxt_xmit_fixed_burst_vec() local 379 uint16_t tx_prod, tx_raw_prod = txr->tx_raw_prod; in bnxt_xmit_fixed_burst_vec() 396 tx_prod = RING_IDX(txr->tx_ring_struct, tx_raw_prod); in bnxt_xmit_fixed_burst_vec() 397 tx_buf = &txr->tx_buf_ring[tx_prod]; in bnxt_xmit_fixed_burst_vec() 400 txbd = &txr->tx_desc_ring[tx_prod]; in bnxt_xmit_fixed_burst_vec() 416 bnxt_db_write(&txr->tx_db, tx_raw_prod); in bnxt_xmit_fixed_burst_vec() 418 txr->tx_raw_prod = tx_raw_prod; in bnxt_xmit_fixed_burst_vec()
|
| H A D | bnxt_ring.c | 870 struct bnxt_tx_ring_info *txr = txq->tx_ring; in bnxt_alloc_hwrm_tx_ring() local 871 struct bnxt_ring *ring = txr->tx_ring_struct; in bnxt_alloc_hwrm_tx_ring() 901 bnxt_set_db(bp, &txr->tx_db, HWRM_RING_ALLOC_INPUT_RING_TYPE_TX, in bnxt_alloc_hwrm_tx_ring()
|
| H A D | bnxt_hwrm.c | 6204 struct bnxt_tx_ring_info *txr = txq->tx_ring; in bnxt_free_hwrm_tx_ring() local 6205 struct bnxt_ring *ring = txr->tx_ring_struct; in bnxt_free_hwrm_tx_ring() 6211 txr->tx_raw_prod = 0; in bnxt_free_hwrm_tx_ring() 6212 txr->tx_raw_cons = 0; in bnxt_free_hwrm_tx_ring() 6213 memset(txr->tx_desc_ring, 0, in bnxt_free_hwrm_tx_ring() 6214 txr->tx_ring_struct->ring_size * sizeof(*txr->tx_desc_ring)); in bnxt_free_hwrm_tx_ring() 6215 memset(txr->tx_buf_ring, 0, in bnxt_free_hwrm_tx_ring() 6216 txr->tx_ring_struct->ring_size * sizeof(*txr->tx_buf_ring)); in bnxt_free_hwrm_tx_ring()
|
| /dpdk/drivers/net/virtio/ |
| H A D | virtqueue.c | 189 struct virtio_tx_region *txr; in virtqueue_txvq_reset_packed() local 203 txr = txvq->virtio_net_hdr_mz->addr; in virtqueue_txvq_reset_packed() 217 start_dp = txr[desc_idx].tx_packed_indir; in virtqueue_txvq_reset_packed() 219 RTE_DIM(txr[desc_idx].tx_packed_indir)); in virtqueue_txvq_reset_packed() 221 + desc_idx * sizeof(*txr) in virtqueue_txvq_reset_packed()
|
| H A D | virtqueue.h | 714 struct virtio_tx_region *txr = txvq->virtio_net_hdr_mz->addr; in virtqueue_enqueue_xmit_packed() local 757 RTE_PTR_DIFF(&txr[idx].tx_packed_indir, txr); in virtqueue_enqueue_xmit_packed() 766 hdr = (struct virtio_net_hdr *)&txr[idx].tx_hdr; in virtqueue_enqueue_xmit_packed() 769 start_dp = txr[idx].tx_packed_indir; in virtqueue_enqueue_xmit_packed() 776 RTE_PTR_DIFF(&txr[idx].tx_hdr, txr); in virtqueue_enqueue_xmit_packed() 778 hdr = (struct virtio_net_hdr *)&txr[idx].tx_hdr; in virtqueue_enqueue_xmit_packed()
|
| H A D | virtio_rxtx.c | 545 struct virtio_tx_region *txr = txvq->virtio_net_hdr_mz->addr; in virtqueue_enqueue_xmit() local 583 RTE_PTR_DIFF(&txr[idx].tx_indir, txr); in virtqueue_enqueue_xmit() 586 hdr = (struct virtio_net_hdr *)&txr[idx].tx_hdr; in virtqueue_enqueue_xmit() 589 start_dp = txr[idx].tx_indir; in virtqueue_enqueue_xmit() 596 RTE_PTR_DIFF(&txr[idx].tx_hdr, txr); in virtqueue_enqueue_xmit() 599 hdr = (struct virtio_net_hdr *)&txr[idx].tx_hdr; in virtqueue_enqueue_xmit()
|
| H A D | virtio_ethdev.c | 647 struct virtio_tx_region *txr; in virtio_init_queue() local 650 txr = hdr_mz->addr; in virtio_init_queue() 651 memset(txr, 0, vq_size * sizeof(*txr)); in virtio_init_queue() 655 struct vring_desc *start_dp = txr[i].tx_indir; in virtio_init_queue() 657 RTE_DIM(txr[i].tx_indir)); in virtio_init_queue() 659 + i * sizeof(*txr) in virtio_init_queue() 666 txr[i].tx_packed_indir; in virtio_init_queue() 668 RTE_DIM(txr[i].tx_packed_indir)); in virtio_init_queue() 670 + i * sizeof(*txr) in virtio_init_queue()
|
| /dpdk/drivers/net/enetc/ |
| H A D | enetc_ethdev.c | 234 enetc_alloc_txbdr(struct enetc_bdr *txr, uint16_t nb_desc) in enetc_alloc_txbdr() argument 239 txr->q_swbd = rte_malloc(NULL, size, ENETC_BD_RING_ALIGN); in enetc_alloc_txbdr() 240 if (txr->q_swbd == NULL) in enetc_alloc_txbdr() 244 txr->bd_base = rte_malloc(NULL, size, ENETC_BD_RING_ALIGN); in enetc_alloc_txbdr() 245 if (txr->bd_base == NULL) { in enetc_alloc_txbdr() 246 rte_free(txr->q_swbd); in enetc_alloc_txbdr() 247 txr->q_swbd = NULL; in enetc_alloc_txbdr() 251 txr->bd_count = nb_desc; in enetc_alloc_txbdr() 252 txr->next_to_clean = 0; in enetc_alloc_txbdr() 253 txr->next_to_use = 0; in enetc_alloc_txbdr()
|
| /dpdk/drivers/net/e1000/ |
| H A D | em_rxtx.c | 300 volatile struct e1000_data_desc *txr = txq->tx_ring; in em_xmit_cleanup() local 313 if (! (txr[desc_to_clean_to].upper.fields.status & E1000_TXD_STAT_DD)) in em_xmit_cleanup() 343 txr[desc_to_clean_to].upper.fields.status = 0; in em_xmit_cleanup() 372 volatile struct e1000_data_desc *txr; in eth_em_xmit_pkts() local 392 txr = txq->tx_ring; in eth_em_xmit_pkts() 523 &txr[tx_id]; in eth_em_xmit_pkts() 551 txd = &txr[tx_id]; in eth_em_xmit_pkts()
|
| H A D | igb_rxtx.c | 385 volatile union e1000_adv_tx_desc *txr; in eth_igb_xmit_pkts() local 406 txr = txq->tx_ring; in eth_igb_xmit_pkts() 501 if (! (txr[tx_end].wb.status & E1000_TXD_STAT_DD)) { in eth_igb_xmit_pkts() 545 &txr[tx_id]; in eth_igb_xmit_pkts() 571 txd = &txr[tx_id]; in eth_igb_xmit_pkts() 1295 volatile union e1000_adv_tx_desc *txr; in igb_tx_done_cleanup() local 1306 txr = txq->tx_ring; in igb_tx_done_cleanup() 1332 if (txr[tx_last].wb.status & in igb_tx_done_cleanup()
|
| /dpdk/drivers/net/igc/ |
| H A D | igc_txrx.c | 1613 volatile union igc_adv_tx_desc * const txr = txq->tx_ring; in igc_xmit_pkts() local 1721 if (!(txr[tx_end].wb.status & IGC_TXD_STAT_DD)) { in igc_xmit_pkts() 1767 igc_adv_tx_context_desc *)&txr[tx_id]; in igc_xmit_pkts() 1799 txd = &txr[tx_id]; in igc_xmit_pkts() 2046 volatile union igc_adv_tx_desc *txr; in eth_igc_tx_done_cleanup() local 2058 txr = txq->tx_ring; in eth_igc_tx_done_cleanup() 2086 if (!(txr[tx_last].wb.status & in eth_igc_tx_done_cleanup()
|
| /dpdk/drivers/net/i40e/ |
| H A D | i40e_rxtx.c | 1081 volatile struct i40e_tx_desc *txr; in i40e_xmit_pkts() local 1100 txr = txq->tx_ring; in i40e_xmit_pkts() 1186 &txr[tx_id]; in i40e_xmit_pkts() 1241 txd = &txr[tx_id]; in i40e_xmit_pkts() 1267 txd = &txr[tx_id]; in i40e_xmit_pkts() 1439 volatile struct i40e_tx_desc *txr = txq->tx_ring; in tx_xmit_pkts() local 1459 txr[txq->tx_next_rs].cmd_type_offset_bsz |= in tx_xmit_pkts() 1472 txr[txq->tx_next_rs].cmd_type_offset_bsz |= in tx_xmit_pkts()
|
| /dpdk/drivers/net/ngbe/ |
| H A D | ngbe_rxtx.c | 513 volatile struct ngbe_tx_desc *txr = txq->tx_ring; in ngbe_xmit_cleanup() local 527 status = txr[desc_to_clean_to].dw3; in ngbe_xmit_cleanup() 560 txr[desc_to_clean_to].dw3 = 0; in ngbe_xmit_cleanup() 577 volatile struct ngbe_tx_desc *txr; in ngbe_xmit_pkts() local 600 txr = txq->tx_ring; in ngbe_xmit_pkts() 768 &txr[tx_id]; in ngbe_xmit_pkts() 801 txd = &txr[tx_id]; in ngbe_xmit_pkts()
|
| /dpdk/drivers/net/txgbe/ |
| H A D | txgbe_rxtx.c | 602 volatile struct txgbe_tx_desc *txr = txq->tx_ring; in txgbe_xmit_cleanup() local 616 status = txr[desc_to_clean_to].dw3; in txgbe_xmit_cleanup() 650 txr[desc_to_clean_to].dw3 = 0; in txgbe_xmit_cleanup() 725 volatile struct txgbe_tx_desc *txr; in txgbe_xmit_pkts() local 751 txr = txq->tx_ring; in txgbe_xmit_pkts() 944 &txr[tx_id]; in txgbe_xmit_pkts() 982 txd = &txr[tx_id]; in txgbe_xmit_pkts()
|
| /dpdk/drivers/net/ixgbe/ |
| H A D | ixgbe_rxtx.c | 568 volatile union ixgbe_adv_tx_desc *txr = txq->tx_ring; in ixgbe_xmit_cleanup() local 582 status = txr[desc_to_clean_to].wb.status; in ixgbe_xmit_cleanup() 613 txr[desc_to_clean_to].wb.status = 0; in ixgbe_xmit_cleanup() 630 volatile union ixgbe_adv_tx_desc *txr; in ixgbe_xmit_pkts() local 656 txr = txq->tx_ring; in ixgbe_xmit_pkts() 850 &txr[tx_id]; in ixgbe_xmit_pkts() 887 txd = &txr[tx_id]; in ixgbe_xmit_pkts()
|
| /dpdk/drivers/net/iavf/ |
| H A D | iavf_rxtx.c | 2621 volatile struct iavf_tx_desc *txr = txq->tx_ring; in iavf_xmit_pkts() local 2709 &txr[desc_idx]; in iavf_xmit_pkts() 2736 &txr[desc_idx]; in iavf_xmit_pkts() 2759 &txr[desc_idx]; in iavf_xmit_pkts()
|
| /dpdk/drivers/net/ice/ |
| H A D | ice_rxtx.c | 3152 volatile struct ice_tx_desc *txr = txq->tx_ring; in tx_xmit_pkts() local 3172 txr[txq->tx_next_rs].cmd_type_offset_bsz |= in tx_xmit_pkts() 3185 txr[txq->tx_next_rs].cmd_type_offset_bsz |= in tx_xmit_pkts()
|