| /f-stack/dpdk/drivers/net/ice/ |
| H A D | ice_rxtx_vec_avx2.c | 849 ice_vtx(volatile struct ice_tx_desc *txdp, in ice_vtx() argument 857 ice_vtx1(txdp, *pkt, flags); in ice_vtx() 858 nb_pkts--, txdp++, pkt++; in ice_vtx() 898 ice_vtx1(txdp, *pkt, flags); in ice_vtx() 899 txdp++, pkt++, nb_pkts--; in ice_vtx() 908 volatile struct ice_tx_desc *txdp; in ice_xmit_fixed_burst_vec_avx2() local 925 txdp = &txq->tx_ring[tx_id]; in ice_xmit_fixed_burst_vec_avx2() 934 ice_vtx(txdp, tx_pkts, n - 1, flags); in ice_xmit_fixed_burst_vec_avx2() 936 txdp += (n - 1); in ice_xmit_fixed_burst_vec_avx2() 938 ice_vtx1(txdp, *tx_pkts++, rs); in ice_xmit_fixed_burst_vec_avx2() [all …]
|
| H A D | ice_rxtx_vec_sse.c | 630 ice_vtx1(volatile struct ice_tx_desc *txdp, struct rte_mbuf *pkt, in ice_vtx1() argument 640 _mm_store_si128((__m128i *)txdp, descriptor); in ice_vtx1() 649 for (i = 0; i < nb_pkts; ++i, ++txdp, ++pkt) in ice_vtx() 650 ice_vtx1(txdp, *pkt, flags); in ice_vtx() 658 volatile struct ice_tx_desc *txdp; in ice_xmit_fixed_burst_vec() local 677 txdp = &txq->tx_ring[tx_id]; in ice_xmit_fixed_burst_vec() 686 for (i = 0; i < n - 1; ++i, ++tx_pkts, ++txdp) in ice_xmit_fixed_burst_vec() 687 ice_vtx1(txdp, *tx_pkts, flags); in ice_xmit_fixed_burst_vec() 689 ice_vtx1(txdp, *tx_pkts++, rs); in ice_xmit_fixed_burst_vec() 697 txdp = &txq->tx_ring[tx_id]; in ice_xmit_fixed_burst_vec() [all …]
|
| H A D | ice_rxtx_vec_avx512.c | 930 ice_vtx1(volatile struct ice_tx_desc *txdp, in ice_vtx1() argument 944 ice_vtx(volatile struct ice_tx_desc *txdp, in ice_vtx() argument 983 ice_vtx1(txdp, *pkt, flags); in ice_vtx() 984 txdp++, pkt++, nb_pkts--; in ice_vtx() 1003 volatile struct ice_tx_desc *txdp; in ice_xmit_fixed_burst_vec_avx512() local 1020 txdp = &txq->tx_ring[tx_id]; in ice_xmit_fixed_burst_vec_avx512() 1030 ice_vtx(txdp, tx_pkts, n - 1, flags); in ice_xmit_fixed_burst_vec_avx512() 1032 txdp += (n - 1); in ice_xmit_fixed_burst_vec_avx512() 1034 ice_vtx1(txdp, *tx_pkts++, rs); in ice_xmit_fixed_burst_vec_avx512() 1042 txdp = txq->tx_ring; in ice_xmit_fixed_burst_vec_avx512() [all …]
|
| H A D | ice_rxtx.c | 2875 for (i = 0; i < 4; i++, txdp++, pkts++) { in tx4() 2877 txdp->buf_addr = rte_cpu_to_le_64(dma_addr); in tx4() 2878 txdp->cmd_type_offset_bsz = in tx4() 2891 txdp->buf_addr = rte_cpu_to_le_64(dma_addr); in tx1() 2892 txdp->cmd_type_offset_bsz = in tx1() 2918 tx4(txdp + i, pkts + i); in ice_tx_fill_hw_ring() 2924 tx1(txdp + mainpart + i, pkts + mainpart + i); in ice_tx_fill_hw_ring() 3968 volatile struct ice_tx_desc *txdp; in ice_fdir_programming() local 3977 txdp = &txq->tx_ring[txq->tx_tail + 1]; in ice_fdir_programming() 3983 txdp->cmd_type_offset_bsz = in ice_fdir_programming() [all …]
|
| /f-stack/dpdk/drivers/net/i40e/ |
| H A D | i40e_rxtx_vec_altivec.c | 529 vtx1(volatile struct i40e_tx_desc *txdp, in vtx1() argument 538 *(vector unsigned long *)txdp = descriptor; in vtx1() 542 vtx(volatile struct i40e_tx_desc *txdp, in vtx() argument 547 for (i = 0; i < nb_pkts; ++i, ++txdp, ++pkt) in vtx() 548 vtx1(txdp, *pkt, flags); in vtx() 556 volatile struct i40e_tx_desc *txdp; in i40e_xmit_fixed_burst_vec() local 575 txdp = &txq->tx_ring[tx_id]; in i40e_xmit_fixed_burst_vec() 585 vtx1(txdp, *tx_pkts, flags); in i40e_xmit_fixed_burst_vec() 587 vtx1(txdp, *tx_pkts++, rs); in i40e_xmit_fixed_burst_vec() 595 txdp = &txq->tx_ring[tx_id]; in i40e_xmit_fixed_burst_vec() [all …]
|
| H A D | i40e_rxtx_vec_neon.c | 510 vtx1(volatile struct i40e_tx_desc *txdp, in vtx1() argument 518 vst1q_u64((uint64_t *)txdp, descriptor); in vtx1() 527 for (i = 0; i < nb_pkts; ++i, ++txdp, ++pkt) in vtx() 528 vtx1(txdp, *pkt, flags); in vtx() 536 volatile struct i40e_tx_desc *txdp; in i40e_xmit_fixed_burst_vec() local 554 txdp = &txq->tx_ring[tx_id]; in i40e_xmit_fixed_burst_vec() 563 for (i = 0; i < n - 1; ++i, ++tx_pkts, ++txdp) in i40e_xmit_fixed_burst_vec() 564 vtx1(txdp, *tx_pkts, flags); in i40e_xmit_fixed_burst_vec() 566 vtx1(txdp, *tx_pkts++, rs); in i40e_xmit_fixed_burst_vec() 574 txdp = &txq->tx_ring[tx_id]; in i40e_xmit_fixed_burst_vec() [all …]
|
| H A D | i40e_rxtx_vec_avx2.c | 830 vtx1(txdp, *pkt, flags); in vtx() 831 nb_pkts--, txdp++, pkt++; in vtx() 857 vtx1(txdp, *pkt, flags); in vtx() 858 txdp++, pkt++, nb_pkts--; in vtx() 867 volatile struct i40e_tx_desc *txdp; in i40e_xmit_fixed_burst_vec_avx2() local 884 txdp = &txq->tx_ring[tx_id]; in i40e_xmit_fixed_burst_vec_avx2() 893 vtx(txdp, tx_pkts, n - 1, flags); in i40e_xmit_fixed_burst_vec_avx2() 895 txdp += (n - 1); in i40e_xmit_fixed_burst_vec_avx2() 897 vtx1(txdp, *tx_pkts++, rs); in i40e_xmit_fixed_burst_vec_avx2() 905 txdp = &txq->tx_ring[tx_id]; in i40e_xmit_fixed_burst_vec_avx2() [all …]
|
| H A D | i40e_rxtx_vec_sse.c | 676 vtx1(volatile struct i40e_tx_desc *txdp, in vtx1() argument 685 _mm_store_si128((__m128i *)txdp, descriptor); in vtx1() 689 vtx(volatile struct i40e_tx_desc *txdp, in vtx() argument 694 for (i = 0; i < nb_pkts; ++i, ++txdp, ++pkt) in vtx() 695 vtx1(txdp, *pkt, flags); in vtx() 703 volatile struct i40e_tx_desc *txdp; in i40e_xmit_fixed_burst_vec() local 721 txdp = &txq->tx_ring[tx_id]; in i40e_xmit_fixed_burst_vec() 731 vtx1(txdp, *tx_pkts, flags); in i40e_xmit_fixed_burst_vec() 733 vtx1(txdp, *tx_pkts++, rs); in i40e_xmit_fixed_burst_vec() 741 txdp = &txq->tx_ring[tx_id]; in i40e_xmit_fixed_burst_vec() [all …]
|
| H A D | i40e_rxtx.c | 1297 tx4(volatile struct i40e_tx_desc *txdp, struct rte_mbuf **pkts) in tx4() argument 1302 for (i = 0; i < 4; i++, txdp++, pkts++) { in tx4() 1304 txdp->buffer_addr = rte_cpu_to_le_64(dma_addr); in tx4() 1305 txdp->cmd_type_offset_bsz = in tx4() 1313 tx1(volatile struct i40e_tx_desc *txdp, struct rte_mbuf **pkts) in tx1() argument 1318 txdp->buffer_addr = rte_cpu_to_le_64(dma_addr); in tx1() 1319 txdp->cmd_type_offset_bsz = in tx1() 1330 volatile struct i40e_tx_desc *txdp = &(txq->tx_ring[txq->tx_tail]); in i40e_tx_fill_hw_ring() local 1343 tx4(txdp + i, pkts + i); in i40e_tx_fill_hw_ring() 1348 tx1(txdp + mainpart + i, pkts + mainpart + i); in i40e_tx_fill_hw_ring()
|
| H A D | i40e_fdir.c | 1831 volatile struct i40e_tx_desc *txdp; in i40e_flow_fdir_filter_programming() local 1910 txdp = &txq->tx_ring[txq->tx_tail + 1]; in i40e_flow_fdir_filter_programming() 1911 txdp->buffer_addr = rte_cpu_to_le_64(pf->fdir.dma_addr[txq->tx_tail >> 1]); in i40e_flow_fdir_filter_programming() 1917 txdp->cmd_type_offset_bsz = in i40e_flow_fdir_filter_programming() 1933 if ((txdp->cmd_type_offset_bsz & in i40e_flow_fdir_filter_programming()
|
| /f-stack/dpdk/drivers/net/ixgbe/ |
| H A D | ixgbe_rxtx_vec_neon.c | 462 vtx1(volatile union ixgbe_adv_tx_desc *txdp, in vtx1() argument 469 vst1q_u64((uint64_t *)&txdp->read, descriptor); in vtx1() 473 vtx(volatile union ixgbe_adv_tx_desc *txdp, in vtx() argument 478 for (i = 0; i < nb_pkts; ++i, ++txdp, ++pkt) in vtx() 479 vtx1(txdp, *pkt, flags); in vtx() 487 volatile union ixgbe_adv_tx_desc *txdp; in ixgbe_xmit_fixed_burst_vec() local 505 txdp = &txq->tx_ring[tx_id]; in ixgbe_xmit_fixed_burst_vec() 515 vtx1(txdp, *tx_pkts, flags); in ixgbe_xmit_fixed_burst_vec() 517 vtx1(txdp, *tx_pkts++, rs); in ixgbe_xmit_fixed_burst_vec() 525 txdp = &txq->tx_ring[tx_id]; in ixgbe_xmit_fixed_burst_vec() [all …]
|
| H A D | ixgbe_rxtx_vec_sse.c | 636 vtx1(volatile union ixgbe_adv_tx_desc *txdp, in vtx1() argument 646 vtx(volatile union ixgbe_adv_tx_desc *txdp, in vtx() argument 651 for (i = 0; i < nb_pkts; ++i, ++txdp, ++pkt) in vtx() 652 vtx1(txdp, *pkt, flags); in vtx() 660 volatile union ixgbe_adv_tx_desc *txdp; in ixgbe_xmit_fixed_burst_vec() local 678 txdp = &txq->tx_ring[tx_id]; in ixgbe_xmit_fixed_burst_vec() 688 for (i = 0; i < n - 1; ++i, ++tx_pkts, ++txdp) in ixgbe_xmit_fixed_burst_vec() 689 vtx1(txdp, *tx_pkts, flags); in ixgbe_xmit_fixed_burst_vec() 691 vtx1(txdp, *tx_pkts++, rs); in ixgbe_xmit_fixed_burst_vec() 699 txdp = &(txq->tx_ring[tx_id]); in ixgbe_xmit_fixed_burst_vec() [all …]
|
| H A D | ixgbe_rxtx.c | 153 tx4(volatile union ixgbe_adv_tx_desc *txdp, struct rte_mbuf **pkts) in tx4() argument 159 for (i = 0; i < 4; ++i, ++txdp, ++pkts) { in tx4() 164 txdp->read.buffer_addr = rte_cpu_to_le_64(buf_dma_addr); in tx4() 166 txdp->read.cmd_type_len = in tx4() 169 txdp->read.olinfo_status = in tx4() 178 tx1(volatile union ixgbe_adv_tx_desc *txdp, struct rte_mbuf **pkts) in tx1() argument 187 txdp->read.buffer_addr = rte_cpu_to_le_64(buf_dma_addr); in tx1() 188 txdp->read.cmd_type_len = in tx1() 190 txdp->read.olinfo_status = in tx1() 221 tx4(txdp + i, pkts + i); in ixgbe_tx_fill_hw_ring() [all …]
|
| /f-stack/dpdk/drivers/net/fm10k/ |
| H A D | fm10k_rxtx_vec.c | 741 vtx1(volatile struct fm10k_tx_desc *txdp, in vtx1() argument 747 _mm_store_si128((__m128i *)txdp, descriptor); in vtx1() 751 vtx(volatile struct fm10k_tx_desc *txdp, in vtx() argument 756 for (i = 0; i < nb_pkts; ++i, ++txdp, ++pkt) in vtx() 757 vtx1(txdp, *pkt, flags); in vtx() 831 volatile struct fm10k_tx_desc *txdp; in fm10k_xmit_fixed_burst_vec() local 849 txdp = &txq->hw_ring[tx_id]; in fm10k_xmit_fixed_burst_vec() 859 vtx1(txdp, *tx_pkts, flags); in fm10k_xmit_fixed_burst_vec() 861 vtx1(txdp, *tx_pkts++, rs); in fm10k_xmit_fixed_burst_vec() 869 txdp = &(txq->hw_ring[tx_id]); in fm10k_xmit_fixed_burst_vec() [all …]
|
| H A D | fm10k_rxtx.c | 464 volatile struct fm10k_tx_desc *txdp; in fm10k_dev_tx_descriptor_status() local 492 txdp = &txq->hw_ring[next_rs]; in fm10k_dev_tx_descriptor_status() 493 if (txdp->flags & FM10K_TXD_FLAG_DONE) in fm10k_dev_tx_descriptor_status()
|
| /f-stack/dpdk/drivers/net/iavf/ |
| H A D | iavf_rxtx_vec_avx2.c | 1407 iavf_vtx1(txdp, *pkt, flags); in iavf_vtx() 1408 nb_pkts--, txdp++, pkt++; in iavf_vtx() 1448 iavf_vtx1(txdp, *pkt, flags); in iavf_vtx() 1449 txdp++, pkt++, nb_pkts--; in iavf_vtx() 1458 volatile struct iavf_tx_desc *txdp; in iavf_xmit_fixed_burst_vec_avx2() local 1476 txdp = &txq->tx_ring[tx_id]; in iavf_xmit_fixed_burst_vec_avx2() 1485 iavf_vtx(txdp, tx_pkts, n - 1, flags); in iavf_xmit_fixed_burst_vec_avx2() 1487 txdp += (n - 1); in iavf_xmit_fixed_burst_vec_avx2() 1489 iavf_vtx1(txdp, *tx_pkts++, rs); in iavf_xmit_fixed_burst_vec_avx2() 1497 txdp = &txq->tx_ring[tx_id]; in iavf_xmit_fixed_burst_vec_avx2() [all …]
|
| H A D | iavf_rxtx_vec_avx512.c | 1543 iavf_vtx1(txdp, *pkt, flags); in iavf_vtx() 1544 nb_pkts--, txdp++, pkt++; in iavf_vtx() 1577 _mm512_storeu_si512((void *)txdp, desc4); in iavf_vtx() 1582 iavf_vtx1(txdp, *pkt, flags); in iavf_vtx() 1583 txdp++, pkt++, nb_pkts--; in iavf_vtx() 1592 volatile struct iavf_tx_desc *txdp; in iavf_xmit_fixed_burst_vec_avx512() local 1610 txdp = &txq->tx_ring[tx_id]; in iavf_xmit_fixed_burst_vec_avx512() 1620 iavf_vtx(txdp, tx_pkts, n - 1, flags); in iavf_xmit_fixed_burst_vec_avx512() 1622 txdp += (n - 1); in iavf_xmit_fixed_burst_vec_avx512() 1624 iavf_vtx1(txdp, *tx_pkts++, rs); in iavf_xmit_fixed_burst_vec_avx512() [all …]
|
| H A D | iavf_rxtx_vec_sse.c | 1095 _mm_store_si128((__m128i *)txdp, descriptor); in vtx1() 1099 iavf_vtx(volatile struct iavf_tx_desc *txdp, struct rte_mbuf **pkt, in iavf_vtx() argument 1104 for (i = 0; i < nb_pkts; ++i, ++txdp, ++pkt) in iavf_vtx() 1105 vtx1(txdp, *pkt, flags); in iavf_vtx() 1113 volatile struct iavf_tx_desc *txdp; in iavf_xmit_fixed_burst_vec() local 1132 txdp = &txq->tx_ring[tx_id]; in iavf_xmit_fixed_burst_vec() 1141 for (i = 0; i < n - 1; ++i, ++tx_pkts, ++txdp) in iavf_xmit_fixed_burst_vec() 1142 vtx1(txdp, *tx_pkts, flags); in iavf_xmit_fixed_burst_vec() 1144 vtx1(txdp, *tx_pkts++, rs); in iavf_xmit_fixed_burst_vec() 1152 txdp = &txq->tx_ring[tx_id]; in iavf_xmit_fixed_burst_vec() [all …]
|
| /f-stack/dpdk/drivers/net/hns3/ |
| H A D | hns3_rxtx_vec_sve.c | 361 struct hns3_desc *txdp = &txq->tx_ring[txq->next_to_use]; in hns3_tx_fill_hw_ring_sve() local 398 svst1_scatter_u64offset_u64(pg, (uint64_t *)&txdp->addr, in hns3_tx_fill_hw_ring_sve() 401 svst1_scatter_u64offset_u64(pg, (uint64_t *)&txdp->tx.vlan_tag, in hns3_tx_fill_hw_ring_sve() 405 (uint64_t *)&txdp->tx.outer_vlan_tag, in hns3_tx_fill_hw_ring_sve() 408 svst1_scatter_u64offset_u64(pg, (uint64_t *)&txdp->tx.paylen, in hns3_tx_fill_hw_ring_sve() 414 txdp += svcntd(); in hns3_tx_fill_hw_ring_sve()
|
| H A D | hns3_rxtx.c | 3465 for (i = 0; i < PER_LOOP_NUM; i++, txdp++, pkts++) { in hns3_tx_setup_4bd() 3467 txdp->addr = rte_cpu_to_le_64(dma_addr); in hns3_tx_setup_4bd() 3469 txdp->tx.paylen = 0; in hns3_tx_setup_4bd() 3470 txdp->tx.type_cs_vlan_tso_len = 0; in hns3_tx_setup_4bd() 3471 txdp->tx.ol_type_vlan_len_msec = 0; in hns3_tx_setup_4bd() 3483 txdp->addr = rte_cpu_to_le_64(dma_addr); in hns3_tx_setup_1bd() 3485 txdp->tx.paylen = 0; in hns3_tx_setup_1bd() 3486 txdp->tx.type_cs_vlan_tso_len = 0; in hns3_tx_setup_1bd() 3487 txdp->tx.ol_type_vlan_len_msec = 0; in hns3_tx_setup_1bd() 3506 hns3_tx_setup_4bd(txdp + i, pkts + i); in hns3_tx_fill_hw_ring() [all …]
|
| /f-stack/freebsd/contrib/dev/ath/ath_hal/ar9300/ |
| H A D | ar9300_xmit.c | 548 ar9300_set_tx_dp(struct ath_hal *ah, u_int q, u_int32_t txdp) in ar9300_set_tx_dp() argument 552 HALASSERT(txdp != 0); in ar9300_set_tx_dp() 554 OS_REG_WRITE(ah, AR_QTXDP(q), txdp); in ar9300_set_tx_dp()
|
| H A D | ar9300_stub_funcs.h | 172 extern HAL_BOOL ar9300_Stub_SetTxDP(struct ath_hal *ah, u_int q, uint32_t txdp);
|
| H A D | ar9300desc.h | 504 extern HAL_BOOL ar9300_set_tx_dp(struct ath_hal *ah, u_int q, u_int32_t txdp);
|
| H A D | ar9300_stub_funcs.c | 986 ar9300_Stub_SetTxDP(struct ath_hal *ah, u_int q, uint32_t txdp) in ar9300_Stub_SetTxDP() argument
|
| /f-stack/dpdk/drivers/net/txgbe/ |
| H A D | txgbe_rxtx.c | 147 tx4(volatile struct txgbe_tx_desc *txdp, struct rte_mbuf **pkts) in tx4() argument 153 for (i = 0; i < 4; ++i, ++txdp, ++pkts) { in tx4() 158 txdp->qw0 = rte_cpu_to_le_64(buf_dma_addr); in tx4() 159 txdp->dw2 = cpu_to_le32(TXGBE_TXD_FLAGS | in tx4() 161 txdp->dw3 = cpu_to_le32(TXGBE_TXD_PAYLEN(pkt_len)); in tx4() 169 tx1(volatile struct txgbe_tx_desc *txdp, struct rte_mbuf **pkts) in tx1() argument 178 txdp->qw0 = cpu_to_le64(buf_dma_addr); in tx1() 179 txdp->dw2 = cpu_to_le32(TXGBE_TXD_FLAGS | in tx1() 181 txdp->dw3 = cpu_to_le32(TXGBE_TXD_PAYLEN(pkt_len)); in tx1() 211 tx4(txdp + i, pkts + i); in txgbe_tx_fill_hw_ring() [all …]
|