Home
last modified time | relevance | path

Searched refs:tx_rs_thresh (Results 1 – 25 of 62) sorted by relevance

123

/dpdk/drivers/net/hns3/
H A Dhns3_rxtx_vec.h23 txq->tx_rs_thresh); in hns3_tx_bulk_free_buffers()
24 for (i = 0; i < txq->tx_rs_thresh; i++) in hns3_tx_bulk_free_buffers()
29 for (i = 0; i < txq->tx_rs_thresh; i++, tx_entry++) { in hns3_tx_bulk_free_buffers()
49 txq->tx_bd_ready += txq->tx_rs_thresh; in hns3_tx_bulk_free_buffers()
50 txq->next_to_clean += txq->tx_rs_thresh; in hns3_tx_bulk_free_buffers()
66 for (i = 0; i < txq->tx_rs_thresh; i++, tx_desc++) { in hns3_tx_free_buffers()
H A Dhns3_rxtx.c2878 rs_thresh = (conf->tx_rs_thresh > 0) ? in hns3_tx_queue_conf_check()
2902 *tx_rs_thresh = rs_thresh; in hns3_tx_queue_conf_check()
2975 uint16_t tx_rs_thresh, tx_free_thresh; in hns3_tx_queue_setup() local
3024 txq->tx_rs_thresh = tx_rs_thresh; in hns3_tx_queue_setup()
3093 for (i = 0; i < txq->tx_rs_thresh; i++) { in hns3_tx_free_useless_buffer()
3105 txq->tx_bd_ready += txq->tx_rs_thresh; in hns3_tx_free_useless_buffer()
3106 txq->next_to_clean += txq->tx_rs_thresh; in hns3_tx_free_useless_buffer()
4011 for (i = 0; i < txq->tx_rs_thresh; i++) in hns3_tx_free_buffer_simple()
4020 txq->tx_bd_ready += txq->tx_rs_thresh; in hns3_tx_free_buffer_simple()
4468 qinfo->conf.tx_rs_thresh = txq->tx_rs_thresh; in hns3_txq_info_get()
[all …]
H A Dhns3_rxtx_vec_sve.c359 for (i = 0; i < txq->tx_rs_thresh; i += HNS3_SVE_CHECK_DESCS_PER_LOOP, in hns3_tx_free_buffers_sve()
493 new_burst = RTE_MIN(nb_pkts, txq->tx_rs_thresh); in hns3_xmit_pkts_vec_sve()
/dpdk/drivers/net/ixgbe/
H A Dixgbe_rxtx_vec_common.h90 n = txq->tx_rs_thresh; in ixgbe_tx_free_bufs()
124 txq->nb_tx_free = (uint16_t)(txq->nb_tx_free + txq->tx_rs_thresh); in ixgbe_tx_free_bufs()
125 txq->tx_next_dd = (uint16_t)(txq->tx_next_dd + txq->tx_rs_thresh); in ixgbe_tx_free_bufs()
127 txq->tx_next_dd = (uint16_t)(txq->tx_rs_thresh - 1); in ixgbe_tx_free_bufs()
129 return txq->tx_rs_thresh; in ixgbe_tx_free_bufs()
153 for (i = txq->tx_next_dd - (txq->tx_rs_thresh - 1); in _ixgbe_tx_queue_release_mbufs_vec()
228 txq->tx_next_dd = (uint16_t)(txq->tx_rs_thresh - 1); in _ixgbe_reset_tx_queue_vec()
229 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in _ixgbe_reset_tx_queue_vec()
H A Dixgbe_rxtx.c147 return txq->tx_rs_thresh; in ixgbe_tx_free_bufs()
298 txq->tx_rs_thresh); in tx_xmit_pkts()
776 txq->tx_rs_thresh, in ixgbe_xmit_pkts()
2672 if (tx_conf->tx_rs_thresh > 0) in ixgbe_dev_tx_queue_setup()
2673 tx_rs_thresh = tx_conf->tx_rs_thresh; in ixgbe_dev_tx_queue_setup()
2678 (unsigned int)tx_rs_thresh, in ixgbe_dev_tx_queue_setup()
2713 (unsigned int)tx_rs_thresh, in ixgbe_dev_tx_queue_setup()
2767 txq->tx_rs_thresh = tx_rs_thresh; in ixgbe_dev_tx_queue_setup()
3326 desc = ((desc + txq->tx_rs_thresh - 1) / txq->tx_rs_thresh) * in ixgbe_dev_tx_descriptor_status()
3327 txq->tx_rs_thresh; in ixgbe_dev_tx_descriptor_status()
[all …]
H A Dixgbe_rxtx_vec_neon.c583 nb_pkts = RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ixgbe_xmit_fixed_burst_vec()
610 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in ixgbe_xmit_fixed_burst_vec()
626 txq->tx_rs_thresh); in ixgbe_xmit_fixed_burst_vec()
H A Dixgbe_rxtx_vec_sse.c705 nb_pkts = RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ixgbe_xmit_fixed_burst_vec()
733 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in ixgbe_xmit_fixed_burst_vec()
749 txq->tx_rs_thresh); in ixgbe_xmit_fixed_burst_vec()
H A Dixgbe_rxtx.h215 uint16_t tx_rs_thresh; member
/dpdk/drivers/net/ice/
H A Dice_rxtx_vec_common.h94 n = txq->tx_rs_thresh; in ice_tx_free_bufs_vec()
128 txq->nb_tx_free = (uint16_t)(txq->nb_tx_free + txq->tx_rs_thresh); in ice_tx_free_bufs_vec()
129 txq->tx_next_dd = (uint16_t)(txq->tx_next_dd + txq->tx_rs_thresh); in ice_tx_free_bufs_vec()
131 txq->tx_next_dd = (uint16_t)(txq->tx_rs_thresh - 1); in ice_tx_free_bufs_vec()
133 return txq->tx_rs_thresh; in ice_tx_free_bufs_vec()
195 i = txq->tx_next_dd - txq->tx_rs_thresh + 1; in _ice_tx_queue_release_mbufs_vec()
306 if (txq->tx_rs_thresh < ICE_VPMD_TX_BURST || in ice_tx_vec_queue_default()
307 txq->tx_rs_thresh > ICE_TX_MAX_FREE_BUF_SZ) in ice_tx_vec_queue_default()
H A Dice_rxtx.c1256 tx_rs_thresh = in ice_tx_queue_setup()
1259 if (tx_conf->tx_rs_thresh) in ice_tx_queue_setup()
1260 tx_rs_thresh = tx_conf->tx_rs_thresh; in ice_tx_queue_setup()
1265 (unsigned int)tx_rs_thresh, in ice_tx_queue_setup()
1276 (unsigned int)tx_rs_thresh, in ice_tx_queue_setup()
1296 (unsigned int)tx_rs_thresh, in ice_tx_queue_setup()
1351 txq->tx_rs_thresh = tx_rs_thresh; in ice_tx_queue_setup()
1448 qinfo->conf.tx_rs_thresh = txq->tx_rs_thresh; in ice_txq_info_get()
2175 desc = ((desc + txq->tx_rs_thresh - 1) / txq->tx_rs_thresh) * in ice_tx_descriptor_status()
2176 txq->tx_rs_thresh; in ice_tx_descriptor_status()
[all …]
H A Dice_rxtx_vec_avx512.c990 n = txq->tx_rs_thresh; in ice_tx_free_bufs_avx512()
1075 txq->nb_tx_free = (uint16_t)(txq->nb_tx_free + txq->tx_rs_thresh); in ice_tx_free_bufs_avx512()
1076 txq->tx_next_dd = (uint16_t)(txq->tx_next_dd + txq->tx_rs_thresh); in ice_tx_free_bufs_avx512()
1078 txq->tx_next_dd = (uint16_t)(txq->tx_rs_thresh - 1); in ice_tx_free_bufs_avx512()
1080 return txq->tx_rs_thresh; in ice_tx_free_bufs_avx512()
1175 nb_pkts = RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ice_xmit_fixed_burst_vec_avx512()
1204 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in ice_xmit_fixed_burst_vec_avx512()
1221 (uint16_t)(txq->tx_next_rs + txq->tx_rs_thresh); in ice_xmit_fixed_burst_vec_avx512()
1241 num = (uint16_t)RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ice_xmit_pkts_vec_avx512()
1263 num = (uint16_t)RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ice_xmit_pkts_vec_avx512_offload()
H A Dice_rxtx_vec_sse.c702 nb_pkts = RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ice_xmit_fixed_burst_vec()
730 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in ice_xmit_fixed_burst_vec()
747 (uint16_t)(txq->tx_next_rs + txq->tx_rs_thresh); in ice_xmit_fixed_burst_vec()
767 num = (uint16_t)RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ice_xmit_pkts_vec()
H A Dice_rxtx_vec_avx2.c904 nb_pkts = RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ice_xmit_fixed_burst_vec_avx2()
932 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in ice_xmit_fixed_burst_vec_avx2()
949 (uint16_t)(txq->tx_next_rs + txq->tx_rs_thresh); in ice_xmit_fixed_burst_vec_avx2()
969 num = (uint16_t)RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ice_xmit_pkts_vec_avx2_common()
/dpdk/drivers/net/i40e/
H A Di40e_rxtx.c1326 uint16_t tx_rs_thresh = txq->tx_rs_thresh; in i40e_tx_free_bufs() local
1339 for (i = 0; i < tx_rs_thresh; i++) in i40e_tx_free_bufs()
1373 return txq->tx_rs_thresh; in i40e_tx_free_bufs()
2203 desc = ((desc + txq->tx_rs_thresh - 1) / txq->tx_rs_thresh) * in i40e_dev_tx_descriptor_status()
2204 txq->tx_rs_thresh; in i40e_dev_tx_descriptor_status()
2329 if (tx_conf->tx_rs_thresh > 0) in i40e_dev_tx_queue_setup()
2330 tx_rs_thresh = tx_conf->tx_rs_thresh; in i40e_dev_tx_queue_setup()
2335 (unsigned int)tx_rs_thresh, in i40e_dev_tx_queue_setup()
2346 (unsigned int)tx_rs_thresh, in i40e_dev_tx_queue_setup()
2419 txq->tx_rs_thresh = tx_rs_thresh; in i40e_dev_tx_queue_setup()
[all …]
H A Di40e_rxtx_vec_common.h96 n = txq->tx_rs_thresh; in i40e_tx_free_bufs()
141 txq->nb_tx_free = (uint16_t)(txq->nb_tx_free + txq->tx_rs_thresh); in i40e_tx_free_bufs()
142 txq->tx_next_dd = (uint16_t)(txq->tx_next_dd + txq->tx_rs_thresh); in i40e_tx_free_bufs()
144 txq->tx_next_dd = (uint16_t)(txq->tx_rs_thresh - 1); in i40e_tx_free_bufs()
146 return txq->tx_rs_thresh; in i40e_tx_free_bufs()
H A Di40e_rxtx_vec_avx512.c895 n = txq->tx_rs_thresh; in i40e_tx_free_bufs_avx512()
981 txq->nb_tx_free = (uint16_t)(txq->nb_tx_free + txq->tx_rs_thresh); in i40e_tx_free_bufs_avx512()
982 txq->tx_next_dd = (uint16_t)(txq->tx_next_dd + txq->tx_rs_thresh); in i40e_tx_free_bufs_avx512()
984 txq->tx_next_dd = (uint16_t)(txq->tx_rs_thresh - 1); in i40e_tx_free_bufs_avx512()
986 return txq->tx_rs_thresh; in i40e_tx_free_bufs_avx512()
1090 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in i40e_xmit_fixed_burst_vec_avx512()
1107 (uint16_t)(txq->tx_next_rs + txq->tx_rs_thresh); in i40e_xmit_fixed_burst_vec_avx512()
1128 num = (uint16_t)RTE_MIN(nb_pkts, txq->tx_rs_thresh); in i40e_xmit_pkts_vec_avx512()
H A Di40e_rxtx_vec_avx2.c795 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in i40e_xmit_fixed_burst_vec_avx2()
812 (uint16_t)(txq->tx_next_rs + txq->tx_rs_thresh); in i40e_xmit_fixed_burst_vec_avx2()
833 num = (uint16_t)RTE_MIN(nb_pkts, txq->tx_rs_thresh); in i40e_xmit_pkts_vec_avx2()
H A Di40e_rxtx_vec_altivec.c589 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in i40e_xmit_fixed_burst_vec()
606 (uint16_t)(txq->tx_next_rs + txq->tx_rs_thresh); in i40e_xmit_fixed_burst_vec()
/dpdk/drivers/net/e1000/
H A Dem_rxtx.c161 uint16_t tx_rs_thresh; member
582 if (txq->nb_tx_used >= txq->tx_rs_thresh) { in eth_em_xmit_pkts()
1208 uint16_t tx_rs_thresh, tx_free_thresh; in eth_em_tx_queue_setup() local
1231 tx_rs_thresh = tx_conf->tx_rs_thresh; in eth_em_tx_queue_setup()
1232 if (tx_rs_thresh == 0) in eth_em_tx_queue_setup()
1244 if (tx_rs_thresh > tx_free_thresh) { in eth_em_tx_queue_setup()
1249 (unsigned int)tx_rs_thresh, in eth_em_tx_queue_setup()
1302 txq->tx_rs_thresh = tx_rs_thresh; in eth_em_tx_queue_setup()
1550 desc = ((desc + txq->tx_rs_thresh - 1) / txq->tx_rs_thresh) * in eth_em_tx_descriptor_status()
1551 txq->tx_rs_thresh; in eth_em_tx_descriptor_status()
[all …]
/dpdk/app/test/
H A Dtest_pmd_perf.c89 .tx_rs_thresh = 32, /* Use PMD default values */
835 tx_conf.tx_rs_thresh = 32; in test_set_rxtx_conf()
840 tx_conf.tx_rs_thresh = 32; in test_set_rxtx_conf()
849 tx_conf.tx_rs_thresh = 32; in test_set_rxtx_conf()
855 tx_conf.tx_rs_thresh = 32; in test_set_rxtx_conf()
/dpdk/drivers/net/iavf/
H A Diavf_rxtx.c131 if (tx_rs_thresh >= (nb_desc - 2)) { in check_tx_thresh()
134 tx_rs_thresh, nb_desc); in check_tx_thresh()
143 if (tx_rs_thresh > tx_free_thresh) { in check_tx_thresh()
146 tx_rs_thresh, tx_free_thresh); in check_tx_thresh()
149 if ((nb_desc % tx_rs_thresh) != 0) { in check_tx_thresh()
152 tx_rs_thresh, nb_desc); in check_tx_thresh()
717 uint16_t tx_rs_thresh, tx_free_thresh; in iavf_dev_tx_queue_setup() local
732 tx_rs_thresh = (uint16_t)((tx_conf->tx_rs_thresh) ? in iavf_dev_tx_queue_setup()
733 tx_conf->tx_rs_thresh : DEFAULT_TX_RS_THRESH); in iavf_dev_tx_queue_setup()
775 txq->rs_thresh = tx_rs_thresh; in iavf_dev_tx_queue_setup()
[all …]
/dpdk/drivers/net/nfp/
H A Dnfp_rxtx.h135 uint32_t tx_rs_thresh; /* not used by now. Future? */ member
/dpdk/app/test-pipeline/
H A Dinit.c102 .tx_rs_thresh = 0,
/dpdk/doc/guides/prog_guide/
H A Dpoll_mode_drv.rst216 …first descriptor used to transmit the packet, exceeds the transmit RS bit threshold (tx_rs_thresh).
219 The default value for tx_rs_thresh is 32.
222 …that the TX Write-back threshold (TX wthresh) should be set to 0 when tx_rs_thresh is greater than…
225 The following constraints must be satisfied for tx_free_thresh and tx_rs_thresh:
227 * tx_rs_thresh must be greater than 0.
229 * tx_rs_thresh must be less than the size of the ring minus 2.
231 * tx_rs_thresh must be less than or equal to tx_free_thresh.
237 * For optimal performance, TX wthresh should be set to 0 when tx_rs_thresh is greater than 1.
251 either perform a bulk release when the ``tx_rs_thresh`` has been crossed
256 independent of whether or not the ``tx_rs_thresh`` has been crossed.
/dpdk/drivers/net/fm10k/
H A Dfm10k_ethdev.c1420 .tx_rs_thresh = FM10K_TX_RS_THRESH_DEFAULT(0), in fm10k_dev_infos_get()
1904 uint16_t tx_rs_thresh; in handle_txconf() local
1930 if (conf->tx_rs_thresh == 0) in handle_txconf()
1931 tx_rs_thresh = FM10K_TX_RS_THRESH_DEFAULT(q); in handle_txconf()
1933 tx_rs_thresh = conf->tx_rs_thresh; in handle_txconf()
1941 tx_rs_thresh)) { in handle_txconf()
1946 tx_rs_thresh, FM10K_TX_RS_THRESH_MAX(q), in handle_txconf()
1952 q->rs_thresh = tx_rs_thresh; in handle_txconf()

123