| /dpdk/drivers/net/hns3/ |
| H A D | hns3_rxtx_vec.h | 23 txq->tx_rs_thresh); in hns3_tx_bulk_free_buffers() 24 for (i = 0; i < txq->tx_rs_thresh; i++) in hns3_tx_bulk_free_buffers() 29 for (i = 0; i < txq->tx_rs_thresh; i++, tx_entry++) { in hns3_tx_bulk_free_buffers() 49 txq->tx_bd_ready += txq->tx_rs_thresh; in hns3_tx_bulk_free_buffers() 50 txq->next_to_clean += txq->tx_rs_thresh; in hns3_tx_bulk_free_buffers() 66 for (i = 0; i < txq->tx_rs_thresh; i++, tx_desc++) { in hns3_tx_free_buffers()
|
| H A D | hns3_rxtx.c | 2878 rs_thresh = (conf->tx_rs_thresh > 0) ? in hns3_tx_queue_conf_check() 2902 *tx_rs_thresh = rs_thresh; in hns3_tx_queue_conf_check() 2975 uint16_t tx_rs_thresh, tx_free_thresh; in hns3_tx_queue_setup() local 3024 txq->tx_rs_thresh = tx_rs_thresh; in hns3_tx_queue_setup() 3093 for (i = 0; i < txq->tx_rs_thresh; i++) { in hns3_tx_free_useless_buffer() 3105 txq->tx_bd_ready += txq->tx_rs_thresh; in hns3_tx_free_useless_buffer() 3106 txq->next_to_clean += txq->tx_rs_thresh; in hns3_tx_free_useless_buffer() 4011 for (i = 0; i < txq->tx_rs_thresh; i++) in hns3_tx_free_buffer_simple() 4020 txq->tx_bd_ready += txq->tx_rs_thresh; in hns3_tx_free_buffer_simple() 4468 qinfo->conf.tx_rs_thresh = txq->tx_rs_thresh; in hns3_txq_info_get() [all …]
|
| H A D | hns3_rxtx_vec_sve.c | 359 for (i = 0; i < txq->tx_rs_thresh; i += HNS3_SVE_CHECK_DESCS_PER_LOOP, in hns3_tx_free_buffers_sve() 493 new_burst = RTE_MIN(nb_pkts, txq->tx_rs_thresh); in hns3_xmit_pkts_vec_sve()
|
| /dpdk/drivers/net/ixgbe/ |
| H A D | ixgbe_rxtx_vec_common.h | 90 n = txq->tx_rs_thresh; in ixgbe_tx_free_bufs() 124 txq->nb_tx_free = (uint16_t)(txq->nb_tx_free + txq->tx_rs_thresh); in ixgbe_tx_free_bufs() 125 txq->tx_next_dd = (uint16_t)(txq->tx_next_dd + txq->tx_rs_thresh); in ixgbe_tx_free_bufs() 127 txq->tx_next_dd = (uint16_t)(txq->tx_rs_thresh - 1); in ixgbe_tx_free_bufs() 129 return txq->tx_rs_thresh; in ixgbe_tx_free_bufs() 153 for (i = txq->tx_next_dd - (txq->tx_rs_thresh - 1); in _ixgbe_tx_queue_release_mbufs_vec() 228 txq->tx_next_dd = (uint16_t)(txq->tx_rs_thresh - 1); in _ixgbe_reset_tx_queue_vec() 229 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in _ixgbe_reset_tx_queue_vec()
|
| H A D | ixgbe_rxtx.c | 147 return txq->tx_rs_thresh; in ixgbe_tx_free_bufs() 298 txq->tx_rs_thresh); in tx_xmit_pkts() 776 txq->tx_rs_thresh, in ixgbe_xmit_pkts() 2672 if (tx_conf->tx_rs_thresh > 0) in ixgbe_dev_tx_queue_setup() 2673 tx_rs_thresh = tx_conf->tx_rs_thresh; in ixgbe_dev_tx_queue_setup() 2678 (unsigned int)tx_rs_thresh, in ixgbe_dev_tx_queue_setup() 2713 (unsigned int)tx_rs_thresh, in ixgbe_dev_tx_queue_setup() 2767 txq->tx_rs_thresh = tx_rs_thresh; in ixgbe_dev_tx_queue_setup() 3326 desc = ((desc + txq->tx_rs_thresh - 1) / txq->tx_rs_thresh) * in ixgbe_dev_tx_descriptor_status() 3327 txq->tx_rs_thresh; in ixgbe_dev_tx_descriptor_status() [all …]
|
| H A D | ixgbe_rxtx_vec_neon.c | 583 nb_pkts = RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ixgbe_xmit_fixed_burst_vec() 610 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in ixgbe_xmit_fixed_burst_vec() 626 txq->tx_rs_thresh); in ixgbe_xmit_fixed_burst_vec()
|
| H A D | ixgbe_rxtx_vec_sse.c | 705 nb_pkts = RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ixgbe_xmit_fixed_burst_vec() 733 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in ixgbe_xmit_fixed_burst_vec() 749 txq->tx_rs_thresh); in ixgbe_xmit_fixed_burst_vec()
|
| H A D | ixgbe_rxtx.h | 215 uint16_t tx_rs_thresh; member
|
| /dpdk/drivers/net/ice/ |
| H A D | ice_rxtx_vec_common.h | 94 n = txq->tx_rs_thresh; in ice_tx_free_bufs_vec() 128 txq->nb_tx_free = (uint16_t)(txq->nb_tx_free + txq->tx_rs_thresh); in ice_tx_free_bufs_vec() 129 txq->tx_next_dd = (uint16_t)(txq->tx_next_dd + txq->tx_rs_thresh); in ice_tx_free_bufs_vec() 131 txq->tx_next_dd = (uint16_t)(txq->tx_rs_thresh - 1); in ice_tx_free_bufs_vec() 133 return txq->tx_rs_thresh; in ice_tx_free_bufs_vec() 195 i = txq->tx_next_dd - txq->tx_rs_thresh + 1; in _ice_tx_queue_release_mbufs_vec() 306 if (txq->tx_rs_thresh < ICE_VPMD_TX_BURST || in ice_tx_vec_queue_default() 307 txq->tx_rs_thresh > ICE_TX_MAX_FREE_BUF_SZ) in ice_tx_vec_queue_default()
|
| H A D | ice_rxtx.c | 1256 tx_rs_thresh = in ice_tx_queue_setup() 1259 if (tx_conf->tx_rs_thresh) in ice_tx_queue_setup() 1260 tx_rs_thresh = tx_conf->tx_rs_thresh; in ice_tx_queue_setup() 1265 (unsigned int)tx_rs_thresh, in ice_tx_queue_setup() 1276 (unsigned int)tx_rs_thresh, in ice_tx_queue_setup() 1296 (unsigned int)tx_rs_thresh, in ice_tx_queue_setup() 1351 txq->tx_rs_thresh = tx_rs_thresh; in ice_tx_queue_setup() 1448 qinfo->conf.tx_rs_thresh = txq->tx_rs_thresh; in ice_txq_info_get() 2175 desc = ((desc + txq->tx_rs_thresh - 1) / txq->tx_rs_thresh) * in ice_tx_descriptor_status() 2176 txq->tx_rs_thresh; in ice_tx_descriptor_status() [all …]
|
| H A D | ice_rxtx_vec_avx512.c | 990 n = txq->tx_rs_thresh; in ice_tx_free_bufs_avx512() 1075 txq->nb_tx_free = (uint16_t)(txq->nb_tx_free + txq->tx_rs_thresh); in ice_tx_free_bufs_avx512() 1076 txq->tx_next_dd = (uint16_t)(txq->tx_next_dd + txq->tx_rs_thresh); in ice_tx_free_bufs_avx512() 1078 txq->tx_next_dd = (uint16_t)(txq->tx_rs_thresh - 1); in ice_tx_free_bufs_avx512() 1080 return txq->tx_rs_thresh; in ice_tx_free_bufs_avx512() 1175 nb_pkts = RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ice_xmit_fixed_burst_vec_avx512() 1204 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in ice_xmit_fixed_burst_vec_avx512() 1221 (uint16_t)(txq->tx_next_rs + txq->tx_rs_thresh); in ice_xmit_fixed_burst_vec_avx512() 1241 num = (uint16_t)RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ice_xmit_pkts_vec_avx512() 1263 num = (uint16_t)RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ice_xmit_pkts_vec_avx512_offload()
|
| H A D | ice_rxtx_vec_sse.c | 702 nb_pkts = RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ice_xmit_fixed_burst_vec() 730 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in ice_xmit_fixed_burst_vec() 747 (uint16_t)(txq->tx_next_rs + txq->tx_rs_thresh); in ice_xmit_fixed_burst_vec() 767 num = (uint16_t)RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ice_xmit_pkts_vec()
|
| H A D | ice_rxtx_vec_avx2.c | 904 nb_pkts = RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ice_xmit_fixed_burst_vec_avx2() 932 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in ice_xmit_fixed_burst_vec_avx2() 949 (uint16_t)(txq->tx_next_rs + txq->tx_rs_thresh); in ice_xmit_fixed_burst_vec_avx2() 969 num = (uint16_t)RTE_MIN(nb_pkts, txq->tx_rs_thresh); in ice_xmit_pkts_vec_avx2_common()
|
| /dpdk/drivers/net/i40e/ |
| H A D | i40e_rxtx.c | 1326 uint16_t tx_rs_thresh = txq->tx_rs_thresh; in i40e_tx_free_bufs() local 1339 for (i = 0; i < tx_rs_thresh; i++) in i40e_tx_free_bufs() 1373 return txq->tx_rs_thresh; in i40e_tx_free_bufs() 2203 desc = ((desc + txq->tx_rs_thresh - 1) / txq->tx_rs_thresh) * in i40e_dev_tx_descriptor_status() 2204 txq->tx_rs_thresh; in i40e_dev_tx_descriptor_status() 2329 if (tx_conf->tx_rs_thresh > 0) in i40e_dev_tx_queue_setup() 2330 tx_rs_thresh = tx_conf->tx_rs_thresh; in i40e_dev_tx_queue_setup() 2335 (unsigned int)tx_rs_thresh, in i40e_dev_tx_queue_setup() 2346 (unsigned int)tx_rs_thresh, in i40e_dev_tx_queue_setup() 2419 txq->tx_rs_thresh = tx_rs_thresh; in i40e_dev_tx_queue_setup() [all …]
|
| H A D | i40e_rxtx_vec_common.h | 96 n = txq->tx_rs_thresh; in i40e_tx_free_bufs() 141 txq->nb_tx_free = (uint16_t)(txq->nb_tx_free + txq->tx_rs_thresh); in i40e_tx_free_bufs() 142 txq->tx_next_dd = (uint16_t)(txq->tx_next_dd + txq->tx_rs_thresh); in i40e_tx_free_bufs() 144 txq->tx_next_dd = (uint16_t)(txq->tx_rs_thresh - 1); in i40e_tx_free_bufs() 146 return txq->tx_rs_thresh; in i40e_tx_free_bufs()
|
| H A D | i40e_rxtx_vec_avx512.c | 895 n = txq->tx_rs_thresh; in i40e_tx_free_bufs_avx512() 981 txq->nb_tx_free = (uint16_t)(txq->nb_tx_free + txq->tx_rs_thresh); in i40e_tx_free_bufs_avx512() 982 txq->tx_next_dd = (uint16_t)(txq->tx_next_dd + txq->tx_rs_thresh); in i40e_tx_free_bufs_avx512() 984 txq->tx_next_dd = (uint16_t)(txq->tx_rs_thresh - 1); in i40e_tx_free_bufs_avx512() 986 return txq->tx_rs_thresh; in i40e_tx_free_bufs_avx512() 1090 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in i40e_xmit_fixed_burst_vec_avx512() 1107 (uint16_t)(txq->tx_next_rs + txq->tx_rs_thresh); in i40e_xmit_fixed_burst_vec_avx512() 1128 num = (uint16_t)RTE_MIN(nb_pkts, txq->tx_rs_thresh); in i40e_xmit_pkts_vec_avx512()
|
| H A D | i40e_rxtx_vec_avx2.c | 795 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in i40e_xmit_fixed_burst_vec_avx2() 812 (uint16_t)(txq->tx_next_rs + txq->tx_rs_thresh); in i40e_xmit_fixed_burst_vec_avx2() 833 num = (uint16_t)RTE_MIN(nb_pkts, txq->tx_rs_thresh); in i40e_xmit_pkts_vec_avx2()
|
| H A D | i40e_rxtx_vec_altivec.c | 589 txq->tx_next_rs = (uint16_t)(txq->tx_rs_thresh - 1); in i40e_xmit_fixed_burst_vec() 606 (uint16_t)(txq->tx_next_rs + txq->tx_rs_thresh); in i40e_xmit_fixed_burst_vec()
|
| /dpdk/drivers/net/e1000/ |
| H A D | em_rxtx.c | 161 uint16_t tx_rs_thresh; member 582 if (txq->nb_tx_used >= txq->tx_rs_thresh) { in eth_em_xmit_pkts() 1208 uint16_t tx_rs_thresh, tx_free_thresh; in eth_em_tx_queue_setup() local 1231 tx_rs_thresh = tx_conf->tx_rs_thresh; in eth_em_tx_queue_setup() 1232 if (tx_rs_thresh == 0) in eth_em_tx_queue_setup() 1244 if (tx_rs_thresh > tx_free_thresh) { in eth_em_tx_queue_setup() 1249 (unsigned int)tx_rs_thresh, in eth_em_tx_queue_setup() 1302 txq->tx_rs_thresh = tx_rs_thresh; in eth_em_tx_queue_setup() 1550 desc = ((desc + txq->tx_rs_thresh - 1) / txq->tx_rs_thresh) * in eth_em_tx_descriptor_status() 1551 txq->tx_rs_thresh; in eth_em_tx_descriptor_status() [all …]
|
| /dpdk/app/test/ |
| H A D | test_pmd_perf.c | 89 .tx_rs_thresh = 32, /* Use PMD default values */ 835 tx_conf.tx_rs_thresh = 32; in test_set_rxtx_conf() 840 tx_conf.tx_rs_thresh = 32; in test_set_rxtx_conf() 849 tx_conf.tx_rs_thresh = 32; in test_set_rxtx_conf() 855 tx_conf.tx_rs_thresh = 32; in test_set_rxtx_conf()
|
| /dpdk/drivers/net/iavf/ |
| H A D | iavf_rxtx.c | 131 if (tx_rs_thresh >= (nb_desc - 2)) { in check_tx_thresh() 134 tx_rs_thresh, nb_desc); in check_tx_thresh() 143 if (tx_rs_thresh > tx_free_thresh) { in check_tx_thresh() 146 tx_rs_thresh, tx_free_thresh); in check_tx_thresh() 149 if ((nb_desc % tx_rs_thresh) != 0) { in check_tx_thresh() 152 tx_rs_thresh, nb_desc); in check_tx_thresh() 717 uint16_t tx_rs_thresh, tx_free_thresh; in iavf_dev_tx_queue_setup() local 732 tx_rs_thresh = (uint16_t)((tx_conf->tx_rs_thresh) ? in iavf_dev_tx_queue_setup() 733 tx_conf->tx_rs_thresh : DEFAULT_TX_RS_THRESH); in iavf_dev_tx_queue_setup() 775 txq->rs_thresh = tx_rs_thresh; in iavf_dev_tx_queue_setup() [all …]
|
| /dpdk/drivers/net/nfp/ |
| H A D | nfp_rxtx.h | 135 uint32_t tx_rs_thresh; /* not used by now. Future? */ member
|
| /dpdk/app/test-pipeline/ |
| H A D | init.c | 102 .tx_rs_thresh = 0,
|
| /dpdk/doc/guides/prog_guide/ |
| H A D | poll_mode_drv.rst | 216 …first descriptor used to transmit the packet, exceeds the transmit RS bit threshold (tx_rs_thresh). 219 The default value for tx_rs_thresh is 32. 222 …that the TX Write-back threshold (TX wthresh) should be set to 0 when tx_rs_thresh is greater than… 225 The following constraints must be satisfied for tx_free_thresh and tx_rs_thresh: 227 * tx_rs_thresh must be greater than 0. 229 * tx_rs_thresh must be less than the size of the ring minus 2. 231 * tx_rs_thresh must be less than or equal to tx_free_thresh. 237 * For optimal performance, TX wthresh should be set to 0 when tx_rs_thresh is greater than 1. 251 either perform a bulk release when the ``tx_rs_thresh`` has been crossed 256 independent of whether or not the ``tx_rs_thresh`` has been crossed.
|
| /dpdk/drivers/net/fm10k/ |
| H A D | fm10k_ethdev.c | 1420 .tx_rs_thresh = FM10K_TX_RS_THRESH_DEFAULT(0), in fm10k_dev_infos_get() 1904 uint16_t tx_rs_thresh; in handle_txconf() local 1930 if (conf->tx_rs_thresh == 0) in handle_txconf() 1931 tx_rs_thresh = FM10K_TX_RS_THRESH_DEFAULT(q); in handle_txconf() 1933 tx_rs_thresh = conf->tx_rs_thresh; in handle_txconf() 1941 tx_rs_thresh)) { in handle_txconf() 1946 tx_rs_thresh, FM10K_TX_RS_THRESH_MAX(q), in handle_txconf() 1952 q->rs_thresh = tx_rs_thresh; in handle_txconf()
|