| /f-stack/dpdk/lib/librte_ethdev/ |
| H A D | rte_ethdev_trace.h | 25 uint16_t nb_tx_q, const struct rte_eth_conf *dev_conf, int rc), 28 rte_trace_point_emit_u16(nb_tx_q);
|
| H A D | rte_ethdev.c | 1332 if (nb_rx_q == 0 && nb_tx_q == 0) { in rte_eth_dev_configure() 1336 nb_tx_q = dev_info.default_txportconf.nb_queues; in rte_eth_dev_configure() 1337 if (nb_tx_q == 0) in rte_eth_dev_configure() 1338 nb_tx_q = RTE_ETH_DEV_FALLBACK_TX_NBQUEUES; in rte_eth_dev_configure() 1349 if (nb_tx_q > RTE_MAX_QUEUES_PER_PORT) { in rte_eth_dev_configure() 1352 nb_tx_q, RTE_MAX_QUEUES_PER_PORT); in rte_eth_dev_configure() 1369 if (nb_tx_q > dev_info.max_tx_queues) { in rte_eth_dev_configure() 1371 port_id, nb_tx_q, dev_info.max_tx_queues); in rte_eth_dev_configure() 1498 diag = eth_dev_tx_queue_config(dev, nb_tx_q); in rte_eth_dev_configure() 1545 rte_ethdev_trace_configure(port_id, nb_rx_q, nb_tx_q, dev_conf, 0); in rte_eth_dev_configure() [all …]
|
| /f-stack/dpdk/drivers/net/hns3/ |
| H A D | hns3_dcb.c | 653 tx_qnum_per_tc = nb_tx_q / hw->num_tc; in hns3_tc_queue_mapping_cfg() 655 if (used_tx_queues != nb_tx_q) { in hns3_tc_queue_mapping_cfg() 658 nb_tx_q, hw->num_tc); in hns3_tc_queue_mapping_cfg() 692 return hns3_tc_queue_mapping_cfg(hw, nb_tx_q); in hns3_queue_to_tc_mapping() 697 uint16_t nb_tx_q) in hns3_dcb_update_tc_queue_mapping() argument 704 ret = hns3_queue_to_tc_mapping(hw, nb_rx_q, nb_tx_q); in hns3_dcb_update_tc_queue_mapping() 1355 uint16_t nb_tx_q = hw->data->nb_tx_queues; in hns3_dcb_cfg_validate() local 1449 uint16_t nb_tx_q = hw->data->nb_tx_queues; in hns3_dcb_info_update() local 1463 if (nb_tx_q < num_tc) { in hns3_dcb_info_update() 1465 nb_tx_q, num_tc); in hns3_dcb_info_update() [all …]
|
| H A D | hns3_dcb.h | 208 uint16_t nb_tx_q);
|
| H A D | hns3_ethdev_vf.c | 780 uint16_t nb_tx_q = dev->data->nb_tx_queues; in hns3vf_dev_configure() local 786 hw->cfg_max_queues = RTE_MAX(nb_rx_q, nb_tx_q); in hns3vf_dev_configure() 799 ret = hns3_set_fake_rx_or_tx_queues(dev, nb_rx_q, nb_tx_q); in hns3vf_dev_configure() 1416 uint16_t nb_tx_q) in hns3vf_set_tc_queue_mapping() argument 1426 if (nb_tx_q < hw->num_tc) { in hns3vf_set_tc_queue_mapping() 1428 nb_tx_q, hw->num_tc); in hns3vf_set_tc_queue_mapping() 1432 return hns3_queue_to_tc_mapping(hw, nb_rx_q, nb_tx_q); in hns3vf_set_tc_queue_mapping() 2069 uint16_t nb_tx_q = hw->data->nb_tx_queues; in hns3vf_do_start() local 2072 ret = hns3vf_set_tc_queue_mapping(hns, nb_rx_q, nb_tx_q); in hns3vf_do_start()
|
| H A D | hns3_rxtx.c | 223 uint16_t nb_tx_q; in hns3_free_tx_queues() local 226 nb_tx_q = hw->data->nb_tx_queues; in hns3_free_tx_queues() 227 for (i = 0; i < nb_tx_q; i++) { in hns3_free_tx_queues() 335 uint16_t nb_tx_q = hw->data->nb_tx_queues; in hns3_update_all_queues_pvid_proc_en() local 348 if (i < nb_tx_q) { in hns3_update_all_queues_pvid_proc_en() 373 uint16_t nb_tx_q = hw->data->nb_tx_queues; in hns3_enable_all_queues() local 1370 uint16_t nb_tx_q; in hns3_fake_tx_queue_setup() local 1396 nb_tx_q = dev->data->nb_tx_queues; in hns3_fake_tx_queue_setup() 1398 (nb_tx_q + idx) * HNS3_TQP_REG_SIZE); in hns3_fake_tx_queue_setup() 1501 uint16_t nb_tx_q) in hns3_set_fake_rx_or_tx_queues() argument [all …]
|
| H A D | hns3_rxtx.h | 663 uint16_t nb_tx_q);
|
| H A D | hns3_ethdev.c | 2313 uint16_t nb_tx_q = dev->data->nb_tx_queues; in hns3_dev_configure() local 2319 hw->cfg_max_queues = RTE_MAX(nb_rx_q, nb_tx_q); in hns3_dev_configure() 2332 ret = hns3_set_fake_rx_or_tx_queues(dev, nb_rx_q, nb_tx_q); in hns3_dev_configure()
|
| /f-stack/dpdk/lib/librte_pdump/ |
| H A D | rte_pdump.c | 234 uint16_t nb_rx_q = 0, nb_tx_q = 0, end_q, queue; in set_pdump_rxtx_cbs() local 283 nb_tx_q = dev_info.nb_tx_queues; in set_pdump_rxtx_cbs() 289 if (nb_tx_q == 0 && flags & RTE_PDUMP_FLAG_TX) { in set_pdump_rxtx_cbs() 294 if ((nb_tx_q == 0 || nb_rx_q == 0) && in set_pdump_rxtx_cbs() 313 end_q = (queue == RTE_PDUMP_ALL_QUEUES) ? nb_tx_q : queue + 1; in set_pdump_rxtx_cbs()
|
| /f-stack/dpdk/app/test/ |
| H A D | test_link_bonding.c | 72 uint16_t nb_tx_q; member 98 .nb_tx_q = 1, 182 test_params->nb_tx_q, &default_pmd_conf), in configure_ethdev() 191 for (q_id = 0; q_id < test_params->nb_tx_q; q_id++) in configure_ethdev() 4122 test_params->nb_tx_q = 4; in test_reconfigure_bonded_device() 4128 test_params->nb_tx_q = 2; in test_reconfigure_bonded_device()
|
| /f-stack/dpdk/drivers/net/txgbe/ |
| H A D | txgbe_ethdev.c | 1118 uint16_t nb_tx_q = dev->data->nb_tx_queues; in txgbe_check_mq_mode() local 1171 (nb_tx_q > RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool)) { in txgbe_check_mq_mode() 1175 nb_rx_q, nb_tx_q, in txgbe_check_mq_mode() 1206 if (nb_tx_q != TXGBE_VMDQ_DCB_NB_QUEUES) { in txgbe_check_mq_mode()
|
| /f-stack/dpdk/drivers/net/ixgbe/ |
| H A D | ixgbe_ethdev.c | 2220 uint16_t nb_tx_q = dev->data->nb_tx_queues; in ixgbe_check_mq_mode() local 2271 (nb_tx_q > RTE_ETH_DEV_SRIOV(dev).nb_q_per_pool)) { in ixgbe_check_mq_mode() 2275 nb_rx_q, nb_tx_q, in ixgbe_check_mq_mode() 2306 if (nb_tx_q != IXGBE_VMDQ_DCB_NB_QUEUES) { in ixgbe_check_mq_mode() 2355 if (nb_tx_q > IXGBE_NONE_MODE_TX_NB_QUEUES) { in ixgbe_check_mq_mode()
|
| /f-stack/dpdk/doc/guides/nics/ |
| H A D | enic.rst | 34 expected nb_tx_q parameter in the call to
|
| /f-stack/dpdk/drivers/net/e1000/ |
| H A D | igb_ethdev.c | 1085 uint16_t nb_tx_q = dev->data->nb_tx_queues; in igb_check_mq_mode() local 1122 if ((nb_rx_q > 1) || (nb_tx_q > 1)) { in igb_check_mq_mode()
|
| /f-stack/dpdk/doc/guides/rel_notes/ |
| H A D | release_18_05.rst | 420 If both the ``nb_rx_q`` and ``nb_tx_q`` parameters are zero,
|