| /dpdk/drivers/net/txgbe/base/ |
| H A D | txgbe_vf.c | 42 mac->max_tx_queues = 1; in txgbe_init_ops_vf() 198 for (i = 0; i < hw->mac.max_tx_queues; i++) in txgbe_stop_hw_vf() 636 hw->mac.max_tx_queues = msg[TXGBE_VF_TX_QUEUES]; in txgbevf_get_queues() 637 if (hw->mac.max_tx_queues == 0 || in txgbevf_get_queues() 638 hw->mac.max_tx_queues > TXGBE_VF_MAX_TX_QUEUES) in txgbevf_get_queues() 639 hw->mac.max_tx_queues = TXGBE_VF_MAX_TX_QUEUES; in txgbevf_get_queues() 653 if (*default_tc >= hw->mac.max_tx_queues) in txgbevf_get_queues()
|
| /dpdk/drivers/net/hns3/ |
| H A D | hns3_tm.c | 24 return RTE_MIN(dev_info.max_tx_queues, RTE_MAX_QUEUES_PER_PORT); in hns3_tm_max_tx_queues_get() 32 uint32_t max_tx_queues = hns3_tm_max_tx_queues_get(dev); in hns3_tm_conf_init() local 37 pf->tm_conf.nb_leaf_nodes_max = max_tx_queues; in hns3_tm_conf_init() 38 pf->tm_conf.nb_nodes_max = 1 + HNS3_MAX_TC_NUM + max_tx_queues; in hns3_tm_conf_init() 122 uint32_t max_tx_queues = hns3_tm_max_tx_queues_get(dev); in hns3_tm_capabilities_get() local 131 cap->n_nodes_max = 1 + HNS3_MAX_TC_NUM + max_tx_queues; in hns3_tm_capabilities_get() 140 cap->sched_n_children_max = max_tx_queues; in hns3_tm_capabilities_get() 771 cap->nonleaf.sched_n_children_max = max_tx_queues; in hns3_tm_nonleaf_level_capsbilities_get() 785 cap->n_nodes_max = max_tx_queues; in hns3_tm_leaf_level_capabilities_get() 787 cap->n_nodes_leaf_max = max_tx_queues; in hns3_tm_leaf_level_capabilities_get() [all …]
|
| /dpdk/drivers/net/ixgbe/base/ |
| H A D | ixgbe_vf.c | 53 hw->mac.max_tx_queues = 1; in ixgbe_init_ops_vf() 236 for (i = 0; i < hw->mac.max_tx_queues; i++) in ixgbe_stop_adapter_vf() 729 hw->mac.max_tx_queues = msg[IXGBE_VF_TX_QUEUES]; in ixgbevf_get_queues() 730 if (hw->mac.max_tx_queues == 0 || in ixgbevf_get_queues() 731 hw->mac.max_tx_queues > IXGBE_VF_MAX_TX_QUEUES) in ixgbevf_get_queues() 732 hw->mac.max_tx_queues = IXGBE_VF_MAX_TX_QUEUES; in ixgbevf_get_queues() 746 if (*default_tc >= hw->mac.max_tx_queues) in ixgbevf_get_queues()
|
| H A D | ixgbe_82598.c | 130 mac->max_tx_queues = IXGBE_82598_MAX_TX_QUEUES; in ixgbe_init_ops_82598() 237 for (i = 0; ((i < hw->mac.max_tx_queues) && in ixgbe_start_hw_82598() 1340 for (i = 0; ((i < hw->mac.max_tx_queues) && in ixgbe_enable_relaxed_ordering_82598()
|
| /dpdk/drivers/net/octeontx_ep/ |
| H A D | otx_ep_ethdev.c | 37 devinfo->max_tx_queues = otx_epvf->max_tx_queues; in otx_ep_dev_info_get() 145 otx_epvf->max_tx_queues = ethdev_queues; in otx_epdev_init() 166 eth_dev->data->nb_tx_queues > otx_epvf->max_tx_queues) { in otx_ep_dev_configure() 298 if (q_no >= otx_epvf->max_tx_queues) { in otx_ep_tx_queue_setup()
|
| H A D | otx_ep_common.h | 426 uint32_t max_tx_queues; member
|
| /dpdk/drivers/net/ixgbe/ |
| H A D | ixgbe_tm.c | 148 if (tc_nb > hw->mac.max_tx_queues) in ixgbe_tm_capabilities_get() 161 hw->mac.max_tx_queues; in ixgbe_tm_capabilities_get() 181 cap->sched_n_children_max = hw->mac.max_tx_queues; in ixgbe_tm_capabilities_get() 870 cap->n_nodes_max = hw->mac.max_tx_queues; in ixgbe_level_capabilities_get() 872 cap->n_nodes_leaf_max = hw->mac.max_tx_queues; in ixgbe_level_capabilities_get() 894 hw->mac.max_tx_queues; in ixgbe_level_capabilities_get() 973 hw->mac.max_tx_queues; in ixgbe_node_capabilities_get()
|
| H A D | ixgbe_vf_representor.c | 55 dev_info->max_tx_queues = IXGBE_VF_MAX_TX_QUEUES; in ixgbe_vf_representor_dev_infos_get()
|
| /dpdk/drivers/net/txgbe/ |
| H A D | txgbe_tm.c | 147 if (tc_nb > hw->mac.max_tx_queues) in txgbe_tm_capabilities_get() 160 hw->mac.max_tx_queues; in txgbe_tm_capabilities_get() 176 cap->sched_n_children_max = hw->mac.max_tx_queues; in txgbe_tm_capabilities_get() 857 cap->n_nodes_max = hw->mac.max_tx_queues; in txgbe_level_capabilities_get() 859 cap->n_nodes_leaf_max = hw->mac.max_tx_queues; in txgbe_level_capabilities_get() 877 hw->mac.max_tx_queues; in txgbe_level_capabilities_get() 946 hw->mac.max_tx_queues; in txgbe_node_capabilities_get()
|
| /dpdk/drivers/net/ring/ |
| H A D | rte_eth_ring.c | 49 unsigned int max_tx_queues; member 168 dev_info->max_tx_queues = (uint16_t)internals->max_tx_queues; in eth_dev_info() 371 internals->max_tx_queues = nb_tx_queues; in do_eth_dev_ring_create()
|
| /dpdk/drivers/net/enetfec/ |
| H A D | enet_ethdev.h | 101 uint16_t max_tx_queues; member
|
| H A D | enet_ethdev.c | 331 dev_info->max_tx_queues = ENETFEC_MAX_Q; in enetfec_eth_info() 599 fep->max_tx_queues = ENETFEC_MAX_Q; in pmd_enetfec_probe() 613 for (i = 0; i < fep->max_tx_queues; i++) { in pmd_enetfec_probe()
|
| /dpdk/drivers/net/nfp/ |
| H A D | nfp_ethdev_vf.c | 338 hw->max_tx_queues = nn_cfg_readl(hw, NFP_NET_CFG_MAX_TXRINGS); in nfp_netvf_init() 408 hw->max_rx_queues, hw->max_tx_queues); in nfp_netvf_init()
|
| H A D | nfp_common.h | 184 uint32_t max_tx_queues; member
|
| H A D | nfp_ethdev.c | 471 hw->max_tx_queues = nn_cfg_readl(hw, NFP_NET_CFG_MAX_TXRINGS); in nfp_net_init() 541 hw->max_rx_queues, hw->max_tx_queues); in nfp_net_init()
|
| /dpdk/drivers/net/avp/ |
| H A D | rte_avp_common.h | 324 uint8_t max_tx_queues; /**< Maximum supported transmit/alloc queues */ member
|
| H A D | avp_ethdev.c | 172 unsigned int max_tx_queues; /**< Maximum number of transmit queues */ member 849 host_info->min_tx_queues, host_info->max_tx_queues); in avp_dev_create() 870 avp->max_tx_queues = in avp_dev_create() 871 RTE_MIN(host_info->max_tx_queues, RTE_AVP_MAX_QUEUES); in avp_dev_create() 893 for (i = 0; i < avp->max_tx_queues; i++) { in avp_dev_create() 2189 dev_info->max_tx_queues = avp->max_tx_queues; in avp_dev_info_get()
|
| /dpdk/drivers/net/netvsc/ |
| H A D | hn_vf.c | 373 info->max_tx_queues = RTE_MIN(vf_info.max_tx_queues, in hn_vf_info_merge() 374 info->max_tx_queues); in hn_vf_info_merge()
|
| /dpdk/drivers/net/bnx2x/ |
| H A D | bnx2x_ethdev.c | 533 dev_info->max_tx_queues = sc->max_tx_queues; in bnx2x_dev_infos_get() 731 ret = bnx2x_vf_get_resources(sc, sc->max_tx_queues, in bnx2x_common_dev_init()
|
| /dpdk/app/test-pmd/ |
| H A D | testpmd.c | 1300 if (dev_info.max_tx_queues < allowed_max_txq) { in get_allowed_max_nb_txq() 1301 allowed_max_txq = dev_info.max_tx_queues; in get_allowed_max_nb_txq() 1578 for (i = 0; i < port->dev_info.max_tx_queues; i++) in init_config_port_offloads() 1784 if (nb_txq > port->dev_info.max_tx_queues) { in init_fwd_streams() 1787 nb_txq, port->dev_info.max_tx_queues); in init_fwd_streams() 2825 k < port->dev_info.max_tx_queues; in start_port() 4011 nb_txq = rte_port->dev_info.max_tx_queues; in init_port_dcb_config() 4017 nb_txq = rte_port->dev_info.max_tx_queues; in init_port_dcb_config()
|
| /dpdk/drivers/net/failsafe/ |
| H A D | failsafe_ops.c | 1099 info->max_tx_queues = RTE_MIN(info->max_tx_queues, sinfo->max_tx_queues); in fs_dev_merge_info() 1169 infos->max_tx_queues = RTE_MAX_QUEUES_PER_PORT; in fs_dev_infos_get()
|
| /dpdk/drivers/net/liquidio/ |
| H A D | lio_ethdev.c | 402 devinfo->max_tx_queues = lio_dev->max_tx_queues; in lio_dev_info_get() 1741 lio_dev->max_tx_queues = eth_dev->data->nb_tx_queues; in lio_dev_configure() 2021 lio_dev->max_tx_queues = dpdk_queues; in lio_first_time_init()
|
| H A D | lio_struct.h | 653 uint8_t max_tx_queues; member
|
| /dpdk/drivers/net/e1000/ |
| H A D | igb_ethdev.c | 2176 dev_info->max_tx_queues = 4; in eth_igb_infos_get() 2182 dev_info->max_tx_queues = 16; in eth_igb_infos_get() 2189 dev_info->max_tx_queues = 8; in eth_igb_infos_get() 2196 dev_info->max_tx_queues = 8; in eth_igb_infos_get() 2203 dev_info->max_tx_queues = 8; in eth_igb_infos_get() 2208 dev_info->max_tx_queues = 4; in eth_igb_infos_get() 2214 dev_info->max_tx_queues = 2; in eth_igb_infos_get() 2303 dev_info->max_tx_queues = 2; in eth_igbvf_infos_get() 2307 dev_info->max_tx_queues = 1; in eth_igbvf_infos_get() 3197 for (i = 0; i < dev_info.max_tx_queues; i++) in igbvf_stop_adapter()
|
| /dpdk/drivers/net/ice/ |
| H A D | ice_dcf_vf_representor.c | 138 dev_info->max_tx_queues = dcf_hw->vsi_res->num_queue_pairs; in ice_dcf_vf_repr_dev_info_get()
|