| /f-stack/dpdk/drivers/net/ring/ |
| H A D | rte_eth_ring.c | 31 const unsigned int nb_tx_queues; member 188 i < dev->data->nb_tx_queues; i++) { in eth_stats_get() 207 for (i = 0; i < dev->data->nb_tx_queues; i++) in eth_stats_reset() 290 const unsigned int nb_tx_queues, in do_eth_dev_ring_create() argument 347 internals->max_tx_queues = nb_tx_queues; in do_eth_dev_ring_create() 352 for (i = 0; i < nb_tx_queues; i++) { in do_eth_dev_ring_create() 359 data->nb_tx_queues = (uint16_t)nb_tx_queues; in do_eth_dev_ring_create() 390 const unsigned int nb_tx_queues, in rte_eth_from_rings() argument 397 .nb_tx_queues = nb_tx_queues, in rte_eth_from_rings() 411 if (tx_queues == NULL && nb_tx_queues > 0) { in rte_eth_from_rings() [all …]
|
| H A D | rte_eth_ring.h | 36 const unsigned nb_tx_queues,
|
| /f-stack/dpdk/drivers/net/sfc/ |
| H A D | sfc_tx.c | 318 while (--sw_index >= (int)nb_tx_queues) { in sfc_tx_fini_queues() 323 sas->txq_count = nb_tx_queues; in sfc_tx_fini_queues() 332 const unsigned int nb_tx_queues = sa->eth_dev->data->nb_tx_queues; in sfc_tx_configure() local 336 nb_tx_queues, sas->txq_count); in sfc_tx_configure() 352 if (nb_tx_queues == sas->txq_count) in sfc_tx_configure() 374 if (nb_tx_queues < sas->txq_count) in sfc_tx_configure() 375 sfc_tx_fini_queues(sa, nb_tx_queues); in sfc_tx_configure() 390 if (nb_tx_queues > sas->txq_count) { in sfc_tx_configure() 392 (nb_tx_queues - sas->txq_count) * in sfc_tx_configure() 395 (nb_tx_queues - sas->txq_count) * in sfc_tx_configure() [all …]
|
| /f-stack/dpdk/drivers/net/mvpp2/ |
| H A D | mrvl_tm.c | 199 cap->sched_n_children_max = dev->data->nb_tx_queues; in mrvl_capabilities_get() 200 cap->sched_sp_n_priorities_max = dev->data->nb_tx_queues; in mrvl_capabilities_get() 201 cap->sched_wfq_n_children_per_group_max = dev->data->nb_tx_queues; in mrvl_capabilities_get() 254 cap->nonleaf.sched_n_children_max = dev->data->nb_tx_queues; in mrvl_level_capabilities_get() 257 dev->data->nb_tx_queues; in mrvl_level_capabilities_get() 265 cap->n_nodes_max = dev->data->nb_tx_queues; in mrvl_level_capabilities_get() 266 cap->n_nodes_leaf_max = dev->data->nb_tx_queues; in mrvl_level_capabilities_get() 317 cap->nonleaf.sched_n_children_max = dev->data->nb_tx_queues; in mrvl_node_capabilities_get() 320 dev->data->nb_tx_queues; in mrvl_node_capabilities_get() 507 if (node_id >= dev->data->nb_tx_queues) { in mrvl_node_check_params() [all …]
|
| /f-stack/dpdk/drivers/net/nfb/ |
| H A D | nfb_ethdev.c | 121 uint16_t nb_tx = dev->data->nb_tx_queues; in nfb_eth_dev_start() 159 uint16_t nb_tx = dev->data->nb_tx_queues; in nfb_eth_dev_stop() 202 dev_info->max_tx_queues = dev->data->nb_tx_queues; in nfb_eth_dev_info() 222 uint16_t nb_tx = dev->data->nb_tx_queues; in nfb_eth_dev_close() 242 dev->data->nb_tx_queues = 0; in nfb_eth_dev_close() 477 data->nb_tx_queues = ndp_get_tx_queue_available_count(internals->nfb); in nfb_eth_dev_init() 480 data->nb_rx_queues, data->nb_tx_queues); in nfb_eth_dev_init()
|
| H A D | nfb_stats.c | 16 uint16_t nb_tx = dev->data->nb_tx_queues; in nfb_eth_stats_get() 60 uint16_t nb_tx = dev->data->nb_tx_queues; in nfb_eth_stats_reset()
|
| /f-stack/dpdk/drivers/net/thunderx/ |
| H A D | nicvf_ethdev.h | 116 *tx_end = dev->data->nb_tx_queues ? in nicvf_tx_range() 117 RTE_MIN(tmp, dev->data->nb_tx_queues - 1) : 0; in nicvf_tx_range()
|
| /f-stack/dpdk/drivers/net/pcap/ |
| H A D | rte_eth_pcap.c | 567 for (i = 0; i < dev->data->nb_tx_queues; i++) { in eth_dev_start() 602 for (i = 0; i < dev->data->nb_tx_queues; i++) in eth_dev_start() 630 for (i = 0; i < dev->data->nb_tx_queues; i++) { in eth_dev_stop() 653 for (i = 0; i < dev->data->nb_tx_queues; i++) in eth_dev_stop() 677 dev_info->max_tx_queues = dev->data->nb_tx_queues; in eth_dev_info() 701 i < dev->data->nb_tx_queues; i++) { in eth_stats_get() 729 for (i = 0; i < dev->data->nb_tx_queues; i++) { in eth_stats_reset() 1115 const unsigned int nb_tx_queues, in pmd_init_internals() argument 1161 data->nb_tx_queues = (uint16_t)nb_tx_queues; in pmd_init_internals() 1273 const unsigned int nb_tx_queues = tx_queues->num_of_queue; in eth_from_pcaps_common() local [all …]
|
| /f-stack/dpdk/drivers/net/hns3/ |
| H A D | hns3_stats.c | 530 num = RTE_MIN(RTE_ETHDEV_QUEUE_STAT_CNTRS, eth_dev->data->nb_tx_queues); in hns3_stats_get() 598 for (i = 0; i != eth_dev->data->nb_tx_queues; ++i) { in hns3_stats_reset() 640 int tx_err_stats = dev->data->nb_tx_queues * HNS3_NUM_TX_ERRORS_XSTATS; in hns3_xstats_calc_num() 642 int tx_queue_stats = dev->data->nb_tx_queues * HNS3_NUM_TX_QUEUE_STATS; in hns3_xstats_calc_num() 673 for (j = 0; j < dev->data->nb_tx_queues; j++) { in hns3_get_queue_stats() 784 for (j = 0; j < dev->data->nb_tx_queues; j++) { in hns3_dev_xstats_get() 869 for (j = 0; j < dev->data->nb_tx_queues; j++) { in hns3_dev_xstats_get_names() 888 for (j = 0; j < dev->data->nb_tx_queues; j++) { in hns3_dev_xstats_get_names()
|
| /f-stack/dpdk/app/test-pmd/ |
| H A D | util.c | 287 for (queue = 0; queue < dev_info.nb_tx_queues; queue++) in add_tx_md_callback() 308 for (queue = 0; queue < dev_info.nb_tx_queues; queue++) in remove_tx_md_callback() 343 for (queue = 0; queue < dev_info.nb_tx_queues; queue++) in add_tx_dynf_callback() 364 for (queue = 0; queue < dev_info.nb_tx_queues; queue++) in remove_tx_dynf_callback()
|
| /f-stack/dpdk/drivers/net/mlx4/ |
| H A D | mlx4_txq.c | 134 const unsigned int txqs_n = dev->data->nb_tx_queues; in mlx4_tx_uar_init_secondary() 319 if (idx >= dev->data->nb_tx_queues) { in mlx4_tx_queue_setup() 322 (void *)dev, idx, dev->data->nb_tx_queues); in mlx4_tx_queue_setup() 512 for (i = 0; i != ETH_DEV(priv)->data->nb_tx_queues; ++i) in mlx4_tx_queue_release()
|
| /f-stack/dpdk/drivers/net/null/ |
| H A D | rte_eth_null.c | 257 if (tx_queue_id >= dev->data->nb_tx_queues) in eth_tx_queue_setup() 323 RTE_MIN(dev->data->nb_tx_queues, in eth_stats_get() 503 const unsigned int nb_tx_queues = 1; in eth_dev_null_create() local 548 data->nb_tx_queues = (uint16_t)nb_tx_queues; in eth_dev_null_create()
|
| /f-stack/dpdk/drivers/net/octeontx2/ |
| H A D | otx2_link.c | 168 for (i = 0; i < eth_dev->data->nb_tx_queues; i++) in otx2_nix_dev_set_link_up() 185 for (i = 0; i < eth_dev->data->nb_tx_queues; i++) in otx2_nix_dev_set_link_down()
|
| /f-stack/dpdk/drivers/net/softnic/ |
| H A D | rte_eth_softnic_link.c | 95 link->n_txq = port_info.nb_tx_queues; in softnic_link_create()
|
| /f-stack/dpdk/drivers/net/kni/ |
| H A D | rte_eth_kni.c | 316 data->nb_tx_queues); in eth_kni_stats_get() 345 for (i = 0; i < data->nb_tx_queues; i++) { in eth_kni_stats_reset() 390 data->nb_tx_queues = 1; in eth_kni_create()
|
| /f-stack/dpdk/drivers/net/vmxnet3/ |
| H A D | vmxnet3_ethdev.c | 421 if (dev->data->nb_tx_queues > VMXNET3_MAX_TX_QUEUES || in vmxnet3_dev_configure() 433 dev->data->nb_tx_queues * sizeof(struct Vmxnet3_RxQueueDesc); in vmxnet3_dev_configure() 439 hw->num_tx_queues = (uint8_t)dev->data->nb_tx_queues; in vmxnet3_dev_configure() 891 for (i = 0; i < dev->data->nb_tx_queues; i++) { in vmxnet3_free_queues() 896 dev->data->nb_tx_queues = 0; in vmxnet3_free_queues() 1020 dev->data->nb_tx_queues * RTE_DIM(vmxnet3_txq_stat_strings) + in vmxnet3_dev_xstats_get_names() 1039 for (i = 0; i < dev->data->nb_tx_queues; i++) { in vmxnet3_dev_xstats_get_names() 1061 dev->data->nb_tx_queues * RTE_DIM(vmxnet3_txq_stat_strings) + in vmxnet3_dev_xstats_get() 1081 for (i = 0; i < dev->data->nb_tx_queues; i++) { in vmxnet3_dev_xstats_get()
|
| /f-stack/dpdk/drivers/net/vhost/ |
| H A D | rte_eth_vhost.c | 242 for (i = 0; i < dev->data->nb_tx_queues; i++) { in vhost_dev_xstats_reset() 306 for (i = 0; i < dev->data->nb_tx_queues; i++) { in vhost_dev_xstats_get() 742 for (i = 0; i < dev->data->nb_tx_queues; i++) { in update_queuing_status() 766 for (i = 0; i < eth_dev->data->nb_tx_queues; i++) { in queue_setup() 870 for (i = 0; i < eth_dev->data->nb_tx_queues; i++) { in destroy_device() 1198 for (i = 0; i < dev->data->nb_tx_queues; i++) in eth_dev_close() 1304 i < dev->data->nb_tx_queues; i++) { in eth_stats_get() 1336 for (i = 0; i < dev->data->nb_tx_queues; i++) { in eth_stats_reset() 1444 data->nb_tx_queues = queues; in eth_dev_vhost_create()
|
| /f-stack/dpdk/drivers/net/tap/ |
| H A D | rte_eth_tap.c | 915 for (i = 0; i < dev->data->nb_tx_queues; i++) in tap_dev_start() 930 for (i = 0; i < dev->data->nb_tx_queues; i++) in tap_dev_stop() 954 if (dev->data->nb_tx_queues > RTE_PMD_TAP_MAX_QUEUES) { in tap_dev_configure() 958 dev->data->nb_tx_queues, in tap_dev_configure() 964 dev->device->name, pmd->name, dev->data->nb_tx_queues); in tap_dev_configure() 1057 dev->data->nb_tx_queues : RTE_ETHDEV_QUEUE_STAT_CNTRS; in tap_stats_get() 1618 if (tx_queue_id >= dev->data->nb_tx_queues) in tap_tx_queue_setup() 1955 data->nb_tx_queues = 0; in eth_dev_tap_create() 2342 dev->data->nb_tx_queues = reply_param->txq_count; in tap_mp_attach_queues() 2381 if (dev->data->nb_rx_queues + dev->data->nb_tx_queues > in tap_mp_sync_queues() [all …]
|
| /f-stack/dpdk/drivers/net/netvsc/ |
| H A D | hn_ethdev.c | 587 dev->data->nb_tx_queues); in hn_dev_configure() 630 for (i = 0; i < dev->data->nb_tx_queues; i++) { in hn_dev_stats_get() 674 for (i = 0; i < dev->data->nb_tx_queues; i++) { in hn_dev_stats_reset() 711 count = dev->data->nb_tx_queues * RTE_DIM(hn_stat_strings); in hn_dev_xstats_count() 733 for (i = 0; i < dev->data->nb_tx_queues; i++) { in hn_dev_xstats_get_names() 787 for (i = 0; i < dev->data->nb_tx_queues; i++) { in hn_dev_xstats_get()
|
| /f-stack/dpdk/drivers/net/mlx5/ |
| H A D | mlx5_ethdev.c | 65 unsigned int txqs_n = dev->data->nb_tx_queues; in mlx5_dev_configure() 228 dev->data->nb_tx_queues > 2) { in mlx5_set_default_params() 237 dev->data->nb_tx_queues > 2) { in mlx5_set_default_params()
|
| /f-stack/dpdk/drivers/net/liquidio/ |
| H A D | lio_ethdev.c | 303 for (i = 0; i < eth_dev->data->nb_tx_queues; i++) { in lio_dev_stats_get() 351 for (i = 0; i < eth_dev->data->nb_tx_queues; i++) { in lio_dev_stats_reset() 1256 if (q_no >= lio_dev->nb_tx_queues) { in lio_dev_tx_queue_setup() 1713 lio_dev->nb_tx_queues != num_txq) { in lio_reconf_queues() 1717 lio_dev->nb_tx_queues = num_txq; in lio_reconf_queues() 1758 eth_dev->data->nb_tx_queues, in lio_dev_configure() 1762 lio_dev->nb_tx_queues = eth_dev->data->nb_tx_queues; in lio_dev_configure() 1766 lio_dev->max_tx_queues = eth_dev->data->nb_tx_queues; in lio_dev_configure() 1779 if_cfg.s.num_iqueues = lio_dev->nb_tx_queues; in lio_dev_configure()
|
| /f-stack/dpdk/lib/librte_ethdev/ |
| H A D | rte_ethdev_core.h | 138 uint16_t nb_tx_queues; /**< Number of TX queues. */ member
|
| /f-stack/dpdk/drivers/net/octeontx/ |
| H A D | octeontx_ethdev_ops.c | 187 for (i = 0; i < eth_dev->data->nb_tx_queues; i++) in octeontx_dev_set_link_up() 201 for (i = 0; i < eth_dev->data->nb_tx_queues; i++) in octeontx_dev_set_link_down()
|
| /f-stack/dpdk/lib/librte_latencystats/ |
| H A D | rte_latencystats.c | 286 for (qid = 0; qid < dev_info.nb_tx_queues; qid++) { in rte_latencystats_init() 329 for (qid = 0; qid < dev_info.nb_tx_queues; qid++) { in rte_latencystats_uninit()
|
| /f-stack/dpdk/drivers/net/ice/ |
| H A D | ice_dcf_ethdev.c | 407 if (tx_queue_id >= dev->data->nb_tx_queues) in ice_dcf_tx_queue_start() 439 if (tx_queue_id >= dev->data->nb_tx_queues) in ice_dcf_tx_queue_stop() 465 for (nb_txq = 0; nb_txq < dev->data->nb_tx_queues; nb_txq++) { in ice_dcf_start_queues() 510 dev->data->nb_tx_queues); in ice_dcf_dev_start() 574 for (i = 0; i < dev->data->nb_tx_queues; i++) { in ice_dcf_stop_queues()
|