Home
last modified time | relevance | path

Searched refs:nb_tx_queues (Results 1 – 25 of 129) sorted by relevance

123456

/f-stack/dpdk/drivers/net/ring/
H A Drte_eth_ring.c31 const unsigned int nb_tx_queues; member
188 i < dev->data->nb_tx_queues; i++) { in eth_stats_get()
207 for (i = 0; i < dev->data->nb_tx_queues; i++) in eth_stats_reset()
290 const unsigned int nb_tx_queues, in do_eth_dev_ring_create() argument
347 internals->max_tx_queues = nb_tx_queues; in do_eth_dev_ring_create()
352 for (i = 0; i < nb_tx_queues; i++) { in do_eth_dev_ring_create()
359 data->nb_tx_queues = (uint16_t)nb_tx_queues; in do_eth_dev_ring_create()
390 const unsigned int nb_tx_queues, in rte_eth_from_rings() argument
397 .nb_tx_queues = nb_tx_queues, in rte_eth_from_rings()
411 if (tx_queues == NULL && nb_tx_queues > 0) { in rte_eth_from_rings()
[all …]
H A Drte_eth_ring.h36 const unsigned nb_tx_queues,
/f-stack/dpdk/drivers/net/sfc/
H A Dsfc_tx.c318 while (--sw_index >= (int)nb_tx_queues) { in sfc_tx_fini_queues()
323 sas->txq_count = nb_tx_queues; in sfc_tx_fini_queues()
332 const unsigned int nb_tx_queues = sa->eth_dev->data->nb_tx_queues; in sfc_tx_configure() local
336 nb_tx_queues, sas->txq_count); in sfc_tx_configure()
352 if (nb_tx_queues == sas->txq_count) in sfc_tx_configure()
374 if (nb_tx_queues < sas->txq_count) in sfc_tx_configure()
375 sfc_tx_fini_queues(sa, nb_tx_queues); in sfc_tx_configure()
390 if (nb_tx_queues > sas->txq_count) { in sfc_tx_configure()
392 (nb_tx_queues - sas->txq_count) * in sfc_tx_configure()
395 (nb_tx_queues - sas->txq_count) * in sfc_tx_configure()
[all …]
/f-stack/dpdk/drivers/net/mvpp2/
H A Dmrvl_tm.c199 cap->sched_n_children_max = dev->data->nb_tx_queues; in mrvl_capabilities_get()
200 cap->sched_sp_n_priorities_max = dev->data->nb_tx_queues; in mrvl_capabilities_get()
201 cap->sched_wfq_n_children_per_group_max = dev->data->nb_tx_queues; in mrvl_capabilities_get()
254 cap->nonleaf.sched_n_children_max = dev->data->nb_tx_queues; in mrvl_level_capabilities_get()
257 dev->data->nb_tx_queues; in mrvl_level_capabilities_get()
265 cap->n_nodes_max = dev->data->nb_tx_queues; in mrvl_level_capabilities_get()
266 cap->n_nodes_leaf_max = dev->data->nb_tx_queues; in mrvl_level_capabilities_get()
317 cap->nonleaf.sched_n_children_max = dev->data->nb_tx_queues; in mrvl_node_capabilities_get()
320 dev->data->nb_tx_queues; in mrvl_node_capabilities_get()
507 if (node_id >= dev->data->nb_tx_queues) { in mrvl_node_check_params()
[all …]
/f-stack/dpdk/drivers/net/nfb/
H A Dnfb_ethdev.c121 uint16_t nb_tx = dev->data->nb_tx_queues; in nfb_eth_dev_start()
159 uint16_t nb_tx = dev->data->nb_tx_queues; in nfb_eth_dev_stop()
202 dev_info->max_tx_queues = dev->data->nb_tx_queues; in nfb_eth_dev_info()
222 uint16_t nb_tx = dev->data->nb_tx_queues; in nfb_eth_dev_close()
242 dev->data->nb_tx_queues = 0; in nfb_eth_dev_close()
477 data->nb_tx_queues = ndp_get_tx_queue_available_count(internals->nfb); in nfb_eth_dev_init()
480 data->nb_rx_queues, data->nb_tx_queues); in nfb_eth_dev_init()
H A Dnfb_stats.c16 uint16_t nb_tx = dev->data->nb_tx_queues; in nfb_eth_stats_get()
60 uint16_t nb_tx = dev->data->nb_tx_queues; in nfb_eth_stats_reset()
/f-stack/dpdk/drivers/net/thunderx/
H A Dnicvf_ethdev.h116 *tx_end = dev->data->nb_tx_queues ? in nicvf_tx_range()
117 RTE_MIN(tmp, dev->data->nb_tx_queues - 1) : 0; in nicvf_tx_range()
/f-stack/dpdk/drivers/net/pcap/
H A Drte_eth_pcap.c567 for (i = 0; i < dev->data->nb_tx_queues; i++) { in eth_dev_start()
602 for (i = 0; i < dev->data->nb_tx_queues; i++) in eth_dev_start()
630 for (i = 0; i < dev->data->nb_tx_queues; i++) { in eth_dev_stop()
653 for (i = 0; i < dev->data->nb_tx_queues; i++) in eth_dev_stop()
677 dev_info->max_tx_queues = dev->data->nb_tx_queues; in eth_dev_info()
701 i < dev->data->nb_tx_queues; i++) { in eth_stats_get()
729 for (i = 0; i < dev->data->nb_tx_queues; i++) { in eth_stats_reset()
1115 const unsigned int nb_tx_queues, in pmd_init_internals() argument
1161 data->nb_tx_queues = (uint16_t)nb_tx_queues; in pmd_init_internals()
1273 const unsigned int nb_tx_queues = tx_queues->num_of_queue; in eth_from_pcaps_common() local
[all …]
/f-stack/dpdk/drivers/net/hns3/
H A Dhns3_stats.c530 num = RTE_MIN(RTE_ETHDEV_QUEUE_STAT_CNTRS, eth_dev->data->nb_tx_queues); in hns3_stats_get()
598 for (i = 0; i != eth_dev->data->nb_tx_queues; ++i) { in hns3_stats_reset()
640 int tx_err_stats = dev->data->nb_tx_queues * HNS3_NUM_TX_ERRORS_XSTATS; in hns3_xstats_calc_num()
642 int tx_queue_stats = dev->data->nb_tx_queues * HNS3_NUM_TX_QUEUE_STATS; in hns3_xstats_calc_num()
673 for (j = 0; j < dev->data->nb_tx_queues; j++) { in hns3_get_queue_stats()
784 for (j = 0; j < dev->data->nb_tx_queues; j++) { in hns3_dev_xstats_get()
869 for (j = 0; j < dev->data->nb_tx_queues; j++) { in hns3_dev_xstats_get_names()
888 for (j = 0; j < dev->data->nb_tx_queues; j++) { in hns3_dev_xstats_get_names()
/f-stack/dpdk/app/test-pmd/
H A Dutil.c287 for (queue = 0; queue < dev_info.nb_tx_queues; queue++) in add_tx_md_callback()
308 for (queue = 0; queue < dev_info.nb_tx_queues; queue++) in remove_tx_md_callback()
343 for (queue = 0; queue < dev_info.nb_tx_queues; queue++) in add_tx_dynf_callback()
364 for (queue = 0; queue < dev_info.nb_tx_queues; queue++) in remove_tx_dynf_callback()
/f-stack/dpdk/drivers/net/mlx4/
H A Dmlx4_txq.c134 const unsigned int txqs_n = dev->data->nb_tx_queues; in mlx4_tx_uar_init_secondary()
319 if (idx >= dev->data->nb_tx_queues) { in mlx4_tx_queue_setup()
322 (void *)dev, idx, dev->data->nb_tx_queues); in mlx4_tx_queue_setup()
512 for (i = 0; i != ETH_DEV(priv)->data->nb_tx_queues; ++i) in mlx4_tx_queue_release()
/f-stack/dpdk/drivers/net/null/
H A Drte_eth_null.c257 if (tx_queue_id >= dev->data->nb_tx_queues) in eth_tx_queue_setup()
323 RTE_MIN(dev->data->nb_tx_queues, in eth_stats_get()
503 const unsigned int nb_tx_queues = 1; in eth_dev_null_create() local
548 data->nb_tx_queues = (uint16_t)nb_tx_queues; in eth_dev_null_create()
/f-stack/dpdk/drivers/net/octeontx2/
H A Dotx2_link.c168 for (i = 0; i < eth_dev->data->nb_tx_queues; i++) in otx2_nix_dev_set_link_up()
185 for (i = 0; i < eth_dev->data->nb_tx_queues; i++) in otx2_nix_dev_set_link_down()
/f-stack/dpdk/drivers/net/softnic/
H A Drte_eth_softnic_link.c95 link->n_txq = port_info.nb_tx_queues; in softnic_link_create()
/f-stack/dpdk/drivers/net/kni/
H A Drte_eth_kni.c316 data->nb_tx_queues); in eth_kni_stats_get()
345 for (i = 0; i < data->nb_tx_queues; i++) { in eth_kni_stats_reset()
390 data->nb_tx_queues = 1; in eth_kni_create()
/f-stack/dpdk/drivers/net/vmxnet3/
H A Dvmxnet3_ethdev.c421 if (dev->data->nb_tx_queues > VMXNET3_MAX_TX_QUEUES || in vmxnet3_dev_configure()
433 dev->data->nb_tx_queues * sizeof(struct Vmxnet3_RxQueueDesc); in vmxnet3_dev_configure()
439 hw->num_tx_queues = (uint8_t)dev->data->nb_tx_queues; in vmxnet3_dev_configure()
891 for (i = 0; i < dev->data->nb_tx_queues; i++) { in vmxnet3_free_queues()
896 dev->data->nb_tx_queues = 0; in vmxnet3_free_queues()
1020 dev->data->nb_tx_queues * RTE_DIM(vmxnet3_txq_stat_strings) + in vmxnet3_dev_xstats_get_names()
1039 for (i = 0; i < dev->data->nb_tx_queues; i++) { in vmxnet3_dev_xstats_get_names()
1061 dev->data->nb_tx_queues * RTE_DIM(vmxnet3_txq_stat_strings) + in vmxnet3_dev_xstats_get()
1081 for (i = 0; i < dev->data->nb_tx_queues; i++) { in vmxnet3_dev_xstats_get()
/f-stack/dpdk/drivers/net/vhost/
H A Drte_eth_vhost.c242 for (i = 0; i < dev->data->nb_tx_queues; i++) { in vhost_dev_xstats_reset()
306 for (i = 0; i < dev->data->nb_tx_queues; i++) { in vhost_dev_xstats_get()
742 for (i = 0; i < dev->data->nb_tx_queues; i++) { in update_queuing_status()
766 for (i = 0; i < eth_dev->data->nb_tx_queues; i++) { in queue_setup()
870 for (i = 0; i < eth_dev->data->nb_tx_queues; i++) { in destroy_device()
1198 for (i = 0; i < dev->data->nb_tx_queues; i++) in eth_dev_close()
1304 i < dev->data->nb_tx_queues; i++) { in eth_stats_get()
1336 for (i = 0; i < dev->data->nb_tx_queues; i++) { in eth_stats_reset()
1444 data->nb_tx_queues = queues; in eth_dev_vhost_create()
/f-stack/dpdk/drivers/net/tap/
H A Drte_eth_tap.c915 for (i = 0; i < dev->data->nb_tx_queues; i++) in tap_dev_start()
930 for (i = 0; i < dev->data->nb_tx_queues; i++) in tap_dev_stop()
954 if (dev->data->nb_tx_queues > RTE_PMD_TAP_MAX_QUEUES) { in tap_dev_configure()
958 dev->data->nb_tx_queues, in tap_dev_configure()
964 dev->device->name, pmd->name, dev->data->nb_tx_queues); in tap_dev_configure()
1057 dev->data->nb_tx_queues : RTE_ETHDEV_QUEUE_STAT_CNTRS; in tap_stats_get()
1618 if (tx_queue_id >= dev->data->nb_tx_queues) in tap_tx_queue_setup()
1955 data->nb_tx_queues = 0; in eth_dev_tap_create()
2342 dev->data->nb_tx_queues = reply_param->txq_count; in tap_mp_attach_queues()
2381 if (dev->data->nb_rx_queues + dev->data->nb_tx_queues > in tap_mp_sync_queues()
[all …]
/f-stack/dpdk/drivers/net/netvsc/
H A Dhn_ethdev.c587 dev->data->nb_tx_queues); in hn_dev_configure()
630 for (i = 0; i < dev->data->nb_tx_queues; i++) { in hn_dev_stats_get()
674 for (i = 0; i < dev->data->nb_tx_queues; i++) { in hn_dev_stats_reset()
711 count = dev->data->nb_tx_queues * RTE_DIM(hn_stat_strings); in hn_dev_xstats_count()
733 for (i = 0; i < dev->data->nb_tx_queues; i++) { in hn_dev_xstats_get_names()
787 for (i = 0; i < dev->data->nb_tx_queues; i++) { in hn_dev_xstats_get()
/f-stack/dpdk/drivers/net/mlx5/
H A Dmlx5_ethdev.c65 unsigned int txqs_n = dev->data->nb_tx_queues; in mlx5_dev_configure()
228 dev->data->nb_tx_queues > 2) { in mlx5_set_default_params()
237 dev->data->nb_tx_queues > 2) { in mlx5_set_default_params()
/f-stack/dpdk/drivers/net/liquidio/
H A Dlio_ethdev.c303 for (i = 0; i < eth_dev->data->nb_tx_queues; i++) { in lio_dev_stats_get()
351 for (i = 0; i < eth_dev->data->nb_tx_queues; i++) { in lio_dev_stats_reset()
1256 if (q_no >= lio_dev->nb_tx_queues) { in lio_dev_tx_queue_setup()
1713 lio_dev->nb_tx_queues != num_txq) { in lio_reconf_queues()
1717 lio_dev->nb_tx_queues = num_txq; in lio_reconf_queues()
1758 eth_dev->data->nb_tx_queues, in lio_dev_configure()
1762 lio_dev->nb_tx_queues = eth_dev->data->nb_tx_queues; in lio_dev_configure()
1766 lio_dev->max_tx_queues = eth_dev->data->nb_tx_queues; in lio_dev_configure()
1779 if_cfg.s.num_iqueues = lio_dev->nb_tx_queues; in lio_dev_configure()
/f-stack/dpdk/lib/librte_ethdev/
H A Drte_ethdev_core.h138 uint16_t nb_tx_queues; /**< Number of TX queues. */ member
/f-stack/dpdk/drivers/net/octeontx/
H A Docteontx_ethdev_ops.c187 for (i = 0; i < eth_dev->data->nb_tx_queues; i++) in octeontx_dev_set_link_up()
201 for (i = 0; i < eth_dev->data->nb_tx_queues; i++) in octeontx_dev_set_link_down()
/f-stack/dpdk/lib/librte_latencystats/
H A Drte_latencystats.c286 for (qid = 0; qid < dev_info.nb_tx_queues; qid++) { in rte_latencystats_init()
329 for (qid = 0; qid < dev_info.nb_tx_queues; qid++) { in rte_latencystats_uninit()
/f-stack/dpdk/drivers/net/ice/
H A Dice_dcf_ethdev.c407 if (tx_queue_id >= dev->data->nb_tx_queues) in ice_dcf_tx_queue_start()
439 if (tx_queue_id >= dev->data->nb_tx_queues) in ice_dcf_tx_queue_stop()
465 for (nb_txq = 0; nb_txq < dev->data->nb_tx_queues; nb_txq++) { in ice_dcf_start_queues()
510 dev->data->nb_tx_queues); in ice_dcf_dev_start()
574 for (i = 0; i < dev->data->nb_tx_queues; i++) { in ice_dcf_stop_queues()

123456