| /dpdk/drivers/net/mlx5/ |
| H A D | mlx5_txq.c | 190 struct mlx5_txq_data *txq = (*priv->txqs)[idx]; in mlx5_tx_queue_stop_primary() 404 (*priv->txqs)[idx] = &txq_ctrl->txq; in mlx5_tx_queue_setup() 474 (*priv->txqs)[idx] = &txq_ctrl->txq; in mlx5_tx_hairpin_queue_setup() 630 if (!(*priv->txqs)[i]) in mlx5_tx_uar_init_secondary() 632 txq = (*priv->txqs)[i]; in mlx5_tx_uar_init_secondary() 645 if (!(*priv->txqs)[i]) in mlx5_tx_uar_init_secondary() 647 txq = (*priv->txqs)[i]; in mlx5_tx_uar_init_secondary() 1201 if (priv->txqs == NULL || (*priv->txqs)[idx] == NULL) in mlx5_txq_release() 1225 (*priv->txqs)[idx] = NULL; in mlx5_txq_release() 1247 if (!(*priv->txqs)[idx]) in mlx5_txq_releasable() [all …]
|
| H A D | mlx5_tx.c | 537 if (priv->txqs_n && (*priv->txqs)[0]) { in mlx5_select_tx_function() 538 struct mlx5_txq_data *txd = (*priv->txqs)[0]; in mlx5_select_tx_function() 683 struct mlx5_txq_data *txq = (*priv->txqs)[tx_queue_id]; in mlx5_txq_info_get() 721 struct mlx5_txq_data *txq = (*priv->txqs)[tx_queue_id]; in mlx5_tx_burst_mode_get()
|
| H A D | mlx5_stats.c | 136 struct mlx5_txq_data *txq = (*priv->txqs)[i]; in mlx5_stats_get() 198 struct mlx5_txq_data *txq_data = (*priv->txqs)[i]; in mlx5_stats_reset()
|
| H A D | mlx5_rxtx.c | 386 struct mlx5_txq_data *txq = (*priv->txqs)[sm->queue_id]; in mlx5_queue_state_modify_primary()
|
| H A D | mlx5_devx.c | 1131 struct mlx5_txq_data *txq_data = (*priv->txqs)[idx]; in mlx5_txq_obj_hairpin_new() 1213 struct mlx5_txq_data *txq_data = (*priv->txqs)[idx]; in mlx5_txq_create_devx_sq_resources() 1254 struct mlx5_txq_data *txq_data = (*priv->txqs)[idx]; in mlx5_txq_devx_obj_new()
|
| H A D | mlx5_ethdev.c | 120 priv->txqs = (void *)dev->data->tx_queues; in mlx5_dev_configure()
|
| H A D | mlx5.c | 1878 if (priv->txqs != NULL) { in mlx5_dev_close() 1884 priv->txqs = NULL; in mlx5_dev_close()
|
| H A D | mlx5.h | 1470 struct mlx5_txq_data *(*txqs)[]; /* TX queues. */ member
|
| /dpdk/drivers/net/hinic/ |
| H A D | hinic_pmd_tx.c | 1253 if (nic_dev->txqs[q_id] == NULL) in hinic_free_all_tx_resources() 1257 hinic_free_all_tx_mbufs(nic_dev->txqs[q_id]); in hinic_free_all_tx_resources() 1258 hinic_free_tx_resources(nic_dev->txqs[q_id]); in hinic_free_all_tx_resources() 1261 kfree(nic_dev->txqs[q_id]); in hinic_free_all_tx_resources() 1262 nic_dev->txqs[q_id] = NULL; in hinic_free_all_tx_resources() 1274 hinic_free_all_tx_mbufs(nic_dev->txqs[q_id]); in hinic_free_all_tx_mbuf()
|
| H A D | hinic_pmd_ethdev.c | 587 nic_dev->txqs[queue_idx] = txq; in hinic_tx_queue_setup() 1137 nic_dev->txqs[txq->q_id] = NULL; in hinic_tx_queue_release() 1335 txq = nic_dev->txqs[i]; in hinic_dev_stats_get() 1390 txq = nic_dev->txqs[qid]; in hinic_dev_stats_reset() 2199 txq = nic_dev->txqs[qid]; in hinic_dev_xstats_get() 2675 nic_dev->txqs = kzalloc_aligned(txq_size, GFP_KERNEL); in hinic_init_sw_rxtxqs() 2676 if (!nic_dev->txqs) { in hinic_init_sw_rxtxqs() 2686 kfree(nic_dev->txqs); in hinic_init_sw_rxtxqs() 2687 nic_dev->txqs = NULL; in hinic_init_sw_rxtxqs() 2698 kfree(nic_dev->txqs); in hinic_deinit_sw_rxtxqs() [all …]
|
| H A D | hinic_pmd_ethdev.h | 326 struct hinic_txq **txqs; member
|
| /dpdk/drivers/net/thunderx/ |
| H A D | nicvf_ethdev.c | 340 uint16_t rxqs = 0, txqs = 0; in nicvf_dev_stats_reset() local 353 txqs |= (0x3 << (i * 2)); in nicvf_dev_stats_reset() 355 ret = nicvf_mbox_reset_stat_counters(nic, 0x3FFF, 0x1F, rxqs, txqs); in nicvf_dev_stats_reset() 371 txqs |= (0x3 << ((i % MAX_SND_QUEUES_PER_QS) * 2)); in nicvf_dev_stats_reset() 373 ret = nicvf_mbox_reset_stat_counters(snic, 0, 0, rxqs, txqs); in nicvf_dev_stats_reset()
|
| /dpdk/drivers/net/ice/base/ |
| H A D | ice_common.c | 4203 sum_size += ice_struct_size(list, txqs, list->num_txqs); in ice_aq_add_lan_txq() 4204 list = (struct ice_aqc_add_tx_qgrp *)(list->txqs + in ice_aq_add_lan_txq() 4965 buf->txqs[0].info.valid_sections = in ice_ena_vsi_txq() 4968 buf->txqs[0].info.generic = 0; in ice_ena_vsi_txq() 4969 buf->txqs[0].info.cir_bw.bw_profile_idx = in ice_ena_vsi_txq() 4971 buf->txqs[0].info.cir_bw.bw_alloc = in ice_ena_vsi_txq() 4973 buf->txqs[0].info.eir_bw.bw_profile_idx = in ice_ena_vsi_txq() 4975 buf->txqs[0].info.eir_bw.bw_alloc = in ice_ena_vsi_txq() 4982 LE16_TO_CPU(buf->txqs[0].txq_id), in ice_ena_vsi_txq() 4987 node.node_teid = buf->txqs[0].q_teid; in ice_ena_vsi_txq()
|
| H A D | ice_adminq_cmd.h | 2557 struct ice_aqc_add_txqs_perq txqs[STRUCT_HACK_VAR_LEN]; member 2645 struct ice_aqc_move_txqs_elem txqs[STRUCT_HACK_VAR_LEN]; member
|
| /dpdk/drivers/net/ice/ |
| H A D | ice_rxtx.c | 678 buf_len = ice_struct_size(txq_elem, txqs, 1); in ice_tx_queue_start() 688 txq_elem->txqs[0].txq_id = rte_cpu_to_le_16(txq->reg_idx); in ice_tx_queue_start() 701 ice_set_ctx(hw, (uint8_t *)&tx_ctx, txq_elem->txqs[0].txq_ctx, in ice_tx_queue_start() 718 txq->q_teid = txq_elem->txqs[0].q_teid; in ice_tx_queue_start() 859 buf_len = ice_struct_size(txq_elem, txqs, 1); in ice_fdir_tx_queue_start() 869 txq_elem->txqs[0].txq_id = rte_cpu_to_le_16(txq->reg_idx); in ice_fdir_tx_queue_start() 881 ice_set_ctx(hw, (uint8_t *)&tx_ctx, txq_elem->txqs[0].txq_ctx, in ice_fdir_tx_queue_start() 898 txq->q_teid = txq_elem->txqs[0].q_teid; in ice_fdir_tx_queue_start()
|
| /dpdk/drivers/net/mlx5/linux/ |
| H A D | mlx5_verbs.c | 862 struct mlx5_txq_data *txq_data = (*priv->txqs)[idx]; in mlx5_txq_ibv_qp_create() 954 struct mlx5_txq_data *txq_data = (*priv->txqs)[idx]; in mlx5_txq_ibv_obj_new()
|