| /f-stack/dpdk/drivers/net/mlx5/ |
| H A D | mlx5_rxq.c | 107 struct mlx5_rxq_data *rxq = (*priv->rxqs)[i]; in mlx5_mprq_enabled() 458 if (!(*priv->rxqs)[idx]) { in mlx5_rxq_releasable() 523 struct mlx5_rxq_data *rxq = (*priv->rxqs)[idx]; in mlx5_rx_queue_stop_primary() 788 (*priv->rxqs)[idx] = &rxq_ctrl->rxq; in mlx5_rx_queue_setup() 858 (*priv->rxqs)[idx] = &rxq_ctrl->rxq; in mlx5_rx_hairpin_queue_setup() 1782 if (!(*priv->rxqs)[idx]) in mlx5_rxq_release() 1802 (*priv->rxqs)[idx] = NULL; in mlx5_rxq_release() 1849 rxq_ctrl = container_of((*priv->rxqs)[idx], in mlx5_rxq_get_type() 1875 rxq_ctrl = container_of((*priv->rxqs)[idx], in mlx5_rxq_get_hairpin_conf() 2553 if (!(*priv->rxqs)[i]) in mlx5_rxq_timestamp_set() [all …]
|
| H A D | mlx5_stats.c | 109 struct mlx5_rxq_data *rxq = (*priv->rxqs)[i]; in mlx5_stats_get() 183 if ((*priv->rxqs)[i] == NULL) in mlx5_stats_reset() 185 memset(&(*priv->rxqs)[i]->stats, 0, in mlx5_stats_reset()
|
| H A D | mlx5_rss.c | 67 if (!(*priv->rxqs)[i]) in mlx5_rss_hash_update() 69 (*priv->rxqs)[i]->rss_hash = !!rss_conf->rss_hf && in mlx5_rss_hash_update()
|
| H A D | mlx5_devx.c | 324 struct mlx5_rxq_data *rxq_data = (*priv->rxqs)[idx]; in mlx5_rxq_create_devx_rq_resources() 420 struct mlx5_rxq_data *rxq_data = (*priv->rxqs)[idx]; in mlx5_rxq_create_devx_cq_resources() 573 struct mlx5_rxq_data *rxq_data = (*priv->rxqs)[idx]; in mlx5_rxq_obj_hairpin_new() 632 struct mlx5_rxq_data *rxq_data = (*priv->rxqs)[idx]; in mlx5_rxq_devx_obj_new() 731 struct mlx5_rxq_data *rxq = (*priv->rxqs)[queues[i]]; in mlx5_devx_ind_table_create_rqt_attr() 853 struct mlx5_rxq_data *rxq_data = (*priv->rxqs)[ind_tbl->queues[0]]; in mlx5_devx_tir_attr_set() 862 if (!(*priv->rxqs)[ind_tbl->queues[i]]->lro) { in mlx5_devx_tir_attr_set()
|
| H A D | mlx5_vlan.c | 93 struct mlx5_rxq_data *rxq = (*priv->rxqs)[queue]; in mlx5_vlan_strip_queue_set()
|
| H A D | mlx5_ethdev.c | 102 priv->rxqs = (void *)dev->data->rx_queues; in mlx5_dev_configure() 162 rxq_data = (*priv->rxqs)[i]; in mlx5_dev_configure_rss_reta()
|
| H A D | mlx5_rxtx_vec.c | 574 struct mlx5_rxq_data *rxq = (*priv->rxqs)[i]; in mlx5_check_vec_rx_support()
|
| H A D | mlx5_flow.c | 1016 container_of((*priv->rxqs)[idx], in flow_drv_rxq_flags_set() 1099 container_of((*priv->rxqs)[idx], in flow_drv_rxq_flags_trim() 1165 if (!(*priv->rxqs)[i]) in flow_rxq_flags_clear() 1167 rxq_ctrl = container_of((*priv->rxqs)[i], in flow_rxq_flags_clear() 1191 if (!(*priv->rxqs)[i]) in mlx5_flow_rxq_dynf_metadata_set() 1193 data = (*priv->rxqs)[i]; in mlx5_flow_rxq_dynf_metadata_set() 1382 if (!(*priv->rxqs)[queue->index]) in mlx5_flow_validate_action_queue() 1488 if (!(*priv->rxqs)[rss->queue[i]]) in mlx5_validate_action_rss() 1492 rxq_ctrl = container_of((*priv->rxqs)[rss->queue[i]], in mlx5_validate_action_rss()
|
| H A D | mlx5.c | 1329 if (priv->rxqs != NULL) { in mlx5_dev_close() 1335 priv->rxqs = NULL; in mlx5_dev_close()
|
| H A D | mlx5_rxtx.c | 552 struct mlx5_rxq_data *rxq = (*priv->rxqs)[rx_queue_id]; in mlx5_rxq_info_get() 598 rxq = (*priv->rxqs)[rx_queue_id]; in mlx5_rx_burst_mode_get() 657 rxq = (*priv->rxqs)[rx_queue_id]; in mlx5_rx_queue_count() 929 struct mlx5_rxq_data *rxq = (*priv->rxqs)[sm->queue_id]; in mlx5_queue_state_modify_primary()
|
| H A D | mlx5.h | 964 struct mlx5_rxq_data *(*rxqs)[]; /* RX queues. */ member
|
| H A D | mlx5_trigger.c | 176 dev->data->port_id, (*priv->rxqs)[i]->idx); in mlx5_rxq_start()
|
| /f-stack/dpdk/drivers/net/hinic/ |
| H A D | hinic_pmd_rx.c | 198 rxq = nic_dev->rxqs[q_id]; in hinic_get_func_rx_buf_size() 419 if (nic_dev->rxqs[q_id] == NULL) in hinic_free_all_rx_resources() 422 hinic_free_all_rx_mbufs(nic_dev->rxqs[q_id]); in hinic_free_all_rx_resources() 423 hinic_free_rx_resources(nic_dev->rxqs[q_id]); in hinic_free_all_rx_resources() 424 kfree(nic_dev->rxqs[q_id]); in hinic_free_all_rx_resources() 425 nic_dev->rxqs[q_id] = NULL; in hinic_free_all_rx_resources() 436 hinic_free_all_rx_mbufs(nic_dev->rxqs[q_id]); in hinic_free_all_rx_mbuf()
|
| H A D | hinic_pmd_ethdev.c | 438 nic_dev->rxqs[queue_idx] = rxq; in hinic_rx_queue_setup() 1107 nic_dev->rxqs[rxq->q_id] = NULL; in hinic_rx_queue_release() 1321 rxq = nic_dev->rxqs[i]; in hinic_dev_stats_get() 1386 rxq = nic_dev->rxqs[qid]; in hinic_dev_stats_reset() 2200 rxq = nic_dev->rxqs[qid]; in hinic_dev_xstats_get() 2722 rxq_size = nic_dev->nic_cap.max_rqs * sizeof(*nic_dev->rxqs); in hinic_init_sw_rxtxqs() 2723 nic_dev->rxqs = kzalloc_aligned(rxq_size, GFP_KERNEL); in hinic_init_sw_rxtxqs() 2724 if (!nic_dev->rxqs) { in hinic_init_sw_rxtxqs() 2741 kfree(nic_dev->rxqs); in hinic_deinit_sw_rxtxqs() 2742 nic_dev->rxqs = NULL; in hinic_deinit_sw_rxtxqs()
|
| H A D | hinic_pmd_ethdev.h | 311 struct hinic_rxq **rxqs; member
|
| /f-stack/dpdk/drivers/net/mlx5/linux/ |
| H A D | mlx5_verbs.c | 195 struct mlx5_rxq_data *rxq_data = (*priv->rxqs)[idx]; in mlx5_rxq_ibv_cq_create() 262 struct mlx5_rxq_data *rxq_data = (*priv->rxqs)[idx]; in mlx5_rxq_ibv_wq_create() 358 struct mlx5_rxq_data *rxq_data = (*priv->rxqs)[idx]; in mlx5_rxq_ibv_obj_new() 529 struct mlx5_rxq_data *rxq = (*priv->rxqs)[ind_tbl->queues[i]]; in mlx5_ibv_ind_table_new()
|
| /f-stack/dpdk/drivers/net/thunderx/ |
| H A D | nicvf_ethdev.c | 354 uint16_t rxqs = 0, txqs = 0; in nicvf_dev_stats_reset() local 363 rxqs |= (0x3 << (i * 2)); in nicvf_dev_stats_reset() 369 ret = nicvf_mbox_reset_stat_counters(nic, 0x3FFF, 0x1F, rxqs, txqs); in nicvf_dev_stats_reset() 381 rxqs |= (0x3 << ((i % MAX_CMP_QUEUES_PER_QS) * 2)); in nicvf_dev_stats_reset() 387 ret = nicvf_mbox_reset_stat_counters(snic, 0, 0, rxqs, txqs); in nicvf_dev_stats_reset()
|
| /f-stack/dpdk/doc/guides/sample_app_ug/ |
| H A D | l3_forward.rst | 68 [--event-eth-rxqs] 103 * ``--event-eth-rxqs:`` Optional, Number of ethernet RX queues per device. Only valid if --mode=eve…
|
| /f-stack/dpdk/drivers/net/bnxt/ |
| H A D | bnxt_hwrm.c | 4824 struct bnxt_rx_queue **rxqs = bp->rx_queues; in bnxt_vnic_rss_configure_thor() local 4870 rxr = rxqs[k]->rx_ring; in bnxt_vnic_rss_configure_thor() 4871 cpr = rxqs[k]->cp_ring; in bnxt_vnic_rss_configure_thor()
|