| /dpdk/drivers/net/i40e/ |
| H A D | i40e_vf_representor.c | 33 representor->adapter->pf.dev_data; in i40e_vf_representor_dev_infos_get() 221 representor->adapter->pf.dev_data->port_id, in i40e_vf_representor_stats_get() 281 representor->adapter->pf.dev_data->port_id, in i40e_vf_representor_stats_reset() 291 representor->adapter->pf.dev_data->port_id, in i40e_vf_representor_promiscuous_enable() 301 representor->adapter->pf.dev_data->port_id, in i40e_vf_representor_promiscuous_disable() 311 representor->adapter->pf.dev_data->port_id, in i40e_vf_representor_allmulticast_enable() 321 representor->adapter->pf.dev_data->port_id, in i40e_vf_representor_allmulticast_disable() 331 representor->adapter->pf.dev_data->port_id, in i40e_vf_representor_mac_addr_remove() 342 representor->adapter->pf.dev_data->port_id, in i40e_vf_representor_mac_addr_set() 354 representor->adapter->pf.dev_data->port_id, in i40e_vf_representor_vlan_filter_set() [all …]
|
| H A D | i40e_ethdev.h | 99 struct rte_eth_dev_data *dev_data; \ 101 dev_data = ((struct i40e_adapter *)hw->back)->pf.dev_data; \ 102 dev = &rte_eth_devices[dev_data->port_id]; \ 1130 struct rte_eth_dev_data *dev_data; /* Pointer to the device data */ member 1464 (((struct i40e_vsi *)vsi)->adapter->pf.dev_data) 1466 (&rte_eth_devices[((struct i40e_vsi *)vsi)->adapter->pf.dev_data->port_id])
|
| H A D | i40e_hash.c | 746 struct rte_eth_dev *dev = &rte_eth_devices[pf->dev_data->port_id]; in i40e_hash_config_region() 946 if (pf->dev_data->dev_conf.rxmode.mq_mode & RTE_ETH_MQ_RX_VMDQ_FLAG) in i40e_hash_parse_queues() 949 max_queue = pf->dev_data->nb_rx_queues; in i40e_hash_parse_queues() 1298 dev = &rte_eth_devices[pf->dev_data->port_id]; in i40e_hash_reset_conf()
|
| H A D | i40e_ethdev.c | 726 struct rte_eth_dev_data *dev_data = in i40e_write_global_rx_ctl() local 727 ((struct i40e_adapter *)hw->back)->pf.dev_data; in i40e_write_global_rx_ctl() 1321 struct rte_eth_dev_data *dev_data = in i40e_aq_debug_write_global_register() local 1322 ((struct i40e_adapter *)hw->back)->pf.dev_data; in i40e_aq_debug_write_global_register() 1449 pf->dev_data = dev->data; in eth_i40e_dev_init() 6375 struct rte_eth_dev_data *data = pf->dev_data; in i40e_dev_tx_init() 6398 struct rte_eth_dev_data *data = pf->dev_data; in i40e_dev_rx_init() 8765 struct rte_eth_dev_data *data = pf->dev_data; in i40e_pf_calc_configured_queues_num() 8851 num = pf->dev_data->nb_rx_queues; in i40e_pf_reset_rss_reta() 9562 struct rte_eth_dev_data *dev_data = in i40e_check_write_global_reg() local [all …]
|
| /dpdk/drivers/net/qede/ |
| H A D | qede_debug.c | 877 if (dev_data->initialized) in qed_dbg_dev_init() 882 dev_data->chip_id = CHIP_K2; in qed_dbg_dev_init() 888 dev_data->chip_id = CHIP_BB; in qed_dbg_dev_init() 921 dev_data->num_ports = 1; in qed_dbg_dev_init() 923 dev_data->num_ports = 2; in qed_dbg_dev_init() 925 dev_data->num_ports = 4; in qed_dbg_dev_init() 935 dev_data->use_dmae = true; in qed_dbg_dev_init() 936 dev_data->initialized = 1; in qed_dbg_dev_init() 1621 u8 chip_id = dev_data->chip_id; in qed_grc_unreset_blocks() 1877 dev_data->use_dmae = 0; in qed_grc_dump_addr_range() [all …]
|
| /dpdk/drivers/net/iavf/ |
| H A D | iavf_vchnl.c | 826 adapter->dev_data->nb_tx_queues; in iavf_enable_queues_lv() 831 adapter->dev_data->nb_rx_queues; in iavf_enable_queues_lv() 870 adapter->dev_data->nb_tx_queues; in iavf_disable_queues_lv() 875 adapter->dev_data->nb_rx_queues; in iavf_disable_queues_lv() 1006 (struct iavf_rx_queue **)adapter->dev_data->rx_queues; in iavf_configure_queues() 1008 (struct iavf_tx_queue **)adapter->dev_data->tx_queues; in iavf_configure_queues() 1032 if (i < adapter->dev_data->nb_tx_queues) { in iavf_configure_queues() 1041 if (i >= adapter->dev_data->nb_rx_queues) in iavf_configure_queues() 1115 for (i = 0; i < adapter->dev_data->nb_rx_queues; i++) { in iavf_config_irq_map() 1194 addr = &adapter->dev_data->mac_addrs[i]; in iavf_add_del_all_mac_addr() [all …]
|
| H A D | iavf_rxtx.h | 757 for (i = 0; i < (ad)->dev_data->nb_rx_queues; i++) { \ 758 struct iavf_rx_queue *rxq = (ad)->dev_data->rx_queues[i]; \
|
| H A D | iavf_ethdev.c | 499 rss_conf = &adapter->dev_data->dev_conf.rx_adv_conf.rss_conf; in iavf_init_rss() 500 nb_q = RTE_MIN(adapter->dev_data->nb_rx_queues, in iavf_init_rss() 690 struct rte_eth_dev_data *dev_data = dev->data; in iavf_init_rxq() local 725 if ((dev_data->dev_conf.rxmode.offloads & RTE_ETH_RX_OFFLOAD_SCATTER) || in iavf_init_rxq() 727 dev_data->scattered_rx = 1; in iavf_init_rxq() 1533 adapter->dev_data->dev_conf.rx_adv_conf.rss_conf = *rss_conf; in iavf_dev_rss_hash_update() 2558 adapter->dev_data = eth_dev->data; in iavf_dev_init()
|
| /dpdk/drivers/net/mlx4/ |
| H A D | mlx4.h | 157 struct rte_eth_dev_data *dev_data; /* Pointer to device data. */ member 197 #define PORT_ID(priv) ((priv)->dev_data->port_id)
|
| /dpdk/drivers/net/avp/ |
| H A D | avp_ethdev.c | 161 struct rte_eth_dev_data *dev_data; member 216 struct rte_eth_dev_data *dev_data; member 862 avp->dev_data = eth_dev->data; in avp_dev_create() 1142 rxq->dev_data = eth_dev->data; in avp_dev_rx_queue_setup() 1186 txq->dev_data = eth_dev->data; in avp_dev_tx_queue_setup() 1451 rxq->dev_data->rx_mbuf_alloc_failed++; in avp_recv_scattered_pkts() 1558 rxq->dev_data->rx_mbuf_alloc_failed++; in avp_recv_pkts() 1942 struct rte_eth_dev_data *data = avp->dev_data; in avp_dev_rx_queue_release_all() 1966 struct rte_eth_dev_data *data = avp->dev_data; in avp_dev_tx_queue_release_all() 2239 struct avp_queue *rxq = avp->dev_data->rx_queues[i]; in avp_dev_stats_get() [all …]
|
| /dpdk/drivers/net/mlx5/ |
| H A D | mlx5_txq.c | 751 uint64_t dev_txoff = priv->dev_data->dev_conf.txmode.offloads; in txq_set_params() 988 priv->dev_data->port_id, priv->sh->dev_cap.max_qp_wr); in txq_adjust_params() 999 priv->dev_data->port_id, priv->sh->dev_cap.max_qp_wr); in txq_adjust_params() 1010 priv->dev_data->port_id, priv->sh->dev_cap.max_qp_wr); in txq_adjust_params() 1019 priv->dev_data->port_id, priv->sh->dev_cap.max_qp_wr); in txq_adjust_params() 1027 priv->dev_data->port_id); in txq_adjust_params() 1035 priv->dev_data->port_id); in txq_adjust_params()
|
| H A D | mlx5_devx.c | 281 rq_attr.user_index = rte_cpu_to_be_16(priv->dev_data->port_id); in mlx5_rxq_create_devx_rq_resources() 305 uint16_t port_id = priv->dev_data->port_id; in mlx5_rxq_create_devx_cq_resources() 451 priv->dev_data->port_id, idx); in mlx5_rxq_obj_hairpin_new() 455 priv->dev_data->rx_queue_state[idx] = RTE_ETH_QUEUE_STATE_HAIRPIN; in mlx5_rxq_obj_hairpin_new() 508 priv->dev_data->port_id, rxq->idx); in mlx5_rxq_devx_obj_new() 528 priv->dev_data->rx_queue_state[rxq->idx] = RTE_ETH_QUEUE_STATE_STARTED; in mlx5_rxq_devx_obj_new()
|
| H A D | mlx5_trigger.c | 735 uint16_t local_port = priv->dev_data->port_id; in mlx5_hairpin_bind_single_port() 876 uint16_t cur_port = priv->dev_data->port_id; in mlx5_hairpin_unbind_single_port() 1029 priv->dev_data->port_id, i, pp); in mlx5_hairpin_get_peer_ports() 1050 priv->dev_data->port_id, i, pp); in mlx5_hairpin_get_peer_ports() 1332 priv->dev_data->port_id) { in mlx5_traffic_enable()
|
| /dpdk/drivers/net/sfc/ |
| H A D | sfc_repr.c | 501 struct rte_eth_dev_data *dev_data = dev->data; in sfc_repr_dev_configure() local 505 dev_data->nb_rx_queues, dev_data->nb_tx_queues); in sfc_repr_dev_configure() 512 ret = sfc_repr_configure(sr, dev_data->nb_rx_queues, in sfc_repr_dev_configure() 513 &dev_data->dev_conf); in sfc_repr_dev_configure()
|
| H A D | sfc_tx.c | 581 struct rte_eth_dev_data *dev_data; in sfc_tx_qstart() local 587 dev_data = sa->eth_dev->data; in sfc_tx_qstart() 588 dev_data->tx_queue_state[ethdev_qid] = in sfc_tx_qstart() 677 struct rte_eth_dev_data *dev_data; in sfc_tx_qstop() local 683 dev_data = sa->eth_dev->data; in sfc_tx_qstop() 684 dev_data->tx_queue_state[ethdev_qid] = in sfc_tx_qstop()
|
| H A D | sfc_port.c | 384 const struct rte_eth_dev_data *dev_data = sa->eth_dev->data; in sfc_port_configure() local 389 port->pdu = EFX_MAC_PDU(dev_data->mtu); in sfc_port_configure()
|
| H A D | sfc_ef10_essb_rx.c | 184 struct rte_eth_dev_data *dev_data = in sfc_ef10_essb_rx_qrefill() local 191 dev_data->rx_mbuf_alloc_failed += RTE_DIM(mbuf_blocks); in sfc_ef10_essb_rx_qrefill()
|
| H A D | sfc_ef10_rx.c | 129 struct rte_eth_dev_data *dev_data = in sfc_ef10_rx_qrefill() local 136 dev_data->rx_mbuf_alloc_failed += RTE_DIM(objs); in sfc_ef10_rx_qrefill()
|
| /dpdk/drivers/net/e1000/ |
| H A D | igb_pf.c | 195 struct rte_eth_dev_data *dev_data = dev->data; in set_rx_mode() local 210 if (dev_data->promiscuous) { in set_rx_mode() 214 if (dev_data->all_multicast) { in set_rx_mode()
|
| /dpdk/drivers/net/ice/ |
| H A D | ice_rxtx.h | 296 for (i = 0; i < (ad)->pf.dev_data->nb_rx_queues; i++) { \ 297 struct ice_rx_queue *rxq = (ad)->pf.dev_data->rx_queues[i]; \
|
| /dpdk/drivers/net/ngbe/ |
| H A D | ngbe_pf.c | 215 struct rte_eth_dev_data *dev_data = eth_dev->data; in ngbe_set_rx_mode() local 238 if (dev_data->promiscuous) { in ngbe_set_rx_mode() 243 } else if (dev_data->all_multicast) { in ngbe_set_rx_mode()
|
| /dpdk/drivers/net/txgbe/ |
| H A D | txgbe_pf.c | 302 struct rte_eth_dev_data *dev_data = eth_dev->data; in txgbe_set_rx_mode() local 325 if (dev_data->promiscuous) { in txgbe_set_rx_mode() 330 } else if (dev_data->all_multicast) { in txgbe_set_rx_mode()
|
| H A D | txgbe_ethdev_vf.c | 1102 struct rte_eth_dev_data *dev_data = dev->data; in txgbevf_dev_set_mtu() local 1113 if (dev_data->dev_started && !dev_data->scattered_rx && in txgbevf_dev_set_mtu()
|
| /dpdk/drivers/net/mlx5/linux/ |
| H A D | mlx5_verbs.c | 200 priv->dev_data->port_id); in mlx5_rxq_ibv_cq_create() 295 priv->dev_data->port_id, rxq->idx, in mlx5_rxq_ibv_wq_create() 321 uint16_t port_id = priv->dev_data->port_id; in mlx5_rxq_ibv_obj_new() 405 priv->dev_data->rx_queue_state[idx] = RTE_ETH_QUEUE_STATE_STARTED; in mlx5_rxq_ibv_obj_new()
|
| /dpdk/drivers/net/ixgbe/ |
| H A D | ixgbe_pf.c | 312 struct rte_eth_dev_data *dev_data = dev->data; in set_rx_mode() local 327 if (dev_data->promiscuous) { in set_rx_mode() 331 if (dev_data->all_multicast) { in set_rx_mode()
|