Lines Matching refs:uint16_t

203 uint16_t rte_eth_iterator_next(struct rte_dev_iterator *iter);
315 uint16_t link_duplex : 1; /**< ETH_LINK_[HALF/FULL]_DUPLEX */
316 uint16_t link_autoneg : 1; /**< ETH_LINK_[AUTONEG/FIXED] */
317 uint16_t link_status : 1; /**< ETH_LINK_[DOWN/UP] */
406 uint16_t split_hdr_size; /**< hdr buf size (header_split enabled).*/
824 uint16_t vlan_id[ETH_MIRROR_MAX_VLANS];
847 uint16_t reta[RTE_RETA_GROUP_SIZE];
911 uint16_t vlan_id; /**< The vlan id of the received frame */
944 uint16_t vlan_id; /**< The vlan id of the received frame */
961 uint16_t pvid;
1012 uint16_t length; /**< Segment data length, configures split point. */
1013 uint16_t offset; /**< Data offset from beginning of mbuf data buffer. */
1034 uint16_t rx_free_thresh; /**< Drives the freeing of RX descriptors. */
1037 uint16_t rx_nseg; /**< Number of descriptions in rx_seg array. */
1062 uint16_t tx_rs_thresh; /**< Drives the setting of RS bit on TXDs. */
1063 uint16_t tx_free_thresh; /**< Start freeing TX buffers if there are
1086 uint16_t max_nb_queues;
1088 uint16_t max_rx_2_tx;
1090 uint16_t max_tx_2_rx;
1091 uint16_t max_nb_desc; /**< The max num of descriptors. */
1103 uint16_t port; /**< Peer port. */
1104 uint16_t queue; /**< Peer queue. */
1147 uint16_t nb_max; /**< Max allowed number of descriptors. */
1148 uint16_t nb_min; /**< Min allowed number of descriptors. */
1149 uint16_t nb_align; /**< Number of descriptors should be aligned to. */
1159 uint16_t nb_seg_max;
1172 uint16_t nb_mtu_seg_max;
1193 uint16_t pause_time; /**< Pause quota in the Pause frame */
1194 uint16_t send_xon; /**< Is XON frame need be sent */
1272 uint16_t udp_port; /**< UDP port used for the tunnel. */
1453 uint16_t burst_size; /**< Device-preferred burst size */
1454 uint16_t ring_size; /**< Device-preferred size of queue rings */
1455 uint16_t nb_queues; /**< Device-preferred number of queues */
1469 uint16_t domain_id; /**< switch domain id */
1470 uint16_t port_id;
1491 uint16_t max_nseg; /**< Maximum amount of segments to split. */
1492 uint16_t reserved; /**< Reserved field. */
1509 uint16_t min_mtu; /**< Minimum MTU allowed */
1510 uint16_t max_mtu; /**< Maximum MTU allowed */
1516 uint16_t max_rx_queues; /**< Maximum number of RX queues. */
1517 uint16_t max_tx_queues; /**< Maximum number of TX queues. */
1521 uint16_t max_vfs; /**< Maximum number of VFs. */
1522 uint16_t max_vmdq_pools; /**< Maximum number of VMDq pools. */
1532 uint16_t reta_size;
1539 uint16_t vmdq_queue_base; /**< First queue ID for VMDQ pools. */
1540 uint16_t vmdq_queue_num; /**< Queue number for VMDQ pools. */
1541 uint16_t vmdq_pool_base; /**< First ID of VMDQ pools. */
1546 uint16_t nb_rx_queues; /**< Number of RX queues. */
1547 uint16_t nb_tx_queues; /**< Number of TX queues. */
1572 uint16_t nb_desc; /**< configured number of RXDs. */
1573 uint16_t rx_buf_size; /**< hardware receive buffer size. */
1582 uint16_t nb_desc; /**< configured number of TXDs. */
1650 uint16_t base;
1651 uint16_t nb_queue;
1655 uint16_t base;
1656 uint16_t nb_queue;
1747 typedef uint16_t (*rte_rx_callback_fn)(uint16_t port_id, uint16_t queue,
1748 struct rte_mbuf *pkts[], uint16_t nb_pkts, uint16_t max_pkts,
1771 typedef uint16_t (*rte_tx_callback_fn)(uint16_t port_id, uint16_t queue,
1772 struct rte_mbuf *pkts[], uint16_t nb_pkts, void *user_param);
1789 uint16_t def_vmdq_idx; /**< Default pool num used for PF */
1790 uint16_t def_pool_q_idx; /**< Default pool queue start reg index */
1834 uint64_t rte_eth_find_next_owned_by(uint16_t port_id,
1853 uint16_t rte_eth_find_next(uint16_t port_id);
1876 uint16_t
1877 rte_eth_find_next_of(uint16_t port_id_start,
1908 uint16_t
1909 rte_eth_find_next_sibling(uint16_t port_id_start, uint16_t ref_port_id);
1956 int rte_eth_dev_owner_set(const uint16_t port_id,
1973 int rte_eth_dev_owner_unset(const uint16_t port_id,
2004 int rte_eth_dev_owner_get(const uint16_t port_id,
2017 uint16_t rte_eth_dev_count_avail(void);
2027 uint16_t rte_eth_dev_count_total(void);
2101 int rte_eth_dev_configure(uint16_t port_id, uint16_t nb_rx_queue,
2102 uint16_t nb_tx_queue, const struct rte_eth_conf *eth_conf);
2117 rte_eth_dev_is_removed(uint16_t port_id);
2181 int rte_eth_rx_queue_setup(uint16_t port_id, uint16_t rx_queue_id,
2182 uint16_t nb_rx_desc, unsigned int socket_id,
2215 (uint16_t port_id, uint16_t rx_queue_id, uint16_t nb_rx_desc,
2266 int rte_eth_tx_queue_setup(uint16_t port_id, uint16_t tx_queue_id,
2267 uint16_t nb_tx_desc, unsigned int socket_id,
2297 (uint16_t port_id, uint16_t tx_queue_id, uint16_t nb_tx_desc,
2327 int rte_eth_hairpin_get_peer_ports(uint16_t port_id, uint16_t *peer_ports,
2353 int rte_eth_hairpin_bind(uint16_t tx_port, uint16_t rx_port);
2380 int rte_eth_hairpin_unbind(uint16_t tx_port, uint16_t rx_port);
2392 int rte_eth_dev_socket_id(uint16_t port_id);
2403 int rte_eth_dev_is_valid_port(uint16_t port_id);
2422 int rte_eth_dev_rx_queue_start(uint16_t port_id, uint16_t rx_queue_id);
2440 int rte_eth_dev_rx_queue_stop(uint16_t port_id, uint16_t rx_queue_id);
2459 int rte_eth_dev_tx_queue_start(uint16_t port_id, uint16_t tx_queue_id);
2477 int rte_eth_dev_tx_queue_stop(uint16_t port_id, uint16_t tx_queue_id);
2498 int rte_eth_dev_start(uint16_t port_id);
2510 int rte_eth_dev_stop(uint16_t port_id);
2524 int rte_eth_dev_set_link_up(uint16_t port_id);
2535 int rte_eth_dev_set_link_down(uint16_t port_id);
2547 int rte_eth_dev_close(uint16_t port_id);
2586 int rte_eth_dev_reset(uint16_t port_id);
2599 int rte_eth_promiscuous_enable(uint16_t port_id);
2612 int rte_eth_promiscuous_disable(uint16_t port_id);
2624 int rte_eth_promiscuous_get(uint16_t port_id);
2637 int rte_eth_allmulticast_enable(uint16_t port_id);
2650 int rte_eth_allmulticast_disable(uint16_t port_id);
2662 int rte_eth_allmulticast_get(uint16_t port_id);
2680 int rte_eth_link_get(uint16_t port_id, struct rte_eth_link *link);
2695 int rte_eth_link_get_nowait(uint16_t port_id, struct rte_eth_link *link);
2752 int rte_eth_stats_get(uint16_t port_id, struct rte_eth_stats *stats);
2765 int rte_eth_stats_reset(uint16_t port_id);
2796 int rte_eth_xstats_get_names(uint16_t port_id,
2829 int rte_eth_xstats_get(uint16_t port_id, struct rte_eth_xstat *xstats,
2855 rte_eth_xstats_get_names_by_id(uint16_t port_id,
2882 int rte_eth_xstats_get_by_id(uint16_t port_id, const uint64_t *ids,
2903 int rte_eth_xstats_get_id_by_name(uint16_t port_id, const char *xstat_name,
2918 int rte_eth_xstats_reset(uint16_t port_id);
2938 int rte_eth_dev_set_tx_queue_stats_mapping(uint16_t port_id,
2939 uint16_t tx_queue_id, uint8_t stat_idx);
2959 int rte_eth_dev_set_rx_queue_stats_mapping(uint16_t port_id,
2960 uint16_t rx_queue_id,
2975 int rte_eth_macaddr_get(uint16_t port_id, struct rte_ether_addr *mac_addr);
3019 int rte_eth_dev_info_get(uint16_t port_id, struct rte_eth_dev_info *dev_info);
3040 int rte_eth_dev_fw_version_get(uint16_t port_id,
3081 int rte_eth_dev_get_supported_ptypes(uint16_t port_id, uint32_t ptype_mask,
3117 int rte_eth_dev_set_ptypes(uint16_t port_id, uint32_t ptype_mask,
3131 int rte_eth_dev_get_mtu(uint16_t port_id, uint16_t *mtu);
3150 int rte_eth_dev_set_mtu(uint16_t port_id, uint16_t mtu);
3171 int rte_eth_dev_vlan_filter(uint16_t port_id, uint16_t vlan_id, int on);
3191 int rte_eth_dev_set_vlan_strip_on_queue(uint16_t port_id, uint16_t rx_queue_id,
3210 int rte_eth_dev_set_vlan_ether_type(uint16_t port_id,
3212 uint16_t tag_type);
3231 int rte_eth_dev_set_vlan_offload(uint16_t port_id, int offload_mask);
3246 int rte_eth_dev_get_vlan_offload(uint16_t port_id);
3262 int rte_eth_dev_set_vlan_pvid(uint16_t port_id, uint16_t pvid, int on);
3264 typedef void (*buffer_tx_error_fn)(struct rte_mbuf **unsent, uint16_t count,
3274 uint16_t size; /**< Size of buffer for buffered tx */
3275 uint16_t length; /**< Number of packets in the array */
3300 rte_eth_tx_buffer_init(struct rte_eth_dev_tx_buffer *buffer, uint16_t size);
3353 rte_eth_tx_buffer_drop_callback(struct rte_mbuf **pkts, uint16_t unsent,
3380 rte_eth_tx_buffer_count_callback(struct rte_mbuf **pkts, uint16_t unsent,
3409 rte_eth_tx_done_cleanup(uint16_t port_id, uint16_t queue_id, uint32_t free_cnt);
3471 typedef int (*rte_eth_dev_cb_fn)(uint16_t port_id,
3492 int rte_eth_dev_callback_register(uint16_t port_id,
3514 int rte_eth_dev_callback_unregister(uint16_t port_id,
3539 int rte_eth_dev_rx_intr_enable(uint16_t port_id, uint16_t queue_id);
3561 int rte_eth_dev_rx_intr_disable(uint16_t port_id, uint16_t queue_id);
3580 int rte_eth_dev_rx_intr_ctl(uint16_t port_id, int epfd, int op, void *data);
3603 int rte_eth_dev_rx_intr_ctl_q(uint16_t port_id, uint16_t queue_id,
3625 rte_eth_dev_rx_intr_ctl_q_get_fd(uint16_t port_id, uint16_t queue_id);
3640 int rte_eth_led_on(uint16_t port_id);
3655 int rte_eth_led_off(uint16_t port_id);
3686 int rte_eth_fec_get_capability(uint16_t port_id,
3714 int rte_eth_fec_get(uint16_t port_id, uint32_t *fec_capa);
3737 int rte_eth_fec_set(uint16_t port_id, uint32_t fec_capa);
3752 int rte_eth_dev_flow_ctrl_get(uint16_t port_id,
3769 int rte_eth_dev_flow_ctrl_set(uint16_t port_id,
3787 int rte_eth_dev_priority_flow_ctrl_set(uint16_t port_id,
3808 int rte_eth_dev_mac_addr_add(uint16_t port_id, struct rte_ether_addr *mac_addr,
3824 int rte_eth_dev_mac_addr_remove(uint16_t port_id,
3840 int rte_eth_dev_default_mac_addr_set(uint16_t port_id,
3860 int rte_eth_dev_rss_reta_update(uint16_t port_id,
3862 uint16_t reta_size);
3882 int rte_eth_dev_rss_reta_query(uint16_t port_id,
3884 uint16_t reta_size);
3905 int rte_eth_dev_uc_hash_table_set(uint16_t port_id, struct rte_ether_addr *addr,
3926 int rte_eth_dev_uc_all_hash_table_set(uint16_t port_id, uint8_t on);
3950 int rte_eth_mirror_rule_set(uint16_t port_id,
3969 int rte_eth_mirror_rule_reset(uint16_t port_id,
3988 int rte_eth_set_queue_rate_limit(uint16_t port_id, uint16_t queue_idx,
3989 uint16_t tx_rate);
4005 int rte_eth_dev_rss_hash_update(uint16_t port_id,
4023 rte_eth_dev_rss_hash_conf_get(uint16_t port_id,
4045 rte_eth_dev_udp_tunnel_port_add(uint16_t port_id,
4068 rte_eth_dev_udp_tunnel_port_delete(uint16_t port_id,
4084 int rte_eth_dev_get_dcb_info(uint16_t port_id,
4115 rte_eth_add_rx_callback(uint16_t port_id, uint16_t queue_id,
4145 rte_eth_add_first_rx_callback(uint16_t port_id, uint16_t queue_id,
4174 rte_eth_add_tx_callback(uint16_t port_id, uint16_t queue_id,
4210 int rte_eth_remove_rx_callback(uint16_t port_id, uint16_t queue_id,
4246 int rte_eth_remove_tx_callback(uint16_t port_id, uint16_t queue_id,
4268 int rte_eth_rx_queue_info_get(uint16_t port_id, uint16_t queue_id,
4290 int rte_eth_tx_queue_info_get(uint16_t port_id, uint16_t queue_id,
4312 int rte_eth_rx_burst_mode_get(uint16_t port_id, uint16_t queue_id,
4334 int rte_eth_tx_burst_mode_get(uint16_t port_id, uint16_t queue_id,
4354 int rte_eth_dev_get_reg_info(uint16_t port_id, struct rte_dev_reg_info *info);
4368 int rte_eth_dev_get_eeprom_length(uint16_t port_id);
4385 int rte_eth_dev_get_eeprom(uint16_t port_id, struct rte_dev_eeprom_info *info);
4402 int rte_eth_dev_set_eeprom(uint16_t port_id, struct rte_dev_eeprom_info *info);
4423 rte_eth_dev_get_module_info(uint16_t port_id,
4446 rte_eth_dev_get_module_eeprom(uint16_t port_id,
4467 int rte_eth_dev_set_mc_addr_list(uint16_t port_id,
4483 int rte_eth_timesync_enable(uint16_t port_id);
4497 int rte_eth_timesync_disable(uint16_t port_id);
4517 int rte_eth_timesync_read_rx_timestamp(uint16_t port_id,
4535 int rte_eth_timesync_read_tx_timestamp(uint16_t port_id,
4555 int rte_eth_timesync_adjust_time(uint16_t port_id, int64_t delta);
4571 int rte_eth_timesync_read_time(uint16_t port_id, struct timespec *time);
4591 int rte_eth_timesync_write_time(uint16_t port_id, const struct timespec *time);
4639 rte_eth_read_clock(uint16_t port_id, uint64_t *clock);
4657 rte_eth_dev_get_port_by_name(const char *name, uint16_t *port_id);
4675 rte_eth_dev_get_name_by_port(uint16_t port_id, char *name);
4693 int rte_eth_dev_adjust_nb_rx_tx_desc(uint16_t port_id,
4694 uint16_t *nb_rx_desc,
4695 uint16_t *nb_tx_desc);
4712 rte_eth_dev_pool_ops_supported(uint16_t port_id, const char *pool);
4724 rte_eth_dev_get_sec_ctx(uint16_t port_id);
4741 int rte_eth_dev_hairpin_capability_get(uint16_t port_id,
4833 static inline uint16_t
4834 rte_eth_rx_burst(uint16_t port_id, uint16_t queue_id, in rte_eth_rx_burst()
4835 struct rte_mbuf **rx_pkts, const uint16_t nb_pkts) in rte_eth_rx_burst()
4838 uint16_t nb_rx; in rte_eth_rx_burst()
4891 rte_eth_rx_queue_count(uint16_t port_id, uint16_t queue_id) in rte_eth_rx_queue_count()
4922 rte_eth_rx_descriptor_done(uint16_t port_id, uint16_t queue_id, uint16_t offset) in rte_eth_rx_descriptor_done()
4968 rte_eth_rx_descriptor_status(uint16_t port_id, uint16_t queue_id, in rte_eth_rx_descriptor_status()
4969 uint16_t offset) in rte_eth_rx_descriptor_status()
5025 static inline int rte_eth_tx_descriptor_status(uint16_t port_id, in rte_eth_tx_descriptor_status()
5026 uint16_t queue_id, uint16_t offset) in rte_eth_tx_descriptor_status()
5111 static inline uint16_t
5112 rte_eth_tx_burst(uint16_t port_id, uint16_t queue_id, in rte_eth_tx_burst()
5113 struct rte_mbuf **tx_pkts, uint16_t nb_pkts) in rte_eth_tx_burst()
5209 static inline uint16_t
5210 rte_eth_tx_prepare(uint16_t port_id, uint16_t queue_id, in rte_eth_tx_prepare()
5211 struct rte_mbuf **tx_pkts, uint16_t nb_pkts) in rte_eth_tx_prepare()
5251 static inline uint16_t
5252 rte_eth_tx_prepare(__rte_unused uint16_t port_id, in rte_eth_tx_prepare()
5253 __rte_unused uint16_t queue_id, in rte_eth_tx_prepare()
5254 __rte_unused struct rte_mbuf **tx_pkts, uint16_t nb_pkts) in rte_eth_tx_prepare()
5283 static inline uint16_t
5284 rte_eth_tx_buffer_flush(uint16_t port_id, uint16_t queue_id, in rte_eth_tx_buffer_flush()
5287 uint16_t sent; in rte_eth_tx_buffer_flush()
5288 uint16_t to_send = buffer->length; in rte_eth_tx_buffer_flush()
5300 (uint16_t)(to_send - sent), in rte_eth_tx_buffer_flush()
5336 static __rte_always_inline uint16_t
5337 rte_eth_tx_buffer(uint16_t port_id, uint16_t queue_id, in rte_eth_tx_buffer()