1 /* SPDX-License-Identifier: BSD-3-Clause 2 * Copyright(c) 2010-2016 Intel Corporation 3 */ 4 5 #ifndef _IXGBE_ETHDEV_H_ 6 #define _IXGBE_ETHDEV_H_ 7 8 #include <stdint.h> 9 10 #include "base/ixgbe_type.h" 11 #include "base/ixgbe_dcb.h" 12 #include "base/ixgbe_dcb_82599.h" 13 #include "base/ixgbe_dcb_82598.h" 14 #include "ixgbe_bypass.h" 15 #ifdef RTE_LIBRTE_SECURITY 16 #include "ixgbe_ipsec.h" 17 #endif 18 #include <rte_flow.h> 19 #include <rte_time.h> 20 #include <rte_hash.h> 21 #include <rte_pci.h> 22 #include <rte_bus_pci.h> 23 #include <rte_tm_driver.h> 24 25 /* need update link, bit flag */ 26 #define IXGBE_FLAG_NEED_LINK_UPDATE (uint32_t)(1 << 0) 27 #define IXGBE_FLAG_MAILBOX (uint32_t)(1 << 1) 28 #define IXGBE_FLAG_PHY_INTERRUPT (uint32_t)(1 << 2) 29 #define IXGBE_FLAG_MACSEC (uint32_t)(1 << 3) 30 #define IXGBE_FLAG_NEED_LINK_CONFIG (uint32_t)(1 << 4) 31 32 /* 33 * Defines that were not part of ixgbe_type.h as they are not used by the 34 * FreeBSD driver. 35 */ 36 #define IXGBE_ADVTXD_MAC_1588 0x00080000 /* IEEE1588 Timestamp packet */ 37 #define IXGBE_RXD_STAT_TMST 0x10000 /* Timestamped Packet indication */ 38 #define IXGBE_ADVTXD_TUCMD_L4T_RSV 0x00001800 /* L4 Packet TYPE, resvd */ 39 #define IXGBE_RXDADV_ERR_CKSUM_BIT 30 40 #define IXGBE_RXDADV_ERR_CKSUM_MSK 3 41 #define IXGBE_ADVTXD_MACLEN_SHIFT 9 /* Bit shift for l2_len */ 42 #define IXGBE_NB_STAT_MAPPING_REGS 32 43 #define IXGBE_EXTENDED_VLAN (uint32_t)(1 << 26) /* EXTENDED VLAN ENABLE */ 44 #define IXGBE_VFTA_SIZE 128 45 #define IXGBE_VLAN_TAG_SIZE 4 46 #define IXGBE_HKEY_MAX_INDEX 10 47 #define IXGBE_MAX_RX_QUEUE_NUM 128 48 #define IXGBE_MAX_INTR_QUEUE_NUM 15 49 #define IXGBE_VMDQ_DCB_NB_QUEUES IXGBE_MAX_RX_QUEUE_NUM 50 #define IXGBE_DCB_NB_QUEUES IXGBE_MAX_RX_QUEUE_NUM 51 #define IXGBE_NONE_MODE_TX_NB_QUEUES 64 52 53 #ifndef NBBY 54 #define NBBY 8 /* number of bits in a byte */ 55 #endif 56 #define IXGBE_HWSTRIP_BITMAP_SIZE (IXGBE_MAX_RX_QUEUE_NUM / (sizeof(uint32_t) * NBBY)) 57 58 /* EITR Interval is in 2048ns uinits for 1G and 10G link */ 59 #define IXGBE_EITR_INTERVAL_UNIT_NS 2048 60 #define IXGBE_EITR_ITR_INT_SHIFT 3 61 #define IXGBE_EITR_INTERVAL_US(us) \ 62 (((us) * 1000 / IXGBE_EITR_INTERVAL_UNIT_NS << IXGBE_EITR_ITR_INT_SHIFT) & \ 63 IXGBE_EITR_ITR_INT_MASK) 64 65 #define IXGBE_QUEUE_ITR_INTERVAL_DEFAULT 500 /* 500us */ 66 67 /* Loopback operation modes */ 68 #define IXGBE_LPBK_NONE 0x0 /* Default value. Loopback is disabled. */ 69 #define IXGBE_LPBK_TX_RX 0x1 /* Tx->Rx loopback operation is enabled. */ 70 /* X540-X550 specific loopback operations */ 71 #define IXGBE_MII_AUTONEG_ENABLE 0x1000 /* Auto-negociation enable (default = 1) */ 72 73 #define IXGBE_MAX_JUMBO_FRAME_SIZE 0x2600 /* Maximum Jumbo frame size. */ 74 75 #define IXGBE_RTTBCNRC_RF_INT_MASK_BASE 0x000003FF 76 #define IXGBE_RTTBCNRC_RF_INT_MASK_M \ 77 (IXGBE_RTTBCNRC_RF_INT_MASK_BASE << IXGBE_RTTBCNRC_RF_INT_SHIFT) 78 79 #define IXGBE_MAX_QUEUE_NUM_PER_VF 8 80 81 #define IXGBE_SYN_FILTER_ENABLE 0x00000001 /* syn filter enable field */ 82 #define IXGBE_SYN_FILTER_QUEUE 0x000000FE /* syn filter queue field */ 83 #define IXGBE_SYN_FILTER_QUEUE_SHIFT 1 /* syn filter queue field shift */ 84 #define IXGBE_SYN_FILTER_SYNQFP 0x80000000 /* syn filter SYNQFP */ 85 86 #define IXGBE_ETQF_UP 0x00070000 /* ethertype filter priority field */ 87 #define IXGBE_ETQF_SHIFT 16 88 #define IXGBE_ETQF_UP_EN 0x00080000 89 #define IXGBE_ETQF_ETHERTYPE 0x0000FFFF /* ethertype filter ethertype field */ 90 #define IXGBE_ETQF_MAX_PRI 7 91 92 #define IXGBE_SDPQF_DSTPORT 0xFFFF0000 /* dst port field */ 93 #define IXGBE_SDPQF_DSTPORT_SHIFT 16 /* dst port field shift */ 94 #define IXGBE_SDPQF_SRCPORT 0x0000FFFF /* src port field */ 95 96 #define IXGBE_L34T_IMIR_SIZE_BP 0x00001000 97 #define IXGBE_L34T_IMIR_RESERVE 0x00080000 /* bit 13 to 19 must be set to 1000000b. */ 98 #define IXGBE_L34T_IMIR_LLI 0x00100000 99 #define IXGBE_L34T_IMIR_QUEUE 0x0FE00000 100 #define IXGBE_L34T_IMIR_QUEUE_SHIFT 21 101 #define IXGBE_5TUPLE_MAX_PRI 7 102 #define IXGBE_5TUPLE_MIN_PRI 1 103 104 /* The overhead from MTU to max frame size. */ 105 #define IXGBE_ETH_OVERHEAD (RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN) 106 107 /* bit of VXLAN tunnel type | 7 bits of zeros | 8 bits of zeros*/ 108 #define IXGBE_FDIR_VXLAN_TUNNEL_TYPE 0x8000 109 /* bit of NVGRE tunnel type | 7 bits of zeros | 8 bits of zeros*/ 110 #define IXGBE_FDIR_NVGRE_TUNNEL_TYPE 0x0 111 112 #define IXGBE_RSS_OFFLOAD_ALL ( \ 113 ETH_RSS_IPV4 | \ 114 ETH_RSS_NONFRAG_IPV4_TCP | \ 115 ETH_RSS_NONFRAG_IPV4_UDP | \ 116 ETH_RSS_IPV6 | \ 117 ETH_RSS_NONFRAG_IPV6_TCP | \ 118 ETH_RSS_NONFRAG_IPV6_UDP | \ 119 ETH_RSS_IPV6_EX | \ 120 ETH_RSS_IPV6_TCP_EX | \ 121 ETH_RSS_IPV6_UDP_EX) 122 123 #define IXGBE_VF_IRQ_ENABLE_MASK 3 /* vf irq enable mask */ 124 #define IXGBE_VF_MAXMSIVECTOR 1 125 126 #define IXGBE_MISC_VEC_ID RTE_INTR_VEC_ZERO_OFFSET 127 #define IXGBE_RX_VEC_START RTE_INTR_VEC_RXTX_OFFSET 128 129 #define IXGBE_SECTX_MINSECIFG_MASK 0x0000000F 130 131 #define IXGBE_MACSEC_PNTHRSH 0xFFFFFE00 132 133 #define IXGBE_MAX_FDIR_FILTER_NUM (1024 * 32) 134 #define IXGBE_MAX_L2_TN_FILTER_NUM 128 135 136 #define MAC_TYPE_FILTER_SUP_EXT(type) do {\ 137 if ((type) != ixgbe_mac_82599EB && (type) != ixgbe_mac_X540)\ 138 return -ENOTSUP;\ 139 } while (0) 140 141 #define MAC_TYPE_FILTER_SUP(type) do {\ 142 if ((type) != ixgbe_mac_82599EB && (type) != ixgbe_mac_X540 &&\ 143 (type) != ixgbe_mac_X550 && (type) != ixgbe_mac_X550EM_x &&\ 144 (type) != ixgbe_mac_X550EM_a)\ 145 return -ENOTSUP;\ 146 } while (0) 147 148 /* Link speed for X550 auto negotiation */ 149 #define IXGBE_LINK_SPEED_X550_AUTONEG (IXGBE_LINK_SPEED_100_FULL | \ 150 IXGBE_LINK_SPEED_1GB_FULL | \ 151 IXGBE_LINK_SPEED_2_5GB_FULL | \ 152 IXGBE_LINK_SPEED_5GB_FULL | \ 153 IXGBE_LINK_SPEED_10GB_FULL) 154 155 /* 156 * Information about the fdir mode. 157 */ 158 struct ixgbe_hw_fdir_mask { 159 uint16_t vlan_tci_mask; 160 uint32_t src_ipv4_mask; 161 uint32_t dst_ipv4_mask; 162 uint16_t src_ipv6_mask; 163 uint16_t dst_ipv6_mask; 164 uint16_t src_port_mask; 165 uint16_t dst_port_mask; 166 uint16_t flex_bytes_mask; 167 uint8_t mac_addr_byte_mask; 168 uint32_t tunnel_id_mask; 169 uint8_t tunnel_type_mask; 170 }; 171 172 struct ixgbe_fdir_filter { 173 TAILQ_ENTRY(ixgbe_fdir_filter) entries; 174 union ixgbe_atr_input ixgbe_fdir; /* key of fdir filter*/ 175 uint32_t fdirflags; /* drop or forward */ 176 uint32_t fdirhash; /* hash value for fdir */ 177 uint8_t queue; /* assigned rx queue */ 178 }; 179 180 /* list of fdir filters */ 181 TAILQ_HEAD(ixgbe_fdir_filter_list, ixgbe_fdir_filter); 182 183 struct ixgbe_fdir_rule { 184 struct ixgbe_hw_fdir_mask mask; 185 union ixgbe_atr_input ixgbe_fdir; /* key of fdir filter*/ 186 bool b_spec; /* If TRUE, ixgbe_fdir, fdirflags, queue have meaning. */ 187 bool b_mask; /* If TRUE, mask has meaning. */ 188 enum rte_fdir_mode mode; /* IP, MAC VLAN, Tunnel */ 189 uint32_t fdirflags; /* drop or forward */ 190 uint32_t soft_id; /* an unique value for this rule */ 191 uint8_t queue; /* assigned rx queue */ 192 uint8_t flex_bytes_offset; 193 }; 194 195 struct ixgbe_hw_fdir_info { 196 struct ixgbe_hw_fdir_mask mask; 197 uint8_t flex_bytes_offset; 198 uint16_t collision; 199 uint16_t free; 200 uint16_t maxhash; 201 uint8_t maxlen; 202 uint64_t add; 203 uint64_t remove; 204 uint64_t f_add; 205 uint64_t f_remove; 206 struct ixgbe_fdir_filter_list fdir_list; /* filter list*/ 207 /* store the pointers of the filters, index is the hash value. */ 208 struct ixgbe_fdir_filter **hash_map; 209 struct rte_hash *hash_handle; /* cuckoo hash handler */ 210 bool mask_added; /* If already got mask from consistent filter */ 211 }; 212 213 struct ixgbe_rte_flow_rss_conf { 214 struct rte_flow_action_rss conf; /**< RSS parameters. */ 215 uint8_t key[IXGBE_HKEY_MAX_INDEX * sizeof(uint32_t)]; /* Hash key. */ 216 uint16_t queue[IXGBE_MAX_RX_QUEUE_NUM]; /**< Queues indices to use. */ 217 }; 218 219 /* structure for interrupt relative data */ 220 struct ixgbe_interrupt { 221 uint32_t flags; 222 uint32_t mask; 223 /*to save original mask during delayed handler */ 224 uint32_t mask_original; 225 }; 226 227 struct ixgbe_stat_mapping_registers { 228 uint32_t tqsm[IXGBE_NB_STAT_MAPPING_REGS]; 229 uint32_t rqsmr[IXGBE_NB_STAT_MAPPING_REGS]; 230 }; 231 232 struct ixgbe_vfta { 233 uint32_t vfta[IXGBE_VFTA_SIZE]; 234 }; 235 236 struct ixgbe_hwstrip { 237 uint32_t bitmap[IXGBE_HWSTRIP_BITMAP_SIZE]; 238 }; 239 240 /* 241 * VF data which used by PF host only 242 */ 243 #define IXGBE_MAX_VF_MC_ENTRIES 30 244 #define IXGBE_MAX_MR_RULE_ENTRIES 4 /* number of mirroring rules supported */ 245 #define IXGBE_MAX_UTA 128 246 247 struct ixgbe_uta_info { 248 uint8_t uc_filter_type; 249 uint16_t uta_in_use; 250 uint32_t uta_shadow[IXGBE_MAX_UTA]; 251 }; 252 253 #define IXGBE_MAX_MIRROR_RULES 4 /* Maximum nb. of mirror rules. */ 254 255 struct ixgbe_mirror_info { 256 struct rte_eth_mirror_conf mr_conf[IXGBE_MAX_MIRROR_RULES]; 257 /**< store PF mirror rules configuration*/ 258 }; 259 260 struct ixgbe_vf_info { 261 uint8_t vf_mac_addresses[RTE_ETHER_ADDR_LEN]; 262 uint16_t vf_mc_hashes[IXGBE_MAX_VF_MC_ENTRIES]; 263 uint16_t num_vf_mc_hashes; 264 uint16_t default_vf_vlan_id; 265 uint16_t vlans_enabled; 266 bool clear_to_send; 267 uint16_t tx_rate[IXGBE_MAX_QUEUE_NUM_PER_VF]; 268 uint16_t vlan_count; 269 uint8_t spoofchk_enabled; 270 uint8_t api_version; 271 uint16_t switch_domain_id; 272 uint16_t xcast_mode; 273 }; 274 275 /* 276 * Possible l4type of 5tuple filters. 277 */ 278 enum ixgbe_5tuple_protocol { 279 IXGBE_FILTER_PROTOCOL_TCP = 0, 280 IXGBE_FILTER_PROTOCOL_UDP, 281 IXGBE_FILTER_PROTOCOL_SCTP, 282 IXGBE_FILTER_PROTOCOL_NONE, 283 }; 284 285 TAILQ_HEAD(ixgbe_5tuple_filter_list, ixgbe_5tuple_filter); 286 287 struct ixgbe_5tuple_filter_info { 288 uint32_t dst_ip; 289 uint32_t src_ip; 290 uint16_t dst_port; 291 uint16_t src_port; 292 enum ixgbe_5tuple_protocol proto; /* l4 protocol. */ 293 uint8_t priority; /* seven levels (001b-111b), 111b is highest, 294 used when more than one filter matches. */ 295 uint8_t dst_ip_mask:1, /* if mask is 1b, do not compare dst ip. */ 296 src_ip_mask:1, /* if mask is 1b, do not compare src ip. */ 297 dst_port_mask:1, /* if mask is 1b, do not compare dst port. */ 298 src_port_mask:1, /* if mask is 1b, do not compare src port. */ 299 proto_mask:1; /* if mask is 1b, do not compare protocol. */ 300 }; 301 302 /* 5tuple filter structure */ 303 struct ixgbe_5tuple_filter { 304 TAILQ_ENTRY(ixgbe_5tuple_filter) entries; 305 uint16_t index; /* the index of 5tuple filter */ 306 struct ixgbe_5tuple_filter_info filter_info; 307 uint16_t queue; /* rx queue assigned to */ 308 }; 309 310 #define IXGBE_5TUPLE_ARRAY_SIZE \ 311 (RTE_ALIGN(IXGBE_MAX_FTQF_FILTERS, (sizeof(uint32_t) * NBBY)) / \ 312 (sizeof(uint32_t) * NBBY)) 313 314 struct ixgbe_ethertype_filter { 315 uint16_t ethertype; 316 uint32_t etqf; 317 uint32_t etqs; 318 /** 319 * If this filter is added by configuration, 320 * it should not be removed. 321 */ 322 bool conf; 323 }; 324 325 /* 326 * Structure to store filters' info. 327 */ 328 struct ixgbe_filter_info { 329 uint8_t ethertype_mask; /* Bit mask for every used ethertype filter */ 330 /* store used ethertype filters*/ 331 struct ixgbe_ethertype_filter ethertype_filters[IXGBE_MAX_ETQF_FILTERS]; 332 /* Bit mask for every used 5tuple filter */ 333 uint32_t fivetuple_mask[IXGBE_5TUPLE_ARRAY_SIZE]; 334 struct ixgbe_5tuple_filter_list fivetuple_list; 335 /* store the SYN filter info */ 336 uint32_t syn_info; 337 /* store the rss filter info */ 338 struct ixgbe_rte_flow_rss_conf rss_info; 339 }; 340 341 struct ixgbe_l2_tn_key { 342 enum rte_eth_tunnel_type l2_tn_type; 343 uint32_t tn_id; 344 }; 345 346 struct ixgbe_l2_tn_filter { 347 TAILQ_ENTRY(ixgbe_l2_tn_filter) entries; 348 struct ixgbe_l2_tn_key key; 349 uint32_t pool; 350 }; 351 352 TAILQ_HEAD(ixgbe_l2_tn_filter_list, ixgbe_l2_tn_filter); 353 354 struct ixgbe_l2_tn_info { 355 struct ixgbe_l2_tn_filter_list l2_tn_list; 356 struct ixgbe_l2_tn_filter **hash_map; 357 struct rte_hash *hash_handle; 358 bool e_tag_en; /* e-tag enabled */ 359 bool e_tag_fwd_en; /* e-tag based forwarding enabled */ 360 bool e_tag_ether_type; /* ether type for e-tag */ 361 }; 362 363 struct rte_flow { 364 enum rte_filter_type filter_type; 365 void *rule; 366 }; 367 368 struct ixgbe_macsec_setting { 369 uint8_t offload_en; 370 uint8_t encrypt_en; 371 uint8_t replayprotect_en; 372 }; 373 374 /* 375 * Statistics counters collected by the MACsec 376 */ 377 struct ixgbe_macsec_stats { 378 /* TX port statistics */ 379 uint64_t out_pkts_untagged; 380 uint64_t out_pkts_encrypted; 381 uint64_t out_pkts_protected; 382 uint64_t out_octets_encrypted; 383 uint64_t out_octets_protected; 384 385 /* RX port statistics */ 386 uint64_t in_pkts_untagged; 387 uint64_t in_pkts_badtag; 388 uint64_t in_pkts_nosci; 389 uint64_t in_pkts_unknownsci; 390 uint64_t in_octets_decrypted; 391 uint64_t in_octets_validated; 392 393 /* RX SC statistics */ 394 uint64_t in_pkts_unchecked; 395 uint64_t in_pkts_delayed; 396 uint64_t in_pkts_late; 397 398 /* RX SA statistics */ 399 uint64_t in_pkts_ok; 400 uint64_t in_pkts_invalid; 401 uint64_t in_pkts_notvalid; 402 uint64_t in_pkts_unusedsa; 403 uint64_t in_pkts_notusingsa; 404 }; 405 406 /* The configuration of bandwidth */ 407 struct ixgbe_bw_conf { 408 uint8_t tc_num; /* Number of TCs. */ 409 }; 410 411 /* Struct to store Traffic Manager shaper profile. */ 412 struct ixgbe_tm_shaper_profile { 413 TAILQ_ENTRY(ixgbe_tm_shaper_profile) node; 414 uint32_t shaper_profile_id; 415 uint32_t reference_count; 416 struct rte_tm_shaper_params profile; 417 }; 418 419 TAILQ_HEAD(ixgbe_shaper_profile_list, ixgbe_tm_shaper_profile); 420 421 /* node type of Traffic Manager */ 422 enum ixgbe_tm_node_type { 423 IXGBE_TM_NODE_TYPE_PORT, 424 IXGBE_TM_NODE_TYPE_TC, 425 IXGBE_TM_NODE_TYPE_QUEUE, 426 IXGBE_TM_NODE_TYPE_MAX, 427 }; 428 429 /* Struct to store Traffic Manager node configuration. */ 430 struct ixgbe_tm_node { 431 TAILQ_ENTRY(ixgbe_tm_node) node; 432 uint32_t id; 433 uint32_t priority; 434 uint32_t weight; 435 uint32_t reference_count; 436 uint16_t no; 437 struct ixgbe_tm_node *parent; 438 struct ixgbe_tm_shaper_profile *shaper_profile; 439 struct rte_tm_node_params params; 440 }; 441 442 TAILQ_HEAD(ixgbe_tm_node_list, ixgbe_tm_node); 443 444 /* The configuration of Traffic Manager */ 445 struct ixgbe_tm_conf { 446 struct ixgbe_shaper_profile_list shaper_profile_list; 447 struct ixgbe_tm_node *root; /* root node - port */ 448 struct ixgbe_tm_node_list tc_list; /* node list for all the TCs */ 449 struct ixgbe_tm_node_list queue_list; /* node list for all the queues */ 450 /** 451 * The number of added TC nodes. 452 * It should be no more than the TC number of this port. 453 */ 454 uint32_t nb_tc_node; 455 /** 456 * The number of added queue nodes. 457 * It should be no more than the queue number of this port. 458 */ 459 uint32_t nb_queue_node; 460 /** 461 * This flag is used to check if APP can change the TM node 462 * configuration. 463 * When it's true, means the configuration is applied to HW, 464 * APP should not change the configuration. 465 * As we don't support on-the-fly configuration, when starting 466 * the port, APP should call the hierarchy_commit API to set this 467 * flag to true. When stopping the port, this flag should be set 468 * to false. 469 */ 470 bool committed; 471 }; 472 473 /* 474 * Structure to store private data for each driver instance (for each port). 475 */ 476 struct ixgbe_adapter { 477 struct ixgbe_hw hw; 478 struct ixgbe_hw_stats stats; 479 struct ixgbe_macsec_stats macsec_stats; 480 struct ixgbe_macsec_setting macsec_setting; 481 struct ixgbe_hw_fdir_info fdir; 482 struct ixgbe_interrupt intr; 483 struct ixgbe_stat_mapping_registers stat_mappings; 484 struct ixgbe_vfta shadow_vfta; 485 struct ixgbe_hwstrip hwstrip; 486 struct ixgbe_dcb_config dcb_config; 487 struct ixgbe_mirror_info mr_data; 488 struct ixgbe_vf_info *vfdata; 489 struct ixgbe_uta_info uta_info; 490 #ifdef RTE_LIBRTE_IXGBE_BYPASS 491 struct ixgbe_bypass_info bps; 492 #endif /* RTE_LIBRTE_IXGBE_BYPASS */ 493 struct ixgbe_filter_info filter; 494 struct ixgbe_l2_tn_info l2_tn; 495 struct ixgbe_bw_conf bw_conf; 496 #ifdef RTE_LIBRTE_SECURITY 497 struct ixgbe_ipsec ipsec; 498 #endif 499 bool rx_bulk_alloc_allowed; 500 bool rx_vec_allowed; 501 struct rte_timecounter systime_tc; 502 struct rte_timecounter rx_tstamp_tc; 503 struct rte_timecounter tx_tstamp_tc; 504 struct ixgbe_tm_conf tm_conf; 505 506 /* For RSS reta table update */ 507 uint8_t rss_reta_updated; 508 509 /* Used for VF link sync with PF's physical and logical (by checking 510 * mailbox status) link status. 511 */ 512 uint8_t pflink_fullchk; 513 uint8_t mac_ctrl_frame_fwd; 514 rte_atomic32_t link_thread_running; 515 pthread_t link_thread_tid; 516 }; 517 518 struct ixgbe_vf_representor { 519 uint16_t vf_id; 520 uint16_t switch_domain_id; 521 struct rte_eth_dev *pf_ethdev; 522 }; 523 524 int ixgbe_vf_representor_init(struct rte_eth_dev *ethdev, void *init_params); 525 int ixgbe_vf_representor_uninit(struct rte_eth_dev *ethdev); 526 527 #define IXGBE_DEV_PRIVATE_TO_HW(adapter)\ 528 (&((struct ixgbe_adapter *)adapter)->hw) 529 530 #define IXGBE_DEV_PRIVATE_TO_STATS(adapter) \ 531 (&((struct ixgbe_adapter *)adapter)->stats) 532 533 #define IXGBE_DEV_PRIVATE_TO_MACSEC_STATS(adapter) \ 534 (&((struct ixgbe_adapter *)adapter)->macsec_stats) 535 536 #define IXGBE_DEV_PRIVATE_TO_MACSEC_SETTING(adapter) \ 537 (&((struct ixgbe_adapter *)adapter)->macsec_setting) 538 539 #define IXGBE_DEV_PRIVATE_TO_INTR(adapter) \ 540 (&((struct ixgbe_adapter *)adapter)->intr) 541 542 #define IXGBE_DEV_PRIVATE_TO_FDIR_INFO(adapter) \ 543 (&((struct ixgbe_adapter *)adapter)->fdir) 544 545 #define IXGBE_DEV_PRIVATE_TO_STAT_MAPPINGS(adapter) \ 546 (&((struct ixgbe_adapter *)adapter)->stat_mappings) 547 548 #define IXGBE_DEV_PRIVATE_TO_VFTA(adapter) \ 549 (&((struct ixgbe_adapter *)adapter)->shadow_vfta) 550 551 #define IXGBE_DEV_PRIVATE_TO_HWSTRIP_BITMAP(adapter) \ 552 (&((struct ixgbe_adapter *)adapter)->hwstrip) 553 554 #define IXGBE_DEV_PRIVATE_TO_DCB_CFG(adapter) \ 555 (&((struct ixgbe_adapter *)adapter)->dcb_config) 556 557 #define IXGBE_DEV_PRIVATE_TO_P_VFDATA(adapter) \ 558 (&((struct ixgbe_adapter *)adapter)->vfdata) 559 560 #define IXGBE_DEV_PRIVATE_TO_PFDATA(adapter) \ 561 (&((struct ixgbe_adapter *)adapter)->mr_data) 562 563 #define IXGBE_DEV_PRIVATE_TO_UTA(adapter) \ 564 (&((struct ixgbe_adapter *)adapter)->uta_info) 565 566 #define IXGBE_DEV_PRIVATE_TO_FILTER_INFO(adapter) \ 567 (&((struct ixgbe_adapter *)adapter)->filter) 568 569 #define IXGBE_DEV_PRIVATE_TO_L2_TN_INFO(adapter) \ 570 (&((struct ixgbe_adapter *)adapter)->l2_tn) 571 572 #define IXGBE_DEV_PRIVATE_TO_BW_CONF(adapter) \ 573 (&((struct ixgbe_adapter *)adapter)->bw_conf) 574 575 #define IXGBE_DEV_PRIVATE_TO_TM_CONF(adapter) \ 576 (&((struct ixgbe_adapter *)adapter)->tm_conf) 577 578 #define IXGBE_DEV_PRIVATE_TO_IPSEC(adapter)\ 579 (&((struct ixgbe_adapter *)adapter)->ipsec) 580 581 /* 582 * RX/TX function prototypes 583 */ 584 void ixgbe_dev_clear_queues(struct rte_eth_dev *dev); 585 586 void ixgbe_dev_free_queues(struct rte_eth_dev *dev); 587 588 void ixgbe_dev_rx_queue_release(void *rxq); 589 590 void ixgbe_dev_tx_queue_release(void *txq); 591 592 int ixgbe_dev_rx_queue_setup(struct rte_eth_dev *dev, uint16_t rx_queue_id, 593 uint16_t nb_rx_desc, unsigned int socket_id, 594 const struct rte_eth_rxconf *rx_conf, 595 struct rte_mempool *mb_pool); 596 597 int ixgbe_dev_tx_queue_setup(struct rte_eth_dev *dev, uint16_t tx_queue_id, 598 uint16_t nb_tx_desc, unsigned int socket_id, 599 const struct rte_eth_txconf *tx_conf); 600 601 uint32_t ixgbe_dev_rx_queue_count(struct rte_eth_dev *dev, 602 uint16_t rx_queue_id); 603 604 int ixgbe_dev_rx_descriptor_done(void *rx_queue, uint16_t offset); 605 606 int ixgbe_dev_rx_descriptor_status(void *rx_queue, uint16_t offset); 607 int ixgbe_dev_tx_descriptor_status(void *tx_queue, uint16_t offset); 608 609 int ixgbe_dev_rx_init(struct rte_eth_dev *dev); 610 611 void ixgbe_dev_tx_init(struct rte_eth_dev *dev); 612 613 int ixgbe_dev_rxtx_start(struct rte_eth_dev *dev); 614 615 int ixgbe_dev_rx_queue_start(struct rte_eth_dev *dev, uint16_t rx_queue_id); 616 617 int ixgbe_dev_rx_queue_stop(struct rte_eth_dev *dev, uint16_t rx_queue_id); 618 619 int ixgbe_dev_tx_queue_start(struct rte_eth_dev *dev, uint16_t tx_queue_id); 620 621 int ixgbe_dev_tx_queue_stop(struct rte_eth_dev *dev, uint16_t tx_queue_id); 622 623 void ixgbe_rxq_info_get(struct rte_eth_dev *dev, uint16_t queue_id, 624 struct rte_eth_rxq_info *qinfo); 625 626 void ixgbe_txq_info_get(struct rte_eth_dev *dev, uint16_t queue_id, 627 struct rte_eth_txq_info *qinfo); 628 629 int ixgbevf_dev_rx_init(struct rte_eth_dev *dev); 630 631 void ixgbevf_dev_tx_init(struct rte_eth_dev *dev); 632 633 void ixgbevf_dev_rxtx_start(struct rte_eth_dev *dev); 634 635 uint16_t ixgbe_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, 636 uint16_t nb_pkts); 637 638 uint16_t ixgbe_recv_pkts_bulk_alloc(void *rx_queue, struct rte_mbuf **rx_pkts, 639 uint16_t nb_pkts); 640 641 uint16_t ixgbe_recv_pkts_lro_single_alloc(void *rx_queue, 642 struct rte_mbuf **rx_pkts, uint16_t nb_pkts); 643 uint16_t ixgbe_recv_pkts_lro_bulk_alloc(void *rx_queue, 644 struct rte_mbuf **rx_pkts, uint16_t nb_pkts); 645 646 uint16_t ixgbe_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, 647 uint16_t nb_pkts); 648 649 uint16_t ixgbe_xmit_pkts_simple(void *tx_queue, struct rte_mbuf **tx_pkts, 650 uint16_t nb_pkts); 651 652 uint16_t ixgbe_prep_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, 653 uint16_t nb_pkts); 654 655 int ixgbe_dev_rss_hash_update(struct rte_eth_dev *dev, 656 struct rte_eth_rss_conf *rss_conf); 657 658 int ixgbe_dev_rss_hash_conf_get(struct rte_eth_dev *dev, 659 struct rte_eth_rss_conf *rss_conf); 660 661 uint16_t ixgbe_reta_size_get(enum ixgbe_mac_type mac_type); 662 663 uint32_t ixgbe_reta_reg_get(enum ixgbe_mac_type mac_type, uint16_t reta_idx); 664 665 uint32_t ixgbe_mrqc_reg_get(enum ixgbe_mac_type mac_type); 666 667 uint32_t ixgbe_rssrk_reg_get(enum ixgbe_mac_type mac_type, uint8_t i); 668 669 bool ixgbe_rss_update_sp(enum ixgbe_mac_type mac_type); 670 671 int ixgbe_add_del_ntuple_filter(struct rte_eth_dev *dev, 672 struct rte_eth_ntuple_filter *filter, 673 bool add); 674 int ixgbe_add_del_ethertype_filter(struct rte_eth_dev *dev, 675 struct rte_eth_ethertype_filter *filter, 676 bool add); 677 int ixgbe_syn_filter_set(struct rte_eth_dev *dev, 678 struct rte_eth_syn_filter *filter, 679 bool add); 680 int 681 ixgbe_dev_l2_tunnel_filter_add(struct rte_eth_dev *dev, 682 struct rte_eth_l2_tunnel_conf *l2_tunnel, 683 bool restore); 684 int 685 ixgbe_dev_l2_tunnel_filter_del(struct rte_eth_dev *dev, 686 struct rte_eth_l2_tunnel_conf *l2_tunnel); 687 void ixgbe_filterlist_init(void); 688 void ixgbe_filterlist_flush(void); 689 /* 690 * Flow director function prototypes 691 */ 692 int ixgbe_fdir_configure(struct rte_eth_dev *dev); 693 int ixgbe_fdir_set_input_mask(struct rte_eth_dev *dev); 694 int ixgbe_fdir_set_flexbytes_offset(struct rte_eth_dev *dev, 695 uint16_t offset); 696 int ixgbe_fdir_filter_program(struct rte_eth_dev *dev, 697 struct ixgbe_fdir_rule *rule, 698 bool del, bool update); 699 700 void ixgbe_configure_dcb(struct rte_eth_dev *dev); 701 702 int 703 ixgbe_dev_link_update_share(struct rte_eth_dev *dev, 704 int wait_to_complete, int vf); 705 706 /* 707 * misc function prototypes 708 */ 709 void ixgbe_vlan_hw_filter_enable(struct rte_eth_dev *dev); 710 711 void ixgbe_vlan_hw_filter_disable(struct rte_eth_dev *dev); 712 713 void ixgbe_vlan_hw_strip_config(struct rte_eth_dev *dev); 714 715 void ixgbe_pf_host_init(struct rte_eth_dev *eth_dev); 716 717 void ixgbe_pf_host_uninit(struct rte_eth_dev *eth_dev); 718 719 void ixgbe_pf_mbx_process(struct rte_eth_dev *eth_dev); 720 721 int ixgbe_pf_host_configure(struct rte_eth_dev *eth_dev); 722 723 uint32_t ixgbe_convert_vm_rx_mask_to_val(uint16_t rx_mask, uint32_t orig_val); 724 725 int ixgbe_fdir_ctrl_func(struct rte_eth_dev *dev, 726 enum rte_filter_op filter_op, void *arg); 727 void ixgbe_fdir_filter_restore(struct rte_eth_dev *dev); 728 int ixgbe_clear_all_fdir_filter(struct rte_eth_dev *dev); 729 730 extern const struct rte_flow_ops ixgbe_flow_ops; 731 732 void ixgbe_clear_all_ethertype_filter(struct rte_eth_dev *dev); 733 void ixgbe_clear_all_ntuple_filter(struct rte_eth_dev *dev); 734 void ixgbe_clear_syn_filter(struct rte_eth_dev *dev); 735 int ixgbe_clear_all_l2_tn_filter(struct rte_eth_dev *dev); 736 737 int ixgbe_disable_sec_tx_path_generic(struct ixgbe_hw *hw); 738 739 int ixgbe_enable_sec_tx_path_generic(struct ixgbe_hw *hw); 740 741 int ixgbe_vt_check(struct ixgbe_hw *hw); 742 int ixgbe_set_vf_rate_limit(struct rte_eth_dev *dev, uint16_t vf, 743 uint16_t tx_rate, uint64_t q_msk); 744 bool is_ixgbe_supported(struct rte_eth_dev *dev); 745 int ixgbe_tm_ops_get(struct rte_eth_dev *dev, void *ops); 746 void ixgbe_tm_conf_init(struct rte_eth_dev *dev); 747 void ixgbe_tm_conf_uninit(struct rte_eth_dev *dev); 748 int ixgbe_set_queue_rate_limit(struct rte_eth_dev *dev, uint16_t queue_idx, 749 uint16_t tx_rate); 750 int ixgbe_rss_conf_init(struct ixgbe_rte_flow_rss_conf *out, 751 const struct rte_flow_action_rss *in); 752 int ixgbe_action_rss_same(const struct rte_flow_action_rss *comp, 753 const struct rte_flow_action_rss *with); 754 int ixgbe_config_rss_filter(struct rte_eth_dev *dev, 755 struct ixgbe_rte_flow_rss_conf *conf, bool add); 756 757 void ixgbe_dev_macsec_register_enable(struct rte_eth_dev *dev, 758 struct ixgbe_macsec_setting *macsec_setting); 759 760 void ixgbe_dev_macsec_register_disable(struct rte_eth_dev *dev); 761 762 void ixgbe_dev_macsec_setting_save(struct rte_eth_dev *dev, 763 struct ixgbe_macsec_setting *macsec_setting); 764 765 void ixgbe_dev_macsec_setting_reset(struct rte_eth_dev *dev); 766 767 static inline int 768 ixgbe_ethertype_filter_lookup(struct ixgbe_filter_info *filter_info, 769 uint16_t ethertype) 770 { 771 int i; 772 773 for (i = 0; i < IXGBE_MAX_ETQF_FILTERS; i++) { 774 if (filter_info->ethertype_filters[i].ethertype == ethertype && 775 (filter_info->ethertype_mask & (1 << i))) 776 return i; 777 } 778 return -1; 779 } 780 781 static inline int 782 ixgbe_ethertype_filter_insert(struct ixgbe_filter_info *filter_info, 783 struct ixgbe_ethertype_filter *ethertype_filter) 784 { 785 int i; 786 787 for (i = 0; i < IXGBE_MAX_ETQF_FILTERS; i++) { 788 if (!(filter_info->ethertype_mask & (1 << i))) { 789 filter_info->ethertype_mask |= 1 << i; 790 filter_info->ethertype_filters[i].ethertype = 791 ethertype_filter->ethertype; 792 filter_info->ethertype_filters[i].etqf = 793 ethertype_filter->etqf; 794 filter_info->ethertype_filters[i].etqs = 795 ethertype_filter->etqs; 796 filter_info->ethertype_filters[i].conf = 797 ethertype_filter->conf; 798 return i; 799 } 800 } 801 return -1; 802 } 803 804 static inline int 805 ixgbe_ethertype_filter_remove(struct ixgbe_filter_info *filter_info, 806 uint8_t idx) 807 { 808 if (idx >= IXGBE_MAX_ETQF_FILTERS) 809 return -1; 810 filter_info->ethertype_mask &= ~(1 << idx); 811 filter_info->ethertype_filters[idx].ethertype = 0; 812 filter_info->ethertype_filters[idx].etqf = 0; 813 filter_info->ethertype_filters[idx].etqs = 0; 814 filter_info->ethertype_filters[idx].etqs = FALSE; 815 return idx; 816 } 817 818 #endif /* _IXGBE_ETHDEV_H_ */ 819