xref: /f-stack/dpdk/drivers/net/e1000/e1000_ethdev.h (revision 2d9fd380)
1 /* SPDX-License-Identifier: BSD-3-Clause
2  * Copyright(c) 2010-2015 Intel Corporation
3  */
4 
5 #ifndef _E1000_ETHDEV_H_
6 #define _E1000_ETHDEV_H_
7 
8 #include <stdint.h>
9 
10 #include <rte_flow.h>
11 #include <rte_time.h>
12 #include <rte_pci.h>
13 
14 #define E1000_INTEL_VENDOR_ID 0x8086
15 
16 /* need update link, bit flag */
17 #define E1000_FLAG_NEED_LINK_UPDATE (uint32_t)(1 << 0)
18 #define E1000_FLAG_MAILBOX          (uint32_t)(1 << 1)
19 
20 /*
21  * Defines that were not part of e1000_hw.h as they are not used by the FreeBSD
22  * driver.
23  */
24 #define E1000_ADVTXD_POPTS_TXSM     0x00000200 /* L4 Checksum offload request */
25 #define E1000_ADVTXD_POPTS_IXSM     0x00000100 /* IP Checksum offload request */
26 #define E1000_ADVTXD_TUCMD_L4T_RSV  0x00001800 /* L4 Packet TYPE of Reserved */
27 #define E1000_RXD_STAT_TMST         0x10000    /* Timestamped Packet indication */
28 #define E1000_RXD_ERR_CKSUM_BIT     29
29 #define E1000_RXD_ERR_CKSUM_MSK     3
30 #define E1000_ADVTXD_MACLEN_SHIFT   9          /* Bit shift for l2_len */
31 #define E1000_CTRL_EXT_EXTEND_VLAN  (1<<26)    /* EXTENDED VLAN */
32 #define IGB_VFTA_SIZE 128
33 
34 #define IGB_HKEY_MAX_INDEX             10
35 #define IGB_MAX_RX_QUEUE_NUM           8
36 #define IGB_MAX_RX_QUEUE_NUM_82576     16
37 
38 #define E1000_I219_MAX_RX_QUEUE_NUM		2
39 #define E1000_I219_MAX_TX_QUEUE_NUM		2
40 
41 #define E1000_SYN_FILTER_ENABLE        0x00000001 /* syn filter enable field */
42 #define E1000_SYN_FILTER_QUEUE         0x0000000E /* syn filter queue field */
43 #define E1000_SYN_FILTER_QUEUE_SHIFT   1          /* syn filter queue field */
44 #define E1000_RFCTL_SYNQFP             0x00080000 /* SYNQFP in RFCTL register */
45 
46 #define E1000_ETQF_ETHERTYPE           0x0000FFFF
47 #define E1000_ETQF_QUEUE               0x00070000
48 #define E1000_ETQF_QUEUE_SHIFT         16
49 #define E1000_MAX_ETQF_FILTERS         8
50 
51 #define E1000_IMIR_DSTPORT             0x0000FFFF
52 #define E1000_IMIR_PRIORITY            0xE0000000
53 #define E1000_MAX_TTQF_FILTERS         8
54 #define E1000_2TUPLE_MAX_PRI           7
55 
56 #define E1000_MAX_FLEX_FILTERS           8
57 #define E1000_MAX_FHFT                   4
58 #define E1000_MAX_FHFT_EXT               4
59 #define E1000_FHFT_SIZE_IN_DWD           64
60 #define E1000_MAX_FLEX_FILTER_PRI        7
61 #define E1000_MAX_FLEX_FILTER_LEN        128
62 #define E1000_MAX_FLEX_FILTER_DWDS \
63 	(E1000_MAX_FLEX_FILTER_LEN / sizeof(uint32_t))
64 #define E1000_FLEX_FILTERS_MASK_SIZE \
65 	(E1000_MAX_FLEX_FILTER_DWDS / 2)
66 #define E1000_FHFT_QUEUEING_LEN          0x0000007F
67 #define E1000_FHFT_QUEUEING_QUEUE        0x00000700
68 #define E1000_FHFT_QUEUEING_PRIO         0x00070000
69 #define E1000_FHFT_QUEUEING_OFFSET       0xFC
70 #define E1000_FHFT_QUEUEING_QUEUE_SHIFT  8
71 #define E1000_FHFT_QUEUEING_PRIO_SHIFT   16
72 #define E1000_WUFC_FLEX_HQ               0x00004000
73 
74 #define E1000_SPQF_SRCPORT               0x0000FFFF
75 
76 #define E1000_MAX_FTQF_FILTERS           8
77 #define E1000_FTQF_PROTOCOL_MASK         0x000000FF
78 #define E1000_FTQF_5TUPLE_MASK_SHIFT     28
79 #define E1000_FTQF_QUEUE_MASK            0x03ff0000
80 #define E1000_FTQF_QUEUE_SHIFT           16
81 #define E1000_FTQF_QUEUE_ENABLE          0x00000100
82 
83 #define IGB_RSS_OFFLOAD_ALL ( \
84 	ETH_RSS_IPV4 | \
85 	ETH_RSS_NONFRAG_IPV4_TCP | \
86 	ETH_RSS_NONFRAG_IPV4_UDP | \
87 	ETH_RSS_IPV6 | \
88 	ETH_RSS_NONFRAG_IPV6_TCP | \
89 	ETH_RSS_NONFRAG_IPV6_UDP | \
90 	ETH_RSS_IPV6_EX | \
91 	ETH_RSS_IPV6_TCP_EX | \
92 	ETH_RSS_IPV6_UDP_EX)
93 
94 /*
95  * The overhead from MTU to max frame size.
96  * Considering VLAN so a tag needs to be counted.
97  */
98 #define E1000_ETH_OVERHEAD (RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + \
99 				VLAN_TAG_SIZE)
100 
101 /*
102  * Maximum number of Ring Descriptors.
103  *
104  * Since RDLEN/TDLEN should be multiple of 128 bytes, the number of ring
105  * desscriptors should meet the following condition:
106  * (num_ring_desc * sizeof(struct e1000_rx/tx_desc)) % 128 == 0
107  */
108 #define	E1000_MIN_RING_DESC	32
109 #define	E1000_MAX_RING_DESC	4096
110 
111 /*
112  * TDBA/RDBA should be aligned on 16 byte boundary. But TDLEN/RDLEN should be
113  * multiple of 128 bytes. So we align TDBA/RDBA on 128 byte boundary.
114  * This will also optimize cache line size effect.
115  * H/W supports up to cache line size 128.
116  */
117 #define	E1000_ALIGN	128
118 
119 #define	IGB_RXD_ALIGN	(E1000_ALIGN / sizeof(union e1000_adv_rx_desc))
120 #define	IGB_TXD_ALIGN	(E1000_ALIGN / sizeof(union e1000_adv_tx_desc))
121 
122 #define	EM_RXD_ALIGN	(E1000_ALIGN / sizeof(struct e1000_rx_desc))
123 #define	EM_TXD_ALIGN	(E1000_ALIGN / sizeof(struct e1000_data_desc))
124 
125 #define E1000_MISC_VEC_ID               RTE_INTR_VEC_ZERO_OFFSET
126 #define E1000_RX_VEC_START              RTE_INTR_VEC_RXTX_OFFSET
127 
128 #define IGB_TX_MAX_SEG     UINT8_MAX
129 #define IGB_TX_MAX_MTU_SEG UINT8_MAX
130 #define EM_TX_MAX_SEG      UINT8_MAX
131 #define EM_TX_MAX_MTU_SEG  UINT8_MAX
132 
133 #define MAC_TYPE_FILTER_SUP(type)    do {\
134 	if ((type) != e1000_82580 && (type) != e1000_i350 &&\
135 		(type) != e1000_82576 && (type) != e1000_i210 &&\
136 		(type) != e1000_i211)\
137 		return -ENOTSUP;\
138 } while (0)
139 
140 #define MAC_TYPE_FILTER_SUP_EXT(type)    do {\
141 	if ((type) != e1000_82580 && (type) != e1000_i350 &&\
142 		(type) != e1000_i210 && (type) != e1000_i211)\
143 		return -ENOTSUP; \
144 } while (0)
145 
146 /* structure for interrupt relative data */
147 struct e1000_interrupt {
148 	uint32_t flags;
149 	uint32_t mask;
150 };
151 
152 /* local vfta copy */
153 struct e1000_vfta {
154 	uint32_t vfta[IGB_VFTA_SIZE];
155 };
156 
157 /*
158  * VF data which used by PF host only
159  */
160 #define E1000_MAX_VF_MC_ENTRIES         30
161 struct e1000_vf_info {
162 	uint8_t vf_mac_addresses[RTE_ETHER_ADDR_LEN];
163 	uint16_t vf_mc_hashes[E1000_MAX_VF_MC_ENTRIES];
164 	uint16_t num_vf_mc_hashes;
165 	uint16_t default_vf_vlan_id;
166 	uint16_t vlans_enabled;
167 	uint16_t pf_qos;
168 	uint16_t vlan_count;
169 	uint16_t tx_rate;
170 };
171 
172 TAILQ_HEAD(e1000_flex_filter_list, e1000_flex_filter);
173 
174 struct e1000_flex_filter_info {
175 	uint16_t len;
176 	uint32_t dwords[E1000_MAX_FLEX_FILTER_DWDS]; /* flex bytes in dword. */
177 	/* if mask bit is 1b, do not compare corresponding byte in dwords. */
178 	uint8_t mask[E1000_FLEX_FILTERS_MASK_SIZE];
179 	uint8_t priority;
180 };
181 
182 /* Flex filter structure */
183 struct e1000_flex_filter {
184 	TAILQ_ENTRY(e1000_flex_filter) entries;
185 	uint16_t index; /* index of flex filter */
186 	struct e1000_flex_filter_info filter_info;
187 	uint16_t queue; /* rx queue assigned to */
188 };
189 
190 TAILQ_HEAD(e1000_5tuple_filter_list, e1000_5tuple_filter);
191 TAILQ_HEAD(e1000_2tuple_filter_list, e1000_2tuple_filter);
192 
193 struct e1000_5tuple_filter_info {
194 	uint32_t dst_ip;
195 	uint32_t src_ip;
196 	uint16_t dst_port;
197 	uint16_t src_port;
198 	uint8_t proto;           /* l4 protocol. */
199 	/* the packet matched above 5tuple and contain any set bit will hit this filter. */
200 	uint8_t tcp_flags;
201 	uint8_t priority;        /* seven levels (001b-111b), 111b is highest,
202 				      used when more than one filter matches. */
203 	uint8_t dst_ip_mask:1,   /* if mask is 1b, do not compare dst ip. */
204 		src_ip_mask:1,   /* if mask is 1b, do not compare src ip. */
205 		dst_port_mask:1, /* if mask is 1b, do not compare dst port. */
206 		src_port_mask:1, /* if mask is 1b, do not compare src port. */
207 		proto_mask:1;    /* if mask is 1b, do not compare protocol. */
208 };
209 
210 struct e1000_2tuple_filter_info {
211 	uint16_t dst_port;
212 	uint8_t proto;           /* l4 protocol. */
213 	/* the packet matched above 2tuple and contain any set bit will hit this filter. */
214 	uint8_t tcp_flags;
215 	uint8_t priority;        /* seven levels (001b-111b), 111b is highest,
216 				      used when more than one filter matches. */
217 	uint8_t dst_ip_mask:1,   /* if mask is 1b, do not compare dst ip. */
218 		src_ip_mask:1,   /* if mask is 1b, do not compare src ip. */
219 		dst_port_mask:1, /* if mask is 1b, do not compare dst port. */
220 		src_port_mask:1, /* if mask is 1b, do not compare src port. */
221 		proto_mask:1;    /* if mask is 1b, do not compare protocol. */
222 };
223 
224 /* 5tuple filter structure */
225 struct e1000_5tuple_filter {
226 	TAILQ_ENTRY(e1000_5tuple_filter) entries;
227 	uint16_t index;       /* the index of 5tuple filter */
228 	struct e1000_5tuple_filter_info filter_info;
229 	uint16_t queue;       /* rx queue assigned to */
230 };
231 
232 /* 2tuple filter structure */
233 struct e1000_2tuple_filter {
234 	TAILQ_ENTRY(e1000_2tuple_filter) entries;
235 	uint16_t index;         /* the index of 2tuple filter */
236 	struct e1000_2tuple_filter_info filter_info;
237 	uint16_t queue;       /* rx queue assigned to */
238 };
239 
240 /* ethertype filter structure */
241 struct igb_ethertype_filter {
242 	uint16_t ethertype;
243 	uint32_t etqf;
244 };
245 
246 struct igb_rte_flow_rss_conf {
247 	struct rte_flow_action_rss conf; /**< RSS parameters. */
248 	uint8_t key[IGB_HKEY_MAX_INDEX * sizeof(uint32_t)]; /* Hash key. */
249 	/* Queues indices to use. */
250 	uint16_t queue[IGB_MAX_RX_QUEUE_NUM_82576];
251 };
252 
253 /*
254  * Structure to store filters'info.
255  */
256 struct e1000_filter_info {
257 	uint8_t ethertype_mask; /* Bit mask for every used ethertype filter */
258 	/* store used ethertype filters*/
259 	struct igb_ethertype_filter ethertype_filters[E1000_MAX_ETQF_FILTERS];
260 	uint8_t flex_mask;	/* Bit mask for every used flex filter */
261 	struct e1000_flex_filter_list flex_list;
262 	/* Bit mask for every used 5tuple filter */
263 	uint8_t fivetuple_mask;
264 	struct e1000_5tuple_filter_list fivetuple_list;
265 	/* Bit mask for every used 2tuple filter */
266 	uint8_t twotuple_mask;
267 	struct e1000_2tuple_filter_list twotuple_list;
268 	/* store the SYN filter info */
269 	uint32_t syn_info;
270 	/* store the rss filter info */
271 	struct igb_rte_flow_rss_conf rss_info;
272 };
273 
274 /*
275  * Structure to store private data for each driver instance (for each port).
276  */
277 struct e1000_adapter {
278 	struct e1000_hw         hw;
279 	struct e1000_hw_stats   stats;
280 	struct e1000_interrupt  intr;
281 	struct e1000_vfta       shadow_vfta;
282 	struct e1000_vf_info    *vfdata;
283 	struct e1000_filter_info filter;
284 	bool stopped;
285 	struct rte_timecounter  systime_tc;
286 	struct rte_timecounter  rx_tstamp_tc;
287 	struct rte_timecounter  tx_tstamp_tc;
288 };
289 
290 #define E1000_DEV_PRIVATE(adapter) \
291 	((struct e1000_adapter *)adapter)
292 
293 #define E1000_DEV_PRIVATE_TO_HW(adapter) \
294 	(&((struct e1000_adapter *)adapter)->hw)
295 
296 #define E1000_DEV_PRIVATE_TO_STATS(adapter) \
297 	(&((struct e1000_adapter *)adapter)->stats)
298 
299 #define E1000_DEV_PRIVATE_TO_INTR(adapter) \
300 	(&((struct e1000_adapter *)adapter)->intr)
301 
302 #define E1000_DEV_PRIVATE_TO_VFTA(adapter) \
303 	(&((struct e1000_adapter *)adapter)->shadow_vfta)
304 
305 #define E1000_DEV_PRIVATE_TO_P_VFDATA(adapter) \
306         (&((struct e1000_adapter *)adapter)->vfdata)
307 
308 #define E1000_DEV_PRIVATE_TO_FILTER_INFO(adapter) \
309 	(&((struct e1000_adapter *)adapter)->filter)
310 
311 struct rte_flow {
312 	enum rte_filter_type filter_type;
313 	void *rule;
314 };
315 
316 /* ntuple filter list structure */
317 struct igb_ntuple_filter_ele {
318 	TAILQ_ENTRY(igb_ntuple_filter_ele) entries;
319 	struct rte_eth_ntuple_filter filter_info;
320 };
321 
322 /* ethertype filter list structure */
323 struct igb_ethertype_filter_ele {
324 	TAILQ_ENTRY(igb_ethertype_filter_ele) entries;
325 	struct rte_eth_ethertype_filter filter_info;
326 };
327 
328 /* syn filter list structure */
329 struct igb_eth_syn_filter_ele {
330 	TAILQ_ENTRY(igb_eth_syn_filter_ele) entries;
331 	struct rte_eth_syn_filter filter_info;
332 };
333 
334 #define IGB_FLEX_FILTER_MAXLEN	128	/**< bytes to use in flex filter. */
335 #define IGB_FLEX_FILTER_MASK_SIZE	\
336 	(RTE_ALIGN(IGB_FLEX_FILTER_MAXLEN, CHAR_BIT) / CHAR_BIT)
337 					/**< mask bytes in flex filter. */
338 
339 /**
340  * A structure used to define the flex filter entry
341  * to support RTE_ETH_FILTER_FLEXIBLE data representation.
342  */
343 struct igb_flex_filter {
344 	uint16_t len;
345 	uint8_t bytes[IGB_FLEX_FILTER_MAXLEN]; /**< flex bytes in big endian. */
346 	uint8_t mask[IGB_FLEX_FILTER_MASK_SIZE];
347 		/**< if mask bit is 1b, do not compare corresponding byte. */
348 	uint8_t priority;
349 	uint16_t queue;       /**< Queue assigned to when match. */
350 };
351 
352 /* flex filter list structure */
353 struct igb_flex_filter_ele {
354 	TAILQ_ENTRY(igb_flex_filter_ele) entries;
355 	struct igb_flex_filter filter_info;
356 };
357 
358 /* rss filter  list structure */
359 struct igb_rss_conf_ele {
360 	TAILQ_ENTRY(igb_rss_conf_ele) entries;
361 	struct igb_rte_flow_rss_conf filter_info;
362 };
363 
364 /* igb_flow memory list structure */
365 struct igb_flow_mem {
366 	TAILQ_ENTRY(igb_flow_mem) entries;
367 	struct rte_flow *flow;
368 	struct rte_eth_dev *dev;
369 };
370 
371 TAILQ_HEAD(igb_ntuple_filter_list, igb_ntuple_filter_ele);
372 extern struct igb_ntuple_filter_list igb_filter_ntuple_list;
373 TAILQ_HEAD(igb_ethertype_filter_list, igb_ethertype_filter_ele);
374 extern struct igb_ethertype_filter_list igb_filter_ethertype_list;
375 TAILQ_HEAD(igb_syn_filter_list, igb_eth_syn_filter_ele);
376 extern struct igb_syn_filter_list igb_filter_syn_list;
377 TAILQ_HEAD(igb_flex_filter_list, igb_flex_filter_ele);
378 extern struct igb_flex_filter_list igb_filter_flex_list;
379 TAILQ_HEAD(igb_rss_filter_list, igb_rss_conf_ele);
380 extern struct igb_rss_filter_list igb_filter_rss_list;
381 TAILQ_HEAD(igb_flow_mem_list, igb_flow_mem);
382 extern struct igb_flow_mem_list igb_flow_list;
383 
384 extern const struct rte_flow_ops igb_flow_ops;
385 
386 /*
387  * RX/TX IGB function prototypes
388  */
389 void eth_igb_tx_queue_release(void *txq);
390 void eth_igb_rx_queue_release(void *rxq);
391 void igb_dev_clear_queues(struct rte_eth_dev *dev);
392 void igb_dev_free_queues(struct rte_eth_dev *dev);
393 
394 uint64_t igb_get_rx_port_offloads_capa(struct rte_eth_dev *dev);
395 uint64_t igb_get_rx_queue_offloads_capa(struct rte_eth_dev *dev);
396 
397 int eth_igb_rx_queue_setup(struct rte_eth_dev *dev, uint16_t rx_queue_id,
398 		uint16_t nb_rx_desc, unsigned int socket_id,
399 		const struct rte_eth_rxconf *rx_conf,
400 		struct rte_mempool *mb_pool);
401 
402 uint32_t eth_igb_rx_queue_count(struct rte_eth_dev *dev,
403 		uint16_t rx_queue_id);
404 
405 int eth_igb_rx_descriptor_done(void *rx_queue, uint16_t offset);
406 
407 int eth_igb_rx_descriptor_status(void *rx_queue, uint16_t offset);
408 int eth_igb_tx_descriptor_status(void *tx_queue, uint16_t offset);
409 
410 uint64_t igb_get_tx_port_offloads_capa(struct rte_eth_dev *dev);
411 uint64_t igb_get_tx_queue_offloads_capa(struct rte_eth_dev *dev);
412 
413 int eth_igb_tx_queue_setup(struct rte_eth_dev *dev, uint16_t tx_queue_id,
414 		uint16_t nb_tx_desc, unsigned int socket_id,
415 		const struct rte_eth_txconf *tx_conf);
416 
417 int eth_igb_tx_done_cleanup(void *txq, uint32_t free_cnt);
418 
419 int eth_igb_rx_init(struct rte_eth_dev *dev);
420 
421 void eth_igb_tx_init(struct rte_eth_dev *dev);
422 
423 uint16_t eth_igb_xmit_pkts(void *txq, struct rte_mbuf **tx_pkts,
424 		uint16_t nb_pkts);
425 
426 uint16_t eth_igb_prep_pkts(void *txq, struct rte_mbuf **tx_pkts,
427 		uint16_t nb_pkts);
428 
429 uint16_t eth_igb_recv_pkts(void *rxq, struct rte_mbuf **rx_pkts,
430 		uint16_t nb_pkts);
431 
432 uint16_t eth_igb_recv_scattered_pkts(void *rxq,
433 		struct rte_mbuf **rx_pkts, uint16_t nb_pkts);
434 
435 int eth_igb_rss_hash_update(struct rte_eth_dev *dev,
436 			    struct rte_eth_rss_conf *rss_conf);
437 
438 int eth_igb_rss_hash_conf_get(struct rte_eth_dev *dev,
439 			      struct rte_eth_rss_conf *rss_conf);
440 
441 int eth_igbvf_rx_init(struct rte_eth_dev *dev);
442 
443 void eth_igbvf_tx_init(struct rte_eth_dev *dev);
444 
445 /*
446  * misc function prototypes
447  */
448 void igb_pf_host_init(struct rte_eth_dev *eth_dev);
449 
450 void igb_pf_mbx_process(struct rte_eth_dev *eth_dev);
451 
452 int igb_pf_host_configure(struct rte_eth_dev *eth_dev);
453 
454 void igb_rxq_info_get(struct rte_eth_dev *dev, uint16_t queue_id,
455 	struct rte_eth_rxq_info *qinfo);
456 
457 void igb_txq_info_get(struct rte_eth_dev *dev, uint16_t queue_id,
458 	struct rte_eth_txq_info *qinfo);
459 
460 uint32_t em_get_max_pktlen(struct rte_eth_dev *dev);
461 
462 /*
463  * RX/TX EM function prototypes
464  */
465 void eth_em_tx_queue_release(void *txq);
466 void eth_em_rx_queue_release(void *rxq);
467 
468 void em_dev_clear_queues(struct rte_eth_dev *dev);
469 void em_dev_free_queues(struct rte_eth_dev *dev);
470 
471 uint64_t em_get_rx_port_offloads_capa(struct rte_eth_dev *dev);
472 uint64_t em_get_rx_queue_offloads_capa(struct rte_eth_dev *dev);
473 
474 int eth_em_rx_queue_setup(struct rte_eth_dev *dev, uint16_t rx_queue_id,
475 		uint16_t nb_rx_desc, unsigned int socket_id,
476 		const struct rte_eth_rxconf *rx_conf,
477 		struct rte_mempool *mb_pool);
478 
479 uint32_t eth_em_rx_queue_count(struct rte_eth_dev *dev,
480 		uint16_t rx_queue_id);
481 
482 int eth_em_rx_descriptor_done(void *rx_queue, uint16_t offset);
483 
484 int eth_em_rx_descriptor_status(void *rx_queue, uint16_t offset);
485 int eth_em_tx_descriptor_status(void *tx_queue, uint16_t offset);
486 
487 uint64_t em_get_tx_port_offloads_capa(struct rte_eth_dev *dev);
488 uint64_t em_get_tx_queue_offloads_capa(struct rte_eth_dev *dev);
489 
490 int eth_em_tx_queue_setup(struct rte_eth_dev *dev, uint16_t tx_queue_id,
491 		uint16_t nb_tx_desc, unsigned int socket_id,
492 		const struct rte_eth_txconf *tx_conf);
493 
494 int eth_em_rx_init(struct rte_eth_dev *dev);
495 
496 void eth_em_tx_init(struct rte_eth_dev *dev);
497 
498 uint16_t eth_em_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
499 		uint16_t nb_pkts);
500 
501 uint16_t eth_em_prep_pkts(void *txq, struct rte_mbuf **tx_pkts,
502 		uint16_t nb_pkts);
503 
504 uint16_t eth_em_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
505 		uint16_t nb_pkts);
506 
507 uint16_t eth_em_recv_scattered_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
508 		uint16_t nb_pkts);
509 
510 void em_rxq_info_get(struct rte_eth_dev *dev, uint16_t queue_id,
511 	struct rte_eth_rxq_info *qinfo);
512 
513 void em_txq_info_get(struct rte_eth_dev *dev, uint16_t queue_id,
514 	struct rte_eth_txq_info *qinfo);
515 
516 void igb_pf_host_uninit(struct rte_eth_dev *dev);
517 
518 void igb_filterlist_flush(struct rte_eth_dev *dev);
519 int igb_delete_5tuple_filter_82576(struct rte_eth_dev *dev,
520 		struct e1000_5tuple_filter *filter);
521 int igb_delete_2tuple_filter(struct rte_eth_dev *dev,
522 		struct e1000_2tuple_filter *filter);
523 void igb_remove_flex_filter(struct rte_eth_dev *dev,
524 			struct e1000_flex_filter *filter);
525 int igb_ethertype_filter_remove(struct e1000_filter_info *filter_info,
526 	uint8_t idx);
527 int igb_add_del_ntuple_filter(struct rte_eth_dev *dev,
528 		struct rte_eth_ntuple_filter *ntuple_filter, bool add);
529 int igb_add_del_ethertype_filter(struct rte_eth_dev *dev,
530 			struct rte_eth_ethertype_filter *filter,
531 			bool add);
532 int eth_igb_syn_filter_set(struct rte_eth_dev *dev,
533 			struct rte_eth_syn_filter *filter,
534 			bool add);
535 int eth_igb_add_del_flex_filter(struct rte_eth_dev *dev,
536 			struct igb_flex_filter *filter,
537 			bool add);
538 int igb_rss_conf_init(struct rte_eth_dev *dev,
539 		      struct igb_rte_flow_rss_conf *out,
540 		      const struct rte_flow_action_rss *in);
541 int igb_action_rss_same(const struct rte_flow_action_rss *comp,
542 			const struct rte_flow_action_rss *with);
543 int igb_config_rss_filter(struct rte_eth_dev *dev,
544 			struct igb_rte_flow_rss_conf *conf,
545 			bool add);
546 void em_flush_desc_rings(struct rte_eth_dev *dev);
547 
548 #endif /* _E1000_ETHDEV_H_ */
549