Home
last modified time | relevance | path

Searched refs:rx_queue (Results 1 – 25 of 158) sorted by relevance

1234567

/dpdk/drivers/net/nfb/
H A Dnfb_stats.c23 struct ndp_rx_queue *rx_queue = *((struct ndp_rx_queue **) in nfb_eth_stats_get() local
30 stats->q_ipackets[i] = rx_queue[i].rx_pkts; in nfb_eth_stats_get()
31 stats->q_ibytes[i] = rx_queue[i].rx_bytes; in nfb_eth_stats_get()
33 rx_total += rx_queue[i].rx_pkts; in nfb_eth_stats_get()
34 rx_total_bytes += rx_queue[i].rx_bytes; in nfb_eth_stats_get()
62 struct ndp_rx_queue *rx_queue = *((struct ndp_rx_queue **) in nfb_eth_stats_reset() local
68 rx_queue[i].rx_pkts = 0; in nfb_eth_stats_reset()
69 rx_queue[i].rx_bytes = 0; in nfb_eth_stats_reset()
70 rx_queue[i].err_pkts = 0; in nfb_eth_stats_reset()
/dpdk/drivers/net/af_packet/
H A Drte_eth_af_packet.c80 struct pkt_rx_queue *rx_queue; member
341 internals->rx_queue[i].sockfd = -1; in eth_dev_stop()
420 internal->rx_queue[i].rx_pkts = 0; in eth_stats_reset()
449 munmap(internals->rx_queue[q].map, in eth_dev_close()
455 rte_free(internals->rx_queue); in eth_dev_close()
673 struct pkt_rx_queue *rx_queue; in rte_pmd_init_internals() local
827 rx_queue = &((*internals)->rx_queue[q]); in rte_pmd_init_internals()
844 if (rx_queue->rd == NULL) in rte_pmd_init_internals()
847 rx_queue->rd[i].iov_base = rx_queue->map + (i * framesize); in rte_pmd_init_internals()
850 rx_queue->sockfd = qsockfd; in rte_pmd_init_internals()
[all …]
/dpdk/drivers/net/iavf/
H A Diavf_rxtx.h610 uint16_t iavf_recv_pkts_flex_rxd(void *rx_queue,
613 uint16_t iavf_recv_scattered_pkts(void *rx_queue,
616 uint16_t iavf_recv_scattered_pkts_flex_rxd(void *rx_queue,
629 uint32_t iavf_dev_rxq_count(void *rx_queue);
630 int iavf_dev_rx_desc_status(void *rx_queue, uint16_t offset);
637 uint16_t iavf_recv_scattered_pkts_vec(void *rx_queue,
647 uint16_t iavf_recv_pkts_vec_avx2_flex_rxd(void *rx_queue,
650 uint16_t iavf_recv_scattered_pkts_vec_avx2(void *rx_queue,
667 uint16_t iavf_recv_pkts_vec_avx512_offload(void *rx_queue,
670 uint16_t iavf_recv_pkts_vec_avx512_flex_rxd(void *rx_queue,
[all …]
H A Diavf_rxtx_vec_avx512.c1514 struct iavf_rx_queue *rxq = rx_queue; in iavf_recv_scattered_burst_vec_avx512()
1566 return retval + iavf_recv_scattered_burst_vec_avx512(rx_queue, in iavf_recv_scattered_pkts_vec_avx512_cmn()
1585 iavf_recv_scattered_burst_vec_avx512_flex_rxd(void *rx_queue, in iavf_recv_scattered_burst_vec_avx512_flex_rxd() argument
1590 struct iavf_rx_queue *rxq = rx_queue; in iavf_recv_scattered_burst_vec_avx512_flex_rxd()
1629 iavf_recv_scattered_pkts_vec_avx512_flex_rxd_cmn(void *rx_queue, in iavf_recv_scattered_pkts_vec_avx512_flex_rxd_cmn() argument
1639 (rx_queue, rx_pkts + retval, in iavf_recv_scattered_pkts_vec_avx512_flex_rxd_cmn()
1651 iavf_recv_scattered_pkts_vec_avx512_flex_rxd(void *rx_queue, in iavf_recv_scattered_pkts_vec_avx512_flex_rxd() argument
1665 return _iavf_recv_raw_pkts_vec_avx512(rx_queue, rx_pkts, in iavf_recv_pkts_vec_avx512_offload()
1670 iavf_recv_scattered_pkts_vec_avx512_offload(void *rx_queue, in iavf_recv_scattered_pkts_vec_avx512_offload() argument
1679 iavf_recv_pkts_vec_avx512_flex_rxd_offload(void *rx_queue, in iavf_recv_pkts_vec_avx512_flex_rxd_offload() argument
[all …]
H A Diavf_rxtx_vec_sse.c932 iavf_recv_pkts_vec(void *rx_queue, struct rte_mbuf **rx_pkts, in iavf_recv_pkts_vec() argument
935 return _recv_raw_pkts_vec(rx_queue, rx_pkts, nb_pkts, NULL); in iavf_recv_pkts_vec()
944 iavf_recv_pkts_vec_flex_rxd(void *rx_queue, struct rte_mbuf **rx_pkts, in iavf_recv_pkts_vec_flex_rxd() argument
960 struct iavf_rx_queue *rxq = rx_queue; in iavf_recv_scattered_burst_vec()
1003 burst = iavf_recv_scattered_burst_vec(rx_queue, in iavf_recv_scattered_pkts_vec()
1012 return retval + iavf_recv_scattered_burst_vec(rx_queue, in iavf_recv_scattered_pkts_vec()
1025 iavf_recv_scattered_burst_vec_flex_rxd(void *rx_queue, in iavf_recv_scattered_burst_vec_flex_rxd() argument
1029 struct iavf_rx_queue *rxq = rx_queue; in iavf_recv_scattered_burst_vec_flex_rxd()
1064 iavf_recv_scattered_pkts_vec_flex_rxd(void *rx_queue, in iavf_recv_scattered_pkts_vec_flex_rxd() argument
1073 burst = iavf_recv_scattered_burst_vec_flex_rxd(rx_queue, in iavf_recv_scattered_pkts_vec_flex_rxd()
[all …]
H A Diavf_rxtx_vec_avx2.c1220 iavf_recv_pkts_vec_avx2(void *rx_queue, struct rte_mbuf **rx_pkts, in iavf_recv_pkts_vec_avx2() argument
1223 return _iavf_recv_raw_pkts_vec_avx2(rx_queue, rx_pkts, nb_pkts, NULL); in iavf_recv_pkts_vec_avx2()
1234 return _iavf_recv_raw_pkts_vec_avx2_flex_rxd(rx_queue, rx_pkts, in iavf_recv_pkts_vec_avx2_flex_rxd()
1247 struct iavf_rx_queue *rxq = rx_queue; in iavf_recv_scattered_burst_vec_avx2()
1292 uint16_t burst = iavf_recv_scattered_burst_vec_avx2(rx_queue, in iavf_recv_scattered_pkts_vec_avx2()
1299 return retval + iavf_recv_scattered_burst_vec_avx2(rx_queue, in iavf_recv_scattered_pkts_vec_avx2()
1310 iavf_recv_scattered_burst_vec_avx2_flex_rxd(void *rx_queue, in iavf_recv_scattered_burst_vec_avx2_flex_rxd() argument
1314 struct iavf_rx_queue *rxq = rx_queue; in iavf_recv_scattered_burst_vec_avx2_flex_rxd()
1353 iavf_recv_scattered_pkts_vec_avx2_flex_rxd(void *rx_queue, in iavf_recv_scattered_pkts_vec_avx2_flex_rxd() argument
1362 (rx_queue, rx_pkts + retval, IAVF_VPMD_RX_MAX_BURST); in iavf_recv_scattered_pkts_vec_avx2_flex_rxd()
[all …]
/dpdk/drivers/net/ice/
H A Dice_rxtx.h224 uint16_t ice_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
234 uint32_t ice_rx_queue_count(void *rx_queue);
243 int ice_rx_descriptor_status(void *rx_queue, uint16_t offset);
254 uint16_t ice_recv_pkts_vec(void *rx_queue, struct rte_mbuf **rx_pkts,
260 uint16_t ice_recv_pkts_vec_avx2(void *rx_queue, struct rte_mbuf **rx_pkts,
264 uint16_t ice_recv_scattered_pkts_vec_avx2(void *rx_queue,
267 uint16_t ice_recv_scattered_pkts_vec_avx2_offload(void *rx_queue,
274 uint16_t ice_recv_pkts_vec_avx512(void *rx_queue, struct rte_mbuf **rx_pkts,
276 uint16_t ice_recv_pkts_vec_avx512_offload(void *rx_queue,
279 uint16_t ice_recv_scattered_pkts_vec_avx512(void *rx_queue,
[all …]
H A Dice_rxtx_vec_avx2.c709 ice_recv_pkts_vec_avx2(void *rx_queue, struct rte_mbuf **rx_pkts, in ice_recv_pkts_vec_avx2() argument
712 return _ice_recv_raw_pkts_vec_avx2(rx_queue, rx_pkts, in ice_recv_pkts_vec_avx2()
720 return _ice_recv_raw_pkts_vec_avx2(rx_queue, rx_pkts, in ice_recv_pkts_vec_avx2_offload()
733 struct ice_rx_queue *rxq = rx_queue; in ice_recv_scattered_burst_vec_avx2()
772 ice_recv_scattered_pkts_vec_avx2_common(void *rx_queue, in ice_recv_scattered_pkts_vec_avx2_common() argument
780 uint16_t burst = ice_recv_scattered_burst_vec_avx2(rx_queue, in ice_recv_scattered_pkts_vec_avx2_common()
787 return retval + ice_recv_scattered_burst_vec_avx2(rx_queue, in ice_recv_scattered_pkts_vec_avx2_common()
792 ice_recv_scattered_pkts_vec_avx2(void *rx_queue, in ice_recv_scattered_pkts_vec_avx2() argument
796 return ice_recv_scattered_pkts_vec_avx2_common(rx_queue, in ice_recv_scattered_pkts_vec_avx2()
803 ice_recv_scattered_pkts_vec_avx2_offload(void *rx_queue, in ice_recv_scattered_pkts_vec_avx2_offload() argument
[all …]
H A Dice_rxtx_vec_avx512.c824 ice_recv_pkts_vec_avx512(void *rx_queue, struct rte_mbuf **rx_pkts, in ice_recv_pkts_vec_avx512() argument
835 ice_recv_pkts_vec_avx512_offload(void *rx_queue, struct rte_mbuf **rx_pkts, in ice_recv_pkts_vec_avx512_offload() argument
838 return _ice_recv_raw_pkts_vec_avx512(rx_queue, rx_pkts, in ice_recv_pkts_vec_avx512_offload()
851 struct ice_rx_queue *rxq = rx_queue; in ice_recv_scattered_burst_vec_avx512()
889 ice_recv_scattered_burst_vec_avx512_offload(void *rx_queue, in ice_recv_scattered_burst_vec_avx512_offload() argument
893 struct ice_rx_queue *rxq = rx_queue; in ice_recv_scattered_burst_vec_avx512_offload()
938 uint16_t burst = ice_recv_scattered_burst_vec_avx512(rx_queue, in ice_recv_scattered_pkts_vec_avx512()
945 return retval + ice_recv_scattered_burst_vec_avx512(rx_queue, in ice_recv_scattered_pkts_vec_avx512()
956 ice_recv_scattered_pkts_vec_avx512_offload(void *rx_queue, in ice_recv_scattered_pkts_vec_avx512_offload() argument
964 ice_recv_scattered_burst_vec_avx512_offload(rx_queue, in ice_recv_scattered_pkts_vec_avx512_offload()
[all …]
/dpdk/drivers/net/virtio/
H A Dvirtio_ethdev.h77 uint16_t virtio_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
79 uint16_t virtio_recv_pkts_packed(void *rx_queue, struct rte_mbuf **rx_pkts,
82 uint16_t virtio_recv_mergeable_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
85 uint16_t virtio_recv_mergeable_pkts_packed(void *rx_queue,
88 uint16_t virtio_recv_pkts_inorder(void *rx_queue,
102 uint16_t virtio_recv_pkts_vec(void *rx_queue, struct rte_mbuf **rx_pkts,
105 uint16_t virtio_recv_pkts_packed_vec(void *rx_queue, struct rte_mbuf **rx_pkts,
/dpdk/drivers/net/thunderx/
H A Dnicvf_rxtx.h86 uint32_t nicvf_dev_rx_queue_count(void *rx_queue);
93 uint16_t nicvf_recv_pkts_vlan_strip(void *rx_queue, struct rte_mbuf **rx_pkts,
95 uint16_t nicvf_recv_pkts_cksum_vlan_strip(void *rx_queue,
98 uint16_t nicvf_recv_pkts_multiseg_no_offload(void *rx_queue,
100 uint16_t nicvf_recv_pkts_multiseg_cksum(void *rx_queue,
102 uint16_t nicvf_recv_pkts_multiseg_vlan_strip(void *rx_queue,
104 uint16_t nicvf_recv_pkts_multiseg_cksum_vlan_strip(void *rx_queue,
H A Dnicvf_rxtx.c428 struct nicvf_rxq *rxq = rx_queue; in nicvf_recv_pkts()
491 return nicvf_recv_pkts(rx_queue, rx_pkts, nb_pkts, in nicvf_recv_pkts_no_offload()
499 return nicvf_recv_pkts(rx_queue, rx_pkts, nb_pkts, in nicvf_recv_pkts_cksum()
507 return nicvf_recv_pkts(rx_queue, rx_pkts, nb_pkts, in nicvf_recv_pkts_vlan_strip()
515 return nicvf_recv_pkts(rx_queue, rx_pkts, nb_pkts, in nicvf_recv_pkts_cksum_vlan_strip()
581 struct nicvf_rxq *rxq = rx_queue; in nicvf_recv_pkts_multiseg()
623 return nicvf_recv_pkts_multiseg(rx_queue, rx_pkts, nb_pkts, in nicvf_recv_pkts_multiseg_no_offload()
631 return nicvf_recv_pkts_multiseg(rx_queue, rx_pkts, nb_pkts, in nicvf_recv_pkts_multiseg_cksum()
644 nicvf_recv_pkts_multiseg_cksum_vlan_strip(void *rx_queue, in nicvf_recv_pkts_multiseg_cksum_vlan_strip() argument
652 nicvf_dev_rx_queue_count(void *rx_queue) in nicvf_dev_rx_queue_count() argument
[all …]
/dpdk/examples/l3fwd-power/
H A Dmain.c884 port_id = rx_queue->port_id; in turn_on_off_intr()
907 portid = rx_queue->port_id; in event_register()
994 rx_queue->idle_hint = 0; in main_intr_loop()
995 portid = rx_queue->port_id; in main_intr_loop()
1151 portid = rx_queue->port_id; in main_telemetry_loop()
1265 rx_queue->idle_hint = 0; in main_empty_poll_loop()
1266 portid = rx_queue->port_id; in main_empty_poll_loop()
1392 rx_queue->idle_hint = 0; in main_legacy_loop()
1393 portid = rx_queue->port_id; in main_legacy_loop()
1424 rx_queue->freq_up_hint = in main_legacy_loop()
[all …]
/dpdk/drivers/crypto/qat/dev/
H A Dqat_sym_pmd_gen1.c948 head = (head + rx_queue->msg_size) & rx_queue->modulo_mask; in qat_sym_dp_dequeue_burst_gen1()
967 head = (head + rx_queue->msg_size) & in qat_sym_dp_dequeue_burst_gen1()
968 rx_queue->modulo_mask; in qat_sym_dp_dequeue_burst_gen1()
981 head = (head + rx_queue->msg_size) & in qat_sym_dp_dequeue_burst_gen1()
982 rx_queue->modulo_mask; in qat_sym_dp_dequeue_burst_gen1()
1009 rx_queue->modulo_mask; in qat_sym_dp_dequeue_single_gen1()
1053 rx_queue->head = dp_ctx->head; in qat_sym_dp_dequeue_done_gen1()
1061 old_head = rx_queue->csr_head; in qat_sym_dp_dequeue_done_gen1()
1062 new_head = rx_queue->head; in qat_sym_dp_dequeue_done_gen1()
1078 rx_queue->csr_head = new_head; in qat_sym_dp_dequeue_done_gen1()
[all …]
/dpdk/drivers/net/ark/
H A Dark_ethdev_rx.h20 uint32_t eth_ark_dev_rx_queue_count(void *rx_queue);
23 uint16_t eth_ark_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
25 void eth_ark_dev_rx_queue_release(void *rx_queue);
/dpdk/lib/eventdev/
H A Drte_event_eth_rx_adapter.c387 return dev_info->rx_queue && in rxa_intr_queue()
400 dev_info->rx_queue && in rxa_polled_queue()
666 &dev_info->rx_queue[q]; in rxa_calc_wrr_sequence()
2174 rx_queue = dev_info->rx_queue; in rxa_sw_add()
2178 dev_info->rx_queue = in rxa_sw_add()
2265 if (rx_queue == NULL) { in rxa_sw_add()
2266 rte_free(dev_info->rx_queue); in rxa_sw_add()
2267 dev_info->rx_queue = NULL; in rxa_sw_add()
2661 dev_info->rx_queue = in rte_event_eth_rx_adapter_queue_add()
2765 dev_info->rx_queue = NULL; in rte_event_eth_rx_adapter_queue_del()
[all …]
/dpdk/drivers/net/i40e/
H A Di40e_rxtx.h206 uint16_t i40e_recv_pkts(void *rx_queue,
209 uint16_t i40e_recv_scattered_pkts(void *rx_queue,
232 uint32_t i40e_dev_rx_queue_count(void *rx_queue);
233 int i40e_dev_rx_descriptor_status(void *rx_queue, uint16_t offset);
236 uint16_t i40e_recv_pkts_vec(void *rx_queue, struct rte_mbuf **rx_pkts,
238 uint16_t i40e_recv_scattered_pkts_vec(void *rx_queue,
253 uint16_t i40e_recv_pkts_vec_avx2(void *rx_queue, struct rte_mbuf **rx_pkts,
255 uint16_t i40e_recv_scattered_pkts_vec_avx2(void *rx_queue,
259 int i40e_get_monitor_addr(void *rx_queue, struct rte_power_monitor_cond *pmc);
260 uint16_t i40e_recv_pkts_vec_avx512(void *rx_queue,
[all …]
/dpdk/drivers/net/ngbe/
H A Dngbe_ethdev.h204 uint32_t ngbe_dev_rx_queue_count(void *rx_queue);
206 int ngbe_dev_rx_descriptor_status(void *rx_queue, uint16_t offset);
241 uint16_t ngbe_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
244 uint16_t ngbe_recv_pkts_bulk_alloc(void *rx_queue, struct rte_mbuf **rx_pkts,
247 uint16_t ngbe_recv_pkts_sc_single_alloc(void *rx_queue,
249 uint16_t ngbe_recv_pkts_sc_bulk_alloc(void *rx_queue,
/dpdk/drivers/net/axgbe/
H A Daxgbe_rxtx.h183 uint16_t axgbe_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
185 uint16_t eth_axgbe_recv_scattered_pkts(void *rx_queue,
187 uint16_t axgbe_recv_pkts_threshold_refresh(void *rx_queue,
191 int axgbe_dev_rx_descriptor_status(void *rx_queue, uint16_t offset);
/dpdk/drivers/common/qat/
H A Dqat_qp.c810 struct qat_queue *rx_queue; in qat_dequeue_op_burst() local
817 rx_queue = &(tmp_qp->rx_q); in qat_dequeue_op_burst()
818 head = rx_queue->head; in qat_dequeue_op_burst()
819 resp_msg = (uint8_t *)rx_queue->base_addr + rx_queue->head; in qat_dequeue_op_burst()
828 tmp_qp->op_cookies[head >> rx_queue->trailz], in qat_dequeue_op_burst()
831 head = adf_modulo(head + rx_queue->msg_size, in qat_dequeue_op_burst()
832 rx_queue->modulo_mask); in qat_dequeue_op_burst()
834 resp_msg = (uint8_t *)rx_queue->base_addr + head; in qat_dequeue_op_burst()
852 rx_queue->nb_processed_responses++; in qat_dequeue_op_burst()
858 rx_queue->head = head; in qat_dequeue_op_burst()
[all …]
/dpdk/app/test-pmd/
H A Dshared_rxq_fwd.c35 fsm[sm_id]->rx_queue == fs->rx_queue) in forward_stream_get()
101 nb_rx = rte_eth_rx_burst(fs->rx_port, fs->rx_queue, pkts_burst, in shared_rxq_fwd()
/dpdk/drivers/net/fm10k/
H A Dfm10k.h318 uint16_t fm10k_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
321 uint16_t fm10k_recv_scattered_pkts(void *rx_queue,
325 fm10k_dev_rx_queue_count(void *rx_queue);
328 fm10k_dev_rx_descriptor_status(void *rx_queue, uint16_t offset);
331 fm10k_dev_tx_descriptor_status(void *rx_queue, uint16_t offset);
/dpdk/drivers/net/atlantic/
H A Datl_ethdev.h69 uint32_t atl_rx_queue_count(void *rx_queue);
71 int atl_dev_rx_descriptor_status(void *rx_queue, uint16_t offset);
98 uint16_t atl_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
/dpdk/drivers/bus/fslmc/portal/
H A Ddpaa2_hw_dpci.c84 rxq = &(dpci_node->rx_queue[i]); in rte_dpaa2_create_dpci_device()
121 dpci_node->rx_queue[i].fqid = rx_attr.fqid; in rte_dpaa2_create_dpci_device()
144 struct dpaa2_queue *rxq = &(dpci_node->rx_queue[i]); in rte_dpaa2_create_dpci_device()
/dpdk/drivers/net/e1000/
H A De1000_ethdev.h403 uint32_t eth_igb_rx_queue_count(void *rx_queue);
405 int eth_igb_rx_descriptor_status(void *rx_queue, uint16_t offset);
477 uint32_t eth_em_rx_queue_count(void *rx_queue);
479 int eth_em_rx_descriptor_status(void *rx_queue, uint16_t offset);
499 uint16_t eth_em_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
502 uint16_t eth_em_recv_scattered_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,

1234567