| /dpdk/drivers/net/nfb/ |
| H A D | nfb_stats.c | 23 struct ndp_rx_queue *rx_queue = *((struct ndp_rx_queue **) in nfb_eth_stats_get() local 30 stats->q_ipackets[i] = rx_queue[i].rx_pkts; in nfb_eth_stats_get() 31 stats->q_ibytes[i] = rx_queue[i].rx_bytes; in nfb_eth_stats_get() 33 rx_total += rx_queue[i].rx_pkts; in nfb_eth_stats_get() 34 rx_total_bytes += rx_queue[i].rx_bytes; in nfb_eth_stats_get() 62 struct ndp_rx_queue *rx_queue = *((struct ndp_rx_queue **) in nfb_eth_stats_reset() local 68 rx_queue[i].rx_pkts = 0; in nfb_eth_stats_reset() 69 rx_queue[i].rx_bytes = 0; in nfb_eth_stats_reset() 70 rx_queue[i].err_pkts = 0; in nfb_eth_stats_reset()
|
| /dpdk/drivers/net/af_packet/ |
| H A D | rte_eth_af_packet.c | 80 struct pkt_rx_queue *rx_queue; member 341 internals->rx_queue[i].sockfd = -1; in eth_dev_stop() 420 internal->rx_queue[i].rx_pkts = 0; in eth_stats_reset() 449 munmap(internals->rx_queue[q].map, in eth_dev_close() 455 rte_free(internals->rx_queue); in eth_dev_close() 673 struct pkt_rx_queue *rx_queue; in rte_pmd_init_internals() local 827 rx_queue = &((*internals)->rx_queue[q]); in rte_pmd_init_internals() 844 if (rx_queue->rd == NULL) in rte_pmd_init_internals() 847 rx_queue->rd[i].iov_base = rx_queue->map + (i * framesize); in rte_pmd_init_internals() 850 rx_queue->sockfd = qsockfd; in rte_pmd_init_internals() [all …]
|
| /dpdk/drivers/net/iavf/ |
| H A D | iavf_rxtx.h | 610 uint16_t iavf_recv_pkts_flex_rxd(void *rx_queue, 613 uint16_t iavf_recv_scattered_pkts(void *rx_queue, 616 uint16_t iavf_recv_scattered_pkts_flex_rxd(void *rx_queue, 629 uint32_t iavf_dev_rxq_count(void *rx_queue); 630 int iavf_dev_rx_desc_status(void *rx_queue, uint16_t offset); 637 uint16_t iavf_recv_scattered_pkts_vec(void *rx_queue, 647 uint16_t iavf_recv_pkts_vec_avx2_flex_rxd(void *rx_queue, 650 uint16_t iavf_recv_scattered_pkts_vec_avx2(void *rx_queue, 667 uint16_t iavf_recv_pkts_vec_avx512_offload(void *rx_queue, 670 uint16_t iavf_recv_pkts_vec_avx512_flex_rxd(void *rx_queue, [all …]
|
| H A D | iavf_rxtx_vec_avx512.c | 1514 struct iavf_rx_queue *rxq = rx_queue; in iavf_recv_scattered_burst_vec_avx512() 1566 return retval + iavf_recv_scattered_burst_vec_avx512(rx_queue, in iavf_recv_scattered_pkts_vec_avx512_cmn() 1585 iavf_recv_scattered_burst_vec_avx512_flex_rxd(void *rx_queue, in iavf_recv_scattered_burst_vec_avx512_flex_rxd() argument 1590 struct iavf_rx_queue *rxq = rx_queue; in iavf_recv_scattered_burst_vec_avx512_flex_rxd() 1629 iavf_recv_scattered_pkts_vec_avx512_flex_rxd_cmn(void *rx_queue, in iavf_recv_scattered_pkts_vec_avx512_flex_rxd_cmn() argument 1639 (rx_queue, rx_pkts + retval, in iavf_recv_scattered_pkts_vec_avx512_flex_rxd_cmn() 1651 iavf_recv_scattered_pkts_vec_avx512_flex_rxd(void *rx_queue, in iavf_recv_scattered_pkts_vec_avx512_flex_rxd() argument 1665 return _iavf_recv_raw_pkts_vec_avx512(rx_queue, rx_pkts, in iavf_recv_pkts_vec_avx512_offload() 1670 iavf_recv_scattered_pkts_vec_avx512_offload(void *rx_queue, in iavf_recv_scattered_pkts_vec_avx512_offload() argument 1679 iavf_recv_pkts_vec_avx512_flex_rxd_offload(void *rx_queue, in iavf_recv_pkts_vec_avx512_flex_rxd_offload() argument [all …]
|
| H A D | iavf_rxtx_vec_sse.c | 932 iavf_recv_pkts_vec(void *rx_queue, struct rte_mbuf **rx_pkts, in iavf_recv_pkts_vec() argument 935 return _recv_raw_pkts_vec(rx_queue, rx_pkts, nb_pkts, NULL); in iavf_recv_pkts_vec() 944 iavf_recv_pkts_vec_flex_rxd(void *rx_queue, struct rte_mbuf **rx_pkts, in iavf_recv_pkts_vec_flex_rxd() argument 960 struct iavf_rx_queue *rxq = rx_queue; in iavf_recv_scattered_burst_vec() 1003 burst = iavf_recv_scattered_burst_vec(rx_queue, in iavf_recv_scattered_pkts_vec() 1012 return retval + iavf_recv_scattered_burst_vec(rx_queue, in iavf_recv_scattered_pkts_vec() 1025 iavf_recv_scattered_burst_vec_flex_rxd(void *rx_queue, in iavf_recv_scattered_burst_vec_flex_rxd() argument 1029 struct iavf_rx_queue *rxq = rx_queue; in iavf_recv_scattered_burst_vec_flex_rxd() 1064 iavf_recv_scattered_pkts_vec_flex_rxd(void *rx_queue, in iavf_recv_scattered_pkts_vec_flex_rxd() argument 1073 burst = iavf_recv_scattered_burst_vec_flex_rxd(rx_queue, in iavf_recv_scattered_pkts_vec_flex_rxd() [all …]
|
| H A D | iavf_rxtx_vec_avx2.c | 1220 iavf_recv_pkts_vec_avx2(void *rx_queue, struct rte_mbuf **rx_pkts, in iavf_recv_pkts_vec_avx2() argument 1223 return _iavf_recv_raw_pkts_vec_avx2(rx_queue, rx_pkts, nb_pkts, NULL); in iavf_recv_pkts_vec_avx2() 1234 return _iavf_recv_raw_pkts_vec_avx2_flex_rxd(rx_queue, rx_pkts, in iavf_recv_pkts_vec_avx2_flex_rxd() 1247 struct iavf_rx_queue *rxq = rx_queue; in iavf_recv_scattered_burst_vec_avx2() 1292 uint16_t burst = iavf_recv_scattered_burst_vec_avx2(rx_queue, in iavf_recv_scattered_pkts_vec_avx2() 1299 return retval + iavf_recv_scattered_burst_vec_avx2(rx_queue, in iavf_recv_scattered_pkts_vec_avx2() 1310 iavf_recv_scattered_burst_vec_avx2_flex_rxd(void *rx_queue, in iavf_recv_scattered_burst_vec_avx2_flex_rxd() argument 1314 struct iavf_rx_queue *rxq = rx_queue; in iavf_recv_scattered_burst_vec_avx2_flex_rxd() 1353 iavf_recv_scattered_pkts_vec_avx2_flex_rxd(void *rx_queue, in iavf_recv_scattered_pkts_vec_avx2_flex_rxd() argument 1362 (rx_queue, rx_pkts + retval, IAVF_VPMD_RX_MAX_BURST); in iavf_recv_scattered_pkts_vec_avx2_flex_rxd() [all …]
|
| /dpdk/drivers/net/ice/ |
| H A D | ice_rxtx.h | 224 uint16_t ice_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, 234 uint32_t ice_rx_queue_count(void *rx_queue); 243 int ice_rx_descriptor_status(void *rx_queue, uint16_t offset); 254 uint16_t ice_recv_pkts_vec(void *rx_queue, struct rte_mbuf **rx_pkts, 260 uint16_t ice_recv_pkts_vec_avx2(void *rx_queue, struct rte_mbuf **rx_pkts, 264 uint16_t ice_recv_scattered_pkts_vec_avx2(void *rx_queue, 267 uint16_t ice_recv_scattered_pkts_vec_avx2_offload(void *rx_queue, 274 uint16_t ice_recv_pkts_vec_avx512(void *rx_queue, struct rte_mbuf **rx_pkts, 276 uint16_t ice_recv_pkts_vec_avx512_offload(void *rx_queue, 279 uint16_t ice_recv_scattered_pkts_vec_avx512(void *rx_queue, [all …]
|
| H A D | ice_rxtx_vec_avx2.c | 709 ice_recv_pkts_vec_avx2(void *rx_queue, struct rte_mbuf **rx_pkts, in ice_recv_pkts_vec_avx2() argument 712 return _ice_recv_raw_pkts_vec_avx2(rx_queue, rx_pkts, in ice_recv_pkts_vec_avx2() 720 return _ice_recv_raw_pkts_vec_avx2(rx_queue, rx_pkts, in ice_recv_pkts_vec_avx2_offload() 733 struct ice_rx_queue *rxq = rx_queue; in ice_recv_scattered_burst_vec_avx2() 772 ice_recv_scattered_pkts_vec_avx2_common(void *rx_queue, in ice_recv_scattered_pkts_vec_avx2_common() argument 780 uint16_t burst = ice_recv_scattered_burst_vec_avx2(rx_queue, in ice_recv_scattered_pkts_vec_avx2_common() 787 return retval + ice_recv_scattered_burst_vec_avx2(rx_queue, in ice_recv_scattered_pkts_vec_avx2_common() 792 ice_recv_scattered_pkts_vec_avx2(void *rx_queue, in ice_recv_scattered_pkts_vec_avx2() argument 796 return ice_recv_scattered_pkts_vec_avx2_common(rx_queue, in ice_recv_scattered_pkts_vec_avx2() 803 ice_recv_scattered_pkts_vec_avx2_offload(void *rx_queue, in ice_recv_scattered_pkts_vec_avx2_offload() argument [all …]
|
| H A D | ice_rxtx_vec_avx512.c | 824 ice_recv_pkts_vec_avx512(void *rx_queue, struct rte_mbuf **rx_pkts, in ice_recv_pkts_vec_avx512() argument 835 ice_recv_pkts_vec_avx512_offload(void *rx_queue, struct rte_mbuf **rx_pkts, in ice_recv_pkts_vec_avx512_offload() argument 838 return _ice_recv_raw_pkts_vec_avx512(rx_queue, rx_pkts, in ice_recv_pkts_vec_avx512_offload() 851 struct ice_rx_queue *rxq = rx_queue; in ice_recv_scattered_burst_vec_avx512() 889 ice_recv_scattered_burst_vec_avx512_offload(void *rx_queue, in ice_recv_scattered_burst_vec_avx512_offload() argument 893 struct ice_rx_queue *rxq = rx_queue; in ice_recv_scattered_burst_vec_avx512_offload() 938 uint16_t burst = ice_recv_scattered_burst_vec_avx512(rx_queue, in ice_recv_scattered_pkts_vec_avx512() 945 return retval + ice_recv_scattered_burst_vec_avx512(rx_queue, in ice_recv_scattered_pkts_vec_avx512() 956 ice_recv_scattered_pkts_vec_avx512_offload(void *rx_queue, in ice_recv_scattered_pkts_vec_avx512_offload() argument 964 ice_recv_scattered_burst_vec_avx512_offload(rx_queue, in ice_recv_scattered_pkts_vec_avx512_offload() [all …]
|
| /dpdk/drivers/net/virtio/ |
| H A D | virtio_ethdev.h | 77 uint16_t virtio_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, 79 uint16_t virtio_recv_pkts_packed(void *rx_queue, struct rte_mbuf **rx_pkts, 82 uint16_t virtio_recv_mergeable_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, 85 uint16_t virtio_recv_mergeable_pkts_packed(void *rx_queue, 88 uint16_t virtio_recv_pkts_inorder(void *rx_queue, 102 uint16_t virtio_recv_pkts_vec(void *rx_queue, struct rte_mbuf **rx_pkts, 105 uint16_t virtio_recv_pkts_packed_vec(void *rx_queue, struct rte_mbuf **rx_pkts,
|
| /dpdk/drivers/net/thunderx/ |
| H A D | nicvf_rxtx.h | 86 uint32_t nicvf_dev_rx_queue_count(void *rx_queue); 93 uint16_t nicvf_recv_pkts_vlan_strip(void *rx_queue, struct rte_mbuf **rx_pkts, 95 uint16_t nicvf_recv_pkts_cksum_vlan_strip(void *rx_queue, 98 uint16_t nicvf_recv_pkts_multiseg_no_offload(void *rx_queue, 100 uint16_t nicvf_recv_pkts_multiseg_cksum(void *rx_queue, 102 uint16_t nicvf_recv_pkts_multiseg_vlan_strip(void *rx_queue, 104 uint16_t nicvf_recv_pkts_multiseg_cksum_vlan_strip(void *rx_queue,
|
| H A D | nicvf_rxtx.c | 428 struct nicvf_rxq *rxq = rx_queue; in nicvf_recv_pkts() 491 return nicvf_recv_pkts(rx_queue, rx_pkts, nb_pkts, in nicvf_recv_pkts_no_offload() 499 return nicvf_recv_pkts(rx_queue, rx_pkts, nb_pkts, in nicvf_recv_pkts_cksum() 507 return nicvf_recv_pkts(rx_queue, rx_pkts, nb_pkts, in nicvf_recv_pkts_vlan_strip() 515 return nicvf_recv_pkts(rx_queue, rx_pkts, nb_pkts, in nicvf_recv_pkts_cksum_vlan_strip() 581 struct nicvf_rxq *rxq = rx_queue; in nicvf_recv_pkts_multiseg() 623 return nicvf_recv_pkts_multiseg(rx_queue, rx_pkts, nb_pkts, in nicvf_recv_pkts_multiseg_no_offload() 631 return nicvf_recv_pkts_multiseg(rx_queue, rx_pkts, nb_pkts, in nicvf_recv_pkts_multiseg_cksum() 644 nicvf_recv_pkts_multiseg_cksum_vlan_strip(void *rx_queue, in nicvf_recv_pkts_multiseg_cksum_vlan_strip() argument 652 nicvf_dev_rx_queue_count(void *rx_queue) in nicvf_dev_rx_queue_count() argument [all …]
|
| /dpdk/examples/l3fwd-power/ |
| H A D | main.c | 884 port_id = rx_queue->port_id; in turn_on_off_intr() 907 portid = rx_queue->port_id; in event_register() 994 rx_queue->idle_hint = 0; in main_intr_loop() 995 portid = rx_queue->port_id; in main_intr_loop() 1151 portid = rx_queue->port_id; in main_telemetry_loop() 1265 rx_queue->idle_hint = 0; in main_empty_poll_loop() 1266 portid = rx_queue->port_id; in main_empty_poll_loop() 1392 rx_queue->idle_hint = 0; in main_legacy_loop() 1393 portid = rx_queue->port_id; in main_legacy_loop() 1424 rx_queue->freq_up_hint = in main_legacy_loop() [all …]
|
| /dpdk/drivers/crypto/qat/dev/ |
| H A D | qat_sym_pmd_gen1.c | 948 head = (head + rx_queue->msg_size) & rx_queue->modulo_mask; in qat_sym_dp_dequeue_burst_gen1() 967 head = (head + rx_queue->msg_size) & in qat_sym_dp_dequeue_burst_gen1() 968 rx_queue->modulo_mask; in qat_sym_dp_dequeue_burst_gen1() 981 head = (head + rx_queue->msg_size) & in qat_sym_dp_dequeue_burst_gen1() 982 rx_queue->modulo_mask; in qat_sym_dp_dequeue_burst_gen1() 1009 rx_queue->modulo_mask; in qat_sym_dp_dequeue_single_gen1() 1053 rx_queue->head = dp_ctx->head; in qat_sym_dp_dequeue_done_gen1() 1061 old_head = rx_queue->csr_head; in qat_sym_dp_dequeue_done_gen1() 1062 new_head = rx_queue->head; in qat_sym_dp_dequeue_done_gen1() 1078 rx_queue->csr_head = new_head; in qat_sym_dp_dequeue_done_gen1() [all …]
|
| /dpdk/drivers/net/ark/ |
| H A D | ark_ethdev_rx.h | 20 uint32_t eth_ark_dev_rx_queue_count(void *rx_queue); 23 uint16_t eth_ark_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, 25 void eth_ark_dev_rx_queue_release(void *rx_queue);
|
| /dpdk/lib/eventdev/ |
| H A D | rte_event_eth_rx_adapter.c | 387 return dev_info->rx_queue && in rxa_intr_queue() 400 dev_info->rx_queue && in rxa_polled_queue() 666 &dev_info->rx_queue[q]; in rxa_calc_wrr_sequence() 2174 rx_queue = dev_info->rx_queue; in rxa_sw_add() 2178 dev_info->rx_queue = in rxa_sw_add() 2265 if (rx_queue == NULL) { in rxa_sw_add() 2266 rte_free(dev_info->rx_queue); in rxa_sw_add() 2267 dev_info->rx_queue = NULL; in rxa_sw_add() 2661 dev_info->rx_queue = in rte_event_eth_rx_adapter_queue_add() 2765 dev_info->rx_queue = NULL; in rte_event_eth_rx_adapter_queue_del() [all …]
|
| /dpdk/drivers/net/i40e/ |
| H A D | i40e_rxtx.h | 206 uint16_t i40e_recv_pkts(void *rx_queue, 209 uint16_t i40e_recv_scattered_pkts(void *rx_queue, 232 uint32_t i40e_dev_rx_queue_count(void *rx_queue); 233 int i40e_dev_rx_descriptor_status(void *rx_queue, uint16_t offset); 236 uint16_t i40e_recv_pkts_vec(void *rx_queue, struct rte_mbuf **rx_pkts, 238 uint16_t i40e_recv_scattered_pkts_vec(void *rx_queue, 253 uint16_t i40e_recv_pkts_vec_avx2(void *rx_queue, struct rte_mbuf **rx_pkts, 255 uint16_t i40e_recv_scattered_pkts_vec_avx2(void *rx_queue, 259 int i40e_get_monitor_addr(void *rx_queue, struct rte_power_monitor_cond *pmc); 260 uint16_t i40e_recv_pkts_vec_avx512(void *rx_queue, [all …]
|
| /dpdk/drivers/net/ngbe/ |
| H A D | ngbe_ethdev.h | 204 uint32_t ngbe_dev_rx_queue_count(void *rx_queue); 206 int ngbe_dev_rx_descriptor_status(void *rx_queue, uint16_t offset); 241 uint16_t ngbe_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, 244 uint16_t ngbe_recv_pkts_bulk_alloc(void *rx_queue, struct rte_mbuf **rx_pkts, 247 uint16_t ngbe_recv_pkts_sc_single_alloc(void *rx_queue, 249 uint16_t ngbe_recv_pkts_sc_bulk_alloc(void *rx_queue,
|
| /dpdk/drivers/net/axgbe/ |
| H A D | axgbe_rxtx.h | 183 uint16_t axgbe_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, 185 uint16_t eth_axgbe_recv_scattered_pkts(void *rx_queue, 187 uint16_t axgbe_recv_pkts_threshold_refresh(void *rx_queue, 191 int axgbe_dev_rx_descriptor_status(void *rx_queue, uint16_t offset);
|
| /dpdk/drivers/common/qat/ |
| H A D | qat_qp.c | 810 struct qat_queue *rx_queue; in qat_dequeue_op_burst() local 817 rx_queue = &(tmp_qp->rx_q); in qat_dequeue_op_burst() 818 head = rx_queue->head; in qat_dequeue_op_burst() 819 resp_msg = (uint8_t *)rx_queue->base_addr + rx_queue->head; in qat_dequeue_op_burst() 828 tmp_qp->op_cookies[head >> rx_queue->trailz], in qat_dequeue_op_burst() 831 head = adf_modulo(head + rx_queue->msg_size, in qat_dequeue_op_burst() 832 rx_queue->modulo_mask); in qat_dequeue_op_burst() 834 resp_msg = (uint8_t *)rx_queue->base_addr + head; in qat_dequeue_op_burst() 852 rx_queue->nb_processed_responses++; in qat_dequeue_op_burst() 858 rx_queue->head = head; in qat_dequeue_op_burst() [all …]
|
| /dpdk/app/test-pmd/ |
| H A D | shared_rxq_fwd.c | 35 fsm[sm_id]->rx_queue == fs->rx_queue) in forward_stream_get() 101 nb_rx = rte_eth_rx_burst(fs->rx_port, fs->rx_queue, pkts_burst, in shared_rxq_fwd()
|
| /dpdk/drivers/net/fm10k/ |
| H A D | fm10k.h | 318 uint16_t fm10k_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, 321 uint16_t fm10k_recv_scattered_pkts(void *rx_queue, 325 fm10k_dev_rx_queue_count(void *rx_queue); 328 fm10k_dev_rx_descriptor_status(void *rx_queue, uint16_t offset); 331 fm10k_dev_tx_descriptor_status(void *rx_queue, uint16_t offset);
|
| /dpdk/drivers/net/atlantic/ |
| H A D | atl_ethdev.h | 69 uint32_t atl_rx_queue_count(void *rx_queue); 71 int atl_dev_rx_descriptor_status(void *rx_queue, uint16_t offset); 98 uint16_t atl_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
|
| /dpdk/drivers/bus/fslmc/portal/ |
| H A D | dpaa2_hw_dpci.c | 84 rxq = &(dpci_node->rx_queue[i]); in rte_dpaa2_create_dpci_device() 121 dpci_node->rx_queue[i].fqid = rx_attr.fqid; in rte_dpaa2_create_dpci_device() 144 struct dpaa2_queue *rxq = &(dpci_node->rx_queue[i]); in rte_dpaa2_create_dpci_device()
|
| /dpdk/drivers/net/e1000/ |
| H A D | e1000_ethdev.h | 403 uint32_t eth_igb_rx_queue_count(void *rx_queue); 405 int eth_igb_rx_descriptor_status(void *rx_queue, uint16_t offset); 477 uint32_t eth_em_rx_queue_count(void *rx_queue); 479 int eth_em_rx_descriptor_status(void *rx_queue, uint16_t offset); 499 uint16_t eth_em_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, 502 uint16_t eth_em_recv_scattered_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
|