Home
last modified time | relevance | path

Searched refs:rx_ring (Results 1 – 25 of 70) sorted by relevance

123

/dpdk/drivers/net/enetc/
H A Denetc_rxtx.c134 i = rx_ring->next_to_use; in enetc_refill_rx_ring()
135 mb_pool = rx_ring->mb_pool; in enetc_refill_rx_ring()
136 rx_swbd = &rx_ring->q_swbd[i]; in enetc_refill_rx_ring()
137 rxbd = ENETC_RXBD(*rx_ring, i); in enetc_refill_rx_ring()
166 rx_ring->next_to_alloc = i; in enetc_refill_rx_ring()
167 rx_ring->next_to_use = i; in enetc_refill_rx_ring()
331 i = rx_ring->next_to_clean; in enetc_clean_rx_ring()
335 bd_count = rx_ring->bd_count; in enetc_clean_rx_ring()
362 rx_ring->crc_len; in enetc_clean_rx_ring()
364 rx_ring->crc_len; in enetc_clean_rx_ring()
[all …]
H A Denetc_ethdev.c412 int idx = rx_ring->index; in enetc_setup_rxbdr()
425 rx_ring->mb_pool = mb_pool; in enetc_setup_rxbdr()
428 enetc_refill_rx_ring(rx_ring, (enetc_bd_unused(rx_ring))); in enetc_setup_rxbdr()
444 struct enetc_bdr *rx_ring; in enetc_rx_queue_setup() local
455 if (rx_ring == NULL) { in enetc_rx_queue_setup()
466 rx_ring->ndev = dev; in enetc_rx_queue_setup()
486 rte_free(rx_ring); in enetc_rx_queue_setup()
527 enetc_free_bdr(rx_ring); in enetc_rx_queue_release()
528 rte_free(rx_ring); in enetc_rx_queue_release()
748 struct enetc_bdr *rx_ring; in enetc_rx_queue_start() local
[all …]
H A Denetc.h108 int enetc_refill_rx_ring(struct enetc_bdr *rx_ring, const int buff_cnt);
/dpdk/drivers/net/bnxt/
H A Dbnxt_ring.c112 struct bnxt_ring *tx_ring, *rx_ring; in bnxt_alloc_rings() local
243 rx_ring = rx_ring_info->rx_ring_struct; in bnxt_alloc_rings()
250 rx_ring->mem_zone = (const void *)mz; in bnxt_alloc_rings()
252 if (!rx_ring->bd) in bnxt_alloc_rings()
254 if (rx_ring->vmem_size) { in bnxt_alloc_rings()
255 rx_ring->vmem = in bnxt_alloc_rings()
269 rx_ring->mem_zone = (const void *)mz; in bnxt_alloc_rings()
271 if (!rx_ring->bd) in bnxt_alloc_rings()
273 if (rx_ring->vmem_size) { in bnxt_alloc_rings()
274 rx_ring->vmem = in bnxt_alloc_rings()
[all …]
H A Dbnxt_rxq.c206 if (!rxq || !rxq->rx_ring) in bnxt_rx_queue_release_mbufs()
209 sw_ring = rxq->rx_ring->rx_buf_ring; in bnxt_rx_queue_release_mbufs()
236 sw_ring = rxq->rx_ring->ag_buf_ring; in bnxt_rx_queue_release_mbufs()
248 tpa_info = rxq->rx_ring->tpa_info; in bnxt_rx_queue_release_mbufs()
278 if (rxq->rx_ring) { in bnxt_free_rxq_mem()
280 rte_free(rxq->rx_ring->rx_ring_struct); in bnxt_free_rxq_mem()
281 rxq->rx_ring->rx_ring_struct = NULL; in bnxt_free_rxq_mem()
284 rte_free(rxq->rx_ring->ag_ring_struct); in bnxt_free_rxq_mem()
285 rxq->rx_ring->ag_ring_struct = NULL; in bnxt_free_rxq_mem()
287 rte_free(rxq->rx_ring); in bnxt_free_rxq_mem()
[all …]
H A Dbnxt_rxr.c150 rxr = rxq->rx_ring; in bnxt_rx_ring_reset()
232 struct bnxt_rx_ring_info *rxr = rxq->rx_ring; in bnxt_tpa_start()
1164 rte_free(rxq->rx_ring->rx_ring_struct); in bnxt_free_rx_rings()
1168 rte_free(rxq->rx_ring->ag_ring_struct); in bnxt_free_rx_rings()
1169 rxq->rx_ring->ag_ring_struct = NULL; in bnxt_free_rx_rings()
1171 rte_free(rxq->rx_ring); in bnxt_free_rx_rings()
1193 if (rxq->rx_ring != NULL) { in bnxt_init_rx_ring_struct()
1194 rxr = rxq->rx_ring; in bnxt_init_rx_ring_struct()
1202 rxq->rx_ring = rxr; in bnxt_init_rx_ring_struct()
1261 rxr = rxq->rx_ring; in bnxt_init_rx_ring_struct()
[all …]
H A Dbnxt_reps.c58 rep_rxr = rep_rxq->rx_ring; in bnxt_vfr_recv()
92 rxr = rxq->rx_ring; in bnxt_rep_rx_burst()
403 if (!rxq || !rxq->rx_ring) in bnxt_vfr_rx_queue_release_mbufs()
406 sw_ring = rxq->rx_ring->rx_buf_ring; in bnxt_vfr_rx_queue_release_mbufs()
408 for (i = 0; i < rxq->rx_ring->rx_ring_struct->ring_size; i++) { in bnxt_vfr_rx_queue_release_mbufs()
605 rxq->rx_ring = rxr; in bnxt_init_rep_rx_ring()
685 rxq->rx_ring->rx_ring_struct->ring_size, in bnxt_rep_rx_queue_setup_op()
693 rxq->rx_ring->rx_buf_ring = buf_ring; in bnxt_rep_rx_queue_setup_op()
715 bnxt_free_ring(rxq->rx_ring->rx_ring_struct); in bnxt_rep_rx_queue_release_op()
716 rte_free(rxq->rx_ring->rx_ring_struct); in bnxt_rep_rx_queue_release_op()
[all …]
/dpdk/drivers/net/ena/
H A Dena_ethdev.c945 queues = adapter->rx_ring; in ena_queue_start_all()
1359 queues = adapter->rx_ring; in ena_queue_stop_all()
1523 rxq = &adapter->rx_ring[queue_idx]; in ena_rx_queue_setup()
2523 mbuf_head->port = rx_ring->port_id; in ena_rx_mbuf()
2611 descs_in_use = rx_ring->ring_size - in eth_ena_recv_pkts()
2622 rx_ring->ena_com_io_sq, in eth_ena_recv_pkts()
2629 ++rx_ring->rx_stats.bad_desc_num; in eth_ena_recv_pkts()
2633 ++rx_ring->rx_stats.bad_req_id; in eth_ena_recv_pkts()
2640 mbuf = ena_rx_mbuf(rx_ring, in eth_ena_recv_pkts()
2648 rx_ring->ena_bufs[i].req_id; in eth_ena_recv_pkts()
[all …]
/dpdk/examples/multi_process/client_server_mp/mp_client/
H A Dclient.c206 struct rte_ring *rx_ring; in main() local
225 rx_ring = rte_ring_lookup(get_rx_queue_name(client_id)); in main()
226 if (rx_ring == NULL) in main()
249 rx_pkts = rte_ring_dequeue_burst(rx_ring, pkts, in main()
/dpdk/examples/server_node_efd/node/
H A Dnode.c309 struct rte_ring *rx_ring; in main() local
331 rx_ring = rte_ring_lookup(get_rx_queue_name(node_id)); in main()
332 if (rx_ring == NULL) in main()
369 unlikely(rte_ring_dequeue_bulk(rx_ring, pkts, in main()
371 rx_pkts = (uint16_t)RTE_MIN(rte_ring_count(rx_ring), in main()
/dpdk/app/pdump/
H A Dmain.c123 struct rte_ring *rx_ring; member
508 rte_ring_free(pt->rx_ring); in cleanup_rings()
662 pt->rx_ring = rte_ring_create(ring_name, pt->ring_size, in create_mp_ring_vdev()
664 if (pt->rx_ring == NULL) { in create_mp_ring_vdev()
745 pt->rx_ring = rte_ring_create(ring_name, pt->ring_size, in create_mp_ring_vdev()
747 if (pt->rx_ring == NULL) { in create_mp_ring_vdev()
834 pt->rx_ring, in enable_pdump()
845 pt->rx_ring, pt->mp, NULL); in enable_pdump()
855 pt->dir, pt->rx_ring, in enable_pdump()
860 pt->rx_ring, pt->mp, NULL); in enable_pdump()
[all …]
/dpdk/examples/qos_sched/
H A Dmain.c56 flow->rx_thread.rx_ring = flow->rx_ring; in app_main_loop()
74 flow->wt_thread.rx_ring = flow->rx_ring; in app_main_loop()
H A Dapp_thread.c92 if (unlikely(rte_ring_sp_enqueue_bulk(conf->rx_ring, in app_rx_thread()
203 nb_pkt = rte_ring_sc_dequeue_burst(conf->rx_ring, (void **)mbufs, in app_worker_thread()
239 nb_pkt = rte_ring_sc_dequeue_burst(conf->rx_ring, (void **)mbufs, in app_mixed_thread()
H A Dmain.h86 struct rte_ring *rx_ring; member
105 struct rte_ring *rx_ring; member
/dpdk/drivers/net/e1000/
H A Dem_rxtx.c701 rx_ring = rxq->rx_ring; in eth_em_recv_pkts()
712 rxdp = &rx_ring[rx_id]; in eth_em_recv_pkts()
775 rte_em_prefetch(&rx_ring[rx_id]); in eth_em_recv_pkts()
877 rx_ring = rxq->rx_ring; in eth_em_recv_scattered_pkts()
896 rxdp = &rx_ring[rx_id]; in eth_em_recv_scattered_pkts()
954 rte_em_prefetch(&rx_ring[rx_id]); in eth_em_recv_scattered_pkts()
1500 rxdp = &(rxq->rx_ring[rxq->rx_tail]); in eth_em_rx_queue_count()
1531 status = &rxq->rx_ring[desc].status; in eth_em_rx_descriptor_status()
1701 rxq->rx_ring[i] = rxd_init; in em_alloc_rx_queue_mbufs()
1703 rxd = &rxq->rx_ring[i]; in em_alloc_rx_queue_mbufs()
[all …]
H A Digb_rxtx.c809 volatile union e1000_adv_rx_desc *rx_ring; in eth_igb_recv_pkts() local
829 rx_ring = rxq->rx_ring; in eth_igb_recv_pkts()
840 rxdp = &rx_ring[rx_id]; in eth_igb_recv_pkts()
902 rte_igb_prefetch(&rx_ring[rx_id]); in eth_igb_recv_pkts()
1016 rx_ring = rxq->rx_ring; in eth_igb_recv_scattered_pkts()
1035 rxdp = &rx_ring[rx_id]; in eth_igb_recv_scattered_pkts()
1093 rte_igb_prefetch(&rx_ring[rx_id]); in eth_igb_recv_scattered_pkts()
1626 rxq->rx_ring[i] = zeroed_desc; in igb_reset_rx_queue()
1785 rxdp = &(rxq->rx_ring[rxq->rx_tail]); in eth_igb_rx_queue_count()
1792 rxdp = &(rxq->rx_ring[rxq->rx_tail + in eth_igb_rx_queue_count()
[all …]
/dpdk/drivers/net/bnx2x/
H A Dbnx2x_rxtx.c98 rxq->rx_ring = (uint64_t*)dma->addr; in bnx2x_dev_rx_queue_setup()
99 memset((void *)rxq->rx_ring, 0, dma_size); in bnx2x_dev_rx_queue_setup()
103 rx_bd = &rxq->rx_ring[TOTAL_RX_BD_PER_PAGE * j - 2]; in bnx2x_dev_rx_queue_setup()
129 rxq->rx_ring[idx] = in bnx2x_dev_rx_queue_setup()
412 rxq->rx_ring[bd_prod] = in bnx2x_recv_pkts()
418 rte_prefetch0(&rxq->rx_ring[rx_pref]); in bnx2x_recv_pkts()
H A Dbnx2x_rxtx.h28 uint64_t *rx_ring; /**< RX ring virtual address. */ member
/dpdk/drivers/net/igc/
H A Digc_txrx.c108 volatile union igc_adv_rx_desc *rx_ring; member
352 volatile union igc_adv_rx_desc * const rx_ring = rxq->rx_ring; in igc_recv_pkts() local
375 rxdp = &rx_ring[rx_id]; in igc_recv_pkts()
438 rte_igc_prefetch(&rx_ring[rx_id]); in igc_recv_pkts()
496 volatile union igc_adv_rx_desc * const rx_ring = rxq->rx_ring; in igc_recv_scattered_pkts() local
523 rxdp = &rx_ring[rx_id]; in igc_recv_scattered_pkts()
582 rte_igc_prefetch(&rx_ring[rx_id]); in igc_recv_scattered_pkts()
740 rxdp = &rxq->rx_ring[rxq->rx_tail]; in eth_igc_rx_queue_count()
801 rxd = &rxq->rx_ring[i]; in igc_alloc_rx_queue_mbufs()
1307 rxq->rx_ring[i] = zeroed_desc; in igc_reset_rx_queue()
[all …]
/dpdk/drivers/net/iavf/
H A Diavf_rxtx.c84 rxdp = &rxq->rx_ring[desc]; in iavf_get_monitor_addr()
306 rxd = &rxq->rx_ring[i]; in alloc_rxq_mbufs()
1322 rx_ring = rxq->rx_ring; in iavf_recv_pkts()
1326 rxdp = &rx_ring[rx_id]; in iavf_recv_pkts()
1361 rte_prefetch0(&rx_ring[rx_id]); in iavf_recv_pkts()
1430 rx_ring = rxq->rx_ring; in iavf_recv_pkts_flex_rxd()
1479 rte_prefetch0(&rx_ring[rx_id]); in iavf_recv_pkts_flex_rxd()
1551 volatile union iavf_rx_desc *rx_ring = rxq->rx_ring; in iavf_recv_scattered_pkts_flex_rxd() local
1727 volatile union iavf_rx_desc *rx_ring = rxq->rx_ring; in iavf_recv_scattered_pkts() local
1732 rxdp = &rx_ring[rx_id]; in iavf_recv_scattered_pkts()
[all …]
/dpdk/drivers/net/netvsc/
H A Dhn_rxtx.c921 qinfo->nb_desc = rxq->rx_ring->size; in hn_dev_rx_queue_info()
960 rxq->rx_ring = rte_ring_create(ring_name, in hn_dev_rx_queue_setup()
963 if (!rxq->rx_ring) in hn_dev_rx_queue_setup()
975 rte_ring_free(rxq->rx_ring); in hn_dev_rx_queue_setup()
989 rte_ring_free(rxq->rx_ring); in hn_rx_queue_free()
990 rxq->rx_ring = NULL; in hn_rx_queue_free()
1023 return rte_ring_count(rxq->rx_ring); in hn_dev_rx_queue_count()
1039 if (offset >= rxq->rx_ring->capacity) in hn_dev_rx_queue_status()
1042 if (offset < rte_ring_count(rxq->rx_ring)) in hn_dev_rx_queue_status()
1618 if (likely(rte_ring_count(rxq->rx_ring) < nb_pkts)) in hn_recv_pkts()
[all …]
/dpdk/drivers/net/hns3/
H A Dhns3_rxtx_vec.c60 struct hns3_desc *rxdp = rxq->rx_ring + rxq->rx_rearm_start; in hns3_rxq_rearm_mbuf()
112 struct hns3_desc *rxdp = &rxq->rx_ring[rxq->next_to_use]; in hns3_recv_pkts_vec()
202 memset(&rxq->rx_ring[rxq->nb_rx_desc], 0, in hns3_rxq_vec_setup()
/dpdk/drivers/net/ngbe/
H A Dngbe_rxtx.c1007 rxdp = &rxq->rx_ring[rxq->rx_tail]; in ngbe_rx_scan_hw_ring()
1096 rxdp = &rxq->rx_ring[alloc_idx]; in ngbe_rx_alloc_bufs()
1237 volatile struct ngbe_rx_desc *rx_ring; in ngbe_recv_pkts() local
1257 rx_ring = rxq->rx_ring; in ngbe_recv_pkts()
1269 rxdp = &rx_ring[rx_id]; in ngbe_recv_pkts()
1332 rte_ngbe_prefetch(&rx_ring[rx_id]); in ngbe_recv_pkts()
1469 volatile struct ngbe_rx_desc *rx_ring = rxq->rx_ring; in ngbe_recv_pkts_sc() local
1493 rxdp = &rx_ring[rx_id]; in ngbe_recv_pkts_sc()
2204 rxq->rx_ring[i] = zeroed_desc; in ngbe_reset_rx_queue()
2385 rxdp = &rxq->rx_ring[rxq->rx_tail]; in ngbe_dev_rx_queue_count()
[all …]
/dpdk/drivers/net/sfc/
H A Dsfc_repr_proxy_api.h30 uint16_t queue_id, struct rte_ring *rx_ring,
/dpdk/drivers/net/i40e/
H A Di40e_rxtx.c102 rxdp = &rxq->rx_ring[desc]; in i40e_get_monitor_addr()
466 rxdp = &rxq->rx_ring[rxq->rx_tail]; in i40e_rx_scan_hw_ring()
594 rxdp = &rxq->rx_ring[alloc_idx]; in i40e_rx_alloc_bufs()
705 volatile union i40e_rx_desc *rx_ring; in i40e_recv_pkts() local
726 rx_ring = rxq->rx_ring; in i40e_recv_pkts()
731 rxdp = &rx_ring[rx_id]; in i40e_recv_pkts()
769 rte_prefetch0(&rx_ring[rx_id]); in i40e_recv_pkts()
835 volatile union i40e_rx_desc *rx_ring = rxq->rx_ring; in i40e_recv_scattered_pkts() local
853 rxdp = &rx_ring[rx_id]; in i40e_recv_scattered_pkts()
891 rte_prefetch0(&rx_ring[rx_id]); in i40e_recv_scattered_pkts()
[all …]

123