Home
last modified time | relevance | path

Searched refs:rxq (Results 1 – 25 of 276) sorted by relevance

12345678910>>...12

/f-stack/dpdk/drivers/net/sfc/
H A Dsfc_ef10_essb_rx.c149 rxd = &rxq->sw_ring[rxq->completed & rxq->rxq_ptr_mask]; in sfc_ef10_essb_maybe_next_completed()
154 rxd = &rxq->sw_ring[rxq->completed & rxq->rxq_ptr_mask]; in sfc_ef10_essb_maybe_next_completed()
155 rxq->left_in_completed = rxq->block_size; in sfc_ef10_essb_maybe_next_completed()
229 *rx_ev = rxq->evq_hw_ring[rxq->evq_read_ptr & rxq->evq_ptr_mask]; in sfc_ef10_essb_rx_event_get()
270 rxd = &rxq->sw_ring[rxq->pending_id]; in sfc_ef10_essb_rx_process_ev()
272 rxq->block_size - rxq->left_in_pending); in sfc_ef10_essb_rx_process_ev()
281 rxq->left_in_pending = rxq->block_size; in sfc_ef10_essb_rx_process_ev()
282 if (rxq->pending_id != rxq->rxq_ptr_mask) in sfc_ef10_essb_rx_process_ev()
476 if (offset < (rxq->added - rxq->completed) * rxq->block_size + in sfc_ef10_essb_rx_qdesc_status()
673 rxq->completed = rxq->pending_id = rxq->added = 0; in sfc_ef10_essb_rx_qstart()
[all …]
H A Dsfc_ef10_rx.c98 sfc_ef10_ev_qprime(rxq->evq_prime, rxq->evq_read_ptr, rxq->ptr_mask); in sfc_ef10_rx_qprime()
99 rxq->evq_read_ptr_primed = rxq->evq_read_ptr; in sfc_ef10_rx_qprime()
210 SFC_ASSERT(rxq->pending == rxq->completed || rxq->scatter_pkt == NULL); in sfc_ef10_rx_pending()
219 rxq->sw_ring[completed++ & rxq->ptr_mask].mbuf; in sfc_ef10_rx_pending()
412 *rx_ev = rxq->evq_hw_ring[rxq->evq_read_ptr & rxq->ptr_mask]; in sfc_ef10_rx_get_event()
466 rxq->evq_read_ptr_primed != rxq->evq_read_ptr) in sfc_ef10_recv_pkts()
576 if (offset < (rxq->added - rxq->completed)) in sfc_ef10_rx_qdesc_status()
655 rxq = rte_zmalloc_socket("sfc-ef10-rxq", sizeof(*rxq), in sfc_ef10_rx_qcreate()
779 for (i = rxq->completed; i != rxq->added; ++i) { in sfc_ef10_rx_qpurge()
780 rxd = &rxq->sw_ring[i & rxq->ptr_mask]; in sfc_ef10_rx_qpurge()
[all …]
H A Dsfc_ef100_rx.c99 rxq->evq_read_ptr & rxq->ptr_mask); in sfc_ef100_rx_qprime()
100 rxq->evq_read_ptr_primed = rxq->evq_read_ptr; in sfc_ef100_rx_qprime()
432 m = rxq->sw_ring[rxq->completed++ & rxq->ptr_mask].mbuf; in sfc_ef100_rx_next_mbuf()
435 if (unlikely(rxq->completed == rxq->added)) in sfc_ef100_rx_next_mbuf()
444 id = rxq->completed & rxq->ptr_mask; in sfc_ef100_rx_next_mbuf()
447 if (unlikely(rxq->completed + 1 == rxq->added)) in sfc_ef100_rx_next_mbuf()
533 *ev = rxq->evq_hw_ring[rxq->evq_read_ptr & rxq->ptr_mask]; in sfc_ef100_rx_get_event()
547 rxq->evq_read_ptr, rxq->evq_read_ptr & rxq->ptr_mask, in sfc_ef100_rx_get_event()
557 rxq->evq_read_ptr & rxq->ptr_mask); in sfc_ef100_rx_get_event()
856 rxd = &rxq->sw_ring[i & rxq->ptr_mask]; in sfc_ef100_rx_qpurge()
[all …]
H A Dsfc_rx.c60 if (rxq->evq->read_ptr_primed != rxq->evq->read_ptr) { in sfc_efx_rx_qprime()
63 rxq->evq->read_ptr_primed = rxq->evq->read_ptr; in sfc_efx_rx_qprime()
123 efx_rx_qpost(rxq->common, addr, rxq->buf_size, in sfc_efx_rx_qrefill()
130 efx_rx_qpush(rxq->common, added, &rxq->pushed); in sfc_efx_rx_qrefill()
339 return rxq->pending - rxq->completed; in sfc_efx_rx_qdesc_npending()
370 if (offset < (rxq->pending - rxq->completed)) in sfc_efx_rx_qdesc_status()
373 if (offset < (rxq->added - rxq->completed)) in sfc_efx_rx_qdesc_status()
491 rxq->evq = sfc_rxq_by_dp_rxq(&rxq->dp)->evq; in sfc_efx_rx_qcreate()
549 rxq->pending = rxq->completed = rxq->added = rxq->pushed = 0; in sfc_efx_rx_qstart()
590 for (i = rxq->completed; i != rxq->added; ++i) { in sfc_efx_rx_qpurge()
[all …]
/f-stack/dpdk/drivers/net/mlx5/
H A Dmlx5_rxtx_vec.c91 uint16_t n = q_n - (rxq->rq_ci - rxq->rq_pi); in mlx5_rx_replenish_bulk_mbuf()
129 (*rxq->elts)[elts_idx + i] = &rxq->fake_mbuf; in mlx5_rx_replenish_bulk_mbuf()
131 *rxq->rq_db = rte_cpu_to_be_32(rxq->rq_ci); in mlx5_rx_replenish_bulk_mbuf()
148 uint32_t n = elts_n - (rxq->elts_ci - rxq->rq_pi); in mlx5_rx_mprq_replenish_bulk_mbuf()
154 rxq->elts_ci - rxq->rq_pi <= rxq->rq_repl_thresh) { in mlx5_rx_mprq_replenish_bulk_mbuf()
170 (*rxq->elts)[elts_idx + i] = &rxq->fake_mbuf; in mlx5_rx_mprq_replenish_bulk_mbuf()
250 *rxq->cq_db = rte_cpu_to_be_32(rxq->cq_ci); in rxq_copy_mprq_mbuf_v()
254 *rxq->rq_db = rte_cpu_to_be_32(rxq->rq_ci); in rxq_copy_mprq_mbuf_v()
296 MLX5_ASSERT(rxq->cqe_n == rxq->elts_n); in rxq_burst_v()
344 rxq->cq_ci += rxq->decompressed; in rxq_burst_v()
[all …]
H A Dmlx5_rxq.c109 (rxq, struct mlx5_rxq_ctrl, rxq); in mlx5_mprq_enabled()
156 struct mlx5_rxq_data *rxq = &rxq_ctrl->rxq; in rxq_alloc_elts_mprq() local
177 rxq->port_id, rxq->idx, wqe_n); in rxq_alloc_elts_mprq()
189 rxq->port_id, rxq->idx); in rxq_alloc_elts_mprq()
240 struct mlx5_rxq_data *rxq = &rxq_ctrl->rxq; in rxq_alloc_elts_sprq() local
317 struct mlx5_rxq_data *rxq = &rxq_ctrl->rxq; in rxq_free_elts_mprq() local
321 rxq->port_id, rxq->idx, (1u << rxq->elts_n)); in rxq_free_elts_mprq()
344 struct mlx5_rxq_data *rxq = &rxq_ctrl->rxq; in rxq_free_elts_sprq() local
363 rxq->rq_pi = rxq->rq_ci; in rxq_free_elts_sprq()
1249 (rxq, struct mlx5_rxq_ctrl, rxq); in mlx5_mprq_alloc_mp()
[all …]
/f-stack/dpdk/drivers/net/hns3/
H A Dhns3_rxtx_vec.c53 struct hns3_entry *rxep = &rxq->sw_ring[rxq->rx_rearm_start]; in hns3_rxq_rearm_mbuf()
54 struct hns3_desc *rxdp = rxq->rx_ring + rxq->rx_rearm_start; in hns3_rxq_rearm_mbuf()
92 if (rxq->rx_rearm_start >= rxq->nb_rx_desc) in hns3_rxq_rearm_mbuf()
106 struct hns3_desc *rxdp = &rxq->rx_ring[rxq->next_to_use]; in hns3_recv_pkts_vec()
122 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 0].mbuf); in hns3_recv_pkts_vec()
123 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 1].mbuf); in hns3_recv_pkts_vec()
124 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 2].mbuf); in hns3_recv_pkts_vec()
125 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 3].mbuf); in hns3_recv_pkts_vec()
158 memset(&rxq->rx_ring[rxq->nb_rx_desc], 0, in hns3_rxq_vec_setup()
161 memset(&rxq->fake_mbuf, 0, sizeof(rxq->fake_mbuf)); in hns3_rxq_vec_setup()
[all …]
/f-stack/dpdk/drivers/net/bnxt/
H A Dbnxt_rxq.c25 if (rxq && rxq->cp_ring && rxq->cp_ring->hw_stats) in bnxt_free_rxq_stats()
55 rxq->vnic = vnic; in bnxt_mq_rx_configure()
210 if (!rxq || !rxq->rx_ring) in bnxt_rx_queue_release_mbufs()
266 if (rxq) { in bnxt_rx_queue_release_op()
293 rte_free(rxq); in bnxt_rx_queue_release_op()
328 if (rxq) in bnxt_rx_queue_setup_op()
333 if (!rxq) { in bnxt_rx_queue_setup_op()
337 rxq->bp = bp; in bnxt_rx_queue_setup_op()
414 if (!rxq) in bnxt_rx_queue_intr_enable_op()
437 if (!rxq) in bnxt_rx_queue_intr_disable_op()
[all …]
H A Dbnxt_rxtx_vec_common.h44 mb_def.port = rxq->port_id; in bnxt_rxq_vec_setup_common()
50 rxq->mbuf_initializer = *(uint64_t *)p; in bnxt_rxq_vec_setup_common()
51 rxq->rxrearm_nb = 0; in bnxt_rxq_vec_setup_common()
52 rxq->rxrearm_start = 0; in bnxt_rxq_vec_setup_common()
67 nb = RTE_MIN(rxq->rxrearm_nb & ~0x3, in bnxt_rxq_rearm()
68 rxq->nb_rx_desc - rxq->rxrearm_start); in bnxt_rxq_rearm()
75 rx_bufs[i] = &rxq->fake_mbuf; in bnxt_rxq_rearm()
90 rxq->rxrearm_start += nb; in bnxt_rxq_rearm()
92 if (rxq->rxrearm_start >= rxq->nb_rx_desc) in bnxt_rxq_rearm()
93 rxq->rxrearm_start = 0; in bnxt_rxq_rearm()
[all …]
/f-stack/dpdk/drivers/net/mlx4/
H A Dmlx4_rxq.c190 struct rxq *rxq = NULL; in mlx4_rss_attach() local
438 if (rxq) in mlx4_rss_init()
487 mlx4_rxq_attach(struct rxq *rxq) in mlx4_rxq_attach() argument
568 rxq->mp->name, rxq->mp->nb_mem_chunks); in mlx4_rxq_attach()
569 mlx4_mr_update_mp(dev, &rxq->mr_ctrl, rxq->mp); in mlx4_rxq_attach()
623 *rxq->rq_db = rte_cpu_to_be_32(rxq->rq_ci); in mlx4_rxq_attach()
649 mlx4_rxq_detach(struct rxq *rxq) in mlx4_rxq_detach() argument
657 memset(&rxq->mcq, 0, sizeof(rxq->mcq)); in mlx4_rxq_detach()
740 struct rxq *rxq; in mlx4_rx_queue_setup() local
810 *rxq = (struct rxq){ in mlx4_rx_queue_setup()
[all …]
/f-stack/dpdk/drivers/net/bnx2x/
H A Dbnx2x_rxtx.c67 if (NULL == rxq) { in bnx2x_dev_rx_queue_setup()
71 rxq->sc = sc; in bnx2x_dev_rx_queue_setup()
72 rxq->mb_pool = mp; in bnx2x_dev_rx_queue_setup()
76 rxq->nb_rx_pages = 1; in bnx2x_dev_rx_queue_setup()
80 rxq->nb_rx_desc = TOTAL_RX_BD(rxq); in bnx2x_dev_rx_queue_setup()
82 rxq->nb_cq_pages = RCQ_BD_PAGES(rxq); in bnx2x_dev_rx_queue_setup()
87 (unsigned long)TOTAL_RX_BD(rxq), rxq->nb_rx_pages, in bnx2x_dev_rx_queue_setup()
130 rxq->rx_ring[idx] = in bnx2x_dev_rx_queue_setup()
136 rxq->rx_bd_tail = rxq->nb_rx_desc; in bnx2x_dev_rx_queue_setup()
156 rxq->rx_cq_tail = TOTAL_RCQ_ENTRIES(rxq); in bnx2x_dev_rx_queue_setup()
[all …]
/f-stack/dpdk/drivers/net/nfb/
H A Dnfb_rx.c122 if (rxq == NULL) { in nfb_eth_rx_queue_setup()
128 rxq->flags = 0; in nfb_eth_rx_queue_setup()
134 rxq); in nfb_eth_rx_queue_setup()
139 rte_free(rxq); in nfb_eth_rx_queue_setup()
161 if (rxq->queue == NULL) in nfb_eth_rx_queue_init()
164 rxq->nfb = nfb; in nfb_eth_rx_queue_init()
171 rxq->rx_pkts = 0; in nfb_eth_rx_queue_init()
172 rxq->rx_bytes = 0; in nfb_eth_rx_queue_init()
173 rxq->err_pkts = 0; in nfb_eth_rx_queue_init()
184 rte_free(rxq); in nfb_eth_rx_queue_release()
[all …]
/f-stack/dpdk/drivers/net/iavf/
H A Diavf_rxtx_vec_common.h72 rxq->pkt_first_seg = start; in reassemble_packets()
73 rxq->pkt_last_seg = end; in reassemble_packets()
151 if (!rxq->sw_ring || rxq->rxrearm_nb >= rxq->nb_rx_desc) in _iavf_rx_queue_release_mbufs_vec()
157 if (rxq->sw_ring[i]) in _iavf_rx_queue_release_mbufs_vec()
161 for (i = rxq->rx_tail; in _iavf_rx_queue_release_mbufs_vec()
164 if (rxq->sw_ring[i]) in _iavf_rx_queue_release_mbufs_vec()
169 rxq->rxrearm_nb = rxq->nb_rx_desc; in _iavf_rx_queue_release_mbufs_vec()
172 memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); in _iavf_rx_queue_release_mbufs_vec()
215 if (!rxq) in iavf_rx_vec_queue_default()
224 if (rxq->nb_rx_desc % rxq->rx_free_thresh) in iavf_rx_vec_queue_default()
[all …]
H A Diavf_rxtx.c186 rxq->sw_ring[rxq->nb_rx_desc + i] = &rxq->fake_mbuf; in reset_rx_queue()
605 rxq->port_id, rxq->queue_id); in iavf_dev_rx_queue_setup()
610 rxq->port_id, rxq->queue_id); in iavf_dev_rx_queue_setup()
823 rxq->ops->release_mbufs(rxq); in iavf_dev_rx_queue_stop()
922 rxq->ops->release_mbufs(rxq); in iavf_stop_queues()
1631 rxep = &rxq->sw_ring[rxq->rx_tail]; in iavf_rx_scan_hw_ring_flex_rxd()
1707 rxdp = &rxq->rx_ring[rxq->rx_tail]; in iavf_rx_scan_hw_ring()
1708 rxep = &rxq->sw_ring[rxq->rx_tail]; in iavf_rx_scan_hw_ring()
1887 if (rxq->rx_tail >= rxq->nb_rx_desc) in rx_recv_pkts()
1891 rxq->port_id, rxq->queue_id, in rx_recv_pkts()
[all …]
/f-stack/dpdk/drivers/net/axgbe/
H A Daxgbe_rxtx.c62 if (!rxq) { in axgbe_dev_rx_queue_setup()
67 rxq->cur = 0; in axgbe_dev_rx_queue_setup()
68 rxq->dirty = 0; in axgbe_dev_rx_queue_setup()
87 if (rxq->free_thresh > rxq->nb_desc) in axgbe_dev_rx_queue_setup()
88 rxq->free_thresh = rxq->nb_desc >> 3; in axgbe_dev_rx_queue_setup()
216 idx = AXGBE_GET_DESC_IDX(rxq, rxq->cur); in axgbe_recv_pkts()
306 idx = AXGBE_GET_DESC_IDX(rxq, rxq->dirty - 1); in axgbe_recv_pkts()
331 idx = AXGBE_GET_DESC_IDX(rxq, rxq->cur); in eth_axgbe_recv_scattered_pkts()
453 idx = AXGBE_GET_DESC_IDX(rxq, rxq->dirty - 1); in eth_axgbe_recv_scattered_pkts()
844 if (offset >= rxq->nb_desc - rxq->dirty) in axgbe_dev_rx_descriptor_status()
[all …]
/f-stack/dpdk/drivers/net/i40e/
H A Di40e_rxtx_vec_common.h71 rxq->pkt_first_seg = start; in reassemble_packets()
72 rxq->pkt_last_seg = end; in reassemble_packets()
150 if (rxq->sw_ring == NULL || rxq->rxrearm_nb >= rxq->nb_rx_desc) in _i40e_rx_queue_release_mbufs_vec()
160 for (i = rxq->rx_tail; in _i40e_rx_queue_release_mbufs_vec()
168 rxq->rxrearm_nb = rxq->nb_rx_desc; in _i40e_rx_queue_release_mbufs_vec()
171 memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); in _i40e_rx_queue_release_mbufs_vec()
200 struct i40e_rx_queue *rxq; in i40e_rx_vec_dev_conf_condition_check_default() local
224 if (!rxq) in i40e_rx_vec_dev_conf_condition_check_default()
226 desc = rxq->nb_rx_desc; in i40e_rx_vec_dev_conf_condition_check_default()
241 if (!rxq) in i40e_rx_vec_dev_conf_condition_check_default()
[all …]
H A Di40e_rxtx.c426 rxdp = &rxq->rx_ring[rxq->rx_tail]; in i40e_rx_scan_hw_ring()
427 rxep = &rxq->sw_ring[rxq->rx_tail]; in i40e_rx_scan_hw_ring()
607 if (rxq->rx_tail >= rxq->nb_rx_desc) in rx_recv_pkts()
1575 if (!rxq || !rxq->q_set) { in i40e_dev_rx_queue_start()
1616 if (!rxq || !rxq->q_set) { in i40e_dev_rx_queue_stop()
1802 rxq->nb_rx_desc, rxq->queue_id); in i40e_dev_rx_queue_setup_runtime()
1975 rxq->port_id, rxq->queue_id); in i40e_dev_rx_queue_setup()
1983 rxq->port_id, rxq->queue_id); in i40e_dev_rx_queue_setup()
2016 rxdp = &(rxq->rx_ring[rxq->rx_tail]); in i40e_dev_rx_queue_count()
2474 rxq->sw_ring[rxq->nb_rx_desc + i].mbuf = &rxq->fake_mbuf; in i40e_reset_rx_queue()
[all …]
/f-stack/dpdk/drivers/net/hinic/
H A Dhinic_pmd_rx.c21 ((rxq)->wq->mask)
24 (((rxq)->wq->cons_idx) & HINIC_GET_RQ_WQE_MASK(rxq))
27 (((rxq)->wq->prod_idx) & HINIC_GET_RQ_WQE_MASK(rxq))
39 #define HINIC_GET_RQ_FREE_WQEBBS(rxq) ((rxq)->wq->delta - 1) argument
206 buf_size = buf_size > rxq->buf_len ? rxq->buf_len : buf_size; in hinic_get_func_rx_buf_size()
293 memcpy(stats, &rxq->rxq_stats, sizeof(rxq->rxq_stats)); in hinic_rxq_get_stats()
300 if (rxq == NULL) in hinic_rxq_stats_reset()
319 rxq->rx_cqe = (struct hinic_rq_cqe *)rxq->cqe_start_vaddr; in hinic_rx_alloc_cqe()
330 rxq->cqe_start_vaddr, rxq->cqe_start_paddr); in hinic_rx_free_cqe()
368 rx_info_sz = rxq->q_depth * sizeof(*rxq->rx_info); in hinic_setup_rx_resources()
[all …]
/f-stack/dpdk/drivers/net/ice/
H A Dice_rxtx_vec_common.h69 rxq->pkt_first_seg = start; in ice_rx_reassemble_packets()
70 rxq->pkt_last_seg = end; in ice_rx_reassemble_packets()
153 if (rxq->rxrearm_nb >= rxq->nb_rx_desc) in _ice_rx_queue_release_mbufs_vec()
157 if (rxq->rxrearm_nb == 0) { in _ice_rx_queue_release_mbufs_vec()
159 if (rxq->sw_ring[i].mbuf) in _ice_rx_queue_release_mbufs_vec()
163 for (i = rxq->rx_tail; in _ice_rx_queue_release_mbufs_vec()
171 rxq->rxrearm_nb = rxq->nb_rx_desc; in _ice_rx_queue_release_mbufs_vec()
174 memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); in _ice_rx_queue_release_mbufs_vec()
248 if (!rxq) in ice_rx_vec_queue_default()
257 if (rxq->nb_rx_desc % rxq->rx_free_thresh) in ice_rx_vec_queue_default()
[all …]
H A Dice_rxtx.c397 if (!rxq || !rxq->sw_ring) { in _ice_rx_queue_release_mbufs()
516 rxq->sw_ring[rxq->nb_rx_desc + i].mbuf = &rxq->fake_mbuf; in ice_reset_rx_queue()
547 if (!rxq || !rxq->q_set) { in ice_rx_queue_start()
574 rxq->rx_rel_mbufs(rxq); in ice_rx_queue_start()
601 rxq->rx_rel_mbufs(rxq); in ice_rx_queue_stop()
766 rxq = pf->fdir.rxq; in ice_fdir_rx_queue_start()
767 if (!rxq || !rxq->q_set) { in ice_fdir_rx_queue_start()
969 rxq = pf->fdir.rxq; in ice_fdir_rx_queue_stop()
977 rxq->rx_rel_mbufs(rxq); in ice_fdir_rx_queue_stop()
2168 pf->fdir.rxq = rxq; in ice_fdir_setup_rx_resources()
[all …]
/f-stack/dpdk/drivers/net/atlantic/
H A Datl_rxtx.c133 rxq = rte_zmalloc_socket("atlantic Rx queue", sizeof(*rxq), in atl_rx_queue_setup()
161 rxq->port_id, rxq->queue_id); in atl_rx_queue_setup()
162 rte_free(rxq); in atl_rx_queue_setup()
178 rxq->port_id, rxq->queue_id); in atl_rx_queue_setup()
363 rxq->port_id, rxq->queue_id); in atl_rx_init()
369 rxq->port_id, rxq->queue_id); in atl_rx_init()
380 rxq->port_id, rxq->queue_id); in atl_rx_init()
409 rxq->port_id, rxq->queue_id); in atl_alloc_rx_queue_mbufs()
457 rxq->port_id, rxq->queue_id); in atl_rx_queue_start()
708 return rxq->nb_rx_desc - rxq->nb_rx_hold; in atl_rx_queue_count()
[all …]
/f-stack/dpdk/drivers/net/qede/
H A Dqede_rxtx.c15 uint16_t idx = rxq->sw_rx_prod & NUM_RX_BDS(rxq); in qede_alloc_rx_buffer()
22 idx, rxq->sw_rx_cons & NUM_RX_BDS(rxq), in qede_alloc_rx_buffer()
58 rxq->sw_rx_prod & NUM_RX_BDS(rxq), in qede_alloc_rx_bulk_mbufs()
59 rxq->sw_rx_cons & NUM_RX_BDS(rxq), in qede_alloc_rx_bulk_mbufs()
70 idx = rxq->sw_rx_prod & NUM_RX_BDS(rxq); in qede_alloc_rx_bulk_mbufs()
269 qdev->fp_array[qid * 2].rxq = rxq; in qede_rx_queue_setup()
275 qdev->fp_array[qid * 2 + 1].rxq = rxq; in qede_rx_queue_setup()
285 qdev->fp_array[qid].rxq = rxq; in qede_rx_queue_setup()
363 rxq = qdev->fp_array[rx_queue_id].rxq; in qede_rx_queue_stop()
765 rxq = fp->rxq; in qede_rx_queue_start()
[all …]
/f-stack/dpdk/drivers/net/e1000/
H A Dem_rxtx.c695 rxq = rx_queue; in eth_em_recv_pkts()
745 (unsigned) rxq->port_id, (unsigned) rxq->queue_id, in eth_em_recv_pkts()
838 (unsigned) rxq->port_id, (unsigned) rxq->queue_id, in eth_em_recv_pkts()
925 (unsigned) rxq->port_id, (unsigned) rxq->queue_id, in eth_em_recv_scattered_pkts()
1083 (unsigned) rxq->port_id, (unsigned) rxq->queue_id, in eth_em_recv_scattered_pkts()
1451 if ((rxq = rte_zmalloc("ethdev RX queue", sizeof(*rxq), in eth_em_rx_queue_setup()
1482 rxq->sw_ring, rxq->rx_ring, rxq->rx_ring_phys_addr); in eth_em_rx_queue_setup()
1500 rxdp = &(rxq->rx_ring[rxq->rx_tail]); in eth_em_rx_queue_count()
1506 if (rxq->rx_tail + desc >= rxq->nb_rx_desc) in eth_em_rx_queue_count()
1507 rxdp = &(rxq->rx_ring[rxq->rx_tail + in eth_em_rx_queue_count()
[all …]
/f-stack/dpdk/drivers/net/igc/
H A Digc_txrx.c407 rxq->port_id, rxq->queue_id, rx_id, staterr, in igc_recv_pkts()
415 rxq->port_id, rxq->queue_id); in igc_recv_pkts()
551 rxq->port_id, rxq->queue_id, rx_id, staterr, in igc_recv_scattered_pkts()
559 rxq->port_id, rxq->queue_id); in igc_recv_scattered_pkts()
720 if (rxq) in eth_igc_rx_queue_release()
739 rxdp = &rxq->rx_ring[rxq->rx_tail]; in eth_igc_rx_queue_count()
741 while (desc < rxq->nb_rx_desc - rxq->rx_tail) { in eth_igc_rx_queue_count()
748 rxdp = &rxq->rx_ring[rxq->rx_tail + desc - rxq->nb_rx_desc]; in eth_igc_rx_queue_count()
765 if (unlikely(!rxq || offset >= rxq->nb_rx_desc)) in eth_igc_rx_descriptor_done()
783 if (unlikely(!rxq || offset >= rxq->nb_rx_desc)) in eth_igc_rx_descriptor_status()
[all …]
/f-stack/dpdk/drivers/net/ixgbe/
H A Dixgbe_rxtx_vec_common.h37 start->pkt_len -= rxq->crc_len; in reassemble_packets()
39 end->data_len -= rxq->crc_len; in reassemble_packets()
69 rxq->pkt_first_seg = start; in reassemble_packets()
70 rxq->pkt_last_seg = end; in reassemble_packets()
174 if (rxq->sw_ring == NULL || rxq->rxrearm_nb >= rxq->nb_rx_desc) in _ixgbe_rx_queue_release_mbufs_vec()
178 if (rxq->rxrearm_nb == 0) { in _ixgbe_rx_queue_release_mbufs_vec()
184 for (i = rxq->rx_tail; in _ixgbe_rx_queue_release_mbufs_vec()
185 i != rxq->rxrearm_start; in _ixgbe_rx_queue_release_mbufs_vec()
192 rxq->rxrearm_nb = rxq->nb_rx_desc; in _ixgbe_rx_queue_release_mbufs_vec()
195 memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); in _ixgbe_rx_queue_release_mbufs_vec()
[all …]

12345678910>>...12