| /f-stack/dpdk/drivers/net/sfc/ |
| H A D | sfc_ef10_essb_rx.c | 149 rxd = &rxq->sw_ring[rxq->completed & rxq->rxq_ptr_mask]; in sfc_ef10_essb_maybe_next_completed() 154 rxd = &rxq->sw_ring[rxq->completed & rxq->rxq_ptr_mask]; in sfc_ef10_essb_maybe_next_completed() 155 rxq->left_in_completed = rxq->block_size; in sfc_ef10_essb_maybe_next_completed() 229 *rx_ev = rxq->evq_hw_ring[rxq->evq_read_ptr & rxq->evq_ptr_mask]; in sfc_ef10_essb_rx_event_get() 270 rxd = &rxq->sw_ring[rxq->pending_id]; in sfc_ef10_essb_rx_process_ev() 272 rxq->block_size - rxq->left_in_pending); in sfc_ef10_essb_rx_process_ev() 281 rxq->left_in_pending = rxq->block_size; in sfc_ef10_essb_rx_process_ev() 282 if (rxq->pending_id != rxq->rxq_ptr_mask) in sfc_ef10_essb_rx_process_ev() 476 if (offset < (rxq->added - rxq->completed) * rxq->block_size + in sfc_ef10_essb_rx_qdesc_status() 673 rxq->completed = rxq->pending_id = rxq->added = 0; in sfc_ef10_essb_rx_qstart() [all …]
|
| H A D | sfc_ef10_rx.c | 98 sfc_ef10_ev_qprime(rxq->evq_prime, rxq->evq_read_ptr, rxq->ptr_mask); in sfc_ef10_rx_qprime() 99 rxq->evq_read_ptr_primed = rxq->evq_read_ptr; in sfc_ef10_rx_qprime() 210 SFC_ASSERT(rxq->pending == rxq->completed || rxq->scatter_pkt == NULL); in sfc_ef10_rx_pending() 219 rxq->sw_ring[completed++ & rxq->ptr_mask].mbuf; in sfc_ef10_rx_pending() 412 *rx_ev = rxq->evq_hw_ring[rxq->evq_read_ptr & rxq->ptr_mask]; in sfc_ef10_rx_get_event() 466 rxq->evq_read_ptr_primed != rxq->evq_read_ptr) in sfc_ef10_recv_pkts() 576 if (offset < (rxq->added - rxq->completed)) in sfc_ef10_rx_qdesc_status() 655 rxq = rte_zmalloc_socket("sfc-ef10-rxq", sizeof(*rxq), in sfc_ef10_rx_qcreate() 779 for (i = rxq->completed; i != rxq->added; ++i) { in sfc_ef10_rx_qpurge() 780 rxd = &rxq->sw_ring[i & rxq->ptr_mask]; in sfc_ef10_rx_qpurge() [all …]
|
| H A D | sfc_ef100_rx.c | 99 rxq->evq_read_ptr & rxq->ptr_mask); in sfc_ef100_rx_qprime() 100 rxq->evq_read_ptr_primed = rxq->evq_read_ptr; in sfc_ef100_rx_qprime() 432 m = rxq->sw_ring[rxq->completed++ & rxq->ptr_mask].mbuf; in sfc_ef100_rx_next_mbuf() 435 if (unlikely(rxq->completed == rxq->added)) in sfc_ef100_rx_next_mbuf() 444 id = rxq->completed & rxq->ptr_mask; in sfc_ef100_rx_next_mbuf() 447 if (unlikely(rxq->completed + 1 == rxq->added)) in sfc_ef100_rx_next_mbuf() 533 *ev = rxq->evq_hw_ring[rxq->evq_read_ptr & rxq->ptr_mask]; in sfc_ef100_rx_get_event() 547 rxq->evq_read_ptr, rxq->evq_read_ptr & rxq->ptr_mask, in sfc_ef100_rx_get_event() 557 rxq->evq_read_ptr & rxq->ptr_mask); in sfc_ef100_rx_get_event() 856 rxd = &rxq->sw_ring[i & rxq->ptr_mask]; in sfc_ef100_rx_qpurge() [all …]
|
| H A D | sfc_rx.c | 60 if (rxq->evq->read_ptr_primed != rxq->evq->read_ptr) { in sfc_efx_rx_qprime() 63 rxq->evq->read_ptr_primed = rxq->evq->read_ptr; in sfc_efx_rx_qprime() 123 efx_rx_qpost(rxq->common, addr, rxq->buf_size, in sfc_efx_rx_qrefill() 130 efx_rx_qpush(rxq->common, added, &rxq->pushed); in sfc_efx_rx_qrefill() 339 return rxq->pending - rxq->completed; in sfc_efx_rx_qdesc_npending() 370 if (offset < (rxq->pending - rxq->completed)) in sfc_efx_rx_qdesc_status() 373 if (offset < (rxq->added - rxq->completed)) in sfc_efx_rx_qdesc_status() 491 rxq->evq = sfc_rxq_by_dp_rxq(&rxq->dp)->evq; in sfc_efx_rx_qcreate() 549 rxq->pending = rxq->completed = rxq->added = rxq->pushed = 0; in sfc_efx_rx_qstart() 590 for (i = rxq->completed; i != rxq->added; ++i) { in sfc_efx_rx_qpurge() [all …]
|
| /f-stack/dpdk/drivers/net/mlx5/ |
| H A D | mlx5_rxtx_vec.c | 91 uint16_t n = q_n - (rxq->rq_ci - rxq->rq_pi); in mlx5_rx_replenish_bulk_mbuf() 129 (*rxq->elts)[elts_idx + i] = &rxq->fake_mbuf; in mlx5_rx_replenish_bulk_mbuf() 131 *rxq->rq_db = rte_cpu_to_be_32(rxq->rq_ci); in mlx5_rx_replenish_bulk_mbuf() 148 uint32_t n = elts_n - (rxq->elts_ci - rxq->rq_pi); in mlx5_rx_mprq_replenish_bulk_mbuf() 154 rxq->elts_ci - rxq->rq_pi <= rxq->rq_repl_thresh) { in mlx5_rx_mprq_replenish_bulk_mbuf() 170 (*rxq->elts)[elts_idx + i] = &rxq->fake_mbuf; in mlx5_rx_mprq_replenish_bulk_mbuf() 250 *rxq->cq_db = rte_cpu_to_be_32(rxq->cq_ci); in rxq_copy_mprq_mbuf_v() 254 *rxq->rq_db = rte_cpu_to_be_32(rxq->rq_ci); in rxq_copy_mprq_mbuf_v() 296 MLX5_ASSERT(rxq->cqe_n == rxq->elts_n); in rxq_burst_v() 344 rxq->cq_ci += rxq->decompressed; in rxq_burst_v() [all …]
|
| H A D | mlx5_rxq.c | 109 (rxq, struct mlx5_rxq_ctrl, rxq); in mlx5_mprq_enabled() 156 struct mlx5_rxq_data *rxq = &rxq_ctrl->rxq; in rxq_alloc_elts_mprq() local 177 rxq->port_id, rxq->idx, wqe_n); in rxq_alloc_elts_mprq() 189 rxq->port_id, rxq->idx); in rxq_alloc_elts_mprq() 240 struct mlx5_rxq_data *rxq = &rxq_ctrl->rxq; in rxq_alloc_elts_sprq() local 317 struct mlx5_rxq_data *rxq = &rxq_ctrl->rxq; in rxq_free_elts_mprq() local 321 rxq->port_id, rxq->idx, (1u << rxq->elts_n)); in rxq_free_elts_mprq() 344 struct mlx5_rxq_data *rxq = &rxq_ctrl->rxq; in rxq_free_elts_sprq() local 363 rxq->rq_pi = rxq->rq_ci; in rxq_free_elts_sprq() 1249 (rxq, struct mlx5_rxq_ctrl, rxq); in mlx5_mprq_alloc_mp() [all …]
|
| /f-stack/dpdk/drivers/net/hns3/ |
| H A D | hns3_rxtx_vec.c | 53 struct hns3_entry *rxep = &rxq->sw_ring[rxq->rx_rearm_start]; in hns3_rxq_rearm_mbuf() 54 struct hns3_desc *rxdp = rxq->rx_ring + rxq->rx_rearm_start; in hns3_rxq_rearm_mbuf() 92 if (rxq->rx_rearm_start >= rxq->nb_rx_desc) in hns3_rxq_rearm_mbuf() 106 struct hns3_desc *rxdp = &rxq->rx_ring[rxq->next_to_use]; in hns3_recv_pkts_vec() 122 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 0].mbuf); in hns3_recv_pkts_vec() 123 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 1].mbuf); in hns3_recv_pkts_vec() 124 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 2].mbuf); in hns3_recv_pkts_vec() 125 rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 3].mbuf); in hns3_recv_pkts_vec() 158 memset(&rxq->rx_ring[rxq->nb_rx_desc], 0, in hns3_rxq_vec_setup() 161 memset(&rxq->fake_mbuf, 0, sizeof(rxq->fake_mbuf)); in hns3_rxq_vec_setup() [all …]
|
| /f-stack/dpdk/drivers/net/bnxt/ |
| H A D | bnxt_rxq.c | 25 if (rxq && rxq->cp_ring && rxq->cp_ring->hw_stats) in bnxt_free_rxq_stats() 55 rxq->vnic = vnic; in bnxt_mq_rx_configure() 210 if (!rxq || !rxq->rx_ring) in bnxt_rx_queue_release_mbufs() 266 if (rxq) { in bnxt_rx_queue_release_op() 293 rte_free(rxq); in bnxt_rx_queue_release_op() 328 if (rxq) in bnxt_rx_queue_setup_op() 333 if (!rxq) { in bnxt_rx_queue_setup_op() 337 rxq->bp = bp; in bnxt_rx_queue_setup_op() 414 if (!rxq) in bnxt_rx_queue_intr_enable_op() 437 if (!rxq) in bnxt_rx_queue_intr_disable_op() [all …]
|
| H A D | bnxt_rxtx_vec_common.h | 44 mb_def.port = rxq->port_id; in bnxt_rxq_vec_setup_common() 50 rxq->mbuf_initializer = *(uint64_t *)p; in bnxt_rxq_vec_setup_common() 51 rxq->rxrearm_nb = 0; in bnxt_rxq_vec_setup_common() 52 rxq->rxrearm_start = 0; in bnxt_rxq_vec_setup_common() 67 nb = RTE_MIN(rxq->rxrearm_nb & ~0x3, in bnxt_rxq_rearm() 68 rxq->nb_rx_desc - rxq->rxrearm_start); in bnxt_rxq_rearm() 75 rx_bufs[i] = &rxq->fake_mbuf; in bnxt_rxq_rearm() 90 rxq->rxrearm_start += nb; in bnxt_rxq_rearm() 92 if (rxq->rxrearm_start >= rxq->nb_rx_desc) in bnxt_rxq_rearm() 93 rxq->rxrearm_start = 0; in bnxt_rxq_rearm() [all …]
|
| /f-stack/dpdk/drivers/net/mlx4/ |
| H A D | mlx4_rxq.c | 190 struct rxq *rxq = NULL; in mlx4_rss_attach() local 438 if (rxq) in mlx4_rss_init() 487 mlx4_rxq_attach(struct rxq *rxq) in mlx4_rxq_attach() argument 568 rxq->mp->name, rxq->mp->nb_mem_chunks); in mlx4_rxq_attach() 569 mlx4_mr_update_mp(dev, &rxq->mr_ctrl, rxq->mp); in mlx4_rxq_attach() 623 *rxq->rq_db = rte_cpu_to_be_32(rxq->rq_ci); in mlx4_rxq_attach() 649 mlx4_rxq_detach(struct rxq *rxq) in mlx4_rxq_detach() argument 657 memset(&rxq->mcq, 0, sizeof(rxq->mcq)); in mlx4_rxq_detach() 740 struct rxq *rxq; in mlx4_rx_queue_setup() local 810 *rxq = (struct rxq){ in mlx4_rx_queue_setup() [all …]
|
| /f-stack/dpdk/drivers/net/bnx2x/ |
| H A D | bnx2x_rxtx.c | 67 if (NULL == rxq) { in bnx2x_dev_rx_queue_setup() 71 rxq->sc = sc; in bnx2x_dev_rx_queue_setup() 72 rxq->mb_pool = mp; in bnx2x_dev_rx_queue_setup() 76 rxq->nb_rx_pages = 1; in bnx2x_dev_rx_queue_setup() 80 rxq->nb_rx_desc = TOTAL_RX_BD(rxq); in bnx2x_dev_rx_queue_setup() 82 rxq->nb_cq_pages = RCQ_BD_PAGES(rxq); in bnx2x_dev_rx_queue_setup() 87 (unsigned long)TOTAL_RX_BD(rxq), rxq->nb_rx_pages, in bnx2x_dev_rx_queue_setup() 130 rxq->rx_ring[idx] = in bnx2x_dev_rx_queue_setup() 136 rxq->rx_bd_tail = rxq->nb_rx_desc; in bnx2x_dev_rx_queue_setup() 156 rxq->rx_cq_tail = TOTAL_RCQ_ENTRIES(rxq); in bnx2x_dev_rx_queue_setup() [all …]
|
| /f-stack/dpdk/drivers/net/nfb/ |
| H A D | nfb_rx.c | 122 if (rxq == NULL) { in nfb_eth_rx_queue_setup() 128 rxq->flags = 0; in nfb_eth_rx_queue_setup() 134 rxq); in nfb_eth_rx_queue_setup() 139 rte_free(rxq); in nfb_eth_rx_queue_setup() 161 if (rxq->queue == NULL) in nfb_eth_rx_queue_init() 164 rxq->nfb = nfb; in nfb_eth_rx_queue_init() 171 rxq->rx_pkts = 0; in nfb_eth_rx_queue_init() 172 rxq->rx_bytes = 0; in nfb_eth_rx_queue_init() 173 rxq->err_pkts = 0; in nfb_eth_rx_queue_init() 184 rte_free(rxq); in nfb_eth_rx_queue_release() [all …]
|
| /f-stack/dpdk/drivers/net/iavf/ |
| H A D | iavf_rxtx_vec_common.h | 72 rxq->pkt_first_seg = start; in reassemble_packets() 73 rxq->pkt_last_seg = end; in reassemble_packets() 151 if (!rxq->sw_ring || rxq->rxrearm_nb >= rxq->nb_rx_desc) in _iavf_rx_queue_release_mbufs_vec() 157 if (rxq->sw_ring[i]) in _iavf_rx_queue_release_mbufs_vec() 161 for (i = rxq->rx_tail; in _iavf_rx_queue_release_mbufs_vec() 164 if (rxq->sw_ring[i]) in _iavf_rx_queue_release_mbufs_vec() 169 rxq->rxrearm_nb = rxq->nb_rx_desc; in _iavf_rx_queue_release_mbufs_vec() 172 memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); in _iavf_rx_queue_release_mbufs_vec() 215 if (!rxq) in iavf_rx_vec_queue_default() 224 if (rxq->nb_rx_desc % rxq->rx_free_thresh) in iavf_rx_vec_queue_default() [all …]
|
| H A D | iavf_rxtx.c | 186 rxq->sw_ring[rxq->nb_rx_desc + i] = &rxq->fake_mbuf; in reset_rx_queue() 605 rxq->port_id, rxq->queue_id); in iavf_dev_rx_queue_setup() 610 rxq->port_id, rxq->queue_id); in iavf_dev_rx_queue_setup() 823 rxq->ops->release_mbufs(rxq); in iavf_dev_rx_queue_stop() 922 rxq->ops->release_mbufs(rxq); in iavf_stop_queues() 1631 rxep = &rxq->sw_ring[rxq->rx_tail]; in iavf_rx_scan_hw_ring_flex_rxd() 1707 rxdp = &rxq->rx_ring[rxq->rx_tail]; in iavf_rx_scan_hw_ring() 1708 rxep = &rxq->sw_ring[rxq->rx_tail]; in iavf_rx_scan_hw_ring() 1887 if (rxq->rx_tail >= rxq->nb_rx_desc) in rx_recv_pkts() 1891 rxq->port_id, rxq->queue_id, in rx_recv_pkts() [all …]
|
| /f-stack/dpdk/drivers/net/axgbe/ |
| H A D | axgbe_rxtx.c | 62 if (!rxq) { in axgbe_dev_rx_queue_setup() 67 rxq->cur = 0; in axgbe_dev_rx_queue_setup() 68 rxq->dirty = 0; in axgbe_dev_rx_queue_setup() 87 if (rxq->free_thresh > rxq->nb_desc) in axgbe_dev_rx_queue_setup() 88 rxq->free_thresh = rxq->nb_desc >> 3; in axgbe_dev_rx_queue_setup() 216 idx = AXGBE_GET_DESC_IDX(rxq, rxq->cur); in axgbe_recv_pkts() 306 idx = AXGBE_GET_DESC_IDX(rxq, rxq->dirty - 1); in axgbe_recv_pkts() 331 idx = AXGBE_GET_DESC_IDX(rxq, rxq->cur); in eth_axgbe_recv_scattered_pkts() 453 idx = AXGBE_GET_DESC_IDX(rxq, rxq->dirty - 1); in eth_axgbe_recv_scattered_pkts() 844 if (offset >= rxq->nb_desc - rxq->dirty) in axgbe_dev_rx_descriptor_status() [all …]
|
| /f-stack/dpdk/drivers/net/i40e/ |
| H A D | i40e_rxtx_vec_common.h | 71 rxq->pkt_first_seg = start; in reassemble_packets() 72 rxq->pkt_last_seg = end; in reassemble_packets() 150 if (rxq->sw_ring == NULL || rxq->rxrearm_nb >= rxq->nb_rx_desc) in _i40e_rx_queue_release_mbufs_vec() 160 for (i = rxq->rx_tail; in _i40e_rx_queue_release_mbufs_vec() 168 rxq->rxrearm_nb = rxq->nb_rx_desc; in _i40e_rx_queue_release_mbufs_vec() 171 memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); in _i40e_rx_queue_release_mbufs_vec() 200 struct i40e_rx_queue *rxq; in i40e_rx_vec_dev_conf_condition_check_default() local 224 if (!rxq) in i40e_rx_vec_dev_conf_condition_check_default() 226 desc = rxq->nb_rx_desc; in i40e_rx_vec_dev_conf_condition_check_default() 241 if (!rxq) in i40e_rx_vec_dev_conf_condition_check_default() [all …]
|
| H A D | i40e_rxtx.c | 426 rxdp = &rxq->rx_ring[rxq->rx_tail]; in i40e_rx_scan_hw_ring() 427 rxep = &rxq->sw_ring[rxq->rx_tail]; in i40e_rx_scan_hw_ring() 607 if (rxq->rx_tail >= rxq->nb_rx_desc) in rx_recv_pkts() 1575 if (!rxq || !rxq->q_set) { in i40e_dev_rx_queue_start() 1616 if (!rxq || !rxq->q_set) { in i40e_dev_rx_queue_stop() 1802 rxq->nb_rx_desc, rxq->queue_id); in i40e_dev_rx_queue_setup_runtime() 1975 rxq->port_id, rxq->queue_id); in i40e_dev_rx_queue_setup() 1983 rxq->port_id, rxq->queue_id); in i40e_dev_rx_queue_setup() 2016 rxdp = &(rxq->rx_ring[rxq->rx_tail]); in i40e_dev_rx_queue_count() 2474 rxq->sw_ring[rxq->nb_rx_desc + i].mbuf = &rxq->fake_mbuf; in i40e_reset_rx_queue() [all …]
|
| /f-stack/dpdk/drivers/net/hinic/ |
| H A D | hinic_pmd_rx.c | 21 ((rxq)->wq->mask) 24 (((rxq)->wq->cons_idx) & HINIC_GET_RQ_WQE_MASK(rxq)) 27 (((rxq)->wq->prod_idx) & HINIC_GET_RQ_WQE_MASK(rxq)) 39 #define HINIC_GET_RQ_FREE_WQEBBS(rxq) ((rxq)->wq->delta - 1) argument 206 buf_size = buf_size > rxq->buf_len ? rxq->buf_len : buf_size; in hinic_get_func_rx_buf_size() 293 memcpy(stats, &rxq->rxq_stats, sizeof(rxq->rxq_stats)); in hinic_rxq_get_stats() 300 if (rxq == NULL) in hinic_rxq_stats_reset() 319 rxq->rx_cqe = (struct hinic_rq_cqe *)rxq->cqe_start_vaddr; in hinic_rx_alloc_cqe() 330 rxq->cqe_start_vaddr, rxq->cqe_start_paddr); in hinic_rx_free_cqe() 368 rx_info_sz = rxq->q_depth * sizeof(*rxq->rx_info); in hinic_setup_rx_resources() [all …]
|
| /f-stack/dpdk/drivers/net/ice/ |
| H A D | ice_rxtx_vec_common.h | 69 rxq->pkt_first_seg = start; in ice_rx_reassemble_packets() 70 rxq->pkt_last_seg = end; in ice_rx_reassemble_packets() 153 if (rxq->rxrearm_nb >= rxq->nb_rx_desc) in _ice_rx_queue_release_mbufs_vec() 157 if (rxq->rxrearm_nb == 0) { in _ice_rx_queue_release_mbufs_vec() 159 if (rxq->sw_ring[i].mbuf) in _ice_rx_queue_release_mbufs_vec() 163 for (i = rxq->rx_tail; in _ice_rx_queue_release_mbufs_vec() 171 rxq->rxrearm_nb = rxq->nb_rx_desc; in _ice_rx_queue_release_mbufs_vec() 174 memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); in _ice_rx_queue_release_mbufs_vec() 248 if (!rxq) in ice_rx_vec_queue_default() 257 if (rxq->nb_rx_desc % rxq->rx_free_thresh) in ice_rx_vec_queue_default() [all …]
|
| H A D | ice_rxtx.c | 397 if (!rxq || !rxq->sw_ring) { in _ice_rx_queue_release_mbufs() 516 rxq->sw_ring[rxq->nb_rx_desc + i].mbuf = &rxq->fake_mbuf; in ice_reset_rx_queue() 547 if (!rxq || !rxq->q_set) { in ice_rx_queue_start() 574 rxq->rx_rel_mbufs(rxq); in ice_rx_queue_start() 601 rxq->rx_rel_mbufs(rxq); in ice_rx_queue_stop() 766 rxq = pf->fdir.rxq; in ice_fdir_rx_queue_start() 767 if (!rxq || !rxq->q_set) { in ice_fdir_rx_queue_start() 969 rxq = pf->fdir.rxq; in ice_fdir_rx_queue_stop() 977 rxq->rx_rel_mbufs(rxq); in ice_fdir_rx_queue_stop() 2168 pf->fdir.rxq = rxq; in ice_fdir_setup_rx_resources() [all …]
|
| /f-stack/dpdk/drivers/net/atlantic/ |
| H A D | atl_rxtx.c | 133 rxq = rte_zmalloc_socket("atlantic Rx queue", sizeof(*rxq), in atl_rx_queue_setup() 161 rxq->port_id, rxq->queue_id); in atl_rx_queue_setup() 162 rte_free(rxq); in atl_rx_queue_setup() 178 rxq->port_id, rxq->queue_id); in atl_rx_queue_setup() 363 rxq->port_id, rxq->queue_id); in atl_rx_init() 369 rxq->port_id, rxq->queue_id); in atl_rx_init() 380 rxq->port_id, rxq->queue_id); in atl_rx_init() 409 rxq->port_id, rxq->queue_id); in atl_alloc_rx_queue_mbufs() 457 rxq->port_id, rxq->queue_id); in atl_rx_queue_start() 708 return rxq->nb_rx_desc - rxq->nb_rx_hold; in atl_rx_queue_count() [all …]
|
| /f-stack/dpdk/drivers/net/qede/ |
| H A D | qede_rxtx.c | 15 uint16_t idx = rxq->sw_rx_prod & NUM_RX_BDS(rxq); in qede_alloc_rx_buffer() 22 idx, rxq->sw_rx_cons & NUM_RX_BDS(rxq), in qede_alloc_rx_buffer() 58 rxq->sw_rx_prod & NUM_RX_BDS(rxq), in qede_alloc_rx_bulk_mbufs() 59 rxq->sw_rx_cons & NUM_RX_BDS(rxq), in qede_alloc_rx_bulk_mbufs() 70 idx = rxq->sw_rx_prod & NUM_RX_BDS(rxq); in qede_alloc_rx_bulk_mbufs() 269 qdev->fp_array[qid * 2].rxq = rxq; in qede_rx_queue_setup() 275 qdev->fp_array[qid * 2 + 1].rxq = rxq; in qede_rx_queue_setup() 285 qdev->fp_array[qid].rxq = rxq; in qede_rx_queue_setup() 363 rxq = qdev->fp_array[rx_queue_id].rxq; in qede_rx_queue_stop() 765 rxq = fp->rxq; in qede_rx_queue_start() [all …]
|
| /f-stack/dpdk/drivers/net/e1000/ |
| H A D | em_rxtx.c | 695 rxq = rx_queue; in eth_em_recv_pkts() 745 (unsigned) rxq->port_id, (unsigned) rxq->queue_id, in eth_em_recv_pkts() 838 (unsigned) rxq->port_id, (unsigned) rxq->queue_id, in eth_em_recv_pkts() 925 (unsigned) rxq->port_id, (unsigned) rxq->queue_id, in eth_em_recv_scattered_pkts() 1083 (unsigned) rxq->port_id, (unsigned) rxq->queue_id, in eth_em_recv_scattered_pkts() 1451 if ((rxq = rte_zmalloc("ethdev RX queue", sizeof(*rxq), in eth_em_rx_queue_setup() 1482 rxq->sw_ring, rxq->rx_ring, rxq->rx_ring_phys_addr); in eth_em_rx_queue_setup() 1500 rxdp = &(rxq->rx_ring[rxq->rx_tail]); in eth_em_rx_queue_count() 1506 if (rxq->rx_tail + desc >= rxq->nb_rx_desc) in eth_em_rx_queue_count() 1507 rxdp = &(rxq->rx_ring[rxq->rx_tail + in eth_em_rx_queue_count() [all …]
|
| /f-stack/dpdk/drivers/net/igc/ |
| H A D | igc_txrx.c | 407 rxq->port_id, rxq->queue_id, rx_id, staterr, in igc_recv_pkts() 415 rxq->port_id, rxq->queue_id); in igc_recv_pkts() 551 rxq->port_id, rxq->queue_id, rx_id, staterr, in igc_recv_scattered_pkts() 559 rxq->port_id, rxq->queue_id); in igc_recv_scattered_pkts() 720 if (rxq) in eth_igc_rx_queue_release() 739 rxdp = &rxq->rx_ring[rxq->rx_tail]; in eth_igc_rx_queue_count() 741 while (desc < rxq->nb_rx_desc - rxq->rx_tail) { in eth_igc_rx_queue_count() 748 rxdp = &rxq->rx_ring[rxq->rx_tail + desc - rxq->nb_rx_desc]; in eth_igc_rx_queue_count() 765 if (unlikely(!rxq || offset >= rxq->nb_rx_desc)) in eth_igc_rx_descriptor_done() 783 if (unlikely(!rxq || offset >= rxq->nb_rx_desc)) in eth_igc_rx_descriptor_status() [all …]
|
| /f-stack/dpdk/drivers/net/ixgbe/ |
| H A D | ixgbe_rxtx_vec_common.h | 37 start->pkt_len -= rxq->crc_len; in reassemble_packets() 39 end->data_len -= rxq->crc_len; in reassemble_packets() 69 rxq->pkt_first_seg = start; in reassemble_packets() 70 rxq->pkt_last_seg = end; in reassemble_packets() 174 if (rxq->sw_ring == NULL || rxq->rxrearm_nb >= rxq->nb_rx_desc) in _ixgbe_rx_queue_release_mbufs_vec() 178 if (rxq->rxrearm_nb == 0) { in _ixgbe_rx_queue_release_mbufs_vec() 184 for (i = rxq->rx_tail; in _ixgbe_rx_queue_release_mbufs_vec() 185 i != rxq->rxrearm_start; in _ixgbe_rx_queue_release_mbufs_vec() 192 rxq->rxrearm_nb = rxq->nb_rx_desc; in _ixgbe_rx_queue_release_mbufs_vec() 195 memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); in _ixgbe_rx_queue_release_mbufs_vec() [all …]
|