Home
last modified time | relevance | path

Searched refs:nb_desc (Results 1 – 25 of 100) sorted by relevance

1234

/f-stack/dpdk/drivers/net/fm10k/
H A Dfm10k_rxtx.c181 q->nb_desc; in fm10k_recv_pkts()
339 q->nb_desc; in fm10k_recv_scattered_pkts()
390 rxq->nb_desc]; in fm10k_dev_rx_queue_count()
410 if (desc >= rxq->nb_desc) in fm10k_dev_rx_descriptor_done()
411 desc -= rxq->nb_desc; in fm10k_dev_rx_descriptor_done()
450 if (desc >= rxq->nb_desc) in fm10k_dev_rx_descriptor_status()
451 desc -= rxq->nb_desc; in fm10k_dev_rx_descriptor_status()
480 desc -= txq->nb_desc; in fm10k_dev_tx_descriptor_status()
481 if (desc >= txq->nb_desc) in fm10k_dev_tx_descriptor_status()
482 desc -= txq->nb_desc; in fm10k_dev_tx_descriptor_status()
[all …]
H A Dfm10k.h62 #define FM10K_RX_FREE_THRESH_MAX(rxq) ((rxq)->nb_desc - 1)
63 #define FM10K_RX_FREE_THRESH_DIV(rxq) ((rxq)->nb_desc)
67 #define FM10K_TX_FREE_THRESH_MAX(txq) ((txq)->nb_desc - 3)
81 RTE_MIN(((txq)->nb_desc - 2), (txq)->free_thresh)
82 #define FM10K_TX_RS_THRESH_DIV(txq) ((txq)->nb_desc)
168 uint16_t nb_desc; member
213 uint16_t nb_desc; member
H A Dfm10k_rxtx_vec.c340 if (rxq->rxrearm_start >= rxq->nb_desc) in fm10k_rxq_rearm()
346 (rxq->nb_desc - 1) : (rxq->rxrearm_start - 1)); in fm10k_rxq_rearm()
355 const unsigned mask = rxq->nb_desc - 1; in fm10k_rx_queue_release_mbufs_vec()
358 if (rxq->sw_ring == NULL || rxq->rxrearm_nb >= rxq->nb_desc) in fm10k_rx_queue_release_mbufs_vec()
363 for (i = 0; i < rxq->nb_desc; i++) in fm10k_rx_queue_release_mbufs_vec()
371 rxq->rxrearm_nb = rxq->nb_desc; in fm10k_rx_queue_release_mbufs_vec()
810 if (txq->next_dd >= txq->nb_desc) in fm10k_tx_free_bufs()
854 n = (uint16_t)(txq->nb_desc - tx_id); in fm10k_xmit_fixed_burst_vec()
898 for (i = 0; i < txq->nb_desc; i++) in fm10k_reset_tx_queue()
902 for (i = 0; i < txq->nb_desc; i++) in fm10k_reset_tx_queue()
[all …]
H A Dfm10k_ethdev.c207 for (i = 0; i < q->nb_desc; ++i) { in rx_queue_reset()
211 q->nb_desc); in rx_queue_reset()
225 q->hw_ring[q->nb_desc + i] = zero; in rx_queue_reset()
250 for (i = 0; i < q->nb_desc; ++i) in rx_queue_clean()
255 q->hw_ring[q->nb_desc + i] = zero; in rx_queue_clean()
264 for (i = 0; i < q->nb_desc; ++i) { in rx_queue_clean()
328 q->nb_free = q->nb_desc - 1; in tx_queue_reset()
345 for (i = 0; i < q->nb_desc; ++i) in tx_queue_clean()
349 for (i = 0; i < q->nb_desc; ++i) { in tx_queue_clean()
1841 q->nb_desc = nb_desc; in fm10k_rx_queue_setup()
[all …]
/f-stack/dpdk/drivers/net/octeontx/
H A Docteontx_rxtx.h317 uint16_t gaura_id, nb_desc = 0; in __octeontx_xmit_prepare() local
321 cmd_buf[nb_desc++] = 0x0; in __octeontx_xmit_prepare()
349 return nb_desc; in __octeontx_xmit_prepare()
356 uint16_t nb_segs, nb_desc = 0; in __octeontx_xmit_mseg_prepare() local
363 cmd_buf[nb_desc++] = 0x0; in __octeontx_xmit_mseg_prepare()
388 cmd_buf[nb_desc] |= in __octeontx_xmit_mseg_prepare()
395 if (!(cmd_buf[nb_desc] & (1ULL << 57))) { in __octeontx_xmit_mseg_prepare()
400 nb_desc++; in __octeontx_xmit_mseg_prepare()
409 return nb_desc; in __octeontx_xmit_mseg_prepare()
419 uint16_t count = 0, nb_desc; in __octeontx_xmit_pkts() local
[all …]
/f-stack/dpdk/drivers/net/axgbe/
H A Daxgbe_rxtx.c24 for (i = 0; i < rx_queue->nb_desc; i++) { in axgbe_rx_queue_release()
48 uint32_t rx_desc = nb_desc; in axgbe_dev_rx_queue_setup()
73 rxq->nb_desc = rx_desc; in axgbe_dev_rx_queue_setup()
87 if (rxq->free_thresh > rxq->nb_desc) in axgbe_dev_rx_queue_setup()
88 rxq->free_thresh = rxq->nb_desc >> 3; in axgbe_dev_rx_queue_setup()
218 if (unlikely(idx == rxq->nb_desc)) in axgbe_recv_pkts()
237 if (unlikely(pidx == rxq->nb_desc)) in axgbe_recv_pkts()
335 if (unlikely(idx == rxq->nb_desc)) in eth_axgbe_recv_scattered_pkts()
496 tx_desc = nb_desc; in axgbe_dev_tx_queue_setup()
515 txq->nb_desc = tx_desc; in axgbe_dev_tx_queue_setup()
[all …]
H A Daxgbe_rxtx.h39 ((_idx) & ((_queue)->nb_desc - 1)))
42 ((_idx) & ((_queue)->nb_desc - 1)) \
71 uint16_t nb_desc; member
111 uint16_t nb_desc; member
/f-stack/dpdk/drivers/net/bnxt/
H A Dbnxt_txq.c89 uint16_t nb_desc, in bnxt_tx_queue_setup_op() argument
108 if (nb_desc < BNXT_MIN_RING_DESC || nb_desc > MAX_TX_DESC_CNT) { in bnxt_tx_queue_setup_op()
109 PMD_DRV_LOG(ERR, "nb_desc %d is invalid", nb_desc); in bnxt_tx_queue_setup_op()
128 sizeof(struct rte_mbuf *) * nb_desc, in bnxt_tx_queue_setup_op()
136 txq->nb_tx_desc = nb_desc; in bnxt_tx_queue_setup_op()
138 RTE_MIN(rte_align32pow2(nb_desc) / 4, RTE_BNXT_MAX_TX_BURST); in bnxt_tx_queue_setup_op()
H A Dbnxt_reps.c590 uint16_t nb_desc, in bnxt_rep_rx_queue_setup_op() argument
609 if (!nb_desc || nb_desc > MAX_RX_DESC_CNT) { in bnxt_rep_rx_queue_setup_op()
610 PMD_DRV_LOG(ERR, "nb_desc %d is invalid\n", nb_desc); in bnxt_rep_rx_queue_setup_op()
625 if (nb_desc != parent_rxq->nb_rx_desc) { in bnxt_rep_rx_queue_setup_op()
626 PMD_DRV_LOG(ERR, "nb_desc %d do not match parent rxq", nb_desc); in bnxt_rep_rx_queue_setup_op()
644 rxq->nb_rx_desc = nb_desc; in bnxt_rep_rx_queue_setup_op()
692 uint16_t nb_desc, in bnxt_rep_tx_queue_setup_op() argument
708 if (!nb_desc || nb_desc > MAX_TX_DESC_CNT) { in bnxt_rep_tx_queue_setup_op()
709 PMD_DRV_LOG(ERR, "nb_desc %d is invalid", nb_desc); in bnxt_rep_tx_queue_setup_op()
724 if (nb_desc != parent_txq->nb_tx_desc) { in bnxt_rep_tx_queue_setup_op()
[all …]
H A Dbnxt_reps.h34 __rte_unused uint16_t nb_desc,
41 __rte_unused uint16_t nb_desc,
/f-stack/dpdk/drivers/raw/octeontx2_ep/
H A Dotx2_ep_enqdeq.c108 iq->nb_desc = conf->num_iqdef_descs; in sdp_init_instr_queue()
115 (iq->nb_desc * SDP_IQREQ_LIST_SIZE), in sdp_init_instr_queue()
125 iq->nb_desc); in sdp_init_instr_queue()
202 for (idx = 0; idx < droq->nb_desc; idx++) { in sdp_droq_destroy_ring_buffers()
259 for (idx = 0; idx < droq->nb_desc; idx++) { in sdp_droq_setup_ring_buffers()
288 (droq->nb_desc * SDP_DROQ_INFO_SIZE), in sdp_alloc_info_buffer()
319 droq->nb_desc = conf->num_oqdef_descs; in sdp_init_droq()
352 (droq->nb_desc * SDP_DROQ_RECVBUF_SIZE), in sdp_init_droq()
493 (int32_t)(iq->nb_desc - 1)) { in post_iqcmd()
681 droq->nb_desc); in sdp_droq_refill()
[all …]
/f-stack/dpdk/drivers/regex/mlx5/
H A Dmlx5_regex_control.c38 regex_ctrl_get_nb_obj(uint16_t nb_desc) in regex_ctrl_get_nb_obj() argument
40 return ((nb_desc / MLX5_REGEX_NUM_WQE_PER_PAGE) + in regex_ctrl_get_nb_obj()
41 !!(nb_desc % MLX5_REGEX_NUM_WQE_PER_PAGE)); in regex_ctrl_get_nb_obj()
345 qp->cq.log_nb_desc = rte_log2_u32(cfg->nb_desc); in mlx5_regex_qp_setup()
346 qp->nb_desc = 1 << qp->cq.log_nb_desc; in mlx5_regex_qp_setup()
348 qp->nb_obj = regex_ctrl_get_nb_obj(qp->nb_desc); in mlx5_regex_qp_setup()
358 log_desc = rte_log2_u32(qp->nb_desc / qp->nb_obj); in mlx5_regex_qp_setup()
H A Dmlx5_regex_fastpath.c330 void *ptr = rte_calloc(__func__, qp->nb_desc, in setup_buffers()
337 MLX5_REGEX_METADATA_SIZE*qp->nb_desc, in setup_buffers()
345 ptr = rte_calloc(__func__, qp->nb_desc, in setup_buffers()
353 MLX5_REGEX_MAX_OUTPUT * qp->nb_desc, in setup_buffers()
363 for (i = 0; i < qp->nb_desc; i++) { in setup_buffers()
366 (i % qp->nb_desc) * MLX5_REGEX_MAX_OUTPUT; in setup_buffers()
369 (i % qp->nb_desc) * MLX5_REGEX_METADATA_SIZE; in setup_buffers()
386 qp->jobs = rte_calloc(__func__, qp->nb_desc, sizeof(*qp->jobs), 64); in mlx5_regexdev_setup_fastpath()
/f-stack/dpdk/drivers/net/ark/
H A Dark_ethdev_tx.c212 uint16_t nb_desc, in eth_ark_tx_queue_setup() argument
222 if (!rte_is_power_of_2(nb_desc)) { in eth_ark_tx_queue_setup()
226 nb_desc, __func__); in eth_ark_tx_queue_setup()
243 queue->queue_size = nb_desc; in eth_ark_tx_queue_setup()
244 queue->queue_mask = nb_desc - 1; in eth_ark_tx_queue_setup()
251 nb_desc * sizeof(struct ark_tx_meta), in eth_ark_tx_queue_setup()
256 nb_desc * sizeof(struct rte_mbuf *), in eth_ark_tx_queue_setup()
H A Dark_ethdev_rx.c119 uint16_t nb_desc, in eth_ark_dev_rx_queue_setup() argument
152 if (!rte_is_power_of_2(nb_desc)) { in eth_ark_dev_rx_queue_setup()
155 nb_desc, __func__); in eth_ark_dev_rx_queue_setup()
173 queue->queue_size = nb_desc; in eth_ark_dev_rx_queue_setup()
174 queue->queue_mask = nb_desc - 1; in eth_ark_dev_rx_queue_setup()
178 nb_desc * sizeof(struct rte_mbuf *), in eth_ark_dev_rx_queue_setup()
183 nb_desc * sizeof(rte_iova_t), in eth_ark_dev_rx_queue_setup()
204 if (queue->seed_index != nb_desc) { in eth_ark_dev_rx_queue_setup()
206 nb_desc, qidx); in eth_ark_dev_rx_queue_setup()
/f-stack/dpdk/drivers/net/mvneta/
H A Dmvneta_rxtx.c80 uint16_t nb_desc = *num; in mvneta_buffs_refill() local
91 for (i = 0; i < nb_desc; i++) { in mvneta_buffs_refill()
106 for (i = *num; i < nb_desc; i++) in mvneta_buffs_refill()
122 uint16_t nb_desc, nb_desc_burst, sent = 0; in mvneta_buffs_alloc() local
125 nb_desc = *num; in mvneta_buffs_alloc()
129 (nb_desc < MRVL_NETA_BUF_RELEASE_BURST_SIZE_MAX) ? in mvneta_buffs_alloc()
130 nb_desc : MRVL_NETA_BUF_RELEASE_BURST_SIZE_MAX; in mvneta_buffs_alloc()
137 nb_desc -= nb_desc_burst; in mvneta_buffs_alloc()
139 } while (nb_desc); in mvneta_buffs_alloc()
991 qinfo->nb_desc = q->size; in mvneta_rxq_info_get()
[all …]
/f-stack/dpdk/drivers/net/liquidio/
H A Dlio_rxtx.c138 (droq->nb_desc * in lio_alloc_info_buffer()
202 droq->nb_desc); in lio_init_droq()
211 (droq->nb_desc * in lio_init_droq()
311 droq->nb_desc); in lio_droq_refill_pullup_descs()
371 droq->nb_desc); in lio_droq_refill()
418 droq->nb_desc); in lio_droq_fast_process_packet()
431 droq->nb_desc); in lio_droq_fast_process_packet()
525 1, droq->nb_desc); in lio_droq_fast_process_packet()
725 iq->nb_desc); in lio_init_instr_queue()
925 iq->nb_desc); in post_command2()
[all …]
/f-stack/dpdk/drivers/net/ionic/
H A Dionic_rxtx.c65 qinfo->nb_desc = q->num_descs; in ionic_txq_info_get()
159 uint16_t nb_desc, uint32_t socket_id __rte_unused, in ionic_dev_tx_queue_setup() argument
170 tx_queue_id, nb_desc); in ionic_dev_tx_queue_setup()
182 if (!rte_is_power_of_2(nb_desc) || nb_desc < IONIC_MIN_RING_DESC) in ionic_dev_tx_queue_setup()
607 qinfo->nb_desc = q->num_descs; in ionic_rxq_info_get()
643 uint16_t nb_desc, in ionic_dev_rx_queue_setup() argument
656 rx_queue_id, nb_desc); in ionic_dev_rx_queue_setup()
668 if (!rte_is_power_of_2(nb_desc) || in ionic_dev_rx_queue_setup()
669 nb_desc < IONIC_MIN_RING_DESC || in ionic_dev_rx_queue_setup()
670 nb_desc > IONIC_MAX_RING_DESC) { in ionic_dev_rx_queue_setup()
[all …]
H A Dionic_rxtx.h26 uint16_t nb_desc, uint32_t socket_id,
33 uint16_t nb_desc, uint32_t socket_id,
/f-stack/dpdk/drivers/net/e1000/
H A Dem_rxtx.c1133 uint16_t i, nb_desc, prev; in em_reset_tx_queue() local
1138 nb_desc = txq->nb_tx_desc; in em_reset_tx_queue()
1142 prev = (uint16_t) (nb_desc - 1); in em_reset_tx_queue()
1144 for (i = 0; i < nb_desc; i++) { in em_reset_tx_queue()
1198 uint16_t nb_desc, in eth_em_tx_queue_setup() argument
1218 if (nb_desc % EM_TXD_ALIGN != 0 || in eth_em_tx_queue_setup()
1297 txq->nb_tx_desc = nb_desc; in eth_em_tx_queue_setup()
1401 uint16_t nb_desc, in eth_em_rx_queue_setup() argument
1464 rxq->nb_rx_desc = nb_desc; in eth_em_rx_queue_setup()
2005 qinfo->nb_desc = rxq->nb_rx_desc; in em_rxq_info_get()
[all …]
/f-stack/dpdk/drivers/net/netvsc/
H A Dhn_var.h176 uint16_t nb_desc, unsigned int socket_id,
188 uint16_t queue_idx, uint16_t nb_desc,
241 uint16_t queue_idx, uint16_t nb_desc,
248 uint16_t queue_idx, uint16_t nb_desc,
H A Dhn_rxtx.c239 uint16_t queue_idx, uint16_t nb_desc, in hn_dev_tx_queue_setup() argument
254 tx_free_thresh = RTE_MIN(nb_desc / 4, in hn_dev_tx_queue_setup()
257 if (tx_free_thresh + 3 >= nb_desc) { in hn_dev_tx_queue_setup()
261 nb_desc - 3, in hn_dev_tx_queue_setup()
281 name, nb_desc, sizeof(struct hn_txdesc)); in hn_dev_tx_queue_setup()
331 qinfo->nb_desc = txq->txdesc_pool->size; in hn_dev_tx_queue_info()
923 qinfo->nb_desc = rxq->rx_ring->size; in hn_dev_rx_queue_info()
929 uint16_t queue_idx, uint16_t nb_desc, in hn_dev_rx_queue_setup() argument
952 if (nb_desc == 0 || nb_desc > count) in hn_dev_rx_queue_setup()
953 nb_desc = count; in hn_dev_rx_queue_setup()
[all …]
/f-stack/dpdk/drivers/net/cxgbe/
H A Dcxgbe_pfvf.h32 uint16_t nb_desc, unsigned int socket_id,
35 uint16_t nb_desc, unsigned int socket_id,
/f-stack/dpdk/drivers/net/hns3/
H A Dhns3_rxtx.c1285 q_info.nb_desc = nb_desc; in hns3_fake_rx_queue_setup()
1379 q_info.nb_desc = nb_desc; in hns3_fake_tx_queue_setup()
1621 uint16_t nb_desc) in hns3_rxq_conf_runtime_check() argument
1643 if (nb_desc < min_vec_bds || in hns3_rxq_conf_runtime_check()
1663 if (nb_desc > HNS3_MAX_RING_DESC || nb_desc < HNS3_MIN_RING_DESC || in hns3_rx_queue_conf_check()
1666 nb_desc); in hns3_rx_queue_conf_check()
1733 q_info.nb_desc = nb_desc; in hns3_rx_queue_setup()
2502 if (nb_desc > HNS3_MAX_RING_DESC || nb_desc < HNS3_MIN_RING_DESC || in hns3_tx_queue_conf_check()
2505 nb_desc); in hns3_tx_queue_conf_check()
2513 if (rs_thresh + free_thresh > nb_desc || nb_desc % rs_thresh || in hns3_tx_queue_conf_check()
[all …]
/f-stack/dpdk/drivers/net/iavf/
H A Diavf_rxtx.c66 if (thresh >= nb_desc) { in check_rx_thresh()
68 thresh, nb_desc); in check_rx_thresh()
94 if (tx_rs_thresh >= (nb_desc - 2)) { in check_tx_thresh()
97 tx_rs_thresh, nb_desc); in check_tx_thresh()
103 tx_free_thresh, nb_desc); in check_tx_thresh()
115 tx_rs_thresh, nb_desc); in check_tx_thresh()
508 "invalid", nb_desc); in iavf_dev_rx_queue_setup()
549 rxq->nb_rx_desc = nb_desc; in iavf_dev_rx_queue_setup()
623 uint16_t nb_desc, in iavf_dev_tx_queue_setup() argument
642 "invalid", nb_desc); in iavf_dev_tx_queue_setup()
[all …]

1234