| /f-stack/dpdk/drivers/net/fm10k/ |
| H A D | fm10k_rxtx.c | 181 q->nb_desc; in fm10k_recv_pkts() 339 q->nb_desc; in fm10k_recv_scattered_pkts() 390 rxq->nb_desc]; in fm10k_dev_rx_queue_count() 410 if (desc >= rxq->nb_desc) in fm10k_dev_rx_descriptor_done() 411 desc -= rxq->nb_desc; in fm10k_dev_rx_descriptor_done() 450 if (desc >= rxq->nb_desc) in fm10k_dev_rx_descriptor_status() 451 desc -= rxq->nb_desc; in fm10k_dev_rx_descriptor_status() 480 desc -= txq->nb_desc; in fm10k_dev_tx_descriptor_status() 481 if (desc >= txq->nb_desc) in fm10k_dev_tx_descriptor_status() 482 desc -= txq->nb_desc; in fm10k_dev_tx_descriptor_status() [all …]
|
| H A D | fm10k.h | 62 #define FM10K_RX_FREE_THRESH_MAX(rxq) ((rxq)->nb_desc - 1) 63 #define FM10K_RX_FREE_THRESH_DIV(rxq) ((rxq)->nb_desc) 67 #define FM10K_TX_FREE_THRESH_MAX(txq) ((txq)->nb_desc - 3) 81 RTE_MIN(((txq)->nb_desc - 2), (txq)->free_thresh) 82 #define FM10K_TX_RS_THRESH_DIV(txq) ((txq)->nb_desc) 168 uint16_t nb_desc; member 213 uint16_t nb_desc; member
|
| H A D | fm10k_rxtx_vec.c | 340 if (rxq->rxrearm_start >= rxq->nb_desc) in fm10k_rxq_rearm() 346 (rxq->nb_desc - 1) : (rxq->rxrearm_start - 1)); in fm10k_rxq_rearm() 355 const unsigned mask = rxq->nb_desc - 1; in fm10k_rx_queue_release_mbufs_vec() 358 if (rxq->sw_ring == NULL || rxq->rxrearm_nb >= rxq->nb_desc) in fm10k_rx_queue_release_mbufs_vec() 363 for (i = 0; i < rxq->nb_desc; i++) in fm10k_rx_queue_release_mbufs_vec() 371 rxq->rxrearm_nb = rxq->nb_desc; in fm10k_rx_queue_release_mbufs_vec() 810 if (txq->next_dd >= txq->nb_desc) in fm10k_tx_free_bufs() 854 n = (uint16_t)(txq->nb_desc - tx_id); in fm10k_xmit_fixed_burst_vec() 898 for (i = 0; i < txq->nb_desc; i++) in fm10k_reset_tx_queue() 902 for (i = 0; i < txq->nb_desc; i++) in fm10k_reset_tx_queue() [all …]
|
| H A D | fm10k_ethdev.c | 207 for (i = 0; i < q->nb_desc; ++i) { in rx_queue_reset() 211 q->nb_desc); in rx_queue_reset() 225 q->hw_ring[q->nb_desc + i] = zero; in rx_queue_reset() 250 for (i = 0; i < q->nb_desc; ++i) in rx_queue_clean() 255 q->hw_ring[q->nb_desc + i] = zero; in rx_queue_clean() 264 for (i = 0; i < q->nb_desc; ++i) { in rx_queue_clean() 328 q->nb_free = q->nb_desc - 1; in tx_queue_reset() 345 for (i = 0; i < q->nb_desc; ++i) in tx_queue_clean() 349 for (i = 0; i < q->nb_desc; ++i) { in tx_queue_clean() 1841 q->nb_desc = nb_desc; in fm10k_rx_queue_setup() [all …]
|
| /f-stack/dpdk/drivers/net/octeontx/ |
| H A D | octeontx_rxtx.h | 317 uint16_t gaura_id, nb_desc = 0; in __octeontx_xmit_prepare() local 321 cmd_buf[nb_desc++] = 0x0; in __octeontx_xmit_prepare() 349 return nb_desc; in __octeontx_xmit_prepare() 356 uint16_t nb_segs, nb_desc = 0; in __octeontx_xmit_mseg_prepare() local 363 cmd_buf[nb_desc++] = 0x0; in __octeontx_xmit_mseg_prepare() 388 cmd_buf[nb_desc] |= in __octeontx_xmit_mseg_prepare() 395 if (!(cmd_buf[nb_desc] & (1ULL << 57))) { in __octeontx_xmit_mseg_prepare() 400 nb_desc++; in __octeontx_xmit_mseg_prepare() 409 return nb_desc; in __octeontx_xmit_mseg_prepare() 419 uint16_t count = 0, nb_desc; in __octeontx_xmit_pkts() local [all …]
|
| /f-stack/dpdk/drivers/net/axgbe/ |
| H A D | axgbe_rxtx.c | 24 for (i = 0; i < rx_queue->nb_desc; i++) { in axgbe_rx_queue_release() 48 uint32_t rx_desc = nb_desc; in axgbe_dev_rx_queue_setup() 73 rxq->nb_desc = rx_desc; in axgbe_dev_rx_queue_setup() 87 if (rxq->free_thresh > rxq->nb_desc) in axgbe_dev_rx_queue_setup() 88 rxq->free_thresh = rxq->nb_desc >> 3; in axgbe_dev_rx_queue_setup() 218 if (unlikely(idx == rxq->nb_desc)) in axgbe_recv_pkts() 237 if (unlikely(pidx == rxq->nb_desc)) in axgbe_recv_pkts() 335 if (unlikely(idx == rxq->nb_desc)) in eth_axgbe_recv_scattered_pkts() 496 tx_desc = nb_desc; in axgbe_dev_tx_queue_setup() 515 txq->nb_desc = tx_desc; in axgbe_dev_tx_queue_setup() [all …]
|
| H A D | axgbe_rxtx.h | 39 ((_idx) & ((_queue)->nb_desc - 1))) 42 ((_idx) & ((_queue)->nb_desc - 1)) \ 71 uint16_t nb_desc; member 111 uint16_t nb_desc; member
|
| /f-stack/dpdk/drivers/net/bnxt/ |
| H A D | bnxt_txq.c | 89 uint16_t nb_desc, in bnxt_tx_queue_setup_op() argument 108 if (nb_desc < BNXT_MIN_RING_DESC || nb_desc > MAX_TX_DESC_CNT) { in bnxt_tx_queue_setup_op() 109 PMD_DRV_LOG(ERR, "nb_desc %d is invalid", nb_desc); in bnxt_tx_queue_setup_op() 128 sizeof(struct rte_mbuf *) * nb_desc, in bnxt_tx_queue_setup_op() 136 txq->nb_tx_desc = nb_desc; in bnxt_tx_queue_setup_op() 138 RTE_MIN(rte_align32pow2(nb_desc) / 4, RTE_BNXT_MAX_TX_BURST); in bnxt_tx_queue_setup_op()
|
| H A D | bnxt_reps.c | 590 uint16_t nb_desc, in bnxt_rep_rx_queue_setup_op() argument 609 if (!nb_desc || nb_desc > MAX_RX_DESC_CNT) { in bnxt_rep_rx_queue_setup_op() 610 PMD_DRV_LOG(ERR, "nb_desc %d is invalid\n", nb_desc); in bnxt_rep_rx_queue_setup_op() 625 if (nb_desc != parent_rxq->nb_rx_desc) { in bnxt_rep_rx_queue_setup_op() 626 PMD_DRV_LOG(ERR, "nb_desc %d do not match parent rxq", nb_desc); in bnxt_rep_rx_queue_setup_op() 644 rxq->nb_rx_desc = nb_desc; in bnxt_rep_rx_queue_setup_op() 692 uint16_t nb_desc, in bnxt_rep_tx_queue_setup_op() argument 708 if (!nb_desc || nb_desc > MAX_TX_DESC_CNT) { in bnxt_rep_tx_queue_setup_op() 709 PMD_DRV_LOG(ERR, "nb_desc %d is invalid", nb_desc); in bnxt_rep_tx_queue_setup_op() 724 if (nb_desc != parent_txq->nb_tx_desc) { in bnxt_rep_tx_queue_setup_op() [all …]
|
| H A D | bnxt_reps.h | 34 __rte_unused uint16_t nb_desc, 41 __rte_unused uint16_t nb_desc,
|
| /f-stack/dpdk/drivers/raw/octeontx2_ep/ |
| H A D | otx2_ep_enqdeq.c | 108 iq->nb_desc = conf->num_iqdef_descs; in sdp_init_instr_queue() 115 (iq->nb_desc * SDP_IQREQ_LIST_SIZE), in sdp_init_instr_queue() 125 iq->nb_desc); in sdp_init_instr_queue() 202 for (idx = 0; idx < droq->nb_desc; idx++) { in sdp_droq_destroy_ring_buffers() 259 for (idx = 0; idx < droq->nb_desc; idx++) { in sdp_droq_setup_ring_buffers() 288 (droq->nb_desc * SDP_DROQ_INFO_SIZE), in sdp_alloc_info_buffer() 319 droq->nb_desc = conf->num_oqdef_descs; in sdp_init_droq() 352 (droq->nb_desc * SDP_DROQ_RECVBUF_SIZE), in sdp_init_droq() 493 (int32_t)(iq->nb_desc - 1)) { in post_iqcmd() 681 droq->nb_desc); in sdp_droq_refill() [all …]
|
| /f-stack/dpdk/drivers/regex/mlx5/ |
| H A D | mlx5_regex_control.c | 38 regex_ctrl_get_nb_obj(uint16_t nb_desc) in regex_ctrl_get_nb_obj() argument 40 return ((nb_desc / MLX5_REGEX_NUM_WQE_PER_PAGE) + in regex_ctrl_get_nb_obj() 41 !!(nb_desc % MLX5_REGEX_NUM_WQE_PER_PAGE)); in regex_ctrl_get_nb_obj() 345 qp->cq.log_nb_desc = rte_log2_u32(cfg->nb_desc); in mlx5_regex_qp_setup() 346 qp->nb_desc = 1 << qp->cq.log_nb_desc; in mlx5_regex_qp_setup() 348 qp->nb_obj = regex_ctrl_get_nb_obj(qp->nb_desc); in mlx5_regex_qp_setup() 358 log_desc = rte_log2_u32(qp->nb_desc / qp->nb_obj); in mlx5_regex_qp_setup()
|
| H A D | mlx5_regex_fastpath.c | 330 void *ptr = rte_calloc(__func__, qp->nb_desc, in setup_buffers() 337 MLX5_REGEX_METADATA_SIZE*qp->nb_desc, in setup_buffers() 345 ptr = rte_calloc(__func__, qp->nb_desc, in setup_buffers() 353 MLX5_REGEX_MAX_OUTPUT * qp->nb_desc, in setup_buffers() 363 for (i = 0; i < qp->nb_desc; i++) { in setup_buffers() 366 (i % qp->nb_desc) * MLX5_REGEX_MAX_OUTPUT; in setup_buffers() 369 (i % qp->nb_desc) * MLX5_REGEX_METADATA_SIZE; in setup_buffers() 386 qp->jobs = rte_calloc(__func__, qp->nb_desc, sizeof(*qp->jobs), 64); in mlx5_regexdev_setup_fastpath()
|
| /f-stack/dpdk/drivers/net/ark/ |
| H A D | ark_ethdev_tx.c | 212 uint16_t nb_desc, in eth_ark_tx_queue_setup() argument 222 if (!rte_is_power_of_2(nb_desc)) { in eth_ark_tx_queue_setup() 226 nb_desc, __func__); in eth_ark_tx_queue_setup() 243 queue->queue_size = nb_desc; in eth_ark_tx_queue_setup() 244 queue->queue_mask = nb_desc - 1; in eth_ark_tx_queue_setup() 251 nb_desc * sizeof(struct ark_tx_meta), in eth_ark_tx_queue_setup() 256 nb_desc * sizeof(struct rte_mbuf *), in eth_ark_tx_queue_setup()
|
| H A D | ark_ethdev_rx.c | 119 uint16_t nb_desc, in eth_ark_dev_rx_queue_setup() argument 152 if (!rte_is_power_of_2(nb_desc)) { in eth_ark_dev_rx_queue_setup() 155 nb_desc, __func__); in eth_ark_dev_rx_queue_setup() 173 queue->queue_size = nb_desc; in eth_ark_dev_rx_queue_setup() 174 queue->queue_mask = nb_desc - 1; in eth_ark_dev_rx_queue_setup() 178 nb_desc * sizeof(struct rte_mbuf *), in eth_ark_dev_rx_queue_setup() 183 nb_desc * sizeof(rte_iova_t), in eth_ark_dev_rx_queue_setup() 204 if (queue->seed_index != nb_desc) { in eth_ark_dev_rx_queue_setup() 206 nb_desc, qidx); in eth_ark_dev_rx_queue_setup()
|
| /f-stack/dpdk/drivers/net/mvneta/ |
| H A D | mvneta_rxtx.c | 80 uint16_t nb_desc = *num; in mvneta_buffs_refill() local 91 for (i = 0; i < nb_desc; i++) { in mvneta_buffs_refill() 106 for (i = *num; i < nb_desc; i++) in mvneta_buffs_refill() 122 uint16_t nb_desc, nb_desc_burst, sent = 0; in mvneta_buffs_alloc() local 125 nb_desc = *num; in mvneta_buffs_alloc() 129 (nb_desc < MRVL_NETA_BUF_RELEASE_BURST_SIZE_MAX) ? in mvneta_buffs_alloc() 130 nb_desc : MRVL_NETA_BUF_RELEASE_BURST_SIZE_MAX; in mvneta_buffs_alloc() 137 nb_desc -= nb_desc_burst; in mvneta_buffs_alloc() 139 } while (nb_desc); in mvneta_buffs_alloc() 991 qinfo->nb_desc = q->size; in mvneta_rxq_info_get() [all …]
|
| /f-stack/dpdk/drivers/net/liquidio/ |
| H A D | lio_rxtx.c | 138 (droq->nb_desc * in lio_alloc_info_buffer() 202 droq->nb_desc); in lio_init_droq() 211 (droq->nb_desc * in lio_init_droq() 311 droq->nb_desc); in lio_droq_refill_pullup_descs() 371 droq->nb_desc); in lio_droq_refill() 418 droq->nb_desc); in lio_droq_fast_process_packet() 431 droq->nb_desc); in lio_droq_fast_process_packet() 525 1, droq->nb_desc); in lio_droq_fast_process_packet() 725 iq->nb_desc); in lio_init_instr_queue() 925 iq->nb_desc); in post_command2() [all …]
|
| /f-stack/dpdk/drivers/net/ionic/ |
| H A D | ionic_rxtx.c | 65 qinfo->nb_desc = q->num_descs; in ionic_txq_info_get() 159 uint16_t nb_desc, uint32_t socket_id __rte_unused, in ionic_dev_tx_queue_setup() argument 170 tx_queue_id, nb_desc); in ionic_dev_tx_queue_setup() 182 if (!rte_is_power_of_2(nb_desc) || nb_desc < IONIC_MIN_RING_DESC) in ionic_dev_tx_queue_setup() 607 qinfo->nb_desc = q->num_descs; in ionic_rxq_info_get() 643 uint16_t nb_desc, in ionic_dev_rx_queue_setup() argument 656 rx_queue_id, nb_desc); in ionic_dev_rx_queue_setup() 668 if (!rte_is_power_of_2(nb_desc) || in ionic_dev_rx_queue_setup() 669 nb_desc < IONIC_MIN_RING_DESC || in ionic_dev_rx_queue_setup() 670 nb_desc > IONIC_MAX_RING_DESC) { in ionic_dev_rx_queue_setup() [all …]
|
| H A D | ionic_rxtx.h | 26 uint16_t nb_desc, uint32_t socket_id, 33 uint16_t nb_desc, uint32_t socket_id,
|
| /f-stack/dpdk/drivers/net/e1000/ |
| H A D | em_rxtx.c | 1133 uint16_t i, nb_desc, prev; in em_reset_tx_queue() local 1138 nb_desc = txq->nb_tx_desc; in em_reset_tx_queue() 1142 prev = (uint16_t) (nb_desc - 1); in em_reset_tx_queue() 1144 for (i = 0; i < nb_desc; i++) { in em_reset_tx_queue() 1198 uint16_t nb_desc, in eth_em_tx_queue_setup() argument 1218 if (nb_desc % EM_TXD_ALIGN != 0 || in eth_em_tx_queue_setup() 1297 txq->nb_tx_desc = nb_desc; in eth_em_tx_queue_setup() 1401 uint16_t nb_desc, in eth_em_rx_queue_setup() argument 1464 rxq->nb_rx_desc = nb_desc; in eth_em_rx_queue_setup() 2005 qinfo->nb_desc = rxq->nb_rx_desc; in em_rxq_info_get() [all …]
|
| /f-stack/dpdk/drivers/net/netvsc/ |
| H A D | hn_var.h | 176 uint16_t nb_desc, unsigned int socket_id, 188 uint16_t queue_idx, uint16_t nb_desc, 241 uint16_t queue_idx, uint16_t nb_desc, 248 uint16_t queue_idx, uint16_t nb_desc,
|
| H A D | hn_rxtx.c | 239 uint16_t queue_idx, uint16_t nb_desc, in hn_dev_tx_queue_setup() argument 254 tx_free_thresh = RTE_MIN(nb_desc / 4, in hn_dev_tx_queue_setup() 257 if (tx_free_thresh + 3 >= nb_desc) { in hn_dev_tx_queue_setup() 261 nb_desc - 3, in hn_dev_tx_queue_setup() 281 name, nb_desc, sizeof(struct hn_txdesc)); in hn_dev_tx_queue_setup() 331 qinfo->nb_desc = txq->txdesc_pool->size; in hn_dev_tx_queue_info() 923 qinfo->nb_desc = rxq->rx_ring->size; in hn_dev_rx_queue_info() 929 uint16_t queue_idx, uint16_t nb_desc, in hn_dev_rx_queue_setup() argument 952 if (nb_desc == 0 || nb_desc > count) in hn_dev_rx_queue_setup() 953 nb_desc = count; in hn_dev_rx_queue_setup() [all …]
|
| /f-stack/dpdk/drivers/net/cxgbe/ |
| H A D | cxgbe_pfvf.h | 32 uint16_t nb_desc, unsigned int socket_id, 35 uint16_t nb_desc, unsigned int socket_id,
|
| /f-stack/dpdk/drivers/net/hns3/ |
| H A D | hns3_rxtx.c | 1285 q_info.nb_desc = nb_desc; in hns3_fake_rx_queue_setup() 1379 q_info.nb_desc = nb_desc; in hns3_fake_tx_queue_setup() 1621 uint16_t nb_desc) in hns3_rxq_conf_runtime_check() argument 1643 if (nb_desc < min_vec_bds || in hns3_rxq_conf_runtime_check() 1663 if (nb_desc > HNS3_MAX_RING_DESC || nb_desc < HNS3_MIN_RING_DESC || in hns3_rx_queue_conf_check() 1666 nb_desc); in hns3_rx_queue_conf_check() 1733 q_info.nb_desc = nb_desc; in hns3_rx_queue_setup() 2502 if (nb_desc > HNS3_MAX_RING_DESC || nb_desc < HNS3_MIN_RING_DESC || in hns3_tx_queue_conf_check() 2505 nb_desc); in hns3_tx_queue_conf_check() 2513 if (rs_thresh + free_thresh > nb_desc || nb_desc % rs_thresh || in hns3_tx_queue_conf_check() [all …]
|
| /f-stack/dpdk/drivers/net/iavf/ |
| H A D | iavf_rxtx.c | 66 if (thresh >= nb_desc) { in check_rx_thresh() 68 thresh, nb_desc); in check_rx_thresh() 94 if (tx_rs_thresh >= (nb_desc - 2)) { in check_tx_thresh() 97 tx_rs_thresh, nb_desc); in check_tx_thresh() 103 tx_free_thresh, nb_desc); in check_tx_thresh() 115 tx_rs_thresh, nb_desc); in check_tx_thresh() 508 "invalid", nb_desc); in iavf_dev_rx_queue_setup() 549 rxq->nb_rx_desc = nb_desc; in iavf_dev_rx_queue_setup() 623 uint16_t nb_desc, in iavf_dev_tx_queue_setup() argument 642 "invalid", nb_desc); in iavf_dev_tx_queue_setup() [all …]
|