| /dpdk/drivers/crypto/bcmfs/ |
| H A D | bcmfs_qp.c | 53 memset(queue->base_addr, 0x9B, queue->queue_size); in bcmfs_queue_delete() 65 queue_dma_zone_reserve(const char *queue_name, uint32_t queue_size, in queue_dma_zone_reserve() argument 72 if (((size_t)queue_size <= mz->len) && in queue_dma_zone_reserve() 84 mz->socket_id, queue_size, socket_id); in queue_dma_zone_reserve() 89 queue_name, queue_size, socket_id); in queue_dma_zone_reserve() 90 return rte_memzone_reserve_aligned(queue_name, queue_size, in queue_dma_zone_reserve() 153 queue->queue_size = queue_size_bytes; in bcmfs_queue_create()
|
| /dpdk/drivers/net/ark/ |
| H A D | ark_ethdev_rx.c | 47 uint32_t queue_size; member 90 ark_mpu_configure(queue->mpu, phys_addr_q_base, queue->queue_size, 0); in eth_ark_rx_hw_setup() 108 if ((cons_index + queue->queue_size - queue->seed_index) >= ARK_RX_MPU_CHUNK) { in eth_ark_rx_update_cons_index() 175 queue->queue_size = nb_desc; in eth_ark_dev_rx_queue_setup() 451 queue->queue_size; in eth_ark_rx_seed_mbufs() 460 if (unlikely(seed_m + nb > queue->queue_size)) in eth_ark_rx_seed_mbufs() 461 nb = queue->queue_size - seed_m; in eth_ark_rx_seed_mbufs() 552 for (i = 0; i < queue->queue_size; ++i) in eth_ark_dev_rx_queue_release() 633 "queue_size", queue->queue_size, in ark_ethdev_rx_dump()
|
| H A D | ark_ethdev_tx.c | 38 uint32_t queue_size; member 130 prod_index_limit = queue->queue_size + queue->free_index - 4; in eth_ark_xmit_pkts() 264 queue->queue_size = nb_desc; in eth_ark_tx_queue_setup() 327 ark_mpu_configure(queue->mpu, ring_base, queue->queue_size, 1); in eth_ark_tx_hw_queue_config() 334 switch (queue->queue_size) { in eth_ark_tx_hw_queue_config()
|
| /dpdk/drivers/common/qat/ |
| H A D | qat_qp.c | 46 uint32_t queue_size, int socket_id); 117 qp->max_inflights = ADF_MAX_INFLIGHTS(qp->tx_q.queue_size, in qat_qp_setup() 236 &(queue->queue_size)) != 0) { in qat_queue_create() 258 queue->queue_size, queue_size_bytes, in qat_queue_create() 270 queue_dma_zone_reserve(const char *queue_name, uint32_t queue_size, in queue_dma_zone_reserve() argument 277 if (((size_t)queue_size <= mz->len) && in queue_dma_zone_reserve() 289 mz->socket_id, queue_size, socket_id); in queue_dma_zone_reserve() 294 queue_name, queue_size, socket_id); in queue_dma_zone_reserve() 295 return rte_memzone_reserve_aligned(queue_name, queue_size, in queue_dma_zone_reserve() 296 socket_id, RTE_MEMZONE_IOVA_CONTIG, queue_size); in queue_dma_zone_reserve() [all …]
|
| /dpdk/examples/ip_pipeline/ |
| H A D | link.c | 116 (params->rx.queue_size == 0) || in link_create() 118 (params->tx.queue_size == 0)) in link_create() 190 params->rx.queue_size, in link_create() 204 params->tx.queue_size, in link_create()
|
| H A D | link.h | 47 uint32_t queue_size; member 54 uint32_t queue_size; member
|
| H A D | cryptodev.c | 65 (params->queue_size == 0) || in cryptodev_create() 101 queue_conf.nb_descriptors = params->queue_size; in cryptodev_create()
|
| H A D | cryptodev.h | 39 uint32_t queue_size; member
|
| /dpdk/drivers/baseband/la12xx/ |
| H A D | bbdev_la12xx.c | 82 .queue_size = MAX_CHANNEL_DEPTH, 146 channel_id, q_priv->queue_size, msg_size); in ipc_queue_configure() 153 for (i = 0; i < q_priv->queue_size; i++) { in ipc_queue_configure() 216 for (i = 0; i < q_priv->queue_size; i++) { in la12xx_e200_queue_setup() 267 ch->depth = rte_cpu_to_be_32(q_priv->queue_size); in la12xx_e200_queue_setup() 299 q_priv->queue_size = queue_conf->queue_size; in la12xx_queue_setup() 442 pi, ci, pi_flag, ci_flag, q_priv->queue_size); in enqueue_single_op() 507 if (unlikely(q_priv->queue_size == pi)) { in enqueue_single_op() 587 ci, ci_flag, q_priv->queue_size); in dequeue_single_op() 597 if (q_priv->queue_size == ci) { in dequeue_single_op() [all …]
|
| H A D | bbdev_la12xx.h | 32 uint16_t queue_size; /**< Queue depth */ member
|
| /dpdk/drivers/net/virtio/virtio_user/ |
| H A D | virtio_user_dev.h | 37 uint32_t queue_size; member 68 int cq, int queue_size, const char *mac, char **ifname,
|
| H A D | virtio_user_dev.c | 564 int cq, int queue_size, const char *mac, char **ifname, in virtio_user_dev_init() argument 582 dev->queue_size = queue_size; in virtio_user_dev_init() 776 if (idx_data >= dev->queue_size) in virtio_user_handle_ctrl_msg_packed() 777 idx_data -= dev->queue_size; in virtio_user_handle_ctrl_msg_packed() 784 if (idx_status >= dev->queue_size) in virtio_user_handle_ctrl_msg_packed() 785 idx_status -= dev->queue_size; in virtio_user_handle_ctrl_msg_packed() 839 if (vq->used_idx >= dev->queue_size) { in virtio_user_handle_cq_packed() 840 vq->used_idx -= dev->queue_size; in virtio_user_handle_cq_packed()
|
| /dpdk/drivers/common/qat/dev/ |
| H A D | qat_dev_gen1.c | 112 queue->queue_size); in qat_qp_csr_build_ring_base_gen1() 155 q_tx_config = BUILD_RING_CONFIG(q_tx->queue_size); in qat_qp_adf_configure_queues_gen1() 156 q_resp_config = BUILD_RESP_RING_CONFIG(q_rx->queue_size, in qat_qp_adf_configure_queues_gen1()
|
| H A D | qat_dev_gen4.c | 153 queue->queue_size); in qat_qp_build_ring_base_gen4() 198 q_tx_config = BUILD_RING_CONFIG(q_tx->queue_size); in qat_qp_adf_configure_queues_gen4() 199 q_resp_config = BUILD_RESP_RING_CONFIG(q_rx->queue_size, in qat_qp_adf_configure_queues_gen4()
|
| /dpdk/examples/pipeline/ |
| H A D | obj.h | 71 uint32_t queue_size; member 78 uint32_t queue_size; member
|
| H A D | obj.c | 198 (params->rx.queue_size == 0) || in link_create() 200 (params->tx.queue_size == 0)) in link_create() 272 params->rx.queue_size, in link_create() 286 params->tx.queue_size, in link_create()
|
| /dpdk/app/test-bbdev/ |
| H A D | test_bbdev.c | 275 ts_params->qconf.queue_size = info.drv.queue_size_lim; in test_bbdev_configure_stop_queue() 323 TEST_ASSERT(qinfo.conf.queue_size == ts_params->qconf.queue_size, in test_bbdev_configure_stop_queue() 325 "invalid queue_size:%u", qinfo.conf.queue_size); in test_bbdev_configure_stop_queue() 408 ts_params->qconf.queue_size = info.drv.queue_size_lim + 1; in test_bbdev_configure_invalid_queue_configure() 413 ts_params->qconf.queue_size); in test_bbdev_configure_invalid_queue_configure() 415 ts_params->qconf.queue_size = info.drv.queue_size_lim; in test_bbdev_configure_invalid_queue_configure() 439 ts_params->qconf.queue_size); in test_bbdev_configure_invalid_queue_configure() 449 ts_params->qconf.queue_size); in test_bbdev_configure_invalid_queue_configure()
|
| /dpdk/drivers/crypto/ccp/ |
| H A D | ccp_dev.c | 88 uint32_t queue_size, in ccp_queue_dma_zone_reserve() argument 95 if (((size_t)queue_size <= mz->len) && in ccp_queue_dma_zone_reserve() 106 mz->socket_id, queue_size, socket_id); in ccp_queue_dma_zone_reserve() 111 queue_name, queue_size, socket_id); in ccp_queue_dma_zone_reserve() 113 return rte_memzone_reserve_aligned(queue_name, queue_size, in ccp_queue_dma_zone_reserve() 114 socket_id, RTE_MEMZONE_IOVA_CONTIG, queue_size); in ccp_queue_dma_zone_reserve()
|
| /dpdk/doc/guides/howto/ |
| H A D | virtio_user_as_exceptional_path.rst | 59 --vdev=virtio_user0,path=/dev/vhost-net,queue_size=1024 \ 72 * ``queue_size`` 83 --vdev=virtio_user0,path=/dev/vhost-net,queues=2,queue_size=1024 \
|
| /dpdk/drivers/net/softnic/ |
| H A D | rte_eth_softnic_cryptodev.c | 74 (params->queue_size == 0) || in softnic_cryptodev_create() 111 queue_conf.nb_descriptors = params->queue_size; in softnic_cryptodev_create()
|
| /dpdk/drivers/dma/dpaa/ |
| H A D | dpaa_qdma.c | 281 unsigned int queue_size[FSL_QDMA_QUEUE_MAX]; in fsl_qdma_alloc_queue_resources() local 292 queue_size[i] = QDMA_QUEUE_SIZE; in fsl_qdma_alloc_queue_resources() 296 if (queue_size[i] > FSL_QDMA_CIRCULAR_DESC_SIZE_MAX || in fsl_qdma_alloc_queue_resources() 297 queue_size[i] < FSL_QDMA_CIRCULAR_DESC_SIZE_MIN) { in fsl_qdma_alloc_queue_resources() 305 queue_size[i], in fsl_qdma_alloc_queue_resources() 307 queue_size[i], &queue_temp->bus_addr); in fsl_qdma_alloc_queue_resources() 312 memset(queue_temp->cq, 0x0, queue_size[i] * in fsl_qdma_alloc_queue_resources() 317 queue_temp->n_cq = queue_size[i]; in fsl_qdma_alloc_queue_resources()
|
| /dpdk/drivers/raw/ntb/ |
| H A D | rte_pmd_ntb.h | 32 uint16_t queue_size; member
|
| H A D | ntb.c | 103 ret = (*hw->ntb_ops->spad_write)(dev, SPAD_Q_SZ, 1, hw->queue_size); in ntb_handshake_work() 509 uint16_t q_size = hw->queue_size; in ntb_queue_init() 844 if (!hw->queue_size || !hw->queue_pairs) { in ntb_dev_info_get() 850 hw->queue_size * sizeof(struct ntb_desc) + in ntb_dev_info_get() 851 hw->queue_size * sizeof(struct ntb_used), in ntb_dev_info_get() 871 hw->queue_size = conf->queue_size; in ntb_dev_configure() 938 if (peer_val != hw->queue_size) { in ntb_dev_start() 940 hw->queue_size, peer_val); in ntb_dev_start() 1097 hw->queue_size = attr_value; in ntb_attr_set()
|
| /dpdk/lib/vhost/ |
| H A D | vhost_user.c | 1523 get_pervq_shm_size_split(uint16_t queue_size) in get_pervq_shm_size_split() argument 1526 queue_size + sizeof(uint64_t) + in get_pervq_shm_size_split() 1531 get_pervq_shm_size_packed(uint16_t queue_size) in get_pervq_shm_size_packed() argument 1534 * queue_size + sizeof(uint64_t) + in get_pervq_shm_size_packed() 1546 uint16_t num_queues, queue_size; in vhost_user_get_inflight_fd() local 1577 queue_size = ctx->msg.payload.inflight.queue_size; in vhost_user_get_inflight_fd() 1620 for (j = 0; j < queue_size; j++) in vhost_user_get_inflight_fd() 1641 uint16_t num_queues, queue_size; in vhost_user_set_inflight_fd() local 1662 queue_size = ctx->msg.payload.inflight.queue_size; in vhost_user_set_inflight_fd() 1725 vq->inflight_packed->desc_num = queue_size; in vhost_user_set_inflight_fd() [all …]
|
| /dpdk/drivers/net/virtio/ |
| H A D | virtio_user_ethdev.c | 169 return dev->queue_size; in virtio_user_get_queue_num() 467 uint64_t queue_size = VIRTIO_USER_DEF_Q_SZ; in virtio_user_pmd_probe() local 560 &get_integer_arg, &queue_size) < 0) { in virtio_user_pmd_probe() 653 queue_size, mac_addr, &ifname, server_mode, in virtio_user_pmd_probe()
|