| /dpdk/drivers/raw/ioat/ |
| H A D | ioat_rawdev.c | 51 if (params->ring_size > 4096 || params->ring_size < 64 || in ioat_dev_configure() 52 !rte_is_power_of_2(params->ring_size)) in ioat_dev_configure() 55 ioat->ring_size = params->ring_size; in ioat_dev_configure() 68 (DESC_SZ + COMPLETION_SZ) * ioat->ring_size, in ioat_dev_configure() 73 ioat->hdls = (void *)&ioat->desc_ring[ioat->ring_size]; in ioat_dev_configure() 78 for (i = 0; i < ioat->ring_size; i++) { in ioat_dev_configure() 80 (((i + 1) % ioat->ring_size) * DESC_SZ); in ioat_dev_configure() 91 if (ioat->ring_size == 0 || ioat->desc_ring == NULL) in ioat_dev_start() 100 ioat->status = ioat->ring_addr + ((ioat->ring_size - 1) * DESC_SZ); in ioat_dev_start() 120 cfg->ring_size = ioat->ring_size; in ioat_dev_info_get() [all …]
|
| H A D | rte_ioat_rawdev_fns.h | 93 unsigned short ring_size; member 119 unsigned short size = ioat->ring_size - 1; in __ioat_burst_capacity() 135 unsigned short mask = ioat->ring_size - 1; in __ioat_write_desc() 192 unsigned short mask = ioat->ring_size - 1; in __ioat_fence() 210 ioat->desc_ring[(ioat->next_write - 1) & (ioat->ring_size - 1)].u in __ioat_perform_ops() 244 unsigned short mask = (ioat->ring_size - 1); in __ioat_completed_ops()
|
| H A D | ioat_rawdev_test.c | 597 struct rte_ioat_rawdev_config p = { .ring_size = -1 }; in ioat_rawdev_test() 612 if (p.ring_size != expected_ring_size[dev_id]) { in ioat_rawdev_test() 614 (int)p.ring_size, expected_ring_size[dev_id]); in ioat_rawdev_test() 618 p.ring_size = IOAT_TEST_RINGSIZE; in ioat_rawdev_test() 624 if (p.ring_size != IOAT_TEST_RINGSIZE) { in ioat_rawdev_test() 626 IOAT_TEST_RINGSIZE, (int)p.ring_size); in ioat_rawdev_test() 629 expected_ring_size[dev_id] = p.ring_size; in ioat_rawdev_test() 637 p.ring_size * 2, /* n == num elements */ in ioat_rawdev_test()
|
| /dpdk/drivers/event/sw/ |
| H A D | event_ring.h | 25 uint32_t ring_size; member 37 const uint32_t ring_size = rte_align32pow2(size + 1); in rob_ring_create() local 39 (ring_size * sizeof(retval->ring[0])); in rob_ring_create() 44 retval->ring_size = ring_size; in rob_ring_create() 45 retval->mask = ring_size - 1; in rob_ring_create()
|
| /dpdk/drivers/net/bnxt/ |
| H A D | bnxt_rxtx_vec_common.h | 107 uint32_t ring_mask, ring_size, num; in bnxt_tx_cmp_vec_fast() local 111 ring_size = txr->tx_ring_struct->ring_size; in bnxt_tx_cmp_vec_fast() 114 num = RTE_MIN(nr_pkts, ring_size - cons); in bnxt_tx_cmp_vec_fast() 138 uint32_t ring_mask, ring_size, num, blk; in bnxt_tx_cmp_vec() local 142 ring_size = txr->tx_ring_struct->ring_size; in bnxt_tx_cmp_vec() 148 num = RTE_MIN(nr_pkts, ring_size - cons); in bnxt_tx_cmp_vec()
|
| H A D | bnxt_rxr.c | 301 cpr->cp_ring_struct->ring_size); in bnxt_agg_bufs_valid() 880 cpr->cp_ring_struct->ring_size)) in bnxt_rx_pkt() 1068 cpr->cp_ring_struct->ring_size)) in bnxt_recv_pkts() 1213 ring->ring_mask = ring->ring_size - 1; in bnxt_init_rx_ring_struct() 1245 ring->ring_size = rxr->rx_ring_struct->ring_size * 2; in bnxt_init_rx_ring_struct() 1249 ring->ring_size = rte_align32pow2(ring->ring_size); in bnxt_init_rx_ring_struct() 1250 ring->ring_mask = ring->ring_size - 1; in bnxt_init_rx_ring_struct() 1271 ring->ring_mask = ring->ring_size - 1; in bnxt_init_rx_ring_struct() 1289 for (j = 0; j < ring->ring_size; j++) { in bnxt_init_rxbds() 1320 for (i = 0; i < ring->ring_size; i++) { in bnxt_init_one_rx_ring() [all …]
|
| H A D | bnxt_ring.c | 156 int cp_ring_len = RTE_CACHE_LINE_ROUNDUP(cp_ring->ring_size * in bnxt_alloc_rings() 174 RTE_CACHE_LINE_ROUNDUP(rx_ring_info->rx_ring_struct->ring_size * in bnxt_alloc_rings() 187 rx_ring_info->rx_ring_struct->ring_size * in bnxt_alloc_rings() 281 rte_bitmap_init(rx_ring_info->rx_ring_struct->ring_size * in bnxt_alloc_rings() 468 ring->ring_size = rte_align32pow2(DEFAULT_CP_RING_SIZE); in bnxt_alloc_rxtx_nq_ring() 469 ring->ring_mask = ring->ring_size - 1; in bnxt_alloc_rxtx_nq_ring() 600 cp_ring->ring_size = rxr->rx_ring_struct->ring_size * 2; in bnxt_alloc_hwrm_rx_ring() 603 cp_ring->ring_size *= AGG_RING_SIZE_FACTOR; in bnxt_alloc_hwrm_rx_ring() 605 cp_ring->ring_mask = cp_ring->ring_size - 1; in bnxt_alloc_hwrm_rx_ring() 852 ring->ring_size = rte_align32pow2(DEFAULT_CP_RING_SIZE); in bnxt_alloc_async_ring_struct() [all …]
|
| H A D | bnxt_rxtx_vec_sse.c | 159 uint16_t cp_ring_size = cpr->cp_ring_struct->ring_size; in recv_burst_vec_sse() 160 uint16_t rx_ring_size = rxr->rx_ring_struct->ring_size; in recv_burst_vec_sse() 426 uint16_t ring_size = txr->tx_ring_struct->ring_size; in bnxt_xmit_pkts_vec() local 451 num = RTE_MIN(num, ring_size - in bnxt_xmit_pkts_vec() 452 (txr->tx_raw_prod & (ring_size - 1))); in bnxt_xmit_pkts_vec()
|
| H A D | bnxt_txr.c | 54 txq->tx_wake_thresh = ring->ring_size / 2; in bnxt_init_one_tx_ring() 79 ring->ring_size = rte_align32pow2(txq->nb_tx_desc); in bnxt_init_tx_ring_struct() 80 ring->ring_mask = ring->ring_size - 1; in bnxt_init_tx_ring_struct() 83 ring->vmem_size = ring->ring_size * sizeof(struct rte_mbuf *); in bnxt_init_tx_ring_struct() 100 ring->ring_size = txr->tx_ring_struct->ring_size; in bnxt_init_tx_ring_struct() 101 ring->ring_mask = ring->ring_size - 1; in bnxt_init_tx_ring_struct()
|
| H A D | bnxt_rxtx_vec_avx2.c | 30 uint16_t cp_ring_size = cpr->cp_ring_struct->ring_size; in recv_burst_vec_avx2() 31 uint16_t rx_ring_size = rxr->rx_ring_struct->ring_size; in recv_burst_vec_avx2() 552 uint16_t ring_size = txr->tx_ring_struct->ring_size; in bnxt_xmit_pkts_vec_avx2() local 577 num = RTE_MIN(num, ring_size - in bnxt_xmit_pkts_vec_avx2() 578 (txr->tx_raw_prod & (ring_size - 1))); in bnxt_xmit_pkts_vec_avx2()
|
| H A D | bnxt_cpr.h | 139 bnxt_cpr_cmp_valid(const void *cmpl, uint32_t raw_cons, uint32_t ring_size) in bnxt_cpr_cmp_valid() argument 144 expected = !(raw_cons & ring_size); in bnxt_cpr_cmp_valid()
|
| /dpdk/drivers/net/ark/ |
| H A D | ark_mpu.c | 93 ark_mpu_configure(struct ark_mpu_t *mpu, rte_iova_t ring, uint32_t ring_size, in ark_mpu_configure() argument 98 if (!rte_is_power_of_2(ring_size)) { in ark_mpu_configure() 100 ring_size); in ark_mpu_configure() 105 mpu->cfg.ring_size = ring_size; in ark_mpu_configure() 106 mpu->cfg.ring_mask = ring_size - 1; in ark_mpu_configure()
|
| H A D | ark_mpu.h | 50 uint32_t ring_size; member 112 uint32_t ring_size, int is_tx);
|
| /dpdk/drivers/bus/vmbus/linux/ |
| H A D | vmbus_uio.c | 190 void **ring_buf, uint32_t *ring_size) in vmbus_uio_map_primary() argument 206 *ring_size = uio_res->maps[HV_TXRX_RING_MAP].size / 2; in vmbus_uio_map_primary() 213 void **ring_buf, uint32_t *ring_size) in vmbus_uio_map_subchan() argument 308 *ring_size = file_size / 2; in vmbus_uio_map_subchan() 317 uint32_t ring_size; in vmbus_uio_map_rings() local 323 ret = vmbus_uio_map_primary(chan, &ring_buf, &ring_size); in vmbus_uio_map_rings() 325 ret = vmbus_uio_map_subchan(dev, chan, &ring_buf, &ring_size); in vmbus_uio_map_rings() 330 vmbus_br_setup(&chan->txbr, ring_buf, ring_size); in vmbus_uio_map_rings() 331 vmbus_br_setup(&chan->rxbr, (char *)ring_buf + ring_size, ring_size); in vmbus_uio_map_rings()
|
| /dpdk/examples/multi_process/simple_mp/ |
| H A D | main.c | 77 const unsigned ring_size = 64; in main() local 91 send_ring = rte_ring_create(_PRI_2_SEC, ring_size, rte_socket_id(), flags); in main() 92 recv_ring = rte_ring_create(_SEC_2_PRI, ring_size, rte_socket_id(), flags); in main()
|
| /dpdk/drivers/net/enetfec/ |
| H A D | enet_ethdev.c | 165 for (i = 0; i < rxq->bd.ring_size; i++) { in enet_free_buffers() 176 for (i = 0; i < txq->bd.ring_size; i++) { in enet_free_buffers() 389 txq->bd.ring_size = nb_desc; in enetfec_tx_queue_setup() 390 fep->total_tx_ring_size += txq->bd.ring_size; in enetfec_tx_queue_setup() 399 size = dsize * txq->bd.ring_size; in enetfec_tx_queue_setup() 413 for (i = 0; i < txq->bd.ring_size; i++) { in enetfec_tx_queue_setup() 469 rxq->bd.ring_size = nb_rx_desc; in enetfec_rx_queue_setup() 470 fep->total_rx_ring_size += rxq->bd.ring_size; in enetfec_rx_queue_setup() 481 size = dsize * rxq->bd.ring_size; in enetfec_rx_queue_setup() 517 for (i = 0; i < rxq->bd.ring_size; i++) { in enetfec_rx_queue_setup()
|
| /dpdk/app/dumpcap/ |
| H A D | main.c | 62 static unsigned int ring_size = 2048; variable 375 ring_size = get_uint(optarg, "packet_limit", 0); in parse_opts() 544 size = ring_size; in create_ring() 548 if (size != ring_size) { in create_ring() 550 ring_size, size); in create_ring() 551 ring_size = size; in create_ring() 556 ring = rte_ring_create(RING_NAME, ring_size, in create_ring() 568 size_t num_mbufs = 2 * ring_size; in create_mempool()
|
| /dpdk/lib/ring/ |
| H A D | rte_ring.c | 247 ssize_t ring_size; in rte_ring_create_elem() local 259 ring_size = rte_ring_get_memsize_elem(esize, count); in rte_ring_create_elem() 260 if (ring_size < 0) { in rte_ring_create_elem() 261 rte_errno = -ring_size; in rte_ring_create_elem() 284 mz = rte_memzone_reserve_aligned(mz_name, ring_size, socket_id, in rte_ring_create_elem()
|
| /dpdk/examples/dma/ |
| H A D | dmafwd.c | 96 static unsigned short ring_size = 2048; variable 209 "Ring Size = %d", ring_size); in print_stats() 699 ring_size = atoi(optarg); in dma_parse_args() 700 if (ring_size == 0) { in dma_parse_args() 708 if (ring_size > MBUF_RING_SIZE) { in dma_parse_args() 711 ring_size = MBUF_RING_SIZE; in dma_parse_args() 783 .nb_desc = ring_size in configure_dmadev_queue() 844 ring_name, ring_size, in assign_rings() 1036 4 * MAX_PKT_BURST + ring_size) + ring_size + in main()
|
| /dpdk/drivers/net/ionic/ |
| H A D | ionic_dev.c | 350 .q_init.ring_size = rte_log2_u32(q->num_descs), in ionic_dev_cmd_adminq_init() 410 uint32_t ring_size; in ionic_q_init() local 415 ring_size = rte_log2_u32(num_descs); in ionic_q_init() 416 if (ring_size < 2 || ring_size > 16) in ionic_q_init()
|
| /dpdk/drivers/net/memif/ |
| H A D | rte_eth_memif.c | 330 ring_size = 1 << mq->log2_ring_size; in eth_memif_rx() 331 mask = ring_size - 1; in eth_memif_rx() 469 ring_size = 1 << mq->log2_ring_size; in eth_memif_rx_zc() 470 mask = ring_size - 1; in eth_memif_rx_zc() 528 n_slots = ring_size - head + mq->last_tail; in eth_memif_rx_zc() 594 ring_size = 1 << mq->log2_ring_size; in eth_memif_tx() 595 mask = ring_size - 1; in eth_memif_tx() 605 n_free = ring_size - slot + in eth_memif_tx() 760 ring_size = 1 << mq->log2_ring_size; in eth_memif_tx_zc() 761 mask = ring_size - 1; in eth_memif_tx_zc() [all …]
|
| /dpdk/drivers/crypto/armv8/ |
| H A D | rte_armv8_pmd_ops.c | 193 unsigned int ring_size, int socket_id) in armv8_crypto_pmd_qp_create_processed_ops_ring() argument 199 if (rte_ring_get_size(r) >= ring_size) { in armv8_crypto_pmd_qp_create_processed_ops_ring() 212 return rte_ring_create(qp->name, ring_size, socket_id, in armv8_crypto_pmd_qp_create_processed_ops_ring()
|
| /dpdk/drivers/compress/zlib/ |
| H A D | zlib_pmd_ops.c | 157 unsigned int ring_size, int socket_id) in zlib_pmd_qp_create_processed_pkts_ring() argument 162 if (rte_ring_get_size(r) >= ring_size) { in zlib_pmd_qp_create_processed_pkts_ring() 173 return rte_ring_create(qp->name, ring_size, socket_id, in zlib_pmd_qp_create_processed_pkts_ring()
|
| /dpdk/drivers/net/mlx5/ |
| H A D | mlx5_ethdev.c | 241 info->default_rxportconf.ring_size = 256; in mlx5_set_default_params() 242 info->default_txportconf.ring_size = 256; in mlx5_set_default_params() 252 info->default_rxportconf.ring_size = 2048; in mlx5_set_default_params() 253 info->default_txportconf.ring_size = 2048; in mlx5_set_default_params() 261 info->default_rxportconf.ring_size = 4096; in mlx5_set_default_params() 262 info->default_txportconf.ring_size = 4096; in mlx5_set_default_params()
|
| /dpdk/drivers/crypto/null/ |
| H A D | null_crypto_pmd_ops.c | 164 unsigned ring_size, int socket_id) in null_crypto_pmd_qp_create_processed_pkts_ring() argument 170 if (rte_ring_get_size(r) >= ring_size) { in null_crypto_pmd_qp_create_processed_pkts_ring() 183 return rte_ring_create(qp->name, ring_size, socket_id, in null_crypto_pmd_qp_create_processed_pkts_ring()
|