| /f-stack/dpdk/lib/librte_mempool/ |
| H A D | rte_mempool_ops_default.c | 16 size_t obj_per_page, pg_sz, objs_in_last_page; in rte_mempool_op_calc_mem_size_helper() local 25 pg_sz = (size_t)1 << pg_shift; in rte_mempool_op_calc_mem_size_helper() 26 if (chunk_reserve >= pg_sz) in rte_mempool_op_calc_mem_size_helper() 28 obj_per_page = (pg_sz - chunk_reserve) / total_elt_sz; in rte_mempool_op_calc_mem_size_helper() 37 pg_sz) * obj_num; in rte_mempool_op_calc_mem_size_helper() 79 check_obj_bounds(char *obj, size_t pg_sz, size_t elt_sz) in check_obj_bounds() argument 81 if (pg_sz == 0) in check_obj_bounds() 83 if (elt_sz > pg_sz) in check_obj_bounds() 85 if (RTE_PTR_ALIGN(obj, pg_sz) != RTE_PTR_ALIGN(obj + elt_sz - 1, pg_sz)) in check_obj_bounds() 97 size_t total_elt_sz, pg_sz; in rte_mempool_op_populate_helper() local [all …]
|
| H A D | rte_mempool.c | 456 *pg_sz = 0; in rte_mempool_get_page_size() 458 *pg_sz = get_min_page_size(mp->socket_id); in rte_mempool_get_page_size() 460 *pg_sz = rte_mem_page_size(); in rte_mempool_get_page_size() 462 rte_mempool_trace_get_page_size(mp, *pg_sz); in rte_mempool_get_page_size() 477 size_t align, pg_sz, pg_shift = 0; in rte_mempool_populate_default() local 535 if (pg_sz != 0) in rte_mempool_populate_default() 536 pg_shift = rte_bsf32(pg_sz); in rte_mempool_populate_default() 592 mz->len, pg_sz, in rte_mempool_populate_default() 616 size_t pg_sz, pg_shift; in get_anon_size() local 620 pg_sz = rte_mem_page_size(); in get_anon_size() [all …]
|
| H A D | rte_mempool_trace.h | 83 size_t len, size_t pg_sz, void *free_cb, void *opaque), 88 rte_trace_point_emit_size_t(pg_sz); 126 RTE_TRACE_POINT_ARGS(struct rte_mempool *mempool, size_t pg_sz), 129 rte_trace_point_emit_size_t(pg_sz);
|
| H A D | rte_mempool.h | 1121 size_t len, size_t pg_sz, rte_mempool_memchunk_free_cb_t *free_cb, 1765 rte_mempool_get_page_size(struct rte_mempool *mp, size_t *pg_sz);
|
| /f-stack/freebsd/x86/iommu/ |
| H A D | intel_idpgtbl.c | 117 iommu_gaddr_t f, pg_sz; in domain_idmap_nextlvl() local 429 iommu_gaddr_t pg_sz, base1, size1; in domain_map_buf_locked() local 441 for (sf = NULL, pi = 0; size > 0; base += pg_sz, size -= pg_sz, in domain_map_buf_locked() 445 run_sz = pg_sz >> DMAR_PAGE_SHIFT; in domain_map_buf_locked() 458 if (size < pg_sz || (base & (pg_sz - 1)) != 0 || in domain_map_buf_locked() 475 KASSERT(size >= pg_sz, in domain_map_buf_locked() 622 iommu_gaddr_t pg_sz; in domain_unmap_buf_locked() local 648 pg_sz = 0; /* silence gcc */ in domain_unmap_buf_locked() 652 for (sf = NULL; size > 0; base += pg_sz, size -= pg_sz) { in domain_unmap_buf_locked() 658 if (pg_sz > size) in domain_unmap_buf_locked() [all …]
|
| H A D | intel_utils.c | 214 static const iommu_gaddr_t pg_sz[] = { in pglvl_page_size() local 226 KASSERT(rlvl < nitems(pg_sz), ("sizeof pg_sz lvl %d", lvl)); in pglvl_page_size() 227 return (pg_sz[rlvl]); in pglvl_page_size()
|
| /f-stack/dpdk/lib/librte_eal/common/ |
| H A D | malloc_heap.c | 304 alloc_sz = (size_t)pg_sz * n_segs; in alloc_pages_on_heap() 400 MALLOC_ELEM_TRAILER_LEN, pg_sz); in try_expand_heap_primary() 401 n_segs = alloc_sz / pg_sz; in try_expand_heap_primary() 474 req.alloc_req.page_sz = pg_sz; in try_expand_heap_secondary() 582 if (prev_pg_sz != pg_sz) { in alloc_more_mem_on_socket() 584 prev_pg_sz = pg_sz; in alloc_more_mem_on_socket() 592 if (prev_pg_sz != pg_sz) { in alloc_more_mem_on_socket() 593 other_pg_sz[n_other_pg_sz++] = pg_sz; in alloc_more_mem_on_socket() 594 prev_pg_sz = pg_sz; in alloc_more_mem_on_socket() 602 uint64_t pg_sz = requested_pg_sz[i]; in alloc_more_mem_on_socket() local [all …]
|
| H A D | malloc_mp.h | 78 alloc_pages_on_heap(struct malloc_heap *heap, uint64_t pg_sz, size_t elt_size,
|
| /f-stack/dpdk/drivers/crypto/octeontx/ |
| H A D | otx_cryptodev_hw_access.c | 517 uint64_t pg_sz = sysconf(_SC_PAGESIZE); in otx_cpt_get_resource() local 541 len = RTE_ALIGN(len, pg_sz); in otx_cpt_get_resource() 547 len = RTE_ALIGN(len, pg_sz); in otx_cpt_get_resource() 584 mem += RTE_ALIGN(used_len, pg_sz) - used_len; in otx_cpt_get_resource() 585 len -= RTE_ALIGN(used_len, pg_sz) - used_len; in otx_cpt_get_resource() 586 dma_addr += RTE_ALIGN(used_len, pg_sz) - used_len; in otx_cpt_get_resource()
|
| /f-stack/freebsd/contrib/octeon-sdk/ |
| H A D | cvmx-app-hotplug.c | 534 size_t pg_sz = sysconf(_SC_PAGESIZE), size; in cvmx_app_get_hotplug_global_ptr() local 548 size = CVMX_APP_HOTPLUG_INFO_REGION_SIZE + pg_sz-1; in cvmx_app_get_hotplug_global_ptr() 549 offset = block_desc->base_addr & ~(pg_sz-1); in cvmx_app_get_hotplug_global_ptr() 557 hgp = (cvmx_app_hotplug_global_t *)(vaddr + ( block_desc->base_addr & (pg_sz-1))); in cvmx_app_get_hotplug_global_ptr()
|
| /f-stack/dpdk/drivers/net/hinic/base/ |
| H A D | hinic_pmd_nicio.h | 21 #define CI_TABLE_SIZE(num_qps, pg_sz) \ argument 22 (ALIGN((num_qps) * HINIC_CI_Q_ADDR_SIZE, pg_sz))
|
| /f-stack/dpdk/drivers/regex/octeontx2/ |
| H A D | otx2_regexdev.c | 68 uint64_t pg_sz = sysconf(_SC_PAGESIZE); in ree_qp_create() local 99 len = RTE_ALIGN(len, pg_sz); in ree_qp_create() 105 len = RTE_ALIGN(len, pg_sz); in ree_qp_create() 130 used_len = RTE_ALIGN(used_len, pg_sz); in ree_qp_create()
|
| /f-stack/dpdk/drivers/mempool/bucket/ |
| H A D | rte_mempool_bucket.c | 462 size_t pg_sz; in bucket_alloc() local 464 rc = rte_mempool_get_page_size(mp, &pg_sz); in bucket_alloc() 482 bd->bucket_mem_size = RTE_MIN(pg_sz, in bucket_alloc()
|
| /f-stack/dpdk/drivers/crypto/octeontx2/ |
| H A D | otx2_cryptodev_ops.c | 168 uint64_t pg_sz = sysconf(_SC_PAGESIZE); in otx2_cpt_qp_create() local 201 len = RTE_ALIGN(len, pg_sz); in otx2_cpt_qp_create() 207 len = RTE_ALIGN(len, pg_sz); in otx2_cpt_qp_create() 239 used_len = RTE_ALIGN(used_len, pg_sz); in otx2_cpt_qp_create()
|