| /f-stack/dpdk/lib/librte_eal/freebsd/ |
| H A D | eal_memory.c | 69 uint64_t mem_sz, page_sz; in rte_eal_hugepage_init() local 76 page_sz = RTE_PGSIZE_4K; in rte_eal_hugepage_init() 77 n_segs = mem_sz / page_sz; in rte_eal_hugepage_init() 105 uint64_t page_sz, mem_needed; in rte_eal_hugepage_init() local 109 page_sz = hpi->hugepage_sz; in rte_eal_hugepage_init() 112 page_sz); in rte_eal_hugepage_init() 149 if (msl->page_sz != page_sz) in rte_eal_hugepage_init() 184 (size_t)msl->page_sz * ms_idx); in rte_eal_hugepage_init() 201 seg->hugepage_sz = page_sz; in rte_eal_hugepage_init() 202 seg->len = page_sz; in rte_eal_hugepage_init() [all …]
|
| H A D | eal_memalloc.c | 15 int __rte_unused n_segs, size_t __rte_unused page_sz, in eal_memalloc_alloc_seg_bulk() argument 23 eal_memalloc_alloc_seg(size_t __rte_unused page_sz, int __rte_unused socket) in eal_memalloc_alloc_seg() argument
|
| H A D | eal.c | 185 size_t page_sz = sysconf(_SC_PAGE_SIZE); in rte_eal_config_create() local 187 size_t cfg_len_aligned = RTE_ALIGN(cfg_len, page_sz); in rte_eal_config_create() 200 sizeof(struct rte_mem_config), page_sz); in rte_eal_config_create() 233 &cfg_len_aligned, page_sz, 0, 0); in rte_eal_config_create()
|
| /f-stack/dpdk/lib/librte_eal/common/ |
| H A D | eal_common_memory.c | 91 page_sz == system_page_sz; in eal_get_virtual_area() 104 *size -= page_sz; in eal_get_virtual_area() 199 msl->page_sz = page_sz; in eal_memseg_list_init_named() 206 socket_id, page_sz >> 10); in eal_memseg_list_init_named() 227 size_t page_sz, mem_sz; in eal_memseg_list_alloc() local 230 page_sz = msl->page_sz; in eal_memseg_list_alloc() 261 size_t page_sz = msl->page_sz; in eal_memseg_list_populate() local 273 ms->hugepage_sz = page_sz; in eal_memseg_list_populate() 275 ms->len = page_sz; in eal_memseg_list_populate() 682 len = n_segs * msl->page_sz; in rte_memseg_contig_walk_thread_unsafe() [all …]
|
| H A D | malloc_heap.c | 167 elem->msl->page_sz)) in find_suitable_element() 200 elem->msl->page_sz)) in find_biggest_element() 564 msl->page_sz)) in alloc_more_mem_on_socket() 808 size_t page_sz; in malloc_heap_free_pages() local 814 page_sz = (size_t)msl->page_sz; in malloc_heap_free_pages() 846 page_sz = (size_t)msl->page_sz; in malloc_heap_free() 865 if (elem->size < page_sz) in malloc_heap_free() 927 aligned_len -= page_sz; in malloc_heap_free() 940 aligned_len -= page_sz; in malloc_heap_free() 1149 ms->len = page_sz; in malloc_heap_create_external_seg() [all …]
|
| H A D | eal_common_memzone.c | 183 mz->hugepage_sz = elem->msl->page_sz; in memzone_reserve_aligned_thread_unsafe() 322 size_t page_sz; in dump_memzone() local 342 page_sz = (size_t)mz->hugepage_sz; in dump_memzone() 343 cur_addr = RTE_PTR_ALIGN_FLOOR(mz->addr, page_sz); in dump_memzone() 347 ms_idx = RTE_PTR_DIFF(mz->addr, msl->base_va) / page_sz; in dump_memzone() 354 cur_addr, ms->iova, ms->len, page_sz); in dump_memzone() 357 cur_addr = RTE_PTR_ADD(cur_addr, page_sz); in dump_memzone()
|
| H A D | eal_common_fbarray.c | 74 return RTE_ALIGN_CEIL(data_sz + msk_sz, page_sz); in calc_data_size() 712 size_t page_sz, mmap_len; in rte_fbarray_init() local 736 page_sz = rte_mem_page_size(); in rte_fbarray_init() 737 if (page_sz == (size_t)-1) { in rte_fbarray_init() 838 size_t page_sz, mmap_len; in rte_fbarray_attach() local 863 page_sz = rte_mem_page_size(); in rte_fbarray_attach() 864 if (page_sz == (size_t)-1) { in rte_fbarray_attach() 942 size_t page_sz = rte_mem_page_size(); in rte_fbarray_detach() local 943 if (page_sz == (size_t)-1) in rte_fbarray_detach() 998 size_t page_sz = rte_mem_page_size(); in rte_fbarray_destroy() local [all …]
|
| H A D | malloc_elem.c | 45 size_t page_sz, cur, max; in malloc_elem_find_max_iova_contig() local 49 page_sz = (size_t)elem->msl->page_sz; in malloc_elem_find_max_iova_contig() 70 cur_page = RTE_PTR_ALIGN_FLOOR(contig_seg_start, page_sz); in malloc_elem_find_max_iova_contig() 74 page_end = RTE_PTR_ADD(cur_page, page_sz); in malloc_elem_find_max_iova_contig() 79 expected_iova = ms->iova + page_sz; in malloc_elem_find_max_iova_contig() 83 cur_page = RTE_PTR_ADD(cur_page, page_sz); in malloc_elem_find_max_iova_contig() 86 page_end = RTE_PTR_ADD(cur_page, page_sz); in malloc_elem_find_max_iova_contig() 118 expected_iova = ms->iova + page_sz; in malloc_elem_find_max_iova_contig()
|
| H A D | eal_private.h | 257 size_t page_sz, int flags, int reserve_flags); 279 uint64_t page_sz, int n_segs, int socket_id, bool heap); 291 eal_memseg_list_init(struct rte_memseg_list *msl, uint64_t page_sz,
|
| H A D | eal_memalloc.h | 16 eal_memalloc_alloc_seg(size_t page_sz, int socket); 28 eal_memalloc_alloc_seg_bulk(struct rte_memseg **ms, int n_segs, size_t page_sz,
|
| H A D | rte_malloc.c | 398 rte_iova_t iova_addrs[], unsigned int n_pages, size_t page_sz) in rte_malloc_heap_memory_add() argument 406 page_sz == 0 || !rte_is_power_of_2(page_sz) || in rte_malloc_heap_memory_add() 407 RTE_ALIGN(len, page_sz) != len || in rte_malloc_heap_memory_add() 408 !rte_is_aligned(va_addr, page_sz) || in rte_malloc_heap_memory_add() 409 ((len / page_sz) != n_pages && iova_addrs != NULL) || in rte_malloc_heap_memory_add() 431 n = len / page_sz; in rte_malloc_heap_memory_add() 433 msl = malloc_heap_create_external_seg(va_addr, iova_addrs, n, page_sz, in rte_malloc_heap_memory_add()
|
| H A D | eal_common_dynmem.c | 24 uint64_t page_sz; in eal_dynmem_memseg_lists_init() member 100 memtypes[cur_type].page_sz = hugepage_sz; in eal_dynmem_memseg_lists_init() 136 pagesz = type->page_sz; in eal_dynmem_memseg_lists_init() 210 if (msl->page_sz != hpi->hugepage_sz) in hugepage_count_walk()
|
| H A D | malloc_mp.h | 34 uint64_t page_sz; member
|
| H A D | malloc_heap.h | 67 unsigned int n_pages, size_t page_sz, const char *seg_name,
|
| /f-stack/dpdk/lib/librte_eal/windows/ |
| H A D | eal_memalloc.c | 177 size_t page_sz; member 190 size_t page_sz; in alloc_seg_walk() local 194 if (msl->page_sz != wa->page_sz) in alloc_seg_walk() 199 page_sz = (size_t)msl->page_sz; in alloc_seg_walk() 239 map_addr = RTE_PTR_ADD(cur_msl->base_va, cur_idx * page_sz); in alloc_seg_walk() 302 seg_idx = RTE_PTR_DIFF(wa->ms->addr, start_addr) / msl->page_sz; in free_seg_walk() 317 size_t page_sz, int socket, bool exact) in eal_memalloc_alloc_seg_bulk() argument 333 if (page_sz == hpi->hugepage_sz) { in eal_memalloc_alloc_seg_bulk() 348 wa.page_sz = page_sz; in eal_memalloc_alloc_seg_bulk() 365 eal_memalloc_alloc_seg(size_t page_sz, int socket) in eal_memalloc_alloc_seg() argument [all …]
|
| H A D | eal_memory.c | 655 uint64_t mem_sz, page_sz; in eal_nohuge_init() local 668 page_sz = RTE_PGSIZE_4K; in eal_nohuge_init() 669 n_segs = mem_sz / page_sz; in eal_nohuge_init() 672 msl, "nohugemem", page_sz, n_segs, 0, true)) { in eal_nohuge_init()
|
| /f-stack/dpdk/lib/librte_eal/linux/ |
| H A D | eal_memalloc.c | 488 page_sz, grow); in resize_hugefile() 785 size_t page_sz; member 797 size_t page_sz; in alloc_seg_walk() local 803 if (msl->page_sz != wa->page_sz) in alloc_seg_walk() 808 page_sz = (size_t)msl->page_sz; in alloc_seg_walk() 1011 if (page_sz == in eal_memalloc_alloc_seg_bulk() 1035 wa.page_sz = page_sz; in eal_memalloc_alloc_seg_bulk() 1170 size_t len, page_sz; in sync_chunk() local 1174 page_sz = (size_t)primary_msl->page_sz; in sync_chunk() 1210 size_t len, page_sz; in sync_chunk() local [all …]
|
| H A D | eal_memory.c | 673 uint64_t page_sz; local 693 if (msl->page_sz != page_sz) 746 memseg_len = (size_t)page_sz; 767 munmap(hfile->orig_va, page_sz); 779 ms->hugepage_sz = page_sz; 860 uint64_t page_sz = local 886 else if (cur->size != page_sz) 978 page_sz >> 10); 1138 uint64_t page_sz; local 1150 page_sz = RTE_PGSIZE_4K; [all …]
|
| H A D | eal.c | 273 size_t page_sz = sysconf(_SC_PAGE_SIZE); in rte_eal_config_create() local 275 size_t cfg_len_aligned = RTE_ALIGN(cfg_len, page_sz); in rte_eal_config_create() 290 sizeof(struct rte_mem_config), page_sz); in rte_eal_config_create() 323 &cfg_len_aligned, page_sz, 0, 0); in rte_eal_config_create()
|
| /f-stack/dpdk/drivers/common/mlx5/ |
| H A D | mlx5_common_mr.c | 292 MLX5_ASSERT(msl->page_sz == ms->hugepage_sz); in mr_find_next_chunk() 628 data.start = RTE_ALIGN_FLOOR(addr, data.msl->page_sz); in mlx5_mr_create_primary() 629 data.end = data.start + data.msl->page_sz; in mlx5_mr_create_primary() 647 MLX5_ASSERT(msl->page_sz == ms->hugepage_sz); in mlx5_mr_create_primary() 649 ms_n = len / msl->page_sz; in mlx5_mr_create_primary() 689 if (len > msl->page_sz && in mlx5_mr_create_primary() 702 data.start = RTE_ALIGN_FLOOR(addr, msl->page_sz); in mlx5_mr_create_primary() 703 data.end = data.start + msl->page_sz; in mlx5_mr_create_primary() 741 start = data_re.start + n * msl->page_sz; in mlx5_mr_create_primary() 754 data.end = start + msl->page_sz; in mlx5_mr_create_primary() [all …]
|
| /f-stack/dpdk/drivers/net/mlx4/ |
| H A D | mlx4_mr.c | 317 MLX4_ASSERT(msl->page_sz == ms->hugepage_sz); in mr_find_next_chunk() 637 data.end = data.start + data.msl->page_sz; in mlx4_mr_create_primary() 653 MLX4_ASSERT(msl->page_sz == ms->hugepage_sz); in mlx4_mr_create_primary() 655 ms_n = len / msl->page_sz; in mlx4_mr_create_primary() 700 if (len > msl->page_sz && in mlx4_mr_create_primary() 715 data.end = data.start + msl->page_sz; in mlx4_mr_create_primary() 754 start = data_re.start + n * msl->page_sz; in mlx4_mr_create_primary() 766 data.end = start + msl->page_sz; in mlx4_mr_create_primary() 772 mr->ms_bmp_n = len / msl->page_sz; in mlx4_mr_create_primary() 910 ms_n = len / msl->page_sz; in mlx4_mr_mem_event_free_cb() [all …]
|
| /f-stack/dpdk/drivers/net/mlx5/ |
| H A D | mlx5_mr.c | 65 RTE_ALIGN((uintptr_t)addr, msl->page_sz)); in mlx5_mr_mem_event_free_cb() 66 MLX5_ASSERT(len == RTE_ALIGN(len, msl->page_sz)); in mlx5_mr_mem_event_free_cb() 67 ms_n = len / msl->page_sz; in mlx5_mr_mem_event_free_cb() 78 start = (uintptr_t)addr + i * msl->page_sz; in mlx5_mr_mem_event_free_cb() 85 MLX5_ASSERT(msl->page_sz == ms->hugepage_sz); in mlx5_mr_mem_event_free_cb()
|
| /f-stack/dpdk/lib/librte_eal/include/ |
| H A D | rte_memory.h | 72 uint64_t page_sz; /**< Page size for all memsegs in this list. */ member 458 unsigned int n_pages, size_t page_sz);
|
| H A D | rte_malloc.h | 345 rte_iova_t iova_addrs[], unsigned int n_pages, size_t page_sz);
|
| /f-stack/dpdk/app/test/ |
| H A D | test_memzone.c | 121 if (msl->page_sz == RTE_PGSIZE_2M) in find_available_pagesz() 123 if (msl->page_sz == RTE_PGSIZE_1G) in find_available_pagesz() 125 if (msl->page_sz == RTE_PGSIZE_16M) in find_available_pagesz() 127 if (msl->page_sz == RTE_PGSIZE_16G) in find_available_pagesz()
|