Home
last modified time | relevance | path

Searched refs:alloc_sz (Results 1 – 6 of 6) sorted by relevance

/dpdk/lib/eal/windows/
H A Deal_memalloc.c40 size_t alloc_sz; in alloc_seg() local
63 alloc_sz = hi->hugepage_sz; in alloc_seg()
67 addr = eal_mem_alloc_socket(alloc_sz, socket_id); in alloc_seg()
70 "on socket %d\n", alloc_sz, socket_id); in alloc_seg()
75 addr = eal_mem_commit(requested_addr, alloc_sz, socket_id); in alloc_seg()
91 requested_addr, alloc_sz, socket_id); in alloc_seg()
130 ms->len = alloc_sz; in alloc_seg()
140 if (eal_mem_decommit(addr, alloc_sz) && (rte_errno == EADDRNOTAVAIL)) { in alloc_seg()
/dpdk/drivers/event/dlb2/pf/
H A Ddlb2_pf.c268 int alloc_sz, qe_sz; in dlb2_pf_ldb_port_create() local
284 alloc_sz = cq_alloc_depth * qe_sz; in dlb2_pf_ldb_port_create()
285 alloc_sz = RTE_CACHE_LINE_ROUNDUP(alloc_sz); in dlb2_pf_ldb_port_create()
287 port_base = dlb2_alloc_coherent_aligned(&mz, &cq_base, alloc_sz, in dlb2_pf_ldb_port_create()
299 memset(port_base, 0, alloc_sz); in dlb2_pf_ldb_port_create()
344 int alloc_sz, qe_sz; in dlb2_pf_dir_port_create() local
359 alloc_sz = cfg->cq_depth * qe_sz; in dlb2_pf_dir_port_create()
360 alloc_sz = RTE_CACHE_LINE_ROUNDUP(alloc_sz); in dlb2_pf_dir_port_create()
362 port_base = dlb2_alloc_coherent_aligned(&mz, &cq_base, alloc_sz, in dlb2_pf_dir_port_create()
374 memset(port_base, 0, alloc_sz); in dlb2_pf_dir_port_create()
/dpdk/lib/eal/linux/
H A Deal_memalloc.c529 size_t alloc_sz; in alloc_seg() local
535 alloc_sz = hi->hugepage_sz; in alloc_seg()
577 map_offset = seg_idx * alloc_sz; in alloc_seg()
586 if (ftruncate(fd, alloc_sz) < 0) { in alloc_seg()
622 munmap(va, alloc_sz); in alloc_seg()
635 (unsigned int)(alloc_sz >> 20)); in alloc_seg()
683 ms->hugepage_sz = alloc_sz; in alloc_seg()
684 ms->len = alloc_sz; in alloc_seg()
694 munmap(addr, alloc_sz); in alloc_seg()
698 new_addr = eal_get_virtual_area(addr, &alloc_sz, alloc_sz, 0, flags); in alloc_seg()
[all …]
/dpdk/lib/eal/common/
H A Dmalloc_heap.c301 size_t alloc_sz; in alloc_pages_on_heap() local
306 alloc_sz = (size_t)pg_sz * n_segs; in alloc_pages_on_heap()
310 heap->total_size + alloc_sz) < 0) { in alloc_pages_on_heap()
389 rollback_expand_heap(ms, n_segs, elem, map_addr, alloc_sz); in alloc_pages_on_heap()
401 size_t alloc_sz; in try_expand_heap_primary() local
405 alloc_sz = RTE_ALIGN_CEIL(align + elt_size + in try_expand_heap_primary()
407 n_segs = alloc_sz / pg_sz; in try_expand_heap_primary()
441 heap->total_size += alloc_sz; in try_expand_heap_primary()
444 socket, alloc_sz >> 20ULL); in try_expand_heap_primary()
453 map_addr, alloc_sz); in try_expand_heap_primary()
[all …]
H A Dmalloc_mp.c223 size_t alloc_sz; in handle_alloc_request() local
252 alloc_sz = RTE_ALIGN_CEIL(ar->align + ar->elt_size + in handle_alloc_request()
254 n_segs = alloc_sz / ar->page_sz; in handle_alloc_request()
273 eal_memalloc_mem_event_notify(RTE_MEM_EVENT_ALLOC, map_addr, alloc_sz); in handle_alloc_request()
283 req->alloc_state.map_len = alloc_sz; in handle_alloc_request()
/dpdk/app/test-crypto-perf/
H A Dcperf_test_pmd_cyclecount.c102 size_t alloc_sz = sizeof(struct rte_crypto_op *) * in cperf_pmd_cyclecount_test_constructor() local
130 ctx->ops = rte_malloc("ops", alloc_sz, 0); in cperf_pmd_cyclecount_test_constructor()
134 ctx->ops_processed = rte_malloc("ops_processed", alloc_sz, 0); in cperf_pmd_cyclecount_test_constructor()