| /linux-6.15/drivers/android/ |
| H A D | binder_alloc.c | 64 return alloc->vm_start + alloc->buffer_size - buffer->user_data; in binder_alloc_buffer_size() 300 mdata->alloc = alloc; in binder_page_alloc() 343 alloc->pid, addr - alloc->vm_start); in binder_install_single_page() 355 alloc->pid, __func__, addr - alloc->vm_start, ret); in binder_install_single_page() 483 if (alloc->free_async_space >= alloc->buffer_size / 10) { in debug_low_async_space_locked() 601 alloc->pid, size, alloc->free_async_space); in binder_alloc_new_buf_locked() 765 alloc->pid, size, alloc->free_async_space); in binder_free_buf_locked() 919 alloc->pages = kvcalloc(alloc->buffer_size / PAGE_SIZE, in binder_alloc_mmap_handler() 939 alloc->free_async_space = alloc->buffer_size / 2; in binder_alloc_mmap_handler() 1024 if (alloc->mm) in binder_alloc_deferred_release() [all …]
|
| H A D | binder_alloc_selftest.c | 108 if (!alloc->pages[page_index] || in check_buffer_pages_allocated() 111 alloc->pages[page_index] ? in check_buffer_pages_allocated() 143 binder_alloc_free_buf(alloc, buffers[seq[i]]); in binder_selftest_free_buf() 151 if (list_empty(page_to_lru(alloc->pages[i]))) { in binder_selftest_free_buf() 154 alloc->pages[i] ? "alloc" : "free", i); in binder_selftest_free_buf() 171 if (alloc->pages[i]) { in binder_selftest_free_page() 173 list_empty(page_to_lru(alloc->pages[i])) ? in binder_selftest_free_page() 194 binder_selftest_free_page(alloc); in binder_selftest_alloc_free() 250 binder_selftest_free_seq(alloc, back_sizes, seq, 0, alloc->buffer_size); in binder_selftest_alloc_size() 260 binder_selftest_alloc_size(alloc, end_offset); in binder_selftest_alloc_offset() [all …]
|
| H A D | binder_alloc.h | 69 struct binder_alloc *alloc; member 124 void binder_selftest_alloc(struct binder_alloc *alloc); 136 void binder_alloc_init(struct binder_alloc *alloc); 139 void binder_alloc_vma_close(struct binder_alloc *alloc); 141 binder_alloc_prepare_to_free(struct binder_alloc *alloc, 143 void binder_alloc_free_buf(struct binder_alloc *alloc, 150 struct binder_alloc *alloc); 152 struct binder_alloc *alloc); 165 mutex_lock(&alloc->mutex); in binder_alloc_get_free_async_space() 166 free_async_space = alloc->free_async_space; in binder_alloc_get_free_async_space() [all …]
|
| H A D | binder_trace.h | 329 __entry->proc = alloc->pid; 341 TP_ARGS(alloc, page_index), 347 __entry->proc = alloc->pid; 356 TP_ARGS(alloc, page_index)); 360 TP_ARGS(alloc, page_index)); 364 TP_ARGS(alloc, page_index)); 368 TP_ARGS(alloc, page_index)); 372 TP_ARGS(alloc, page_index)); 376 TP_ARGS(alloc, page_index)); 380 TP_ARGS(alloc, page_index)); [all …]
|
| /linux-6.15/drivers/infiniband/hw/cxgb4/ |
| H A D | id_table.c | 51 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); in c4iw_id_alloc() 53 obj = find_first_zero_bit(alloc->table, alloc->max); in c4iw_id_alloc() 60 if (alloc->last >= alloc->max) in c4iw_id_alloc() 61 alloc->last = 0; in c4iw_id_alloc() 63 obj += alloc->start; in c4iw_id_alloc() 75 obj -= alloc->start; in c4iw_id_free() 85 alloc->start = start; in c4iw_id_table_alloc() 86 alloc->flags = flags; in c4iw_id_table_alloc() 90 alloc->last = 0; in c4iw_id_table_alloc() 91 alloc->max = num; in c4iw_id_table_alloc() [all …]
|
| /linux-6.15/sound/isa/gus/ |
| H A D | gus_mem.c | 43 pblock = alloc->first; in snd_gf1_mem_xalloc() 50 alloc->first = nblock; in snd_gf1_mem_xalloc() 59 if (alloc->last == NULL) { in snd_gf1_mem_xalloc() 61 alloc->first = alloc->last = nblock; in snd_gf1_mem_xalloc() 63 nblock->prev = alloc->last; in snd_gf1_mem_xalloc() 65 alloc->last = nblock; in snd_gf1_mem_xalloc() 86 if (alloc->last == block) { in snd_gf1_mem_xfree() 87 alloc->last = block->prev; in snd_gf1_mem_xfree() 132 struct snd_gf1_bank_info *info = w_16 ? alloc->banks_16 : alloc->banks_8; in snd_gf1_mem_find() 236 alloc->first = alloc->last = NULL; in snd_gf1_mem_init() [all …]
|
| /linux-6.15/drivers/infiniband/hw/mthca/ |
| H A D | mthca_allocator.c | 47 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); in mthca_alloc() 49 alloc->top = (alloc->top + alloc->max) & alloc->mask; in mthca_alloc() 50 obj = find_first_zero_bit(alloc->table, alloc->max); in mthca_alloc() 55 obj |= alloc->top; in mthca_alloc() 73 alloc->last = min(alloc->last, obj); in mthca_free() 74 alloc->top = (alloc->top + alloc->max) & alloc->mask; in mthca_free() 86 alloc->last = 0; in mthca_alloc_init() 87 alloc->top = 0; in mthca_alloc_init() 88 alloc->max = num; in mthca_alloc_init() 89 alloc->mask = mask; in mthca_alloc_init() [all …]
|
| H A D | mthca_uar.c | 40 uar->index = mthca_alloc(&dev->uar_table.alloc); in mthca_uar_alloc() 51 mthca_free(&dev->uar_table.alloc, uar->index); in mthca_uar_free() 58 ret = mthca_alloc_init(&dev->uar_table.alloc, in mthca_init_uar_table() 67 mthca_alloc_cleanup(&dev->uar_table.alloc); in mthca_init_uar_table() 77 mthca_alloc_cleanup(&dev->uar_table.alloc); in mthca_cleanup_uar_table()
|
| H A D | mthca_pd.c | 46 pd->pd_num = mthca_alloc(&dev->pd_table.alloc); in mthca_pd_alloc() 56 mthca_free(&dev->pd_table.alloc, pd->pd_num); in mthca_pd_alloc() 66 mthca_free(&dev->pd_table.alloc, pd->pd_num); in mthca_pd_free() 71 return mthca_alloc_init(&dev->pd_table.alloc, in mthca_init_pd_table() 80 mthca_alloc_cleanup(&dev->pd_table.alloc); in mthca_cleanup_pd_table()
|
| /linux-6.15/fs/ocfs2/ |
| H A D | localalloc.c | 38 struct ocfs2_dinode *alloc, 46 struct ocfs2_dinode *alloc, 307 la = OCFS2_LOCAL_ALLOC(alloc); in ocfs2_load_local_alloc() 484 struct ocfs2_dinode *alloc; in ocfs2_begin_local_alloc_recovery() local 620 struct ocfs2_dinode *alloc; in ocfs2_reserve_local_alloc_bits() local 724 struct ocfs2_dinode *alloc; in ocfs2_claim_local_alloc_bits() local 731 la = OCFS2_LOCAL_ALLOC(alloc); in ocfs2_claim_local_alloc_bits() 780 struct ocfs2_dinode *alloc; in ocfs2_free_local_alloc_bits() local 787 la = OCFS2_LOCAL_ALLOC(alloc); in ocfs2_free_local_alloc_bits() 911 alloc->id1.bitmap1.i_used = 0; in ocfs2_clear_local_alloc() [all …]
|
| /linux-6.15/fs/xfs/libxfs/ |
| H A D | xfs_alloc_btree.c | 139 key->alloc.ar_startblock = rec->alloc.ar_startblock; in xfs_allocbt_init_key_from_rec() 140 key->alloc.ar_blockcount = rec->alloc.ar_blockcount; in xfs_allocbt_init_key_from_rec() 153 key->alloc.ar_blockcount = 0; in xfs_bnobt_init_high_key_from_rec() 161 key->alloc.ar_blockcount = rec->alloc.ar_blockcount; in xfs_cntbt_init_high_key_from_rec() 162 key->alloc.ar_startblock = 0; in xfs_cntbt_init_high_key_from_rec() 226 be32_to_cpu(k2->alloc.ar_startblock); in xfs_bnobt_diff_two_keys() 239 mask->alloc.ar_startblock)); in xfs_cntbt_diff_two_keys() 242 be32_to_cpu(k2->alloc.ar_blockcount); in xfs_cntbt_diff_two_keys() 247 be32_to_cpu(k2->alloc.ar_startblock); in xfs_cntbt_diff_two_keys() 388 (k1->alloc.ar_blockcount == k2->alloc.ar_blockcount && in xfs_cntbt_keys_inorder() [all …]
|
| /linux-6.15/drivers/gpu/drm/ttm/ |
| H A D | ttm_pool.c | 221 unsigned int num_pages = alloc->pages - alloc->caching_divide; in ttm_pool_apply_caching() 226 switch (alloc->tt_caching) { in ttm_pool_apply_caching() 235 alloc->caching_divide = alloc->pages; in ttm_pool_apply_caching() 495 if (!alloc->dma_addr) in ttm_pool_allocated_page_commit() 570 alloc, nr); in ttm_pool_restore_commit() 572 alloc->caching_divide = alloc->pages; in ttm_pool_restore_commit() 621 if (alloc->dma_addr) { in ttm_pool_page_allocated() 634 alloc->caching_divide = alloc->pages; in ttm_pool_page_allocated() 682 alloc->pages = tt->pages; in ttm_pool_alloc_state_init() 729 alloc->remaining_pages; in __ttm_pool_alloc() [all …]
|
| /linux-6.15/lib/zstd/compress/ |
| H A D | zstd_cwksp.h | 282 assert(alloc >= bottom); in ZSTD_cwksp_reserve_internal_buffer_space() 283 if (alloc < bottom) { in ZSTD_cwksp_reserve_internal_buffer_space() 293 ws->allocStart = alloc; in ZSTD_cwksp_reserve_internal_buffer_space() 294 return alloc; in ZSTD_cwksp_reserve_internal_buffer_space() 347 void* alloc; in ZSTD_cwksp_reserve_internal() local 356 return alloc; in ZSTD_cwksp_reserve_internal() 414 void* alloc; in ZSTD_cwksp_reserve_table() local 425 alloc = ws->tableEnd; in ZSTD_cwksp_reserve_table() 426 end = (BYTE *)alloc + bytes; in ZSTD_cwksp_reserve_table() 444 return alloc; in ZSTD_cwksp_reserve_table() [all …]
|
| /linux-6.15/tools/perf/util/ |
| H A D | strbuf.c | 22 sb->alloc = sb->len = 0; in strbuf_init() 31 if (sb->alloc) { in strbuf_release() 39 char *res = sb->alloc ? sb->buf : NULL; in strbuf_detach() 51 if (nr < sb->alloc) in strbuf_grow() 57 if (alloc_nr(sb->alloc) > nr) in strbuf_grow() 58 nr = alloc_nr(sb->alloc); in strbuf_grow() 64 buf = realloc(sb->alloc ? sb->buf : NULL, nr * sizeof(*buf)); in strbuf_grow() 69 sb->alloc = nr; in strbuf_grow() 106 len = vsnprintf(sb->buf + sb->len, sb->alloc - sb->len, fmt, ap); in strbuf_addv() 142 size_t oldalloc = sb->alloc; in strbuf_read() [all …]
|
| H A D | strbuf.h | 51 size_t alloc; member 65 return sb->alloc ? sb->alloc - sb->len - 1 : 0; in strbuf_avail() 71 if (!sb->alloc) { in strbuf_setlen() 76 assert(len < sb->alloc); in strbuf_setlen()
|
| H A D | help-unknown-cmd.c | 37 if (nr > cmds->alloc) { in add_cmd_list() 39 if (alloc_nr(cmds->alloc) < nr) in add_cmd_list() 40 cmds->alloc = nr; in add_cmd_list() 42 cmds->alloc = alloc_nr(cmds->alloc); in add_cmd_list() 43 tmp = realloc(cmds->names, cmds->alloc * sizeof(*cmds->names)); in add_cmd_list()
|
| /linux-6.15/tools/lib/subcmd/ |
| H A D | subcmd-util.h | 38 #define ALLOC_GROW(x, nr, alloc) \ argument 40 if ((nr) > alloc) { \ 41 if (alloc_nr(alloc) < (nr)) \ 42 alloc = (nr); \ 44 alloc = alloc_nr(alloc); \ 45 x = xrealloc((x), alloc * sizeof(*(x))); \
|
| /linux-6.15/tools/testing/selftests/mm/ |
| H A D | droppable.c | 22 void *alloc; in main() local 28 alloc = mmap(0, alloc_size, PROT_READ | PROT_WRITE, MAP_ANONYMOUS | MAP_DROPPABLE, -1, 0); in main() 29 assert(alloc != MAP_FAILED); in main() 30 memset(alloc, 'A', alloc_size); in main() 32 assert(*(uint8_t *)(alloc + i)); in main() 43 if (!*(uint8_t *)(alloc + i)) { in main()
|
| /linux-6.15/arch/s390/mm/ |
| H A D | pgalloc.c | 255 unsigned long end, int alloc) in base_page_walk() argument 259 if (!alloc) in base_page_walk() 271 unsigned long end, int alloc) in base_segment_walk() argument 281 if (!alloc) in base_segment_walk() 292 if (!alloc) in base_segment_walk() 310 if (!alloc) in base_region3_walk() 321 if (!alloc) in base_region3_walk() 338 if (!alloc) in base_region2_walk() 349 if (!alloc) in base_region2_walk() 366 if (!alloc) in base_region1_walk() [all …]
|
| /linux-6.15/tools/testing/selftests/bpf/progs/ |
| H A D | verifier_kfunc_prog_types.c | 126 struct bpf_cpumask *alloc, *ref; in cpumask_kfunc_load_test() local 128 alloc = bpf_cpumask_create(); in cpumask_kfunc_load_test() 129 if (!alloc) in cpumask_kfunc_load_test() 132 ref = bpf_cpumask_acquire(alloc); in cpumask_kfunc_load_test() 133 bpf_cpumask_set_cpu(0, alloc); in cpumask_kfunc_load_test() 137 bpf_cpumask_release(alloc); in cpumask_kfunc_load_test()
|
| /linux-6.15/arch/xtensa/variants/dc233c/include/variant/ |
| H A D | tie-asm.h | 77 .macro xchal_ncp_store ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0 85 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0 97 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 115 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 141 .macro xchal_ncp_load ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0 149 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0 161 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 179 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
|
| /linux-6.15/arch/xtensa/variants/csp/include/variant/ |
| H A D | tie-asm.h | 76 .macro xchal_ncp_store ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0 84 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0 96 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 116 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 142 .macro xchal_ncp_load ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0 150 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0 162 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 182 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
|
| /linux-6.15/net/core/ |
| H A D | page_pool.c | 408 pool->alloc.cache[pool->alloc.count++] = netmem; in page_pool_refill_alloc_cache() 424 netmem = pool->alloc.cache[--pool->alloc.count]; in page_pool_refill_alloc_cache() 439 netmem = pool->alloc.cache[--pool->alloc.count]; in __page_pool_get_cached() 542 return pool->alloc.cache[--pool->alloc.count]; in __page_pool_alloc_pages_slow() 563 pool->alloc.cache[pool->alloc.count++] = netmem; in __page_pool_alloc_pages_slow() 572 netmem = pool->alloc.cache[--pool->alloc.count]; in __page_pool_alloc_pages_slow() 739 pool->alloc.cache[pool->alloc.count++] = netmem; in page_pool_recycle_in_cache() 1075 while (pool->alloc.count) { in page_pool_empty_alloc_cache_once() 1076 netmem = pool->alloc.cache[--pool->alloc.count]; in page_pool_empty_alloc_cache_once() 1194 while (pool->alloc.count) { in page_pool_update_nid() [all …]
|
| /linux-6.15/fs/nfs/ |
| H A D | nfs3acl.c | 260 struct posix_acl *orig = acl, *dfacl = NULL, *alloc; in nfs3_set_acl() local 267 alloc = get_inode_acl(inode, ACL_TYPE_DEFAULT); in nfs3_set_acl() 268 if (IS_ERR(alloc)) in nfs3_set_acl() 270 dfacl = alloc; in nfs3_set_acl() 274 alloc = get_inode_acl(inode, ACL_TYPE_ACCESS); in nfs3_set_acl() 275 if (IS_ERR(alloc)) in nfs3_set_acl() 278 acl = alloc; in nfs3_set_acl() 284 alloc = posix_acl_from_mode(inode->i_mode, GFP_KERNEL); in nfs3_set_acl() 285 if (IS_ERR(alloc)) in nfs3_set_acl() 287 acl = alloc; in nfs3_set_acl() [all …]
|
| /linux-6.15/mm/ |
| H A D | mempool.c | 93 if (pool->alloc == mempool_kmalloc) { in poison_element() 95 } else if (pool->alloc == mempool_alloc_slab) { in poison_element() 97 } else if (pool->alloc == mempool_alloc_pages) { in poison_element() 117 if (pool->alloc == mempool_alloc_slab || pool->alloc == mempool_kmalloc) in kasan_poison_element() 119 else if (pool->alloc == mempool_alloc_pages) in kasan_poison_element() 127 if (pool->alloc == mempool_kmalloc) in kasan_unpoison_element() 129 else if (pool->alloc == mempool_alloc_slab) in kasan_unpoison_element() 132 else if (pool->alloc == mempool_alloc_pages) in kasan_unpoison_element() 202 pool->alloc = alloc_fn; in mempool_init_node() 217 element = pool->alloc(gfp_mask, pool->pool_data); in mempool_init_node() [all …]
|