| /linux-6.15/net/core/ |
| H A D | page_pool.c | 203 memcpy(&pool->p, ¶ms->fast, sizeof(pool->p)); in page_pool_init() 291 if (!pool->dma_map || !pool->dma_sync) in page_pool_init() 299 err = pool->mp_ops->init(pool); in page_pool_init() 408 pool->alloc.cache[pool->alloc.count++] = netmem; in page_pool_refill_alloc_cache() 424 netmem = pool->alloc.cache[--pool->alloc.count]; in page_pool_refill_alloc_cache() 439 netmem = pool->alloc.cache[--pool->alloc.count]; in __page_pool_get_cached() 542 return pool->alloc.cache[--pool->alloc.count]; in __page_pool_alloc_pages_slow() 563 pool->alloc.cache[pool->alloc.count++] = netmem; in __page_pool_alloc_pages_slow() 739 pool->alloc.cache[pool->alloc.count++] = netmem; in page_pool_recycle_in_cache() 1051 pool->disconnect(pool); in __page_pool_destroy() [all …]
|
| /linux-6.15/mm/ |
| H A D | mempool.c | 139 BUG_ON(pool->curr_nr >= pool->min_nr); in add_element() 142 pool->elements[pool->curr_nr++] = element; in add_element() 147 void *element = pool->elements[--pool->curr_nr]; in remove_element() 170 pool->free(element, pool->pool_data); in mempool_exit() 214 while (pool->curr_nr < pool->min_nr) { in mempool_init_node() 217 element = pool->alloc(gfp_mask, pool->pool_data); in mempool_init_node() 322 pool->free(element, pool->pool_data); in mempool_resize() 349 while (pool->curr_nr < pool->min_nr) { in mempool_resize() 355 if (pool->curr_nr < pool->min_nr) { in mempool_resize() 540 if (likely(pool->curr_nr < pool->min_nr)) { in mempool_free() [all …]
|
| H A D | dmapool.c | 83 pool->name, pool->nr_active, in pools_show() 84 pool->nr_blocks, pool->size, in pools_show() 187 pool->nr_active++; in pool_block_pop() 306 while (offset + pool->size <= pool->allocation) { in pool_initialise_page() 331 pool->nr_pages++; in pool_initialise_page() 342 page->vaddr = dma_alloc_coherent(pool->dev, pool->allocation, in pool_alloc_page() 378 dev_err(pool->dev, "%s %s busy\n", __func__, pool->name); in dma_pool_destroy() 384 dma_free_coherent(pool->dev, pool->allocation, in dma_pool_destroy() 390 kfree(pool); in dma_pool_destroy() 503 if (pool) in dmam_pool_create() [all …]
|
| /linux-6.15/net/xdp/ |
| H A D | xsk_buff_pool.c | 39 if (!pool) in xp_destroy() 44 kvfree(pool); in xp_destroy() 67 if (!pool) in xp_create_and_assign_umem() 105 xskb->pool = pool; in xp_create_and_assign_umem() 114 return pool; in xp_create_and_assign_umem() 222 bpf.xsk.pool = pool; in xp_assign_dev() 258 if (!pool->fq || !pool->cq) in xp_assign_dev_shared() 274 xsk_clear_pool_at_qid(pool->netdev, pool->queue_id); in xp_clear_dev() 309 if (!pool) in xp_put_pool() 497 *addr + pool->chunk_size > pool->addrs_cnt || in xp_check_unaligned() [all …]
|
| /linux-6.15/drivers/net/ethernet/ti/ |
| H A D | k3-cppi-desc-pool.c | 30 if (!pool) in k3_cppi_desc_pool_destroy() 38 dma_free_coherent(pool->dev, pool->mem_size, pool->cpumem, in k3_cppi_desc_pool_destroy() 45 kfree(pool); in k3_cppi_desc_pool_destroy() 58 pool = kzalloc(sizeof(*pool), GFP_KERNEL); in k3_cppi_desc_pool_create_name() 59 if (!pool) in k3_cppi_desc_pool_create_name() 65 pool->mem_size = pool->num_desc * pool->desc_size; in k3_cppi_desc_pool_create_name() 80 pool->desc_infos = kcalloc(pool->num_desc, in k3_cppi_desc_pool_create_name() 85 pool->cpumem = dma_alloc_coherent(pool->dev, pool->mem_size, in k3_cppi_desc_pool_create_name() 92 (phys_addr_t)pool->dma_addr, pool->mem_size, in k3_cppi_desc_pool_create_name() 99 return pool; in k3_cppi_desc_pool_create_name() [all …]
|
| /linux-6.15/drivers/net/ethernet/mellanox/mlx5/core/steering/hws/ |
| H A D | pool.c | 69 resource->pool = pool; in hws_pool_create_one_resource() 435 (void)pool; in hws_pool_general_element_db_uninit() 550 ret = hws_pool_buddy_db_init(pool, pool->alloc_log_sz); in hws_pool_db_init() 562 pool->p_db_uninit(pool); in hws_pool_db_unint() 571 ret = pool->p_get_chunk(pool, chunk); in mlx5hws_pool_chunk_alloc() 581 pool->p_put_chunk(pool, chunk); in mlx5hws_pool_chunk_free() 591 pool = kzalloc(sizeof(*pool), GFP_KERNEL); in mlx5hws_pool_create() 592 if (!pool) in mlx5hws_pool_create() 619 return pool; in mlx5hws_pool_create() 622 kfree(pool); in mlx5hws_pool_create() [all …]
|
| /linux-6.15/drivers/md/ |
| H A D | dm-thin.c | 621 struct pool *pool = tc->pool; in requeue_deferred_cells() local 674 struct pool *pool = tc->pool; in get_bio_block() local 691 struct pool *pool = tc->pool; in get_bio_block_range() local 716 struct pool *pool = tc->pool; in remap() local 754 struct pool *pool = tc->pool; in issue() local 882 struct pool *pool = tc->pool; in cell_defer_no_holder() local 962 struct pool *pool = tc->pool; in complete_overwrite_bio() local 995 struct pool *pool = tc->pool; in process_prepared_mapping() local 1087 struct pool *pool = tc->pool; in passdown_double_checking_shared_status() local 1151 struct pool *pool = tc->pool; in process_prepared_discard_passdown_pt1() local [all …]
|
| /linux-6.15/net/ceph/ |
| H A D | msgpool.c | 17 msg = ceph_msg_new2(pool->type, pool->front_len, pool->max_data_items, in msgpool_alloc() 23 msg->pool = pool; in msgpool_alloc() 34 msg->pool = NULL; in msgpool_free() 43 pool->type = type; in ceph_msgpool_init() 46 pool->pool = mempool_create(size, msgpool_alloc, msgpool_free, pool); in ceph_msgpool_init() 47 if (!pool->pool) in ceph_msgpool_init() 49 pool->name = name; in ceph_msgpool_init() 56 mempool_destroy(pool->pool); in ceph_msgpool_destroy() 68 pool->front_len, pool->max_data_items); in ceph_msgpool_get() 76 msg = mempool_alloc(pool->pool, GFP_NOFS); in ceph_msgpool_get() [all …]
|
| /linux-6.15/sound/core/seq/ |
| H A D | seq_memory.c | 24 return pool->total_elements - atomic_read(&pool->counter); in snd_seq_pool_available() 29 return snd_seq_pool_available(pool) >= pool->room; in snd_seq_output_ok() 239 pool = cell->pool; in snd_seq_cell_free() 289 while (pool->free == NULL && ! nonblock && ! pool->closing) { in snd_seq_cell_alloc() 462 cellptr->pool = pool; in snd_seq_pool_init() 466 pool->room = (pool->size + 1) / 2; in snd_seq_pool_init() 470 pool->total_elements = pool->size; in snd_seq_pool_init() 521 pool = kzalloc(sizeof(*pool), GFP_KERNEL); in snd_seq_pool_new() 522 if (!pool) in snd_seq_pool_new() 536 return pool; in snd_seq_pool_new() [all …]
|
| /linux-6.15/lib/ |
| H A D | objpool.c | 40 pool->nr_objs++; in objpool_init_percpu_slot() 109 if (!pool->cpu_slots) in objpool_fini_percpu_slots() 146 pool->cpu_slots = kzalloc(slot_size, pool->gfp); in objpool_init() 147 if (!pool->cpu_slots) in objpool_init() 155 refcount_set(&pool->ref, pool->nr_objs + 1); in objpool_init() 164 if (!pool->cpu_slots) in objpool_free() 171 if (pool->release) in objpool_free() 172 pool->release(pool, pool->context); in objpool_free() 179 if (!obj || !pool) in objpool_drop() 183 objpool_free(pool); in objpool_drop() [all …]
|
| /linux-6.15/drivers/gpu/drm/panthor/ |
| H A D | panthor_heap.c | 496 kfree(pool); in panthor_heap_pool_release() 505 if (pool) in panthor_heap_pool_put() 518 if (pool) in panthor_heap_pool_get() 521 return pool; in panthor_heap_pool_get() 542 pool = kzalloc(sizeof(*pool), GFP_KERNEL); in panthor_heap_pool_create() 543 if (!pool) in panthor_heap_pool_create() 568 atomic_add(pool->gpu_contexts->obj->size, &pool->size); in panthor_heap_pool_create() 570 return pool; in panthor_heap_pool_create() 597 if (!pool) in panthor_heap_pool_destroy() 605 atomic_sub(pool->gpu_contexts->obj->size, &pool->size); in panthor_heap_pool_destroy() [all …]
|
| /linux-6.15/include/net/ |
| H A D | xdp_sock_drv.h | 42 return pool->chunk_size; in xsk_pool_get_chunk_size() 47 return xsk_pool_get_chunk_size(pool) - xsk_pool_get_headroom(pool); in xsk_pool_get_rx_frame_size() 53 xp_set_rxq_info(pool, rxq); in xsk_pool_set_rxq_info() 59 xp_fill_cb(pool, desc); in xsk_pool_fill_cb() 65 xp_dma_unmap(pool, attrs); in xsk_pool_dma_unmap() 71 struct xdp_umem *umem = pool->umem; in xsk_pool_dma_map() 92 return xp_alloc(pool); in xsk_buff_alloc() 108 return xp_can_alloc(pool, count); in xsk_buff_can_alloc() 191 return xp_raw_get_dma(pool, addr); in xsk_buff_raw_get_dma() 233 if (!pool->tx_metadata_len) in __xsk_buff_get_metadata() [all …]
|
| H A D | xsk_buff_pool.h | 30 struct xsk_buff_pool *pool; member 124 xskb->xdp.data_hard_start = pool->addrs + addr + pool->headroom; in xp_init_xskb_addr() 167 xskb->pool->frame_len, in xp_dma_sync_for_cpu() 192 return pool->dma_pages && in xp_desc_crosses_non_contig_pg() 203 return addr & pool->chunk_mask; in xp_aligned_extract_addr() 224 return xp_aligned_extract_addr(pool, addr) >> pool->chunk_shift; in xp_aligned_extract_idx() 229 if (xskb->pool->unaligned) in xp_release() 230 xskb->pool->free_heads[xskb->pool->free_heads_cnt++] = xskb; in xp_release() 234 struct xsk_buff_pool *pool) in xp_get_handle() argument 239 if (!pool->unaligned) in xp_get_handle() [all …]
|
| /linux-6.15/kernel/cgroup/ |
| H A D | dmem.c | 109 kfree(pool); in free_cg_pool() 132 return pool ? READ_ONCE(pool->cnt.low) : 0; in get_resource_low() 137 return pool ? READ_ONCE(pool->cnt.min) : 0; in get_resource_min() 147 return pool ? page_counter_read(&pool->cnt) : 0; in get_resource_current() 283 for (pool = test_pool; pool && limit_pool != pool; pool = pool_parent(pool)) in dmem_cgroup_state_evict_valuable() 286 if (!pool) in dmem_cgroup_state_evict_valuable() 327 pool = kzalloc(sizeof(*pool), GFP_NOWAIT); in alloc_pool_single() 328 if (!pool) in alloc_pool_single() 520 if (pool) in dmem_cgroup_pool_state_put() 533 if (pool && !READ_ONCE(pool->inited)) in get_cg_pool_unlocked() [all …]
|
| /linux-6.15/drivers/staging/media/atomisp/pci/runtime/rmgr/src/ |
| H A D | rmgr_vbuf.c | 125 assert(pool); in ia_css_rmgr_init_vbuf() 126 if (!pool) in ia_css_rmgr_init_vbuf() 129 if (pool->recycle && pool->size) { in ia_css_rmgr_init_vbuf() 133 pool->size; in ia_css_rmgr_init_vbuf() 135 if (pool->handles) in ia_css_rmgr_init_vbuf() 141 pool->size = 0; in ia_css_rmgr_init_vbuf() 142 pool->handles = NULL; in ia_css_rmgr_init_vbuf() 157 if (!pool) { in ia_css_rmgr_uninit_vbuf() 161 if (pool->handles) { in ia_css_rmgr_uninit_vbuf() 194 assert(pool); in rmgr_push_handle() [all …]
|
| /linux-6.15/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/ |
| H A D | dr_icm_pool.c | 280 buddy->pool = pool; in dr_icm_buddy_create() 291 pool->dmn->num_buddies[pool->icm_type]++; in dr_icm_buddy_create() 343 return pool->hot_memory_size > pool->th; in dr_icm_pool_is_sync_required() 470 struct mlx5dr_icm_pool *pool = buddy->pool; in mlx5dr_icm_free_chunk() local 481 hot_chunk = &pool->hot_chunks_arr[pool->hot_chunks_num++]; in mlx5dr_icm_free_chunk() 512 pool = kvzalloc(sizeof(*pool), GFP_KERNEL); in mlx5dr_icm_pool_create() 513 if (!pool) in mlx5dr_icm_pool_create() 516 pool->dmn = dmn; in mlx5dr_icm_pool_create() 557 return pool; in mlx5dr_icm_pool_create() 560 kvfree(pool); in mlx5dr_icm_pool_create() [all …]
|
| H A D | dr_arg.c | 62 pool->dmn->pdn, in dr_arg_pool_alloc_objs() 102 mutex_lock(&pool->mutex); in dr_arg_pool_get_arg_obj() 125 mutex_lock(&pool->mutex); in dr_arg_pool_put_arg_obj() 133 struct dr_arg_pool *pool; in dr_arg_pool_create() local 135 pool = kzalloc(sizeof(*pool), GFP_KERNEL); in dr_arg_pool_create() 136 if (!pool) in dr_arg_pool_create() 139 pool->dmn = dmn; in dr_arg_pool_create() 142 mutex_init(&pool->mutex); in dr_arg_pool_create() 148 return pool; in dr_arg_pool_create() 151 kfree(pool); in dr_arg_pool_create() [all …]
|
| /linux-6.15/drivers/gpu/drm/amd/display/dc/resource/dce80/ |
| H A D | dce80_resource.c | 933 *pool = NULL; in dce80_destroy_resource_pool() 1044 if (!pool->base.irqs) in dce80_construct() 1138 if (!pool) in dce80_create_resource_pool() 1142 return &pool->base; in dce80_create_resource_pool() 1144 kfree(pool); in dce80_create_resource_pool() 1338 if (!pool) in dce81_create_resource_pool() 1342 return &pool->base; in dce81_create_resource_pool() 1344 kfree(pool); in dce81_create_resource_pool() 1535 if (!pool) in dce83_create_resource_pool() 1539 return &pool->base; in dce83_create_resource_pool() [all …]
|
| /linux-6.15/drivers/net/ethernet/mellanox/mlx5/core/ |
| H A D | irq_affinity.c | 10 pool->irqs_per_cpu[cpu]--; in cpu_put() 15 pool->irqs_per_cpu[cpu]++; in cpu_get() 27 if (!pool->irqs_per_cpu[cpu]) { in cpu_get_least_loaded() 33 if (pool->irqs_per_cpu[cpu] < pool->irqs_per_cpu[best_cpu]) in cpu_get_least_loaded() 42 pool->irqs_per_cpu[best_cpu]++; in cpu_get_least_loaded() 55 err = xa_alloc(&pool->irqs, &irq_index, NULL, pool->xa_num_irqs, GFP_KERNEL); in irq_pool_request_irq() 58 if (pool->irqs_per_cpu) { in irq_pool_request_irq() 132 mutex_lock(&pool->lock); in mlx5_irq_affinity_request() 162 mutex_unlock(&pool->lock); in mlx5_irq_affinity_request() 188 if (pool->irqs_per_cpu) in mlx5_irq_affinity_irq_release() [all …]
|
| /linux-6.15/drivers/gpu/drm/amd/display/dc/dce60/ |
| H A D | dce60_resource.c | 927 *pool = NULL; in dce60_destroy_resource_pool() 1031 if (!pool->base.irqs) in dce60_construct() 1125 if (!pool) in dce60_create_resource_pool() 1129 return &pool->base; in dce60_create_resource_pool() 1131 kfree(pool); in dce60_create_resource_pool() 1323 if (!pool) in dce61_create_resource_pool() 1327 return &pool->base; in dce61_create_resource_pool() 1329 kfree(pool); in dce61_create_resource_pool() 1517 if (!pool) in dce64_create_resource_pool() 1521 return &pool->base; in dce64_create_resource_pool() [all …]
|
| /linux-6.15/include/net/page_pool/ |
| H A D | helpers.h | 96 return page_pool_alloc_pages(pool, gfp); in page_pool_dev_alloc_pages() 128 return page_pool_alloc_netmems(pool, gfp); in page_pool_alloc_netmem() 139 if (pool->frag_offset + *size > max_size) { in page_pool_alloc_netmem() 141 pool->frag_offset = max_size; in page_pool_alloc_netmem() 214 return page_pool_alloc_va(pool, size, gfp); in page_pool_dev_alloc_va() 227 return pool->p.dma_dir; in page_pool_get_dma_dir() 448 page_pool_get_dma_dir(pool)); in __page_pool_dma_sync_for_cpu() 476 if (!pool->dma_sync_for_cpu) in page_pool_dma_sync_netmem_for_cpu() 479 __page_pool_dma_sync_for_cpu(pool, in page_pool_dma_sync_netmem_for_cpu() 491 if (unlikely(pool->p.nid != new_nid)) in page_pool_nid_changed() [all …]
|
| /linux-6.15/drivers/net/ethernet/mellanox/mlx5/core/lib/ |
| H A D | crypto.c | 19 #define MLX5_CRYPTO_DEK_POOL_CALC_FREED(pool) MLX5_CRYPTO_DEK_CALC_FREED(pool) argument 412 pool->avail_deks--; in mlx5_crypto_dek_pool_pop() 413 pool->in_use_deks++; in mlx5_crypto_dek_pool_pop() 460 if (pool->syncing) in mlx5_crypto_dek_pool_push() 595 err = mlx5_crypto_cmd_sync_crypto(pool->mdev, BIT(pool->key_purpose)); in mlx5_crypto_dek_sync_work_fn() 676 mlx5_crypto_dek_pool_splice_destroy_list(pool, &pool->destroy_list, in mlx5_crypto_dek_destroy_work_fn() 686 pool = kzalloc(sizeof(*pool), GFP_KERNEL); in mlx5_crypto_dek_pool_create() 687 if (!pool) in mlx5_crypto_dek_pool_create() 690 pool->mdev = mdev; in mlx5_crypto_dek_pool_create() 704 return pool; in mlx5_crypto_dek_pool_create() [all …]
|
| /linux-6.15/net/rds/ |
| H A D | ib_rdma.c | 271 struct rds_ib_mr_pool *pool = ibmr->pool; in rds_ib_teardown_mr() local 419 &pool->clean_list); in rds_ib_flush_mr_pool() 446 if (atomic_inc_return(&pool->item_count) <= pool->max_items) in rds_ib_try_reuse_ibmr() 483 struct rds_ib_mr_pool *pool = ibmr->pool; in rds_ib_free_mr() local 505 if (atomic_read(&pool->free_pinned) >= pool->max_free_pinned || in rds_ib_free_mr() 506 atomic_read(&pool->dirty_count) >= pool->max_items / 5) in rds_ib_free_mr() 636 kfree(pool); in rds_ib_destroy_mr_pool() 644 pool = kzalloc(sizeof(*pool), GFP_KERNEL); in rds_ib_create_mr_pool() 645 if (!pool) in rds_ib_create_mr_pool() 667 pool->max_free_pinned = pool->max_items * pool->max_pages / 4; in rds_ib_create_mr_pool() [all …]
|
| /linux-6.15/arch/arm64/kvm/hyp/nvhe/ |
| H A D | page_alloc.c | 45 if (addr < pool->range_start || addr >= pool->range_end) in __find_buddy_nocheck() 103 if (phys < pool->range_start || phys >= pool->range_end) in __hyp_attach_page() 156 __hyp_attach_page(pool, p); in __hyp_put_page() 170 hyp_spin_lock(&pool->lock); in hyp_put_page() 171 __hyp_put_page(pool, p); in hyp_put_page() 172 hyp_spin_unlock(&pool->lock); in hyp_put_page() 179 hyp_spin_lock(&pool->lock); in hyp_get_page() 203 hyp_spin_lock(&pool->lock); in hyp_alloc_pages() 206 while (i <= pool->max_order && list_empty(&pool->free_area[i])) in hyp_alloc_pages() 208 if (i > pool->max_order) { in hyp_alloc_pages() [all …]
|
| /linux-6.15/drivers/net/ethernet/mellanox/mlxsw/ |
| H A D | spectrum_cnt.c | 127 pool = kzalloc(struct_size(pool, sub_pools, sub_pools_count), in mlxsw_sp_counter_pool_init() 129 if (!pool) in mlxsw_sp_counter_pool_init() 134 flex_array_size(pool, sub_pools, pool->sub_pools_count)); in mlxsw_sp_counter_pool_init() 139 &pool->pool_size); in mlxsw_sp_counter_pool_init() 145 pool->usage = bitmap_zalloc(pool->pool_size, GFP_KERNEL); in mlxsw_sp_counter_pool_init() 146 if (!pool->usage) { in mlxsw_sp_counter_pool_init() 158 bitmap_free(pool->usage); in mlxsw_sp_counter_pool_init() 163 kfree(pool); in mlxsw_sp_counter_pool_init() 173 WARN_ON(find_first_bit(pool->usage, pool->pool_size) != in mlxsw_sp_counter_pool_fini() 176 bitmap_free(pool->usage); in mlxsw_sp_counter_pool_fini() [all …]
|