Home
last modified time | relevance | path

Searched refs:folio_order (Results 1 – 25 of 40) sorted by relevance

12

/linux-6.15/include/linux/
H A Dfolio_queue.h238 folioq->orders[slot] = folio_order(folio); in folioq_append()
260 folioq->orders[slot] = folio_order(folio); in folioq_append_mark()
H A Dhuge_mm.h474 return folio_order(folio) >= HPAGE_PMD_ORDER; in folio_test_pmd_mappable()
H A Dmm.h1214 static inline unsigned int folio_order(const struct folio *folio) in folio_order() function
2011 return folio_order(folio) > 1; in folio_has_pincount()
2229 return PAGE_SHIFT + folio_order(folio); in folio_shift()
2242 return PAGE_SIZE << folio_order(folio); in folio_size()
/linux-6.15/mm/
H A Dpage_io.c279 count_mthp_stat(folio_order(folio), MTHP_STAT_ZSWPOUT); in swap_writepage()
300 count_mthp_stat(folio_order(folio), MTHP_STAT_SWPOUT); in count_swpout_vm_event()
497 count_mthp_stat(folio_order(folio), MTHP_STAT_SWPIN); in sio_read_complete()
592 count_mthp_stat(folio_order(folio), MTHP_STAT_SWPIN); in swap_read_folio_bdev_sync()
609 count_mthp_stat(folio_order(folio), MTHP_STAT_SWPIN); in swap_read_folio_bdev_async()
H A Dhuge_memory.c3474 int order = folio_order(folio); in __split_unmapped_folio()
3507 int old_order = folio_order(folio); in __split_unmapped_folio()
3564 mod_mthp_stat(folio_order(release), in __split_unmapped_folio()
3736 int order = folio_order(folio); in __folio_split()
3747 if (new_order >= folio_order(folio)) in __folio_split()
3867 if (folio_order(folio) > 1 && in __folio_split()
3872 mod_mthp_stat(folio_order(folio), in __folio_split()
4064 mod_mthp_stat(folio_order(folio), in __folio_unqueue_deferred_split()
4088 if (folio_order(folio) <= 1) in deferred_split_folio()
4201 mod_mthp_stat(folio_order(folio), in deferred_split_scan()
[all …]
H A Dmigrate.c493 mod_mthp_stat(folio_order(folio), MTHP_STAT_NR_ANON, 1); in __folio_migrate_mapping()
519 mod_mthp_stat(folio_order(folio), MTHP_STAT_NR_ANON, 1); in __folio_migrate_mapping()
2181 order = folio_order(src); in alloc_migration_target()
2627 int order = folio_order(src); in alloc_misplaced_dst_folio()
2690 folio_order(folio), ZONE_MOVABLE); in migrate_misplaced_folio_prepare()
H A Ddebug.c88 folio_order(folio), in __dump_folio()
H A Dslab.h221 return folio_order(slab_folio(slab)); in slab_order()
H A Dreadahead.c632 unsigned int order = folio_order(folio); in page_cache_async_ra()
H A Dshmem.c886 XA_STATE_ORDER(xas, &mapping->i_pages, index, folio_order(folio)); in shmem_add_to_page_cache()
1913 count_mthp_stat(folio_order(folio), MTHP_STAT_SHMEM_FALLBACK); in shmem_alloc_and_add_folio()
1914 count_mthp_stat(folio_order(folio), MTHP_STAT_SHMEM_FALLBACK_CHARGE); in shmem_alloc_and_add_folio()
2062 new = shmem_alloc_folio(gfp, folio_order(old), info, index); in shmem_replace_folio()
2329 } else if (order != folio_order(folio)) { in shmem_swapin_folio()
2360 xa_get_order(&mapping->i_pages, index) != folio_order(folio)) { in shmem_swapin_folio()
2526 count_mthp_stat(folio_order(folio), MTHP_STAT_SHMEM_ALLOC); in shmem_get_folio_gfp()
H A Dswap.c112 free_frozen_pages(&folio->page, folio_order(folio)); in __folio_put()
H A Dkhugepaged.c1546 if (folio_order(folio) != HPAGE_PMD_ORDER) { in collapse_pte_mapped_thp()
1981 if (folio_order(folio) == HPAGE_PMD_ORDER && in collapse_file()
2298 if (folio_order(folio) == HPAGE_PMD_ORDER && in hpage_collapse_scan_file()
H A Dcompaction.c1221 if (unlikely(skip_isolation_on_order(folio_order(folio), in isolate_migratepages_block()
1837 int order = folio_order(src); in compaction_alloc_noprof()
1894 int order = folio_order(dst); in compaction_free()
H A Dfilemap.c136 xas_set_order(&xas, folio->index, folio_order(folio)); in page_cache_delete()
860 XA_STATE_ORDER(xas, &mapping->i_pages, index, folio_order(folio)); in __filemap_add_folio()
863 unsigned int forder = folio_order(folio); in __filemap_add_folio()
867 VM_BUG_ON_FOLIO(folio_order(folio) < mapping_min_folio_order(mapping), in __filemap_add_folio()
H A Dswap_state.c94 XA_STATE_ORDER(xas, &address_space->i_pages, idx, folio_order(folio)); in add_to_swap_cache()
H A Dmempolicy.c1255 order = folio_order(src); in alloc_migration_target_by_mpol()
1412 order = folio_order(folio); in do_mbind()
2809 pol = get_vma_policy(vma, addr, folio_order(folio), &ilx); in mpol_misplaced()
/linux-6.15/virt/kvm/
H A Dguest_memfd.c34 int rc = kvm_arch_gmem_prepare(kvm, gfn, pfn, folio_order(folio)); in __kvm_gmem_prepare_folio()
80 WARN_ON(!IS_ALIGNED(slot->gmem.pgoff, 1 << folio_order(folio))); in kvm_gmem_prepare_folio()
82 index = ALIGN_DOWN(index, 1 << folio_order(folio)); in kvm_gmem_prepare_folio()
370 int order = folio_order(folio); in kvm_gmem_free_folio()
/linux-6.15/fs/bcachefs/
H A Dfs-io-pagecache.h26 return PAGE_SECTORS << folio_order(folio); in folio_sectors()
/linux-6.15/fs/netfs/
H A Drolling_buffer.c137 unsigned int order = folio_order(folio); in rolling_buffer_load_from_ra()
/linux-6.15/include/trace/events/
H A Dfilemap.h38 __entry->order = folio_order(folio);
/linux-6.15/fs/btrfs/
H A Dextent_io.h303 if (folio_order(eb->folios[0])) in num_extent_folios()
H A Dsubpage.c185 ASSERT(folio_order(folio) == 0); in btrfs_subpage_assert()
H A Dbio.c172 ASSERT(folio_order(page_folio(bv->bv_page)) == 0); in btrfs_end_repair_bio()
/linux-6.15/fs/
H A Ddax.c396 order = folio_order(folio); in dax_folio_put()
432 WARN_ON_ONCE(folio_order(folio)); in dax_folio_init()
461 WARN_ON_ONCE(dax_entry_order(entry) != folio_order(folio)); in dax_associate_entry()
/linux-6.15/fs/nfs/
H A Dinternal.h860 pgoff_t index = folio->index >> folio_order(folio); in nfs_folio_length()

12