| /linux-6.15/include/linux/ |
| H A D | page_ref.h | 87 static inline int folio_ref_count(const struct folio *folio) in folio_ref_count() function 94 return folio_ref_count(page_folio(page)); in page_count()
|
| H A D | mm.h | 1531 ((unsigned int) folio_ref_count(folio) + 127u <= 127u) 2052 return ((unsigned int)folio_ref_count(folio)) >= in folio_maybe_dma_pinned()
|
| /linux-6.15/arch/xtensa/mm/ |
| H A D | tlb.c | 262 folio_ref_count(f), folio_mapcount(f)); in check_tlb_entry() 263 if (!folio_ref_count(f)) in check_tlb_entry()
|
| /linux-6.15/mm/ |
| H A D | migrate.c | 606 if (folio_ref_count(folio) != expected_count) in folio_migrate_mapping() 623 if (folio_ref_count(src) != expected_count) in migrate_huge_page_move_mapping() 755 if (folio_ref_count(src) != expected_count) in __migrate_folio() 841 if (folio_ref_count(src) != expected_count) in __buffer_migrate_folio() 1197 if (folio_ref_count(src) == 1) { in migrate_folio_unmap() 1452 if (folio_ref_count(src) == 1) { in unmap_and_move_huge_page()
|
| H A D | hugetlb_vmemmap.c | 461 VM_WARN_ON_ONCE_FOLIO(folio_ref_count(folio), folio); in __hugetlb_vmemmap_restore_folio() 573 VM_WARN_ON_ONCE_FOLIO(folio_ref_count(folio), folio); in __hugetlb_vmemmap_optimize_folio()
|
| H A D | memfd.c | 34 return folio_ref_count(folio) - folio_mapcount(folio) != in memfd_folio_has_extra_refs()
|
| H A D | debug.c | 79 folio_ref_count(folio), mapcount, mapping, in __dump_folio()
|
| H A D | workingset.c | 390 VM_BUG_ON_FOLIO(folio_ref_count(folio), folio); in workingset_eviction()
|
| H A D | huge_memory.c | 1983 if (folio_ref_count(folio) > in do_huge_pmd_wp_page() 1988 if (folio_ref_count(folio) == 1) { in do_huge_pmd_wp_page() 2969 VM_WARN_ON_FOLIO(!folio_ref_count(folio), folio); in __split_huge_pmd_locked() 3203 ref_count = folio_ref_count(folio); in __discard_anon_folio_pmd_locked() 3305 return folio_mapcount(folio) == folio_ref_count(folio) - extra_pins - in can_split_folio() 4055 WARN_ON_ONCE(folio_ref_count(folio)); in __folio_unqueue_deferred_split()
|
| H A D | memory.c | 2113 if (!folio_ref_count(folio)) in validate_page_before_insert() 3749 if (folio_large_mapcount(folio) != folio_ref_count(folio)) in __wp_can_reuse_large_anon_folio() 3754 VM_WARN_ON_ONCE_FOLIO(folio_large_mapcount(folio) > folio_ref_count(folio), folio); in __wp_can_reuse_large_anon_folio() 3758 if (folio_large_mapcount(folio) != folio_ref_count(folio)) in __wp_can_reuse_large_anon_folio() 3797 if (folio_test_ksm(folio) || folio_ref_count(folio) > 3) in wp_can_reuse_anon_folio() 3805 if (folio_ref_count(folio) > 1 + folio_test_swapcache(folio)) in wp_can_reuse_anon_folio() 3811 if (folio_test_ksm(folio) || folio_ref_count(folio) != 1) { in wp_can_reuse_anon_folio() 4132 folio_ref_count(folio) == (1 + folio_nr_pages(folio)); in should_try_to_free_swap() 4752 (exclusive || folio_ref_count(folio) == 1)) { in do_swap_page()
|
| H A D | hugetlb.c | 1286 VM_BUG_ON_FOLIO(folio_ref_count(folio), folio); in enqueue_hugetlb_folio() 1837 VM_BUG_ON_FOLIO(folio_ref_count(folio), folio); in free_huge_folio() 2150 if (!folio_ref_count(folio)) { in dissolve_free_hugetlb_folio() 2834 } else if (folio_ref_count(old_folio)) { in alloc_and_dissolve_hugetlb_folio() 2927 if (folio_ref_count(folio) && folio_isolate_hugetlb(folio, list)) in isolate_or_dissolve_huge_page() 2929 else if (!folio_ref_count(folio)) in isolate_or_dissolve_huge_page() 2967 if (!folio_ref_count(folio)) { in replace_free_hugepage_folios() 3467 WARN_ON(folio_ref_count(folio) != 1); in gather_bootmem_prealloc_node()
|
| H A D | truncate.c | 308 if (folio_ref_count(folio) > in mapping_evict_folio()
|
| H A D | migrate_device.c | 366 if ((folio_ref_count(folio) - extra) > folio_mapcount(folio)) in migrate_vma_check_page()
|
| H A D | gup.c | 84 if (WARN_ON_ONCE(folio_ref_count(folio) < 0)) in try_get_folio() 147 if (WARN_ON_ONCE(folio_ref_count(folio) <= 0)) in try_grab_folio() 231 WARN_ON_ONCE(folio_ref_count(folio) < GUP_PIN_COUNTING_BIAS); in folio_add_pin()
|
| H A D | khugepaged.c | 561 return folio_ref_count(folio) == expected_refcount; in is_refcount_suitable() 2034 if (folio_ref_count(folio) != 2 + folio_nr_pages(folio)) { in collapse_file()
|
| H A D | memcontrol-v1.c | 597 VM_BUG_ON_FOLIO(folio_ref_count(folio), folio); in memcg1_swapout()
|
| H A D | vmscan.c | 489 return folio_ref_count(folio) - folio_test_private(folio) == in is_page_cache_freeable() 1504 if (!mapping && folio_ref_count(folio) == 1) { in shrink_folio_list() 1842 VM_BUG_ON_FOLIO(!folio_ref_count(folio), folio); in folio_isolate_lru()
|
| H A D | memory-failure.c | 2595 if (folio_ref_count(folio) > 1) { in unpoison_memory()
|
| H A D | rmap.c | 2110 ref_count = folio_ref_count(folio); in try_to_unmap_one()
|
| H A D | compaction.c | 1120 if (!mapping && (folio_ref_count(folio) - 1) > folio_mapcount(folio)) in isolate_migratepages_block()
|
| H A D | ksm.c | 1305 if (folio_mapcount(folio) + 1 + swapped != folio_ref_count(folio)) { in write_protect_page()
|
| /linux-6.15/fs/nilfs2/ |
| H A D | page.c | 169 folio, folio_ref_count(folio), in nilfs_folio_bug()
|
| /linux-6.15/fs/ |
| H A D | dax.c | 416 WARN_ON_ONCE(folio_ref_count(new_folio)); in dax_folio_put() 438 WARN_ON_ONCE(folio_ref_count(folio)); in dax_folio_init() 493 if (folio_ref_count(folio) - folio_mapcount(folio)) in dax_busy_page()
|
| H A D | aio.c | 343 folio_ref_count(folio)); in aio_free_ring() 531 folio_ref_count(folio)); in aio_setup_ring()
|
| /linux-6.15/fs/btrfs/ |
| H A D | compression.c | 260 ASSERT(folio_ref_count(folio) == 1); in btrfs_free_compr_folio()
|