Home
last modified time | relevance | path

Searched refs:folio_ref_count (Results 1 – 25 of 27) sorted by relevance

12

/linux-6.15/include/linux/
H A Dpage_ref.h87 static inline int folio_ref_count(const struct folio *folio) in folio_ref_count() function
94 return folio_ref_count(page_folio(page)); in page_count()
H A Dmm.h1531 ((unsigned int) folio_ref_count(folio) + 127u <= 127u)
2052 return ((unsigned int)folio_ref_count(folio)) >= in folio_maybe_dma_pinned()
/linux-6.15/arch/xtensa/mm/
H A Dtlb.c262 folio_ref_count(f), folio_mapcount(f)); in check_tlb_entry()
263 if (!folio_ref_count(f)) in check_tlb_entry()
/linux-6.15/mm/
H A Dmigrate.c606 if (folio_ref_count(folio) != expected_count) in folio_migrate_mapping()
623 if (folio_ref_count(src) != expected_count) in migrate_huge_page_move_mapping()
755 if (folio_ref_count(src) != expected_count) in __migrate_folio()
841 if (folio_ref_count(src) != expected_count) in __buffer_migrate_folio()
1197 if (folio_ref_count(src) == 1) { in migrate_folio_unmap()
1452 if (folio_ref_count(src) == 1) { in unmap_and_move_huge_page()
H A Dhugetlb_vmemmap.c461 VM_WARN_ON_ONCE_FOLIO(folio_ref_count(folio), folio); in __hugetlb_vmemmap_restore_folio()
573 VM_WARN_ON_ONCE_FOLIO(folio_ref_count(folio), folio); in __hugetlb_vmemmap_optimize_folio()
H A Dmemfd.c34 return folio_ref_count(folio) - folio_mapcount(folio) != in memfd_folio_has_extra_refs()
H A Ddebug.c79 folio_ref_count(folio), mapcount, mapping, in __dump_folio()
H A Dworkingset.c390 VM_BUG_ON_FOLIO(folio_ref_count(folio), folio); in workingset_eviction()
H A Dhuge_memory.c1983 if (folio_ref_count(folio) > in do_huge_pmd_wp_page()
1988 if (folio_ref_count(folio) == 1) { in do_huge_pmd_wp_page()
2969 VM_WARN_ON_FOLIO(!folio_ref_count(folio), folio); in __split_huge_pmd_locked()
3203 ref_count = folio_ref_count(folio); in __discard_anon_folio_pmd_locked()
3305 return folio_mapcount(folio) == folio_ref_count(folio) - extra_pins - in can_split_folio()
4055 WARN_ON_ONCE(folio_ref_count(folio)); in __folio_unqueue_deferred_split()
H A Dmemory.c2113 if (!folio_ref_count(folio)) in validate_page_before_insert()
3749 if (folio_large_mapcount(folio) != folio_ref_count(folio)) in __wp_can_reuse_large_anon_folio()
3754 VM_WARN_ON_ONCE_FOLIO(folio_large_mapcount(folio) > folio_ref_count(folio), folio); in __wp_can_reuse_large_anon_folio()
3758 if (folio_large_mapcount(folio) != folio_ref_count(folio)) in __wp_can_reuse_large_anon_folio()
3797 if (folio_test_ksm(folio) || folio_ref_count(folio) > 3) in wp_can_reuse_anon_folio()
3805 if (folio_ref_count(folio) > 1 + folio_test_swapcache(folio)) in wp_can_reuse_anon_folio()
3811 if (folio_test_ksm(folio) || folio_ref_count(folio) != 1) { in wp_can_reuse_anon_folio()
4132 folio_ref_count(folio) == (1 + folio_nr_pages(folio)); in should_try_to_free_swap()
4752 (exclusive || folio_ref_count(folio) == 1)) { in do_swap_page()
H A Dhugetlb.c1286 VM_BUG_ON_FOLIO(folio_ref_count(folio), folio); in enqueue_hugetlb_folio()
1837 VM_BUG_ON_FOLIO(folio_ref_count(folio), folio); in free_huge_folio()
2150 if (!folio_ref_count(folio)) { in dissolve_free_hugetlb_folio()
2834 } else if (folio_ref_count(old_folio)) { in alloc_and_dissolve_hugetlb_folio()
2927 if (folio_ref_count(folio) && folio_isolate_hugetlb(folio, list)) in isolate_or_dissolve_huge_page()
2929 else if (!folio_ref_count(folio)) in isolate_or_dissolve_huge_page()
2967 if (!folio_ref_count(folio)) { in replace_free_hugepage_folios()
3467 WARN_ON(folio_ref_count(folio) != 1); in gather_bootmem_prealloc_node()
H A Dtruncate.c308 if (folio_ref_count(folio) > in mapping_evict_folio()
H A Dmigrate_device.c366 if ((folio_ref_count(folio) - extra) > folio_mapcount(folio)) in migrate_vma_check_page()
H A Dgup.c84 if (WARN_ON_ONCE(folio_ref_count(folio) < 0)) in try_get_folio()
147 if (WARN_ON_ONCE(folio_ref_count(folio) <= 0)) in try_grab_folio()
231 WARN_ON_ONCE(folio_ref_count(folio) < GUP_PIN_COUNTING_BIAS); in folio_add_pin()
H A Dkhugepaged.c561 return folio_ref_count(folio) == expected_refcount; in is_refcount_suitable()
2034 if (folio_ref_count(folio) != 2 + folio_nr_pages(folio)) { in collapse_file()
H A Dmemcontrol-v1.c597 VM_BUG_ON_FOLIO(folio_ref_count(folio), folio); in memcg1_swapout()
H A Dvmscan.c489 return folio_ref_count(folio) - folio_test_private(folio) == in is_page_cache_freeable()
1504 if (!mapping && folio_ref_count(folio) == 1) { in shrink_folio_list()
1842 VM_BUG_ON_FOLIO(!folio_ref_count(folio), folio); in folio_isolate_lru()
H A Dmemory-failure.c2595 if (folio_ref_count(folio) > 1) { in unpoison_memory()
H A Drmap.c2110 ref_count = folio_ref_count(folio); in try_to_unmap_one()
H A Dcompaction.c1120 if (!mapping && (folio_ref_count(folio) - 1) > folio_mapcount(folio)) in isolate_migratepages_block()
H A Dksm.c1305 if (folio_mapcount(folio) + 1 + swapped != folio_ref_count(folio)) { in write_protect_page()
/linux-6.15/fs/nilfs2/
H A Dpage.c169 folio, folio_ref_count(folio), in nilfs_folio_bug()
/linux-6.15/fs/
H A Ddax.c416 WARN_ON_ONCE(folio_ref_count(new_folio)); in dax_folio_put()
438 WARN_ON_ONCE(folio_ref_count(folio)); in dax_folio_init()
493 if (folio_ref_count(folio) - folio_mapcount(folio)) in dax_busy_page()
H A Daio.c343 folio_ref_count(folio)); in aio_free_ring()
531 folio_ref_count(folio)); in aio_setup_ring()
/linux-6.15/fs/btrfs/
H A Dcompression.c260 ASSERT(folio_ref_count(folio) == 1); in btrfs_free_compr_folio()

12