| /linux-6.15/mm/ |
| H A D | memory.c | 3514 vmf->pte = pte_offset_map_lock(mm, vmf->pmd, vmf->address, &vmf->ptl); in wp_page_copy() 3627 vmf->pte = pte_offset_map_lock(vmf->vma->vm_mm, vmf->pmd, vmf->address, in finish_mkwrite_fault() 3636 update_mmu_tlb(vmf->vma, vmf->address, vmf->pte); in finish_mkwrite_fault() 4137 vmf->pte = pte_offset_map_lock(vmf->vma->vm_mm, vmf->pmd, in pte_marker_clear() 4150 pte_clear(vmf->vma->vm_mm, vmf->address, vmf->pte); in pte_marker_clear() 4443 vmf->address, &vmf->ptl); in do_swap_page() 4543 vmf->address, &vmf->ptl); in do_swap_page() 4977 vmf->address, &vmf->ptl); in do_anonymous_page() 5632 vmf->pte = pte_offset_map_lock(vmf->vma->vm_mm, vmf->pmd, in do_fault() 6008 update_mmu_tlb(vmf->vma, vmf->address, vmf->pte); in handle_pte_fault() [all …]
|
| H A D | huge_memory.c | 1236 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in __do_huge_pmd_anonymous_page() 1348 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_anonymous_page() 1865 vmf->ptl = pud_lock(vmf->vma->vm_mm, vmf->pud); in huge_pud_set_accessed() 1869 touch_pud(vmf->vma, vmf->address, vmf->pud, write); in huge_pud_set_accessed() 1879 vmf->ptl = pmd_lock(vmf->vma->vm_mm, vmf->pmd); in huge_pmd_set_accessed() 1880 if (unlikely(!pmd_same(*vmf->pmd, vmf->orig_pmd))) in huge_pmd_set_accessed() 1883 touch_pmd(vmf->vma, vmf->address, vmf->pmd, write); in huge_pmd_set_accessed() 1904 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_zero_wp_pmd() 1930 vmf->ptl = pmd_lockptr(vma->vm_mm, vmf->pmd); in do_huge_pmd_wp_page() 2057 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_numa_page() [all …]
|
| H A D | filemap.c | 3321 ptep = pte_offset_map_ro_nolock(vma->vm_mm, vmf->pmd, vmf->address, in filemap_fault_recheck_pte_none() 3322 &vmf->ptl); in filemap_fault_recheck_pte_none() 3329 spin_lock(vmf->ptl); in filemap_fault_recheck_pte_none() 3415 vmf->gfp_mask); in filemap_fault() 3544 if (pmd_none(*vmf->pmd) && vmf->prealloc_pte) in filemap_map_pmd() 3545 pmd_install(mm, vmf->pmd, &vmf->prealloc_pte); in filemap_map_pmd() 3640 vmf->pte += count; in filemap_map_folio_range() 3653 vmf->pte = old_ptep; in filemap_map_folio_range() 3715 vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, addr, &vmf->ptl); in filemap_map_pages() 3716 if (!vmf->pte) { in filemap_map_pages() [all …]
|
| H A D | hugetlb.c | 6467 if (!hugetlb_pte_stable(h, mm, vmf->address, vmf->pte, vmf->orig_pte)) { in hugetlb_no_page() 6496 if (hugetlb_pte_stable(h, mm, vmf->address, vmf->pte, vmf->orig_pte)) in hugetlb_no_page() 6545 if (!hugetlb_pte_stable(h, mm, vmf->address, vmf->pte, vmf->orig_pte)) { in hugetlb_no_page() 6569 vmf->ptl = huge_pte_lock(h, mm, vmf->pte); in hugetlb_no_page() 6572 if (!pte_same(huge_ptep_get(mm, vmf->address, vmf->pte), vmf->orig_pte)) in hugetlb_no_page() 6700 vmf.orig_pte = huge_ptep_get(mm, vmf.address, vmf.pte); in hugetlb_fault() 6775 vmf.ptl = huge_pte_lock(h, mm, vmf.pte); in hugetlb_fault() 6778 if (unlikely(!pte_same(vmf.orig_pte, huge_ptep_get(mm, vmf.address, vmf.pte)))) in hugetlb_fault() 6796 set_huge_pte_at(mm, vmf.address, vmf.pte, vmf.orig_pte, in hugetlb_fault() 6823 vmf.orig_pte = pte_mkyoung(vmf.orig_pte); in hugetlb_fault() [all …]
|
| H A D | secretmem.c | 50 static vm_fault_t secretmem_fault(struct vm_fault *vmf) in secretmem_fault() argument 52 struct address_space *mapping = vmf->vma->vm_file->f_mapping; in secretmem_fault() 53 struct inode *inode = file_inode(vmf->vma->vm_file); in secretmem_fault() 54 pgoff_t offset = vmf->pgoff; in secretmem_fault() 55 gfp_t gfp = vmf->gfp_mask; in secretmem_fault() 62 if (((loff_t)vmf->pgoff << PAGE_SHIFT) >= i_size_read(inode)) in secretmem_fault() 105 vmf->page = page; in secretmem_fault()
|
| H A D | swap_state.c | 677 static int swap_vma_ra_win(struct vm_fault *vmf, unsigned long *start, in swap_vma_ra_win() argument 680 struct vm_area_struct *vma = vmf->vma; in swap_vma_ra_win() 689 faddr = vmf->address; in swap_vma_ra_win() 732 struct mempolicy *mpol, pgoff_t targ_ilx, struct vm_fault *vmf) in swap_vma_readahead() argument 744 win = swap_vma_ra_win(vmf, &start, &end); in swap_vma_readahead() 748 ilx = targ_ilx - PFN_DOWN(vmf->address - start); in swap_vma_readahead() 753 pte = pte_offset_map(vmf->pmd, addr); in swap_vma_readahead() 771 if (addr != vmf->address) { in swap_vma_readahead() 805 struct vm_fault *vmf) in swapin_readahead() argument 811 mpol = get_vma_policy(vmf->vma, vmf->address, 0, &ilx); in swapin_readahead() [all …]
|
| /linux-6.15/include/trace/events/ |
| H A D | fs_dax.h | 13 TP_ARGS(inode, vmf, max_pgoff, result), 30 __entry->vm_end = vmf->vma->vm_end; 32 __entry->address = vmf->address; 33 __entry->flags = vmf->flags; 34 __entry->pgoff = vmf->pgoff; 80 __entry->address = vmf->address; 123 __entry->address = vmf->address; 155 TP_ARGS(inode, vmf, result), 170 __entry->flags = vmf->flags; 171 __entry->pgoff = vmf->pgoff; [all …]
|
| /linux-6.15/drivers/dax/ |
| H A D | device.c | 80 struct file *filp = vmf->vma->vm_file; in dax_set_mapping() 88 pgoff = linear_page_index(vmf->vma, in dax_set_mapping() 103 struct vm_fault *vmf) in __dev_dax_pte_fault() argument 133 vmf->flags & FAULT_FLAG_WRITE); in __dev_dax_pte_fault() 137 struct vm_fault *vmf) in __dev_dax_pmd_fault() argument 161 if (pmd_addr < vmf->vma->vm_start || in __dev_dax_pmd_fault() 177 vmf->flags & FAULT_FLAG_WRITE); in __dev_dax_pmd_fault() 182 struct vm_fault *vmf) in __dev_dax_pud_fault() argument 223 vmf->flags & FAULT_FLAG_WRITE); in __dev_dax_pud_fault() 227 struct vm_fault *vmf) in __dev_dax_pud_fault() argument [all …]
|
| /linux-6.15/fs/ |
| H A D | dax.c | 1031 copy_user_page(vto, kaddr, vmf->address, vmf->cow_page); in copy_cow_page_dax() 1085 vmf->address, shared); in dax_insert_entry() 1415 ptl = pmd_lock(vmf->vma->vm_mm, vmf->pmd); in dax_pmd_load_hole() 1427 set_pmd_at(vmf->vma->vm_mm, pmd_addr, vmf->pmd, pmd_entry); in dax_pmd_load_hole() 1812 clear_user_highpage(vmf->cow_page, vmf->address); in dax_fault_cow_page() 1827 ret = finish_fault(vmf); in dax_fault_cow_page() 1857 if (!pmd && vmf->cow_page) in dax_fault_iter() 1926 if ((vmf->flags & FAULT_FLAG_WRITE) && !vmf->cow_page) in dax_iomap_pte_fault() 1941 if (pmd_trans_huge(*vmf->pmd) || pmd_devmap(*vmf->pmd)) { in dax_iomap_pte_fault() 2064 if (!pmd_none(*vmf->pmd) && !pmd_trans_huge(*vmf->pmd) && in dax_iomap_pmd_fault() [all …]
|
| H A D | userfaultfd.c | 231 struct vm_fault *vmf, in userfaultfd_huge_must_wait() argument 238 assert_fault_locked(vmf); in userfaultfd_huge_must_wait() 261 struct vm_fault *vmf, in userfaultfd_huge_must_wait() argument 276 struct vm_fault *vmf, in userfaultfd_must_wait() argument 280 unsigned long address = vmf->address; in userfaultfd_must_wait() 289 assert_fault_locked(vmf); in userfaultfd_must_wait() 380 assert_fault_locked(vmf); in handle_userfault() 419 vmf->flags); in handle_userfault() 455 release_fault_lock(vmf); in handle_userfault() 464 uwq.msg = userfault_msg(vmf->address, vmf->real_address, vmf->flags, in handle_userfault() [all …]
|
| /linux-6.15/drivers/gpu/drm/ttm/ |
| H A D | ttm_bo_vm.c | 42 struct vm_fault *vmf) in ttm_bo_vm_fault_idle() argument 58 if (vmf->flags & FAULT_FLAG_RETRY_NOWAIT) in ttm_bo_vm_fault_idle() 62 mmap_read_unlock(vmf->vma->vm_mm); in ttm_bo_vm_fault_idle() 117 struct vm_fault *vmf) in ttm_bo_vm_reserve() argument 134 mmap_read_unlock(vmf->vma->vm_mm); in ttm_bo_vm_reserve() 185 struct vm_area_struct *vma = vmf->vma; in ttm_bo_vm_fault_reserved() 196 unsigned long address = vmf->address; in ttm_bo_vm_fault_reserved() 202 ret = ttm_bo_vm_fault_idle(bo, vmf); in ttm_bo_vm_fault_reserved() 293 struct vm_area_struct *vma = vmf->vma; in ttm_bo_vm_dummy_page() 323 struct vm_area_struct *vma = vmf->vma; in ttm_bo_vm_fault() [all …]
|
| /linux-6.15/fs/ocfs2/ |
| H A D | mmap.c | 31 static vm_fault_t ocfs2_fault(struct vm_fault *vmf) in ocfs2_fault() argument 33 struct vm_area_struct *vma = vmf->vma; in ocfs2_fault() 38 ret = filemap_fault(vmf); in ocfs2_fault() 42 vma, vmf->page, vmf->pgoff); in ocfs2_fault() 113 static vm_fault_t ocfs2_page_mkwrite(struct vm_fault *vmf) in ocfs2_page_mkwrite() argument 115 struct folio *folio = page_folio(vmf->page); in ocfs2_page_mkwrite() 116 struct inode *inode = file_inode(vmf->vma->vm_file); in ocfs2_page_mkwrite() 144 ret = __ocfs2_page_mkwrite(vmf->vma->vm_file, di_bh, folio); in ocfs2_page_mkwrite()
|
| /linux-6.15/drivers/video/fbdev/core/ |
| H A D | fb_defio.c | 121 static vm_fault_t fb_deferred_io_fault(struct vm_fault *vmf) in fb_deferred_io_fault() argument 125 struct fb_info *info = vmf->vma->vm_private_data; in fb_deferred_io_fault() 127 offset = vmf->pgoff << PAGE_SHIFT; in fb_deferred_io_fault() 135 if (!vmf->vma->vm_file) in fb_deferred_io_fault() 140 vmf->page = page; in fb_deferred_io_fault() 221 unsigned long offset = vmf->pgoff << PAGE_SHIFT; in fb_deferred_io_page_mkwrite() 222 struct page *page = vmf->page; in fb_deferred_io_page_mkwrite() 224 file_update_time(vmf->vma->vm_file); in fb_deferred_io_page_mkwrite() 230 static vm_fault_t fb_deferred_io_mkwrite(struct vm_fault *vmf) in fb_deferred_io_mkwrite() argument 232 struct fb_info *info = vmf->vma->vm_private_data; in fb_deferred_io_mkwrite() [all …]
|
| /linux-6.15/arch/x86/entry/vdso/ |
| H A D | vma.c | 53 struct vm_area_struct *vma, struct vm_fault *vmf) in vdso_fault() argument 57 if (!image || (vmf->pgoff << PAGE_SHIFT) >= image->size) in vdso_fault() 60 vmf->page = virt_to_page(image->data + (vmf->pgoff << PAGE_SHIFT)); in vdso_fault() 61 get_page(vmf->page); in vdso_fault() 94 struct vm_area_struct *vma, struct vm_fault *vmf) in vvar_vclock_fault() argument 96 switch (vmf->pgoff) { in vvar_vclock_fault() 104 return vmf_insert_pfn_prot(vma, vmf->address, in vvar_vclock_fault() 115 return vmf_insert_pfn(vma, vmf->address, pfn); in vvar_vclock_fault()
|
| /linux-6.15/fs/xfs/ |
| H A D | xfs_file.c | 1658 struct vm_fault *vmf, in xfs_dax_fault_locked() argument 1680 struct vm_fault *vmf, in xfs_dax_read_fault() argument 1707 struct vm_fault *vmf, in __xfs_write_fault() argument 1746 struct vm_fault *vmf, in xfs_write_fault_zoned() argument 1772 struct vm_fault *vmf, in xfs_write_fault() argument 1782 struct vm_fault *vmf) in xfs_is_write_fault() argument 1790 struct vm_fault *vmf) in xfs_filemap_fault() argument 1802 return filemap_fault(vmf); in xfs_filemap_fault() 1807 struct vm_fault *vmf, in xfs_filemap_huge_fault() argument 1821 struct vm_fault *vmf) in xfs_filemap_page_mkwrite() argument [all …]
|
| /linux-6.15/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_page_dirty.c | 375 vm_fault_t vmw_bo_vm_mkwrite(struct vm_fault *vmf) in vmw_bo_vm_mkwrite() argument 377 struct vm_area_struct *vma = vmf->vma; in vmw_bo_vm_mkwrite() 389 save_flags = vmf->flags; in vmw_bo_vm_mkwrite() 390 vmf->flags &= ~FAULT_FLAG_ALLOW_RETRY; in vmw_bo_vm_mkwrite() 391 ret = ttm_bo_vm_reserve(bo, vmf); in vmw_bo_vm_mkwrite() 392 vmf->flags = save_flags; in vmw_bo_vm_mkwrite() 416 vm_fault_t vmw_bo_vm_fault(struct vm_fault *vmf) in vmw_bo_vm_fault() argument 418 struct vm_area_struct *vma = vmf->vma; in vmw_bo_vm_fault() 426 ret = ttm_bo_vm_reserve(bo, vmf); in vmw_bo_vm_fault() 437 page_offset = vmf->pgoff - in vmw_bo_vm_fault() [all …]
|
| /linux-6.15/lib/vdso/ |
| H A D | datastore.c | 41 struct vm_area_struct *vma, struct vm_fault *vmf) in vvar_fault() argument 47 switch (vmf->pgoff) { in vvar_fault() 57 addr = vmf->address + VDSO_TIMENS_PAGE_OFFSET * PAGE_SIZE; in vvar_fault() 85 vmf->pgoff - VDSO_ARCH_PAGES_START; in vvar_fault() 91 return vmf_insert_pfn(vma, vmf->address, pfn); in vvar_fault()
|
| /linux-6.15/drivers/xen/ |
| H A D | privcmd-buf.c | 118 static vm_fault_t privcmd_buf_vma_fault(struct vm_fault *vmf) in privcmd_buf_vma_fault() argument 121 vmf->vma, vmf->vma->vm_start, vmf->vma->vm_end, in privcmd_buf_vma_fault() 122 vmf->pgoff, (void *)vmf->address); in privcmd_buf_vma_fault()
|
| /linux-6.15/include/linux/ |
| H A D | huge_mm.h | 10 vm_fault_t do_huge_pmd_anonymous_page(struct vm_fault *vmf); 14 void huge_pmd_set_accessed(struct vm_fault *vmf); 20 void huge_pud_set_accessed(struct vm_fault *vmf, pud_t orig_pud); 22 static inline void huge_pud_set_accessed(struct vm_fault *vmf, pud_t orig_pud) in huge_pud_set_accessed() argument 27 vm_fault_t do_huge_pmd_wp_page(struct vm_fault *vmf); 40 vm_fault_t vmf_insert_pfn_pmd(struct vm_fault *vmf, pfn_t pfn, bool write); 41 vm_fault_t vmf_insert_pfn_pud(struct vm_fault *vmf, pfn_t pfn, bool write); 42 vm_fault_t vmf_insert_folio_pmd(struct vm_fault *vmf, struct folio *folio, 44 vm_fault_t vmf_insert_folio_pud(struct vm_fault *vmf, struct folio *folio, 480 vm_fault_t do_huge_pmd_numa_page(struct vm_fault *vmf); [all …]
|
| /linux-6.15/fs/nilfs2/ |
| H A D | file.c | 45 static vm_fault_t nilfs_page_mkwrite(struct vm_fault *vmf) in nilfs_page_mkwrite() argument 47 struct vm_area_struct *vma = vmf->vma; in nilfs_page_mkwrite() 48 struct folio *folio = page_folio(vmf->page); in nilfs_page_mkwrite() 101 ret = block_page_mkwrite(vma, vmf, nilfs_get_block); in nilfs_page_mkwrite()
|
| /linux-6.15/drivers/misc/ocxl/ |
| H A D | sysfs.c | 109 static vm_fault_t global_mmio_fault(struct vm_fault *vmf) in global_mmio_fault() argument 111 struct vm_area_struct *vma = vmf->vma; in global_mmio_fault() 115 if (vmf->pgoff >= (afu->config.global_mmio_size >> PAGE_SHIFT)) in global_mmio_fault() 118 offset = vmf->pgoff; in global_mmio_fault() 120 return vmf_insert_pfn(vma, vmf->address, offset); in global_mmio_fault()
|
| H A D | context.c | 139 static vm_fault_t ocxl_mmap_fault(struct vm_fault *vmf) in ocxl_mmap_fault() argument 141 struct vm_area_struct *vma = vmf->vma; in ocxl_mmap_fault() 146 offset = vmf->pgoff << PAGE_SHIFT; in ocxl_mmap_fault() 148 ctx->pasid, vmf->address, offset); in ocxl_mmap_fault() 151 ret = map_pp_mmio(vma, vmf->address, offset, ctx); in ocxl_mmap_fault() 153 ret = map_afu_irq(vma, vmf->address, offset, ctx); in ocxl_mmap_fault()
|
| /linux-6.15/fs/ext2/ |
| H A D | file.c | 93 static vm_fault_t ext2_dax_fault(struct vm_fault *vmf) in ext2_dax_fault() argument 95 struct inode *inode = file_inode(vmf->vma->vm_file); in ext2_dax_fault() 97 bool write = (vmf->flags & FAULT_FLAG_WRITE) && in ext2_dax_fault() 98 (vmf->vma->vm_flags & VM_SHARED); in ext2_dax_fault() 102 file_update_time(vmf->vma->vm_file); in ext2_dax_fault() 106 ret = dax_iomap_fault(vmf, 0, NULL, NULL, &ext2_iomap_ops); in ext2_dax_fault()
|
| /linux-6.15/drivers/char/agp/ |
| H A D | alpha-agp.c | 14 static vm_fault_t alpha_core_agp_vm_fault(struct vm_fault *vmf) in alpha_core_agp_vm_fault() argument 21 dma_addr = vmf->address - vmf->vma->vm_start + agp->aperture.bus_base; in alpha_core_agp_vm_fault() 32 vmf->page = page; in alpha_core_agp_vm_fault()
|
| /linux-6.15/sound/usb/usx2y/ |
| H A D | usX2Yhwdep.c | 21 static vm_fault_t snd_us428ctls_vm_fault(struct vm_fault *vmf) in snd_us428ctls_vm_fault() argument 27 offset = vmf->pgoff << PAGE_SHIFT; in snd_us428ctls_vm_fault() 28 vaddr = (char *)((struct usx2ydev *)vmf->vma->vm_private_data)->us428ctls_sharedmem + offset; in snd_us428ctls_vm_fault() 31 vmf->page = page; in snd_us428ctls_vm_fault()
|