| /linux-6.15/mm/ |
| H A D | mmap.c | 83 unsigned long vm_flags = vma->vm_flags; in vma_set_page_prot() local 88 vm_flags &= ~VM_SHARED; in vma_set_page_prot() 473 if (vm_flags & VM_EXEC) in do_mmap() 475 vm_flags &= ~VM_MAYEXEC; in do_mmap() 522 vm_flags |= VM_DROPPABLE; in do_mmap() 731 vm_flags); in arch_get_unmapped_area() 799 vm_flags); in arch_get_unmapped_area_topdown() 1100 vm_flags_t vm_flags; in SYSCALL_DEFINE5() local 1142 vm_flags = vma->vm_flags; in SYSCALL_DEFINE5() 1168 if (vma->vm_flags != vm_flags) in SYSCALL_DEFINE5() [all …]
|
| H A D | nommu.c | 850 unsigned long vm_flags; in determine_vm_flags() local 876 vm_flags |= VM_MAYOVERLAY; in determine_vm_flags() 883 return vm_flags; in determine_vm_flags() 963 region->vm_flags = vma->vm_flags; in do_mmap_private() 1014 vm_flags_t vm_flags, in do_mmap() argument 1054 region->vm_flags = vm_flags; in do_mmap() 1057 vm_flags_init(vma, vm_flags); in do_mmap() 1613 if (!(vma->vm_flags & VM_USERMAP)) in remap_vmalloc_range() 1741 if (vma->vm_flags & VM_MAYREAD) { in __copy_remote_vm_str() 1819 if (vma->vm_flags & VM_SHARED) { in nommu_shrink_inode_mappings() [all …]
|
| H A D | userfaultfd.c | 51 else if (!(vma->vm_flags & VM_SHARED) && in find_vma_and_prepare_anon() 680 if (!(dst_vma->vm_flags & VM_SHARED)) { in mfill_atomic_pte() 752 dst_vma->vm_flags & VM_SHARED)) in mfill_atomic() 1467 if ((src_vma->vm_flags & VM_ACCESS_FLAGS) != (dst_vma->vm_flags & VM_ACCESS_FLAGS) || in validate_move_areas() 1472 if ((src_vma->vm_flags & VM_LOCKED) != (dst_vma->vm_flags & VM_LOCKED)) in validate_move_areas() 1479 if (!(src_vma->vm_flags & VM_WRITE)) in validate_move_areas() 1737 if (src_vma->vm_flags & VM_SHARED) in move_pages() 1742 if (dst_vma->vm_flags & VM_SHARED) in move_pages() 1944 unsigned long vm_flags, in userfaultfd_register_range() argument 1969 (vma->vm_flags & vm_flags) == vm_flags) in userfaultfd_register_range() [all …]
|
| H A D | vma.c | 1350 if (next->vm_flags & VM_LOCKED) in vms_gather_munmap_vmas() 1353 if (next->vm_flags & VM_ACCOUNT) in vms_gather_munmap_vmas() 1870 !((a->vm_flags ^ b->vm_flags) & ~(VM_ACCESS_FLAGS | VM_SOFTDIRTY)) && in anon_vma_compatible() 1965 if (vma->vm_flags & VM_PFNMAP) in vma_fs_can_writeback() 2374 (vma->vm_flags & VM_MAYWRITE)); in __mmap_new_file_vma() 2377 map->retry_merge = vma->vm_flags != map->flags && !(vma->vm_flags & VM_SPECIAL); in __mmap_new_file_vma() 2378 map->flags = vma->vm_flags; in __mmap_new_file_vma() 2465 unsigned long vm_flags = vma->vm_flags; in __mmap_complete() local 2473 if (vm_flags & VM_LOCKED) { in __mmap_complete() 2582 if (map_deny_write_exec(vm_flags, vm_flags)) in mmap_region() [all …]
|
| H A D | execmem.c | 29 pgprot_t pgprot, unsigned long vm_flags) in execmem_vmalloc() argument 39 vm_flags |= VM_DEFER_KMEMLEAK; in execmem_vmalloc() 41 if (vm_flags & VM_ALLOW_HUGE_VMAP) in execmem_vmalloc() 45 pgprot, vm_flags, NUMA_NO_NODE, in execmem_vmalloc() 51 pgprot, vm_flags, NUMA_NO_NODE, in execmem_vmalloc() 85 pgprot_t pgprot, unsigned long vm_flags) in execmem_vmalloc() argument 287 unsigned long vm_flags = VM_ALLOW_HUGE_VMAP; in execmem_cache_populate() local 294 p = execmem_vmalloc(range, alloc_size, PAGE_KERNEL, vm_flags); in execmem_cache_populate() 410 unsigned long vm_flags = VM_FLUSH_RESET_PERMS; in execmem_alloc() local 417 p = execmem_vmalloc(range, size, pgprot, vm_flags); in execmem_alloc()
|
| H A D | mseal.c | 43 if (vma->vm_file || vma->vm_flags & VM_SHARED) in is_ro_anon() 50 if (!(vma->vm_flags & VM_WRITE) || in is_ro_anon() 77 vm_flags_t oldflags = vma->vm_flags; in mseal_fixup() 148 newflags = vma->vm_flags | VM_SEALED; in apply_mm_seal()
|
| H A D | mlock.c | 334 if (!(vma->vm_flags & VM_LOCKED)) in allow_mlock_munlock() 373 if (vma->vm_flags & VM_LOCKED) in mlock_pte_range() 398 if (vma->vm_flags & VM_LOCKED) in mlock_pte_range() 475 vm_flags_t oldflags = vma->vm_flags; in mlock_fixup() 547 newflags = vma->vm_flags & ~VM_LOCKED_MASK; in apply_vma_lock_flags() 588 if (vma->vm_flags & VM_LOCKED) { in count_mm_mlocked_page_nr() 668 vm_flags_t vm_flags = VM_LOCKED; in SYSCALL_DEFINE3() local 674 vm_flags |= VM_LOCKONFAULT; in SYSCALL_DEFINE3() 676 return do_mlock(start, len, vm_flags); in SYSCALL_DEFINE3() 733 newflags = vma->vm_flags & ~VM_LOCKED_MASK; in apply_mlockall_flags()
|
| H A D | mremap.c | 915 if (vma->vm_flags & VM_MAYSHARE) in vrm_set_new_addr() 937 if (!(vrm->vma->vm_flags & VM_ACCOUNT)) in vrm_charge() 964 if (!(vrm->vma->vm_flags & VM_ACCOUNT)) in vrm_uncharge() 983 vm_stat_account(mm, vma->vm_flags, pages); in vrm_stat_account() 984 if (vma->vm_flags & VM_LOCKED) { in vrm_stat_account() 1000 unsigned long dummy = vma->vm_flags; in prep_move_vma() 1194 if (unlikely(vma->vm_flags & VM_PFNMAP)) in copy_vma_and_data() 1311 (vma->vm_flags & (VM_DONTEXPAND | VM_PFNMAP))) in resize_is_valid() 1327 if (vma->vm_flags & (VM_DONTEXPAND | VM_PFNMAP)) in resize_is_valid() 1432 vm_flags_t vm_flags = vrm->vma->vm_flags; in mremap_to() local [all …]
|
| H A D | vma.h | 180 .flags = vma_->vm_flags, \ 301 unsigned long len, vm_flags_t vm_flags, unsigned long pgoff, 318 if (vma->vm_flags & VM_SHARED) in vma_wants_manual_pte_write_upgrade() 320 return !!(vma->vm_flags & VM_WRITE); in vma_wants_manual_pte_write_upgrade() 324 static inline pgprot_t vm_pgprot_modify(pgprot_t oldprot, unsigned long vm_flags) in vm_pgprot_modify() argument 326 return pgprot_modify(oldprot, vm_get_page_prot(vm_flags)); in vm_pgprot_modify() 512 return (vma->vm_flags & VM_SEALED); in vma_is_sealed()
|
| /linux-6.15/arch/arm64/include/asm/ |
| H A D | mman.h | 71 static inline bool arch_validate_flags(unsigned long vm_flags) in arch_validate_flags() argument 78 if ((vm_flags & VM_MTE) && !(vm_flags & VM_MTE_ALLOWED)) in arch_validate_flags() 82 if (system_supports_gcs() && (vm_flags & VM_SHADOW_STACK)) { in arch_validate_flags() 84 if (vm_flags & VM_EXEC) in arch_validate_flags() 88 VM_WARN_ON(vm_flags & VM_SHARED); in arch_validate_flags() 94 #define arch_validate_flags(vm_flags) arch_validate_flags(vm_flags) argument
|
| /linux-6.15/include/trace/events/ |
| H A D | fs_dax.h | 18 __field(unsigned long, vm_flags) 31 __entry->vm_flags = vmf->vma->vm_flags; 43 __entry->vm_flags & VM_SHARED ? "shared" : "private", 70 __field(unsigned long, vm_flags) 79 __entry->vm_flags = vmf->vma->vm_flags; 111 __field(unsigned long, vm_flags) 122 __entry->vm_flags = vmf->vma->vm_flags; 158 __field(unsigned long, vm_flags) 168 __entry->vm_flags = vmf->vma->vm_flags; 202 __field(unsigned long, vm_flags) [all …]
|
| /linux-6.15/arch/arm64/mm/ |
| H A D | mmap.c | 84 pgprot_t vm_get_page_prot(unsigned long vm_flags) in vm_get_page_prot() argument 89 if (system_supports_gcs() && (vm_flags & VM_SHADOW_STACK)) { in vm_get_page_prot() 92 prot = pgprot_val(protection_map[vm_flags & in vm_get_page_prot() 96 if (vm_flags & VM_ARM64_BTI) in vm_get_page_prot() 109 if (vm_flags & VM_MTE) in vm_get_page_prot() 114 if (vm_flags & VM_PKEY_BIT0) in vm_get_page_prot() 116 if (vm_flags & VM_PKEY_BIT1) in vm_get_page_prot() 118 if (vm_flags & VM_PKEY_BIT2) in vm_get_page_prot()
|
| H A D | fault.c | 536 if (!(vma->vm_flags & VM_SHADOW_STACK)) in is_invalid_gcs_access() 552 unsigned long vm_flags; in do_page_fault() local 580 vm_flags = VM_EXEC; in do_page_fault() 588 vm_flags = VM_WRITE; in do_page_fault() 592 vm_flags = VM_WRITE; in do_page_fault() 596 vm_flags = VM_READ; in do_page_fault() 598 vm_flags |= VM_WRITE; in do_page_fault() 601 vm_flags |= VM_EXEC; in do_page_fault() 630 if (!(vma->vm_flags & vm_flags)) { in do_page_fault() 675 if (!(vma->vm_flags & vm_flags)) { in do_page_fault() [all …]
|
| /linux-6.15/arch/powerpc/include/asm/book3s/64/ |
| H A D | hash-pkey.h | 8 static inline u64 hash__vmflag_to_pte_pkey_bits(u64 vm_flags) in hash__vmflag_to_pte_pkey_bits() argument 10 return (((vm_flags & VM_PKEY_BIT0) ? H_PTE_PKEY_BIT0 : 0x0UL) | in hash__vmflag_to_pte_pkey_bits() 11 ((vm_flags & VM_PKEY_BIT1) ? H_PTE_PKEY_BIT1 : 0x0UL) | in hash__vmflag_to_pte_pkey_bits() 12 ((vm_flags & VM_PKEY_BIT2) ? H_PTE_PKEY_BIT2 : 0x0UL) | in hash__vmflag_to_pte_pkey_bits() 13 ((vm_flags & VM_PKEY_BIT3) ? H_PTE_PKEY_BIT3 : 0x0UL) | in hash__vmflag_to_pte_pkey_bits() 14 ((vm_flags & VM_PKEY_BIT4) ? H_PTE_PKEY_BIT4 : 0x0UL)); in hash__vmflag_to_pte_pkey_bits()
|
| H A D | pkeys.h | 8 static inline u64 vmflag_to_pte_pkey_bits(u64 vm_flags) in vmflag_to_pte_pkey_bits() argument 15 return hash__vmflag_to_pte_pkey_bits(vm_flags); in vmflag_to_pte_pkey_bits()
|
| /linux-6.15/arch/sparc/include/asm/ |
| H A D | mman.h | 57 #define arch_validate_flags(vm_flags) arch_validate_flags(vm_flags) argument 61 static inline bool arch_validate_flags(unsigned long vm_flags) in arch_validate_flags() argument 67 if (vm_flags & VM_SPARC_ADI) { in arch_validate_flags() 72 if (vm_flags & (VM_PFNMAP | VM_MIXEDMAP)) in arch_validate_flags() 83 if (vm_flags & VM_MERGEABLE) in arch_validate_flags()
|
| /linux-6.15/include/linux/ |
| H A D | userfaultfd_k.h | 168 return vma->vm_flags & (VM_UFFD_WP | VM_UFFD_MINOR); in uffd_disable_huge_pmd_share() 185 return vma->vm_flags & VM_UFFD_MISSING; in userfaultfd_missing() 190 return vma->vm_flags & VM_UFFD_WP; in userfaultfd_wp() 195 return vma->vm_flags & VM_UFFD_MINOR; in userfaultfd_minor() 212 return vma->vm_flags & __VM_UFFD_FLAGS; in userfaultfd_armed() 216 unsigned long vm_flags, in vma_can_userfault() argument 219 vm_flags &= __VM_UFFD_FLAGS; in vma_can_userfault() 221 if (vm_flags & VM_DROPPABLE) in vma_can_userfault() 224 if ((vm_flags & VM_UFFD_MINOR) && in vma_can_userfault() 232 if (wp_async && (vm_flags == VM_UFFD_WP)) in vma_can_userfault() [all …]
|
| H A D | huge_mm.h | 99 #define thp_vma_allowable_order(vma, vm_flags, tva_flags, order) \ argument 264 unsigned long vm_flags, 285 unsigned long vm_flags, in thp_vma_allowable_orders() argument 293 if (vm_flags & VM_HUGEPAGE) in thp_vma_allowable_orders() 296 ((vm_flags & VM_HUGEPAGE) && hugepage_global_enabled())) in thp_vma_allowable_orders() 320 unsigned long vm_flags) in vma_thp_disabled() argument 327 return (vm_flags & VM_NOHUGEPAGE) || in vma_thp_disabled() 341 vm_flags_t vm_flags); 530 unsigned long vm_flags, in thp_vma_allowable_orders() argument 544 unsigned long flags, vm_flags_t vm_flags) in thp_get_unmapped_area_vmflags() argument [all …]
|
| /linux-6.15/arch/x86/mm/ |
| H A D | pgprot.c | 35 pgprot_t vm_get_page_prot(unsigned long vm_flags) in vm_get_page_prot() argument 37 unsigned long val = pgprot_val(protection_map[vm_flags & in vm_get_page_prot() 48 if (vm_flags & VM_PKEY_BIT0) in vm_get_page_prot() 50 if (vm_flags & VM_PKEY_BIT1) in vm_get_page_prot() 52 if (vm_flags & VM_PKEY_BIT2) in vm_get_page_prot() 54 if (vm_flags & VM_PKEY_BIT3) in vm_get_page_prot()
|
| /linux-6.15/tools/testing/vma/ |
| H A D | vma_internal.h | 260 const vm_flags_t vm_flags; member 436 return __pgprot(vm_flags); in vm_get_page_prot() 447 return is_shared_maywrite(vma->vm_flags); in vma_is_shared_maywrite() 938 unsigned long vm_flags) in khugepaged_enter_vma() argument 941 (void)vm_flags; in khugepaged_enter_vma() 1059 unsigned long vm_flags = vma->vm_flags; in vma_set_page_prot() local 1066 vm_flags &= ~VM_SHARED; in vma_set_page_prot() 1090 if (vma->vm_flags & VM_GROWSDOWN) in stack_guard_start_gap() 1094 if (vma->vm_flags & VM_SHADOW_STACK) in stack_guard_start_gap() 1115 if (vma->vm_flags & VM_GROWSUP) { in vm_end_gap() [all …]
|
| /linux-6.15/tools/testing/selftests/bpf/progs/ |
| H A D | bpf_iter_task_vmas.c | 43 perm_str[0] = (vma->vm_flags & VM_READ) ? 'r' : '-'; in proc_maps() 44 perm_str[1] = (vma->vm_flags & VM_WRITE) ? 'w' : '-'; in proc_maps() 45 perm_str[2] = (vma->vm_flags & VM_EXEC) ? 'x' : '-'; in proc_maps() 46 perm_str[3] = (vma->vm_flags & VM_MAYSHARE) ? 's' : 'p'; in proc_maps()
|
| /linux-6.15/arch/x86/kernel/ |
| H A D | sys_x86_64.c | 118 static inline unsigned long stack_guard_placement(vm_flags_t vm_flags) in stack_guard_placement() argument 120 if (vm_flags & VM_SHADOW_STACK) in stack_guard_placement() 128 unsigned long pgoff, unsigned long flags, vm_flags_t vm_flags) in arch_get_unmapped_area() argument 156 info.start_gap = stack_guard_placement(vm_flags); in arch_get_unmapped_area() 169 unsigned long flags, vm_flags_t vm_flags) in arch_get_unmapped_area_topdown() argument 209 info.start_gap = stack_guard_placement(vm_flags); in arch_get_unmapped_area_topdown()
|
| /linux-6.15/arch/nios2/mm/ |
| H A D | cacheflush.c | 90 if (!(vma->vm_flags & VM_MAYSHARE)) in flush_aliases() 138 if (vma == NULL || (vma->vm_flags & VM_EXEC)) in flush_cache_range() 159 if (vma->vm_flags & VM_EXEC) in flush_cache_page() 236 if (vma->vm_flags & VM_EXEC) in update_mmu_cache_range() 268 if (vma->vm_flags & VM_EXEC) in copy_from_user_page() 279 if (vma->vm_flags & VM_EXEC) in copy_to_user_page()
|
| /linux-6.15/arch/hexagon/mm/ |
| H A D | vm_fault.c | 70 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault() 74 if (!(vma->vm_flags & VM_READ)) in do_page_fault() 78 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault()
|
| /linux-6.15/drivers/sbus/char/ |
| H A D | flash.c | 44 if ((vma->vm_flags & VM_READ) && in flash_mmap() 45 (vma->vm_flags & VM_WRITE)) { in flash_mmap() 49 if (vma->vm_flags & VM_READ) { in flash_mmap() 52 } else if (vma->vm_flags & VM_WRITE) { in flash_mmap()
|