Lines Matching refs:vm_area_struct

233 struct vm_area_struct *vm_area_alloc(struct mm_struct *);
234 struct vm_area_struct *vm_area_dup(struct vm_area_struct *);
235 void vm_area_free(struct vm_area_struct *);
536 struct vm_area_struct *vma; /* Target VMA */
586 void (*open)(struct vm_area_struct * area);
591 void (*close)(struct vm_area_struct * area);
593 int (*may_split)(struct vm_area_struct *area, unsigned long addr);
594 int (*mremap)(struct vm_area_struct *area);
600 int (*mprotect)(struct vm_area_struct *vma, unsigned long start,
606 unsigned long (*pagesize)(struct vm_area_struct * area);
619 int (*access)(struct vm_area_struct *vma, unsigned long addr,
625 const char *(*name)(struct vm_area_struct *vma);
635 int (*set_policy)(struct vm_area_struct *vma, struct mempolicy *new);
647 struct mempolicy *(*get_policy)(struct vm_area_struct *vma,
655 struct page *(*find_special_page)(struct vm_area_struct *vma,
660 static inline void vma_numab_state_init(struct vm_area_struct *vma) in vma_numab_state_init()
664 static inline void vma_numab_state_free(struct vm_area_struct *vma) in vma_numab_state_free()
669 static inline void vma_numab_state_init(struct vm_area_struct *vma) {} in vma_numab_state_init()
670 static inline void vma_numab_state_free(struct vm_area_struct *vma) {} in vma_numab_state_free()
674 static inline void vma_lock_init(struct vm_area_struct *vma, bool reset_refcnt) in vma_lock_init()
697 static inline void vma_refcount_put(struct vm_area_struct *vma) in vma_refcount_put()
720 static inline struct vm_area_struct *vma_start_read(struct mm_struct *mm, in vma_start_read()
721 struct vm_area_struct *vma) in vma_start_read()
773 static inline bool vma_start_read_locked_nested(struct vm_area_struct *vma, int subclass) in vma_start_read_locked_nested()
792 static inline bool vma_start_read_locked(struct vm_area_struct *vma) in vma_start_read_locked()
797 static inline void vma_end_read(struct vm_area_struct *vma) in vma_end_read()
803 static bool __is_vma_write_locked(struct vm_area_struct *vma, unsigned int *mm_lock_seq) in __is_vma_write_locked()
815 void __vma_start_write(struct vm_area_struct *vma, unsigned int mm_lock_seq);
822 static inline void vma_start_write(struct vm_area_struct *vma) in vma_start_write()
832 static inline void vma_assert_write_locked(struct vm_area_struct *vma) in vma_assert_write_locked()
839 static inline void vma_assert_locked(struct vm_area_struct *vma) in vma_assert_locked()
852 static inline void vma_assert_attached(struct vm_area_struct *vma) in vma_assert_attached()
857 static inline void vma_assert_detached(struct vm_area_struct *vma) in vma_assert_detached()
862 static inline void vma_mark_attached(struct vm_area_struct *vma) in vma_mark_attached()
869 void vma_mark_detached(struct vm_area_struct *vma);
887 struct vm_area_struct *lock_vma_under_rcu(struct mm_struct *mm,
892 static inline void vma_lock_init(struct vm_area_struct *vma, bool reset_refcnt) {} in vma_lock_init()
893 static inline struct vm_area_struct *vma_start_read(struct mm_struct *mm, in vma_start_read()
894 struct vm_area_struct *vma) in vma_start_read()
896 static inline void vma_end_read(struct vm_area_struct *vma) {} in vma_end_read()
897 static inline void vma_start_write(struct vm_area_struct *vma) {} in vma_start_write()
898 static inline void vma_assert_write_locked(struct vm_area_struct *vma) in vma_assert_write_locked()
900 static inline void vma_assert_attached(struct vm_area_struct *vma) {} in vma_assert_attached()
901 static inline void vma_assert_detached(struct vm_area_struct *vma) {} in vma_assert_detached()
902 static inline void vma_mark_attached(struct vm_area_struct *vma) {} in vma_mark_attached()
903 static inline void vma_mark_detached(struct vm_area_struct *vma) {} in vma_mark_detached()
905 static inline struct vm_area_struct *lock_vma_under_rcu(struct mm_struct *mm, in lock_vma_under_rcu()
911 static inline void vma_assert_locked(struct vm_area_struct *vma) in vma_assert_locked()
930 static inline void vma_init(struct vm_area_struct *vma, struct mm_struct *mm) in vma_init()
940 static inline void vm_flags_init(struct vm_area_struct *vma, in vm_flags_init()
951 static inline void vm_flags_reset(struct vm_area_struct *vma, in vm_flags_reset()
958 static inline void vm_flags_reset_once(struct vm_area_struct *vma, in vm_flags_reset_once()
965 static inline void vm_flags_set(struct vm_area_struct *vma, in vm_flags_set()
972 static inline void vm_flags_clear(struct vm_area_struct *vma, in vm_flags_clear()
983 static inline void __vm_flags_mod(struct vm_area_struct *vma, in __vm_flags_mod()
993 static inline void vm_flags_mod(struct vm_area_struct *vma, in vm_flags_mod()
1000 static inline void vma_set_anonymous(struct vm_area_struct *vma) in vma_set_anonymous()
1005 static inline bool vma_is_anonymous(struct vm_area_struct *vma) in vma_is_anonymous()
1014 static inline bool vma_is_initial_heap(const struct vm_area_struct *vma) in vma_is_initial_heap()
1024 static inline bool vma_is_initial_stack(const struct vm_area_struct *vma) in vma_is_initial_stack()
1035 static inline bool vma_is_temporary_stack(struct vm_area_struct *vma) in vma_is_temporary_stack()
1049 static inline bool vma_is_foreign(struct vm_area_struct *vma) in vma_is_foreign()
1060 static inline bool vma_is_accessible(struct vm_area_struct *vma) in vma_is_accessible()
1071 static inline bool vma_is_shared_maywrite(struct vm_area_struct *vma) in vma_is_shared_maywrite()
1077 struct vm_area_struct *vma_find(struct vma_iterator *vmi, unsigned long max) in vma_find()
1082 static inline struct vm_area_struct *vma_next(struct vma_iterator *vmi) in vma_next()
1092 struct vm_area_struct *vma_iter_next_range(struct vma_iterator *vmi) in vma_iter_next_range()
1098 static inline struct vm_area_struct *vma_prev(struct vma_iterator *vmi) in vma_prev()
1121 struct vm_area_struct *vma) in vma_iter_bulk_store()
1155 bool vma_is_shmem(struct vm_area_struct *vma);
1156 bool vma_is_anon_shmem(struct vm_area_struct *vma);
1158 static inline bool vma_is_shmem(struct vm_area_struct *vma) { return false; } in vma_is_shmem()
1159 static inline bool vma_is_anon_shmem(struct vm_area_struct *vma) { return false; } in vma_is_anon_shmem()
1162 int vma_is_stack_for_current(struct vm_area_struct *vma);
1455 static inline pte_t maybe_mkwrite(pte_t pte, struct vm_area_struct *vma) in maybe_mkwrite()
1831 static inline void vma_set_access_pid_bit(struct vm_area_struct *vma) in vma_set_access_pid_bit()
1892 static inline void vma_set_access_pid_bit(struct vm_area_struct *vma) in vma_set_access_pid_bit()
2062 static inline bool folio_needs_cow_for_dma(struct vm_area_struct *vma, in folio_needs_cow_for_dma()
2465 struct folio *vm_normal_folio(struct vm_area_struct *vma, unsigned long addr,
2467 struct page *vm_normal_page(struct vm_area_struct *vma, unsigned long addr,
2469 struct folio *vm_normal_folio_pmd(struct vm_area_struct *vma,
2471 struct page *vm_normal_page_pmd(struct vm_area_struct *vma, unsigned long addr,
2474 void zap_vma_ptes(struct vm_area_struct *vma, unsigned long address,
2476 void zap_page_range_single(struct vm_area_struct *vma, unsigned long address,
2478 static inline void zap_vma_pages(struct vm_area_struct *vma) in zap_vma_pages()
2484 struct vm_area_struct *start_vma, unsigned long start,
2492 copy_page_range(struct vm_area_struct *dst_vma, struct vm_area_struct *src_vma);
2493 int generic_access_phys(struct vm_area_struct *vma, unsigned long addr,
2502 struct vm_area_struct *vma;
2536 struct vm_area_struct *lock_mm_and_find_vma(struct mm_struct *mm,
2540 extern vm_fault_t handle_mm_fault(struct vm_area_struct *vma,
2551 static inline vm_fault_t handle_mm_fault(struct vm_area_struct *vma, in handle_mm_fault()
2578 static inline struct vm_area_struct *vma_lookup(struct mm_struct *mm,
2606 struct vm_area_struct **vmap) in get_user_page_vma_remote()
2609 struct vm_area_struct *vma; in get_user_page_vma_remote()
2683 bool can_change_pte_writable(struct vm_area_struct *vma, unsigned long addr,
2686 struct vm_area_struct *vma, unsigned long start,
2689 struct vm_area_struct *vma, struct vm_area_struct **pprev,
3380 void vma_interval_tree_insert(struct vm_area_struct *node,
3382 void vma_interval_tree_insert_after(struct vm_area_struct *node,
3383 struct vm_area_struct *prev,
3385 void vma_interval_tree_remove(struct vm_area_struct *node,
3387 struct vm_area_struct *vma_interval_tree_iter_first(struct rb_root_cached *root,
3389 struct vm_area_struct *vma_interval_tree_iter_next(struct vm_area_struct *node,
3415 extern int insert_vm_struct(struct mm_struct *, struct vm_area_struct *);
3417 int relocate_vma_down(struct vm_area_struct *vma, unsigned long shift);
3418 bool mmap_read_lock_maybe_expand(struct mm_struct *mm, struct vm_area_struct *vma,
3446 extern bool vma_is_special_mapping(const struct vm_area_struct *vma,
3448 extern struct vm_area_struct *_install_special_mapping(struct mm_struct *mm,
3474 int do_vmi_align_munmap(struct vma_iterator *vmi, struct vm_area_struct *vma,
3527 int expand_stack_locked(struct vm_area_struct *vma, unsigned long address);
3528 struct vm_area_struct *expand_stack(struct mm_struct * mm, unsigned long addr);
3531 extern struct vm_area_struct * find_vma(struct mm_struct * mm, unsigned long addr);
3532 extern struct vm_area_struct * find_vma_prev(struct mm_struct * mm, unsigned long addr,
3533 struct vm_area_struct **pprev);
3539 struct vm_area_struct *find_vma_intersection(struct mm_struct *mm,
3550 struct vm_area_struct *vma_lookup(struct mm_struct *mm, unsigned long addr) in vma_lookup()
3555 static inline unsigned long stack_guard_start_gap(struct vm_area_struct *vma) in stack_guard_start_gap()
3567 static inline unsigned long vm_start_gap(struct vm_area_struct *vma) in vm_start_gap()
3578 static inline unsigned long vm_end_gap(struct vm_area_struct *vma) in vm_end_gap()
3590 static inline unsigned long vma_pages(struct vm_area_struct *vma) in vma_pages()
3596 static inline struct vm_area_struct *find_exact_vma(struct mm_struct *mm, in find_exact_vma()
3599 struct vm_area_struct *vma = vma_lookup(mm, vm_start); in find_exact_vma()
3607 static inline bool range_in_vma(struct vm_area_struct *vma, in range_in_vma()
3615 void vma_set_page_prot(struct vm_area_struct *vma);
3621 static inline void vma_set_page_prot(struct vm_area_struct *vma) in vma_set_page_prot()
3627 void vma_set_file(struct vm_area_struct *vma, struct file *file);
3630 unsigned long change_prot_numa(struct vm_area_struct *vma,
3634 struct vm_area_struct *find_extend_vma_locked(struct mm_struct *,
3636 int remap_pfn_range(struct vm_area_struct *, unsigned long addr,
3638 int remap_pfn_range_notrack(struct vm_area_struct *vma, unsigned long addr,
3640 int vm_insert_page(struct vm_area_struct *, unsigned long addr, struct page *);
3641 int vm_insert_pages(struct vm_area_struct *vma, unsigned long addr,
3643 int vm_map_pages(struct vm_area_struct *vma, struct page **pages,
3645 int vm_map_pages_zero(struct vm_area_struct *vma, struct page **pages,
3649 vm_fault_t vmf_insert_pfn(struct vm_area_struct *vma, unsigned long addr,
3651 vm_fault_t vmf_insert_pfn_prot(struct vm_area_struct *vma, unsigned long addr,
3653 vm_fault_t vmf_insert_mixed(struct vm_area_struct *vma, unsigned long addr,
3655 vm_fault_t vmf_insert_mixed_mkwrite(struct vm_area_struct *vma,
3657 int vm_iomap_memory(struct vm_area_struct *vma, phys_addr_t start, unsigned long len);
3659 static inline vm_fault_t vmf_insert_page(struct vm_area_struct *vma, in vmf_insert_page()
3673 static inline int io_remap_pfn_range(struct vm_area_struct *vma, in io_remap_pfn_range()
3723 static inline bool gup_can_follow_protnone(struct vm_area_struct *vma, in gup_can_follow_protnone()
3892 extern struct vm_area_struct *get_gate_vma(struct mm_struct *mm);
3896 static inline struct vm_area_struct *get_gate_vma(struct mm_struct *mm) in get_gate_vma()
3917 const char * arch_vma_name(struct vm_area_struct *vma);
4137 struct vm_area_struct *vma);
4152 static inline bool vma_is_special_huge(const struct vm_area_struct *vma) in vma_is_special_huge()
4221 void vma_pgtable_walk_begin(struct vm_area_struct *vma);
4222 void vma_pgtable_walk_end(struct vm_area_struct *vma);