Lines Matching refs:mm_lock_seq
732 if (READ_ONCE(vma->vm_lock_seq) == READ_ONCE(mm->mm_lock_seq.sequence)) in vma_start_read()
759 if (unlikely(vma->vm_lock_seq == raw_read_seqcount(&mm->mm_lock_seq))) { in vma_start_read()
803 static bool __is_vma_write_locked(struct vm_area_struct *vma, unsigned int *mm_lock_seq) in __is_vma_write_locked() argument
811 *mm_lock_seq = vma->vm_mm->mm_lock_seq.sequence; in __is_vma_write_locked()
812 return (vma->vm_lock_seq == *mm_lock_seq); in __is_vma_write_locked()
815 void __vma_start_write(struct vm_area_struct *vma, unsigned int mm_lock_seq);
824 unsigned int mm_lock_seq; in vma_start_write() local
826 if (__is_vma_write_locked(vma, &mm_lock_seq)) in vma_start_write()
829 __vma_start_write(vma, mm_lock_seq); in vma_start_write()
834 unsigned int mm_lock_seq; in vma_assert_write_locked() local
836 VM_BUG_ON_VMA(!__is_vma_write_locked(vma, &mm_lock_seq), vma); in vma_assert_write_locked()
841 unsigned int mm_lock_seq; in vma_assert_locked() local
844 !__is_vma_write_locked(vma, &mm_lock_seq), vma); in vma_assert_locked()