Lines Matching refs:prot

126 	vm_prot_t	prot;  member
256 if (((fs->prot & VM_PROT_WRITE) == 0 && in vm_fault_dirty()
313 if (m == NULL || ((fs->prot & VM_PROT_WRITE) != 0 && in vm_fault_soft_fast()
329 if ((fs->prot & VM_PROT_WRITE) != 0) { in vm_fault_soft_fast()
350 rv = pmap_enter(fs->map->pmap, vaddr, m_map, fs->prot, fs->fault_type | in vm_fault_soft_fast()
504 rv = pmap_enter(fs->map->pmap, vaddr, m, fs->prot, in vm_fault_populate()
560 rv = pmap_enter(fs->map->pmap, vaddr, m, fs->prot, fs->fault_type | in vm_fault_populate()
566 &m[i], fs->prot, fs->fault_type | in vm_fault_populate()
802 &fs->first_pindex, &fs->prot, &fs->wired); in vm_fault_lookup()
832 fs->fault_type = fs->prot | (fs->fault_type & VM_PROT_COPY); in vm_fault_lookup()
880 fs->prot &= retry_prot; in vm_fault_relookup()
882 if (fs->prot == 0) in vm_fault_relookup()
1541 fs.prot &= ~VM_PROT_WRITE; in vm_fault()
1585 pmap_enter(fs.map->pmap, vaddr, fs.m, fs.prot, in vm_fault()
1797 vm_prot_t prot, vm_page_t *ma, int max_count) in vm_fault_quick_hold_pages() argument
1822 *mp = pmap_extract_and_hold(map->pmap, va, prot); in vm_fault_quick_hold_pages()
1825 else if ((prot & VM_PROT_WRITE) != 0 && in vm_fault_quick_hold_pages()
1854 if ((prot & VM_PROT_QUICK_NOFAULT) != 0 && in vm_fault_quick_hold_pages()
1858 if (*mp == NULL && vm_fault(map, va, prot, in vm_fault_quick_hold_pages()
1891 vm_prot_t access, prot; in vm_fault_copy_entry() local
1902 access = prot = dst_entry->protection; in vm_fault_copy_entry()
2053 pmap_enter(dst_map->pmap, vaddr, dst_m, prot, in vm_fault_copy_entry()