| /f-stack/freebsd/amd64/vmm/io/ |
| H A D | iommu.c | 101 IOMMU_CREATE_MAPPING(void *domain, vm_paddr_t gpa, vm_paddr_t hpa, uint64_t len) in IOMMU_CREATE_MAPPING() argument 105 return ((*ops->create_mapping)(domain, gpa, hpa, len)); in IOMMU_CREATE_MAPPING() 288 iommu_create_mapping(void *dom, vm_paddr_t gpa, vm_paddr_t hpa, size_t len) in iommu_create_mapping() argument 295 mapped = IOMMU_CREATE_MAPPING(dom, gpa, hpa, remaining); in iommu_create_mapping() 297 hpa += mapped; in iommu_create_mapping()
|
| H A D | iommu.h | 41 vm_paddr_t hpa, uint64_t len); 70 void iommu_create_mapping(void *dom, vm_paddr_t gpa, vm_paddr_t hpa,
|
| H A D | ppt.h | 36 vm_paddr_t gpa, size_t len, vm_paddr_t hpa);
|
| H A D | ppt.c | 445 vm_paddr_t gpa, size_t len, vm_paddr_t hpa) in ppt_map_mmio() argument 458 error = vm_map_mmio(vm, gpa, len, hpa); in ppt_map_mmio()
|
| /f-stack/freebsd/amd64/vmm/ |
| H A D | vmm_mem.c | 62 vm_paddr_t hpa) in vmm_mmio_alloc() argument 69 error = sglist_append_phys(sg, hpa, len); in vmm_mmio_alloc()
|
| H A D | vmm_mem.h | 39 vm_paddr_t hpa);
|
| H A D | vmm_dev.c | 200 void *hpa, *cookie; in vmmdev_rw() local 235 hpa = vm_gpa_hold(sc->vm, lastcpu, gpa, c, in vmmdev_rw() 237 if (hpa == NULL) { in vmmdev_rw() 244 error = uiomove(hpa, c, uio); in vmmdev_rw() 526 pptmmio->hpa); in vmmdev_ioctl()
|
| H A D | vmm.c | 620 vm_map_mmio(struct vm *vm, vm_paddr_t gpa, size_t len, vm_paddr_t hpa) in vm_map_mmio() argument 624 if ((obj = vmm_mmio_alloc(vm->vmspace, gpa, len, hpa)) == NULL) in vm_map_mmio() 882 vm_paddr_t gpa, hpa; in vm_iommu_modify() local 919 hpa = DMAP_TO_PHYS((uintptr_t)vp); in vm_iommu_modify() 921 iommu_create_mapping(vm->iommu, gpa, hpa, sz); in vm_iommu_modify() 922 iommu_remove_mapping(host_domain, hpa, sz); in vm_iommu_modify() 925 iommu_create_mapping(host_domain, hpa, hpa, sz); in vm_iommu_modify()
|
| /f-stack/freebsd/amd64/vmm/intel/ |
| H A D | vtd.c | 518 vtd_update_mapping(void *arg, vm_paddr_t gpa, vm_paddr_t hpa, uint64_t len, in vtd_update_mapping() argument 537 if (hpa & PAGE_MASK) in vtd_update_mapping() 538 panic("vtd_create_mapping: unaligned hpa 0x%0lx", hpa); in vtd_update_mapping() 556 (hpa & (spsize - 1)) == 0 && in vtd_update_mapping() 597 ptp[ptpindex] = hpa | VTD_PTE_RD | VTD_PTE_WR; in vtd_update_mapping() 607 vtd_create_mapping(void *arg, vm_paddr_t gpa, vm_paddr_t hpa, uint64_t len) in vtd_create_mapping() argument 610 return (vtd_update_mapping(arg, gpa, hpa, len, CREATE_MAPPING)); in vtd_create_mapping()
|
| /f-stack/freebsd/amd64/vmm/amd/ |
| H A D | amdv.c | 86 amd_iommu_create_mapping(void *domain, vm_paddr_t gpa, vm_paddr_t hpa, in amd_iommu_create_mapping() argument
|
| H A D | amdvi_hw.c | 1146 vm_paddr_t hpa, uint64_t pg_size, bool create) in amdvi_set_pt() argument 1156 if (hpa & (pg_size - 1)) { in amdvi_set_pt() 1190 pt[index] = hpa | AMDVI_PT_RW | AMDVI_PT_PRESENT; in amdvi_set_pt() 1204 vm_paddr_t hpa, uint64_t size, bool create) in amdvi_update_mapping() argument 1217 len = amdvi_set_pt(ptp, level, gpa + mapped, hpa + mapped, in amdvi_update_mapping() 1221 hpa, gpa); in amdvi_update_mapping() 1231 amdvi_create_mapping(void *arg, vm_paddr_t gpa, vm_paddr_t hpa, in amdvi_create_mapping() argument 1248 return (amdvi_update_mapping(domain, gpa, hpa, len, true)); in amdvi_create_mapping()
|
| /f-stack/dpdk/lib/librte_vhost/ |
| H A D | vhost.h | 620 rte_iova_t hpa; in gpa_to_hpa() local 623 hpa = gpa_to_first_hpa(dev, gpa, size, &hpa_size); in gpa_to_hpa() 624 return hpa_size == size ? hpa : 0; in gpa_to_hpa()
|
| H A D | virtio_net.c | 987 void *hpa; in async_mbuf_to_desc() local 1077 hpa = (void *)(uintptr_t)gpa_to_first_hpa(dev, in async_mbuf_to_desc() 1081 if (unlikely(!hpa || mapped_len < cpy_threshold)) in async_mbuf_to_desc() 1089 hpa, (size_t)mapped_len); in async_mbuf_to_desc()
|
| /f-stack/freebsd/amd64/include/ |
| H A D | vmm_dev.h | 140 vm_paddr_t hpa; member
|
| H A D | vmm.h | 236 int vm_map_mmio(struct vm *vm, vm_paddr_t gpa, size_t len, vm_paddr_t hpa);
|