| /linux-6.15/arch/powerpc/kvm/ |
| H A D | book3s_32_mmu_host.c | 55 pteg[0] = 0; in kvmppc_mmu_invalidate_pte() 184 dprintk_mmu("KVM: %08x - %08x\n", pteg[0], pteg[1]); in kvmppc_mmu_map_page() 185 dprintk_mmu("KVM: %08x - %08x\n", pteg[2], pteg[3]); in kvmppc_mmu_map_page() 186 dprintk_mmu("KVM: %08x - %08x\n", pteg[4], pteg[5]); in kvmppc_mmu_map_page() 187 dprintk_mmu("KVM: %08x - %08x\n", pteg[6], pteg[7]); in kvmppc_mmu_map_page() 188 dprintk_mmu("KVM: %08x - %08x\n", pteg[8], pteg[9]); in kvmppc_mmu_map_page() 220 dprintk_mmu("KVM: %08x - %08x\n", pteg[0], pteg[1]); in kvmppc_mmu_map_page() 221 dprintk_mmu("KVM: %08x - %08x\n", pteg[2], pteg[3]); in kvmppc_mmu_map_page() 222 dprintk_mmu("KVM: %08x - %08x\n", pteg[4], pteg[5]); in kvmppc_mmu_map_page() 223 dprintk_mmu("KVM: %08x - %08x\n", pteg[6], pteg[7]); in kvmppc_mmu_map_page() [all …]
|
| H A D | book3s_pr_papr.c | 40 __be64 pteg[2 * 8]; in kvmppc_h_pr_enter() local 51 if (copy_from_user(pteg, (void __user *)pteg_addr, sizeof(pteg))) in kvmppc_h_pr_enter() 53 hpte = pteg; in kvmppc_h_pr_enter() 91 unsigned long v = 0, pteg, rb; in kvmppc_h_pr_remove() local 95 pteg = get_pteg_addr(vcpu, pte_index); in kvmppc_h_pr_remove() 98 if (copy_from_user(pte, (void __user *)pteg, sizeof(pte))) in kvmppc_h_pr_remove() 110 if (copy_to_user((void __user *)pteg, &v, sizeof(v))) in kvmppc_h_pr_remove() 155 unsigned long pteg, rb, flags; in kvmppc_h_pr_bulk_remove() local 178 pteg = get_pteg_addr(vcpu, tsh & H_BULK_REMOVE_PTEX); in kvmppc_h_pr_bulk_remove() 219 unsigned long rb, pteg, r, v; in kvmppc_h_pr_protect() local [all …]
|
| H A D | book3s_32_mmu.c | 98 u32 page, hash, pteg, htabmask; in kvmppc_mmu_book3s_32_get_pteg() local 109 pteg = (vcpu_book3s->sdr1 & 0xffff0000) | hash; in kvmppc_mmu_book3s_32_get_pteg() 112 kvmppc_get_pc(vcpu), eaddr, vcpu_book3s->sdr1, pteg, in kvmppc_mmu_book3s_32_get_pteg() 115 r = gfn_to_hva(vcpu->kvm, pteg >> PAGE_SHIFT); in kvmppc_mmu_book3s_32_get_pteg() 118 return r | (pteg & ~PAGE_MASK); in kvmppc_mmu_book3s_32_get_pteg() 188 u32 pteg[16]; in kvmppc_mmu_book3s_32_xlate_pte() local 209 if(copy_from_user(pteg, (void __user *)ptegp, sizeof(pteg))) { in kvmppc_mmu_book3s_32_xlate_pte() 216 pte0 = be32_to_cpu(pteg[i]); in kvmppc_mmu_book3s_32_xlate_pte() 217 pte1 = be32_to_cpu(pteg[i + 1]); in kvmppc_mmu_book3s_32_xlate_pte() 282 i, be32_to_cpu(pteg[i]), in kvmppc_mmu_book3s_32_xlate_pte() [all …]
|
| H A D | book3s_64_mmu.c | 121 u64 hash, pteg, htabsize; in kvmppc_mmu_book3s_64_get_pteg() local 138 pteg |= hash; in kvmppc_mmu_book3s_64_get_pteg() 141 page, vcpu_book3s->sdr1, pteg, slbe->vsid); in kvmppc_mmu_book3s_64_get_pteg() 146 r = pteg; in kvmppc_mmu_book3s_64_get_pteg() 148 r = gfn_to_hva(vcpu->kvm, pteg >> PAGE_SHIFT); in kvmppc_mmu_book3s_64_get_pteg() 152 return r | (pteg & ~PAGE_MASK); in kvmppc_mmu_book3s_64_get_pteg() 197 u64 pteg[16]; in kvmppc_mmu_book3s_64_xlate() local 251 if(copy_from_user(pteg, (void __user *)ptegp, sizeof(pteg))) { in kvmppc_mmu_book3s_64_xlate() 263 u64 pte0 = be64_to_cpu(pteg[i]); in kvmppc_mmu_book3s_64_xlate() 264 u64 pte1 = be64_to_cpu(pteg[i + 1]); in kvmppc_mmu_book3s_64_xlate() [all …]
|
| H A D | book3s_64_mmu_hv.c | 1209 unsigned long apsize, avpn, pteg, hash; in resize_hpt_rehash_hpte() local 1282 pteg = idx / HPTES_PER_GROUP; in resize_hpt_rehash_hpte() 1284 pteg = ~pteg; in resize_hpt_rehash_hpte() 1294 offset |= ((vsid ^ pteg) & old_hash_mask) << pshift; in resize_hpt_rehash_hpte() 1304 offset |= ((vsid ^ (vsid << 25) ^ pteg) & old_hash_mask) << pshift; in resize_hpt_rehash_hpte()
|
| /linux-6.15/arch/powerpc/mm/book3s64/ |
| H A D | hash_native.c | 663 unsigned long avpn, pteg, vpi; in hpte_decode() local 687 pteg = slot / HPTES_PER_GROUP; in hpte_decode() 689 pteg = ~pteg; in hpte_decode() 698 vpi = (vsid ^ pteg) & htab_hash_mask; in hpte_decode() 708 vpi = (vsid ^ (vsid << 25) ^ pteg) & htab_hash_mask; in hpte_decode()
|