Home
last modified time | relevance | path

Searched refs:PMD_SIZE (Results 1 – 25 of 149) sorted by relevance

123456

/linux-6.15/arch/m68k/mm/
H A Dkmap.c50 #define IO_SIZE PMD_SIZE
85 virtaddr += PMD_SIZE; in __free_io_area()
86 size -= PMD_SIZE; in __free_io_area()
254 if (!(virtaddr & (PMD_SIZE-1))) in __ioremap()
269 physaddr += PMD_SIZE; in __ioremap()
270 virtaddr += PMD_SIZE; in __ioremap()
271 size -= PMD_SIZE; in __ioremap()
379 virtaddr += PMD_SIZE; in kernel_set_cachemode()
380 size -= PMD_SIZE; in kernel_set_cachemode()
/linux-6.15/arch/s390/mm/
H A Dvmem.c126 if (likely(IS_ALIGNED(unused_sub_pmd_start, PMD_SIZE))) in vmemmap_use_sub_pmd()
136 unsigned long page = ALIGN_DOWN(start, PMD_SIZE); in vmemmap_use_new_sub_pmd()
144 if (!IS_ALIGNED(start, PMD_SIZE)) in vmemmap_use_new_sub_pmd()
151 if (!IS_ALIGNED(end, PMD_SIZE)) in vmemmap_use_new_sub_pmd()
158 unsigned long page = ALIGN_DOWN(start, PMD_SIZE); in vmemmap_unuse_sub_pmd()
238 if (IS_ALIGNED(addr, PMD_SIZE) && in modify_pmd_table()
239 IS_ALIGNED(next, PMD_SIZE)) { in modify_pmd_table()
251 if (IS_ALIGNED(addr, PMD_SIZE) && in modify_pmd_table()
252 IS_ALIGNED(next, PMD_SIZE) && in modify_pmd_table()
271 if (!IS_ALIGNED(addr, PMD_SIZE) || in modify_pmd_table()
[all …]
/linux-6.15/arch/x86/mm/
H A Dinit.c358 unsigned long end = round_up(mr[i].end, PMD_SIZE); in adjust_range_page_size_mask()
424 end_pfn = PFN_DOWN(PMD_SIZE); in split_mem_range()
426 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
428 end_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
438 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
465 start_pfn = round_up(pfn, PFN_DOWN(PMD_SIZE)); in split_mem_range()
646 addr = memblock_phys_alloc_range(PMD_SIZE, PMD_SIZE, map_start, in memory_map_top_down()
652 memblock_phys_free(addr, PMD_SIZE); in memory_map_top_down()
653 real_end = addr + PMD_SIZE; in memory_map_top_down()
657 step_size = PMD_SIZE; in memory_map_top_down()
[all …]
H A Dinit_64.c379 for (; size; phys += PMD_SIZE, size -= PMD_SIZE) { in __init_extra_mapping()
442 for (; vaddr + PMD_SIZE - 1 < vaddr_end; pmd++, vaddr += PMD_SIZE) { in cleanup_highmap()
888 if (likely(IS_ALIGNED(end, PMD_SIZE))) in vmemmap_use_sub_pmd()
919 if (!IS_ALIGNED(start, PMD_SIZE)) in vmemmap_use_new_sub_pmd()
927 if (!IS_ALIGNED(end, PMD_SIZE)) in vmemmap_use_new_sub_pmd()
1140 if (IS_ALIGNED(addr, PMD_SIZE) && in remove_pmd_table()
1141 IS_ALIGNED(next, PMD_SIZE)) { in remove_pmd_table()
1525 addr_end = addr + PMD_SIZE; in vmemmap_set_pmd()
1526 p_end = p + PMD_SIZE; in vmemmap_set_pmd()
1528 if (!IS_ALIGNED(addr, PMD_SIZE) || in vmemmap_set_pmd()
[all …]
H A Dmem_encrypt_identity.c94 static char sme_workarea[2 * PMD_SIZE] __section(".init.scratch");
195 ppd->vaddr += PMD_SIZE; in __sme_map_range_pmd()
196 ppd->paddr += PMD_SIZE; in __sme_map_range_pmd()
222 ppd->vaddr_end = ALIGN(ppd->vaddr, PMD_SIZE); in __sme_map_range()
322 kernel_end = ALIGN((unsigned long)RIP_REL_REF(_end), PMD_SIZE); in sme_encrypt_kernel()
349 execute_end = execute_start + (PAGE_SIZE * 2) + PMD_SIZE; in sme_encrypt_kernel()
372 workarea_end = ALIGN(workarea_start + workarea_len, PMD_SIZE); in sme_encrypt_kernel()
/linux-6.15/arch/sh/include/asm/
H A Dpgtable-3level.h23 #define PMD_SIZE (1UL << PMD_SHIFT) macro
24 #define PMD_MASK (~(PMD_SIZE-1))
26 #define PTRS_PER_PMD ((1 << PGDIR_SHIFT) / PMD_SIZE)
/linux-6.15/arch/riscv/mm/
H A Dinit.c537 if (sz == PMD_SIZE) { in create_pmd_mapping()
755 !(pa & (PMD_SIZE - 1)) && !(va & (PMD_SIZE - 1)) && size >= PMD_SIZE) in best_map_size()
756 return PMD_SIZE; in best_map_size()
878 PMD_SIZE, PAGE_KERNEL_EXEC); in set_satp_mode()
880 set_satp_mode_pmd + PMD_SIZE, in set_satp_mode()
881 set_satp_mode_pmd + PMD_SIZE, in set_satp_mode()
882 PMD_SIZE, PAGE_KERNEL_EXEC); in set_satp_mode()
950 PMD_SIZE, PAGE_KERNEL); in create_kernel_page_table()
961 PMD_SIZE, in create_kernel_page_table()
987 pa, PMD_SIZE, PAGE_KERNEL); in create_fdt_early_page_table()
[all …]
H A Dhugetlbpage.c56 if (sz == PMD_SIZE) { in huge_pte_alloc()
112 if (sz == PMD_SIZE) in huge_pte_offset()
137 case PMD_SIZE: in hugetlb_mask_last_page()
138 return PUD_SIZE - PMD_SIZE; in hugetlb_mask_last_page()
140 return PMD_SIZE - napot_cont_size(NAPOT_CONT64KB_ORDER); in hugetlb_mask_last_page()
228 else if (sz >= PMD_SIZE) in num_contig_ptes_from_size()
H A Dtlbflush.c161 else if (stride_size >= PMD_SIZE) in flush_tlb_range()
162 stride_size = PMD_SIZE; in flush_tlb_range()
183 start, end - start, PMD_SIZE); in flush_pmd_tlb_range()
/linux-6.15/arch/parisc/kernel/
H A Dpci-dma.c85 if (end > PMD_SIZE) in map_pte_uncached()
86 end = PMD_SIZE; in map_pte_uncached()
120 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; in map_pmd_uncached()
121 orig_vaddr += PMD_SIZE; in map_pmd_uncached()
170 if (end > PMD_SIZE) in unmap_uncached_pte()
171 end = PMD_SIZE; in unmap_uncached_pte()
210 vaddr = (vaddr + PMD_SIZE) & PMD_MASK; in unmap_uncached_pmd()
211 orig_vaddr += PMD_SIZE; in unmap_uncached_pmd()
/linux-6.15/arch/x86/include/asm/
H A Dpgtable_32_types.h12 # define PMD_SIZE (1UL << PMD_SHIFT) macro
13 # define PMD_MASK (~(PMD_SIZE - 1))
/linux-6.15/arch/arm64/mm/
H A Dhugetlbpage.c60 case PMD_SIZE: in __hugetlb_valid_size()
95 *pgsize = PMD_SIZE; in find_num_contig()
109 *pgsize = PMD_SIZE; in num_contig_ptes()
274 } else if (sz == PMD_SIZE) { in huge_pte_alloc()
318 if (!(sz == PMD_SIZE || sz == CONT_PMD_SIZE) && in huge_pte_offset()
343 case PMD_SIZE: in hugetlb_mask_last_page()
344 return PUD_SIZE - PMD_SIZE; in hugetlb_mask_last_page()
346 return PMD_SIZE - CONT_PTE_SIZE; in hugetlb_mask_last_page()
367 case PMD_SIZE: in arch_make_huge_pte()
/linux-6.15/arch/powerpc/mm/book3s64/
H A Dradix_pgtable.c102 if (map_page_size == PMD_SIZE) { in early_map_kernel_page()
165 if (map_page_size == PMD_SIZE) { in __map_kernel_page()
329 } else if (IS_ALIGNED(addr, PMD_SIZE) && gap >= PMD_SIZE && in create_physical_mapping()
331 mapping_size = PMD_SIZE; in create_physical_mapping()
742 return !vmemmap_populated(start, PMD_SIZE); in vmemmap_pmd_is_unused()
835 if (IS_ALIGNED(addr, PMD_SIZE) && in remove_pmd_table()
836 IS_ALIGNED(next, PMD_SIZE)) { in remove_pmd_table()
1006 VM_BUG_ON(!IS_ALIGNED(addr, PMD_SIZE)); in vmemmap_set_pmd()
1162 if (!IS_ALIGNED(addr, PMD_SIZE) || (altmap && in radix__vmemmap_populate()
1330 addr_pfn += (PMD_SIZE >> PAGE_SHIFT); in vmemmap_populate_compound_pages()
[all …]
/linux-6.15/arch/loongarch/kvm/
H A Dmmu.c399 if (IS_ALIGNED(size, PMD_SIZE) && IS_ALIGNED(gpa_start, PMD_SIZE) in kvm_arch_prepare_memory_region()
400 && IS_ALIGNED(hva_start, PMD_SIZE)) in kvm_arch_prepare_memory_region()
427 gpa_offset = gpa_start & (PMD_SIZE - 1); in kvm_arch_prepare_memory_region()
428 hva_offset = hva_start & (PMD_SIZE - 1); in kvm_arch_prepare_memory_region()
433 gpa_offset = PMD_SIZE; in kvm_arch_prepare_memory_region()
434 if ((size + gpa_offset) < (PMD_SIZE * 2)) in kvm_arch_prepare_memory_region()
638 return (hva >= ALIGN(start, PMD_SIZE)) && (hva < ALIGN_DOWN(end, PMD_SIZE)); in fault_supports_huge_mapping()
/linux-6.15/arch/nios2/mm/
H A Dioremap.c33 if (end > PMD_SIZE) in remap_area_pte()
34 end = PMD_SIZE; in remap_area_pte()
70 address = (address + PMD_SIZE) & PMD_MASK; in remap_area_pmd()
/linux-6.15/include/asm-generic/
H A Dpgtable-nopmd.h22 #define PMD_SIZE (1UL << PMD_SHIFT) macro
23 #define PMD_MASK (~(PMD_SIZE-1))
/linux-6.15/arch/loongarch/include/asm/
H A Dpgtable.h29 #define PMD_SIZE (1UL << PMD_SHIFT) macro
30 #define PMD_MASK (~(PMD_SIZE-1))
34 #define PMD_SIZE (1UL << PMD_SHIFT) macro
35 #define PMD_MASK (~(PMD_SIZE-1))
96 …R_PUD * PTRS_PER_PMD * PTRS_PER_PTE * PAGE_SIZE, (1UL << cpu_vabits)) - PMD_SIZE - VMEMMAP_SIZE - …
100 …D * PTRS_PER_PMD * PTRS_PER_PTE * PAGE_SIZE, (1UL << cpu_vabits) / 2) - PMD_SIZE - VMEMMAP_SIZE - …
103 #define vmemmap ((struct page *)((VMALLOC_END + PMD_SIZE) & PMD_MASK))
/linux-6.15/arch/sparc/mm/
H A Dhugetlbpage.c188 if (sz >= PMD_SIZE) in huge_pte_alloc()
232 else if (size >= PMD_SIZE) in __set_huge_pte_at()
275 else if (size >= PMD_SIZE) in huge_ptep_get_and_clear()
394 addr += PMD_SIZE; in hugetlb_free_pgd_range()
404 end -= PMD_SIZE; in hugetlb_free_pgd_range()
/linux-6.15/arch/powerpc/include/asm/nohash/64/
H A Dpgtable-4k.h31 #define PMD_SIZE (1UL << PMD_SHIFT) macro
32 #define PMD_MASK (~(PMD_SIZE-1))
/linux-6.15/arch/arm64/include/asm/
H A Dkernel-pgtable.h21 #if defined(PMD_SIZE) && PMD_SIZE <= MIN_KIMG_ALIGN
H A Dhugetlb.h86 case PMD_SIZE: in flush_hugetlb_tlb_range()
87 __flush_tlb_range(vma, start, end, PMD_SIZE, false, 2); in flush_hugetlb_tlb_range()
/linux-6.15/arch/x86/kernel/
H A Dvmlinux.lds.S68 #define ALIGN_ENTRY_TEXT_BEGIN . = ALIGN(PMD_SIZE);
69 #define ALIGN_ENTRY_TEXT_END . = ALIGN(PMD_SIZE);
80 . = ALIGN(PMD_SIZE); \
85 . = ALIGN(PMD_SIZE); \
/linux-6.15/arch/x86/virt/svm/
H A Dsev.c177 if (IS_ALIGNED(pa, PMD_SIZE)) in __snp_fixup_e820_tables()
196 pa = ALIGN_DOWN(pa, PMD_SIZE); in __snp_fixup_e820_tables()
197 if (e820__mapped_any(pa, pa + PMD_SIZE, E820_TYPE_RAM)) { in __snp_fixup_e820_tables()
199 e820__range_update(pa, PMD_SIZE, E820_TYPE_RAM, E820_TYPE_RESERVED); in __snp_fixup_e820_tables()
200 e820__range_update_table(e820_table_kexec, pa, PMD_SIZE, E820_TYPE_RAM, E820_TYPE_RESERVED); in __snp_fixup_e820_tables()
201 if (!memblock_is_region_reserved(pa, PMD_SIZE)) in __snp_fixup_e820_tables()
202 memblock_reserve(pa, PMD_SIZE); in __snp_fixup_e820_tables()
/linux-6.15/arch/m68k/include/asm/
H A Dpgtable_mm.h39 #define PMD_SIZE (1UL << PMD_SHIFT) macro
40 #define PMD_MASK (~(PMD_SIZE-1))
/linux-6.15/arch/arm/mm/
H A Dmmu.c1114 next = (addr + PMD_SIZE - 1) & PMD_MASK; in fill_pmd_gaps()
1206 if (!IS_ALIGNED(block_start, PMD_SIZE)) { in adjust_lowmem_bounds()
1209 len = round_up(block_start, PMD_SIZE) - block_start; in adjust_lowmem_bounds()
1242 if (!IS_ALIGNED(block_start, PMD_SIZE)) in adjust_lowmem_bounds()
1244 else if (!IS_ALIGNED(block_end, PMD_SIZE)) in adjust_lowmem_bounds()
1263 memblock_limit = round_down(memblock_limit, PMD_SIZE); in adjust_lowmem_bounds()
1293 for (addr = 0; addr < KASAN_SHADOW_START; addr += PMD_SIZE) in prepare_page_table()
1304 for (addr = 0; addr < MODULES_VADDR; addr += PMD_SIZE) in prepare_page_table()
1310 addr = ((unsigned long)_exiprom + PMD_SIZE - 1) & PMD_MASK; in prepare_page_table()
1312 for ( ; addr < PAGE_OFFSET; addr += PMD_SIZE) in prepare_page_table()
[all …]

123456