Home
last modified time | relevance | path

Searched refs:kaddr (Results 1 – 25 of 185) sorted by relevance

12345678

/linux-6.15/arch/arm64/kernel/
H A Dparavirt.c38 struct pvclock_vcpu_stolen_time __rcu *kaddr; member
67 kaddr = rcu_dereference(reg->kaddr); in para_steal_clock()
68 if (!kaddr) { in para_steal_clock()
84 if (!reg->kaddr) in stolen_time_cpu_down_prepare()
87 kaddr = rcu_replace_pointer(reg->kaddr, NULL, true); in stolen_time_cpu_down_prepare()
89 memunmap(kaddr); in stolen_time_cpu_down_prepare()
107 kaddr = memremap(res.a0, in stolen_time_cpu_online()
111 rcu_assign_pointer(reg->kaddr, kaddr); in stolen_time_cpu_online()
113 if (!reg->kaddr) { in stolen_time_cpu_online()
118 if (le32_to_cpu(kaddr->revision) != 0 || in stolen_time_cpu_online()
[all …]
/linux-6.15/fs/nilfs2/
H A Ddir.c191 void *kaddr; in nilfs_get_folio() local
203 return kaddr; in nilfs_get_folio()
247 char *kaddr, *limit; in nilfs_readdir() local
252 if (IS_ERR(kaddr)) { in nilfs_readdir()
315 if (IS_ERR(kaddr)) in nilfs_find_entry()
451 if (IS_ERR(kaddr)) in nilfs_add_link()
573 void *kaddr; in nilfs_make_empty() local
598 kunmap_local(kaddr); in nilfs_make_empty()
611 char *kaddr; in nilfs_empty_dir() local
618 if (IS_ERR(kaddr)) in nilfs_empty_dir()
[all …]
/linux-6.15/fs/btrfs/
H A Daccessors.c31 token->kaddr = folio_address(eb->folios[0]); in btrfs_init_map_token()
74 ASSERT(token->kaddr); \
85 memcpy(lebytes, token->kaddr + oil, part); \
106 return get_unaligned_le##bits(kaddr + oil); \
108 memcpy(lebytes, kaddr + oil, part); \
109 kaddr = folio_address(eb->folios[idx + 1]); \
110 memcpy(lebytes + part, kaddr, size - part); \
128 ASSERT(token->kaddr); \
143 memcpy(token->kaddr + oil, lebytes, part); \
164 put_unaligned_le##bits(val, kaddr + oil); \
[all …]
H A Dlzo.c141 char *kaddr; in copy_compressed_data_to_page() local
161 kaddr = kmap_local_folio(cur_folio, 0); in copy_compressed_data_to_page()
173 kunmap_local(kaddr); in copy_compressed_data_to_page()
186 kaddr = kmap_local_folio(cur_folio, 0); in copy_compressed_data_to_page()
188 memcpy(kaddr + offset_in_page(*cur_out), in copy_compressed_data_to_page()
208 kunmap_local(kaddr); in copy_compressed_data_to_page()
336 char *kaddr; in lzo_decompress_bio() local
346 len_in = read_compress_length(kaddr); in lzo_decompress_bio()
347 kunmap_local(kaddr); in lzo_decompress_bio()
384 kaddr = kmap_local_folio(cur_folio, 0); in lzo_decompress_bio()
[all …]
/linux-6.15/fs/ext2/
H A Ddir.c194 void *kaddr; in ext2_get_folio() local
204 return kaddr; in ext2_get_folio()
281 if (IS_ERR(kaddr)) { in ext2_readdir()
286 return PTR_ERR(kaddr); in ext2_readdir()
362 if (IS_ERR(kaddr)) in ext2_find_entry()
500 if (IS_ERR(kaddr)) in ext2_add_link()
575 char *kaddr; in ext2_delete_entry() local
623 void *kaddr; in ext2_make_empty() local
648 kunmap_local(kaddr); in ext2_make_empty()
662 char *kaddr; in ext2_empty_dir() local
[all …]
/linux-6.15/arch/loongarch/include/asm/
H A Dpage.h73 struct page *dmw_virt_to_page(unsigned long kaddr);
74 struct page *tlb_virt_to_page(unsigned long kaddr);
82 #define virt_to_page(kaddr) phys_to_page(__pa(kaddr)) argument
94 #define virt_to_page(kaddr) \ argument
96 (likely((unsigned long)kaddr < vm_map_base)) ? \
97 dmw_virt_to_page((unsigned long)kaddr) : tlb_virt_to_page((unsigned long)kaddr);\
103 #define virt_to_pfn(kaddr) page_to_pfn(virt_to_page(kaddr)) argument
105 extern int __virt_addr_valid(volatile void *kaddr);
106 #define virt_addr_valid(kaddr) __virt_addr_valid((volatile void *)(kaddr)) argument
H A Dkfence.h20 char *kaddr, *vaddr; in arch_kfence_init_pool() local
40 kaddr = kfence_pool; in arch_kfence_init_pool()
42 while (kaddr < kfence_pool + KFENCE_POOL_SIZE) { in arch_kfence_init_pool()
43 set_page_address(virt_to_page(kaddr), vaddr); in arch_kfence_init_pool()
44 kaddr += PAGE_SIZE; in arch_kfence_init_pool()
H A Dio.h68 #define virt_to_phys(kaddr) \ argument
70 (likely((unsigned long)kaddr < vm_map_base)) ? __pa((unsigned long)kaddr) : \
71 page_to_phys(tlb_virt_to_page((unsigned long)kaddr)) + offset_in_page((unsigned long)kaddr);\
/linux-6.15/fs/ufs/
H A Ddir.c191 void *kaddr; in ufs_get_folio() local
201 return kaddr; in ufs_get_folio()
273 if (!IS_ERR(kaddr)) { in ufs_find_entry()
324 if (IS_ERR(kaddr)) in ufs_add_link()
325 return PTR_ERR(kaddr); in ufs_add_link()
434 if (IS_ERR(kaddr)) { in ufs_readdir()
487 char *kaddr; in ufs_delete_entry() local
540 char *kaddr; in ufs_make_empty() local
568 kunmap_local(kaddr); in ufs_make_empty()
584 char *kaddr; in ufs_empty_dir() local
[all …]
/linux-6.15/fs/minix/
H A Ddir.c102 char *p, *kaddr, *limit; in minix_readdir() local
106 if (IS_ERR(kaddr)) in minix_readdir()
108 p = kaddr+offset; in minix_readdir()
169 char *kaddr, *limit; in minix_find_entry() local
172 if (IS_ERR(kaddr)) in minix_find_entry()
209 char *kaddr, *p; in minix_add_link() local
226 if (IS_ERR(kaddr)) in minix_add_link()
314 char *kaddr; in minix_make_empty() local
345 kunmap_local(kaddr); in minix_make_empty()
362 char *name, *kaddr; in minix_empty_dir() local
[all …]
/linux-6.15/arch/m68k/sun3/
H A Ddvma.c23 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) in dvma_page() argument
29 j = *(volatile unsigned long *)kaddr; in dvma_page()
30 *(volatile unsigned long *)kaddr = j; in dvma_page()
32 ptep = pfn_pte(virt_to_pfn((void *)kaddr), PAGE_KERNEL); in dvma_page()
40 return (vaddr + (kaddr & ~PAGE_MASK)); in dvma_page()
44 int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, in dvma_map_iommu() argument
56 dvma_page(kaddr, vaddr); in dvma_map_iommu()
57 kaddr += PAGE_SIZE; in dvma_map_iommu()
H A Dsun3dvma.c262 unsigned long dvma_map_align(unsigned long kaddr, int len, int align) in dvma_map_align() argument
271 if(!kaddr || !len) { in dvma_map_align()
277 pr_debug("dvma_map request %08x bytes from %08lx\n", len, kaddr); in dvma_map_align()
278 off = kaddr & ~DVMA_PAGE_MASK; in dvma_map_align()
279 kaddr &= PAGE_MASK; in dvma_map_align()
291 if(!dvma_map_iommu(kaddr, baddr, len)) in dvma_map_align()
319 unsigned long kaddr; in dvma_malloc_align() local
333 free_pages(kaddr, get_order(len)); in dvma_malloc_align()
339 if(dvma_map_cpu(kaddr, vaddr, len) < 0) { in dvma_malloc_align()
341 free_pages(kaddr, get_order(len)); in dvma_malloc_align()
[all …]
/linux-6.15/drivers/acpi/
H A Dnvs.c78 void *kaddr; member
138 if (entry->kaddr) { in suspend_nvs_free()
140 iounmap(entry->kaddr); in suspend_nvs_free()
143 acpi_os_unmap_iomem(entry->kaddr, in suspend_nvs_free()
146 entry->kaddr = NULL; in suspend_nvs_free()
182 entry->kaddr = acpi_os_get_iomem(phys, size); in suspend_nvs_save()
183 if (!entry->kaddr) { in suspend_nvs_save()
184 entry->kaddr = acpi_os_ioremap(phys, size); in suspend_nvs_save()
185 entry->unmap = !!entry->kaddr; in suspend_nvs_save()
187 if (!entry->kaddr) { in suspend_nvs_save()
[all …]
/linux-6.15/fs/erofs/
H A Dxattr.c15 void *kaddr; member
85 if (IS_ERR(it.kaddr)) { in erofs_init_inode_xattrs()
86 ret = PTR_ERR(it.kaddr); in erofs_init_inode_xattrs()
90 ih = it.kaddr; in erofs_init_inode_xattrs()
106 if (IS_ERR(it.kaddr)) { in erofs_init_inode_xattrs()
187 if (IS_ERR(it->kaddr)) in erofs_xattr_copy_to_buffer()
190 src = it->kaddr; in erofs_xattr_copy_to_buffer()
290 if (IS_ERR(it->kaddr)) in erofs_getxattr_foreach()
297 it->kaddr, slice)) in erofs_getxattr_foreach()
334 if (IS_ERR(it->kaddr)) in erofs_xattr_iter_inline()
[all …]
/linux-6.15/arch/m68k/sun3x/
H A Ddvma.c78 inline int dvma_map_cpu(unsigned long kaddr, in dvma_map_cpu() argument
87 kaddr &= PAGE_MASK; in dvma_map_cpu()
92 pr_debug("dvma: mapping kern %08lx to virt %08lx\n", kaddr, vaddr); in dvma_map_cpu()
127 __pa(kaddr), vaddr); in dvma_map_cpu()
128 set_pte(pte, pfn_pte(virt_to_pfn((void *)kaddr), in dvma_map_cpu()
131 kaddr += PAGE_SIZE; in dvma_map_cpu()
146 int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, int len) in dvma_map_iommu() argument
162 dvma_entry_set(index, __pa(kaddr)); in dvma_map_iommu()
167 kaddr += DVMA_PAGE_SIZE; in dvma_map_iommu()
/linux-6.15/arch/arm/include/asm/
H A Dtlbflush.h477 tlb_op(TLB_V4_U_PAGE, "c8, c7, 1", kaddr); in __local_flush_tlb_kernel_page()
478 tlb_op(TLB_V4_D_PAGE, "c8, c6, 1", kaddr); in __local_flush_tlb_kernel_page()
479 tlb_op(TLB_V4_I_PAGE, "c8, c5, 1", kaddr); in __local_flush_tlb_kernel_page()
483 tlb_op(TLB_V6_U_PAGE, "c8, c7, 1", kaddr); in __local_flush_tlb_kernel_page()
484 tlb_op(TLB_V6_D_PAGE, "c8, c6, 1", kaddr); in __local_flush_tlb_kernel_page()
485 tlb_op(TLB_V6_I_PAGE, "c8, c5, 1", kaddr); in __local_flush_tlb_kernel_page()
492 kaddr &= PAGE_MASK; in local_flush_tlb_kernel_page()
497 __local_flush_tlb_kernel_page(kaddr); in local_flush_tlb_kernel_page()
498 tlb_op(TLB_V7_UIS_PAGE, "c8, c7, 1", kaddr); in local_flush_tlb_kernel_page()
510 kaddr &= PAGE_MASK; in __flush_tlb_kernel_page()
[all …]
H A Dmemory.h292 unsigned long kaddr = (unsigned long)p; in virt_to_pfn() local
293 return (((kaddr - PAGE_OFFSET) >> PAGE_SHIFT) + in virt_to_pfn()
384 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument
385 #define virt_addr_valid(kaddr) (((unsigned long)(kaddr) >= PAGE_OFFSET && (unsigned long)(kaddr) < … argument
386 && pfn_valid(virt_to_pfn(kaddr)))
/linux-6.15/include/linux/
H A Dhighmem.h239 void *kaddr = kmap_local_page(page); in clear_highpage() local
240 clear_page(kaddr); in clear_highpage()
241 kunmap_local(kaddr); in clear_highpage()
248 clear_page(kasan_reset_tag(kaddr)); in clear_highpage_kasan_tagged()
249 kunmap_local(kaddr); in clear_highpage_kasan_tagged()
283 kunmap_local(kaddr); in zero_user_segments()
521 size_t offset, void *kaddr) in folio_zero_tail() argument
529 memset(kaddr, 0, max); in folio_zero_tail()
530 kunmap_local(kaddr); in folio_zero_tail()
538 memset(kaddr, 0, len); in folio_zero_tail()
[all …]
/linux-6.15/arch/hexagon/include/asm/
H A Dpage.h90 #define virt_to_page(kaddr) pfn_to_page(PFN_DOWN(__pa(kaddr))) argument
95 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) argument
119 static inline unsigned long virt_to_pfn(const void *kaddr) in virt_to_pfn() argument
121 return __pa(kaddr) >> PAGE_SHIFT; in virt_to_pfn()
/linux-6.15/arch/arc/include/asm/
H A Dpage.h123 static inline unsigned long virt_to_pfn(const void *kaddr) in virt_to_pfn() argument
125 return __pa(kaddr) >> PAGE_SHIFT; in virt_to_pfn()
128 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument
129 #define virt_addr_valid(kaddr) pfn_valid(virt_to_pfn(kaddr)) argument
/linux-6.15/arch/m68k/include/asm/
H A Dpage_no.h22 static inline unsigned long virt_to_pfn(const void *kaddr) in virt_to_pfn() argument
24 return __pa(kaddr) >> PAGE_SHIFT; in virt_to_pfn()
35 #define virt_addr_valid(kaddr) (((unsigned long)(kaddr) >= PAGE_OFFSET) && \ argument
36 ((unsigned long)(kaddr) < memory_end))
/linux-6.15/tools/testing/nvdimm/
H A Dpmem-dax.c12 long nr_pages, enum dax_access_mode mode, void **kaddr, in __pmem_direct_access() argument
28 if (kaddr) in __pmem_direct_access()
29 *kaddr = pmem->virt_addr + offset; in __pmem_direct_access()
39 if (kaddr) in __pmem_direct_access()
40 *kaddr = pmem->virt_addr + offset; in __pmem_direct_access()
/linux-6.15/arch/mips/include/asm/
H A Dpage.h208 #define virt_to_pfn(kaddr) PFN_DOWN(virt_to_phys((void *)(kaddr))) argument
209 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument
211 extern bool __virt_addr_valid(const volatile void *kaddr);
212 #define virt_addr_valid(kaddr) \ argument
213 __virt_addr_valid((const volatile void *) (kaddr))
/linux-6.15/arch/x86/include/asm/
H A Dpage.h68 #define virt_to_page(kaddr) pfn_to_page(__pa(kaddr) >> PAGE_SHIFT) argument
69 extern bool __virt_addr_valid(unsigned long kaddr);
70 #define virt_addr_valid(kaddr) __virt_addr_valid((unsigned long) (kaddr)) argument
/linux-6.15/arch/csky/include/asm/
H A Dpage.h33 #define virt_addr_valid(kaddr) ((void *)(kaddr) >= (void *)PAGE_OFFSET && \ argument
34 (void *)(kaddr) < high_memory)
73 static inline unsigned long virt_to_pfn(const void *kaddr) in virt_to_pfn() argument
75 return __pa(kaddr) >> PAGE_SHIFT; in virt_to_pfn()

12345678