1 #ifndef _LINUX_MEMBLOCK_H 2 #define _LINUX_MEMBLOCK_H 3 #ifdef __KERNEL__ 4 5 #define MEMBLOCK_ERROR 0 6 7 #ifdef CONFIG_HAVE_MEMBLOCK 8 /* 9 * Logical memory blocks. 10 * 11 * Copyright (C) 2001 Peter Bergner, IBM Corp. 12 * 13 * This program is free software; you can redistribute it and/or 14 * modify it under the terms of the GNU General Public License 15 * as published by the Free Software Foundation; either version 16 * 2 of the License, or (at your option) any later version. 17 */ 18 19 #include <linux/init.h> 20 #include <linux/mm.h> 21 22 #include <asm/memblock.h> 23 24 #define INIT_MEMBLOCK_REGIONS 128 25 26 struct memblock_region { 27 phys_addr_t base; 28 phys_addr_t size; 29 }; 30 31 struct memblock_type { 32 unsigned long cnt; /* number of regions */ 33 unsigned long max; /* size of the allocated array */ 34 struct memblock_region *regions; 35 }; 36 37 struct memblock { 38 phys_addr_t current_limit; 39 phys_addr_t memory_size; /* Updated by memblock_analyze() */ 40 struct memblock_type memory; 41 struct memblock_type reserved; 42 }; 43 44 extern struct memblock memblock; 45 extern int memblock_debug; 46 extern int memblock_can_resize; 47 48 #define memblock_dbg(fmt, ...) \ 49 if (memblock_debug) printk(KERN_INFO pr_fmt(fmt), ##__VA_ARGS__) 50 51 u64 memblock_find_in_range(u64 start, u64 end, u64 size, u64 align); 52 int memblock_free_reserved_regions(void); 53 int memblock_reserve_reserved_regions(void); 54 55 extern void memblock_init(void); 56 extern void memblock_analyze(void); 57 extern long memblock_add(phys_addr_t base, phys_addr_t size); 58 extern long memblock_remove(phys_addr_t base, phys_addr_t size); 59 extern long memblock_free(phys_addr_t base, phys_addr_t size); 60 extern long memblock_reserve(phys_addr_t base, phys_addr_t size); 61 62 /* The numa aware allocator is only available if 63 * CONFIG_ARCH_POPULATES_NODE_MAP is set 64 */ 65 extern phys_addr_t memblock_alloc_nid(phys_addr_t size, phys_addr_t align, 66 int nid); 67 extern phys_addr_t memblock_alloc_try_nid(phys_addr_t size, phys_addr_t align, 68 int nid); 69 70 extern phys_addr_t memblock_alloc(phys_addr_t size, phys_addr_t align); 71 72 /* Flags for memblock_alloc_base() amd __memblock_alloc_base() */ 73 #define MEMBLOCK_ALLOC_ANYWHERE (~(phys_addr_t)0) 74 #define MEMBLOCK_ALLOC_ACCESSIBLE 0 75 76 extern phys_addr_t memblock_alloc_base(phys_addr_t size, 77 phys_addr_t align, 78 phys_addr_t max_addr); 79 extern phys_addr_t __memblock_alloc_base(phys_addr_t size, 80 phys_addr_t align, 81 phys_addr_t max_addr); 82 extern phys_addr_t memblock_phys_mem_size(void); 83 extern phys_addr_t memblock_end_of_DRAM(void); 84 extern void memblock_enforce_memory_limit(phys_addr_t memory_limit); 85 extern int memblock_is_memory(phys_addr_t addr); 86 extern int memblock_is_region_memory(phys_addr_t base, phys_addr_t size); 87 extern int memblock_is_reserved(phys_addr_t addr); 88 extern int memblock_is_region_reserved(phys_addr_t base, phys_addr_t size); 89 90 extern void memblock_dump_all(void); 91 92 /* Provided by the architecture */ 93 extern phys_addr_t memblock_nid_range(phys_addr_t start, phys_addr_t end, int *nid); 94 extern int memblock_memory_can_coalesce(phys_addr_t addr1, phys_addr_t size1, 95 phys_addr_t addr2, phys_addr_t size2); 96 97 /** 98 * memblock_set_current_limit - Set the current allocation limit to allow 99 * limiting allocations to what is currently 100 * accessible during boot 101 * @limit: New limit value (physical address) 102 */ 103 extern void memblock_set_current_limit(phys_addr_t limit); 104 105 106 /* 107 * pfn conversion functions 108 * 109 * While the memory MEMBLOCKs should always be page aligned, the reserved 110 * MEMBLOCKs may not be. This accessor attempt to provide a very clear 111 * idea of what they return for such non aligned MEMBLOCKs. 112 */ 113 114 /** 115 * memblock_region_memory_base_pfn - Return the lowest pfn intersecting with the memory region 116 * @reg: memblock_region structure 117 */ 118 static inline unsigned long memblock_region_memory_base_pfn(const struct memblock_region *reg) 119 { 120 return PFN_UP(reg->base); 121 } 122 123 /** 124 * memblock_region_memory_end_pfn - Return the end_pfn this region 125 * @reg: memblock_region structure 126 */ 127 static inline unsigned long memblock_region_memory_end_pfn(const struct memblock_region *reg) 128 { 129 return PFN_DOWN(reg->base + reg->size); 130 } 131 132 /** 133 * memblock_region_reserved_base_pfn - Return the lowest pfn intersecting with the reserved region 134 * @reg: memblock_region structure 135 */ 136 static inline unsigned long memblock_region_reserved_base_pfn(const struct memblock_region *reg) 137 { 138 return PFN_DOWN(reg->base); 139 } 140 141 /** 142 * memblock_region_reserved_end_pfn - Return the end_pfn this region 143 * @reg: memblock_region structure 144 */ 145 static inline unsigned long memblock_region_reserved_end_pfn(const struct memblock_region *reg) 146 { 147 return PFN_UP(reg->base + reg->size); 148 } 149 150 #define for_each_memblock(memblock_type, region) \ 151 for (region = memblock.memblock_type.regions; \ 152 region < (memblock.memblock_type.regions + memblock.memblock_type.cnt); \ 153 region++) 154 155 156 #ifdef ARCH_DISCARD_MEMBLOCK 157 #define __init_memblock __init 158 #define __initdata_memblock __initdata 159 #else 160 #define __init_memblock 161 #define __initdata_memblock 162 #endif 163 164 #else 165 static inline phys_addr_t memblock_alloc(phys_addr_t size, phys_addr_t align) 166 { 167 return MEMBLOCK_ERROR; 168 } 169 170 #endif /* CONFIG_HAVE_MEMBLOCK */ 171 172 #endif /* __KERNEL__ */ 173 174 #endif /* _LINUX_MEMBLOCK_H */ 175