1 /* memcontrol.h - Memory Controller 2 * 3 * Copyright IBM Corporation, 2007 4 * Author Balbir Singh <[email protected]> 5 * 6 * Copyright 2007 OpenVZ SWsoft Inc 7 * Author: Pavel Emelianov <[email protected]> 8 * 9 * This program is free software; you can redistribute it and/or modify 10 * it under the terms of the GNU General Public License as published by 11 * the Free Software Foundation; either version 2 of the License, or 12 * (at your option) any later version. 13 * 14 * This program is distributed in the hope that it will be useful, 15 * but WITHOUT ANY WARRANTY; without even the implied warranty of 16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 17 * GNU General Public License for more details. 18 */ 19 20 #ifndef _LINUX_MEMCONTROL_H 21 #define _LINUX_MEMCONTROL_H 22 23 struct mem_cgroup; 24 struct page_cgroup; 25 struct page; 26 struct mm_struct; 27 28 #ifdef CONFIG_CGROUP_MEM_CONT 29 30 extern void mm_init_cgroup(struct mm_struct *mm, struct task_struct *p); 31 extern void mm_free_cgroup(struct mm_struct *mm); 32 extern void page_assign_page_cgroup(struct page *page, 33 struct page_cgroup *pc); 34 extern struct page_cgroup *page_get_page_cgroup(struct page *page); 35 extern int mem_cgroup_charge(struct page *page, struct mm_struct *mm, 36 gfp_t gfp_mask); 37 extern void mem_cgroup_uncharge(struct page_cgroup *pc); 38 extern void mem_cgroup_uncharge_page(struct page *page); 39 extern void mem_cgroup_move_lists(struct page_cgroup *pc, bool active); 40 extern unsigned long mem_cgroup_isolate_pages(unsigned long nr_to_scan, 41 struct list_head *dst, 42 unsigned long *scanned, int order, 43 int mode, struct zone *z, 44 struct mem_cgroup *mem_cont, 45 int active); 46 extern void mem_cgroup_out_of_memory(struct mem_cgroup *mem, gfp_t gfp_mask); 47 extern int mem_cgroup_cache_charge(struct page *page, struct mm_struct *mm, 48 gfp_t gfp_mask); 49 int task_in_mem_cgroup(struct task_struct *task, const struct mem_cgroup *mem); 50 51 #define vm_match_cgroup(mm, cgroup) \ 52 ((cgroup) == rcu_dereference((mm)->mem_cgroup)) 53 54 extern int mem_cgroup_prepare_migration(struct page *page); 55 extern void mem_cgroup_end_migration(struct page *page); 56 extern void mem_cgroup_page_migration(struct page *page, struct page *newpage); 57 58 /* 59 * For memory reclaim. 60 */ 61 extern int mem_cgroup_calc_mapped_ratio(struct mem_cgroup *mem); 62 extern long mem_cgroup_reclaim_imbalance(struct mem_cgroup *mem); 63 64 extern int mem_cgroup_get_reclaim_priority(struct mem_cgroup *mem); 65 extern void mem_cgroup_note_reclaim_priority(struct mem_cgroup *mem, 66 int priority); 67 extern void mem_cgroup_record_reclaim_priority(struct mem_cgroup *mem, 68 int priority); 69 70 extern long mem_cgroup_calc_reclaim_active(struct mem_cgroup *mem, 71 struct zone *zone, int priority); 72 extern long mem_cgroup_calc_reclaim_inactive(struct mem_cgroup *mem, 73 struct zone *zone, int priority); 74 75 #else /* CONFIG_CGROUP_MEM_CONT */ 76 static inline void mm_init_cgroup(struct mm_struct *mm, 77 struct task_struct *p) 78 { 79 } 80 81 static inline void mm_free_cgroup(struct mm_struct *mm) 82 { 83 } 84 85 static inline void page_assign_page_cgroup(struct page *page, 86 struct page_cgroup *pc) 87 { 88 } 89 90 static inline struct page_cgroup *page_get_page_cgroup(struct page *page) 91 { 92 return NULL; 93 } 94 95 static inline int mem_cgroup_charge(struct page *page, struct mm_struct *mm, 96 gfp_t gfp_mask) 97 { 98 return 0; 99 } 100 101 static inline void mem_cgroup_uncharge(struct page_cgroup *pc) 102 { 103 } 104 105 static inline void mem_cgroup_uncharge_page(struct page *page) 106 { 107 } 108 109 static inline void mem_cgroup_move_lists(struct page_cgroup *pc, 110 bool active) 111 { 112 } 113 114 static inline int mem_cgroup_cache_charge(struct page *page, 115 struct mm_struct *mm, 116 gfp_t gfp_mask) 117 { 118 return 0; 119 } 120 121 static inline int vm_match_cgroup(struct mm_struct *mm, struct mem_cgroup *mem) 122 { 123 return 1; 124 } 125 126 static inline int task_in_mem_cgroup(struct task_struct *task, 127 const struct mem_cgroup *mem) 128 { 129 return 1; 130 } 131 132 static inline int mem_cgroup_prepare_migration(struct page *page) 133 { 134 return 0; 135 } 136 137 static inline void mem_cgroup_end_migration(struct page *page) 138 { 139 } 140 141 static inline void 142 mem_cgroup_page_migration(struct page *page, struct page *newpage) 143 { 144 } 145 146 static inline int mem_cgroup_calc_mapped_ratio(struct mem_cgroup *mem) 147 { 148 return 0; 149 } 150 151 static inline int mem_cgroup_reclaim_imbalance(struct mem_cgroup *mem) 152 { 153 return 0; 154 } 155 156 static inline int mem_cgroup_get_reclaim_priority(struct mem_cgroup *mem) 157 { 158 return 0; 159 } 160 161 static inline void mem_cgroup_note_reclaim_priority(struct mem_cgroup *mem, 162 int priority) 163 { 164 } 165 166 static inline void mem_cgroup_record_reclaim_priority(struct mem_cgroup *mem, 167 int priority) 168 { 169 } 170 171 static inline long mem_cgroup_calc_reclaim_active(struct mem_cgroup *mem, 172 struct zone *zone, int priority) 173 { 174 return 0; 175 } 176 177 static inline long mem_cgroup_calc_reclaim_inactive(struct mem_cgroup *mem, 178 struct zone *zone, int priority) 179 { 180 return 0; 181 } 182 #endif /* CONFIG_CGROUP_MEM_CONT */ 183 184 #endif /* _LINUX_MEMCONTROL_H */ 185 186