Commit | Line | Data |
---|---|---|
208d54e5 DH |
1 | #ifndef __LINUX_MEMORY_HOTPLUG_H |
2 | #define __LINUX_MEMORY_HOTPLUG_H | |
3 | ||
4 | #include <linux/mmzone.h> | |
5 | #include <linux/spinlock.h> | |
3947be19 | 6 | #include <linux/notifier.h> |
187f1882 | 7 | #include <linux/bug.h> |
208d54e5 | 8 | |
78679302 KH |
9 | struct page; |
10 | struct zone; | |
11 | struct pglist_data; | |
ea01ea93 | 12 | struct mem_section; |
e90bdb7f | 13 | struct memory_block; |
78679302 | 14 | |
208d54e5 | 15 | #ifdef CONFIG_MEMORY_HOTPLUG |
04753278 YG |
16 | |
17 | /* | |
5f24ce5f AA |
18 | * Types for free bootmem stored in page->lru.next. These have to be in |
19 | * some random range in unsigned long space for debugging purposes. | |
04753278 | 20 | */ |
5f24ce5f AA |
21 | enum { |
22 | MEMORY_HOTPLUG_MIN_BOOTMEM_TYPE = 12, | |
23 | SECTION_INFO = MEMORY_HOTPLUG_MIN_BOOTMEM_TYPE, | |
24 | MIX_SECTION_INFO, | |
25 | NODE_INFO, | |
26 | MEMORY_HOTPLUG_MAX_BOOTMEM_TYPE = NODE_INFO, | |
27 | }; | |
04753278 | 28 | |
4f7c6b49 | 29 | /* Types for control the zone type of onlined and offlined memory */ |
511c2aba | 30 | enum { |
4f7c6b49 TC |
31 | MMOP_OFFLINE = -1, |
32 | MMOP_ONLINE_KEEP, | |
33 | MMOP_ONLINE_KERNEL, | |
34 | MMOP_ONLINE_MOVABLE, | |
511c2aba LJ |
35 | }; |
36 | ||
208d54e5 DH |
37 | /* |
38 | * pgdat resizing functions | |
39 | */ | |
40 | static inline | |
41 | void pgdat_resize_lock(struct pglist_data *pgdat, unsigned long *flags) | |
42 | { | |
43 | spin_lock_irqsave(&pgdat->node_size_lock, *flags); | |
44 | } | |
45 | static inline | |
46 | void pgdat_resize_unlock(struct pglist_data *pgdat, unsigned long *flags) | |
47 | { | |
bdc8cb98 | 48 | spin_unlock_irqrestore(&pgdat->node_size_lock, *flags); |
208d54e5 DH |
49 | } |
50 | static inline | |
51 | void pgdat_resize_init(struct pglist_data *pgdat) | |
52 | { | |
53 | spin_lock_init(&pgdat->node_size_lock); | |
54 | } | |
bdc8cb98 DH |
55 | /* |
56 | * Zone resizing functions | |
511c2aba LJ |
57 | * |
58 | * Note: any attempt to resize a zone should has pgdat_resize_lock() | |
59 | * zone_span_writelock() both held. This ensure the size of a zone | |
60 | * can't be changed while pgdat_resize_lock() held. | |
bdc8cb98 DH |
61 | */ |
62 | static inline unsigned zone_span_seqbegin(struct zone *zone) | |
63 | { | |
64 | return read_seqbegin(&zone->span_seqlock); | |
65 | } | |
66 | static inline int zone_span_seqretry(struct zone *zone, unsigned iv) | |
67 | { | |
68 | return read_seqretry(&zone->span_seqlock, iv); | |
69 | } | |
70 | static inline void zone_span_writelock(struct zone *zone) | |
71 | { | |
72 | write_seqlock(&zone->span_seqlock); | |
73 | } | |
74 | static inline void zone_span_writeunlock(struct zone *zone) | |
75 | { | |
76 | write_sequnlock(&zone->span_seqlock); | |
77 | } | |
78 | static inline void zone_seqlock_init(struct zone *zone) | |
79 | { | |
80 | seqlock_init(&zone->span_seqlock); | |
81 | } | |
3947be19 DH |
82 | extern int zone_grow_free_lists(struct zone *zone, unsigned long new_nr_pages); |
83 | extern int zone_grow_waitqueues(struct zone *zone, unsigned long nr_pages); | |
84 | extern int add_one_highpage(struct page *page, int pfn, int bad_ppro); | |
3947be19 | 85 | /* VM interface that may be used by firmware interface */ |
511c2aba | 86 | extern int online_pages(unsigned long, unsigned long, int); |
ed2f2400 | 87 | extern int test_pages_in_a_zone(unsigned long, unsigned long); |
0c0e6195 | 88 | extern void __offline_isolated_pages(unsigned long, unsigned long); |
48e94196 | 89 | |
9d0ad8ca DK |
90 | typedef void (*online_page_callback_t)(struct page *page); |
91 | ||
92 | extern int set_online_page_callback(online_page_callback_t callback); | |
93 | extern int restore_online_page_callback(online_page_callback_t callback); | |
94 | ||
95 | extern void __online_page_set_limits(struct page *page); | |
96 | extern void __online_page_increment_counters(struct page *page); | |
97 | extern void __online_page_free(struct page *page); | |
98 | ||
01b0f197 TK |
99 | extern int try_online_node(int nid); |
100 | ||
49ac8255 KH |
101 | #ifdef CONFIG_MEMORY_HOTREMOVE |
102 | extern bool is_pageblock_removable_nolock(struct page *page); | |
24d335ca | 103 | extern int arch_remove_memory(u64 start, u64 size); |
4edd7cef DR |
104 | extern int __remove_pages(struct zone *zone, unsigned long start_pfn, |
105 | unsigned long nr_pages); | |
49ac8255 KH |
106 | #endif /* CONFIG_MEMORY_HOTREMOVE */ |
107 | ||
3947be19 | 108 | /* reasonably generic interface to expand the physical pages in a zone */ |
c04fc586 | 109 | extern int __add_pages(int nid, struct zone *zone, unsigned long start_pfn, |
3947be19 | 110 | unsigned long nr_pages); |
bc02af93 YG |
111 | |
112 | #ifdef CONFIG_NUMA | |
113 | extern int memory_add_physaddr_to_nid(u64 start); | |
114 | #else | |
115 | static inline int memory_add_physaddr_to_nid(u64 start) | |
116 | { | |
117 | return 0; | |
118 | } | |
119 | #endif | |
120 | ||
306d6cbe YG |
121 | #ifdef CONFIG_HAVE_ARCH_NODEDATA_EXTENSION |
122 | /* | |
123 | * For supporting node-hotadd, we have to allocate a new pgdat. | |
124 | * | |
125 | * If an arch has generic style NODE_DATA(), | |
126 | * node_data[nid] = kzalloc() works well. But it depends on the architecture. | |
127 | * | |
128 | * In general, generic_alloc_nodedata() is used. | |
129 | * Now, arch_free_nodedata() is just defined for error path of node_hot_add. | |
130 | * | |
131 | */ | |
dd0932d9 YG |
132 | extern pg_data_t *arch_alloc_nodedata(int nid); |
133 | extern void arch_free_nodedata(pg_data_t *pgdat); | |
7049027c | 134 | extern void arch_refresh_nodedata(int nid, pg_data_t *pgdat); |
306d6cbe YG |
135 | |
136 | #else /* CONFIG_HAVE_ARCH_NODEDATA_EXTENSION */ | |
137 | ||
138 | #define arch_alloc_nodedata(nid) generic_alloc_nodedata(nid) | |
139 | #define arch_free_nodedata(pgdat) generic_free_nodedata(pgdat) | |
140 | ||
141 | #ifdef CONFIG_NUMA | |
142 | /* | |
143 | * If ARCH_HAS_NODEDATA_EXTENSION=n, this func is used to allocate pgdat. | |
144 | * XXX: kmalloc_node() can't work well to get new node's memory at this time. | |
145 | * Because, pgdat for the new node is not allocated/initialized yet itself. | |
146 | * To use new node's memory, more consideration will be necessary. | |
147 | */ | |
148 | #define generic_alloc_nodedata(nid) \ | |
149 | ({ \ | |
150 | kzalloc(sizeof(pg_data_t), GFP_KERNEL); \ | |
151 | }) | |
152 | /* | |
153 | * This definition is just for error path in node hotadd. | |
154 | * For node hotremove, we have to replace this. | |
155 | */ | |
156 | #define generic_free_nodedata(pgdat) kfree(pgdat) | |
157 | ||
10ad400b YG |
158 | extern pg_data_t *node_data[]; |
159 | static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) | |
160 | { | |
161 | node_data[nid] = pgdat; | |
162 | } | |
163 | ||
306d6cbe YG |
164 | #else /* !CONFIG_NUMA */ |
165 | ||
166 | /* never called */ | |
167 | static inline pg_data_t *generic_alloc_nodedata(int nid) | |
168 | { | |
169 | BUG(); | |
170 | return NULL; | |
171 | } | |
172 | static inline void generic_free_nodedata(pg_data_t *pgdat) | |
173 | { | |
174 | } | |
10ad400b YG |
175 | static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) |
176 | { | |
177 | } | |
306d6cbe YG |
178 | #endif /* CONFIG_NUMA */ |
179 | #endif /* CONFIG_HAVE_ARCH_NODEDATA_EXTENSION */ | |
180 | ||
46723bfa YI |
181 | #ifdef CONFIG_HAVE_BOOTMEM_INFO_NODE |
182 | extern void register_page_bootmem_info_node(struct pglist_data *pgdat); | |
183 | #else | |
04753278 YG |
184 | static inline void register_page_bootmem_info_node(struct pglist_data *pgdat) |
185 | { | |
186 | } | |
04753278 | 187 | #endif |
46723bfa YI |
188 | extern void put_page_bootmem(struct page *page); |
189 | extern void get_page_bootmem(unsigned long ingo, struct page *page, | |
190 | unsigned long type); | |
04753278 | 191 | |
bfc8c901 VD |
192 | void get_online_mems(void); |
193 | void put_online_mems(void); | |
20d6c96b | 194 | |
30467e0b DR |
195 | void mem_hotplug_begin(void); |
196 | void mem_hotplug_done(void); | |
197 | ||
208d54e5 DH |
198 | #else /* ! CONFIG_MEMORY_HOTPLUG */ |
199 | /* | |
200 | * Stub functions for when hotplug is off | |
201 | */ | |
202 | static inline void pgdat_resize_lock(struct pglist_data *p, unsigned long *f) {} | |
203 | static inline void pgdat_resize_unlock(struct pglist_data *p, unsigned long *f) {} | |
204 | static inline void pgdat_resize_init(struct pglist_data *pgdat) {} | |
bdc8cb98 DH |
205 | |
206 | static inline unsigned zone_span_seqbegin(struct zone *zone) | |
207 | { | |
208 | return 0; | |
209 | } | |
210 | static inline int zone_span_seqretry(struct zone *zone, unsigned iv) | |
211 | { | |
212 | return 0; | |
213 | } | |
214 | static inline void zone_span_writelock(struct zone *zone) {} | |
215 | static inline void zone_span_writeunlock(struct zone *zone) {} | |
216 | static inline void zone_seqlock_init(struct zone *zone) {} | |
3947be19 DH |
217 | |
218 | static inline int mhp_notimplemented(const char *func) | |
219 | { | |
220 | printk(KERN_WARNING "%s() called, with CONFIG_MEMORY_HOTPLUG disabled\n", func); | |
221 | dump_stack(); | |
222 | return -ENOSYS; | |
223 | } | |
224 | ||
04753278 YG |
225 | static inline void register_page_bootmem_info_node(struct pglist_data *pgdat) |
226 | { | |
227 | } | |
228 | ||
01b0f197 TK |
229 | static inline int try_online_node(int nid) |
230 | { | |
231 | return 0; | |
232 | } | |
233 | ||
bfc8c901 VD |
234 | static inline void get_online_mems(void) {} |
235 | static inline void put_online_mems(void) {} | |
20d6c96b | 236 | |
30467e0b DR |
237 | static inline void mem_hotplug_begin(void) {} |
238 | static inline void mem_hotplug_done(void) {} | |
239 | ||
bdc8cb98 | 240 | #endif /* ! CONFIG_MEMORY_HOTPLUG */ |
9d99aaa3 | 241 | |
5c755e9f BP |
242 | #ifdef CONFIG_MEMORY_HOTREMOVE |
243 | ||
244 | extern int is_mem_section_removable(unsigned long pfn, unsigned long nr_pages); | |
90b30cdc | 245 | extern void try_offline_node(int nid); |
aba6efc4 RW |
246 | extern int offline_pages(unsigned long start_pfn, unsigned long nr_pages); |
247 | extern void remove_memory(int nid, u64 start, u64 size); | |
5c755e9f BP |
248 | |
249 | #else | |
250 | static inline int is_mem_section_removable(unsigned long pfn, | |
251 | unsigned long nr_pages) | |
252 | { | |
253 | return 0; | |
254 | } | |
90b30cdc WC |
255 | |
256 | static inline void try_offline_node(int nid) {} | |
aba6efc4 RW |
257 | |
258 | static inline int offline_pages(unsigned long start_pfn, unsigned long nr_pages) | |
259 | { | |
260 | return -EINVAL; | |
261 | } | |
262 | ||
263 | static inline void remove_memory(int nid, u64 start, u64 size) {} | |
5c755e9f BP |
264 | #endif /* CONFIG_MEMORY_HOTREMOVE */ |
265 | ||
e2ff3940 RW |
266 | extern int walk_memory_range(unsigned long start_pfn, unsigned long end_pfn, |
267 | void *arg, int (*func)(struct memory_block *, void *)); | |
bc02af93 | 268 | extern int add_memory(int nid, u64 start, u64 size); |
033fbae9 DW |
269 | extern int zone_for_memory(int nid, u64 start, u64 size, int zone_default, |
270 | bool for_device); | |
271 | extern int arch_add_memory(int nid, u64 start, u64 size, bool for_device); | |
a16cee10 | 272 | extern int offline_pages(unsigned long start_pfn, unsigned long nr_pages); |
6677e3ea | 273 | extern bool is_memblock_offlined(struct memory_block *mem); |
242831eb | 274 | extern void remove_memory(int nid, u64 start, u64 size); |
85b35fea | 275 | extern int sparse_add_one_section(struct zone *zone, unsigned long start_pfn); |
ea01ea93 | 276 | extern void sparse_remove_one_section(struct zone *zone, struct mem_section *ms); |
04753278 YG |
277 | extern struct page *sparse_decode_mem_map(unsigned long coded_mem_map, |
278 | unsigned long pnum); | |
9d99aaa3 | 279 | |
208d54e5 | 280 | #endif /* __LINUX_MEMORY_HOTPLUG_H */ |