Blame view

mm/nobootmem.c 11 KB
093258732   Yinghai Lu   bootmem: Separate...
1
2
3
4
5
6
7
8
9
10
11
12
13
14
  /*
   *  bootmem - A boot-time physical memory allocator and configurator
   *
   *  Copyright (C) 1999 Ingo Molnar
   *                1999 Kanoj Sarcar, SGI
   *                2008 Johannes Weiner
   *
   * Access to this subsystem has to be serialized externally (which is true
   * for the boot process anyway).
   */
  #include <linux/init.h>
  #include <linux/pfn.h>
  #include <linux/slab.h>
  #include <linux/bootmem.h>
b95f1b31b   Paul Gortmaker   mm: Map most file...
15
  #include <linux/export.h>
093258732   Yinghai Lu   bootmem: Separate...
16
17
18
19
20
21
22
23
24
  #include <linux/kmemleak.h>
  #include <linux/range.h>
  #include <linux/memblock.h>
  
  #include <asm/bug.h>
  #include <asm/io.h>
  #include <asm/processor.h>
  
  #include "internal.h"
e782ab421   Yinghai Lu   bootmem: Move con...
25
26
27
28
  #ifndef CONFIG_NEED_MULTIPLE_NODES
  struct pglist_data __refdata contig_page_data;
  EXPORT_SYMBOL(contig_page_data);
  #endif
093258732   Yinghai Lu   bootmem: Separate...
29
30
31
  unsigned long max_low_pfn;
  unsigned long min_low_pfn;
  unsigned long max_pfn;
8bc1f91e1   Yinghai Lu   bootmem: Move __a...
32
33
34
35
36
37
38
39
  static void * __init __alloc_memory_core_early(int nid, u64 size, u64 align,
  					u64 goal, u64 limit)
  {
  	void *ptr;
  	u64 addr;
  
  	if (limit > memblock.current_limit)
  		limit = memblock.current_limit;
eb40c4c27   Tejun Heo   memblock, x86: Re...
40
  	addr = memblock_find_in_range_node(goal, limit, size, align, nid);
1f5026a7e   Tejun Heo   memblock: Kill ME...
41
  	if (!addr)
8bc1f91e1   Yinghai Lu   bootmem: Move __a...
42
  		return NULL;
b476e2951   Joonsoo Kim   mm, nobootmem: do...
43
  	memblock_reserve(addr, size);
8bc1f91e1   Yinghai Lu   bootmem: Move __a...
44
45
  	ptr = phys_to_virt(addr);
  	memset(ptr, 0, size);
8bc1f91e1   Yinghai Lu   bootmem: Move __a...
46
47
48
49
50
51
52
  	/*
  	 * The min_count is set to 0 so that bootmem allocated blocks
  	 * are never reported as leaks.
  	 */
  	kmemleak_alloc(ptr, size, 0, 0);
  	return ptr;
  }
093258732   Yinghai Lu   bootmem: Separate...
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
  /*
   * free_bootmem_late - free bootmem pages directly to page allocator
   * @addr: starting address of the range
   * @size: size of the range in bytes
   *
   * This is only useful when the bootmem allocator has already been torn
   * down, but we are still initializing the system.  Pages are given directly
   * to the page allocator, no bootmem metadata is updated because it is gone.
   */
  void __init free_bootmem_late(unsigned long addr, unsigned long size)
  {
  	unsigned long cursor, end;
  
  	kmemleak_free_part(__va(addr), size);
  
  	cursor = PFN_UP(addr);
  	end = PFN_DOWN(addr + size);
  
  	for (; cursor < end; cursor++) {
  		__free_pages_bootmem(pfn_to_page(cursor), 0);
  		totalram_pages++;
  	}
  }
  
  static void __init __free_pages_memory(unsigned long start, unsigned long end)
  {
6bc2e853c   Russ Anderson   mm: nobootmem: fi...
79
  	unsigned long i, start_aligned, end_aligned;
093258732   Yinghai Lu   bootmem: Separate...
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
  	int order = ilog2(BITS_PER_LONG);
  
  	start_aligned = (start + (BITS_PER_LONG - 1)) & ~(BITS_PER_LONG - 1);
  	end_aligned = end & ~(BITS_PER_LONG - 1);
  
  	if (end_aligned <= start_aligned) {
  		for (i = start; i < end; i++)
  			__free_pages_bootmem(pfn_to_page(i), 0);
  
  		return;
  	}
  
  	for (i = start; i < start_aligned; i++)
  		__free_pages_bootmem(pfn_to_page(i), 0);
  
  	for (i = start_aligned; i < end_aligned; i += BITS_PER_LONG)
  		__free_pages_bootmem(pfn_to_page(i), order);
  
  	for (i = end_aligned; i < end; i++)
  		__free_pages_bootmem(pfn_to_page(i), 0);
  }
29f673860   Yinghai Lu   memblock: free al...
101
102
103
104
105
106
107
108
109
110
111
112
113
114
  static unsigned long __init __free_memory_core(phys_addr_t start,
  				 phys_addr_t end)
  {
  	unsigned long start_pfn = PFN_UP(start);
  	unsigned long end_pfn = min_t(unsigned long,
  				      PFN_DOWN(end), max_low_pfn);
  
  	if (start_pfn > end_pfn)
  		return 0;
  
  	__free_pages_memory(start_pfn, end_pfn);
  
  	return end_pfn - start_pfn;
  }
b4def3509   Joonsoo Kim   mm, nobootmem: cl...
115
  static unsigned long __init free_low_memory_core_early(void)
093258732   Yinghai Lu   bootmem: Separate...
116
  {
093258732   Yinghai Lu   bootmem: Separate...
117
  	unsigned long count = 0;
29f673860   Yinghai Lu   memblock: free al...
118
  	phys_addr_t start, end, size;
8a9ca34c1   Tejun Heo   memblock, x86: Re...
119
  	u64 i;
29f673860   Yinghai Lu   memblock: free al...
120
121
122
123
124
125
126
  	for_each_free_mem_range(i, MAX_NUMNODES, &start, &end, NULL)
  		count += __free_memory_core(start, end);
  
  	/* free range that is used for reserved array if we allocate it */
  	size = get_allocated_memblock_reserved_regions_info(&start);
  	if (size)
  		count += __free_memory_core(start, start + size);
093258732   Yinghai Lu   bootmem: Separate...
127
128
129
  
  	return count;
  }
9feedc9d8   Jiang Liu   mm: introduce new...
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
  static void reset_node_lowmem_managed_pages(pg_data_t *pgdat)
  {
  	struct zone *z;
  
  	/*
  	 * In free_area_init_core(), highmem zone's managed_pages is set to
  	 * present_pages, and bootmem allocator doesn't allocate from highmem
  	 * zones. So there's no need to recalculate managed_pages because all
  	 * highmem pages will be managed by the buddy system. Here highmem
  	 * zone also includes highmem movable zone.
  	 */
  	for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++)
  		if (!is_highmem(z))
  			z->managed_pages = 0;
  }
093258732   Yinghai Lu   bootmem: Separate...
145
  /**
093258732   Yinghai Lu   bootmem: Separate...
146
147
148
149
150
151
   * free_all_bootmem - release free pages to the buddy allocator
   *
   * Returns the number of pages actually released.
   */
  unsigned long __init free_all_bootmem(void)
  {
9feedc9d8   Jiang Liu   mm: introduce new...
152
153
154
155
  	struct pglist_data *pgdat;
  
  	for_each_online_pgdat(pgdat)
  		reset_node_lowmem_managed_pages(pgdat);
093258732   Yinghai Lu   bootmem: Separate...
156
157
  	/*
  	 * We need to use MAX_NUMNODES instead of NODE_DATA(0)->node_id
25985edce   Lucas De Marchi   Fix common misspe...
158
  	 *  because in some case like Node0 doesn't have RAM installed
093258732   Yinghai Lu   bootmem: Separate...
159
  	 *  low ram will be on Node1
093258732   Yinghai Lu   bootmem: Separate...
160
  	 */
b4def3509   Joonsoo Kim   mm, nobootmem: cl...
161
  	return free_low_memory_core_early();
093258732   Yinghai Lu   bootmem: Separate...
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
  }
  
  /**
   * free_bootmem_node - mark a page range as usable
   * @pgdat: node the range resides on
   * @physaddr: starting address of the range
   * @size: size of the range in bytes
   *
   * Partial pages will be considered reserved and left as they are.
   *
   * The range must reside completely on the specified node.
   */
  void __init free_bootmem_node(pg_data_t *pgdat, unsigned long physaddr,
  			      unsigned long size)
  {
  	kmemleak_free_part(__va(physaddr), size);
24aa07882   Tejun Heo   memblock, x86: Re...
178
  	memblock_free(physaddr, size);
093258732   Yinghai Lu   bootmem: Separate...
179
180
181
182
183
184
185
186
187
188
189
190
191
192
  }
  
  /**
   * free_bootmem - mark a page range as usable
   * @addr: starting address of the range
   * @size: size of the range in bytes
   *
   * Partial pages will be considered reserved and left as they are.
   *
   * The range must be contiguous but may span node boundaries.
   */
  void __init free_bootmem(unsigned long addr, unsigned long size)
  {
  	kmemleak_free_part(__va(addr), size);
24aa07882   Tejun Heo   memblock, x86: Re...
193
  	memblock_free(addr, size);
093258732   Yinghai Lu   bootmem: Separate...
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
  }
  
  static void * __init ___alloc_bootmem_nopanic(unsigned long size,
  					unsigned long align,
  					unsigned long goal,
  					unsigned long limit)
  {
  	void *ptr;
  
  	if (WARN_ON_ONCE(slab_is_available()))
  		return kzalloc(size, GFP_NOWAIT);
  
  restart:
  
  	ptr = __alloc_memory_core_early(MAX_NUMNODES, size, align, goal, limit);
  
  	if (ptr)
  		return ptr;
  
  	if (goal != 0) {
  		goal = 0;
  		goto restart;
  	}
  
  	return NULL;
  }
  
  /**
   * __alloc_bootmem_nopanic - allocate boot memory without panicking
   * @size: size of the request in bytes
   * @align: alignment of the region
   * @goal: preferred starting address of the region
   *
   * The goal is dropped if it can not be satisfied and the allocation will
   * fall back to memory below @goal.
   *
   * Allocation may happen on any node in the system.
   *
   * Returns NULL on failure.
   */
  void * __init __alloc_bootmem_nopanic(unsigned long size, unsigned long align,
  					unsigned long goal)
  {
  	unsigned long limit = -1UL;
  
  	return ___alloc_bootmem_nopanic(size, align, goal, limit);
  }
  
  static void * __init ___alloc_bootmem(unsigned long size, unsigned long align,
  					unsigned long goal, unsigned long limit)
  {
  	void *mem = ___alloc_bootmem_nopanic(size, align, goal, limit);
  
  	if (mem)
  		return mem;
  	/*
  	 * Whoops, we cannot satisfy the allocation request.
  	 */
  	printk(KERN_ALERT "bootmem alloc of %lu bytes failed!
  ", size);
  	panic("Out of memory");
  	return NULL;
  }
  
  /**
   * __alloc_bootmem - allocate boot memory
   * @size: size of the request in bytes
   * @align: alignment of the region
   * @goal: preferred starting address of the region
   *
   * The goal is dropped if it can not be satisfied and the allocation will
   * fall back to memory below @goal.
   *
   * Allocation may happen on any node in the system.
   *
   * The function panics if the request can not be satisfied.
   */
  void * __init __alloc_bootmem(unsigned long size, unsigned long align,
  			      unsigned long goal)
  {
  	unsigned long limit = -1UL;
  
  	return ___alloc_bootmem(size, align, goal, limit);
  }
99ab7b194   Yinghai Lu   mm: sparse: fix u...
278
  void * __init ___alloc_bootmem_node_nopanic(pg_data_t *pgdat,
ba5398683   Johannes Weiner   mm: nobootmem: un...
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
  						   unsigned long size,
  						   unsigned long align,
  						   unsigned long goal,
  						   unsigned long limit)
  {
  	void *ptr;
  
  again:
  	ptr = __alloc_memory_core_early(pgdat->node_id, size, align,
  					goal, limit);
  	if (ptr)
  		return ptr;
  
  	ptr = __alloc_memory_core_early(MAX_NUMNODES, size, align,
  					goal, limit);
  	if (ptr)
  		return ptr;
  
  	if (goal) {
  		goal = 0;
  		goto again;
  	}
  
  	return NULL;
  }
  
  void * __init __alloc_bootmem_node_nopanic(pg_data_t *pgdat, unsigned long size,
  				   unsigned long align, unsigned long goal)
  {
  	if (WARN_ON_ONCE(slab_is_available()))
  		return kzalloc_node(size, GFP_NOWAIT, pgdat->node_id);
  
  	return ___alloc_bootmem_node_nopanic(pgdat, size, align, goal, 0);
  }
  
  void * __init ___alloc_bootmem_node(pg_data_t *pgdat, unsigned long size,
  				    unsigned long align, unsigned long goal,
  				    unsigned long limit)
  {
  	void *ptr;
  
  	ptr = ___alloc_bootmem_node_nopanic(pgdat, size, align, goal, limit);
  	if (ptr)
  		return ptr;
  
  	printk(KERN_ALERT "bootmem alloc of %lu bytes failed!
  ", size);
  	panic("Out of memory");
  	return NULL;
  }
093258732   Yinghai Lu   bootmem: Separate...
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
  /**
   * __alloc_bootmem_node - allocate boot memory from a specific node
   * @pgdat: node to allocate from
   * @size: size of the request in bytes
   * @align: alignment of the region
   * @goal: preferred starting address of the region
   *
   * The goal is dropped if it can not be satisfied and the allocation will
   * fall back to memory below @goal.
   *
   * Allocation may fall back to any node in the system if the specified node
   * can not hold the requested memory.
   *
   * The function panics if the request can not be satisfied.
   */
  void * __init __alloc_bootmem_node(pg_data_t *pgdat, unsigned long size,
  				   unsigned long align, unsigned long goal)
  {
093258732   Yinghai Lu   bootmem: Separate...
347
348
  	if (WARN_ON_ONCE(slab_is_available()))
  		return kzalloc_node(size, GFP_NOWAIT, pgdat->node_id);
ba5398683   Johannes Weiner   mm: nobootmem: un...
349
  	return ___alloc_bootmem_node(pgdat, size, align, goal, 0);
093258732   Yinghai Lu   bootmem: Separate...
350
351
352
353
354
  }
  
  void * __init __alloc_bootmem_node_high(pg_data_t *pgdat, unsigned long size,
  				   unsigned long align, unsigned long goal)
  {
093258732   Yinghai Lu   bootmem: Separate...
355
  	return __alloc_bootmem_node(pgdat, size, align, goal);
093258732   Yinghai Lu   bootmem: Separate...
356
  }
093258732   Yinghai Lu   bootmem: Separate...
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
  #ifndef ARCH_LOW_ADDRESS_LIMIT
  #define ARCH_LOW_ADDRESS_LIMIT	0xffffffffUL
  #endif
  
  /**
   * __alloc_bootmem_low - allocate low boot memory
   * @size: size of the request in bytes
   * @align: alignment of the region
   * @goal: preferred starting address of the region
   *
   * The goal is dropped if it can not be satisfied and the allocation will
   * fall back to memory below @goal.
   *
   * Allocation may happen on any node in the system.
   *
   * The function panics if the request can not be satisfied.
   */
  void * __init __alloc_bootmem_low(unsigned long size, unsigned long align,
  				  unsigned long goal)
  {
  	return ___alloc_bootmem(size, align, goal, ARCH_LOW_ADDRESS_LIMIT);
  }
38fa4175e   Yinghai Lu   mm: Add alloc_boo...
379
380
381
382
383
384
385
  void * __init __alloc_bootmem_low_nopanic(unsigned long size,
  					  unsigned long align,
  					  unsigned long goal)
  {
  	return ___alloc_bootmem_nopanic(size, align, goal,
  					ARCH_LOW_ADDRESS_LIMIT);
  }
093258732   Yinghai Lu   bootmem: Separate...
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
  /**
   * __alloc_bootmem_low_node - allocate low boot memory from a specific node
   * @pgdat: node to allocate from
   * @size: size of the request in bytes
   * @align: alignment of the region
   * @goal: preferred starting address of the region
   *
   * The goal is dropped if it can not be satisfied and the allocation will
   * fall back to memory below @goal.
   *
   * Allocation may fall back to any node in the system if the specified node
   * can not hold the requested memory.
   *
   * The function panics if the request can not be satisfied.
   */
  void * __init __alloc_bootmem_low_node(pg_data_t *pgdat, unsigned long size,
  				       unsigned long align, unsigned long goal)
  {
093258732   Yinghai Lu   bootmem: Separate...
404
405
  	if (WARN_ON_ONCE(slab_is_available()))
  		return kzalloc_node(size, GFP_NOWAIT, pgdat->node_id);
ba5398683   Johannes Weiner   mm: nobootmem: un...
406
407
  	return ___alloc_bootmem_node(pgdat, size, align, goal,
  				     ARCH_LOW_ADDRESS_LIMIT);
093258732   Yinghai Lu   bootmem: Separate...
408
  }