Blame view
mm/bootmem.c
21 KB
1da177e4c
|
1 |
/* |
57cfc29ef
|
2 |
* bootmem - A boot-time physical memory allocator and configurator |
1da177e4c
|
3 4 |
* * Copyright (C) 1999 Ingo Molnar |
57cfc29ef
|
5 6 |
* 1999 Kanoj Sarcar, SGI * 2008 Johannes Weiner |
1da177e4c
|
7 |
* |
57cfc29ef
|
8 9 |
* Access to this subsystem has to be serialized externally (which is true * for the boot process anyway). |
1da177e4c
|
10 |
*/ |
1da177e4c
|
11 |
#include <linux/init.h> |
bbc7b92e3
|
12 |
#include <linux/pfn.h> |
5a0e3ad6a
|
13 |
#include <linux/slab.h> |
1da177e4c
|
14 |
#include <linux/bootmem.h> |
b95f1b31b
|
15 |
#include <linux/export.h> |
ec3a354bd
|
16 |
#include <linux/kmemleak.h> |
08677214e
|
17 |
#include <linux/range.h> |
72d7c3b33
|
18 |
#include <linux/memblock.h> |
d85fbee89
|
19 20 |
#include <linux/bug.h> #include <linux/io.h> |
e786e86a5
|
21 |
|
dfd54cbcc
|
22 |
#include <asm/processor.h> |
e786e86a5
|
23 |
|
1da177e4c
|
24 |
#include "internal.h" |
e782ab421
|
25 26 27 28 29 30 |
#ifndef CONFIG_NEED_MULTIPLE_NODES struct pglist_data __refdata contig_page_data = { .bdata = &bootmem_node_data[0] }; EXPORT_SYMBOL(contig_page_data); #endif |
1da177e4c
|
31 32 33 |
unsigned long max_low_pfn; unsigned long min_low_pfn; unsigned long max_pfn; |
8dd330300
|
34 |
unsigned long long max_possible_pfn; |
1da177e4c
|
35 |
|
b61bfa3c4
|
36 |
bootmem_data_t bootmem_node_data[MAX_NUMNODES] __initdata; |
636cc40cb
|
37 |
static struct list_head bdata_list __initdata = LIST_HEAD_INIT(bdata_list); |
2e5237daf
|
38 39 40 41 42 43 44 45 46 47 48 |
static int bootmem_debug; static int __init bootmem_debug_setup(char *buf) { bootmem_debug = 1; return 0; } early_param("bootmem_debug", bootmem_debug_setup); #define bdebug(fmt, args...) ({ \ if (unlikely(bootmem_debug)) \ |
1170532bb
|
49 |
pr_info("bootmem::%s " fmt, \ |
80a914dc0
|
50 |
__func__, ## args); \ |
2e5237daf
|
51 |
}) |
df049a5f4
|
52 |
static unsigned long __init bootmap_bytes(unsigned long pages) |
223e8dc92
|
53 |
{ |
9571a9829
|
54 |
unsigned long bytes = DIV_ROUND_UP(pages, 8); |
223e8dc92
|
55 |
|
df049a5f4
|
56 |
return ALIGN(bytes, sizeof(long)); |
223e8dc92
|
57 |
} |
a66fd7dae
|
58 59 60 61 |
/** * bootmem_bootmap_pages - calculate bitmap size in pages * @pages: number of pages the bitmap has to represent */ |
f71bf0cac
|
62 |
unsigned long __init bootmem_bootmap_pages(unsigned long pages) |
1da177e4c
|
63 |
{ |
df049a5f4
|
64 |
unsigned long bytes = bootmap_bytes(pages); |
1da177e4c
|
65 |
|
df049a5f4
|
66 |
return PAGE_ALIGN(bytes) >> PAGE_SHIFT; |
1da177e4c
|
67 |
} |
f71bf0cac
|
68 |
|
679bc9fbb
|
69 70 71 |
/* * link bdata in order */ |
69d49e681
|
72 |
static void __init link_bootmem(bootmem_data_t *bdata) |
679bc9fbb
|
73 |
{ |
5c2b8a162
|
74 |
bootmem_data_t *ent; |
f71bf0cac
|
75 |
|
5c2b8a162
|
76 77 78 79 80 |
list_for_each_entry(ent, &bdata_list, list) { if (bdata->node_min_pfn < ent->node_min_pfn) { list_add_tail(&bdata->list, &ent->list); return; } |
679bc9fbb
|
81 |
} |
5c2b8a162
|
82 83 |
list_add_tail(&bdata->list, &bdata_list); |
679bc9fbb
|
84 |
} |
bbc7b92e3
|
85 |
/* |
1da177e4c
|
86 87 |
* Called once to set up the allocator itself. */ |
8ae044630
|
88 |
static unsigned long __init init_bootmem_core(bootmem_data_t *bdata, |
1da177e4c
|
89 90 |
unsigned long mapstart, unsigned long start, unsigned long end) { |
bbc7b92e3
|
91 |
unsigned long mapsize; |
1da177e4c
|
92 |
|
2dbb51c49
|
93 |
mminit_validate_memmodel_limits(&start, &end); |
bbc7b92e3
|
94 |
bdata->node_bootmem_map = phys_to_virt(PFN_PHYS(mapstart)); |
3560e249a
|
95 |
bdata->node_min_pfn = start; |
1da177e4c
|
96 |
bdata->node_low_pfn = end; |
679bc9fbb
|
97 |
link_bootmem(bdata); |
1da177e4c
|
98 99 100 101 102 |
/* * Initially all pages are reserved - setup_arch() has to * register free RAM areas explicitly. */ |
df049a5f4
|
103 |
mapsize = bootmap_bytes(end - start); |
1da177e4c
|
104 |
memset(bdata->node_bootmem_map, 0xff, mapsize); |
2e5237daf
|
105 106 107 |
bdebug("nid=%td start=%lx map=%lx end=%lx mapsize=%lx ", bdata - bootmem_node_data, start, mapstart, end, mapsize); |
1da177e4c
|
108 109 |
return mapsize; } |
a66fd7dae
|
110 111 112 113 114 115 116 117 118 |
/** * init_bootmem_node - register a node as boot memory * @pgdat: node to register * @freepfn: pfn where the bitmap for this node is to be placed * @startpfn: first pfn on the node * @endpfn: first pfn after the node * * Returns the number of bytes needed to hold the bitmap for this node. */ |
223e8dc92
|
119 120 121 122 123 |
unsigned long __init init_bootmem_node(pg_data_t *pgdat, unsigned long freepfn, unsigned long startpfn, unsigned long endpfn) { return init_bootmem_core(pgdat->bdata, freepfn, startpfn, endpfn); } |
a66fd7dae
|
124 125 126 127 128 129 130 |
/** * init_bootmem - register boot memory * @start: pfn where the bitmap is to be placed * @pages: number of available physical pages * * Returns the number of bytes needed to hold the bitmap. */ |
223e8dc92
|
131 132 133 134 135 136 |
unsigned long __init init_bootmem(unsigned long start, unsigned long pages) { max_low_pfn = pages; min_low_pfn = start; return init_bootmem_core(NODE_DATA(0)->bdata, start, 0, pages); } |
093258732
|
137 |
|
9f993ac3f
|
138 139 |
/* * free_bootmem_late - free bootmem pages directly to page allocator |
81df9bff2
|
140 |
* @addr: starting physical address of the range |
9f993ac3f
|
141 142 143 144 145 146 |
* @size: size of the range in bytes * * This is only useful when the bootmem allocator has already been torn * down, but we are still initializing the system. Pages are given directly * to the page allocator, no bootmem metadata is updated because it is gone. */ |
81df9bff2
|
147 |
void __init free_bootmem_late(unsigned long physaddr, unsigned long size) |
9f993ac3f
|
148 149 |
{ unsigned long cursor, end; |
81df9bff2
|
150 |
kmemleak_free_part(__va(physaddr), size); |
9f993ac3f
|
151 |
|
81df9bff2
|
152 153 |
cursor = PFN_UP(physaddr); end = PFN_DOWN(physaddr + size); |
9f993ac3f
|
154 155 |
for (; cursor < end; cursor++) { |
d70ddd7a5
|
156 |
__free_pages_bootmem(pfn_to_page(cursor), cursor, 0); |
9f993ac3f
|
157 158 159 |
totalram_pages++; } } |
223e8dc92
|
160 161 162 |
static unsigned long __init free_all_bootmem_core(bootmem_data_t *bdata) { struct page *page; |
d70ddd7a5
|
163 |
unsigned long *map, start, end, pages, cur, count = 0; |
41546c174
|
164 165 166 |
if (!bdata->node_bootmem_map) return 0; |
4a099fb4b
|
167 |
map = bdata->node_bootmem_map; |
3560e249a
|
168 |
start = bdata->node_min_pfn; |
41546c174
|
169 |
end = bdata->node_low_pfn; |
799f933a8
|
170 171 172 |
bdebug("nid=%td start=%lx end=%lx ", bdata - bootmem_node_data, start, end); |
223e8dc92
|
173 |
|
41546c174
|
174 |
while (start < end) { |
4a099fb4b
|
175 |
unsigned long idx, vec; |
10d73e655
|
176 |
unsigned shift; |
223e8dc92
|
177 |
|
3560e249a
|
178 |
idx = start - bdata->node_min_pfn; |
10d73e655
|
179 180 181 182 183 |
shift = idx & (BITS_PER_LONG - 1); /* * vec holds at most BITS_PER_LONG map bits, * bit 0 corresponds to start. */ |
41546c174
|
184 |
vec = ~map[idx / BITS_PER_LONG]; |
10d73e655
|
185 186 187 188 189 190 191 |
if (shift) { vec >>= shift; if (end - start >= BITS_PER_LONG) vec |= ~map[idx / BITS_PER_LONG + 1] << (BITS_PER_LONG - shift); } |
799f933a8
|
192 193 194 195 196 197 |
/* * If we have a properly aligned and fully unreserved * BITS_PER_LONG block of pages in front of us, free * it in one go. */ if (IS_ALIGNED(start, BITS_PER_LONG) && vec == ~0UL) { |
41546c174
|
198 |
int order = ilog2(BITS_PER_LONG); |
d70ddd7a5
|
199 |
__free_pages_bootmem(pfn_to_page(start), start, order); |
223e8dc92
|
200 |
count += BITS_PER_LONG; |
799f933a8
|
201 |
start += BITS_PER_LONG; |
41546c174
|
202 |
} else { |
d70ddd7a5
|
203 |
cur = start; |
41546c174
|
204 |
|
10d73e655
|
205 206 |
start = ALIGN(start + 1, BITS_PER_LONG); while (vec && cur != start) { |
41546c174
|
207 |
if (vec & 1) { |
10d73e655
|
208 |
page = pfn_to_page(cur); |
d70ddd7a5
|
209 |
__free_pages_bootmem(page, cur, 0); |
41546c174
|
210 |
count++; |
223e8dc92
|
211 |
} |
41546c174
|
212 |
vec >>= 1; |
10d73e655
|
213 |
++cur; |
223e8dc92
|
214 |
} |
223e8dc92
|
215 |
} |
223e8dc92
|
216 |
} |
d70ddd7a5
|
217 |
cur = bdata->node_min_pfn; |
223e8dc92
|
218 |
page = virt_to_page(bdata->node_bootmem_map); |
3560e249a
|
219 |
pages = bdata->node_low_pfn - bdata->node_min_pfn; |
41546c174
|
220 221 |
pages = bootmem_bootmap_pages(pages); count += pages; |
5576646f3
|
222 |
while (pages--) |
d70ddd7a5
|
223 |
__free_pages_bootmem(page++, cur++, 0); |
1b4ace414
|
224 |
bdata->node_bootmem_map = NULL; |
223e8dc92
|
225 |
|
2e5237daf
|
226 227 |
bdebug("nid=%td released=%lx ", bdata - bootmem_node_data, count); |
223e8dc92
|
228 229 |
return count; } |
7b4b2a0d6
|
230 |
static int reset_managed_pages_done __initdata; |
f784a3f19
|
231 |
void reset_node_managed_pages(pg_data_t *pgdat) |
9feedc9d8
|
232 233 |
{ struct zone *z; |
9feedc9d8
|
234 |
for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++) |
7b4b2a0d6
|
235 236 237 238 239 240 |
z->managed_pages = 0; } void __init reset_all_zones_managed_pages(void) { struct pglist_data *pgdat; |
f784a3f19
|
241 242 |
if (reset_managed_pages_done) return; |
7b4b2a0d6
|
243 244 |
for_each_online_pgdat(pgdat) reset_node_managed_pages(pgdat); |
f784a3f19
|
245 |
|
7b4b2a0d6
|
246 |
reset_managed_pages_done = 1; |
9feedc9d8
|
247 |
} |
a66fd7dae
|
248 |
/** |
a66fd7dae
|
249 250 251 252 |
* free_all_bootmem - release free pages to the buddy allocator * * Returns the number of pages actually released. */ |
223e8dc92
|
253 254 |
unsigned long __init free_all_bootmem(void) { |
aa235fc71
|
255 256 |
unsigned long total_pages = 0; bootmem_data_t *bdata; |
9feedc9d8
|
257 |
|
7b4b2a0d6
|
258 |
reset_all_zones_managed_pages(); |
aa235fc71
|
259 260 261 |
list_for_each_entry(bdata, &bdata_list, list) total_pages += free_all_bootmem_core(bdata); |
0c9885347
|
262 |
totalram_pages += total_pages; |
aa235fc71
|
263 |
return total_pages; |
223e8dc92
|
264 |
} |
d747fa4bc
|
265 266 267 268 269 270 271 |
static void __init __free(bootmem_data_t *bdata, unsigned long sidx, unsigned long eidx) { unsigned long idx; bdebug("nid=%td start=%lx end=%lx ", bdata - bootmem_node_data, |
3560e249a
|
272 273 |
sidx + bdata->node_min_pfn, eidx + bdata->node_min_pfn); |
d747fa4bc
|
274 |
|
1b4ace414
|
275 276 |
if (WARN_ON(bdata->node_bootmem_map == NULL)) return; |
e2bf3cae5
|
277 278 |
if (bdata->hint_idx > sidx) bdata->hint_idx = sidx; |
d747fa4bc
|
279 280 281 282 283 284 285 286 287 288 289 290 291 292 |
for (idx = sidx; idx < eidx; idx++) if (!test_and_clear_bit(idx, bdata->node_bootmem_map)) BUG(); } static int __init __reserve(bootmem_data_t *bdata, unsigned long sidx, unsigned long eidx, int flags) { unsigned long idx; int exclusive = flags & BOOTMEM_EXCLUSIVE; bdebug("nid=%td start=%lx end=%lx flags=%x ", bdata - bootmem_node_data, |
3560e249a
|
293 294 |
sidx + bdata->node_min_pfn, eidx + bdata->node_min_pfn, |
d747fa4bc
|
295 |
flags); |
1b4ace414
|
296 297 |
if (WARN_ON(bdata->node_bootmem_map == NULL)) return 0; |
d747fa4bc
|
298 299 300 301 302 303 304 305 |
for (idx = sidx; idx < eidx; idx++) if (test_and_set_bit(idx, bdata->node_bootmem_map)) { if (exclusive) { __free(bdata, sidx, idx); return -EBUSY; } bdebug("silent double reserve of PFN %lx ", |
3560e249a
|
306 |
idx + bdata->node_min_pfn); |
d747fa4bc
|
307 308 309 |
} return 0; } |
e2bf3cae5
|
310 311 312 |
static int __init mark_bootmem_node(bootmem_data_t *bdata, unsigned long start, unsigned long end, int reserve, int flags) |
223e8dc92
|
313 314 |
{ unsigned long sidx, eidx; |
223e8dc92
|
315 |
|
e2bf3cae5
|
316 317 318 |
bdebug("nid=%td start=%lx end=%lx reserve=%d flags=%x ", bdata - bootmem_node_data, start, end, reserve, flags); |
223e8dc92
|
319 |
|
3560e249a
|
320 |
BUG_ON(start < bdata->node_min_pfn); |
e2bf3cae5
|
321 |
BUG_ON(end > bdata->node_low_pfn); |
223e8dc92
|
322 |
|
3560e249a
|
323 324 |
sidx = start - bdata->node_min_pfn; eidx = end - bdata->node_min_pfn; |
223e8dc92
|
325 |
|
e2bf3cae5
|
326 327 |
if (reserve) return __reserve(bdata, sidx, eidx, flags); |
223e8dc92
|
328 |
else |
e2bf3cae5
|
329 330 331 332 333 334 335 336 337 338 339 340 341 342 |
__free(bdata, sidx, eidx); return 0; } static int __init mark_bootmem(unsigned long start, unsigned long end, int reserve, int flags) { unsigned long pos; bootmem_data_t *bdata; pos = start; list_for_each_entry(bdata, &bdata_list, list) { int err; unsigned long max; |
3560e249a
|
343 344 |
if (pos < bdata->node_min_pfn || pos >= bdata->node_low_pfn) { |
e2bf3cae5
|
345 346 347 348 349 |
BUG_ON(pos != start); continue; } max = min(bdata->node_low_pfn, end); |
223e8dc92
|
350 |
|
e2bf3cae5
|
351 352 353 354 355 |
err = mark_bootmem_node(bdata, pos, max, reserve, flags); if (reserve && err) { mark_bootmem(start, pos, 0, 0); return err; } |
223e8dc92
|
356 |
|
e2bf3cae5
|
357 358 359 360 361 |
if (max == end) return 0; pos = bdata->node_low_pfn; } BUG(); |
223e8dc92
|
362 |
} |
a66fd7dae
|
363 364 365 366 367 368 369 370 |
/** * free_bootmem_node - mark a page range as usable * @pgdat: node the range resides on * @physaddr: starting address of the range * @size: size of the range in bytes * * Partial pages will be considered reserved and left as they are. * |
e2bf3cae5
|
371 |
* The range must reside completely on the specified node. |
a66fd7dae
|
372 |
*/ |
223e8dc92
|
373 374 375 |
void __init free_bootmem_node(pg_data_t *pgdat, unsigned long physaddr, unsigned long size) { |
e2bf3cae5
|
376 |
unsigned long start, end; |
ec3a354bd
|
377 |
kmemleak_free_part(__va(physaddr), size); |
e2bf3cae5
|
378 379 380 381 |
start = PFN_UP(physaddr); end = PFN_DOWN(physaddr + size); mark_bootmem_node(pgdat->bdata, start, end, 0, 0); |
223e8dc92
|
382 |
} |
a66fd7dae
|
383 384 |
/** * free_bootmem - mark a page range as usable |
81df9bff2
|
385 |
* @addr: starting physical address of the range |
a66fd7dae
|
386 387 388 389 |
* @size: size of the range in bytes * * Partial pages will be considered reserved and left as they are. * |
e2bf3cae5
|
390 |
* The range must be contiguous but may span node boundaries. |
a66fd7dae
|
391 |
*/ |
81df9bff2
|
392 |
void __init free_bootmem(unsigned long physaddr, unsigned long size) |
223e8dc92
|
393 |
{ |
e2bf3cae5
|
394 |
unsigned long start, end; |
a5645a61b
|
395 |
|
81df9bff2
|
396 |
kmemleak_free_part(__va(physaddr), size); |
ec3a354bd
|
397 |
|
81df9bff2
|
398 399 |
start = PFN_UP(physaddr); end = PFN_DOWN(physaddr + size); |
1da177e4c
|
400 |
|
e2bf3cae5
|
401 |
mark_bootmem(start, end, 0, 0); |
1da177e4c
|
402 |
} |
a66fd7dae
|
403 404 405 406 407 408 409 410 411 |
/** * reserve_bootmem_node - mark a page range as reserved * @pgdat: node the range resides on * @physaddr: starting address of the range * @size: size of the range in bytes * @flags: reservation flags (see linux/bootmem.h) * * Partial pages will be reserved. * |
e2bf3cae5
|
412 |
* The range must reside completely on the specified node. |
a66fd7dae
|
413 |
*/ |
223e8dc92
|
414 415 |
int __init reserve_bootmem_node(pg_data_t *pgdat, unsigned long physaddr, unsigned long size, int flags) |
1da177e4c
|
416 |
{ |
e2bf3cae5
|
417 |
unsigned long start, end; |
1da177e4c
|
418 |
|
e2bf3cae5
|
419 420 421 422 |
start = PFN_DOWN(physaddr); end = PFN_UP(physaddr + size); return mark_bootmem_node(pgdat->bdata, start, end, 1, flags); |
223e8dc92
|
423 |
} |
5a982cbc7
|
424 |
|
a66fd7dae
|
425 |
/** |
0d4ba4d7b
|
426 |
* reserve_bootmem - mark a page range as reserved |
a66fd7dae
|
427 428 429 430 431 432 |
* @addr: starting address of the range * @size: size of the range in bytes * @flags: reservation flags (see linux/bootmem.h) * * Partial pages will be reserved. * |
e2bf3cae5
|
433 |
* The range must be contiguous but may span node boundaries. |
a66fd7dae
|
434 |
*/ |
223e8dc92
|
435 436 437 |
int __init reserve_bootmem(unsigned long addr, unsigned long size, int flags) { |
e2bf3cae5
|
438 |
unsigned long start, end; |
1da177e4c
|
439 |
|
e2bf3cae5
|
440 441 |
start = PFN_DOWN(addr); end = PFN_UP(addr + size); |
223e8dc92
|
442 |
|
e2bf3cae5
|
443 |
return mark_bootmem(start, end, 1, flags); |
1da177e4c
|
444 |
} |
8aa043d74
|
445 446 |
static unsigned long __init align_idx(struct bootmem_data *bdata, unsigned long idx, unsigned long step) |
481ebd0d7
|
447 448 449 450 451 452 453 454 455 456 |
{ unsigned long base = bdata->node_min_pfn; /* * Align the index with respect to the node start so that the * combination of both satisfies the requested alignment. */ return ALIGN(base + idx, step) - base; } |
8aa043d74
|
457 458 |
static unsigned long __init align_off(struct bootmem_data *bdata, unsigned long off, unsigned long align) |
481ebd0d7
|
459 460 461 462 463 464 465 |
{ unsigned long base = PFN_PHYS(bdata->node_min_pfn); /* Same as align_idx for byte offsets */ return ALIGN(base + off, align) - base; } |
c6785b6bf
|
466 |
static void * __init alloc_bootmem_bdata(struct bootmem_data *bdata, |
d0c4f5702
|
467 468 |
unsigned long size, unsigned long align, unsigned long goal, unsigned long limit) |
1da177e4c
|
469 |
{ |
0f3caba21
|
470 |
unsigned long fallback = 0; |
5f2809e69
|
471 |
unsigned long min, max, start, sidx, midx, step; |
594fe1a04
|
472 473 474 475 |
bdebug("nid=%td size=%lx [%lu pages] align=%lx goal=%lx limit=%lx ", bdata - bootmem_node_data, size, PAGE_ALIGN(size) >> PAGE_SHIFT, align, goal, limit); |
5f2809e69
|
476 477 478 |
BUG_ON(!size); BUG_ON(align & (align - 1)); BUG_ON(limit && goal + size > limit); |
1da177e4c
|
479 |
|
7c309a64d
|
480 481 |
if (!bdata->node_bootmem_map) return NULL; |
3560e249a
|
482 |
min = bdata->node_min_pfn; |
5f2809e69
|
483 |
max = bdata->node_low_pfn; |
9a2dc04cf
|
484 |
|
5f2809e69
|
485 486 487 488 489 490 |
goal >>= PAGE_SHIFT; limit >>= PAGE_SHIFT; if (limit && max > limit) max = limit; if (max <= min) |
9a2dc04cf
|
491 |
return NULL; |
5f2809e69
|
492 |
step = max(align >> PAGE_SHIFT, 1UL); |
281dd25cd
|
493 |
|
5f2809e69
|
494 495 496 497 |
if (goal && min < goal && goal < max) start = ALIGN(goal, step); else start = ALIGN(min, step); |
1da177e4c
|
498 |
|
481ebd0d7
|
499 |
sidx = start - bdata->node_min_pfn; |
3560e249a
|
500 |
midx = max - bdata->node_min_pfn; |
1da177e4c
|
501 |
|
5f2809e69
|
502 |
if (bdata->hint_idx > sidx) { |
0f3caba21
|
503 504 505 506 507 |
/* * Handle the valid case of sidx being zero and still * catch the fallback below. */ fallback = sidx + 1; |
481ebd0d7
|
508 |
sidx = align_idx(bdata, bdata->hint_idx, step); |
5f2809e69
|
509 |
} |
1da177e4c
|
510 |
|
5f2809e69
|
511 512 513 514 515 516 |
while (1) { int merge; void *region; unsigned long eidx, i, start_off, end_off; find_block: sidx = find_next_zero_bit(bdata->node_bootmem_map, midx, sidx); |
481ebd0d7
|
517 |
sidx = align_idx(bdata, sidx, step); |
5f2809e69
|
518 |
eidx = sidx + PFN_UP(size); |
ad09315ca
|
519 |
|
5f2809e69
|
520 |
if (sidx >= midx || eidx > midx) |
66d43e98e
|
521 |
break; |
1da177e4c
|
522 |
|
5f2809e69
|
523 524 |
for (i = sidx; i < eidx; i++) if (test_bit(i, bdata->node_bootmem_map)) { |
481ebd0d7
|
525 |
sidx = align_idx(bdata, i, step); |
5f2809e69
|
526 527 528 529 |
if (sidx == i) sidx += step; goto find_block; } |
1da177e4c
|
530 |
|
627240aaa
|
531 |
if (bdata->last_end_off & (PAGE_SIZE - 1) && |
5f2809e69
|
532 |
PFN_DOWN(bdata->last_end_off) + 1 == sidx) |
481ebd0d7
|
533 |
start_off = align_off(bdata, bdata->last_end_off, align); |
5f2809e69
|
534 535 536 537 538 539 540 541 542 543 544 545 |
else start_off = PFN_PHYS(sidx); merge = PFN_DOWN(start_off) < sidx; end_off = start_off + size; bdata->last_end_off = end_off; bdata->hint_idx = PFN_UP(end_off); /* * Reserve the area now: */ |
d747fa4bc
|
546 547 548 |
if (__reserve(bdata, PFN_DOWN(start_off) + merge, PFN_UP(end_off), BOOTMEM_EXCLUSIVE)) BUG(); |
5f2809e69
|
549 |
|
3560e249a
|
550 551 |
region = phys_to_virt(PFN_PHYS(bdata->node_min_pfn) + start_off); |
5f2809e69
|
552 |
memset(region, 0, size); |
008139d91
|
553 554 555 556 557 |
/* * The min_count is set to 0 so that bootmem allocated blocks * are never reported as leaks. */ kmemleak_alloc(region, size, 0, 0); |
5f2809e69
|
558 |
return region; |
1da177e4c
|
559 |
} |
0f3caba21
|
560 |
if (fallback) { |
481ebd0d7
|
561 |
sidx = align_idx(bdata, fallback - 1, step); |
0f3caba21
|
562 563 564 565 566 567 |
fallback = 0; goto find_block; } return NULL; } |
c12ab504a
|
568 |
static void * __init alloc_bootmem_core(unsigned long size, |
0f3caba21
|
569 570 571 572 573 |
unsigned long align, unsigned long goal, unsigned long limit) { bootmem_data_t *bdata; |
d0c4f5702
|
574 |
void *region; |
0f3caba21
|
575 |
|
3f7dfe24b
|
576 577 |
if (WARN_ON_ONCE(slab_is_available())) return kzalloc(size, GFP_NOWAIT); |
0f3caba21
|
578 |
|
d0c4f5702
|
579 |
list_for_each_entry(bdata, &bdata_list, list) { |
0f3caba21
|
580 581 |
if (goal && bdata->node_low_pfn <= PFN_DOWN(goal)) continue; |
3560e249a
|
582 |
if (limit && bdata->node_min_pfn >= PFN_DOWN(limit)) |
0f3caba21
|
583 |
break; |
c6785b6bf
|
584 |
region = alloc_bootmem_bdata(bdata, size, align, goal, limit); |
0f3caba21
|
585 586 587 |
if (region) return region; } |
c12ab504a
|
588 589 590 591 592 593 594 595 596 597 598 599 600 601 |
return NULL; } static void * __init ___alloc_bootmem_nopanic(unsigned long size, unsigned long align, unsigned long goal, unsigned long limit) { void *ptr; restart: ptr = alloc_bootmem_core(size, align, goal, limit); if (ptr) return ptr; |
5f2809e69
|
602 603 |
if (goal) { goal = 0; |
0f3caba21
|
604 |
goto restart; |
5f2809e69
|
605 |
} |
2e5237daf
|
606 |
|
5f2809e69
|
607 |
return NULL; |
1da177e4c
|
608 |
} |
a66fd7dae
|
609 610 611 612 613 614 615 616 617 618 619 620 621 |
/** * __alloc_bootmem_nopanic - allocate boot memory without panicking * @size: size of the request in bytes * @align: alignment of the region * @goal: preferred starting address of the region * * The goal is dropped if it can not be satisfied and the allocation will * fall back to memory below @goal. * * Allocation may happen on any node in the system. * * Returns NULL on failure. */ |
bb0923a66
|
622 |
void * __init __alloc_bootmem_nopanic(unsigned long size, unsigned long align, |
0f3caba21
|
623 |
unsigned long goal) |
1da177e4c
|
624 |
{ |
08677214e
|
625 |
unsigned long limit = 0; |
08677214e
|
626 |
return ___alloc_bootmem_nopanic(size, align, goal, limit); |
0f3caba21
|
627 |
} |
1da177e4c
|
628 |
|
0f3caba21
|
629 630 631 632 633 634 635 636 637 638 |
static void * __init ___alloc_bootmem(unsigned long size, unsigned long align, unsigned long goal, unsigned long limit) { void *mem = ___alloc_bootmem_nopanic(size, align, goal, limit); if (mem) return mem; /* * Whoops, we cannot satisfy the allocation request. */ |
1170532bb
|
639 640 |
pr_alert("bootmem alloc of %lu bytes failed! ", size); |
0f3caba21
|
641 |
panic("Out of memory"); |
a8062231d
|
642 643 |
return NULL; } |
1da177e4c
|
644 |
|
a66fd7dae
|
645 646 647 648 649 650 651 652 653 654 655 656 657 |
/** * __alloc_bootmem - allocate boot memory * @size: size of the request in bytes * @align: alignment of the region * @goal: preferred starting address of the region * * The goal is dropped if it can not be satisfied and the allocation will * fall back to memory below @goal. * * Allocation may happen on any node in the system. * * The function panics if the request can not be satisfied. */ |
bb0923a66
|
658 659 |
void * __init __alloc_bootmem(unsigned long size, unsigned long align, unsigned long goal) |
a8062231d
|
660 |
{ |
08677214e
|
661 |
unsigned long limit = 0; |
08677214e
|
662 |
return ___alloc_bootmem(size, align, goal, limit); |
1da177e4c
|
663 |
} |
99ab7b194
|
664 |
void * __init ___alloc_bootmem_node_nopanic(pg_data_t *pgdat, |
4cc278b72
|
665 666 667 668 |
unsigned long size, unsigned long align, unsigned long goal, unsigned long limit) { void *ptr; |
3f7dfe24b
|
669 670 |
if (WARN_ON_ONCE(slab_is_available())) return kzalloc(size, GFP_NOWAIT); |
ab3818432
|
671 |
again: |
d0c4f5702
|
672 |
|
c8f4a2d09
|
673 674 675 |
/* do not panic in alloc_bootmem_bdata() */ if (limit && goal + size > limit) limit = 0; |
e9079911e
|
676 |
ptr = alloc_bootmem_bdata(pgdat->bdata, size, align, goal, limit); |
4cc278b72
|
677 678 |
if (ptr) return ptr; |
ab3818432
|
679 680 681 682 683 684 685 686 |
ptr = alloc_bootmem_core(size, align, goal, limit); if (ptr) return ptr; if (goal) { goal = 0; goto again; } |
421456edd
|
687 688 689 690 691 692 693 694 |
return NULL; } void * __init __alloc_bootmem_node_nopanic(pg_data_t *pgdat, unsigned long size, unsigned long align, unsigned long goal) { if (WARN_ON_ONCE(slab_is_available())) return kzalloc_node(size, GFP_NOWAIT, pgdat->node_id); |
e9079911e
|
695 |
return ___alloc_bootmem_node_nopanic(pgdat, size, align, goal, 0); |
421456edd
|
696 |
} |
e9079911e
|
697 |
void * __init ___alloc_bootmem_node(pg_data_t *pgdat, unsigned long size, |
421456edd
|
698 699 700 701 |
unsigned long align, unsigned long goal, unsigned long limit) { void *ptr; |
e9079911e
|
702 |
ptr = ___alloc_bootmem_node_nopanic(pgdat, size, align, goal, 0); |
421456edd
|
703 704 |
if (ptr) return ptr; |
1170532bb
|
705 706 |
pr_alert("bootmem alloc of %lu bytes failed! ", size); |
ab3818432
|
707 708 |
panic("Out of memory"); return NULL; |
4cc278b72
|
709 |
} |
a66fd7dae
|
710 711 712 713 714 715 716 717 718 719 720 721 722 723 724 |
/** * __alloc_bootmem_node - allocate boot memory from a specific node * @pgdat: node to allocate from * @size: size of the request in bytes * @align: alignment of the region * @goal: preferred starting address of the region * * The goal is dropped if it can not be satisfied and the allocation will * fall back to memory below @goal. * * Allocation may fall back to any node in the system if the specified node * can not hold the requested memory. * * The function panics if the request can not be satisfied. */ |
bb0923a66
|
725 726 |
void * __init __alloc_bootmem_node(pg_data_t *pgdat, unsigned long size, unsigned long align, unsigned long goal) |
1da177e4c
|
727 |
{ |
c91c4773b
|
728 729 |
if (WARN_ON_ONCE(slab_is_available())) return kzalloc_node(size, GFP_NOWAIT, pgdat->node_id); |
e9079911e
|
730 |
return ___alloc_bootmem_node(pgdat, size, align, goal, 0); |
08677214e
|
731 732 733 734 735 736 737 738 739 740 741 742 |
} void * __init __alloc_bootmem_node_high(pg_data_t *pgdat, unsigned long size, unsigned long align, unsigned long goal) { #ifdef MAX_DMA32_PFN unsigned long end_pfn; if (WARN_ON_ONCE(slab_is_available())) return kzalloc_node(size, GFP_NOWAIT, pgdat->node_id); /* update goal according ...MAX_DMA32_PFN */ |
83285c72e
|
743 |
end_pfn = pgdat_end_pfn(pgdat); |
08677214e
|
744 745 746 747 748 749 750 |
if (end_pfn > MAX_DMA32_PFN + (128 >> (20 - PAGE_SHIFT)) && (goal >> PAGE_SHIFT) < MAX_DMA32_PFN) { void *ptr; unsigned long new_goal; new_goal = MAX_DMA32_PFN << PAGE_SHIFT; |
c6785b6bf
|
751 |
ptr = alloc_bootmem_bdata(pgdat->bdata, size, align, |
08677214e
|
752 |
new_goal, 0); |
08677214e
|
753 754 755 756 757 758 |
if (ptr) return ptr; } #endif return __alloc_bootmem_node(pgdat, size, align, goal); |
1da177e4c
|
759 |
} |
dfd54cbcc
|
760 761 762 |
#ifndef ARCH_LOW_ADDRESS_LIMIT #define ARCH_LOW_ADDRESS_LIMIT 0xffffffffUL #endif |
008857c1a
|
763 |
|
a66fd7dae
|
764 765 766 767 768 769 770 771 772 773 774 775 776 |
/** * __alloc_bootmem_low - allocate low boot memory * @size: size of the request in bytes * @align: alignment of the region * @goal: preferred starting address of the region * * The goal is dropped if it can not be satisfied and the allocation will * fall back to memory below @goal. * * Allocation may happen on any node in the system. * * The function panics if the request can not be satisfied. */ |
bb0923a66
|
777 778 |
void * __init __alloc_bootmem_low(unsigned long size, unsigned long align, unsigned long goal) |
008857c1a
|
779 |
{ |
0f3caba21
|
780 |
return ___alloc_bootmem(size, align, goal, ARCH_LOW_ADDRESS_LIMIT); |
008857c1a
|
781 |
} |
38fa4175e
|
782 783 784 785 786 787 788 |
void * __init __alloc_bootmem_low_nopanic(unsigned long size, unsigned long align, unsigned long goal) { return ___alloc_bootmem_nopanic(size, align, goal, ARCH_LOW_ADDRESS_LIMIT); } |
a66fd7dae
|
789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 |
/** * __alloc_bootmem_low_node - allocate low boot memory from a specific node * @pgdat: node to allocate from * @size: size of the request in bytes * @align: alignment of the region * @goal: preferred starting address of the region * * The goal is dropped if it can not be satisfied and the allocation will * fall back to memory below @goal. * * Allocation may fall back to any node in the system if the specified node * can not hold the requested memory. * * The function panics if the request can not be satisfied. */ |
008857c1a
|
804 805 806 |
void * __init __alloc_bootmem_low_node(pg_data_t *pgdat, unsigned long size, unsigned long align, unsigned long goal) { |
c91c4773b
|
807 808 |
if (WARN_ON_ONCE(slab_is_available())) return kzalloc_node(size, GFP_NOWAIT, pgdat->node_id); |
e9079911e
|
809 810 |
return ___alloc_bootmem_node(pgdat, size, align, goal, ARCH_LOW_ADDRESS_LIMIT); |
008857c1a
|
811 |
} |