Commit a587d2daebcd2bc159d4348b6a7b028950a6d803

Authored by Yinghai Lu
Committed by H. Peter Anvin
1 parent a9ce6bc151

x86: Remove not used early_res code

and some functions in e820.c that are not used anymore

Signed-off-by: Yinghai Lu <yinghai@kernel.org>
Signed-off-by: H. Peter Anvin <hpa@zytor.com>

Showing 5 changed files with 0 additions and 680 deletions Side-by-side Diff

arch/x86/include/asm/e820.h
... ... @@ -112,30 +112,16 @@
112 112 }
113 113 #endif
114 114  
115   -extern unsigned long end_user_pfn;
116   -
117   -extern u64 find_e820_area(u64 start, u64 end, u64 size, u64 align);
118   -extern u64 find_e820_area_size(u64 start, u64 *sizep, u64 align);
119   -extern u64 early_reserve_e820(u64 startt, u64 sizet, u64 align);
120   -
121 115 extern unsigned long e820_end_of_ram_pfn(void);
122 116 extern unsigned long e820_end_of_low_ram_pfn(void);
123   -extern void e820_register_active_regions(int nid, unsigned long start_pfn,
124   - unsigned long end_pfn);
125   -extern u64 e820_hole_size(u64 start, u64 end);
126   -
127 117 extern u64 early_reserve_e820(u64 startt, u64 sizet, u64 align);
128 118  
129 119 void memblock_x86_fill(void);
130   -
131 120 extern void finish_e820_parsing(void);
132 121 extern void e820_reserve_resources(void);
133 122 extern void e820_reserve_resources_late(void);
134 123 extern void setup_memory_map(void);
135 124 extern char *default_machine_specific_memory_setup(void);
136   -
137   -void reserve_early(u64 start, u64 end, char *name);
138   -void free_early(u64 start, u64 end);
139 125  
140 126 /*
141 127 * Returns true iff the specified range [s,e) is completely contained inside
arch/x86/kernel/e820.c
... ... @@ -739,32 +739,6 @@
739 739 #endif
740 740  
741 741 /*
742   - * Find a free area with specified alignment in a specific range.
743   - */
744   -u64 __init find_e820_area(u64 start, u64 end, u64 size, u64 align)
745   -{
746   - u64 mem = memblock_find_in_range(start, end, size, align);
747   -
748   - if (mem == MEMBLOCK_ERROR)
749   - return -1ULL;
750   -
751   - return mem;
752   -}
753   -
754   -/*
755   - * Find next free range after *start
756   - */
757   -u64 __init find_e820_area_size(u64 start, u64 *sizep, u64 align)
758   -{
759   - u64 mem = memblock_x86_find_in_range_size(start, sizep, align);
760   -
761   - if (mem == MEMBLOCK_ERROR)
762   - return -1ULL
763   -
764   - return mem;
765   -}
766   -
767   -/*
768 742 * pre allocated 4k and reserved it in memblock and e820_saved
769 743 */
770 744 u64 __init early_reserve_e820(u64 startt, u64 sizet, u64 align)
... ... @@ -854,32 +828,6 @@
854 828 unsigned long __init e820_end_of_low_ram_pfn(void)
855 829 {
856 830 return e820_end_pfn(1UL<<(32 - PAGE_SHIFT), E820_RAM);
857   -}
858   -
859   -/* Walk the e820 map and register active regions within a node */
860   -void __init e820_register_active_regions(int nid, unsigned long start_pfn,
861   - unsigned long last_pfn)
862   -{
863   - memblock_x86_register_active_regions(nid, start_pfn, last_pfn);
864   -}
865   -
866   -/*
867   - * Find the hole size (in bytes) in the memory range.
868   - * @start: starting address of the memory range to scan
869   - * @end: ending address of the memory range to scan
870   - */
871   -u64 __init e820_hole_size(u64 start, u64 end)
872   -{
873   - return memblock_x86_hole_size(start, end);
874   -}
875   -
876   -void reserve_early(u64 start, u64 end, char *name)
877   -{
878   - memblock_x86_reserve_range(start, end, name);
879   -}
880   -void free_early(u64 start, u64 end)
881   -{
882   - memblock_x86_free_range(start, end);
883 831 }
884 832  
885 833 static void early_panic(char *msg)
include/linux/early_res.h
1   -#ifndef _LINUX_EARLY_RES_H
2   -#define _LINUX_EARLY_RES_H
3   -#ifdef __KERNEL__
4   -
5   -extern void reserve_early(u64 start, u64 end, char *name);
6   -extern void reserve_early_overlap_ok(u64 start, u64 end, char *name);
7   -extern void free_early(u64 start, u64 end);
8   -void free_early_partial(u64 start, u64 end);
9   -extern void early_res_to_bootmem(u64 start, u64 end);
10   -
11   -void reserve_early_without_check(u64 start, u64 end, char *name);
12   -u64 find_early_area(u64 ei_start, u64 ei_last, u64 start, u64 end,
13   - u64 size, u64 align);
14   -u64 find_early_area_size(u64 ei_start, u64 ei_last, u64 start,
15   - u64 *sizep, u64 align);
16   -u64 find_fw_memmap_area(u64 start, u64 end, u64 size, u64 align);
17   -u64 get_max_mapped(void);
18   -#include <linux/range.h>
19   -int get_free_all_memory_range(struct range **rangep, int nodeid);
20   -
21   -#endif /* __KERNEL__ */
22   -
23   -#endif /* _LINUX_EARLY_RES_H */
... ... @@ -11,7 +11,6 @@
11 11 hrtimer.o rwsem.o nsproxy.o srcu.o semaphore.o \
12 12 notifier.o ksysfs.o pm_qos_params.o sched_clock.o cred.o \
13 13 async.o range.o
14   -obj-$(CONFIG_HAVE_EARLY_RES) += early_res.o
15 14 obj-y += groups.o
16 15  
17 16 ifdef CONFIG_FUNCTION_TRACER
kernel/early_res.c
1   -/*
2   - * early_res, could be used to replace bootmem
3   - */
4   -#include <linux/kernel.h>
5   -#include <linux/types.h>
6   -#include <linux/init.h>
7   -#include <linux/bootmem.h>
8   -#include <linux/mm.h>
9   -#include <linux/early_res.h>
10   -#include <linux/slab.h>
11   -#include <linux/kmemleak.h>
12   -
13   -/*
14   - * Early reserved memory areas.
15   - */
16   -/*
17   - * need to make sure this one is bigger enough before
18   - * find_fw_memmap_area could be used
19   - */
20   -#define MAX_EARLY_RES_X 32
21   -
22   -struct early_res {
23   - u64 start, end;
24   - char name[15];
25   - char overlap_ok;
26   -};
27   -static struct early_res early_res_x[MAX_EARLY_RES_X] __initdata;
28   -
29   -static int max_early_res __initdata = MAX_EARLY_RES_X;
30   -static struct early_res *early_res __initdata = &early_res_x[0];
31   -static int early_res_count __initdata;
32   -
33   -static int __init find_overlapped_early(u64 start, u64 end)
34   -{
35   - int i;
36   - struct early_res *r;
37   -
38   - for (i = 0; i < max_early_res && early_res[i].end; i++) {
39   - r = &early_res[i];
40   - if (end > r->start && start < r->end)
41   - break;
42   - }
43   -
44   - return i;
45   -}
46   -
47   -/*
48   - * Drop the i-th range from the early reservation map,
49   - * by copying any higher ranges down one over it, and
50   - * clearing what had been the last slot.
51   - */
52   -static void __init drop_range(int i)
53   -{
54   - int j;
55   -
56   - for (j = i + 1; j < max_early_res && early_res[j].end; j++)
57   - ;
58   -
59   - memmove(&early_res[i], &early_res[i + 1],
60   - (j - 1 - i) * sizeof(struct early_res));
61   -
62   - early_res[j - 1].end = 0;
63   - early_res_count--;
64   -}
65   -
66   -static void __init drop_range_partial(int i, u64 start, u64 end)
67   -{
68   - u64 common_start, common_end;
69   - u64 old_start, old_end;
70   -
71   - old_start = early_res[i].start;
72   - old_end = early_res[i].end;
73   - common_start = max(old_start, start);
74   - common_end = min(old_end, end);
75   -
76   - /* no overlap ? */
77   - if (common_start >= common_end)
78   - return;
79   -
80   - if (old_start < common_start) {
81   - /* make head segment */
82   - early_res[i].end = common_start;
83   - if (old_end > common_end) {
84   - char name[15];
85   -
86   - /*
87   - * Save a local copy of the name, since the
88   - * early_res array could get resized inside
89   - * reserve_early_without_check() ->
90   - * __check_and_double_early_res(), which would
91   - * make the current name pointer invalid.
92   - */
93   - strncpy(name, early_res[i].name,
94   - sizeof(early_res[i].name) - 1);
95   - /* add another for left over on tail */
96   - reserve_early_without_check(common_end, old_end, name);
97   - }
98   - return;
99   - } else {
100   - if (old_end > common_end) {
101   - /* reuse the entry for tail left */
102   - early_res[i].start = common_end;
103   - return;
104   - }
105   - /* all covered */
106   - drop_range(i);
107   - }
108   -}
109   -
110   -/*
111   - * Split any existing ranges that:
112   - * 1) are marked 'overlap_ok', and
113   - * 2) overlap with the stated range [start, end)
114   - * into whatever portion (if any) of the existing range is entirely
115   - * below or entirely above the stated range. Drop the portion
116   - * of the existing range that overlaps with the stated range,
117   - * which will allow the caller of this routine to then add that
118   - * stated range without conflicting with any existing range.
119   - */
120   -static void __init drop_overlaps_that_are_ok(u64 start, u64 end)
121   -{
122   - int i;
123   - struct early_res *r;
124   - u64 lower_start, lower_end;
125   - u64 upper_start, upper_end;
126   - char name[15];
127   -
128   - for (i = 0; i < max_early_res && early_res[i].end; i++) {
129   - r = &early_res[i];
130   -
131   - /* Continue past non-overlapping ranges */
132   - if (end <= r->start || start >= r->end)
133   - continue;
134   -
135   - /*
136   - * Leave non-ok overlaps as is; let caller
137   - * panic "Overlapping early reservations"
138   - * when it hits this overlap.
139   - */
140   - if (!r->overlap_ok)
141   - return;
142   -
143   - /*
144   - * We have an ok overlap. We will drop it from the early
145   - * reservation map, and add back in any non-overlapping
146   - * portions (lower or upper) as separate, overlap_ok,
147   - * non-overlapping ranges.
148   - */
149   -
150   - /* 1. Note any non-overlapping (lower or upper) ranges. */
151   - strncpy(name, r->name, sizeof(name) - 1);
152   -
153   - lower_start = lower_end = 0;
154   - upper_start = upper_end = 0;
155   - if (r->start < start) {
156   - lower_start = r->start;
157   - lower_end = start;
158   - }
159   - if (r->end > end) {
160   - upper_start = end;
161   - upper_end = r->end;
162   - }
163   -
164   - /* 2. Drop the original ok overlapping range */
165   - drop_range(i);
166   -
167   - i--; /* resume for-loop on copied down entry */
168   -
169   - /* 3. Add back in any non-overlapping ranges. */
170   - if (lower_end)
171   - reserve_early_overlap_ok(lower_start, lower_end, name);
172   - if (upper_end)
173   - reserve_early_overlap_ok(upper_start, upper_end, name);
174   - }
175   -}
176   -
177   -static void __init __reserve_early(u64 start, u64 end, char *name,
178   - int overlap_ok)
179   -{
180   - int i;
181   - struct early_res *r;
182   -
183   - i = find_overlapped_early(start, end);
184   - if (i >= max_early_res)
185   - panic("Too many early reservations");
186   - r = &early_res[i];
187   - if (r->end)
188   - panic("Overlapping early reservations "
189   - "%llx-%llx %s to %llx-%llx %s\n",
190   - start, end - 1, name ? name : "", r->start,
191   - r->end - 1, r->name);
192   - r->start = start;
193   - r->end = end;
194   - r->overlap_ok = overlap_ok;
195   - if (name)
196   - strncpy(r->name, name, sizeof(r->name) - 1);
197   - early_res_count++;
198   -}
199   -
200   -/*
201   - * A few early reservtations come here.
202   - *
203   - * The 'overlap_ok' in the name of this routine does -not- mean it
204   - * is ok for these reservations to overlap an earlier reservation.
205   - * Rather it means that it is ok for subsequent reservations to
206   - * overlap this one.
207   - *
208   - * Use this entry point to reserve early ranges when you are doing
209   - * so out of "Paranoia", reserving perhaps more memory than you need,
210   - * just in case, and don't mind a subsequent overlapping reservation
211   - * that is known to be needed.
212   - *
213   - * The drop_overlaps_that_are_ok() call here isn't really needed.
214   - * It would be needed if we had two colliding 'overlap_ok'
215   - * reservations, so that the second such would not panic on the
216   - * overlap with the first. We don't have any such as of this
217   - * writing, but might as well tolerate such if it happens in
218   - * the future.
219   - */
220   -void __init reserve_early_overlap_ok(u64 start, u64 end, char *name)
221   -{
222   - drop_overlaps_that_are_ok(start, end);
223   - __reserve_early(start, end, name, 1);
224   -}
225   -
226   -static void __init __check_and_double_early_res(u64 ex_start, u64 ex_end)
227   -{
228   - u64 start, end, size, mem;
229   - struct early_res *new;
230   -
231   - /* do we have enough slots left ? */
232   - if ((max_early_res - early_res_count) > max(max_early_res/8, 2))
233   - return;
234   -
235   - /* double it */
236   - mem = -1ULL;
237   - size = sizeof(struct early_res) * max_early_res * 2;
238   - if (early_res == early_res_x)
239   - start = 0;
240   - else
241   - start = early_res[0].end;
242   - end = ex_start;
243   - if (start + size < end)
244   - mem = find_fw_memmap_area(start, end, size,
245   - sizeof(struct early_res));
246   - if (mem == -1ULL) {
247   - start = ex_end;
248   - end = get_max_mapped();
249   - if (start + size < end)
250   - mem = find_fw_memmap_area(start, end, size,
251   - sizeof(struct early_res));
252   - }
253   - if (mem == -1ULL)
254   - panic("can not find more space for early_res array");
255   -
256   - new = __va(mem);
257   - /* save the first one for own */
258   - new[0].start = mem;
259   - new[0].end = mem + size;
260   - new[0].overlap_ok = 0;
261   - /* copy old to new */
262   - if (early_res == early_res_x) {
263   - memcpy(&new[1], &early_res[0],
264   - sizeof(struct early_res) * max_early_res);
265   - memset(&new[max_early_res+1], 0,
266   - sizeof(struct early_res) * (max_early_res - 1));
267   - early_res_count++;
268   - } else {
269   - memcpy(&new[1], &early_res[1],
270   - sizeof(struct early_res) * (max_early_res - 1));
271   - memset(&new[max_early_res], 0,
272   - sizeof(struct early_res) * max_early_res);
273   - }
274   - memset(&early_res[0], 0, sizeof(struct early_res) * max_early_res);
275   - early_res = new;
276   - max_early_res *= 2;
277   - printk(KERN_DEBUG "early_res array is doubled to %d at [%llx - %llx]\n",
278   - max_early_res, mem, mem + size - 1);
279   -}
280   -
281   -/*
282   - * Most early reservations come here.
283   - *
284   - * We first have drop_overlaps_that_are_ok() drop any pre-existing
285   - * 'overlap_ok' ranges, so that we can then reserve this memory
286   - * range without risk of panic'ing on an overlapping overlap_ok
287   - * early reservation.
288   - */
289   -void __init reserve_early(u64 start, u64 end, char *name)
290   -{
291   - if (start >= end)
292   - return;
293   -
294   - __check_and_double_early_res(start, end);
295   -
296   - drop_overlaps_that_are_ok(start, end);
297   - __reserve_early(start, end, name, 0);
298   -}
299   -
300   -void __init reserve_early_without_check(u64 start, u64 end, char *name)
301   -{
302   - struct early_res *r;
303   -
304   - if (start >= end)
305   - return;
306   -
307   - __check_and_double_early_res(start, end);
308   -
309   - r = &early_res[early_res_count];
310   -
311   - r->start = start;
312   - r->end = end;
313   - r->overlap_ok = 0;
314   - if (name)
315   - strncpy(r->name, name, sizeof(r->name) - 1);
316   - early_res_count++;
317   -}
318   -
319   -void __init free_early(u64 start, u64 end)
320   -{
321   - struct early_res *r;
322   - int i;
323   -
324   - kmemleak_free_part(__va(start), end - start);
325   -
326   - i = find_overlapped_early(start, end);
327   - r = &early_res[i];
328   - if (i >= max_early_res || r->end != end || r->start != start)
329   - panic("free_early on not reserved area: %llx-%llx!",
330   - start, end - 1);
331   -
332   - drop_range(i);
333   -}
334   -
335   -void __init free_early_partial(u64 start, u64 end)
336   -{
337   - struct early_res *r;
338   - int i;
339   -
340   - kmemleak_free_part(__va(start), end - start);
341   -
342   - if (start == end)
343   - return;
344   -
345   - if (WARN_ONCE(start > end, " wrong range [%#llx, %#llx]\n", start, end))
346   - return;
347   -
348   -try_next:
349   - i = find_overlapped_early(start, end);
350   - if (i >= max_early_res)
351   - return;
352   -
353   - r = &early_res[i];
354   - /* hole ? */
355   - if (r->end >= end && r->start <= start) {
356   - drop_range_partial(i, start, end);
357   - return;
358   - }
359   -
360   - drop_range_partial(i, start, end);
361   - goto try_next;
362   -}
363   -
364   -#ifdef CONFIG_NO_BOOTMEM
365   -static void __init subtract_early_res(struct range *range, int az)
366   -{
367   - int i, count;
368   - u64 final_start, final_end;
369   - int idx = 0;
370   -
371   - count = 0;
372   - for (i = 0; i < max_early_res && early_res[i].end; i++)
373   - count++;
374   -
375   - /* need to skip first one ?*/
376   - if (early_res != early_res_x)
377   - idx = 1;
378   -
379   -#define DEBUG_PRINT_EARLY_RES 1
380   -
381   -#if DEBUG_PRINT_EARLY_RES
382   - printk(KERN_INFO "Subtract (%d early reservations)\n", count);
383   -#endif
384   - for (i = idx; i < count; i++) {
385   - struct early_res *r = &early_res[i];
386   -#if DEBUG_PRINT_EARLY_RES
387   - printk(KERN_INFO " #%d [%010llx - %010llx] %15s\n", i,
388   - r->start, r->end, r->name);
389   -#endif
390   - final_start = PFN_DOWN(r->start);
391   - final_end = PFN_UP(r->end);
392   - if (final_start >= final_end)
393   - continue;
394   - subtract_range(range, az, final_start, final_end);
395   - }
396   -
397   -}
398   -
399   -int __init get_free_all_memory_range(struct range **rangep, int nodeid)
400   -{
401   - int i, count;
402   - u64 start = 0, end;
403   - u64 size;
404   - u64 mem;
405   - struct range *range;
406   - int nr_range;
407   -
408   - count = 0;
409   - for (i = 0; i < max_early_res && early_res[i].end; i++)
410   - count++;
411   -
412   - count *= 2;
413   -
414   - size = sizeof(struct range) * count;
415   - end = get_max_mapped();
416   -#ifdef MAX_DMA32_PFN
417   - if (end > (MAX_DMA32_PFN << PAGE_SHIFT))
418   - start = MAX_DMA32_PFN << PAGE_SHIFT;
419   -#endif
420   - mem = find_fw_memmap_area(start, end, size, sizeof(struct range));
421   - if (mem == -1ULL)
422   - panic("can not find more space for range free");
423   -
424   - range = __va(mem);
425   - /* use early_node_map[] and early_res to get range array at first */
426   - memset(range, 0, size);
427   - nr_range = 0;
428   -
429   - /* need to go over early_node_map to find out good range for node */
430   - nr_range = add_from_early_node_map(range, count, nr_range, nodeid);
431   -#ifdef CONFIG_X86_32
432   - subtract_range(range, count, max_low_pfn, -1ULL);
433   -#endif
434   - subtract_early_res(range, count);
435   - nr_range = clean_sort_range(range, count);
436   -
437   - /* need to clear it ? */
438   - if (nodeid == MAX_NUMNODES) {
439   - memset(&early_res[0], 0,
440   - sizeof(struct early_res) * max_early_res);
441   - early_res = NULL;
442   - max_early_res = 0;
443   - }
444   -
445   - *rangep = range;
446   - return nr_range;
447   -}
448   -#else
449   -void __init early_res_to_bootmem(u64 start, u64 end)
450   -{
451   - int i, count;
452   - u64 final_start, final_end;
453   - int idx = 0;
454   -
455   - count = 0;
456   - for (i = 0; i < max_early_res && early_res[i].end; i++)
457   - count++;
458   -
459   - /* need to skip first one ?*/
460   - if (early_res != early_res_x)
461   - idx = 1;
462   -
463   - printk(KERN_INFO "(%d/%d early reservations) ==> bootmem [%010llx - %010llx]\n",
464   - count - idx, max_early_res, start, end);
465   - for (i = idx; i < count; i++) {
466   - struct early_res *r = &early_res[i];
467   - printk(KERN_INFO " #%d [%010llx - %010llx] %16s", i,
468   - r->start, r->end, r->name);
469   - final_start = max(start, r->start);
470   - final_end = min(end, r->end);
471   - if (final_start >= final_end) {
472   - printk(KERN_CONT "\n");
473   - continue;
474   - }
475   - printk(KERN_CONT " ==> [%010llx - %010llx]\n",
476   - final_start, final_end);
477   - reserve_bootmem_generic(final_start, final_end - final_start,
478   - BOOTMEM_DEFAULT);
479   - }
480   - /* clear them */
481   - memset(&early_res[0], 0, sizeof(struct early_res) * max_early_res);
482   - early_res = NULL;
483   - max_early_res = 0;
484   - early_res_count = 0;
485   -}
486   -#endif
487   -
488   -/* Check for already reserved areas */
489   -static inline int __init bad_addr(u64 *addrp, u64 size, u64 align)
490   -{
491   - int i;
492   - u64 addr = *addrp;
493   - int changed = 0;
494   - struct early_res *r;
495   -again:
496   - i = find_overlapped_early(addr, addr + size);
497   - r = &early_res[i];
498   - if (i < max_early_res && r->end) {
499   - *addrp = addr = round_up(r->end, align);
500   - changed = 1;
501   - goto again;
502   - }
503   - return changed;
504   -}
505   -
506   -/* Check for already reserved areas */
507   -static inline int __init bad_addr_size(u64 *addrp, u64 *sizep, u64 align)
508   -{
509   - int i;
510   - u64 addr = *addrp, last;
511   - u64 size = *sizep;
512   - int changed = 0;
513   -again:
514   - last = addr + size;
515   - for (i = 0; i < max_early_res && early_res[i].end; i++) {
516   - struct early_res *r = &early_res[i];
517   - if (last > r->start && addr < r->start) {
518   - size = r->start - addr;
519   - changed = 1;
520   - goto again;
521   - }
522   - if (last > r->end && addr < r->end) {
523   - addr = round_up(r->end, align);
524   - size = last - addr;
525   - changed = 1;
526   - goto again;
527   - }
528   - if (last <= r->end && addr >= r->start) {
529   - (*sizep)++;
530   - return 0;
531   - }
532   - }
533   - if (changed) {
534   - *addrp = addr;
535   - *sizep = size;
536   - }
537   - return changed;
538   -}
539   -
540   -/*
541   - * Find a free area with specified alignment in a specific range.
542   - * only with the area.between start to end is active range from early_node_map
543   - * so they are good as RAM
544   - */
545   -u64 __init find_early_area(u64 ei_start, u64 ei_last, u64 start, u64 end,
546   - u64 size, u64 align)
547   -{
548   - u64 addr, last;
549   -
550   - addr = round_up(ei_start, align);
551   - if (addr < start)
552   - addr = round_up(start, align);
553   - if (addr >= ei_last)
554   - goto out;
555   - while (bad_addr(&addr, size, align) && addr+size <= ei_last)
556   - ;
557   - last = addr + size;
558   - if (last > ei_last)
559   - goto out;
560   - if (last > end)
561   - goto out;
562   -
563   - return addr;
564   -
565   -out:
566   - return -1ULL;
567   -}
568   -
569   -u64 __init find_early_area_size(u64 ei_start, u64 ei_last, u64 start,
570   - u64 *sizep, u64 align)
571   -{
572   - u64 addr, last;
573   -
574   - addr = round_up(ei_start, align);
575   - if (addr < start)
576   - addr = round_up(start, align);
577   - if (addr >= ei_last)
578   - goto out;
579   - *sizep = ei_last - addr;
580   - while (bad_addr_size(&addr, sizep, align) && addr + *sizep <= ei_last)
581   - ;
582   - last = addr + *sizep;
583   - if (last > ei_last)
584   - goto out;
585   -
586   - return addr;
587   -
588   -out:
589   - return -1ULL;
590   -}