Commit dc2786f0c19a779395ef69189dd5e7df2573b29b
Committed by
Jiri Slaby
1 parent
ee1760b2b4
Exists in
ti-linux-3.12.y
and in
2 other branches
mm: page_alloc: calculate classzone_idx once from the zonelist ref
commit d8846374a85f4290a473a4e2a64c1ba046c4a0e1 upstream. There is no need to calculate zone_idx(preferred_zone) multiple times or use the pgdat to figure it out. Signed-off-by: Mel Gorman <mgorman@suse.de> Acked-by: Rik van Riel <riel@redhat.com> Acked-by: David Rientjes <rientjes@google.com> Cc: Johannes Weiner <hannes@cmpxchg.org> Cc: Vlastimil Babka <vbabka@suse.cz> Cc: Jan Kara <jack@suse.cz> Cc: Michal Hocko <mhocko@suse.cz> Cc: Hugh Dickins <hughd@google.com> Cc: Dave Hansen <dave.hansen@intel.com> Cc: Theodore Ts'o <tytso@mit.edu> Cc: "Paul E. McKenney" <paulmck@linux.vnet.ibm.com> Cc: Oleg Nesterov <oleg@redhat.com> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Dan Carpenter <dan.carpenter@oracle.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org> Signed-off-by: Mel Gorman <mgorman@suse.de> Signed-off-by: Jiri Slaby <jslaby@suse.cz>
Showing 1 changed file with 35 additions and 25 deletions Side-by-side Diff
mm/page_alloc.c
... | ... | @@ -1897,17 +1897,15 @@ |
1897 | 1897 | static struct page * |
1898 | 1898 | get_page_from_freelist(gfp_t gfp_mask, nodemask_t *nodemask, unsigned int order, |
1899 | 1899 | struct zonelist *zonelist, int high_zoneidx, int alloc_flags, |
1900 | - struct zone *preferred_zone, int migratetype) | |
1900 | + struct zone *preferred_zone, int classzone_idx, int migratetype) | |
1901 | 1901 | { |
1902 | 1902 | struct zoneref *z; |
1903 | 1903 | struct page *page = NULL; |
1904 | - int classzone_idx; | |
1905 | 1904 | struct zone *zone; |
1906 | 1905 | nodemask_t *allowednodes = NULL;/* zonelist_cache approximation */ |
1907 | 1906 | int zlc_active = 0; /* set if using zonelist_cache */ |
1908 | 1907 | int did_zlc_setup = 0; /* just call zlc_setup() one time */ |
1909 | 1908 | |
1910 | - classzone_idx = zone_idx(preferred_zone); | |
1911 | 1909 | zonelist_scan: |
1912 | 1910 | /* |
1913 | 1911 | * Scan zonelist, looking for a zone with enough free. |
... | ... | @@ -2171,7 +2169,7 @@ |
2171 | 2169 | __alloc_pages_may_oom(gfp_t gfp_mask, unsigned int order, |
2172 | 2170 | struct zonelist *zonelist, enum zone_type high_zoneidx, |
2173 | 2171 | nodemask_t *nodemask, struct zone *preferred_zone, |
2174 | - int migratetype) | |
2172 | + int classzone_idx, int migratetype) | |
2175 | 2173 | { |
2176 | 2174 | struct page *page; |
2177 | 2175 | |
... | ... | @@ -2189,7 +2187,7 @@ |
2189 | 2187 | page = get_page_from_freelist(gfp_mask|__GFP_HARDWALL, nodemask, |
2190 | 2188 | order, zonelist, high_zoneidx, |
2191 | 2189 | ALLOC_WMARK_HIGH|ALLOC_CPUSET, |
2192 | - preferred_zone, migratetype); | |
2190 | + preferred_zone, classzone_idx, migratetype); | |
2193 | 2191 | if (page) |
2194 | 2192 | goto out; |
2195 | 2193 | |
... | ... | @@ -2224,7 +2222,7 @@ |
2224 | 2222 | __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, |
2225 | 2223 | struct zonelist *zonelist, enum zone_type high_zoneidx, |
2226 | 2224 | nodemask_t *nodemask, int alloc_flags, struct zone *preferred_zone, |
2227 | - int migratetype, enum migrate_mode mode, | |
2225 | + int classzone_idx, int migratetype, enum migrate_mode mode, | |
2228 | 2226 | bool *contended_compaction, bool *deferred_compaction, |
2229 | 2227 | unsigned long *did_some_progress) |
2230 | 2228 | { |
... | ... | @@ -2252,7 +2250,7 @@ |
2252 | 2250 | page = get_page_from_freelist(gfp_mask, nodemask, |
2253 | 2251 | order, zonelist, high_zoneidx, |
2254 | 2252 | alloc_flags & ~ALLOC_NO_WATERMARKS, |
2255 | - preferred_zone, migratetype); | |
2253 | + preferred_zone, classzone_idx, migratetype); | |
2256 | 2254 | if (page) { |
2257 | 2255 | preferred_zone->compact_blockskip_flush = false; |
2258 | 2256 | compaction_defer_reset(preferred_zone, order, true); |
... | ... | @@ -2284,7 +2282,8 @@ |
2284 | 2282 | __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order, |
2285 | 2283 | struct zonelist *zonelist, enum zone_type high_zoneidx, |
2286 | 2284 | nodemask_t *nodemask, int alloc_flags, struct zone *preferred_zone, |
2287 | - int migratetype, enum migrate_mode mode, bool *contended_compaction, | |
2285 | + int classzone_idx, int migratetype, | |
2286 | + enum migrate_mode mode, bool *contended_compaction, | |
2288 | 2287 | bool *deferred_compaction, unsigned long *did_some_progress) |
2289 | 2288 | { |
2290 | 2289 | return NULL; |
... | ... | @@ -2324,7 +2323,7 @@ |
2324 | 2323 | __alloc_pages_direct_reclaim(gfp_t gfp_mask, unsigned int order, |
2325 | 2324 | struct zonelist *zonelist, enum zone_type high_zoneidx, |
2326 | 2325 | nodemask_t *nodemask, int alloc_flags, struct zone *preferred_zone, |
2327 | - int migratetype, unsigned long *did_some_progress) | |
2326 | + int classzone_idx, int migratetype, unsigned long *did_some_progress) | |
2328 | 2327 | { |
2329 | 2328 | struct page *page = NULL; |
2330 | 2329 | bool drained = false; |
... | ... | @@ -2342,7 +2341,8 @@ |
2342 | 2341 | page = get_page_from_freelist(gfp_mask, nodemask, order, |
2343 | 2342 | zonelist, high_zoneidx, |
2344 | 2343 | alloc_flags & ~ALLOC_NO_WATERMARKS, |
2345 | - preferred_zone, migratetype); | |
2344 | + preferred_zone, classzone_idx, | |
2345 | + migratetype); | |
2346 | 2346 | |
2347 | 2347 | /* |
2348 | 2348 | * If an allocation failed after direct reclaim, it could be because |
2349 | 2349 | |
... | ... | @@ -2365,14 +2365,14 @@ |
2365 | 2365 | __alloc_pages_high_priority(gfp_t gfp_mask, unsigned int order, |
2366 | 2366 | struct zonelist *zonelist, enum zone_type high_zoneidx, |
2367 | 2367 | nodemask_t *nodemask, struct zone *preferred_zone, |
2368 | - int migratetype) | |
2368 | + int classzone_idx, int migratetype) | |
2369 | 2369 | { |
2370 | 2370 | struct page *page; |
2371 | 2371 | |
2372 | 2372 | do { |
2373 | 2373 | page = get_page_from_freelist(gfp_mask, nodemask, order, |
2374 | 2374 | zonelist, high_zoneidx, ALLOC_NO_WATERMARKS, |
2375 | - preferred_zone, migratetype); | |
2375 | + preferred_zone, classzone_idx, migratetype); | |
2376 | 2376 | |
2377 | 2377 | if (!page && gfp_mask & __GFP_NOFAIL) |
2378 | 2378 | wait_iff_congested(preferred_zone, BLK_RW_ASYNC, HZ/50); |
... | ... | @@ -2473,7 +2473,7 @@ |
2473 | 2473 | __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order, |
2474 | 2474 | struct zonelist *zonelist, enum zone_type high_zoneidx, |
2475 | 2475 | nodemask_t *nodemask, struct zone *preferred_zone, |
2476 | - int migratetype) | |
2476 | + int classzone_idx, int migratetype) | |
2477 | 2477 | { |
2478 | 2478 | const gfp_t wait = gfp_mask & __GFP_WAIT; |
2479 | 2479 | struct page *page = NULL; |
2480 | 2480 | |
... | ... | @@ -2522,15 +2522,19 @@ |
2522 | 2522 | * Find the true preferred zone if the allocation is unconstrained by |
2523 | 2523 | * cpusets. |
2524 | 2524 | */ |
2525 | - if (!(alloc_flags & ALLOC_CPUSET) && !nodemask) | |
2526 | - first_zones_zonelist(zonelist, high_zoneidx, NULL, | |
2527 | - &preferred_zone); | |
2525 | + if (!(alloc_flags & ALLOC_CPUSET) && !nodemask) { | |
2526 | + struct zoneref *preferred_zoneref; | |
2527 | + preferred_zoneref = first_zones_zonelist(zonelist, high_zoneidx, | |
2528 | + NULL, | |
2529 | + &preferred_zone); | |
2530 | + classzone_idx = zonelist_zone_idx(preferred_zoneref); | |
2531 | + } | |
2528 | 2532 | |
2529 | 2533 | rebalance: |
2530 | 2534 | /* This is the last chance, in general, before the goto nopage. */ |
2531 | 2535 | page = get_page_from_freelist(gfp_mask, nodemask, order, zonelist, |
2532 | 2536 | high_zoneidx, alloc_flags & ~ALLOC_NO_WATERMARKS, |
2533 | - preferred_zone, migratetype); | |
2537 | + preferred_zone, classzone_idx, migratetype); | |
2534 | 2538 | if (page) |
2535 | 2539 | goto got_pg; |
2536 | 2540 | |
... | ... | @@ -2545,7 +2549,7 @@ |
2545 | 2549 | |
2546 | 2550 | page = __alloc_pages_high_priority(gfp_mask, order, |
2547 | 2551 | zonelist, high_zoneidx, nodemask, |
2548 | - preferred_zone, migratetype); | |
2552 | + preferred_zone, classzone_idx, migratetype); | |
2549 | 2553 | if (page) { |
2550 | 2554 | goto got_pg; |
2551 | 2555 | } |
... | ... | @@ -2569,7 +2573,8 @@ |
2569 | 2573 | */ |
2570 | 2574 | page = __alloc_pages_direct_compact(gfp_mask, order, zonelist, |
2571 | 2575 | high_zoneidx, nodemask, alloc_flags, |
2572 | - preferred_zone, migratetype, | |
2576 | + preferred_zone, | |
2577 | + classzone_idx, migratetype, | |
2573 | 2578 | migration_mode, &contended_compaction, |
2574 | 2579 | &deferred_compaction, |
2575 | 2580 | &did_some_progress); |
... | ... | @@ -2592,7 +2597,8 @@ |
2592 | 2597 | zonelist, high_zoneidx, |
2593 | 2598 | nodemask, |
2594 | 2599 | alloc_flags, preferred_zone, |
2595 | - migratetype, &did_some_progress); | |
2600 | + classzone_idx, migratetype, | |
2601 | + &did_some_progress); | |
2596 | 2602 | if (page) |
2597 | 2603 | goto got_pg; |
2598 | 2604 | |
... | ... | @@ -2611,7 +2617,7 @@ |
2611 | 2617 | page = __alloc_pages_may_oom(gfp_mask, order, |
2612 | 2618 | zonelist, high_zoneidx, |
2613 | 2619 | nodemask, preferred_zone, |
2614 | - migratetype); | |
2620 | + classzone_idx, migratetype); | |
2615 | 2621 | if (page) |
2616 | 2622 | goto got_pg; |
2617 | 2623 | |
... | ... | @@ -2652,7 +2658,8 @@ |
2652 | 2658 | */ |
2653 | 2659 | page = __alloc_pages_direct_compact(gfp_mask, order, zonelist, |
2654 | 2660 | high_zoneidx, nodemask, alloc_flags, |
2655 | - preferred_zone, migratetype, | |
2661 | + preferred_zone, | |
2662 | + classzone_idx, migratetype, | |
2656 | 2663 | migration_mode, &contended_compaction, |
2657 | 2664 | &deferred_compaction, |
2658 | 2665 | &did_some_progress); |
2659 | 2666 | |
... | ... | @@ -2679,11 +2686,13 @@ |
2679 | 2686 | { |
2680 | 2687 | enum zone_type high_zoneidx = gfp_zone(gfp_mask); |
2681 | 2688 | struct zone *preferred_zone; |
2689 | + struct zoneref *preferred_zoneref; | |
2682 | 2690 | struct page *page = NULL; |
2683 | 2691 | int migratetype = allocflags_to_migratetype(gfp_mask); |
2684 | 2692 | unsigned int cpuset_mems_cookie; |
2685 | 2693 | int alloc_flags = ALLOC_WMARK_LOW|ALLOC_CPUSET|ALLOC_FAIR; |
2686 | 2694 | struct mem_cgroup *memcg = NULL; |
2695 | + int classzone_idx; | |
2687 | 2696 | |
2688 | 2697 | gfp_mask &= gfp_allowed_mask; |
2689 | 2698 | |
2690 | 2699 | |
... | ... | @@ -2713,11 +2722,12 @@ |
2713 | 2722 | cpuset_mems_cookie = read_mems_allowed_begin(); |
2714 | 2723 | |
2715 | 2724 | /* The preferred zone is used for statistics later */ |
2716 | - first_zones_zonelist(zonelist, high_zoneidx, | |
2725 | + preferred_zoneref = first_zones_zonelist(zonelist, high_zoneidx, | |
2717 | 2726 | nodemask ? : &cpuset_current_mems_allowed, |
2718 | 2727 | &preferred_zone); |
2719 | 2728 | if (!preferred_zone) |
2720 | 2729 | goto out; |
2730 | + classzone_idx = zonelist_zone_idx(preferred_zoneref); | |
2721 | 2731 | |
2722 | 2732 | #ifdef CONFIG_CMA |
2723 | 2733 | if (allocflags_to_migratetype(gfp_mask) == MIGRATE_MOVABLE) |
... | ... | @@ -2727,7 +2737,7 @@ |
2727 | 2737 | /* First allocation attempt */ |
2728 | 2738 | page = get_page_from_freelist(gfp_mask|__GFP_HARDWALL, nodemask, order, |
2729 | 2739 | zonelist, high_zoneidx, alloc_flags, |
2730 | - preferred_zone, migratetype); | |
2740 | + preferred_zone, classzone_idx, migratetype); | |
2731 | 2741 | if (unlikely(!page)) { |
2732 | 2742 | /* |
2733 | 2743 | * The first pass makes sure allocations are spread |
... | ... | @@ -2753,7 +2763,7 @@ |
2753 | 2763 | gfp_mask = memalloc_noio_flags(gfp_mask); |
2754 | 2764 | page = __alloc_pages_slowpath(gfp_mask, order, |
2755 | 2765 | zonelist, high_zoneidx, nodemask, |
2756 | - preferred_zone, migratetype); | |
2766 | + preferred_zone, classzone_idx, migratetype); | |
2757 | 2767 | } |
2758 | 2768 | |
2759 | 2769 | trace_mm_page_alloc(page, order, gfp_mask, migratetype); |