Blame view

mm/swap_state.c 12.7 KB
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1
2
3
4
5
6
7
8
  /*
   *  linux/mm/swap_state.c
   *
   *  Copyright (C) 1991, 1992, 1993, 1994  Linus Torvalds
   *  Swap reorganised 29.12.95, Stephen Tweedie
   *
   *  Rewritten to use page cache, (C) 1998 Stephen Tweedie
   */
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
9
  #include <linux/mm.h>
5a0e3ad6a   Tejun Heo   include cleanup: ...
10
  #include <linux/gfp.h>
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
11
12
  #include <linux/kernel_stat.h>
  #include <linux/swap.h>
46017e954   Hugh Dickins   swapin_readahead:...
13
  #include <linux/swapops.h>
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
14
15
  #include <linux/init.h>
  #include <linux/pagemap.h>
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
16
  #include <linux/backing-dev.h>
3fb5c298b   Christian Ehrhardt   swap: allow swap ...
17
  #include <linux/blkdev.h>
c484d4104   Hugh Dickins   [PATCH] mm: free_...
18
  #include <linux/pagevec.h>
b20a35035   Christoph Lameter   [PATCH] page migr...
19
  #include <linux/migrate.h>
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
20
21
22
23
24
  
  #include <asm/pgtable.h>
  
  /*
   * swapper_space is a fiction, retained to simplify the path through
7eaceacca   Jens Axboe   block: remove per...
25
   * vmscan's shrink_page_list.
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
26
   */
f5e54d6e5   Christoph Hellwig   [PATCH] mark addr...
27
  static const struct address_space_operations swap_aops = {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
28
  	.writepage	= swap_writepage,
62c230bc1   Mel Gorman   mm: add support f...
29
  	.set_page_dirty	= swap_set_page_dirty,
1c93923cc   Andrew Morton   include/linux/mig...
30
  #ifdef CONFIG_MIGRATION
e965f9630   Christoph Lameter   [PATCH] Direct Mi...
31
  	.migratepage	= migrate_page,
1c93923cc   Andrew Morton   include/linux/mig...
32
  #endif
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
33
  };
33806f06d   Shaohua Li   swap: make each s...
34
35
36
  struct address_space swapper_spaces[MAX_SWAPFILES] = {
  	[0 ... MAX_SWAPFILES - 1] = {
  		.page_tree	= RADIX_TREE_INIT(GFP_ATOMIC|__GFP_NOWARN),
4bb5f5d93   David Herrmann   mm: allow drivers...
37
  		.i_mmap_writable = ATOMIC_INIT(0),
33806f06d   Shaohua Li   swap: make each s...
38
  		.a_ops		= &swap_aops,
33806f06d   Shaohua Li   swap: make each s...
39
  	}
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
40
  };
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
41
42
43
44
45
46
47
48
  
  #define INC_CACHE_INFO(x)	do { swap_cache_info.x++; } while (0)
  
  static struct {
  	unsigned long add_total;
  	unsigned long del_total;
  	unsigned long find_success;
  	unsigned long find_total;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
49
  } swap_cache_info;
33806f06d   Shaohua Li   swap: make each s...
50
51
52
53
54
55
56
57
58
  unsigned long total_swapcache_pages(void)
  {
  	int i;
  	unsigned long ret = 0;
  
  	for (i = 0; i < MAX_SWAPFILES; i++)
  		ret += swapper_spaces[i].nrpages;
  	return ret;
  }
579f82901   Shaohua Li   swap: add a simpl...
59
  static atomic_t swapin_readahead_hits = ATOMIC_INIT(4);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
60
61
  void show_swap_cache_info(void)
  {
33806f06d   Shaohua Li   swap: make each s...
62
63
  	printk("%lu pages in swap cache
  ", total_swapcache_pages());
2c97b7fc0   Johannes Weiner   mm: print swapcac...
64
65
  	printk("Swap cache stats: add %lu, delete %lu, find %lu/%lu
  ",
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
66
  		swap_cache_info.add_total, swap_cache_info.del_total,
bb63be0a0   Hugh Dickins   tmpfs: move swap_...
67
  		swap_cache_info.find_success, swap_cache_info.find_total);
ec8acf20a   Shaohua Li   swap: add per-par...
68
69
70
  	printk("Free swap  = %ldkB
  ",
  		get_nr_swap_pages() << (PAGE_SHIFT - 10));
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
71
72
73
74
75
  	printk("Total swap = %lukB
  ", total_swap_pages << (PAGE_SHIFT - 10));
  }
  
  /*
31a563962   Daisuke Nishimura   mm: add_to_swap_c...
76
   * __add_to_swap_cache resembles add_to_page_cache_locked on swapper_space,
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
77
78
   * but sets SwapCache flag and private instead of mapping and index.
   */
2f772e6ca   Seth Jennings   mm: break up swap...
79
  int __add_to_swap_cache(struct page *page, swp_entry_t entry)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
80
81
  {
  	int error;
33806f06d   Shaohua Li   swap: make each s...
82
  	struct address_space *address_space;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
83

309381fea   Sasha Levin   mm: dump page whe...
84
85
86
  	VM_BUG_ON_PAGE(!PageLocked(page), page);
  	VM_BUG_ON_PAGE(PageSwapCache(page), page);
  	VM_BUG_ON_PAGE(!PageSwapBacked(page), page);
51726b122   Hugh Dickins   mm: replace some ...
87

31a563962   Daisuke Nishimura   mm: add_to_swap_c...
88
89
90
  	page_cache_get(page);
  	SetPageSwapCache(page);
  	set_page_private(page, entry.val);
33806f06d   Shaohua Li   swap: make each s...
91
92
93
94
  	address_space = swap_address_space(entry);
  	spin_lock_irq(&address_space->tree_lock);
  	error = radix_tree_insert(&address_space->page_tree,
  					entry.val, page);
31a563962   Daisuke Nishimura   mm: add_to_swap_c...
95
  	if (likely(!error)) {
33806f06d   Shaohua Li   swap: make each s...
96
  		address_space->nrpages++;
31a563962   Daisuke Nishimura   mm: add_to_swap_c...
97
98
99
  		__inc_zone_page_state(page, NR_FILE_PAGES);
  		INC_CACHE_INFO(add_total);
  	}
33806f06d   Shaohua Li   swap: make each s...
100
  	spin_unlock_irq(&address_space->tree_lock);
31a563962   Daisuke Nishimura   mm: add_to_swap_c...
101
102
  
  	if (unlikely(error)) {
2ca4532a4   Daisuke Nishimura   mm: add_to_swap_c...
103
104
105
106
107
108
  		/*
  		 * Only the context which have set SWAP_HAS_CACHE flag
  		 * would call add_to_swap_cache().
  		 * So add_to_swap_cache() doesn't returns -EEXIST.
  		 */
  		VM_BUG_ON(error == -EEXIST);
31a563962   Daisuke Nishimura   mm: add_to_swap_c...
109
110
111
112
113
114
115
116
117
118
119
120
  		set_page_private(page, 0UL);
  		ClearPageSwapCache(page);
  		page_cache_release(page);
  	}
  
  	return error;
  }
  
  
  int add_to_swap_cache(struct page *page, swp_entry_t entry, gfp_t gfp_mask)
  {
  	int error;
5e4c0d974   Jan Kara   lib/radix-tree.c:...
121
  	error = radix_tree_maybe_preload(gfp_mask);
35c754d79   Balbir Singh   memory controller...
122
  	if (!error) {
31a563962   Daisuke Nishimura   mm: add_to_swap_c...
123
  		error = __add_to_swap_cache(page, entry);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
124
  		radix_tree_preload_end();
fa1de9008   Hugh Dickins   memcgroup: revert...
125
  	}
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
126
127
  	return error;
  }
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
128
129
130
131
132
133
  /*
   * This must be called only on pages that have
   * been verified to be in the swap cache.
   */
  void __delete_from_swap_cache(struct page *page)
  {
33806f06d   Shaohua Li   swap: make each s...
134
135
  	swp_entry_t entry;
  	struct address_space *address_space;
309381fea   Sasha Levin   mm: dump page whe...
136
137
138
  	VM_BUG_ON_PAGE(!PageLocked(page), page);
  	VM_BUG_ON_PAGE(!PageSwapCache(page), page);
  	VM_BUG_ON_PAGE(PageWriteback(page), page);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
139

33806f06d   Shaohua Li   swap: make each s...
140
141
142
  	entry.val = page_private(page);
  	address_space = swap_address_space(entry);
  	radix_tree_delete(&address_space->page_tree, page_private(page));
4c21e2f24   Hugh Dickins   [PATCH] mm: split...
143
  	set_page_private(page, 0);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
144
  	ClearPageSwapCache(page);
33806f06d   Shaohua Li   swap: make each s...
145
  	address_space->nrpages--;
347ce434d   Christoph Lameter   [PATCH] zoned vm ...
146
  	__dec_zone_page_state(page, NR_FILE_PAGES);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
147
148
149
150
151
152
153
154
155
156
  	INC_CACHE_INFO(del_total);
  }
  
  /**
   * add_to_swap - allocate swap space for a page
   * @page: page we want to move to swap
   *
   * Allocate swap space for the page and add the page to the
   * swap cache.  Caller needs to hold the page lock. 
   */
5bc7b8aca   Shaohua Li   mm: thp: add spli...
157
  int add_to_swap(struct page *page, struct list_head *list)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
158
159
  {
  	swp_entry_t entry;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
160
  	int err;
309381fea   Sasha Levin   mm: dump page whe...
161
162
  	VM_BUG_ON_PAGE(!PageLocked(page), page);
  	VM_BUG_ON_PAGE(!PageUptodate(page), page);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
163

2ca4532a4   Daisuke Nishimura   mm: add_to_swap_c...
164
165
166
  	entry = get_swap_page();
  	if (!entry.val)
  		return 0;
3f04f62f9   Andrea Arcangeli   thp: split_huge_p...
167
  	if (unlikely(PageTransHuge(page)))
5bc7b8aca   Shaohua Li   mm: thp: add spli...
168
  		if (unlikely(split_huge_page_to_list(page, list))) {
0a31bc97c   Johannes Weiner   mm: memcontrol: r...
169
  			swapcache_free(entry);
3f04f62f9   Andrea Arcangeli   thp: split_huge_p...
170
171
  			return 0;
  		}
2ca4532a4   Daisuke Nishimura   mm: add_to_swap_c...
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
  	/*
  	 * Radix-tree node allocations from PF_MEMALLOC contexts could
  	 * completely exhaust the page allocator. __GFP_NOMEMALLOC
  	 * stops emergency reserves from being allocated.
  	 *
  	 * TODO: this could cause a theoretical memory reclaim
  	 * deadlock in the swap out path.
  	 */
  	/*
  	 * Add it to the swap cache and mark it dirty
  	 */
  	err = add_to_swap_cache(page, entry,
  			__GFP_HIGH|__GFP_NOMEMALLOC|__GFP_NOWARN);
  
  	if (!err) {	/* Success */
  		SetPageDirty(page);
  		return 1;
  	} else {	/* -ENOMEM radix-tree allocation failure */
bd53b714d   Nick Piggin   [PATCH] mm: use _...
190
  		/*
2ca4532a4   Daisuke Nishimura   mm: add_to_swap_c...
191
192
  		 * add_to_swap_cache() doesn't return -EEXIST, so we can safely
  		 * clear SWAP_HAS_CACHE flag.
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
193
  		 */
0a31bc97c   Johannes Weiner   mm: memcontrol: r...
194
  		swapcache_free(entry);
2ca4532a4   Daisuke Nishimura   mm: add_to_swap_c...
195
  		return 0;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
196
197
198
199
200
201
202
203
204
205
206
207
  	}
  }
  
  /*
   * This must be called only on pages that have
   * been verified to be in the swap cache and locked.
   * It will never put the page into the free list,
   * the caller has a reference on the page.
   */
  void delete_from_swap_cache(struct page *page)
  {
  	swp_entry_t entry;
33806f06d   Shaohua Li   swap: make each s...
208
  	struct address_space *address_space;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
209

4c21e2f24   Hugh Dickins   [PATCH] mm: split...
210
  	entry.val = page_private(page);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
211

33806f06d   Shaohua Li   swap: make each s...
212
213
  	address_space = swap_address_space(entry);
  	spin_lock_irq(&address_space->tree_lock);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
214
  	__delete_from_swap_cache(page);
33806f06d   Shaohua Li   swap: make each s...
215
  	spin_unlock_irq(&address_space->tree_lock);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
216

0a31bc97c   Johannes Weiner   mm: memcontrol: r...
217
  	swapcache_free(entry);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
218
219
  	page_cache_release(page);
  }
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
220
221
222
223
  /* 
   * If we are the only user, then try to free up the swap cache. 
   * 
   * Its ok to check for PageSwapCache without the page lock
a2c43eed8   Hugh Dickins   mm: try_to_free_s...
224
225
   * here because we are going to recheck again inside
   * try_to_free_swap() _with_ the lock.
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
226
227
228
229
   * 					- Marcelo
   */
  static inline void free_swap_cache(struct page *page)
  {
a2c43eed8   Hugh Dickins   mm: try_to_free_s...
230
231
  	if (PageSwapCache(page) && !page_mapped(page) && trylock_page(page)) {
  		try_to_free_swap(page);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
232
233
234
235
236
237
  		unlock_page(page);
  	}
  }
  
  /* 
   * Perform a free_page(), also freeing any swap cache associated with
b8072f099   Hugh Dickins   [PATCH] mm: updat...
238
   * this page if it is the last user of the page.
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
239
240
241
242
243
244
245
246
247
248
249
250
251
   */
  void free_page_and_swap_cache(struct page *page)
  {
  	free_swap_cache(page);
  	page_cache_release(page);
  }
  
  /*
   * Passed an array of pages, drop them all from swapcache and then release
   * them.  They are removed from the LRU and freed if this is their last use.
   */
  void free_pages_and_swap_cache(struct page **pages, int nr)
  {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
252
  	struct page **pagep = pages;
aabfb5729   Michal Hocko   mm: memcontrol: d...
253
  	int i;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
254
255
  
  	lru_add_drain();
aabfb5729   Michal Hocko   mm: memcontrol: d...
256
257
258
  	for (i = 0; i < nr; i++)
  		free_swap_cache(pagep[i]);
  	release_pages(pagep, nr, false);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
259
260
261
262
263
264
265
266
267
268
269
  }
  
  /*
   * Lookup a swap entry in the swap cache. A found page will be returned
   * unlocked and with its refcount incremented - we rely on the kernel
   * lock getting page table operations atomic even if we drop the page
   * lock before returning.
   */
  struct page * lookup_swap_cache(swp_entry_t entry)
  {
  	struct page *page;
33806f06d   Shaohua Li   swap: make each s...
270
  	page = find_get_page(swap_address_space(entry), entry.val);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
271

579f82901   Shaohua Li   swap: add a simpl...
272
  	if (page) {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
273
  		INC_CACHE_INFO(find_success);
579f82901   Shaohua Li   swap: add a simpl...
274
275
276
  		if (TestClearPageReadahead(page))
  			atomic_inc(&swapin_readahead_hits);
  	}
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
277
278
279
280
281
282
283
284
285
286
287
  
  	INC_CACHE_INFO(find_total);
  	return page;
  }
  
  /* 
   * Locate a page of swap in physical memory, reserving swap cache space
   * and reading the disk if it is not already cached.
   * A failure return means that either the page allocation failed or that
   * the swap entry is no longer in use.
   */
02098feaa   Hugh Dickins   swapin needs gfp_...
288
  struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask,
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
289
290
291
292
293
294
295
296
297
298
299
  			struct vm_area_struct *vma, unsigned long addr)
  {
  	struct page *found_page, *new_page = NULL;
  	int err;
  
  	do {
  		/*
  		 * First check the swap cache.  Since this is normally
  		 * called after lookup_swap_cache() failed, re-calling
  		 * that would confuse statistics.
  		 */
33806f06d   Shaohua Li   swap: make each s...
300
301
  		found_page = find_get_page(swap_address_space(entry),
  					entry.val);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
302
303
304
305
306
307
308
  		if (found_page)
  			break;
  
  		/*
  		 * Get a new page to read into from swap.
  		 */
  		if (!new_page) {
02098feaa   Hugh Dickins   swapin needs gfp_...
309
  			new_page = alloc_page_vma(gfp_mask, vma, addr);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
310
311
312
313
314
  			if (!new_page)
  				break;		/* Out of memory */
  		}
  
  		/*
31a563962   Daisuke Nishimura   mm: add_to_swap_c...
315
316
  		 * call radix_tree_preload() while we can wait.
  		 */
5e4c0d974   Jan Kara   lib/radix-tree.c:...
317
  		err = radix_tree_maybe_preload(gfp_mask & GFP_KERNEL);
31a563962   Daisuke Nishimura   mm: add_to_swap_c...
318
319
320
321
  		if (err)
  			break;
  
  		/*
f000944d0   Hugh Dickins   tmpfs: shuffle ad...
322
323
  		 * Swap entry may have been freed since our caller observed it.
  		 */
355cfa73d   KAMEZAWA Hiroyuki   mm: modify swap_m...
324
  		err = swapcache_prepare(entry);
cbab0e4ee   Rafael Aquini   swap: avoid read_...
325
  		if (err == -EEXIST) {
31a563962   Daisuke Nishimura   mm: add_to_swap_c...
326
  			radix_tree_preload_end();
cbab0e4ee   Rafael Aquini   swap: avoid read_...
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
  			/*
  			 * We might race against get_swap_page() and stumble
  			 * across a SWAP_HAS_CACHE swap_map entry whose page
  			 * has not been brought into the swapcache yet, while
  			 * the other end is scheduled away waiting on discard
  			 * I/O completion at scan_swap_map().
  			 *
  			 * In order to avoid turning this transitory state
  			 * into a permanent loop around this -EEXIST case
  			 * if !CONFIG_PREEMPT and the I/O completion happens
  			 * to be waiting on the CPU waitqueue where we are now
  			 * busy looping, we just conditionally invoke the
  			 * scheduler here, if there are some more important
  			 * tasks to run.
  			 */
  			cond_resched();
355cfa73d   KAMEZAWA Hiroyuki   mm: modify swap_m...
343
  			continue;
31a563962   Daisuke Nishimura   mm: add_to_swap_c...
344
345
346
  		}
  		if (err) {		/* swp entry is obsolete ? */
  			radix_tree_preload_end();
f000944d0   Hugh Dickins   tmpfs: shuffle ad...
347
  			break;
31a563962   Daisuke Nishimura   mm: add_to_swap_c...
348
  		}
f000944d0   Hugh Dickins   tmpfs: shuffle ad...
349

2ca4532a4   Daisuke Nishimura   mm: add_to_swap_c...
350
  		/* May fail (-ENOMEM) if radix-tree node allocation failed. */
f45840b5c   Nick Piggin   mm: pagecache ins...
351
  		__set_page_locked(new_page);
b2e185384   Rik van Riel   define page_file_...
352
  		SetPageSwapBacked(new_page);
31a563962   Daisuke Nishimura   mm: add_to_swap_c...
353
  		err = __add_to_swap_cache(new_page, entry);
529ae9aaa   Nick Piggin   mm: rename page t...
354
  		if (likely(!err)) {
31a563962   Daisuke Nishimura   mm: add_to_swap_c...
355
  			radix_tree_preload_end();
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
356
357
358
  			/*
  			 * Initiate read into locked page and return.
  			 */
c5fdae469   Rik van Riel   vmscan: add newly...
359
  			lru_cache_add_anon(new_page);
aca8bf323   Minchan Kim   mm: remove file a...
360
  			swap_readpage(new_page);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
361
362
  			return new_page;
  		}
31a563962   Daisuke Nishimura   mm: add_to_swap_c...
363
  		radix_tree_preload_end();
b2e185384   Rik van Riel   define page_file_...
364
  		ClearPageSwapBacked(new_page);
f45840b5c   Nick Piggin   mm: pagecache ins...
365
  		__clear_page_locked(new_page);
2ca4532a4   Daisuke Nishimura   mm: add_to_swap_c...
366
367
368
369
  		/*
  		 * add_to_swap_cache() doesn't return -EEXIST, so we can safely
  		 * clear SWAP_HAS_CACHE flag.
  		 */
0a31bc97c   Johannes Weiner   mm: memcontrol: r...
370
  		swapcache_free(entry);
f000944d0   Hugh Dickins   tmpfs: shuffle ad...
371
  	} while (err != -ENOMEM);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
372
373
374
375
376
  
  	if (new_page)
  		page_cache_release(new_page);
  	return found_page;
  }
46017e954   Hugh Dickins   swapin_readahead:...
377

579f82901   Shaohua Li   swap: add a simpl...
378
379
380
381
382
  static unsigned long swapin_nr_pages(unsigned long offset)
  {
  	static unsigned long prev_offset;
  	unsigned int pages, max_pages, last_ra;
  	static atomic_t last_readahead_pages;
4db0c3c29   Jason Low   mm: remove rest o...
383
  	max_pages = 1 << READ_ONCE(page_cluster);
579f82901   Shaohua Li   swap: add a simpl...
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
  	if (max_pages <= 1)
  		return 1;
  
  	/*
  	 * This heuristic has been found to work well on both sequential and
  	 * random loads, swapping to hard disk or to SSD: please don't ask
  	 * what the "+ 2" means, it just happens to work well, that's all.
  	 */
  	pages = atomic_xchg(&swapin_readahead_hits, 0) + 2;
  	if (pages == 2) {
  		/*
  		 * We can have no readahead hits to judge by: but must not get
  		 * stuck here forever, so check for an adjacent offset instead
  		 * (and don't even bother to check whether swap type is same).
  		 */
  		if (offset != prev_offset + 1 && offset != prev_offset - 1)
  			pages = 1;
  		prev_offset = offset;
  	} else {
  		unsigned int roundup = 4;
  		while (roundup < pages)
  			roundup <<= 1;
  		pages = roundup;
  	}
  
  	if (pages > max_pages)
  		pages = max_pages;
  
  	/* Don't shrink readahead too fast */
  	last_ra = atomic_read(&last_readahead_pages) / 2;
  	if (pages < last_ra)
  		pages = last_ra;
  	atomic_set(&last_readahead_pages, pages);
  
  	return pages;
  }
46017e954   Hugh Dickins   swapin_readahead:...
420
421
422
  /**
   * swapin_readahead - swap in pages in hope we need them soon
   * @entry: swap entry of this memory
7682486b3   Randy Dunlap   mm: fix various k...
423
   * @gfp_mask: memory allocation flags
46017e954   Hugh Dickins   swapin_readahead:...
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
   * @vma: user vma this address belongs to
   * @addr: target address for mempolicy
   *
   * Returns the struct page for entry and addr, after queueing swapin.
   *
   * Primitive swap readahead code. We simply read an aligned block of
   * (1 << page_cluster) entries in the swap area. This method is chosen
   * because it doesn't cost us any seek time.  We also make sure to queue
   * the 'original' request together with the readahead ones...
   *
   * This has been extended to use the NUMA policies from the mm triggering
   * the readahead.
   *
   * Caller must hold down_read on the vma->vm_mm if vma is not NULL.
   */
02098feaa   Hugh Dickins   swapin needs gfp_...
439
  struct page *swapin_readahead(swp_entry_t entry, gfp_t gfp_mask,
46017e954   Hugh Dickins   swapin_readahead:...
440
441
  			struct vm_area_struct *vma, unsigned long addr)
  {
46017e954   Hugh Dickins   swapin_readahead:...
442
  	struct page *page;
579f82901   Shaohua Li   swap: add a simpl...
443
444
  	unsigned long entry_offset = swp_offset(entry);
  	unsigned long offset = entry_offset;
67f96aa25   Rik van Riel   mm: make swapin r...
445
  	unsigned long start_offset, end_offset;
579f82901   Shaohua Li   swap: add a simpl...
446
  	unsigned long mask;
3fb5c298b   Christian Ehrhardt   swap: allow swap ...
447
  	struct blk_plug plug;
46017e954   Hugh Dickins   swapin_readahead:...
448

579f82901   Shaohua Li   swap: add a simpl...
449
450
451
  	mask = swapin_nr_pages(offset) - 1;
  	if (!mask)
  		goto skip;
67f96aa25   Rik van Riel   mm: make swapin r...
452
453
454
455
456
  	/* Read a page_cluster sized and aligned cluster around offset. */
  	start_offset = offset & ~mask;
  	end_offset = offset | mask;
  	if (!start_offset)	/* First page is swap header. */
  		start_offset++;
3fb5c298b   Christian Ehrhardt   swap: allow swap ...
457
  	blk_start_plug(&plug);
67f96aa25   Rik van Riel   mm: make swapin r...
458
  	for (offset = start_offset; offset <= end_offset ; offset++) {
46017e954   Hugh Dickins   swapin_readahead:...
459
460
  		/* Ok, do the async read-ahead now */
  		page = read_swap_cache_async(swp_entry(swp_type(entry), offset),
02098feaa   Hugh Dickins   swapin needs gfp_...
461
  						gfp_mask, vma, addr);
46017e954   Hugh Dickins   swapin_readahead:...
462
  		if (!page)
67f96aa25   Rik van Riel   mm: make swapin r...
463
  			continue;
579f82901   Shaohua Li   swap: add a simpl...
464
465
  		if (offset != entry_offset)
  			SetPageReadahead(page);
46017e954   Hugh Dickins   swapin_readahead:...
466
467
  		page_cache_release(page);
  	}
3fb5c298b   Christian Ehrhardt   swap: allow swap ...
468
  	blk_finish_plug(&plug);
46017e954   Hugh Dickins   swapin_readahead:...
469
  	lru_add_drain();	/* Push any new pages onto the LRU now */
579f82901   Shaohua Li   swap: add a simpl...
470
  skip:
02098feaa   Hugh Dickins   swapin needs gfp_...
471
  	return read_swap_cache_async(entry, gfp_mask, vma, addr);
46017e954   Hugh Dickins   swapin_readahead:...
472
  }