Blame view

mm/page_cgroup.c 11.8 KB
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
1
2
3
4
5
6
  #include <linux/mm.h>
  #include <linux/mmzone.h>
  #include <linux/bootmem.h>
  #include <linux/bit_spinlock.h>
  #include <linux/page_cgroup.h>
  #include <linux/hash.h>
94b6da5ab   KAMEZAWA Hiroyuki   memcg: fix page_c...
7
  #include <linux/slab.h>
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
8
  #include <linux/memory.h>
4c8210427   Paul Mundt   mm: page_cgroup n...
9
  #include <linux/vmalloc.h>
94b6da5ab   KAMEZAWA Hiroyuki   memcg: fix page_c...
10
  #include <linux/cgroup.h>
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
11
  #include <linux/swapops.h>
7952f9881   Catalin Marinas   kmemleak: Annotat...
12
  #include <linux/kmemleak.h>
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
13

52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
14
15
16
  static unsigned long total_usage;
  
  #if !defined(CONFIG_SPARSEMEM)
31168481c   Al Viro   meminit section w...
17
  void __meminit pgdat_page_cgroup_init(struct pglist_data *pgdat)
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
18
19
20
21
22
23
24
25
26
27
28
  {
  	pgdat->node_page_cgroup = NULL;
  }
  
  struct page_cgroup *lookup_page_cgroup(struct page *page)
  {
  	unsigned long pfn = page_to_pfn(page);
  	unsigned long offset;
  	struct page_cgroup *base;
  
  	base = NODE_DATA(page_to_nid(page))->node_page_cgroup;
00c54c0ba   Johannes Weiner   mm: page_cgroup: ...
29
30
31
32
33
34
35
  #ifdef CONFIG_DEBUG_VM
  	/*
  	 * The sanity checks the page allocator does upon freeing a
  	 * page can reach here before the page_cgroup arrays are
  	 * allocated when feeding a range of pages to the allocator
  	 * for the first time during bootup or memory hotplug.
  	 */
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
36
37
  	if (unlikely(!base))
  		return NULL;
00c54c0ba   Johannes Weiner   mm: page_cgroup: ...
38
  #endif
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
39
40
41
42
43
44
  	offset = pfn - NODE_DATA(page_to_nid(page))->node_start_pfn;
  	return base + offset;
  }
  
  static int __init alloc_node_page_cgroup(int nid)
  {
6b208e3f6   Johannes Weiner   mm: memcg: remove...
45
  	struct page_cgroup *base;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
46
  	unsigned long table_size;
6b208e3f6   Johannes Weiner   mm: memcg: remove...
47
  	unsigned long nr_pages;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
48

52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
49
  	nr_pages = NODE_DATA(nid)->node_spanned_pages;
653d22c0f   KAMEZAWA Hiroyuki   page_cgroup shoul...
50
51
  	if (!nr_pages)
  		return 0;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
52
  	table_size = sizeof(struct page_cgroup) * nr_pages;
ca371c0d7   KAMEZAWA Hiroyuki   memcg: fix page_c...
53
54
55
56
  
  	base = __alloc_bootmem_node_nopanic(NODE_DATA(nid),
  			table_size, PAGE_SIZE, __pa(MAX_DMA_ADDRESS));
  	if (!base)
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
57
  		return -ENOMEM;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
58
59
60
61
  	NODE_DATA(nid)->node_page_cgroup = base;
  	total_usage += table_size;
  	return 0;
  }
ca371c0d7   KAMEZAWA Hiroyuki   memcg: fix page_c...
62
  void __init page_cgroup_init_flatmem(void)
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
63
64
65
  {
  
  	int nid, fail;
f8d665422   Hirokazu Takahashi   memcg: add mem_cg...
66
  	if (mem_cgroup_disabled())
94b6da5ab   KAMEZAWA Hiroyuki   memcg: fix page_c...
67
  		return;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
68
69
70
71
72
73
74
  	for_each_online_node(nid)  {
  		fail = alloc_node_page_cgroup(nid);
  		if (fail)
  			goto fail;
  	}
  	printk(KERN_INFO "allocated %ld bytes of page_cgroup
  ", total_usage);
8ca739e36   Randy Dunlap   cgroups: make mes...
75
76
77
  	printk(KERN_INFO "please try 'cgroup_disable=memory' option if you"
  	" don't want memory cgroups
  ");
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
78
79
  	return;
  fail:
8ca739e36   Randy Dunlap   cgroups: make mes...
80
81
82
83
  	printk(KERN_CRIT "allocation of page_cgroup failed.
  ");
  	printk(KERN_CRIT "please try 'cgroup_disable=memory' boot option
  ");
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
84
85
86
87
88
89
90
91
92
  	panic("Out of memory");
  }
  
  #else /* CONFIG_FLAT_NODE_MEM_MAP */
  
  struct page_cgroup *lookup_page_cgroup(struct page *page)
  {
  	unsigned long pfn = page_to_pfn(page);
  	struct mem_section *section = __pfn_to_section(pfn);
00c54c0ba   Johannes Weiner   mm: page_cgroup: ...
93
94
95
96
97
98
99
  #ifdef CONFIG_DEBUG_VM
  	/*
  	 * The sanity checks the page allocator does upon freeing a
  	 * page can reach here before the page_cgroup arrays are
  	 * allocated when feeding a range of pages to the allocator
  	 * for the first time during bootup or memory hotplug.
  	 */
d69b042f3   Balbir Singh   memcg: add file-b...
100
101
  	if (!section->page_cgroup)
  		return NULL;
00c54c0ba   Johannes Weiner   mm: page_cgroup: ...
102
  #endif
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
103
104
  	return section->page_cgroup + pfn;
  }
268433b8e   Namhyung Kim   memcg: mark init_...
105
  static void *__meminit alloc_page_cgroup(size_t size, int nid)
dde79e005   Michal Hocko   page_cgroup: redu...
106
  {
6b208e3f6   Johannes Weiner   mm: memcg: remove...
107
  	gfp_t flags = GFP_KERNEL | __GFP_ZERO | __GFP_NOWARN;
dde79e005   Michal Hocko   page_cgroup: redu...
108
  	void *addr = NULL;
ff7ee93f4   Steven Rostedt   cgroup/kmemleak: ...
109
110
111
  	addr = alloc_pages_exact_nid(nid, size, flags);
  	if (addr) {
  		kmemleak_alloc(addr, size, 1, flags);
dde79e005   Michal Hocko   page_cgroup: redu...
112
  		return addr;
ff7ee93f4   Steven Rostedt   cgroup/kmemleak: ...
113
  	}
dde79e005   Michal Hocko   page_cgroup: redu...
114
115
  
  	if (node_state(nid, N_HIGH_MEMORY))
6b208e3f6   Johannes Weiner   mm: memcg: remove...
116
  		addr = vzalloc_node(size, nid);
dde79e005   Michal Hocko   page_cgroup: redu...
117
  	else
6b208e3f6   Johannes Weiner   mm: memcg: remove...
118
  		addr = vzalloc(size);
dde79e005   Michal Hocko   page_cgroup: redu...
119
120
121
  
  	return addr;
  }
37573e8c7   KAMEZAWA Hiroyuki   memcg: fix init_p...
122
  static int __meminit init_section_page_cgroup(unsigned long pfn, int nid)
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
123
  {
6b3ae58ef   Johannes Weiner   memcg: remove dir...
124
  	struct mem_section *section;
6b208e3f6   Johannes Weiner   mm: memcg: remove...
125
  	struct page_cgroup *base;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
126
  	unsigned long table_size;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
127

6b208e3f6   Johannes Weiner   mm: memcg: remove...
128
  	section = __pfn_to_section(pfn);
6b3ae58ef   Johannes Weiner   memcg: remove dir...
129
130
131
  
  	if (section->page_cgroup)
  		return 0;
6b3ae58ef   Johannes Weiner   memcg: remove dir...
132
  	table_size = sizeof(struct page_cgroup) * PAGES_PER_SECTION;
dde79e005   Michal Hocko   page_cgroup: redu...
133
  	base = alloc_page_cgroup(table_size, nid);
6b3ae58ef   Johannes Weiner   memcg: remove dir...
134
135
136
137
138
139
  	/*
  	 * The value stored in section->page_cgroup is (base - pfn)
  	 * and it does not point to the memory block allocated above,
  	 * causing kmemleak false positives.
  	 */
  	kmemleak_not_leak(base);
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
140
141
142
143
144
145
  
  	if (!base) {
  		printk(KERN_ERR "page cgroup allocation failure
  ");
  		return -ENOMEM;
  	}
37573e8c7   KAMEZAWA Hiroyuki   memcg: fix init_p...
146
147
148
149
150
  	/*
  	 * The passed "pfn" may not be aligned to SECTION.  For the calculation
  	 * we need to apply a mask.
  	 */
  	pfn &= PAGE_SECTION_MASK;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
151
152
153
154
155
  	section->page_cgroup = base - pfn;
  	total_usage += table_size;
  	return 0;
  }
  #ifdef CONFIG_MEMORY_HOTPLUG
0efc8eb9c   Bob Liu   page_cgroup: drop...
156
157
158
159
160
161
162
163
164
165
166
167
168
  static void free_page_cgroup(void *addr)
  {
  	if (is_vmalloc_addr(addr)) {
  		vfree(addr);
  	} else {
  		struct page *page = virt_to_page(addr);
  		size_t table_size =
  			sizeof(struct page_cgroup) * PAGES_PER_SECTION;
  
  		BUG_ON(PageReserved(page));
  		free_pages_exact(addr, table_size);
  	}
  }
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
169
170
171
172
173
174
175
176
177
  void __free_page_cgroup(unsigned long pfn)
  {
  	struct mem_section *ms;
  	struct page_cgroup *base;
  
  	ms = __pfn_to_section(pfn);
  	if (!ms || !ms->page_cgroup)
  		return;
  	base = ms->page_cgroup + pfn;
dde79e005   Michal Hocko   page_cgroup: redu...
178
179
  	free_page_cgroup(base);
  	ms->page_cgroup = NULL;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
180
  }
31168481c   Al Viro   meminit section w...
181
  int __meminit online_page_cgroup(unsigned long start_pfn,
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
182
183
184
185
186
  			unsigned long nr_pages,
  			int nid)
  {
  	unsigned long start, end, pfn;
  	int fail = 0;
1bb36fbd4   Daniel Kiper   mm/page_cgroup.c:...
187
188
  	start = SECTION_ALIGN_DOWN(start_pfn);
  	end = SECTION_ALIGN_UP(start_pfn + nr_pages);
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
189

37573e8c7   KAMEZAWA Hiroyuki   memcg: fix init_p...
190
191
192
193
194
195
196
197
198
  	if (nid == -1) {
  		/*
  		 * In this case, "nid" already exists and contains valid memory.
  		 * "start_pfn" passed to us is a pfn which is an arg for
  		 * online__pages(), and start_pfn should exist.
  		 */
  		nid = pfn_to_nid(start_pfn);
  		VM_BUG_ON(!node_state(nid, N_ONLINE));
  	}
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
199
200
201
  	for (pfn = start; !fail && pfn < end; pfn += PAGES_PER_SECTION) {
  		if (!pfn_present(pfn))
  			continue;
37573e8c7   KAMEZAWA Hiroyuki   memcg: fix init_p...
202
  		fail = init_section_page_cgroup(pfn, nid);
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
203
204
205
206
207
208
209
210
211
212
  	}
  	if (!fail)
  		return 0;
  
  	/* rollback */
  	for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION)
  		__free_page_cgroup(pfn);
  
  	return -ENOMEM;
  }
31168481c   Al Viro   meminit section w...
213
  int __meminit offline_page_cgroup(unsigned long start_pfn,
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
214
215
216
  		unsigned long nr_pages, int nid)
  {
  	unsigned long start, end, pfn;
1bb36fbd4   Daniel Kiper   mm/page_cgroup.c:...
217
218
  	start = SECTION_ALIGN_DOWN(start_pfn);
  	end = SECTION_ALIGN_UP(start_pfn + nr_pages);
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
219
220
221
222
223
224
  
  	for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION)
  		__free_page_cgroup(pfn);
  	return 0;
  
  }
31168481c   Al Viro   meminit section w...
225
  static int __meminit page_cgroup_callback(struct notifier_block *self,
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
226
227
228
229
230
231
232
233
234
  			       unsigned long action, void *arg)
  {
  	struct memory_notify *mn = arg;
  	int ret = 0;
  	switch (action) {
  	case MEM_GOING_ONLINE:
  		ret = online_page_cgroup(mn->start_pfn,
  				   mn->nr_pages, mn->status_change_nid);
  		break;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
235
236
237
238
  	case MEM_OFFLINE:
  		offline_page_cgroup(mn->start_pfn,
  				mn->nr_pages, mn->status_change_nid);
  		break;
dc19f9db3   KAMEZAWA Hiroyuki   memcg: memory hot...
239
  	case MEM_CANCEL_ONLINE:
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
240
241
242
243
244
245
  	case MEM_GOING_OFFLINE:
  		break;
  	case MEM_ONLINE:
  	case MEM_CANCEL_OFFLINE:
  		break;
  	}
dc19f9db3   KAMEZAWA Hiroyuki   memcg: memory hot...
246

5fda1bd5b   Prarit Bhargava   mm: notifier_from...
247
  	return notifier_from_errno(ret);
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
248
249
250
251
252
253
254
  }
  
  #endif
  
  void __init page_cgroup_init(void)
  {
  	unsigned long pfn;
37573e8c7   KAMEZAWA Hiroyuki   memcg: fix init_p...
255
  	int nid;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
256

f8d665422   Hirokazu Takahashi   memcg: add mem_cg...
257
  	if (mem_cgroup_disabled())
94b6da5ab   KAMEZAWA Hiroyuki   memcg: fix page_c...
258
  		return;
37573e8c7   KAMEZAWA Hiroyuki   memcg: fix init_p...
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
  	for_each_node_state(nid, N_HIGH_MEMORY) {
  		unsigned long start_pfn, end_pfn;
  
  		start_pfn = node_start_pfn(nid);
  		end_pfn = node_end_pfn(nid);
  		/*
  		 * start_pfn and end_pfn may not be aligned to SECTION and the
  		 * page->flags of out of node pages are not initialized.  So we
  		 * scan [start_pfn, the biggest section's pfn < end_pfn) here.
  		 */
  		for (pfn = start_pfn;
  		     pfn < end_pfn;
                       pfn = ALIGN(pfn + 1, PAGES_PER_SECTION)) {
  
  			if (!pfn_valid(pfn))
  				continue;
  			/*
  			 * Nodes's pfns can be overlapping.
  			 * We know some arch can have a nodes layout such as
  			 * -------------pfn-------------->
  			 * N0 | N1 | N2 | N0 | N1 | N2|....
  			 */
  			if (pfn_to_nid(pfn) != nid)
  				continue;
  			if (init_section_page_cgroup(pfn, nid))
  				goto oom;
  		}
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
286
  	}
37573e8c7   KAMEZAWA Hiroyuki   memcg: fix init_p...
287
  	hotplug_memory_notifier(page_cgroup_callback, 0);
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
288
289
  	printk(KERN_INFO "allocated %ld bytes of page_cgroup
  ", total_usage);
37573e8c7   KAMEZAWA Hiroyuki   memcg: fix init_p...
290
291
292
293
294
295
296
297
  	printk(KERN_INFO "please try 'cgroup_disable=memory' option if you "
  			 "don't want memory cgroups
  ");
  	return;
  oom:
  	printk(KERN_CRIT "try 'cgroup_disable=memory' boot option
  ");
  	panic("Out of memory");
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
298
  }
31168481c   Al Viro   meminit section w...
299
  void __meminit pgdat_page_cgroup_init(struct pglist_data *pgdat)
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
300
301
302
303
304
  {
  	return;
  }
  
  #endif
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
305
306
307
308
309
310
311
312
  
  
  #ifdef CONFIG_CGROUP_MEM_RES_CTLR_SWAP
  
  static DEFINE_MUTEX(swap_cgroup_mutex);
  struct swap_cgroup_ctrl {
  	struct page **map;
  	unsigned long length;
e9e58a4ec   KAMEZAWA Hiroyuki   memcg: avoid use ...
313
  	spinlock_t	lock;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
314
  };
61600f578   H Hartley Sweeten   mm/page_cgroup.c:...
315
  static struct swap_cgroup_ctrl swap_cgroup_ctrl[MAX_SWAPFILES];
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
316

27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
317
  struct swap_cgroup {
a3b2d6926   KAMEZAWA Hiroyuki   cgroups: use css ...
318
  	unsigned short		id;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
319
320
  };
  #define SC_PER_PAGE	(PAGE_SIZE/sizeof(struct swap_cgroup))
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
  
  /*
   * SwapCgroup implements "lookup" and "exchange" operations.
   * In typical usage, this swap_cgroup is accessed via memcg's charge/uncharge
   * against SwapCache. At swap_free(), this is accessed directly from swap.
   *
   * This means,
   *  - we have no race in "exchange" when we're accessed via SwapCache because
   *    SwapCache(and its swp_entry) is under lock.
   *  - When called via swap_free(), there is no user of this entry and no race.
   * Then, we don't need lock around "exchange".
   *
   * TODO: we can push these buffers out to HIGHMEM.
   */
  
  /*
   * allocate buffer for swap_cgroup.
   */
  static int swap_cgroup_prepare(int type)
  {
  	struct page *page;
  	struct swap_cgroup_ctrl *ctrl;
  	unsigned long idx, max;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
  	ctrl = &swap_cgroup_ctrl[type];
  
  	for (idx = 0; idx < ctrl->length; idx++) {
  		page = alloc_page(GFP_KERNEL | __GFP_ZERO);
  		if (!page)
  			goto not_enough_page;
  		ctrl->map[idx] = page;
  	}
  	return 0;
  not_enough_page:
  	max = idx;
  	for (idx = 0; idx < max; idx++)
  		__free_page(ctrl->map[idx]);
  
  	return -ENOMEM;
  }
9fb4b7cc0   Bob Liu   page_cgroup: add ...
360
361
362
363
364
365
366
367
368
369
370
371
372
373
  static struct swap_cgroup *lookup_swap_cgroup(swp_entry_t ent,
  					struct swap_cgroup_ctrl **ctrlp)
  {
  	pgoff_t offset = swp_offset(ent);
  	struct swap_cgroup_ctrl *ctrl;
  	struct page *mappage;
  
  	ctrl = &swap_cgroup_ctrl[swp_type(ent)];
  	if (ctrlp)
  		*ctrlp = ctrl;
  
  	mappage = ctrl->map[offset / SC_PER_PAGE];
  	return page_address(mappage) + offset % SC_PER_PAGE;
  }
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
374
  /**
024914477   Daisuke Nishimura   memcg: move charg...
375
376
377
378
379
380
   * swap_cgroup_cmpxchg - cmpxchg mem_cgroup's id for this swp_entry.
   * @end: swap entry to be cmpxchged
   * @old: old id
   * @new: new id
   *
   * Returns old id at success, 0 at failure.
25985edce   Lucas De Marchi   Fix common misspe...
381
   * (There is no mem_cgroup using 0 as its id)
024914477   Daisuke Nishimura   memcg: move charg...
382
383
384
385
   */
  unsigned short swap_cgroup_cmpxchg(swp_entry_t ent,
  					unsigned short old, unsigned short new)
  {
024914477   Daisuke Nishimura   memcg: move charg...
386
  	struct swap_cgroup_ctrl *ctrl;
024914477   Daisuke Nishimura   memcg: move charg...
387
  	struct swap_cgroup *sc;
e9e58a4ec   KAMEZAWA Hiroyuki   memcg: avoid use ...
388
389
  	unsigned long flags;
  	unsigned short retval;
024914477   Daisuke Nishimura   memcg: move charg...
390

9fb4b7cc0   Bob Liu   page_cgroup: add ...
391
  	sc = lookup_swap_cgroup(ent, &ctrl);
024914477   Daisuke Nishimura   memcg: move charg...
392

e9e58a4ec   KAMEZAWA Hiroyuki   memcg: avoid use ...
393
394
395
396
  	spin_lock_irqsave(&ctrl->lock, flags);
  	retval = sc->id;
  	if (retval == old)
  		sc->id = new;
024914477   Daisuke Nishimura   memcg: move charg...
397
  	else
e9e58a4ec   KAMEZAWA Hiroyuki   memcg: avoid use ...
398
399
400
  		retval = 0;
  	spin_unlock_irqrestore(&ctrl->lock, flags);
  	return retval;
024914477   Daisuke Nishimura   memcg: move charg...
401
402
403
  }
  
  /**
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
404
405
406
407
   * swap_cgroup_record - record mem_cgroup for this swp_entry.
   * @ent: swap entry to be recorded into
   * @mem: mem_cgroup to be recorded
   *
a3b2d6926   KAMEZAWA Hiroyuki   cgroups: use css ...
408
409
   * Returns old value at success, 0 at failure.
   * (Of course, old value can be 0.)
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
410
   */
a3b2d6926   KAMEZAWA Hiroyuki   cgroups: use css ...
411
  unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id)
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
412
  {
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
413
  	struct swap_cgroup_ctrl *ctrl;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
414
  	struct swap_cgroup *sc;
a3b2d6926   KAMEZAWA Hiroyuki   cgroups: use css ...
415
  	unsigned short old;
e9e58a4ec   KAMEZAWA Hiroyuki   memcg: avoid use ...
416
  	unsigned long flags;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
417

9fb4b7cc0   Bob Liu   page_cgroup: add ...
418
  	sc = lookup_swap_cgroup(ent, &ctrl);
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
419

e9e58a4ec   KAMEZAWA Hiroyuki   memcg: avoid use ...
420
421
422
423
  	spin_lock_irqsave(&ctrl->lock, flags);
  	old = sc->id;
  	sc->id = id;
  	spin_unlock_irqrestore(&ctrl->lock, flags);
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
424
425
426
427
428
  
  	return old;
  }
  
  /**
9fb4b7cc0   Bob Liu   page_cgroup: add ...
429
   * lookup_swap_cgroup_id - lookup mem_cgroup id tied to swap entry
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
430
431
   * @ent: swap entry to be looked up.
   *
a3b2d6926   KAMEZAWA Hiroyuki   cgroups: use css ...
432
   * Returns CSS ID of mem_cgroup at success. 0 at failure. (0 is invalid ID)
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
433
   */
9fb4b7cc0   Bob Liu   page_cgroup: add ...
434
  unsigned short lookup_swap_cgroup_id(swp_entry_t ent)
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
435
  {
9fb4b7cc0   Bob Liu   page_cgroup: add ...
436
  	return lookup_swap_cgroup(ent, NULL)->id;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
437
438
439
440
441
442
443
444
445
446
447
  }
  
  int swap_cgroup_swapon(int type, unsigned long max_pages)
  {
  	void *array;
  	unsigned long array_size;
  	unsigned long length;
  	struct swap_cgroup_ctrl *ctrl;
  
  	if (!do_swap_account)
  		return 0;
33278f7f0   Namhyung Kim   memcg: fix off-by...
448
  	length = DIV_ROUND_UP(max_pages, SC_PER_PAGE);
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
449
  	array_size = length * sizeof(void *);
8c1fec1ba   Joe Perches   mm: Convert vmall...
450
  	array = vzalloc(array_size);
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
451
452
  	if (!array)
  		goto nomem;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
453
454
455
456
  	ctrl = &swap_cgroup_ctrl[type];
  	mutex_lock(&swap_cgroup_mutex);
  	ctrl->length = length;
  	ctrl->map = array;
e9e58a4ec   KAMEZAWA Hiroyuki   memcg: avoid use ...
457
  	spin_lock_init(&ctrl->lock);
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
458
459
460
461
  	if (swap_cgroup_prepare(type)) {
  		/* memory shortage */
  		ctrl->map = NULL;
  		ctrl->length = 0;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
462
  		mutex_unlock(&swap_cgroup_mutex);
6a5b18d2b   Namhyung Kim   memcg: move page-...
463
  		vfree(array);
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
464
465
466
  		goto nomem;
  	}
  	mutex_unlock(&swap_cgroup_mutex);
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
467
468
469
470
471
  	return 0;
  nomem:
  	printk(KERN_INFO "couldn't allocate enough memory for swap_cgroup.
  ");
  	printk(KERN_INFO
00a66d297   WANG Cong   mm: remove the le...
472
473
  		"swap_cgroup can be disabled by swapaccount=0 boot option
  ");
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
474
475
476
477
478
  	return -ENOMEM;
  }
  
  void swap_cgroup_swapoff(int type)
  {
6a5b18d2b   Namhyung Kim   memcg: move page-...
479
480
  	struct page **map;
  	unsigned long i, length;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
481
482
483
484
485
486
487
  	struct swap_cgroup_ctrl *ctrl;
  
  	if (!do_swap_account)
  		return;
  
  	mutex_lock(&swap_cgroup_mutex);
  	ctrl = &swap_cgroup_ctrl[type];
6a5b18d2b   Namhyung Kim   memcg: move page-...
488
489
490
491
492
493
494
495
496
  	map = ctrl->map;
  	length = ctrl->length;
  	ctrl->map = NULL;
  	ctrl->length = 0;
  	mutex_unlock(&swap_cgroup_mutex);
  
  	if (map) {
  		for (i = 0; i < length; i++) {
  			struct page *page = map[i];
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
497
498
499
  			if (page)
  				__free_page(page);
  		}
6a5b18d2b   Namhyung Kim   memcg: move page-...
500
  		vfree(map);
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
501
  	}
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
502
503
504
  }
  
  #endif