Blame view

mm/page_cgroup.c 12 KB
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
1
2
3
4
5
6
  #include <linux/mm.h>
  #include <linux/mmzone.h>
  #include <linux/bootmem.h>
  #include <linux/bit_spinlock.h>
  #include <linux/page_cgroup.h>
  #include <linux/hash.h>
94b6da5ab   KAMEZAWA Hiroyuki   memcg: fix page_c...
7
  #include <linux/slab.h>
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
8
  #include <linux/memory.h>
4c8210427   Paul Mundt   mm: page_cgroup n...
9
  #include <linux/vmalloc.h>
94b6da5ab   KAMEZAWA Hiroyuki   memcg: fix page_c...
10
  #include <linux/cgroup.h>
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
11
  #include <linux/swapops.h>
7952f9881   Catalin Marinas   kmemleak: Annotat...
12
  #include <linux/kmemleak.h>
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
13

52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
14
15
16
  static unsigned long total_usage;
  
  #if !defined(CONFIG_SPARSEMEM)
31168481c   Al Viro   meminit section w...
17
  void __meminit pgdat_page_cgroup_init(struct pglist_data *pgdat)
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
18
19
20
21
22
23
24
25
26
27
28
  {
  	pgdat->node_page_cgroup = NULL;
  }
  
  struct page_cgroup *lookup_page_cgroup(struct page *page)
  {
  	unsigned long pfn = page_to_pfn(page);
  	unsigned long offset;
  	struct page_cgroup *base;
  
  	base = NODE_DATA(page_to_nid(page))->node_page_cgroup;
00c54c0ba   Johannes Weiner   mm: page_cgroup: ...
29
30
31
32
33
34
35
  #ifdef CONFIG_DEBUG_VM
  	/*
  	 * The sanity checks the page allocator does upon freeing a
  	 * page can reach here before the page_cgroup arrays are
  	 * allocated when feeding a range of pages to the allocator
  	 * for the first time during bootup or memory hotplug.
  	 */
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
36
37
  	if (unlikely(!base))
  		return NULL;
00c54c0ba   Johannes Weiner   mm: page_cgroup: ...
38
  #endif
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
39
40
41
42
43
44
  	offset = pfn - NODE_DATA(page_to_nid(page))->node_start_pfn;
  	return base + offset;
  }
  
  static int __init alloc_node_page_cgroup(int nid)
  {
6b208e3f6   Johannes Weiner   mm: memcg: remove...
45
  	struct page_cgroup *base;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
46
  	unsigned long table_size;
6b208e3f6   Johannes Weiner   mm: memcg: remove...
47
  	unsigned long nr_pages;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
48

52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
49
  	nr_pages = NODE_DATA(nid)->node_spanned_pages;
653d22c0f   KAMEZAWA Hiroyuki   page_cgroup shoul...
50
51
  	if (!nr_pages)
  		return 0;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
52
  	table_size = sizeof(struct page_cgroup) * nr_pages;
ca371c0d7   KAMEZAWA Hiroyuki   memcg: fix page_c...
53

0d036e9e3   Grygorii Strashko   mm/page_cgroup.c:...
54
55
56
  	base = memblock_virt_alloc_try_nid_nopanic(
  			table_size, PAGE_SIZE, __pa(MAX_DMA_ADDRESS),
  			BOOTMEM_ALLOC_ACCESSIBLE, nid);
ca371c0d7   KAMEZAWA Hiroyuki   memcg: fix page_c...
57
  	if (!base)
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
58
  		return -ENOMEM;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
59
60
61
62
  	NODE_DATA(nid)->node_page_cgroup = base;
  	total_usage += table_size;
  	return 0;
  }
ca371c0d7   KAMEZAWA Hiroyuki   memcg: fix page_c...
63
  void __init page_cgroup_init_flatmem(void)
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
64
65
66
  {
  
  	int nid, fail;
f8d665422   Hirokazu Takahashi   memcg: add mem_cg...
67
  	if (mem_cgroup_disabled())
94b6da5ab   KAMEZAWA Hiroyuki   memcg: fix page_c...
68
  		return;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
69
70
71
72
73
74
75
  	for_each_online_node(nid)  {
  		fail = alloc_node_page_cgroup(nid);
  		if (fail)
  			goto fail;
  	}
  	printk(KERN_INFO "allocated %ld bytes of page_cgroup
  ", total_usage);
8ca739e36   Randy Dunlap   cgroups: make mes...
76
77
78
  	printk(KERN_INFO "please try 'cgroup_disable=memory' option if you"
  	" don't want memory cgroups
  ");
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
79
80
  	return;
  fail:
8ca739e36   Randy Dunlap   cgroups: make mes...
81
82
83
84
  	printk(KERN_CRIT "allocation of page_cgroup failed.
  ");
  	printk(KERN_CRIT "please try 'cgroup_disable=memory' boot option
  ");
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
85
86
87
88
89
90
91
92
93
  	panic("Out of memory");
  }
  
  #else /* CONFIG_FLAT_NODE_MEM_MAP */
  
  struct page_cgroup *lookup_page_cgroup(struct page *page)
  {
  	unsigned long pfn = page_to_pfn(page);
  	struct mem_section *section = __pfn_to_section(pfn);
00c54c0ba   Johannes Weiner   mm: page_cgroup: ...
94
95
96
97
98
99
100
  #ifdef CONFIG_DEBUG_VM
  	/*
  	 * The sanity checks the page allocator does upon freeing a
  	 * page can reach here before the page_cgroup arrays are
  	 * allocated when feeding a range of pages to the allocator
  	 * for the first time during bootup or memory hotplug.
  	 */
d69b042f3   Balbir Singh   memcg: add file-b...
101
102
  	if (!section->page_cgroup)
  		return NULL;
00c54c0ba   Johannes Weiner   mm: page_cgroup: ...
103
  #endif
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
104
105
  	return section->page_cgroup + pfn;
  }
268433b8e   Namhyung Kim   memcg: mark init_...
106
  static void *__meminit alloc_page_cgroup(size_t size, int nid)
dde79e005   Michal Hocko   page_cgroup: redu...
107
  {
6b208e3f6   Johannes Weiner   mm: memcg: remove...
108
  	gfp_t flags = GFP_KERNEL | __GFP_ZERO | __GFP_NOWARN;
dde79e005   Michal Hocko   page_cgroup: redu...
109
  	void *addr = NULL;
ff7ee93f4   Steven Rostedt   cgroup/kmemleak: ...
110
111
112
  	addr = alloc_pages_exact_nid(nid, size, flags);
  	if (addr) {
  		kmemleak_alloc(addr, size, 1, flags);
dde79e005   Michal Hocko   page_cgroup: redu...
113
  		return addr;
ff7ee93f4   Steven Rostedt   cgroup/kmemleak: ...
114
  	}
dde79e005   Michal Hocko   page_cgroup: redu...
115
116
  
  	if (node_state(nid, N_HIGH_MEMORY))
6b208e3f6   Johannes Weiner   mm: memcg: remove...
117
  		addr = vzalloc_node(size, nid);
dde79e005   Michal Hocko   page_cgroup: redu...
118
  	else
6b208e3f6   Johannes Weiner   mm: memcg: remove...
119
  		addr = vzalloc(size);
dde79e005   Michal Hocko   page_cgroup: redu...
120
121
122
  
  	return addr;
  }
37573e8c7   KAMEZAWA Hiroyuki   memcg: fix init_p...
123
  static int __meminit init_section_page_cgroup(unsigned long pfn, int nid)
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
124
  {
6b3ae58ef   Johannes Weiner   memcg: remove dir...
125
  	struct mem_section *section;
6b208e3f6   Johannes Weiner   mm: memcg: remove...
126
  	struct page_cgroup *base;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
127
  	unsigned long table_size;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
128

6b208e3f6   Johannes Weiner   mm: memcg: remove...
129
  	section = __pfn_to_section(pfn);
6b3ae58ef   Johannes Weiner   memcg: remove dir...
130
131
132
  
  	if (section->page_cgroup)
  		return 0;
6b3ae58ef   Johannes Weiner   memcg: remove dir...
133
  	table_size = sizeof(struct page_cgroup) * PAGES_PER_SECTION;
dde79e005   Michal Hocko   page_cgroup: redu...
134
  	base = alloc_page_cgroup(table_size, nid);
6b3ae58ef   Johannes Weiner   memcg: remove dir...
135
136
137
138
139
140
  	/*
  	 * The value stored in section->page_cgroup is (base - pfn)
  	 * and it does not point to the memory block allocated above,
  	 * causing kmemleak false positives.
  	 */
  	kmemleak_not_leak(base);
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
141
142
143
144
145
146
  
  	if (!base) {
  		printk(KERN_ERR "page cgroup allocation failure
  ");
  		return -ENOMEM;
  	}
37573e8c7   KAMEZAWA Hiroyuki   memcg: fix init_p...
147
148
149
150
151
  	/*
  	 * The passed "pfn" may not be aligned to SECTION.  For the calculation
  	 * we need to apply a mask.
  	 */
  	pfn &= PAGE_SECTION_MASK;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
152
153
154
155
156
  	section->page_cgroup = base - pfn;
  	total_usage += table_size;
  	return 0;
  }
  #ifdef CONFIG_MEMORY_HOTPLUG
0efc8eb9c   Bob Liu   page_cgroup: drop...
157
158
159
160
161
162
163
164
165
166
167
168
169
  static void free_page_cgroup(void *addr)
  {
  	if (is_vmalloc_addr(addr)) {
  		vfree(addr);
  	} else {
  		struct page *page = virt_to_page(addr);
  		size_t table_size =
  			sizeof(struct page_cgroup) * PAGES_PER_SECTION;
  
  		BUG_ON(PageReserved(page));
  		free_pages_exact(addr, table_size);
  	}
  }
d20199e1f   Rashika Kheria   mm/page_cgroup.c:...
170
  static void __free_page_cgroup(unsigned long pfn)
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
171
172
173
174
175
176
177
178
  {
  	struct mem_section *ms;
  	struct page_cgroup *base;
  
  	ms = __pfn_to_section(pfn);
  	if (!ms || !ms->page_cgroup)
  		return;
  	base = ms->page_cgroup + pfn;
dde79e005   Michal Hocko   page_cgroup: redu...
179
180
  	free_page_cgroup(base);
  	ms->page_cgroup = NULL;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
181
  }
d20199e1f   Rashika Kheria   mm/page_cgroup.c:...
182
183
184
  static int __meminit online_page_cgroup(unsigned long start_pfn,
  				unsigned long nr_pages,
  				int nid)
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
185
186
187
  {
  	unsigned long start, end, pfn;
  	int fail = 0;
1bb36fbd4   Daniel Kiper   mm/page_cgroup.c:...
188
189
  	start = SECTION_ALIGN_DOWN(start_pfn);
  	end = SECTION_ALIGN_UP(start_pfn + nr_pages);
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
190

37573e8c7   KAMEZAWA Hiroyuki   memcg: fix init_p...
191
192
193
194
195
196
197
198
199
  	if (nid == -1) {
  		/*
  		 * In this case, "nid" already exists and contains valid memory.
  		 * "start_pfn" passed to us is a pfn which is an arg for
  		 * online__pages(), and start_pfn should exist.
  		 */
  		nid = pfn_to_nid(start_pfn);
  		VM_BUG_ON(!node_state(nid, N_ONLINE));
  	}
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
200
201
202
  	for (pfn = start; !fail && pfn < end; pfn += PAGES_PER_SECTION) {
  		if (!pfn_present(pfn))
  			continue;
37573e8c7   KAMEZAWA Hiroyuki   memcg: fix init_p...
203
  		fail = init_section_page_cgroup(pfn, nid);
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
204
205
206
207
208
209
210
211
212
213
  	}
  	if (!fail)
  		return 0;
  
  	/* rollback */
  	for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION)
  		__free_page_cgroup(pfn);
  
  	return -ENOMEM;
  }
d20199e1f   Rashika Kheria   mm/page_cgroup.c:...
214
215
  static int __meminit offline_page_cgroup(unsigned long start_pfn,
  				unsigned long nr_pages, int nid)
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
216
217
  {
  	unsigned long start, end, pfn;
1bb36fbd4   Daniel Kiper   mm/page_cgroup.c:...
218
219
  	start = SECTION_ALIGN_DOWN(start_pfn);
  	end = SECTION_ALIGN_UP(start_pfn + nr_pages);
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
220
221
222
223
224
225
  
  	for (pfn = start; pfn < end; pfn += PAGES_PER_SECTION)
  		__free_page_cgroup(pfn);
  	return 0;
  
  }
31168481c   Al Viro   meminit section w...
226
  static int __meminit page_cgroup_callback(struct notifier_block *self,
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
227
228
229
230
231
232
233
234
235
  			       unsigned long action, void *arg)
  {
  	struct memory_notify *mn = arg;
  	int ret = 0;
  	switch (action) {
  	case MEM_GOING_ONLINE:
  		ret = online_page_cgroup(mn->start_pfn,
  				   mn->nr_pages, mn->status_change_nid);
  		break;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
236
237
238
239
  	case MEM_OFFLINE:
  		offline_page_cgroup(mn->start_pfn,
  				mn->nr_pages, mn->status_change_nid);
  		break;
dc19f9db3   KAMEZAWA Hiroyuki   memcg: memory hot...
240
  	case MEM_CANCEL_ONLINE:
7c72eb327   Wen Congyang   memory-hotplug: a...
241
242
243
  		offline_page_cgroup(mn->start_pfn,
  				mn->nr_pages, mn->status_change_nid);
  		break;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
244
245
246
247
248
249
  	case MEM_GOING_OFFLINE:
  		break;
  	case MEM_ONLINE:
  	case MEM_CANCEL_OFFLINE:
  		break;
  	}
dc19f9db3   KAMEZAWA Hiroyuki   memcg: memory hot...
250

5fda1bd5b   Prarit Bhargava   mm: notifier_from...
251
  	return notifier_from_errno(ret);
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
252
253
254
255
256
257
258
  }
  
  #endif
  
  void __init page_cgroup_init(void)
  {
  	unsigned long pfn;
37573e8c7   KAMEZAWA Hiroyuki   memcg: fix init_p...
259
  	int nid;
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
260

f8d665422   Hirokazu Takahashi   memcg: add mem_cg...
261
  	if (mem_cgroup_disabled())
94b6da5ab   KAMEZAWA Hiroyuki   memcg: fix page_c...
262
  		return;
31aaea4aa   Lai Jiangshan   memcontrol: use N...
263
  	for_each_node_state(nid, N_MEMORY) {
37573e8c7   KAMEZAWA Hiroyuki   memcg: fix init_p...
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
  		unsigned long start_pfn, end_pfn;
  
  		start_pfn = node_start_pfn(nid);
  		end_pfn = node_end_pfn(nid);
  		/*
  		 * start_pfn and end_pfn may not be aligned to SECTION and the
  		 * page->flags of out of node pages are not initialized.  So we
  		 * scan [start_pfn, the biggest section's pfn < end_pfn) here.
  		 */
  		for (pfn = start_pfn;
  		     pfn < end_pfn;
                       pfn = ALIGN(pfn + 1, PAGES_PER_SECTION)) {
  
  			if (!pfn_valid(pfn))
  				continue;
  			/*
  			 * Nodes's pfns can be overlapping.
  			 * We know some arch can have a nodes layout such as
  			 * -------------pfn-------------->
  			 * N0 | N1 | N2 | N0 | N1 | N2|....
  			 */
  			if (pfn_to_nid(pfn) != nid)
  				continue;
  			if (init_section_page_cgroup(pfn, nid))
  				goto oom;
  		}
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
290
  	}
37573e8c7   KAMEZAWA Hiroyuki   memcg: fix init_p...
291
  	hotplug_memory_notifier(page_cgroup_callback, 0);
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
292
293
  	printk(KERN_INFO "allocated %ld bytes of page_cgroup
  ", total_usage);
37573e8c7   KAMEZAWA Hiroyuki   memcg: fix init_p...
294
295
296
297
298
299
300
301
  	printk(KERN_INFO "please try 'cgroup_disable=memory' option if you "
  			 "don't want memory cgroups
  ");
  	return;
  oom:
  	printk(KERN_CRIT "try 'cgroup_disable=memory' boot option
  ");
  	panic("Out of memory");
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
302
  }
31168481c   Al Viro   meminit section w...
303
  void __meminit pgdat_page_cgroup_init(struct pglist_data *pgdat)
52d4b9ac0   KAMEZAWA Hiroyuki   memcg: allocate a...
304
305
306
307
308
  {
  	return;
  }
  
  #endif
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
309

c255a4580   Andrew Morton   memcg: rename con...
310
  #ifdef CONFIG_MEMCG_SWAP
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
311
312
313
314
315
  
  static DEFINE_MUTEX(swap_cgroup_mutex);
  struct swap_cgroup_ctrl {
  	struct page **map;
  	unsigned long length;
e9e58a4ec   KAMEZAWA Hiroyuki   memcg: avoid use ...
316
  	spinlock_t	lock;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
317
  };
61600f578   H Hartley Sweeten   mm/page_cgroup.c:...
318
  static struct swap_cgroup_ctrl swap_cgroup_ctrl[MAX_SWAPFILES];
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
319

27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
320
  struct swap_cgroup {
a3b2d6926   KAMEZAWA Hiroyuki   cgroups: use css ...
321
  	unsigned short		id;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
322
323
  };
  #define SC_PER_PAGE	(PAGE_SIZE/sizeof(struct swap_cgroup))
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
  
  /*
   * SwapCgroup implements "lookup" and "exchange" operations.
   * In typical usage, this swap_cgroup is accessed via memcg's charge/uncharge
   * against SwapCache. At swap_free(), this is accessed directly from swap.
   *
   * This means,
   *  - we have no race in "exchange" when we're accessed via SwapCache because
   *    SwapCache(and its swp_entry) is under lock.
   *  - When called via swap_free(), there is no user of this entry and no race.
   * Then, we don't need lock around "exchange".
   *
   * TODO: we can push these buffers out to HIGHMEM.
   */
  
  /*
   * allocate buffer for swap_cgroup.
   */
  static int swap_cgroup_prepare(int type)
  {
  	struct page *page;
  	struct swap_cgroup_ctrl *ctrl;
  	unsigned long idx, max;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
  	ctrl = &swap_cgroup_ctrl[type];
  
  	for (idx = 0; idx < ctrl->length; idx++) {
  		page = alloc_page(GFP_KERNEL | __GFP_ZERO);
  		if (!page)
  			goto not_enough_page;
  		ctrl->map[idx] = page;
  	}
  	return 0;
  not_enough_page:
  	max = idx;
  	for (idx = 0; idx < max; idx++)
  		__free_page(ctrl->map[idx]);
  
  	return -ENOMEM;
  }
9fb4b7cc0   Bob Liu   page_cgroup: add ...
363
364
365
366
367
368
  static struct swap_cgroup *lookup_swap_cgroup(swp_entry_t ent,
  					struct swap_cgroup_ctrl **ctrlp)
  {
  	pgoff_t offset = swp_offset(ent);
  	struct swap_cgroup_ctrl *ctrl;
  	struct page *mappage;
c09ff089a   Hugh Dickins   page_cgroup: fix ...
369
  	struct swap_cgroup *sc;
9fb4b7cc0   Bob Liu   page_cgroup: add ...
370
371
372
373
374
375
  
  	ctrl = &swap_cgroup_ctrl[swp_type(ent)];
  	if (ctrlp)
  		*ctrlp = ctrl;
  
  	mappage = ctrl->map[offset / SC_PER_PAGE];
c09ff089a   Hugh Dickins   page_cgroup: fix ...
376
377
  	sc = page_address(mappage);
  	return sc + offset % SC_PER_PAGE;
9fb4b7cc0   Bob Liu   page_cgroup: add ...
378
  }
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
379
  /**
024914477   Daisuke Nishimura   memcg: move charg...
380
   * swap_cgroup_cmpxchg - cmpxchg mem_cgroup's id for this swp_entry.
dad7557eb   Wanpeng Li   mm: fix kernel-do...
381
   * @ent: swap entry to be cmpxchged
024914477   Daisuke Nishimura   memcg: move charg...
382
383
384
385
   * @old: old id
   * @new: new id
   *
   * Returns old id at success, 0 at failure.
25985edce   Lucas De Marchi   Fix common misspe...
386
   * (There is no mem_cgroup using 0 as its id)
024914477   Daisuke Nishimura   memcg: move charg...
387
388
389
390
   */
  unsigned short swap_cgroup_cmpxchg(swp_entry_t ent,
  					unsigned short old, unsigned short new)
  {
024914477   Daisuke Nishimura   memcg: move charg...
391
  	struct swap_cgroup_ctrl *ctrl;
024914477   Daisuke Nishimura   memcg: move charg...
392
  	struct swap_cgroup *sc;
e9e58a4ec   KAMEZAWA Hiroyuki   memcg: avoid use ...
393
394
  	unsigned long flags;
  	unsigned short retval;
024914477   Daisuke Nishimura   memcg: move charg...
395

9fb4b7cc0   Bob Liu   page_cgroup: add ...
396
  	sc = lookup_swap_cgroup(ent, &ctrl);
024914477   Daisuke Nishimura   memcg: move charg...
397

e9e58a4ec   KAMEZAWA Hiroyuki   memcg: avoid use ...
398
399
400
401
  	spin_lock_irqsave(&ctrl->lock, flags);
  	retval = sc->id;
  	if (retval == old)
  		sc->id = new;
024914477   Daisuke Nishimura   memcg: move charg...
402
  	else
e9e58a4ec   KAMEZAWA Hiroyuki   memcg: avoid use ...
403
404
405
  		retval = 0;
  	spin_unlock_irqrestore(&ctrl->lock, flags);
  	return retval;
024914477   Daisuke Nishimura   memcg: move charg...
406
407
408
  }
  
  /**
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
409
410
   * swap_cgroup_record - record mem_cgroup for this swp_entry.
   * @ent: swap entry to be recorded into
dad7557eb   Wanpeng Li   mm: fix kernel-do...
411
   * @id: mem_cgroup to be recorded
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
412
   *
a3b2d6926   KAMEZAWA Hiroyuki   cgroups: use css ...
413
414
   * Returns old value at success, 0 at failure.
   * (Of course, old value can be 0.)
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
415
   */
a3b2d6926   KAMEZAWA Hiroyuki   cgroups: use css ...
416
  unsigned short swap_cgroup_record(swp_entry_t ent, unsigned short id)
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
417
  {
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
418
  	struct swap_cgroup_ctrl *ctrl;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
419
  	struct swap_cgroup *sc;
a3b2d6926   KAMEZAWA Hiroyuki   cgroups: use css ...
420
  	unsigned short old;
e9e58a4ec   KAMEZAWA Hiroyuki   memcg: avoid use ...
421
  	unsigned long flags;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
422

9fb4b7cc0   Bob Liu   page_cgroup: add ...
423
  	sc = lookup_swap_cgroup(ent, &ctrl);
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
424

e9e58a4ec   KAMEZAWA Hiroyuki   memcg: avoid use ...
425
426
427
428
  	spin_lock_irqsave(&ctrl->lock, flags);
  	old = sc->id;
  	sc->id = id;
  	spin_unlock_irqrestore(&ctrl->lock, flags);
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
429
430
431
432
433
  
  	return old;
  }
  
  /**
9fb4b7cc0   Bob Liu   page_cgroup: add ...
434
   * lookup_swap_cgroup_id - lookup mem_cgroup id tied to swap entry
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
435
436
   * @ent: swap entry to be looked up.
   *
b3ff8a2f9   Hugh Dickins   cgroup: remove st...
437
   * Returns ID of mem_cgroup at success. 0 at failure. (0 is invalid ID)
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
438
   */
9fb4b7cc0   Bob Liu   page_cgroup: add ...
439
  unsigned short lookup_swap_cgroup_id(swp_entry_t ent)
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
440
  {
9fb4b7cc0   Bob Liu   page_cgroup: add ...
441
  	return lookup_swap_cgroup(ent, NULL)->id;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
442
443
444
445
446
447
448
449
450
451
452
  }
  
  int swap_cgroup_swapon(int type, unsigned long max_pages)
  {
  	void *array;
  	unsigned long array_size;
  	unsigned long length;
  	struct swap_cgroup_ctrl *ctrl;
  
  	if (!do_swap_account)
  		return 0;
33278f7f0   Namhyung Kim   memcg: fix off-by...
453
  	length = DIV_ROUND_UP(max_pages, SC_PER_PAGE);
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
454
  	array_size = length * sizeof(void *);
8c1fec1ba   Joe Perches   mm: Convert vmall...
455
  	array = vzalloc(array_size);
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
456
457
  	if (!array)
  		goto nomem;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
458
459
460
461
  	ctrl = &swap_cgroup_ctrl[type];
  	mutex_lock(&swap_cgroup_mutex);
  	ctrl->length = length;
  	ctrl->map = array;
e9e58a4ec   KAMEZAWA Hiroyuki   memcg: avoid use ...
462
  	spin_lock_init(&ctrl->lock);
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
463
464
465
466
  	if (swap_cgroup_prepare(type)) {
  		/* memory shortage */
  		ctrl->map = NULL;
  		ctrl->length = 0;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
467
  		mutex_unlock(&swap_cgroup_mutex);
6a5b18d2b   Namhyung Kim   memcg: move page-...
468
  		vfree(array);
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
469
470
471
  		goto nomem;
  	}
  	mutex_unlock(&swap_cgroup_mutex);
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
472
473
474
475
476
  	return 0;
  nomem:
  	printk(KERN_INFO "couldn't allocate enough memory for swap_cgroup.
  ");
  	printk(KERN_INFO
00a66d297   WANG Cong   mm: remove the le...
477
478
  		"swap_cgroup can be disabled by swapaccount=0 boot option
  ");
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
479
480
481
482
483
  	return -ENOMEM;
  }
  
  void swap_cgroup_swapoff(int type)
  {
6a5b18d2b   Namhyung Kim   memcg: move page-...
484
485
  	struct page **map;
  	unsigned long i, length;
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
486
487
488
489
490
491
492
  	struct swap_cgroup_ctrl *ctrl;
  
  	if (!do_swap_account)
  		return;
  
  	mutex_lock(&swap_cgroup_mutex);
  	ctrl = &swap_cgroup_ctrl[type];
6a5b18d2b   Namhyung Kim   memcg: move page-...
493
494
495
496
497
498
499
500
501
  	map = ctrl->map;
  	length = ctrl->length;
  	ctrl->map = NULL;
  	ctrl->length = 0;
  	mutex_unlock(&swap_cgroup_mutex);
  
  	if (map) {
  		for (i = 0; i < length; i++) {
  			struct page *page = map[i];
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
502
503
504
  			if (page)
  				__free_page(page);
  		}
6a5b18d2b   Namhyung Kim   memcg: move page-...
505
  		vfree(map);
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
506
  	}
27a7faa07   KAMEZAWA Hiroyuki   memcg: swap cgrou...
507
508
509
  }
  
  #endif