Blame view

mm/shmem.c 111 KB
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1
2
3
4
5
6
7
8
  /*
   * Resizable virtual memory filesystem for Linux.
   *
   * Copyright (C) 2000 Linus Torvalds.
   *		 2000 Transmeta Corp.
   *		 2000-2001 Christoph Rohland
   *		 2000-2001 SAP AG
   *		 2002 Red Hat Inc.
6922c0c7a   Hugh Dickins   tmpfs: convert sh...
9
10
   * Copyright (C) 2002-2011 Hugh Dickins.
   * Copyright (C) 2011 Google Inc.
0edd73b33   Hugh Dickins   [PATCH] shmem: re...
11
   * Copyright (C) 2002-2005 VERITAS Software Corporation.
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
12
13
14
15
16
17
   * Copyright (C) 2004 Andi Kleen, SuSE Labs
   *
   * Extended attribute support for tmpfs:
   * Copyright (c) 2004, Luke Kenneth Casson Leighton <lkcl@lkcl.net>
   * Copyright (c) 2004 Red Hat, Inc., James Morris <jmorris@redhat.com>
   *
853ac43ab   Matt Mackall   shmem: unify regu...
18
19
20
   * tiny-shmem:
   * Copyright (c) 2004, 2008 Matt Mackall <mpm@selenic.com>
   *
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
21
22
   * This file is released under the GPL.
   */
853ac43ab   Matt Mackall   shmem: unify regu...
23
24
25
26
  #include <linux/fs.h>
  #include <linux/init.h>
  #include <linux/vfs.h>
  #include <linux/mount.h>
250297edf   Andrew Morton   mm/shmem.c: remov...
27
  #include <linux/ramfs.h>
caefba174   Hugh Dickins   shmem: respect MA...
28
  #include <linux/pagemap.h>
853ac43ab   Matt Mackall   shmem: unify regu...
29
30
  #include <linux/file.h>
  #include <linux/mm.h>
46c9a946d   Arnd Bergmann   shmem: use monoto...
31
  #include <linux/random.h>
174cd4b1e   Ingo Molnar   sched/headers: Pr...
32
  #include <linux/sched/signal.h>
b95f1b31b   Paul Gortmaker   mm: Map most file...
33
  #include <linux/export.h>
853ac43ab   Matt Mackall   shmem: unify regu...
34
  #include <linux/swap.h>
e2e40f2c1   Christoph Hellwig   fs: move struct k...
35
  #include <linux/uio.h>
f3f0e1d21   Kirill A. Shutemov   khugepaged: add s...
36
  #include <linux/khugepaged.h>
749df87bd   Mike Kravetz   mm/shmem: add hug...
37
  #include <linux/hugetlb.h>
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
38
  #include <linux/frontswap.h>
626c3920a   Al Viro   shmem_parse_one()...
39
  #include <linux/fs_parser.h>
853ac43ab   Matt Mackall   shmem: unify regu...
40

95cc09d66   Andrea Arcangeli   userfaultfd: shme...
41
  #include <asm/tlbflush.h> /* for arch/microblaze update_mmu_cache() */
853ac43ab   Matt Mackall   shmem: unify regu...
42
43
44
  static struct vfsmount *shm_mnt;
  
  #ifdef CONFIG_SHMEM
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
45
46
47
48
49
  /*
   * This virtual memory filesystem is heavily based on the ramfs. It
   * extends ramfs by the ability to use swap and honor resource limits
   * which makes it a completely usable filesystem.
   */
39f0247d3   Andreas Gruenbacher   [PATCH] Access Co...
50
  #include <linux/xattr.h>
a56942551   Christoph Hellwig   knfsd: exportfs: ...
51
  #include <linux/exportfs.h>
1c7c474c3   Christoph Hellwig   make generic_acl ...
52
  #include <linux/posix_acl.h>
feda821e7   Christoph Hellwig   fs: remove generi...
53
  #include <linux/posix_acl_xattr.h>
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
54
  #include <linux/mman.h>
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
55
56
57
58
  #include <linux/string.h>
  #include <linux/slab.h>
  #include <linux/backing-dev.h>
  #include <linux/shmem_fs.h>
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
59
  #include <linux/writeback.h>
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
60
  #include <linux/blkdev.h>
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
61
  #include <linux/pagevec.h>
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
62
  #include <linux/percpu_counter.h>
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
63
  #include <linux/falloc.h>
708e3508c   Hugh Dickins   tmpfs: clone shme...
64
  #include <linux/splice.h>
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
65
66
67
68
  #include <linux/security.h>
  #include <linux/swapops.h>
  #include <linux/mempolicy.h>
  #include <linux/namei.h>
b00dc3ad7   Hugh Dickins   [PATCH] tmpfs: fi...
69
  #include <linux/ctype.h>
304dbdb7a   Lee Schermerhorn   [PATCH] add migra...
70
  #include <linux/migrate.h>
c1f60a5a4   Christoph Lameter   [PATCH] reduce MA...
71
  #include <linux/highmem.h>
680d794ba   akpm@linux-foundation.org   mount options: fi...
72
  #include <linux/seq_file.h>
925629278   Mimi Zohar   integrity: specia...
73
  #include <linux/magic.h>
9183df25f   David Herrmann   shm: add memfd_cr...
74
  #include <linux/syscalls.h>
40e041a2c   David Herrmann   shm: add sealing API
75
  #include <linux/fcntl.h>
9183df25f   David Herrmann   shm: add memfd_cr...
76
  #include <uapi/linux/memfd.h>
cfda05267   Mike Rapoport   userfaultfd: shme...
77
  #include <linux/userfaultfd_k.h>
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
78
  #include <linux/rmap.h>
2b4db7961   Amir Goldstein   tmpfs: generate r...
79
  #include <linux/uuid.h>
304dbdb7a   Lee Schermerhorn   [PATCH] add migra...
80

7c0f6ba68   Linus Torvalds   Replace <asm/uacc...
81
  #include <linux/uaccess.h>
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
82

dd56b0464   Mel Gorman   mm: page_alloc: h...
83
  #include "internal.h"
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
84
85
  #define BLOCKS_PER_PAGE  (PAGE_SIZE/512)
  #define VM_ACCT(size)    (PAGE_ALIGN(size) >> PAGE_SHIFT)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
86

1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
87
88
  /* Pretend that each entry is of this size in directory's i_size */
  #define BOGO_DIRENT_SIZE 20
69f07ec93   Hugh Dickins   tmpfs: use kmemdu...
89
90
  /* Symlink up to this size is kmalloc'ed instead of using a swappable page */
  #define SHORT_SYMLINK_LEN 128
1aac14003   Hugh Dickins   tmpfs: quit when ...
91
  /*
f00cdc6df   Hugh Dickins   shmem: fix faulti...
92
93
94
   * shmem_fallocate communicates with shmem_fault or shmem_writepage via
   * inode->i_private (with i_mutex making sure that it has only one user at
   * a time): we would prefer not to enlarge the shmem inode just for that.
1aac14003   Hugh Dickins   tmpfs: quit when ...
95
96
   */
  struct shmem_falloc {
8e205f779   Hugh Dickins   shmem: fix faulti...
97
  	wait_queue_head_t *waitq; /* faults into hole wait for punch to end */
1aac14003   Hugh Dickins   tmpfs: quit when ...
98
99
100
101
102
  	pgoff_t start;		/* start of range currently being fallocated */
  	pgoff_t next;		/* the next page offset to be fallocated */
  	pgoff_t nr_falloced;	/* how many new pages have been fallocated */
  	pgoff_t nr_unswapped;	/* how often writepage refused to swap out */
  };
0b5071dd3   Al Viro   shmem_parse_optio...
103
104
105
106
107
108
109
  struct shmem_options {
  	unsigned long long blocks;
  	unsigned long long inodes;
  	struct mempolicy *mpol;
  	kuid_t uid;
  	kgid_t gid;
  	umode_t mode;
ea3271f71   Chris Down   tmpfs: support 64...
110
  	bool full_inums;
0b5071dd3   Al Viro   shmem_parse_optio...
111
112
113
114
115
  	int huge;
  	int seen;
  #define SHMEM_SEEN_BLOCKS 1
  #define SHMEM_SEEN_INODES 2
  #define SHMEM_SEEN_HUGE 4
ea3271f71   Chris Down   tmpfs: support 64...
116
  #define SHMEM_SEEN_INUMS 8
0b5071dd3   Al Viro   shmem_parse_optio...
117
  };
b76db7354   Andrew Morton   mount-options-fix...
118
  #ifdef CONFIG_TMPFS
680d794ba   akpm@linux-foundation.org   mount options: fi...
119
120
  static unsigned long shmem_default_max_blocks(void)
  {
ca79b0c21   Arun KS   mm: convert total...
121
  	return totalram_pages() / 2;
680d794ba   akpm@linux-foundation.org   mount options: fi...
122
123
124
125
  }
  
  static unsigned long shmem_default_max_inodes(void)
  {
ca79b0c21   Arun KS   mm: convert total...
126
127
128
  	unsigned long nr_pages = totalram_pages();
  
  	return min(nr_pages - totalhigh_pages(), nr_pages / 2);
680d794ba   akpm@linux-foundation.org   mount options: fi...
129
  }
b76db7354   Andrew Morton   mount-options-fix...
130
  #endif
680d794ba   akpm@linux-foundation.org   mount options: fi...
131

bde05d1cc   Hugh Dickins   shmem: replace pa...
132
133
134
  static bool shmem_should_replace_page(struct page *page, gfp_t gfp);
  static int shmem_replace_page(struct page **pagep, gfp_t gfp,
  				struct shmem_inode_info *info, pgoff_t index);
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
135
136
137
138
  static int shmem_swapin_page(struct inode *inode, pgoff_t index,
  			     struct page **pagep, enum sgp_type sgp,
  			     gfp_t gfp, struct vm_area_struct *vma,
  			     vm_fault_t *fault_type);
68da9f055   Hugh Dickins   tmpfs: pass gfp t...
139
  static int shmem_getpage_gfp(struct inode *inode, pgoff_t index,
9e18eb293   Andres Lagar-Cavilla   tmpfs: mem_cgroup...
140
  		struct page **pagep, enum sgp_type sgp,
cfda05267   Mike Rapoport   userfaultfd: shme...
141
  		gfp_t gfp, struct vm_area_struct *vma,
2b7403035   Souptick Joarder   mm: Change return...
142
  		struct vm_fault *vmf, vm_fault_t *fault_type);
68da9f055   Hugh Dickins   tmpfs: pass gfp t...
143

f3f0e1d21   Kirill A. Shutemov   khugepaged: add s...
144
  int shmem_getpage(struct inode *inode, pgoff_t index,
9e18eb293   Andres Lagar-Cavilla   tmpfs: mem_cgroup...
145
  		struct page **pagep, enum sgp_type sgp)
68da9f055   Hugh Dickins   tmpfs: pass gfp t...
146
147
  {
  	return shmem_getpage_gfp(inode, index, pagep, sgp,
cfda05267   Mike Rapoport   userfaultfd: shme...
148
  		mapping_gfp_mask(inode->i_mapping), NULL, NULL, NULL);
68da9f055   Hugh Dickins   tmpfs: pass gfp t...
149
  }
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
150

1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
151
152
153
154
155
156
157
158
159
160
161
162
163
  static inline struct shmem_sb_info *SHMEM_SB(struct super_block *sb)
  {
  	return sb->s_fs_info;
  }
  
  /*
   * shmem_file_setup pre-accounts the whole fixed size of a VM object,
   * for shared memory and for shared anonymous (/dev/zero) mappings
   * (unless MAP_NORESERVE and sysctl_overcommit_memory <= 1),
   * consistent with the pre-accounting of private mappings ...
   */
  static inline int shmem_acct_size(unsigned long flags, loff_t size)
  {
0b0a0806b   Hugh Dickins   shmem: fix shared...
164
  	return (flags & VM_NORESERVE) ?
191c54244   Al Viro   mm: collapse secu...
165
  		0 : security_vm_enough_memory_mm(current->mm, VM_ACCT(size));
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
166
167
168
169
  }
  
  static inline void shmem_unacct_size(unsigned long flags, loff_t size)
  {
0b0a0806b   Hugh Dickins   shmem: fix shared...
170
  	if (!(flags & VM_NORESERVE))
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
171
172
  		vm_unacct_memory(VM_ACCT(size));
  }
771425179   Konstantin Khlebnikov   shmem: update mem...
173
174
175
176
177
178
179
180
181
182
183
184
  static inline int shmem_reacct_size(unsigned long flags,
  		loff_t oldsize, loff_t newsize)
  {
  	if (!(flags & VM_NORESERVE)) {
  		if (VM_ACCT(newsize) > VM_ACCT(oldsize))
  			return security_vm_enough_memory_mm(current->mm,
  					VM_ACCT(newsize) - VM_ACCT(oldsize));
  		else if (VM_ACCT(newsize) < VM_ACCT(oldsize))
  			vm_unacct_memory(VM_ACCT(oldsize) - VM_ACCT(newsize));
  	}
  	return 0;
  }
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
185
186
  /*
   * ... whereas tmpfs objects are accounted incrementally as
75edd345e   Hugh Dickins   tmpfs: preliminar...
187
   * pages are allocated, in order to allow large sparse files.
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
188
189
190
   * shmem_getpage reports shmem_acct_block failure as -ENOSPC not -ENOMEM,
   * so that a failure on a sparse tmpfs mapping will give SIGBUS not OOM.
   */
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
191
  static inline int shmem_acct_block(unsigned long flags, long pages)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
192
  {
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
193
194
195
196
197
  	if (!(flags & VM_NORESERVE))
  		return 0;
  
  	return security_vm_enough_memory_mm(current->mm,
  			pages * VM_ACCT(PAGE_SIZE));
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
198
199
200
201
  }
  
  static inline void shmem_unacct_blocks(unsigned long flags, long pages)
  {
0b0a0806b   Hugh Dickins   shmem: fix shared...
202
  	if (flags & VM_NORESERVE)
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
203
  		vm_unacct_memory(pages * VM_ACCT(PAGE_SIZE));
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
204
  }
0f0796945   Mike Rapoport   shmem: introduce ...
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
  static inline bool shmem_inode_acct_block(struct inode *inode, long pages)
  {
  	struct shmem_inode_info *info = SHMEM_I(inode);
  	struct shmem_sb_info *sbinfo = SHMEM_SB(inode->i_sb);
  
  	if (shmem_acct_block(info->flags, pages))
  		return false;
  
  	if (sbinfo->max_blocks) {
  		if (percpu_counter_compare(&sbinfo->used_blocks,
  					   sbinfo->max_blocks - pages) > 0)
  			goto unacct;
  		percpu_counter_add(&sbinfo->used_blocks, pages);
  	}
  
  	return true;
  
  unacct:
  	shmem_unacct_blocks(info->flags, pages);
  	return false;
  }
  
  static inline void shmem_inode_unacct_blocks(struct inode *inode, long pages)
  {
  	struct shmem_inode_info *info = SHMEM_I(inode);
  	struct shmem_sb_info *sbinfo = SHMEM_SB(inode->i_sb);
  
  	if (sbinfo->max_blocks)
  		percpu_counter_sub(&sbinfo->used_blocks, pages);
  	shmem_unacct_blocks(info->flags, pages);
  }
759b9775c   Hugh Dickins   [PATCH] shmem and...
236
  static const struct super_operations shmem_ops;
f5e54d6e5   Christoph Hellwig   [PATCH] mark addr...
237
  static const struct address_space_operations shmem_aops;
15ad7cdcf   Helge Deller   [PATCH] struct se...
238
  static const struct file_operations shmem_file_operations;
92e1d5be9   Arjan van de Ven   [PATCH] mark stru...
239
240
241
  static const struct inode_operations shmem_inode_operations;
  static const struct inode_operations shmem_dir_inode_operations;
  static const struct inode_operations shmem_special_inode_operations;
f0f37e2f7   Alexey Dobriyan   const: mark struc...
242
  static const struct vm_operations_struct shmem_vm_ops;
779750d20   Kirill A. Shutemov   shmem: split huge...
243
  static struct file_system_type shmem_fs_type;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
244

b0506e488   Mike Rapoport   userfaultfd: shme...
245
246
247
248
  bool vma_is_shmem(struct vm_area_struct *vma)
  {
  	return vma->vm_ops == &shmem_vm_ops;
  }
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
249
  static LIST_HEAD(shmem_swaplist);
cb5f7b9a4   Hugh Dickins   tmpfs: make shmem...
250
  static DEFINE_MUTEX(shmem_swaplist_mutex);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
251

e809d5f0b   Chris Down   tmpfs: per-superb...
252
253
254
255
256
257
258
259
260
261
262
  /*
   * shmem_reserve_inode() performs bookkeeping to reserve a shmem inode, and
   * produces a novel ino for the newly allocated inode.
   *
   * It may also be called when making a hard link to permit the space needed by
   * each dentry. However, in that case, no new inode number is needed since that
   * internally draws from another pool of inode numbers (currently global
   * get_next_ino()). This case is indicated by passing NULL as inop.
   */
  #define SHMEM_INO_BATCH 1024
  static int shmem_reserve_inode(struct super_block *sb, ino_t *inop)
5b04c6890   Pavel Emelyanov   shmem: factor out...
263
264
  {
  	struct shmem_sb_info *sbinfo = SHMEM_SB(sb);
e809d5f0b   Chris Down   tmpfs: per-superb...
265
266
267
  	ino_t ino;
  
  	if (!(sb->s_flags & SB_KERNMOUNT)) {
5b04c6890   Pavel Emelyanov   shmem: factor out...
268
  		spin_lock(&sbinfo->stat_lock);
bb3e96d63   Byron Stanoszek   tmpfs: restore fu...
269
270
271
272
273
274
  		if (sbinfo->max_inodes) {
  			if (!sbinfo->free_inodes) {
  				spin_unlock(&sbinfo->stat_lock);
  				return -ENOSPC;
  			}
  			sbinfo->free_inodes--;
5b04c6890   Pavel Emelyanov   shmem: factor out...
275
  		}
e809d5f0b   Chris Down   tmpfs: per-superb...
276
277
278
279
  		if (inop) {
  			ino = sbinfo->next_ino++;
  			if (unlikely(is_zero_ino(ino)))
  				ino = sbinfo->next_ino++;
ea3271f71   Chris Down   tmpfs: support 64...
280
281
  			if (unlikely(!sbinfo->full_inums &&
  				     ino > UINT_MAX)) {
e809d5f0b   Chris Down   tmpfs: per-superb...
282
283
284
285
  				/*
  				 * Emulate get_next_ino uint wraparound for
  				 * compatibility
  				 */
ea3271f71   Chris Down   tmpfs: support 64...
286
287
288
289
290
291
  				if (IS_ENABLED(CONFIG_64BIT))
  					pr_warn("%s: inode number overflow on device %d, consider using inode64 mount option
  ",
  						__func__, MINOR(sb->s_dev));
  				sbinfo->next_ino = 1;
  				ino = sbinfo->next_ino++;
e809d5f0b   Chris Down   tmpfs: per-superb...
292
293
294
  			}
  			*inop = ino;
  		}
5b04c6890   Pavel Emelyanov   shmem: factor out...
295
  		spin_unlock(&sbinfo->stat_lock);
e809d5f0b   Chris Down   tmpfs: per-superb...
296
297
298
299
300
301
302
303
  	} else if (inop) {
  		/*
  		 * __shmem_file_setup, one of our callers, is lock-free: it
  		 * doesn't hold stat_lock in shmem_reserve_inode since
  		 * max_inodes is always 0, and is called from potentially
  		 * unknown contexts. As such, use a per-cpu batched allocator
  		 * which doesn't require the per-sb stat_lock unless we are at
  		 * the batch boundary.
ea3271f71   Chris Down   tmpfs: support 64...
304
305
306
307
  		 *
  		 * We don't need to worry about inode{32,64} since SB_KERNMOUNT
  		 * shmem mounts are not exposed to userspace, so we don't need
  		 * to worry about things like glibc compatibility.
e809d5f0b   Chris Down   tmpfs: per-superb...
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
  		 */
  		ino_t *next_ino;
  		next_ino = per_cpu_ptr(sbinfo->ino_batch, get_cpu());
  		ino = *next_ino;
  		if (unlikely(ino % SHMEM_INO_BATCH == 0)) {
  			spin_lock(&sbinfo->stat_lock);
  			ino = sbinfo->next_ino;
  			sbinfo->next_ino += SHMEM_INO_BATCH;
  			spin_unlock(&sbinfo->stat_lock);
  			if (unlikely(is_zero_ino(ino)))
  				ino++;
  		}
  		*inop = ino;
  		*next_ino = ++ino;
  		put_cpu();
5b04c6890   Pavel Emelyanov   shmem: factor out...
323
  	}
e809d5f0b   Chris Down   tmpfs: per-superb...
324

5b04c6890   Pavel Emelyanov   shmem: factor out...
325
326
327
328
329
330
331
332
333
334
335
336
  	return 0;
  }
  
  static void shmem_free_inode(struct super_block *sb)
  {
  	struct shmem_sb_info *sbinfo = SHMEM_SB(sb);
  	if (sbinfo->max_inodes) {
  		spin_lock(&sbinfo->stat_lock);
  		sbinfo->free_inodes++;
  		spin_unlock(&sbinfo->stat_lock);
  	}
  }
467118102   Randy Dunlap   mm/shmem and tiny...
337
  /**
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
338
   * shmem_recalc_inode - recalculate the block usage of an inode
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
   * @inode: inode to recalc
   *
   * We have to calculate the free blocks since the mm can drop
   * undirtied hole pages behind our back.
   *
   * But normally   info->alloced == inode->i_mapping->nrpages + info->swapped
   * So mm freed is info->alloced - (inode->i_mapping->nrpages + info->swapped)
   *
   * It has to be called with the spinlock held.
   */
  static void shmem_recalc_inode(struct inode *inode)
  {
  	struct shmem_inode_info *info = SHMEM_I(inode);
  	long freed;
  
  	freed = info->alloced - info->swapped - inode->i_mapping->nrpages;
  	if (freed > 0) {
  		info->alloced -= freed;
54af60421   Hugh Dickins   tmpfs: convert sh...
357
  		inode->i_blocks -= freed * BLOCKS_PER_PAGE;
0f0796945   Mike Rapoport   shmem: introduce ...
358
  		shmem_inode_unacct_blocks(inode, freed);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
359
360
  	}
  }
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
361
362
363
  bool shmem_charge(struct inode *inode, long pages)
  {
  	struct shmem_inode_info *info = SHMEM_I(inode);
4595ef88d   Kirill A. Shutemov   shmem: make shmem...
364
  	unsigned long flags;
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
365

0f0796945   Mike Rapoport   shmem: introduce ...
366
  	if (!shmem_inode_acct_block(inode, pages))
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
367
  		return false;
b1cc94ab2   Mike Rapoport   shmem: shmem_char...
368

aaa52e340   Hugh Dickins   mm/khugepaged: fi...
369
370
  	/* nrpages adjustment first, then shmem_recalc_inode() when balanced */
  	inode->i_mapping->nrpages += pages;
4595ef88d   Kirill A. Shutemov   shmem: make shmem...
371
  	spin_lock_irqsave(&info->lock, flags);
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
372
373
374
  	info->alloced += pages;
  	inode->i_blocks += pages * BLOCKS_PER_PAGE;
  	shmem_recalc_inode(inode);
4595ef88d   Kirill A. Shutemov   shmem: make shmem...
375
  	spin_unlock_irqrestore(&info->lock, flags);
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
376

800d8c63b   Kirill A. Shutemov   shmem: add huge p...
377
378
379
380
381
382
  	return true;
  }
  
  void shmem_uncharge(struct inode *inode, long pages)
  {
  	struct shmem_inode_info *info = SHMEM_I(inode);
4595ef88d   Kirill A. Shutemov   shmem: make shmem...
383
  	unsigned long flags;
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
384

aaa52e340   Hugh Dickins   mm/khugepaged: fi...
385
  	/* nrpages adjustment done by __delete_from_page_cache() or caller */
4595ef88d   Kirill A. Shutemov   shmem: make shmem...
386
  	spin_lock_irqsave(&info->lock, flags);
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
387
388
389
  	info->alloced -= pages;
  	inode->i_blocks -= pages * BLOCKS_PER_PAGE;
  	shmem_recalc_inode(inode);
4595ef88d   Kirill A. Shutemov   shmem: make shmem...
390
  	spin_unlock_irqrestore(&info->lock, flags);
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
391

0f0796945   Mike Rapoport   shmem: introduce ...
392
  	shmem_inode_unacct_blocks(inode, pages);
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
393
  }
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
394
  /*
62f945b6a   Matthew Wilcox   shmem: Convert sh...
395
   * Replace item expected in xarray by a new item, while holding xa_lock.
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
396
   */
62f945b6a   Matthew Wilcox   shmem: Convert sh...
397
  static int shmem_replace_entry(struct address_space *mapping,
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
398
399
  			pgoff_t index, void *expected, void *replacement)
  {
62f945b6a   Matthew Wilcox   shmem: Convert sh...
400
  	XA_STATE(xas, &mapping->i_pages, index);
6dbaf22ce   Johannes Weiner   mm: shmem: save o...
401
  	void *item;
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
402
403
  
  	VM_BUG_ON(!expected);
6dbaf22ce   Johannes Weiner   mm: shmem: save o...
404
  	VM_BUG_ON(!replacement);
62f945b6a   Matthew Wilcox   shmem: Convert sh...
405
  	item = xas_load(&xas);
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
406
407
  	if (item != expected)
  		return -ENOENT;
62f945b6a   Matthew Wilcox   shmem: Convert sh...
408
  	xas_store(&xas, replacement);
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
409
410
411
412
  	return 0;
  }
  
  /*
d18992286   Hugh Dickins   shmem: fix negati...
413
414
415
416
417
418
419
420
421
   * Sometimes, before we decide whether to proceed or to fail, we must check
   * that an entry was not already brought back from swap by a racing thread.
   *
   * Checking page is not enough: by the time a SwapCache page is locked, it
   * might be reused, and again be SwapCache, using the same swap as before.
   */
  static bool shmem_confirm_swap(struct address_space *mapping,
  			       pgoff_t index, swp_entry_t swap)
  {
a12831bf4   Matthew Wilcox   shmem: Convert sh...
422
  	return xa_load(&mapping->i_pages, index) == swp_to_radix_entry(swap);
d18992286   Hugh Dickins   shmem: fix negati...
423
424
425
  }
  
  /*
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
   * Definitions for "huge tmpfs": tmpfs mounted with the huge= option
   *
   * SHMEM_HUGE_NEVER:
   *	disables huge pages for the mount;
   * SHMEM_HUGE_ALWAYS:
   *	enables huge pages for the mount;
   * SHMEM_HUGE_WITHIN_SIZE:
   *	only allocate huge pages if the page will be fully within i_size,
   *	also respect fadvise()/madvise() hints;
   * SHMEM_HUGE_ADVISE:
   *	only allocate huge pages if requested with fadvise()/madvise();
   */
  
  #define SHMEM_HUGE_NEVER	0
  #define SHMEM_HUGE_ALWAYS	1
  #define SHMEM_HUGE_WITHIN_SIZE	2
  #define SHMEM_HUGE_ADVISE	3
  
  /*
   * Special values.
   * Only can be set via /sys/kernel/mm/transparent_hugepage/shmem_enabled:
   *
   * SHMEM_HUGE_DENY:
   *	disables huge on shm_mnt and all mounts, for emergency use;
   * SHMEM_HUGE_FORCE:
   *	enables huge on shm_mnt and all mounts, w/o needing option, for testing;
   *
   */
  #define SHMEM_HUGE_DENY		(-1)
  #define SHMEM_HUGE_FORCE	(-2)
396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
456
  #ifdef CONFIG_TRANSPARENT_HUGEPAGE
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
457
  /* ifdef here to avoid bloating shmem.o when not necessary */
5b9c98f30   Mike Kravetz   mm/shmem: add __r...
458
  static int shmem_huge __read_mostly;
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
459

e5f2249ab   Arnd Bergmann   mm/shmem.c: fix u...
460
  #if defined(CONFIG_SYSFS)
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
  static int shmem_parse_huge(const char *str)
  {
  	if (!strcmp(str, "never"))
  		return SHMEM_HUGE_NEVER;
  	if (!strcmp(str, "always"))
  		return SHMEM_HUGE_ALWAYS;
  	if (!strcmp(str, "within_size"))
  		return SHMEM_HUGE_WITHIN_SIZE;
  	if (!strcmp(str, "advise"))
  		return SHMEM_HUGE_ADVISE;
  	if (!strcmp(str, "deny"))
  		return SHMEM_HUGE_DENY;
  	if (!strcmp(str, "force"))
  		return SHMEM_HUGE_FORCE;
  	return -EINVAL;
  }
e5f2249ab   Arnd Bergmann   mm/shmem.c: fix u...
477
  #endif
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
478

e5f2249ab   Arnd Bergmann   mm/shmem.c: fix u...
479
  #if defined(CONFIG_SYSFS) || defined(CONFIG_TMPFS)
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
  static const char *shmem_format_huge(int huge)
  {
  	switch (huge) {
  	case SHMEM_HUGE_NEVER:
  		return "never";
  	case SHMEM_HUGE_ALWAYS:
  		return "always";
  	case SHMEM_HUGE_WITHIN_SIZE:
  		return "within_size";
  	case SHMEM_HUGE_ADVISE:
  		return "advise";
  	case SHMEM_HUGE_DENY:
  		return "deny";
  	case SHMEM_HUGE_FORCE:
  		return "force";
  	default:
  		VM_BUG_ON(1);
  		return "bad_val";
  	}
  }
f1f5929cd   Jérémy Lefaure   shmem: fix compil...
500
  #endif
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
501

779750d20   Kirill A. Shutemov   shmem: split huge...
502
503
504
505
  static unsigned long shmem_unused_huge_shrink(struct shmem_sb_info *sbinfo,
  		struct shrink_control *sc, unsigned long nr_to_split)
  {
  	LIST_HEAD(list), *pos, *next;
253fd0f02   Kirill A. Shutemov   shmem: fix sleepi...
506
  	LIST_HEAD(to_remove);
779750d20   Kirill A. Shutemov   shmem: split huge...
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
  	struct inode *inode;
  	struct shmem_inode_info *info;
  	struct page *page;
  	unsigned long batch = sc ? sc->nr_to_scan : 128;
  	int removed = 0, split = 0;
  
  	if (list_empty(&sbinfo->shrinklist))
  		return SHRINK_STOP;
  
  	spin_lock(&sbinfo->shrinklist_lock);
  	list_for_each_safe(pos, next, &sbinfo->shrinklist) {
  		info = list_entry(pos, struct shmem_inode_info, shrinklist);
  
  		/* pin the inode */
  		inode = igrab(&info->vfs_inode);
  
  		/* inode is about to be evicted */
  		if (!inode) {
  			list_del_init(&info->shrinklist);
  			removed++;
  			goto next;
  		}
  
  		/* Check if there's anything to gain */
  		if (round_up(inode->i_size, PAGE_SIZE) ==
  				round_up(inode->i_size, HPAGE_PMD_SIZE)) {
253fd0f02   Kirill A. Shutemov   shmem: fix sleepi...
533
  			list_move(&info->shrinklist, &to_remove);
779750d20   Kirill A. Shutemov   shmem: split huge...
534
  			removed++;
779750d20   Kirill A. Shutemov   shmem: split huge...
535
536
537
538
539
540
541
542
543
  			goto next;
  		}
  
  		list_move(&info->shrinklist, &list);
  next:
  		if (!--batch)
  			break;
  	}
  	spin_unlock(&sbinfo->shrinklist_lock);
253fd0f02   Kirill A. Shutemov   shmem: fix sleepi...
544
545
546
547
548
549
  	list_for_each_safe(pos, next, &to_remove) {
  		info = list_entry(pos, struct shmem_inode_info, shrinklist);
  		inode = &info->vfs_inode;
  		list_del_init(&info->shrinklist);
  		iput(inode);
  	}
779750d20   Kirill A. Shutemov   shmem: split huge...
550
551
552
553
554
  	list_for_each_safe(pos, next, &list) {
  		int ret;
  
  		info = list_entry(pos, struct shmem_inode_info, shrinklist);
  		inode = &info->vfs_inode;
b3cd54b25   Kirill A. Shutemov   mm/shmem: do not ...
555
556
  		if (nr_to_split && split >= nr_to_split)
  			goto leave;
779750d20   Kirill A. Shutemov   shmem: split huge...
557

b3cd54b25   Kirill A. Shutemov   mm/shmem: do not ...
558
  		page = find_get_page(inode->i_mapping,
779750d20   Kirill A. Shutemov   shmem: split huge...
559
560
561
  				(inode->i_size & HPAGE_PMD_MASK) >> PAGE_SHIFT);
  		if (!page)
  			goto drop;
b3cd54b25   Kirill A. Shutemov   mm/shmem: do not ...
562
  		/* No huge page at the end of the file: nothing to split */
779750d20   Kirill A. Shutemov   shmem: split huge...
563
  		if (!PageTransHuge(page)) {
779750d20   Kirill A. Shutemov   shmem: split huge...
564
565
566
  			put_page(page);
  			goto drop;
  		}
b3cd54b25   Kirill A. Shutemov   mm/shmem: do not ...
567
568
569
570
571
572
573
574
575
576
577
  		/*
  		 * Leave the inode on the list if we failed to lock
  		 * the page at this time.
  		 *
  		 * Waiting for the lock may lead to deadlock in the
  		 * reclaim path.
  		 */
  		if (!trylock_page(page)) {
  			put_page(page);
  			goto leave;
  		}
779750d20   Kirill A. Shutemov   shmem: split huge...
578
579
580
  		ret = split_huge_page(page);
  		unlock_page(page);
  		put_page(page);
b3cd54b25   Kirill A. Shutemov   mm/shmem: do not ...
581
582
583
  		/* If split failed leave the inode on the list */
  		if (ret)
  			goto leave;
779750d20   Kirill A. Shutemov   shmem: split huge...
584
585
586
587
588
  
  		split++;
  drop:
  		list_del_init(&info->shrinklist);
  		removed++;
b3cd54b25   Kirill A. Shutemov   mm/shmem: do not ...
589
  leave:
779750d20   Kirill A. Shutemov   shmem: split huge...
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
  		iput(inode);
  	}
  
  	spin_lock(&sbinfo->shrinklist_lock);
  	list_splice_tail(&list, &sbinfo->shrinklist);
  	sbinfo->shrinklist_len -= removed;
  	spin_unlock(&sbinfo->shrinklist_lock);
  
  	return split;
  }
  
  static long shmem_unused_huge_scan(struct super_block *sb,
  		struct shrink_control *sc)
  {
  	struct shmem_sb_info *sbinfo = SHMEM_SB(sb);
  
  	if (!READ_ONCE(sbinfo->shrinklist_len))
  		return SHRINK_STOP;
  
  	return shmem_unused_huge_shrink(sbinfo, sc, 0);
  }
  
  static long shmem_unused_huge_count(struct super_block *sb,
  		struct shrink_control *sc)
  {
  	struct shmem_sb_info *sbinfo = SHMEM_SB(sb);
  	return READ_ONCE(sbinfo->shrinklist_len);
  }
396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
618
  #else /* !CONFIG_TRANSPARENT_HUGEPAGE */
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
619
620
  
  #define shmem_huge SHMEM_HUGE_DENY
779750d20   Kirill A. Shutemov   shmem: split huge...
621
622
623
624
625
  static unsigned long shmem_unused_huge_shrink(struct shmem_sb_info *sbinfo,
  		struct shrink_control *sc, unsigned long nr_to_split)
  {
  	return 0;
  }
396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
626
  #endif /* CONFIG_TRANSPARENT_HUGEPAGE */
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
627

89fdcd262   Yang Shi   mm: shmem: make s...
628
629
  static inline bool is_huge_enabled(struct shmem_sb_info *sbinfo)
  {
396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
630
  	if (IS_ENABLED(CONFIG_TRANSPARENT_HUGEPAGE) &&
89fdcd262   Yang Shi   mm: shmem: make s...
631
632
633
634
635
  	    (shmem_huge == SHMEM_HUGE_FORCE || sbinfo->huge) &&
  	    shmem_huge != SHMEM_HUGE_DENY)
  		return true;
  	return false;
  }
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
636
  /*
46f65ec15   Hugh Dickins   tmpfs: convert sh...
637
638
639
640
   * Like add_to_page_cache_locked, but error if expected item has gone.
   */
  static int shmem_add_to_page_cache(struct page *page,
  				   struct address_space *mapping,
3fea5a499   Johannes Weiner   mm: memcontrol: c...
641
642
  				   pgoff_t index, void *expected, gfp_t gfp,
  				   struct mm_struct *charge_mm)
46f65ec15   Hugh Dickins   tmpfs: convert sh...
643
  {
552446a41   Matthew Wilcox   shmem: Convert sh...
644
645
  	XA_STATE_ORDER(xas, &mapping->i_pages, index, compound_order(page));
  	unsigned long i = 0;
d8c6546b1   Matthew Wilcox (Oracle)   mm: introduce com...
646
  	unsigned long nr = compound_nr(page);
3fea5a499   Johannes Weiner   mm: memcontrol: c...
647
  	int error;
46f65ec15   Hugh Dickins   tmpfs: convert sh...
648

800d8c63b   Kirill A. Shutemov   shmem: add huge p...
649
650
  	VM_BUG_ON_PAGE(PageTail(page), page);
  	VM_BUG_ON_PAGE(index != round_down(index, nr), page);
309381fea   Sasha Levin   mm: dump page whe...
651
652
  	VM_BUG_ON_PAGE(!PageLocked(page), page);
  	VM_BUG_ON_PAGE(!PageSwapBacked(page), page);
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
653
  	VM_BUG_ON(expected && PageTransHuge(page));
46f65ec15   Hugh Dickins   tmpfs: convert sh...
654

800d8c63b   Kirill A. Shutemov   shmem: add huge p...
655
  	page_ref_add(page, nr);
b065b4321   Hugh Dickins   shmem: cleanup sh...
656
657
  	page->mapping = mapping;
  	page->index = index;
4c6355b25   Johannes Weiner   mm: memcontrol: c...
658
  	if (!PageSwapCache(page)) {
d9eb1ea2b   Johannes Weiner   mm: memcontrol: d...
659
  		error = mem_cgroup_charge(page, charge_mm, gfp);
4c6355b25   Johannes Weiner   mm: memcontrol: c...
660
661
662
663
664
665
  		if (error) {
  			if (PageTransHuge(page)) {
  				count_vm_event(THP_FILE_FALLBACK);
  				count_vm_event(THP_FILE_FALLBACK_CHARGE);
  			}
  			goto error;
3fea5a499   Johannes Weiner   mm: memcontrol: c...
666
  		}
3fea5a499   Johannes Weiner   mm: memcontrol: c...
667
668
  	}
  	cgroup_throttle_swaprate(page, gfp);
552446a41   Matthew Wilcox   shmem: Convert sh...
669
670
671
672
673
674
675
676
677
678
  	do {
  		void *entry;
  		xas_lock_irq(&xas);
  		entry = xas_find_conflict(&xas);
  		if (entry != expected)
  			xas_set_err(&xas, -EEXIST);
  		xas_create_range(&xas);
  		if (xas_error(&xas))
  			goto unlock;
  next:
4101196b1   Matthew Wilcox (Oracle)   mm: page cache: s...
679
  		xas_store(&xas, page);
552446a41   Matthew Wilcox   shmem: Convert sh...
680
681
682
  		if (++i < nr) {
  			xas_next(&xas);
  			goto next;
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
683
  		}
552446a41   Matthew Wilcox   shmem: Convert sh...
684
  		if (PageTransHuge(page)) {
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
685
  			count_vm_event(THP_FILE_ALLOC);
552446a41   Matthew Wilcox   shmem: Convert sh...
686
  			__inc_node_page_state(page, NR_SHMEM_THPS);
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
687
  		}
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
688
  		mapping->nrpages += nr;
0d1c20722   Johannes Weiner   mm: memcontrol: s...
689
690
  		__mod_lruvec_page_state(page, NR_FILE_PAGES, nr);
  		__mod_lruvec_page_state(page, NR_SHMEM, nr);
552446a41   Matthew Wilcox   shmem: Convert sh...
691
692
693
694
695
  unlock:
  		xas_unlock_irq(&xas);
  	} while (xas_nomem(&xas, gfp));
  
  	if (xas_error(&xas)) {
3fea5a499   Johannes Weiner   mm: memcontrol: c...
696
697
  		error = xas_error(&xas);
  		goto error;
46f65ec15   Hugh Dickins   tmpfs: convert sh...
698
  	}
552446a41   Matthew Wilcox   shmem: Convert sh...
699
700
  
  	return 0;
3fea5a499   Johannes Weiner   mm: memcontrol: c...
701
702
703
704
  error:
  	page->mapping = NULL;
  	page_ref_sub(page, nr);
  	return error;
46f65ec15   Hugh Dickins   tmpfs: convert sh...
705
706
707
  }
  
  /*
6922c0c7a   Hugh Dickins   tmpfs: convert sh...
708
709
710
711
712
713
   * Like delete_from_page_cache, but substitutes swap for page.
   */
  static void shmem_delete_from_page_cache(struct page *page, void *radswap)
  {
  	struct address_space *mapping = page->mapping;
  	int error;
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
714
  	VM_BUG_ON_PAGE(PageCompound(page), page);
b93b01631   Matthew Wilcox   page cache: use x...
715
  	xa_lock_irq(&mapping->i_pages);
62f945b6a   Matthew Wilcox   shmem: Convert sh...
716
  	error = shmem_replace_entry(mapping, page->index, page, radswap);
6922c0c7a   Hugh Dickins   tmpfs: convert sh...
717
718
  	page->mapping = NULL;
  	mapping->nrpages--;
0d1c20722   Johannes Weiner   mm: memcontrol: s...
719
720
  	__dec_lruvec_page_state(page, NR_FILE_PAGES);
  	__dec_lruvec_page_state(page, NR_SHMEM);
b93b01631   Matthew Wilcox   page cache: use x...
721
  	xa_unlock_irq(&mapping->i_pages);
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
722
  	put_page(page);
6922c0c7a   Hugh Dickins   tmpfs: convert sh...
723
724
725
726
  	BUG_ON(error);
  }
  
  /*
c121d3bb7   Matthew Wilcox   shmem: Convert sh...
727
   * Remove swap entry from page cache, free the swap and its page cache.
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
728
729
730
731
   */
  static int shmem_free_swap(struct address_space *mapping,
  			   pgoff_t index, void *radswap)
  {
6dbaf22ce   Johannes Weiner   mm: shmem: save o...
732
  	void *old;
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
733

55f3f7eab   Matthew Wilcox   XArray: Add xa_cm...
734
  	old = xa_cmpxchg_irq(&mapping->i_pages, index, radswap, NULL, 0);
6dbaf22ce   Johannes Weiner   mm: shmem: save o...
735
736
737
738
  	if (old != radswap)
  		return -ENOENT;
  	free_swap_and_cache(radix_to_swp_entry(radswap));
  	return 0;
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
739
740
741
  }
  
  /*
6a15a3709   Vlastimil Babka   mm, proc: reduce ...
742
   * Determine (in bytes) how many of the shmem object's pages mapped by the
48131e03c   Vlastimil Babka   mm, proc: reduce ...
743
   * given offsets are swapped out.
6a15a3709   Vlastimil Babka   mm, proc: reduce ...
744
   *
b93b01631   Matthew Wilcox   page cache: use x...
745
   * This is safe to call without i_mutex or the i_pages lock thanks to RCU,
6a15a3709   Vlastimil Babka   mm, proc: reduce ...
746
747
   * as long as the inode doesn't go away and racy results are not a problem.
   */
48131e03c   Vlastimil Babka   mm, proc: reduce ...
748
749
  unsigned long shmem_partial_swap_usage(struct address_space *mapping,
  						pgoff_t start, pgoff_t end)
6a15a3709   Vlastimil Babka   mm, proc: reduce ...
750
  {
7ae3424fb   Matthew Wilcox   shmem: Convert sh...
751
  	XA_STATE(xas, &mapping->i_pages, start);
6a15a3709   Vlastimil Babka   mm, proc: reduce ...
752
  	struct page *page;
48131e03c   Vlastimil Babka   mm, proc: reduce ...
753
  	unsigned long swapped = 0;
6a15a3709   Vlastimil Babka   mm, proc: reduce ...
754
755
  
  	rcu_read_lock();
7ae3424fb   Matthew Wilcox   shmem: Convert sh...
756
757
  	xas_for_each(&xas, page, end - 1) {
  		if (xas_retry(&xas, page))
2cf938aae   Matthew Wilcox   mm: use radix_tre...
758
  			continue;
3159f943a   Matthew Wilcox   xarray: Replace e...
759
  		if (xa_is_value(page))
6a15a3709   Vlastimil Babka   mm, proc: reduce ...
760
761
762
  			swapped++;
  
  		if (need_resched()) {
7ae3424fb   Matthew Wilcox   shmem: Convert sh...
763
  			xas_pause(&xas);
6a15a3709   Vlastimil Babka   mm, proc: reduce ...
764
  			cond_resched_rcu();
6a15a3709   Vlastimil Babka   mm, proc: reduce ...
765
766
767
768
769
770
771
772
773
  		}
  	}
  
  	rcu_read_unlock();
  
  	return swapped << PAGE_SHIFT;
  }
  
  /*
48131e03c   Vlastimil Babka   mm, proc: reduce ...
774
775
776
   * Determine (in bytes) how many of the shmem object's pages mapped by the
   * given vma is swapped out.
   *
b93b01631   Matthew Wilcox   page cache: use x...
777
   * This is safe to call without i_mutex or the i_pages lock thanks to RCU,
48131e03c   Vlastimil Babka   mm, proc: reduce ...
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
   * as long as the inode doesn't go away and racy results are not a problem.
   */
  unsigned long shmem_swap_usage(struct vm_area_struct *vma)
  {
  	struct inode *inode = file_inode(vma->vm_file);
  	struct shmem_inode_info *info = SHMEM_I(inode);
  	struct address_space *mapping = inode->i_mapping;
  	unsigned long swapped;
  
  	/* Be careful as we don't hold info->lock */
  	swapped = READ_ONCE(info->swapped);
  
  	/*
  	 * The easier cases are when the shmem object has nothing in swap, or
  	 * the vma maps it whole. Then we can simply use the stats that we
  	 * already track.
  	 */
  	if (!swapped)
  		return 0;
  
  	if (!vma->vm_pgoff && vma->vm_end - vma->vm_start >= inode->i_size)
  		return swapped << PAGE_SHIFT;
  
  	/* Here comes the more involved part */
  	return shmem_partial_swap_usage(mapping,
  			linear_page_index(vma, vma->vm_start),
  			linear_page_index(vma, vma->vm_end));
  }
  
  /*
245132643   Hugh Dickins   SHM_UNLOCK: fix U...
808
809
810
811
812
813
814
   * SysV IPC SHM_UNLOCK restore Unevictable pages to their evictable lists.
   */
  void shmem_unlock_mapping(struct address_space *mapping)
  {
  	struct pagevec pvec;
  	pgoff_t indices[PAGEVEC_SIZE];
  	pgoff_t index = 0;
866798201   Mel Gorman   mm, pagevec: remo...
815
  	pagevec_init(&pvec);
245132643   Hugh Dickins   SHM_UNLOCK: fix U...
816
817
818
819
820
821
822
823
  	/*
  	 * Minor point, but we might as well stop if someone else SHM_LOCKs it.
  	 */
  	while (!mapping_unevictable(mapping)) {
  		/*
  		 * Avoid pagevec_lookup(): find_get_pages() returns 0 as if it
  		 * has finished, if it hits a row of PAGEVEC_SIZE swap entries.
  		 */
0cd6144aa   Johannes Weiner   mm + fs: prepare ...
824
825
  		pvec.nr = find_get_entries(mapping, index,
  					   PAGEVEC_SIZE, pvec.pages, indices);
245132643   Hugh Dickins   SHM_UNLOCK: fix U...
826
827
828
  		if (!pvec.nr)
  			break;
  		index = indices[pvec.nr - 1] + 1;
0cd6144aa   Johannes Weiner   mm + fs: prepare ...
829
  		pagevec_remove_exceptionals(&pvec);
64e3d12f7   Kuo-Hsin Yang   mm, drm/i915: mar...
830
  		check_move_unevictable_pages(&pvec);
245132643   Hugh Dickins   SHM_UNLOCK: fix U...
831
832
833
  		pagevec_release(&pvec);
  		cond_resched();
  	}
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
834
835
836
  }
  
  /*
71725ed10   Hugh Dickins   mm: huge tmpfs: t...
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
   * Check whether a hole-punch or truncation needs to split a huge page,
   * returning true if no split was required, or the split has been successful.
   *
   * Eviction (or truncation to 0 size) should never need to split a huge page;
   * but in rare cases might do so, if shmem_undo_range() failed to trylock on
   * head, and then succeeded to trylock on tail.
   *
   * A split can only succeed when there are no additional references on the
   * huge page: so the split below relies upon find_get_entries() having stopped
   * when it found a subpage of the huge page, without getting further references.
   */
  static bool shmem_punch_compound(struct page *page, pgoff_t start, pgoff_t end)
  {
  	if (!PageTransCompound(page))
  		return true;
  
  	/* Just proceed to delete a huge page wholly within the range punched */
  	if (PageHead(page) &&
  	    page->index >= start && page->index + HPAGE_PMD_NR <= end)
  		return true;
  
  	/* Try to split huge page, so we can truly punch the hole or truncate */
  	return split_huge_page(page) >= 0;
  }
  
  /*
7f4446eef   Matthew Wilcox   shmem: Comment fi...
863
   * Remove range of pages and swap entries from page cache, and free them.
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
864
   * If !unfalloc, truncate or punch hole; if unfalloc, undo failed fallocate.
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
865
   */
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
866
867
  static void shmem_undo_range(struct inode *inode, loff_t lstart, loff_t lend,
  								 bool unfalloc)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
868
  {
285b2c4fd   Hugh Dickins   tmpfs: demolish o...
869
  	struct address_space *mapping = inode->i_mapping;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
870
  	struct shmem_inode_info *info = SHMEM_I(inode);
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
871
872
873
874
  	pgoff_t start = (lstart + PAGE_SIZE - 1) >> PAGE_SHIFT;
  	pgoff_t end = (lend + 1) >> PAGE_SHIFT;
  	unsigned int partial_start = lstart & (PAGE_SIZE - 1);
  	unsigned int partial_end = (lend + 1) & (PAGE_SIZE - 1);
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
875
  	struct pagevec pvec;
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
876
877
  	pgoff_t indices[PAGEVEC_SIZE];
  	long nr_swaps_freed = 0;
285b2c4fd   Hugh Dickins   tmpfs: demolish o...
878
  	pgoff_t index;
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
879
  	int i;
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
880
881
  	if (lend == -1)
  		end = -1;	/* unsigned, so actually very big */
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
882

866798201   Mel Gorman   mm, pagevec: remo...
883
  	pagevec_init(&pvec);
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
884
  	index = start;
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
885
  	while (index < end) {
0cd6144aa   Johannes Weiner   mm + fs: prepare ...
886
887
888
  		pvec.nr = find_get_entries(mapping, index,
  			min(end - index, (pgoff_t)PAGEVEC_SIZE),
  			pvec.pages, indices);
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
889
890
  		if (!pvec.nr)
  			break;
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
891
892
  		for (i = 0; i < pagevec_count(&pvec); i++) {
  			struct page *page = pvec.pages[i];
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
893
  			index = indices[i];
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
894
  			if (index >= end)
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
895
  				break;
3159f943a   Matthew Wilcox   xarray: Replace e...
896
  			if (xa_is_value(page)) {
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
897
898
  				if (unfalloc)
  					continue;
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
899
900
  				nr_swaps_freed += !shmem_free_swap(mapping,
  								index, page);
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
901
  				continue;
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
902
  			}
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
903
  			VM_BUG_ON_PAGE(page_to_pgoff(page) != index, page);
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
904
  			if (!trylock_page(page))
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
905
  				continue;
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
906

71725ed10   Hugh Dickins   mm: huge tmpfs: t...
907
908
909
910
  			if ((!unfalloc || !PageUptodate(page)) &&
  			    page_mapping(page) == mapping) {
  				VM_BUG_ON_PAGE(PageWriteback(page), page);
  				if (shmem_punch_compound(page, start, end))
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
911
  					truncate_inode_page(mapping, page);
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
912
  			}
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
913
914
  			unlock_page(page);
  		}
0cd6144aa   Johannes Weiner   mm + fs: prepare ...
915
  		pagevec_remove_exceptionals(&pvec);
245132643   Hugh Dickins   SHM_UNLOCK: fix U...
916
  		pagevec_release(&pvec);
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
917
918
919
  		cond_resched();
  		index++;
  	}
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
920

83e4fa9c1   Hugh Dickins   tmpfs: support fa...
921
  	if (partial_start) {
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
922
  		struct page *page = NULL;
9e18eb293   Andres Lagar-Cavilla   tmpfs: mem_cgroup...
923
  		shmem_getpage(inode, start - 1, &page, SGP_READ);
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
924
  		if (page) {
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
925
  			unsigned int top = PAGE_SIZE;
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
926
927
928
929
930
931
932
  			if (start > end) {
  				top = partial_end;
  				partial_end = 0;
  			}
  			zero_user_segment(page, partial_start, top);
  			set_page_dirty(page);
  			unlock_page(page);
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
933
  			put_page(page);
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
934
935
936
937
  		}
  	}
  	if (partial_end) {
  		struct page *page = NULL;
9e18eb293   Andres Lagar-Cavilla   tmpfs: mem_cgroup...
938
  		shmem_getpage(inode, end, &page, SGP_READ);
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
939
940
  		if (page) {
  			zero_user_segment(page, 0, partial_end);
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
941
942
  			set_page_dirty(page);
  			unlock_page(page);
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
943
  			put_page(page);
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
944
945
  		}
  	}
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
946
947
  	if (start >= end)
  		return;
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
948
949
  
  	index = start;
b1a366500   Hugh Dickins   shmem: fix splici...
950
  	while (index < end) {
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
951
  		cond_resched();
0cd6144aa   Johannes Weiner   mm + fs: prepare ...
952
953
  
  		pvec.nr = find_get_entries(mapping, index,
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
954
  				min(end - index, (pgoff_t)PAGEVEC_SIZE),
0cd6144aa   Johannes Weiner   mm + fs: prepare ...
955
  				pvec.pages, indices);
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
956
  		if (!pvec.nr) {
b1a366500   Hugh Dickins   shmem: fix splici...
957
958
  			/* If all gone or hole-punch or unfalloc, we're done */
  			if (index == start || end != -1)
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
959
  				break;
b1a366500   Hugh Dickins   shmem: fix splici...
960
  			/* But if truncating, restart to make sure all gone */
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
961
962
963
  			index = start;
  			continue;
  		}
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
964
965
  		for (i = 0; i < pagevec_count(&pvec); i++) {
  			struct page *page = pvec.pages[i];
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
966
  			index = indices[i];
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
967
  			if (index >= end)
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
968
  				break;
3159f943a   Matthew Wilcox   xarray: Replace e...
969
  			if (xa_is_value(page)) {
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
970
971
  				if (unfalloc)
  					continue;
b1a366500   Hugh Dickins   shmem: fix splici...
972
973
974
975
976
977
  				if (shmem_free_swap(mapping, index, page)) {
  					/* Swap was replaced by page: retry */
  					index--;
  					break;
  				}
  				nr_swaps_freed++;
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
978
979
  				continue;
  			}
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
980
  			lock_page(page);
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
981

1635f6a74   Hugh Dickins   tmpfs: undo fallo...
982
  			if (!unfalloc || !PageUptodate(page)) {
71725ed10   Hugh Dickins   mm: huge tmpfs: t...
983
  				if (page_mapping(page) != mapping) {
b1a366500   Hugh Dickins   shmem: fix splici...
984
985
986
987
  					/* Page was replaced by swap: retry */
  					unlock_page(page);
  					index--;
  					break;
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
988
  				}
71725ed10   Hugh Dickins   mm: huge tmpfs: t...
989
990
991
  				VM_BUG_ON_PAGE(PageWriteback(page), page);
  				if (shmem_punch_compound(page, start, end))
  					truncate_inode_page(mapping, page);
0783ac95b   Hugh Dickins   mm/shmem: fix bui...
992
  				else if (IS_ENABLED(CONFIG_TRANSPARENT_HUGEPAGE)) {
71725ed10   Hugh Dickins   mm: huge tmpfs: t...
993
994
995
996
997
998
999
1000
  					/* Wipe the page and don't get stuck */
  					clear_highpage(page);
  					flush_dcache_page(page);
  					set_page_dirty(page);
  					if (index <
  					    round_up(start, HPAGE_PMD_NR))
  						start = index + 1;
  				}
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
1001
  			}
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
1002
1003
  			unlock_page(page);
  		}
0cd6144aa   Johannes Weiner   mm + fs: prepare ...
1004
  		pagevec_remove_exceptionals(&pvec);
245132643   Hugh Dickins   SHM_UNLOCK: fix U...
1005
  		pagevec_release(&pvec);
bda97eab0   Hugh Dickins   tmpfs: copy trunc...
1006
1007
  		index++;
  	}
94c1e62df   Hugh Dickins   tmpfs: take contr...
1008

4595ef88d   Kirill A. Shutemov   shmem: make shmem...
1009
  	spin_lock_irq(&info->lock);
7a5d0fbb2   Hugh Dickins   tmpfs: convert sh...
1010
  	info->swapped -= nr_swaps_freed;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1011
  	shmem_recalc_inode(inode);
4595ef88d   Kirill A. Shutemov   shmem: make shmem...
1012
  	spin_unlock_irq(&info->lock);
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
1013
  }
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1014

1635f6a74   Hugh Dickins   tmpfs: undo fallo...
1015
1016
1017
  void shmem_truncate_range(struct inode *inode, loff_t lstart, loff_t lend)
  {
  	shmem_undo_range(inode, lstart, lend, false);
078cd8279   Deepa Dinamani   fs: Replace CURRE...
1018
  	inode->i_ctime = inode->i_mtime = current_time(inode);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1019
  }
94c1e62df   Hugh Dickins   tmpfs: take contr...
1020
  EXPORT_SYMBOL_GPL(shmem_truncate_range);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1021

a528d35e8   David Howells   statx: Add a syst...
1022
1023
  static int shmem_getattr(const struct path *path, struct kstat *stat,
  			 u32 request_mask, unsigned int query_flags)
44a30220b   Yu Zhao   shmem: recalculat...
1024
  {
a528d35e8   David Howells   statx: Add a syst...
1025
  	struct inode *inode = path->dentry->d_inode;
44a30220b   Yu Zhao   shmem: recalculat...
1026
  	struct shmem_inode_info *info = SHMEM_I(inode);
89fdcd262   Yang Shi   mm: shmem: make s...
1027
  	struct shmem_sb_info *sb_info = SHMEM_SB(inode->i_sb);
44a30220b   Yu Zhao   shmem: recalculat...
1028

d0424c429   Hugh Dickins   tmpfs: avoid a li...
1029
  	if (info->alloced - info->swapped != inode->i_mapping->nrpages) {
4595ef88d   Kirill A. Shutemov   shmem: make shmem...
1030
  		spin_lock_irq(&info->lock);
d0424c429   Hugh Dickins   tmpfs: avoid a li...
1031
  		shmem_recalc_inode(inode);
4595ef88d   Kirill A. Shutemov   shmem: make shmem...
1032
  		spin_unlock_irq(&info->lock);
d0424c429   Hugh Dickins   tmpfs: avoid a li...
1033
  	}
44a30220b   Yu Zhao   shmem: recalculat...
1034
  	generic_fillattr(inode, stat);
89fdcd262   Yang Shi   mm: shmem: make s...
1035
1036
1037
  
  	if (is_huge_enabled(sb_info))
  		stat->blksize = HPAGE_PMD_SIZE;
44a30220b   Yu Zhao   shmem: recalculat...
1038
1039
  	return 0;
  }
94c1e62df   Hugh Dickins   tmpfs: take contr...
1040
  static int shmem_setattr(struct dentry *dentry, struct iattr *attr)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1041
  {
75c3cfa85   David Howells   VFS: assorted wei...
1042
  	struct inode *inode = d_inode(dentry);
40e041a2c   David Herrmann   shm: add sealing API
1043
  	struct shmem_inode_info *info = SHMEM_I(inode);
779750d20   Kirill A. Shutemov   shmem: split huge...
1044
  	struct shmem_sb_info *sbinfo = SHMEM_SB(inode->i_sb);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1045
  	int error;
31051c85b   Jan Kara   fs: Give dentry t...
1046
  	error = setattr_prepare(dentry, attr);
db78b877f   Christoph Hellwig   always call inode...
1047
1048
  	if (error)
  		return error;
94c1e62df   Hugh Dickins   tmpfs: take contr...
1049
1050
1051
  	if (S_ISREG(inode->i_mode) && (attr->ia_valid & ATTR_SIZE)) {
  		loff_t oldsize = inode->i_size;
  		loff_t newsize = attr->ia_size;
3889e6e76   npiggin@suse.de   tmpfs: convert to...
1052

40e041a2c   David Herrmann   shm: add sealing API
1053
1054
1055
1056
  		/* protected by i_mutex */
  		if ((newsize < oldsize && (info->seals & F_SEAL_SHRINK)) ||
  		    (newsize > oldsize && (info->seals & F_SEAL_GROW)))
  			return -EPERM;
94c1e62df   Hugh Dickins   tmpfs: take contr...
1057
  		if (newsize != oldsize) {
771425179   Konstantin Khlebnikov   shmem: update mem...
1058
1059
1060
1061
  			error = shmem_reacct_size(SHMEM_I(inode)->flags,
  					oldsize, newsize);
  			if (error)
  				return error;
94c1e62df   Hugh Dickins   tmpfs: take contr...
1062
  			i_size_write(inode, newsize);
078cd8279   Deepa Dinamani   fs: Replace CURRE...
1063
  			inode->i_ctime = inode->i_mtime = current_time(inode);
94c1e62df   Hugh Dickins   tmpfs: take contr...
1064
  		}
afa2db2fb   Josef Bacik   tmpfs: truncate p...
1065
  		if (newsize <= oldsize) {
94c1e62df   Hugh Dickins   tmpfs: take contr...
1066
  			loff_t holebegin = round_up(newsize, PAGE_SIZE);
d0424c429   Hugh Dickins   tmpfs: avoid a li...
1067
1068
1069
1070
1071
1072
  			if (oldsize > holebegin)
  				unmap_mapping_range(inode->i_mapping,
  							holebegin, 0, 1);
  			if (info->alloced)
  				shmem_truncate_range(inode,
  							newsize, (loff_t)-1);
94c1e62df   Hugh Dickins   tmpfs: take contr...
1073
  			/* unmap again to remove racily COWed private pages */
d0424c429   Hugh Dickins   tmpfs: avoid a li...
1074
1075
1076
  			if (oldsize > holebegin)
  				unmap_mapping_range(inode->i_mapping,
  							holebegin, 0, 1);
779750d20   Kirill A. Shutemov   shmem: split huge...
1077
1078
1079
1080
1081
  
  			/*
  			 * Part of the huge page can be beyond i_size: subject
  			 * to shrink under memory pressure.
  			 */
396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
1082
  			if (IS_ENABLED(CONFIG_TRANSPARENT_HUGEPAGE)) {
779750d20   Kirill A. Shutemov   shmem: split huge...
1083
  				spin_lock(&sbinfo->shrinklist_lock);
d041353dc   Cong Wang   mm: fix list corr...
1084
1085
1086
1087
1088
  				/*
  				 * _careful to defend against unlocked access to
  				 * ->shrink_list in shmem_unused_huge_shrink()
  				 */
  				if (list_empty_careful(&info->shrinklist)) {
779750d20   Kirill A. Shutemov   shmem: split huge...
1089
1090
1091
1092
1093
1094
  					list_add_tail(&info->shrinklist,
  							&sbinfo->shrinklist);
  					sbinfo->shrinklist_len++;
  				}
  				spin_unlock(&sbinfo->shrinklist_lock);
  			}
94c1e62df   Hugh Dickins   tmpfs: take contr...
1095
  		}
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1096
  	}
db78b877f   Christoph Hellwig   always call inode...
1097
  	setattr_copy(inode, attr);
db78b877f   Christoph Hellwig   always call inode...
1098
  	if (attr->ia_valid & ATTR_MODE)
feda821e7   Christoph Hellwig   fs: remove generi...
1099
  		error = posix_acl_chmod(inode, inode->i_mode);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1100
1101
  	return error;
  }
1f895f75d   Al Viro   switch shmem.c to...
1102
  static void shmem_evict_inode(struct inode *inode)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1103
  {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1104
  	struct shmem_inode_info *info = SHMEM_I(inode);
779750d20   Kirill A. Shutemov   shmem: split huge...
1105
  	struct shmem_sb_info *sbinfo = SHMEM_SB(inode->i_sb);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1106

3889e6e76   npiggin@suse.de   tmpfs: convert to...
1107
  	if (inode->i_mapping->a_ops == &shmem_aops) {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1108
1109
  		shmem_unacct_size(info->flags, inode->i_size);
  		inode->i_size = 0;
3889e6e76   npiggin@suse.de   tmpfs: convert to...
1110
  		shmem_truncate_range(inode, 0, (loff_t)-1);
779750d20   Kirill A. Shutemov   shmem: split huge...
1111
1112
1113
1114
1115
1116
1117
1118
  		if (!list_empty(&info->shrinklist)) {
  			spin_lock(&sbinfo->shrinklist_lock);
  			if (!list_empty(&info->shrinklist)) {
  				list_del_init(&info->shrinklist);
  				sbinfo->shrinklist_len--;
  			}
  			spin_unlock(&sbinfo->shrinklist_lock);
  		}
af53d3e9e   Hugh Dickins   mm: swapoff: shme...
1119
1120
1121
1122
  		while (!list_empty(&info->swaplist)) {
  			/* Wait while shmem_unuse() is scanning this inode... */
  			wait_var_event(&info->stop_eviction,
  				       !atomic_read(&info->stop_eviction));
cb5f7b9a4   Hugh Dickins   tmpfs: make shmem...
1123
  			mutex_lock(&shmem_swaplist_mutex);
af53d3e9e   Hugh Dickins   mm: swapoff: shme...
1124
1125
1126
  			/* ...but beware of the race if we peeked too early */
  			if (!atomic_read(&info->stop_eviction))
  				list_del_init(&info->swaplist);
cb5f7b9a4   Hugh Dickins   tmpfs: make shmem...
1127
  			mutex_unlock(&shmem_swaplist_mutex);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1128
  		}
3ed47db34   Al Viro   make sure that fr...
1129
  	}
b09e0fa4b   Eric Paris   tmpfs: implement ...
1130

38f386574   Aristeu Rozanski   xattr: extract si...
1131
  	simple_xattrs_free(&info->xattrs);
0f3c42f52   Hugh Dickins   tmpfs: change fin...
1132
  	WARN_ON(inode->i_blocks);
5b04c6890   Pavel Emelyanov   shmem: factor out...
1133
  	shmem_free_inode(inode->i_sb);
dbd5768f8   Jan Kara   vfs: Rename end_w...
1134
  	clear_inode(inode);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1135
  }
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1136
1137
1138
1139
1140
  extern struct swap_info_struct *swap_info[];
  
  static int shmem_find_swap_entries(struct address_space *mapping,
  				   pgoff_t start, unsigned int nr_entries,
  				   struct page **entries, pgoff_t *indices,
870395465   Hugh Dickins   mm: swapoff: shme...
1141
  				   unsigned int type, bool frontswap)
478922e2b   Matthew Wilcox   radix-tree: delet...
1142
  {
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1143
1144
  	XA_STATE(xas, &mapping->i_pages, start);
  	struct page *page;
870395465   Hugh Dickins   mm: swapoff: shme...
1145
  	swp_entry_t entry;
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1146
1147
1148
1149
  	unsigned int ret = 0;
  
  	if (!nr_entries)
  		return 0;
478922e2b   Matthew Wilcox   radix-tree: delet...
1150
1151
  
  	rcu_read_lock();
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1152
1153
  	xas_for_each(&xas, page, ULONG_MAX) {
  		if (xas_retry(&xas, page))
5b9c98f30   Mike Kravetz   mm/shmem: add __r...
1154
  			continue;
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1155
1156
  
  		if (!xa_is_value(page))
478922e2b   Matthew Wilcox   radix-tree: delet...
1157
  			continue;
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1158

870395465   Hugh Dickins   mm: swapoff: shme...
1159
1160
1161
1162
1163
1164
  		entry = radix_to_swp_entry(page);
  		if (swp_type(entry) != type)
  			continue;
  		if (frontswap &&
  		    !frontswap_test(swap_info[type], swp_offset(entry)))
  			continue;
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
  
  		indices[ret] = xas.xa_index;
  		entries[ret] = page;
  
  		if (need_resched()) {
  			xas_pause(&xas);
  			cond_resched_rcu();
  		}
  		if (++ret == nr_entries)
  			break;
478922e2b   Matthew Wilcox   radix-tree: delet...
1175
  	}
478922e2b   Matthew Wilcox   radix-tree: delet...
1176
  	rcu_read_unlock();
e21a29552   Matthew Wilcox   shmem: Convert fi...
1177

b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1178
  	return ret;
478922e2b   Matthew Wilcox   radix-tree: delet...
1179
  }
46f65ec15   Hugh Dickins   tmpfs: convert sh...
1180
  /*
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1181
1182
   * Move the swapped pages for an inode to page cache. Returns the count
   * of pages swapped in, or the error in case of failure.
46f65ec15   Hugh Dickins   tmpfs: convert sh...
1183
   */
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1184
1185
  static int shmem_unuse_swap_entries(struct inode *inode, struct pagevec pvec,
  				    pgoff_t *indices)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1186
  {
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1187
1188
  	int i = 0;
  	int ret = 0;
bde05d1cc   Hugh Dickins   shmem: replace pa...
1189
  	int error = 0;
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1190
  	struct address_space *mapping = inode->i_mapping;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1191

b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1192
1193
  	for (i = 0; i < pvec.nr; i++) {
  		struct page *page = pvec.pages[i];
2e0e26c76   Hugh Dickins   tmpfs: open a win...
1194

b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
  		if (!xa_is_value(page))
  			continue;
  		error = shmem_swapin_page(inode, indices[i],
  					  &page, SGP_CACHE,
  					  mapping_gfp_mask(mapping),
  					  NULL, NULL);
  		if (error == 0) {
  			unlock_page(page);
  			put_page(page);
  			ret++;
  		}
  		if (error == -ENOMEM)
  			break;
  		error = 0;
bde05d1cc   Hugh Dickins   shmem: replace pa...
1209
  	}
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1210
1211
  	return error ? error : ret;
  }
bde05d1cc   Hugh Dickins   shmem: replace pa...
1212

b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
  /*
   * If swap found in inode, free it and move page from swapcache to filecache.
   */
  static int shmem_unuse_inode(struct inode *inode, unsigned int type,
  			     bool frontswap, unsigned long *fs_pages_to_unuse)
  {
  	struct address_space *mapping = inode->i_mapping;
  	pgoff_t start = 0;
  	struct pagevec pvec;
  	pgoff_t indices[PAGEVEC_SIZE];
  	bool frontswap_partial = (frontswap && *fs_pages_to_unuse > 0);
  	int ret = 0;
  
  	pagevec_init(&pvec);
  	do {
  		unsigned int nr_entries = PAGEVEC_SIZE;
  
  		if (frontswap_partial && *fs_pages_to_unuse < PAGEVEC_SIZE)
  			nr_entries = *fs_pages_to_unuse;
  
  		pvec.nr = shmem_find_swap_entries(mapping, start, nr_entries,
  						  pvec.pages, indices,
870395465   Hugh Dickins   mm: swapoff: shme...
1235
  						  type, frontswap);
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1236
1237
1238
  		if (pvec.nr == 0) {
  			ret = 0;
  			break;
46f65ec15   Hugh Dickins   tmpfs: convert sh...
1239
  		}
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
  
  		ret = shmem_unuse_swap_entries(inode, pvec, indices);
  		if (ret < 0)
  			break;
  
  		if (frontswap_partial) {
  			*fs_pages_to_unuse -= ret;
  			if (*fs_pages_to_unuse == 0) {
  				ret = FRONTSWAP_PAGES_UNUSED;
  				break;
  			}
  		}
  
  		start = indices[pvec.nr - 1];
  	} while (true);
  
  	return ret;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1257
1258
1259
  }
  
  /*
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1260
1261
1262
   * Read all the shared memory data that resides in the swap
   * device 'type' back into memory, so the swap device can be
   * unused.
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1263
   */
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1264
1265
  int shmem_unuse(unsigned int type, bool frontswap,
  		unsigned long *fs_pages_to_unuse)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1266
  {
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1267
  	struct shmem_inode_info *info, *next;
bde05d1cc   Hugh Dickins   shmem: replace pa...
1268
  	int error = 0;
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1269
1270
1271
1272
  	if (list_empty(&shmem_swaplist))
  		return 0;
  
  	mutex_lock(&shmem_swaplist_mutex);
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1273
1274
  	list_for_each_entry_safe(info, next, &shmem_swaplist, swaplist) {
  		if (!info->swapped) {
6922c0c7a   Hugh Dickins   tmpfs: convert sh...
1275
  			list_del_init(&info->swaplist);
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1276
1277
  			continue;
  		}
af53d3e9e   Hugh Dickins   mm: swapoff: shme...
1278
1279
1280
1281
1282
1283
1284
  		/*
  		 * Drop the swaplist mutex while searching the inode for swap;
  		 * but before doing so, make sure shmem_evict_inode() will not
  		 * remove placeholder inode from swaplist, nor let it be freed
  		 * (igrab() would protect from unlink, but not from unmount).
  		 */
  		atomic_inc(&info->stop_eviction);
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1285
  		mutex_unlock(&shmem_swaplist_mutex);
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1286

af53d3e9e   Hugh Dickins   mm: swapoff: shme...
1287
  		error = shmem_unuse_inode(&info->vfs_inode, type, frontswap,
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1288
  					  fs_pages_to_unuse);
cb5f7b9a4   Hugh Dickins   tmpfs: make shmem...
1289
  		cond_resched();
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1290
1291
1292
1293
1294
  
  		mutex_lock(&shmem_swaplist_mutex);
  		next = list_next_entry(info, swaplist);
  		if (!info->swapped)
  			list_del_init(&info->swaplist);
af53d3e9e   Hugh Dickins   mm: swapoff: shme...
1295
1296
  		if (atomic_dec_and_test(&info->stop_eviction))
  			wake_up_var(&info->stop_eviction);
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1297
  		if (error)
778dd893a   Hugh Dickins   tmpfs: fix race b...
1298
  			break;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1299
  	}
cb5f7b9a4   Hugh Dickins   tmpfs: make shmem...
1300
  	mutex_unlock(&shmem_swaplist_mutex);
778dd893a   Hugh Dickins   tmpfs: fix race b...
1301

778dd893a   Hugh Dickins   tmpfs: fix race b...
1302
  	return error;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1303
1304
1305
1306
1307
1308
1309
1310
  }
  
  /*
   * Move the page from the page cache to the swap cache.
   */
  static int shmem_writepage(struct page *page, struct writeback_control *wbc)
  {
  	struct shmem_inode_info *info;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1311
  	struct address_space *mapping;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1312
  	struct inode *inode;
6922c0c7a   Hugh Dickins   tmpfs: convert sh...
1313
1314
  	swp_entry_t swap;
  	pgoff_t index;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1315

800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1316
  	VM_BUG_ON_PAGE(PageCompound(page), page);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1317
  	BUG_ON(!PageLocked(page));
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1318
1319
1320
1321
1322
1323
  	mapping = page->mapping;
  	index = page->index;
  	inode = mapping->host;
  	info = SHMEM_I(inode);
  	if (info->flags & VM_LOCKED)
  		goto redirty;
d9fe526a8   Hugh Dickins   tmpfs: allow file...
1324
  	if (!total_swap_pages)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1325
  		goto redirty;
d9fe526a8   Hugh Dickins   tmpfs: allow file...
1326
  	/*
97b713ba3   Christoph Hellwig   fs: kill BDI_CAP_...
1327
1328
1329
1330
1331
  	 * Our capabilities prevent regular writeback or sync from ever calling
  	 * shmem_writepage; but a stacking filesystem might use ->writepage of
  	 * its underlying filesystem, in which case tmpfs should write out to
  	 * swap only in response to memory pressure, and not for the writeback
  	 * threads or sync.
d9fe526a8   Hugh Dickins   tmpfs: allow file...
1332
  	 */
48f170fb7   Hugh Dickins   tmpfs: simplify u...
1333
1334
1335
1336
  	if (!wbc->for_reclaim) {
  		WARN_ON_ONCE(1);	/* Still happens? Tell us about it! */
  		goto redirty;
  	}
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
1337
1338
1339
1340
1341
  
  	/*
  	 * This is somewhat ridiculous, but without plumbing a SWAP_MAP_FALLOC
  	 * value into swapfile.c, the only way we can correctly account for a
  	 * fallocated page arriving here is now to initialize it and write it.
1aac14003   Hugh Dickins   tmpfs: quit when ...
1342
1343
1344
1345
1346
1347
  	 *
  	 * That's okay for a page already fallocated earlier, but if we have
  	 * not yet completed the fallocation, then (a) we want to keep track
  	 * of this page in case we have to undo it, and (b) it may not be a
  	 * good idea to continue anyway, once we're pushing into swap.  So
  	 * reactivate the page, and let shmem_fallocate() quit when too many.
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
1348
1349
  	 */
  	if (!PageUptodate(page)) {
1aac14003   Hugh Dickins   tmpfs: quit when ...
1350
1351
1352
1353
1354
  		if (inode->i_private) {
  			struct shmem_falloc *shmem_falloc;
  			spin_lock(&inode->i_lock);
  			shmem_falloc = inode->i_private;
  			if (shmem_falloc &&
8e205f779   Hugh Dickins   shmem: fix faulti...
1355
  			    !shmem_falloc->waitq &&
1aac14003   Hugh Dickins   tmpfs: quit when ...
1356
1357
1358
1359
1360
1361
1362
1363
1364
  			    index >= shmem_falloc->start &&
  			    index < shmem_falloc->next)
  				shmem_falloc->nr_unswapped++;
  			else
  				shmem_falloc = NULL;
  			spin_unlock(&inode->i_lock);
  			if (shmem_falloc)
  				goto redirty;
  		}
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
1365
1366
1367
1368
  		clear_highpage(page);
  		flush_dcache_page(page);
  		SetPageUptodate(page);
  	}
38d8b4e6b   Huang Ying   mm, THP, swap: de...
1369
  	swap = get_swap_page(page);
48f170fb7   Hugh Dickins   tmpfs: simplify u...
1370
1371
  	if (!swap.val)
  		goto redirty;
d9fe526a8   Hugh Dickins   tmpfs: allow file...
1372

b1dea800a   Hugh Dickins   tmpfs: fix race b...
1373
1374
  	/*
  	 * Add inode to shmem_unuse()'s list of swapped-out inodes,
6922c0c7a   Hugh Dickins   tmpfs: convert sh...
1375
1376
  	 * if it's not already there.  Do it now before the page is
  	 * moved to swap cache, when its pagelock no longer protects
b1dea800a   Hugh Dickins   tmpfs: fix race b...
1377
  	 * the inode from eviction.  But don't unlock the mutex until
6922c0c7a   Hugh Dickins   tmpfs: convert sh...
1378
1379
  	 * we've incremented swapped, because shmem_unuse_inode() will
  	 * prune a !swapped inode from the swaplist under this mutex.
b1dea800a   Hugh Dickins   tmpfs: fix race b...
1380
  	 */
48f170fb7   Hugh Dickins   tmpfs: simplify u...
1381
1382
  	mutex_lock(&shmem_swaplist_mutex);
  	if (list_empty(&info->swaplist))
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
1383
  		list_add(&info->swaplist, &shmem_swaplist);
b1dea800a   Hugh Dickins   tmpfs: fix race b...
1384

4afab1cd2   Yang Shi   mm: shmem: use pr...
1385
  	if (add_to_swap_cache(page, swap,
3852f6768   Joonsoo Kim   mm/swapcache: sup...
1386
1387
  			__GFP_HIGH | __GFP_NOMEMALLOC | __GFP_NOWARN,
  			NULL) == 0) {
4595ef88d   Kirill A. Shutemov   shmem: make shmem...
1388
  		spin_lock_irq(&info->lock);
6922c0c7a   Hugh Dickins   tmpfs: convert sh...
1389
  		shmem_recalc_inode(inode);
267a4c76b   Hugh Dickins   tmpfs: fix shmem_...
1390
  		info->swapped++;
4595ef88d   Kirill A. Shutemov   shmem: make shmem...
1391
  		spin_unlock_irq(&info->lock);
6922c0c7a   Hugh Dickins   tmpfs: convert sh...
1392

267a4c76b   Hugh Dickins   tmpfs: fix shmem_...
1393
1394
  		swap_shmem_alloc(swap);
  		shmem_delete_from_page_cache(page, swp_to_radix_entry(swap));
6922c0c7a   Hugh Dickins   tmpfs: convert sh...
1395
  		mutex_unlock(&shmem_swaplist_mutex);
d9fe526a8   Hugh Dickins   tmpfs: allow file...
1396
  		BUG_ON(page_mapped(page));
9fab5619b   Hugh Dickins   shmem: writepage ...
1397
  		swap_writepage(page, wbc);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1398
1399
  		return 0;
  	}
6922c0c7a   Hugh Dickins   tmpfs: convert sh...
1400
  	mutex_unlock(&shmem_swaplist_mutex);
75f6d6d29   Minchan Kim   mm, THP, swap: un...
1401
  	put_swap_page(page, swap);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1402
1403
  redirty:
  	set_page_dirty(page);
d9fe526a8   Hugh Dickins   tmpfs: allow file...
1404
1405
1406
1407
  	if (wbc->for_reclaim)
  		return AOP_WRITEPAGE_ACTIVATE;	/* Return with page locked */
  	unlock_page(page);
  	return 0;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1408
  }
75edd345e   Hugh Dickins   tmpfs: preliminar...
1409
  #if defined(CONFIG_NUMA) && defined(CONFIG_TMPFS)
71fe804b6   Lee Schermerhorn   mempolicy: use st...
1410
  static void shmem_show_mpol(struct seq_file *seq, struct mempolicy *mpol)
680d794ba   akpm@linux-foundation.org   mount options: fi...
1411
  {
095f1fc4e   Lee Schermerhorn   mempolicy: rework...
1412
  	char buffer[64];
680d794ba   akpm@linux-foundation.org   mount options: fi...
1413

71fe804b6   Lee Schermerhorn   mempolicy: use st...
1414
  	if (!mpol || mpol->mode == MPOL_DEFAULT)
095f1fc4e   Lee Schermerhorn   mempolicy: rework...
1415
  		return;		/* show nothing */
680d794ba   akpm@linux-foundation.org   mount options: fi...
1416

a7a88b237   Hugh Dickins   mempolicy: remove...
1417
  	mpol_to_str(buffer, sizeof(buffer), mpol);
095f1fc4e   Lee Schermerhorn   mempolicy: rework...
1418
1419
  
  	seq_printf(seq, ",mpol=%s", buffer);
680d794ba   akpm@linux-foundation.org   mount options: fi...
1420
  }
71fe804b6   Lee Schermerhorn   mempolicy: use st...
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
  
  static struct mempolicy *shmem_get_sbmpol(struct shmem_sb_info *sbinfo)
  {
  	struct mempolicy *mpol = NULL;
  	if (sbinfo->mpol) {
  		spin_lock(&sbinfo->stat_lock);	/* prevent replace/use races */
  		mpol = sbinfo->mpol;
  		mpol_get(mpol);
  		spin_unlock(&sbinfo->stat_lock);
  	}
  	return mpol;
  }
75edd345e   Hugh Dickins   tmpfs: preliminar...
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
  #else /* !CONFIG_NUMA || !CONFIG_TMPFS */
  static inline void shmem_show_mpol(struct seq_file *seq, struct mempolicy *mpol)
  {
  }
  static inline struct mempolicy *shmem_get_sbmpol(struct shmem_sb_info *sbinfo)
  {
  	return NULL;
  }
  #endif /* CONFIG_NUMA && CONFIG_TMPFS */
  #ifndef CONFIG_NUMA
  #define vm_policy vm_private_data
  #endif
680d794ba   akpm@linux-foundation.org   mount options: fi...
1445

800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1446
1447
1448
1449
  static void shmem_pseudo_vma_init(struct vm_area_struct *vma,
  		struct shmem_inode_info *info, pgoff_t index)
  {
  	/* Create a pseudo vma that just contains the policy */
2c4541e24   Kirill A. Shutemov   mm: use vma_init(...
1450
  	vma_init(vma, NULL);
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1451
1452
  	/* Bias interleave by inode number to distribute better across nodes */
  	vma->vm_pgoff = index + info->vfs_inode.i_ino;
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1453
1454
1455
1456
1457
1458
1459
1460
  	vma->vm_policy = mpol_shared_policy_lookup(&info->policy, index);
  }
  
  static void shmem_pseudo_vma_destroy(struct vm_area_struct *vma)
  {
  	/* Drop reference taken by mpol_shared_policy_lookup() */
  	mpol_cond_put(vma->vm_policy);
  }
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
1461
1462
  static struct page *shmem_swapin(swp_entry_t swap, gfp_t gfp,
  			struct shmem_inode_info *info, pgoff_t index)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1463
  {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1464
  	struct vm_area_struct pvma;
18a2f371f   Mel Gorman   tmpfs: fix shared...
1465
  	struct page *page;
e9e9b7ece   Minchan Kim   mm: swap: unify c...
1466
  	struct vm_fault vmf;
52cd3b074   Lee Schermerhorn   mempolicy: rework...
1467

800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1468
  	shmem_pseudo_vma_init(&pvma, info, index);
e9e9b7ece   Minchan Kim   mm: swap: unify c...
1469
1470
1471
  	vmf.vma = &pvma;
  	vmf.address = 0;
  	page = swap_cluster_readahead(swap, gfp, &vmf);
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1472
  	shmem_pseudo_vma_destroy(&pvma);
18a2f371f   Mel Gorman   tmpfs: fix shared...
1473

800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1474
1475
1476
1477
1478
1479
1480
  	return page;
  }
  
  static struct page *shmem_alloc_hugepage(gfp_t gfp,
  		struct shmem_inode_info *info, pgoff_t index)
  {
  	struct vm_area_struct pvma;
7b8d046fb   Matthew Wilcox   shmem: Convert sh...
1481
1482
  	struct address_space *mapping = info->vfs_inode.i_mapping;
  	pgoff_t hindex;
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1483
  	struct page *page;
4620a06e4   Geert Uytterhoeven   shmem: Fix link e...
1484
  	hindex = round_down(index, HPAGE_PMD_NR);
7b8d046fb   Matthew Wilcox   shmem: Convert sh...
1485
1486
  	if (xa_find(&mapping->i_pages, &hindex, hindex + HPAGE_PMD_NR - 1,
  								XA_PRESENT))
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1487
  		return NULL;
18a2f371f   Mel Gorman   tmpfs: fix shared...
1488

800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1489
1490
  	shmem_pseudo_vma_init(&pvma, info, hindex);
  	page = alloc_pages_vma(gfp | __GFP_COMP | __GFP_NORETRY | __GFP_NOWARN,
19deb7695   David Rientjes   Revert "Revert "R...
1491
  			HPAGE_PMD_ORDER, &pvma, 0, numa_node_id(), true);
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1492
1493
1494
  	shmem_pseudo_vma_destroy(&pvma);
  	if (page)
  		prep_transhuge_page(page);
dcdf11ee1   David Rientjes   mm, shmem: add vm...
1495
1496
  	else
  		count_vm_event(THP_FILE_FALLBACK);
18a2f371f   Mel Gorman   tmpfs: fix shared...
1497
  	return page;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1498
  }
02098feaa   Hugh Dickins   swapin needs gfp_...
1499
  static struct page *shmem_alloc_page(gfp_t gfp,
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
1500
  			struct shmem_inode_info *info, pgoff_t index)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1501
1502
  {
  	struct vm_area_struct pvma;
18a2f371f   Mel Gorman   tmpfs: fix shared...
1503
  	struct page *page;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1504

800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1505
1506
1507
1508
1509
1510
1511
1512
  	shmem_pseudo_vma_init(&pvma, info, index);
  	page = alloc_page_vma(gfp, &pvma, 0);
  	shmem_pseudo_vma_destroy(&pvma);
  
  	return page;
  }
  
  static struct page *shmem_alloc_and_acct_page(gfp_t gfp,
0f0796945   Mike Rapoport   shmem: introduce ...
1513
  		struct inode *inode,
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1514
1515
  		pgoff_t index, bool huge)
  {
0f0796945   Mike Rapoport   shmem: introduce ...
1516
  	struct shmem_inode_info *info = SHMEM_I(inode);
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1517
1518
1519
  	struct page *page;
  	int nr;
  	int err = -ENOSPC;
52cd3b074   Lee Schermerhorn   mempolicy: rework...
1520

396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
1521
  	if (!IS_ENABLED(CONFIG_TRANSPARENT_HUGEPAGE))
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1522
1523
  		huge = false;
  	nr = huge ? HPAGE_PMD_NR : 1;
0f0796945   Mike Rapoport   shmem: introduce ...
1524
  	if (!shmem_inode_acct_block(inode, nr))
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1525
  		goto failed;
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1526
1527
1528
1529
1530
  
  	if (huge)
  		page = shmem_alloc_hugepage(gfp, info, index);
  	else
  		page = shmem_alloc_page(gfp, info, index);
75edd345e   Hugh Dickins   tmpfs: preliminar...
1531
1532
1533
  	if (page) {
  		__SetPageLocked(page);
  		__SetPageSwapBacked(page);
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1534
  		return page;
75edd345e   Hugh Dickins   tmpfs: preliminar...
1535
  	}
18a2f371f   Mel Gorman   tmpfs: fix shared...
1536

800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1537
  	err = -ENOMEM;
0f0796945   Mike Rapoport   shmem: introduce ...
1538
  	shmem_inode_unacct_blocks(inode, nr);
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1539
1540
  failed:
  	return ERR_PTR(err);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1541
  }
71fe804b6   Lee Schermerhorn   mempolicy: use st...
1542

1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1543
  /*
bde05d1cc   Hugh Dickins   shmem: replace pa...
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
   * When a page is moved from swapcache to shmem filecache (either by the
   * usual swapin of shmem_getpage_gfp(), or by the less common swapoff of
   * shmem_unuse_inode()), it may have been read in earlier from swap, in
   * ignorance of the mapping it belongs to.  If that mapping has special
   * constraints (like the gma500 GEM driver, which requires RAM below 4GB),
   * we may need to copy to a suitable page before moving to filecache.
   *
   * In a future release, this may well be extended to respect cpuset and
   * NUMA mempolicy, and applied also to anonymous pages in do_swap_page();
   * but for now it is a simple matter of zone.
   */
  static bool shmem_should_replace_page(struct page *page, gfp_t gfp)
  {
  	return page_zonenum(page) > gfp_zone(gfp);
  }
  
  static int shmem_replace_page(struct page **pagep, gfp_t gfp,
  				struct shmem_inode_info *info, pgoff_t index)
  {
  	struct page *oldpage, *newpage;
  	struct address_space *swap_mapping;
c1cb20d43   Yu Zhao   mm: use swp_offse...
1565
  	swp_entry_t entry;
bde05d1cc   Hugh Dickins   shmem: replace pa...
1566
1567
1568
1569
  	pgoff_t swap_index;
  	int error;
  
  	oldpage = *pagep;
c1cb20d43   Yu Zhao   mm: use swp_offse...
1570
1571
  	entry.val = page_private(oldpage);
  	swap_index = swp_offset(entry);
bde05d1cc   Hugh Dickins   shmem: replace pa...
1572
1573
1574
1575
1576
1577
1578
1579
1580
1581
  	swap_mapping = page_mapping(oldpage);
  
  	/*
  	 * We have arrived here because our zones are constrained, so don't
  	 * limit chance of success by further cpuset and node constraints.
  	 */
  	gfp &= ~GFP_CONSTRAINT_MASK;
  	newpage = shmem_alloc_page(gfp, info, index);
  	if (!newpage)
  		return -ENOMEM;
bde05d1cc   Hugh Dickins   shmem: replace pa...
1582

09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
1583
  	get_page(newpage);
bde05d1cc   Hugh Dickins   shmem: replace pa...
1584
  	copy_highpage(newpage, oldpage);
0142ef6cd   Hugh Dickins   shmem: replace_pa...
1585
  	flush_dcache_page(newpage);
bde05d1cc   Hugh Dickins   shmem: replace pa...
1586

9956edf37   Hugh Dickins   shmem: fix pagefl...
1587
1588
  	__SetPageLocked(newpage);
  	__SetPageSwapBacked(newpage);
bde05d1cc   Hugh Dickins   shmem: replace pa...
1589
  	SetPageUptodate(newpage);
c1cb20d43   Yu Zhao   mm: use swp_offse...
1590
  	set_page_private(newpage, entry.val);
bde05d1cc   Hugh Dickins   shmem: replace pa...
1591
1592
1593
1594
1595
1596
  	SetPageSwapCache(newpage);
  
  	/*
  	 * Our caller will very soon move newpage out of swapcache, but it's
  	 * a nice clean interface for us to replace oldpage by newpage there.
  	 */
b93b01631   Matthew Wilcox   page cache: use x...
1597
  	xa_lock_irq(&swap_mapping->i_pages);
62f945b6a   Matthew Wilcox   shmem: Convert sh...
1598
  	error = shmem_replace_entry(swap_mapping, swap_index, oldpage, newpage);
0142ef6cd   Hugh Dickins   shmem: replace_pa...
1599
  	if (!error) {
0d1c20722   Johannes Weiner   mm: memcontrol: s...
1600
1601
1602
  		mem_cgroup_migrate(oldpage, newpage);
  		__inc_lruvec_page_state(newpage, NR_FILE_PAGES);
  		__dec_lruvec_page_state(oldpage, NR_FILE_PAGES);
0142ef6cd   Hugh Dickins   shmem: replace_pa...
1603
  	}
b93b01631   Matthew Wilcox   page cache: use x...
1604
  	xa_unlock_irq(&swap_mapping->i_pages);
bde05d1cc   Hugh Dickins   shmem: replace pa...
1605

0142ef6cd   Hugh Dickins   shmem: replace_pa...
1606
1607
1608
1609
1610
1611
1612
1613
  	if (unlikely(error)) {
  		/*
  		 * Is this possible?  I think not, now that our callers check
  		 * both PageSwapCache and page_private after getting page lock;
  		 * but be defensive.  Reverse old to newpage for clear and free.
  		 */
  		oldpage = newpage;
  	} else {
6058eaec8   Johannes Weiner   mm: fold and remo...
1614
  		lru_cache_add(newpage);
0142ef6cd   Hugh Dickins   shmem: replace_pa...
1615
1616
  		*pagep = newpage;
  	}
bde05d1cc   Hugh Dickins   shmem: replace pa...
1617
1618
1619
1620
1621
  
  	ClearPageSwapCache(oldpage);
  	set_page_private(oldpage, 0);
  
  	unlock_page(oldpage);
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
1622
1623
  	put_page(oldpage);
  	put_page(oldpage);
0142ef6cd   Hugh Dickins   shmem: replace_pa...
1624
  	return error;
bde05d1cc   Hugh Dickins   shmem: replace pa...
1625
1626
1627
  }
  
  /*
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1628
1629
1630
   * Swap in the page pointed to by *pagep.
   * Caller has to make sure that *pagep contains a valid swapped page.
   * Returns 0 and the page in pagep if success. On failure, returns the
af44c12fe   Randy Dunlap   mm/shmem.c: delet...
1631
   * error code and NULL in *pagep.
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1632
1633
1634
1635
1636
1637
1638
1639
1640
   */
  static int shmem_swapin_page(struct inode *inode, pgoff_t index,
  			     struct page **pagep, enum sgp_type sgp,
  			     gfp_t gfp, struct vm_area_struct *vma,
  			     vm_fault_t *fault_type)
  {
  	struct address_space *mapping = inode->i_mapping;
  	struct shmem_inode_info *info = SHMEM_I(inode);
  	struct mm_struct *charge_mm = vma ? vma->vm_mm : current->mm;
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1641
1642
1643
1644
1645
1646
1647
1648
1649
1650
1651
1652
1653
1654
1655
1656
1657
1658
1659
1660
1661
1662
1663
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674
1675
1676
1677
  	struct page *page;
  	swp_entry_t swap;
  	int error;
  
  	VM_BUG_ON(!*pagep || !xa_is_value(*pagep));
  	swap = radix_to_swp_entry(*pagep);
  	*pagep = NULL;
  
  	/* Look it up and read it in.. */
  	page = lookup_swap_cache(swap, NULL, 0);
  	if (!page) {
  		/* Or update major stats only when swapin succeeds?? */
  		if (fault_type) {
  			*fault_type |= VM_FAULT_MAJOR;
  			count_vm_event(PGMAJFAULT);
  			count_memcg_event_mm(charge_mm, PGMAJFAULT);
  		}
  		/* Here we actually start the io */
  		page = shmem_swapin(swap, gfp, info, index);
  		if (!page) {
  			error = -ENOMEM;
  			goto failed;
  		}
  	}
  
  	/* We have to do this with page locked to prevent races */
  	lock_page(page);
  	if (!PageSwapCache(page) || page_private(page) != swap.val ||
  	    !shmem_confirm_swap(mapping, index, swap)) {
  		error = -EEXIST;
  		goto unlock;
  	}
  	if (!PageUptodate(page)) {
  		error = -EIO;
  		goto failed;
  	}
  	wait_on_page_writeback(page);
8a84802e2   Steven Price   mm: Add arch hook...
1678
1679
1680
1681
1682
  	/*
  	 * Some architectures may have to restore extra metadata to the
  	 * physical page after reading from swap.
  	 */
  	arch_swap_restore(swap, page);
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1683
1684
1685
1686
1687
  	if (shmem_should_replace_page(page, gfp)) {
  		error = shmem_replace_page(&page, gfp, info, index);
  		if (error)
  			goto failed;
  	}
14235ab36   Johannes Weiner   mm: shmem: remove...
1688
  	error = shmem_add_to_page_cache(page, mapping, index,
3fea5a499   Johannes Weiner   mm: memcontrol: c...
1689
1690
  					swp_to_radix_entry(swap), gfp,
  					charge_mm);
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1691
1692
  	if (error)
  		goto failed;
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1693
1694
1695
1696
1697
1698
1699
1700
1701
1702
1703
1704
1705
1706
1707
1708
1709
1710
1711
1712
1713
1714
1715
1716
1717
1718
1719
  	spin_lock_irq(&info->lock);
  	info->swapped--;
  	shmem_recalc_inode(inode);
  	spin_unlock_irq(&info->lock);
  
  	if (sgp == SGP_WRITE)
  		mark_page_accessed(page);
  
  	delete_from_swap_cache(page);
  	set_page_dirty(page);
  	swap_free(swap);
  
  	*pagep = page;
  	return 0;
  failed:
  	if (!shmem_confirm_swap(mapping, index, swap))
  		error = -EEXIST;
  unlock:
  	if (page) {
  		unlock_page(page);
  		put_page(page);
  	}
  
  	return error;
  }
  
  /*
68da9f055   Hugh Dickins   tmpfs: pass gfp t...
1720
   * shmem_getpage_gfp - find page in cache, or get from swap, or allocate
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1721
1722
1723
   *
   * If we allocate a new one we do not mark it dirty. That's up to the
   * vm. If we swap it in we mark it dirty since we also free the swap
9e18eb293   Andres Lagar-Cavilla   tmpfs: mem_cgroup...
1724
1725
   * entry since a page cannot live in both the swap and page cache.
   *
28eb3c808   Miles Chen   shmem: fix obsole...
1726
   * vmf and fault_type are only supplied by shmem_fault:
9e18eb293   Andres Lagar-Cavilla   tmpfs: mem_cgroup...
1727
   * otherwise they are NULL.
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1728
   */
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
1729
  static int shmem_getpage_gfp(struct inode *inode, pgoff_t index,
9e18eb293   Andres Lagar-Cavilla   tmpfs: mem_cgroup...
1730
  	struct page **pagep, enum sgp_type sgp, gfp_t gfp,
2b7403035   Souptick Joarder   mm: Change return...
1731
1732
  	struct vm_area_struct *vma, struct vm_fault *vmf,
  			vm_fault_t *fault_type)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1733
1734
  {
  	struct address_space *mapping = inode->i_mapping;
23f919d4a   Arnd Bergmann   shmem: avoid mayb...
1735
  	struct shmem_inode_info *info = SHMEM_I(inode);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1736
  	struct shmem_sb_info *sbinfo;
9e18eb293   Andres Lagar-Cavilla   tmpfs: mem_cgroup...
1737
  	struct mm_struct *charge_mm;
27ab70062   Hugh Dickins   tmpfs: simplify f...
1738
  	struct page *page;
657e3038c   Kirill A. Shutemov   shmem, thp: respe...
1739
  	enum sgp_type sgp_huge = sgp;
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1740
  	pgoff_t hindex = index;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1741
  	int error;
54af60421   Hugh Dickins   tmpfs: convert sh...
1742
  	int once = 0;
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
1743
  	int alloced = 0;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1744

09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
1745
  	if (index > (MAX_LFS_FILESIZE >> PAGE_SHIFT))
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1746
  		return -EFBIG;
657e3038c   Kirill A. Shutemov   shmem, thp: respe...
1747
1748
  	if (sgp == SGP_NOHUGE || sgp == SGP_HUGE)
  		sgp = SGP_CACHE;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1749
  repeat:
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1750
1751
1752
1753
1754
1755
1756
  	if (sgp <= SGP_CACHE &&
  	    ((loff_t)index << PAGE_SHIFT) >= i_size_read(inode)) {
  		return -EINVAL;
  	}
  
  	sbinfo = SHMEM_SB(inode->i_sb);
  	charge_mm = vma ? vma->vm_mm : current->mm;
0cd6144aa   Johannes Weiner   mm + fs: prepare ...
1757
  	page = find_lock_entry(mapping, index);
3159f943a   Matthew Wilcox   xarray: Replace e...
1758
  	if (xa_is_value(page)) {
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1759
1760
1761
1762
  		error = shmem_swapin_page(inode, index, &page,
  					  sgp, gfp, vma, fault_type);
  		if (error == -EEXIST)
  			goto repeat;
54af60421   Hugh Dickins   tmpfs: convert sh...
1763

c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1764
1765
  		*pagep = page;
  		return error;
54af60421   Hugh Dickins   tmpfs: convert sh...
1766
  	}
63ec1973d   Matthew Wilcox (Oracle)   mm/shmem: return ...
1767
1768
  	if (page)
  		hindex = page->index;
66d2f4d28   Hugh Dickins   shmem: fix init_p...
1769
1770
  	if (page && sgp == SGP_WRITE)
  		mark_page_accessed(page);
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
1771
1772
1773
1774
1775
  	/* fallocated page? */
  	if (page && !PageUptodate(page)) {
  		if (sgp != SGP_READ)
  			goto clear;
  		unlock_page(page);
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
1776
  		put_page(page);
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
1777
  		page = NULL;
63ec1973d   Matthew Wilcox (Oracle)   mm/shmem: return ...
1778
  		hindex = index;
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
1779
  	}
63ec1973d   Matthew Wilcox (Oracle)   mm/shmem: return ...
1780
1781
  	if (page || sgp == SGP_READ)
  		goto out;
27ab70062   Hugh Dickins   tmpfs: simplify f...
1782
1783
  
  	/*
54af60421   Hugh Dickins   tmpfs: convert sh...
1784
1785
  	 * Fast cache lookup did not find it:
  	 * bring it back from swap or allocate.
27ab70062   Hugh Dickins   tmpfs: simplify f...
1786
  	 */
54af60421   Hugh Dickins   tmpfs: convert sh...
1787

c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1788
1789
1790
1791
  	if (vma && userfaultfd_missing(vma)) {
  		*fault_type = handle_userfault(vmf, VM_UFFD_MISSING);
  		return 0;
  	}
cfda05267   Mike Rapoport   userfaultfd: shme...
1792

c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1793
1794
1795
1796
1797
1798
1799
1800
  	/* shmem_symlink() */
  	if (mapping->a_ops != &shmem_aops)
  		goto alloc_nohuge;
  	if (shmem_huge == SHMEM_HUGE_DENY || sgp_huge == SGP_NOHUGE)
  		goto alloc_nohuge;
  	if (shmem_huge == SHMEM_HUGE_FORCE)
  		goto alloc_huge;
  	switch (sbinfo->huge) {
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1801
1802
  	case SHMEM_HUGE_NEVER:
  		goto alloc_nohuge;
27d80fa24   Kees Cook   mm/shmem.c: distr...
1803
1804
1805
  	case SHMEM_HUGE_WITHIN_SIZE: {
  		loff_t i_size;
  		pgoff_t off;
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1806
1807
1808
1809
  		off = round_up(index, HPAGE_PMD_NR);
  		i_size = round_up(i_size_read(inode), PAGE_SIZE);
  		if (i_size >= HPAGE_PMD_SIZE &&
  		    i_size >> PAGE_SHIFT >= off)
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1810
  			goto alloc_huge;
27d80fa24   Kees Cook   mm/shmem.c: distr...
1811
1812
1813
  
  		fallthrough;
  	}
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1814
1815
1816
1817
1818
1819
  	case SHMEM_HUGE_ADVISE:
  		if (sgp_huge == SGP_HUGE)
  			goto alloc_huge;
  		/* TODO: implement fadvise() hints */
  		goto alloc_nohuge;
  	}
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1820

800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1821
  alloc_huge:
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1822
1823
1824
1825
1826
1827
1828
1829
  	page = shmem_alloc_and_acct_page(gfp, inode, index, true);
  	if (IS_ERR(page)) {
  alloc_nohuge:
  		page = shmem_alloc_and_acct_page(gfp, inode,
  						 index, false);
  	}
  	if (IS_ERR(page)) {
  		int retry = 5;
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1830

c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1831
1832
1833
1834
1835
1836
1837
1838
1839
1840
  		error = PTR_ERR(page);
  		page = NULL;
  		if (error != -ENOSPC)
  			goto unlock;
  		/*
  		 * Try to reclaim some space by splitting a huge page
  		 * beyond i_size on the filesystem.
  		 */
  		while (retry--) {
  			int ret;
66d2f4d28   Hugh Dickins   shmem: fix init_p...
1841

c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1842
1843
1844
1845
1846
  			ret = shmem_unused_huge_shrink(sbinfo, NULL, 1);
  			if (ret == SHRINK_STOP)
  				break;
  			if (ret)
  				goto alloc_nohuge;
b065b4321   Hugh Dickins   shmem: cleanup sh...
1847
  		}
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1848
1849
  		goto unlock;
  	}
54af60421   Hugh Dickins   tmpfs: convert sh...
1850

c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1851
1852
1853
1854
  	if (PageTransHuge(page))
  		hindex = round_down(index, HPAGE_PMD_NR);
  	else
  		hindex = index;
54af60421   Hugh Dickins   tmpfs: convert sh...
1855

c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1856
1857
  	if (sgp == SGP_WRITE)
  		__SetPageReferenced(page);
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1858
  	error = shmem_add_to_page_cache(page, mapping, hindex,
3fea5a499   Johannes Weiner   mm: memcontrol: c...
1859
1860
1861
  					NULL, gfp & GFP_RECLAIM_MASK,
  					charge_mm);
  	if (error)
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1862
  		goto unacct;
6058eaec8   Johannes Weiner   mm: fold and remo...
1863
  	lru_cache_add(page);
779750d20   Kirill A. Shutemov   shmem: split huge...
1864

c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1865
  	spin_lock_irq(&info->lock);
d8c6546b1   Matthew Wilcox (Oracle)   mm: introduce com...
1866
  	info->alloced += compound_nr(page);
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1867
1868
1869
1870
1871
1872
1873
1874
  	inode->i_blocks += BLOCKS_PER_PAGE << compound_order(page);
  	shmem_recalc_inode(inode);
  	spin_unlock_irq(&info->lock);
  	alloced = true;
  
  	if (PageTransHuge(page) &&
  	    DIV_ROUND_UP(i_size_read(inode), PAGE_SIZE) <
  			hindex + HPAGE_PMD_NR - 1) {
ec9516fbc   Hugh Dickins   tmpfs: optimize c...
1875
  		/*
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1876
1877
  		 * Part of the huge page is beyond i_size: subject
  		 * to shrink under memory pressure.
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
1878
  		 */
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1879
  		spin_lock(&sbinfo->shrinklist_lock);
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
1880
  		/*
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1881
1882
  		 * _careful to defend against unlocked access to
  		 * ->shrink_list in shmem_unused_huge_shrink()
ec9516fbc   Hugh Dickins   tmpfs: optimize c...
1883
  		 */
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1884
1885
1886
1887
1888
1889
1890
  		if (list_empty_careful(&info->shrinklist)) {
  			list_add_tail(&info->shrinklist,
  				      &sbinfo->shrinklist);
  			sbinfo->shrinklist_len++;
  		}
  		spin_unlock(&sbinfo->shrinklist_lock);
  	}
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1891

c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1892
1893
1894
1895
1896
1897
1898
1899
1900
1901
1902
1903
  	/*
  	 * Let SGP_FALLOC use the SGP_WRITE optimization on a new page.
  	 */
  	if (sgp == SGP_FALLOC)
  		sgp = SGP_WRITE;
  clear:
  	/*
  	 * Let SGP_WRITE caller clear ends if write does not fill page;
  	 * but SGP_FALLOC on a page fallocated earlier must initialize
  	 * it now, lest undo on failure cancel our earlier guarantee.
  	 */
  	if (sgp != SGP_WRITE && !PageUptodate(page)) {
c5bf121e4   Vineeth Remanan Pillai   mm: refactor swap...
1904
  		int i;
63ec1973d   Matthew Wilcox (Oracle)   mm/shmem: return ...
1905
1906
1907
  		for (i = 0; i < compound_nr(page); i++) {
  			clear_highpage(page + i);
  			flush_dcache_page(page + i);
ec9516fbc   Hugh Dickins   tmpfs: optimize c...
1908
  		}
63ec1973d   Matthew Wilcox (Oracle)   mm/shmem: return ...
1909
  		SetPageUptodate(page);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1910
  	}
bde05d1cc   Hugh Dickins   shmem: replace pa...
1911

54af60421   Hugh Dickins   tmpfs: convert sh...
1912
  	/* Perhaps the file has been truncated since we checked */
75edd345e   Hugh Dickins   tmpfs: preliminar...
1913
  	if (sgp <= SGP_CACHE &&
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
1914
  	    ((loff_t)index << PAGE_SHIFT) >= i_size_read(inode)) {
267a4c76b   Hugh Dickins   tmpfs: fix shmem_...
1915
1916
1917
  		if (alloced) {
  			ClearPageDirty(page);
  			delete_from_page_cache(page);
4595ef88d   Kirill A. Shutemov   shmem: make shmem...
1918
  			spin_lock_irq(&info->lock);
267a4c76b   Hugh Dickins   tmpfs: fix shmem_...
1919
  			shmem_recalc_inode(inode);
4595ef88d   Kirill A. Shutemov   shmem: make shmem...
1920
  			spin_unlock_irq(&info->lock);
267a4c76b   Hugh Dickins   tmpfs: fix shmem_...
1921
  		}
54af60421   Hugh Dickins   tmpfs: convert sh...
1922
  		error = -EINVAL;
267a4c76b   Hugh Dickins   tmpfs: fix shmem_...
1923
  		goto unlock;
e83c32e8f   Hugh Dickins   tmpfs: simplify p...
1924
  	}
63ec1973d   Matthew Wilcox (Oracle)   mm/shmem: return ...
1925
  out:
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1926
  	*pagep = page + index - hindex;
54af60421   Hugh Dickins   tmpfs: convert sh...
1927
  	return 0;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1928

59a16ead5   Hugh Dickins   tmpfs: fix spurio...
1929
  	/*
54af60421   Hugh Dickins   tmpfs: convert sh...
1930
  	 * Error recovery.
59a16ead5   Hugh Dickins   tmpfs: fix spurio...
1931
  	 */
54af60421   Hugh Dickins   tmpfs: convert sh...
1932
  unacct:
d8c6546b1   Matthew Wilcox (Oracle)   mm: introduce com...
1933
  	shmem_inode_unacct_blocks(inode, compound_nr(page));
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
1934
1935
1936
1937
1938
1939
  
  	if (PageTransHuge(page)) {
  		unlock_page(page);
  		put_page(page);
  		goto alloc_nohuge;
  	}
d18992286   Hugh Dickins   shmem: fix negati...
1940
  unlock:
27ab70062   Hugh Dickins   tmpfs: simplify f...
1941
  	if (page) {
54af60421   Hugh Dickins   tmpfs: convert sh...
1942
  		unlock_page(page);
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
1943
  		put_page(page);
54af60421   Hugh Dickins   tmpfs: convert sh...
1944
1945
  	}
  	if (error == -ENOSPC && !once++) {
4595ef88d   Kirill A. Shutemov   shmem: make shmem...
1946
  		spin_lock_irq(&info->lock);
54af60421   Hugh Dickins   tmpfs: convert sh...
1947
  		shmem_recalc_inode(inode);
4595ef88d   Kirill A. Shutemov   shmem: make shmem...
1948
  		spin_unlock_irq(&info->lock);
27ab70062   Hugh Dickins   tmpfs: simplify f...
1949
  		goto repeat;
ff36b8016   Shaohua Li   shmem: reduce pag...
1950
  	}
7f4446eef   Matthew Wilcox   shmem: Comment fi...
1951
  	if (error == -EEXIST)
54af60421   Hugh Dickins   tmpfs: convert sh...
1952
1953
  		goto repeat;
  	return error;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1954
  }
10d20bd25   Linus Torvalds   shmem: fix shm fa...
1955
1956
1957
1958
1959
  /*
   * This is like autoremove_wake_function, but it removes the wait queue
   * entry unconditionally - even if something else had already woken the
   * target.
   */
ac6424b98   Ingo Molnar   sched/wait: Renam...
1960
  static int synchronous_wake_function(wait_queue_entry_t *wait, unsigned mode, int sync, void *key)
10d20bd25   Linus Torvalds   shmem: fix shm fa...
1961
1962
  {
  	int ret = default_wake_function(wait, mode, sync, key);
2055da973   Ingo Molnar   sched/wait: Disam...
1963
  	list_del_init(&wait->entry);
10d20bd25   Linus Torvalds   shmem: fix shm fa...
1964
1965
  	return ret;
  }
20acce679   Souptick Joarder   mm/shmem.c: use n...
1966
  static vm_fault_t shmem_fault(struct vm_fault *vmf)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1967
  {
11bac8000   Dave Jiang   mm, fs: reduce fa...
1968
  	struct vm_area_struct *vma = vmf->vma;
496ad9aa8   Al Viro   new helper: file_...
1969
  	struct inode *inode = file_inode(vma->vm_file);
9e18eb293   Andres Lagar-Cavilla   tmpfs: mem_cgroup...
1970
  	gfp_t gfp = mapping_gfp_mask(inode->i_mapping);
657e3038c   Kirill A. Shutemov   shmem, thp: respe...
1971
  	enum sgp_type sgp;
20acce679   Souptick Joarder   mm/shmem.c: use n...
1972
1973
  	int err;
  	vm_fault_t ret = VM_FAULT_LOCKED;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
1974

f00cdc6df   Hugh Dickins   shmem: fix faulti...
1975
1976
1977
1978
  	/*
  	 * Trinity finds that probing a hole which tmpfs is punching can
  	 * prevent the hole-punch from ever completing: which in turn
  	 * locks writers out with its hold on i_mutex.  So refrain from
8e205f779   Hugh Dickins   shmem: fix faulti...
1979
1980
1981
1982
1983
1984
1985
1986
1987
1988
1989
1990
  	 * faulting pages into the hole while it's being punched.  Although
  	 * shmem_undo_range() does remove the additions, it may be unable to
  	 * keep up, as each new page needs its own unmap_mapping_range() call,
  	 * and the i_mmap tree grows ever slower to scan if new vmas are added.
  	 *
  	 * It does not matter if we sometimes reach this check just before the
  	 * hole-punch begins, so that one fault then races with the punch:
  	 * we just need to make racing faults a rare case.
  	 *
  	 * The implementation below would be much simpler if we just used a
  	 * standard mutex or completion: but we cannot take i_mutex in fault,
  	 * and bloating every shmem inode for this unlikely case would be sad.
f00cdc6df   Hugh Dickins   shmem: fix faulti...
1991
1992
1993
1994
1995
1996
  	 */
  	if (unlikely(inode->i_private)) {
  		struct shmem_falloc *shmem_falloc;
  
  		spin_lock(&inode->i_lock);
  		shmem_falloc = inode->i_private;
8e205f779   Hugh Dickins   shmem: fix faulti...
1997
1998
1999
2000
  		if (shmem_falloc &&
  		    shmem_falloc->waitq &&
  		    vmf->pgoff >= shmem_falloc->start &&
  		    vmf->pgoff < shmem_falloc->next) {
8897c1b1a   Kirill A. Shutemov   shmem: pin the fi...
2001
  			struct file *fpin;
8e205f779   Hugh Dickins   shmem: fix faulti...
2002
  			wait_queue_head_t *shmem_falloc_waitq;
10d20bd25   Linus Torvalds   shmem: fix shm fa...
2003
  			DEFINE_WAIT_FUNC(shmem_fault_wait, synchronous_wake_function);
8e205f779   Hugh Dickins   shmem: fix faulti...
2004
2005
  
  			ret = VM_FAULT_NOPAGE;
8897c1b1a   Kirill A. Shutemov   shmem: pin the fi...
2006
2007
  			fpin = maybe_unlock_mmap_for_io(vmf, NULL);
  			if (fpin)
8e205f779   Hugh Dickins   shmem: fix faulti...
2008
  				ret = VM_FAULT_RETRY;
8e205f779   Hugh Dickins   shmem: fix faulti...
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
  
  			shmem_falloc_waitq = shmem_falloc->waitq;
  			prepare_to_wait(shmem_falloc_waitq, &shmem_fault_wait,
  					TASK_UNINTERRUPTIBLE);
  			spin_unlock(&inode->i_lock);
  			schedule();
  
  			/*
  			 * shmem_falloc_waitq points into the shmem_fallocate()
  			 * stack of the hole-punching task: shmem_falloc_waitq
  			 * is usually invalid by the time we reach here, but
  			 * finish_wait() does not dereference it in that case;
  			 * though i_lock needed lest racing with wake_up_all().
  			 */
  			spin_lock(&inode->i_lock);
  			finish_wait(shmem_falloc_waitq, &shmem_fault_wait);
  			spin_unlock(&inode->i_lock);
8897c1b1a   Kirill A. Shutemov   shmem: pin the fi...
2026
2027
2028
  
  			if (fpin)
  				fput(fpin);
8e205f779   Hugh Dickins   shmem: fix faulti...
2029
  			return ret;
f00cdc6df   Hugh Dickins   shmem: fix faulti...
2030
  		}
8e205f779   Hugh Dickins   shmem: fix faulti...
2031
  		spin_unlock(&inode->i_lock);
f00cdc6df   Hugh Dickins   shmem: fix faulti...
2032
  	}
657e3038c   Kirill A. Shutemov   shmem, thp: respe...
2033
  	sgp = SGP_CACHE;
186003323   Michal Hocko   mm: make PR_SET_T...
2034
2035
2036
  
  	if ((vma->vm_flags & VM_NOHUGEPAGE) ||
  	    test_bit(MMF_DISABLE_THP, &vma->vm_mm->flags))
657e3038c   Kirill A. Shutemov   shmem, thp: respe...
2037
  		sgp = SGP_NOHUGE;
186003323   Michal Hocko   mm: make PR_SET_T...
2038
2039
  	else if (vma->vm_flags & VM_HUGEPAGE)
  		sgp = SGP_HUGE;
657e3038c   Kirill A. Shutemov   shmem, thp: respe...
2040

20acce679   Souptick Joarder   mm/shmem.c: use n...
2041
  	err = shmem_getpage_gfp(inode, vmf->pgoff, &vmf->page, sgp,
cfda05267   Mike Rapoport   userfaultfd: shme...
2042
  				  gfp, vma, vmf, &ret);
20acce679   Souptick Joarder   mm/shmem.c: use n...
2043
2044
  	if (err)
  		return vmf_error(err);
68da9f055   Hugh Dickins   tmpfs: pass gfp t...
2045
  	return ret;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2046
  }
c01d5b300   Hugh Dickins   shmem: get_unmapp...
2047
2048
2049
2050
2051
2052
2053
2054
2055
2056
2057
2058
2059
2060
2061
2062
2063
  unsigned long shmem_get_unmapped_area(struct file *file,
  				      unsigned long uaddr, unsigned long len,
  				      unsigned long pgoff, unsigned long flags)
  {
  	unsigned long (*get_area)(struct file *,
  		unsigned long, unsigned long, unsigned long, unsigned long);
  	unsigned long addr;
  	unsigned long offset;
  	unsigned long inflated_len;
  	unsigned long inflated_addr;
  	unsigned long inflated_offset;
  
  	if (len > TASK_SIZE)
  		return -ENOMEM;
  
  	get_area = current->mm->get_unmapped_area;
  	addr = get_area(file, uaddr, len, pgoff, flags);
396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
2064
  	if (!IS_ENABLED(CONFIG_TRANSPARENT_HUGEPAGE))
c01d5b300   Hugh Dickins   shmem: get_unmapp...
2065
2066
2067
2068
2069
2070
2071
2072
2073
2074
2075
2076
2077
2078
2079
2080
2081
  		return addr;
  	if (IS_ERR_VALUE(addr))
  		return addr;
  	if (addr & ~PAGE_MASK)
  		return addr;
  	if (addr > TASK_SIZE - len)
  		return addr;
  
  	if (shmem_huge == SHMEM_HUGE_DENY)
  		return addr;
  	if (len < HPAGE_PMD_SIZE)
  		return addr;
  	if (flags & MAP_FIXED)
  		return addr;
  	/*
  	 * Our priority is to support MAP_SHARED mapped hugely;
  	 * and support MAP_PRIVATE mapped hugely too, until it is COWed.
991589974   Kirill A. Shutemov   mm/shmem.c: thp, ...
2082
2083
  	 * But if caller specified an address hint and we allocated area there
  	 * successfully, respect that as before.
c01d5b300   Hugh Dickins   shmem: get_unmapp...
2084
  	 */
991589974   Kirill A. Shutemov   mm/shmem.c: thp, ...
2085
  	if (uaddr == addr)
c01d5b300   Hugh Dickins   shmem: get_unmapp...
2086
2087
2088
2089
2090
2091
2092
2093
2094
2095
2096
2097
2098
2099
2100
2101
2102
  		return addr;
  
  	if (shmem_huge != SHMEM_HUGE_FORCE) {
  		struct super_block *sb;
  
  		if (file) {
  			VM_BUG_ON(file->f_op != &shmem_file_operations);
  			sb = file_inode(file)->i_sb;
  		} else {
  			/*
  			 * Called directly from mm/mmap.c, or drivers/char/mem.c
  			 * for "/dev/zero", to create a shared anonymous object.
  			 */
  			if (IS_ERR(shm_mnt))
  				return addr;
  			sb = shm_mnt->mnt_sb;
  		}
3089bf614   Toshi Kani   shmem: fix tmpfs ...
2103
  		if (SHMEM_SB(sb)->huge == SHMEM_HUGE_NEVER)
c01d5b300   Hugh Dickins   shmem: get_unmapp...
2104
2105
2106
2107
2108
2109
2110
2111
2112
2113
2114
2115
2116
2117
  			return addr;
  	}
  
  	offset = (pgoff << PAGE_SHIFT) & (HPAGE_PMD_SIZE-1);
  	if (offset && offset + len < 2 * HPAGE_PMD_SIZE)
  		return addr;
  	if ((addr & (HPAGE_PMD_SIZE-1)) == offset)
  		return addr;
  
  	inflated_len = len + HPAGE_PMD_SIZE - PAGE_SIZE;
  	if (inflated_len > TASK_SIZE)
  		return addr;
  	if (inflated_len < len)
  		return addr;
991589974   Kirill A. Shutemov   mm/shmem.c: thp, ...
2118
  	inflated_addr = get_area(NULL, uaddr, inflated_len, 0, flags);
c01d5b300   Hugh Dickins   shmem: get_unmapp...
2119
2120
2121
2122
2123
2124
2125
2126
2127
2128
2129
2130
2131
2132
  	if (IS_ERR_VALUE(inflated_addr))
  		return addr;
  	if (inflated_addr & ~PAGE_MASK)
  		return addr;
  
  	inflated_offset = inflated_addr & (HPAGE_PMD_SIZE-1);
  	inflated_addr += offset - inflated_offset;
  	if (inflated_offset > offset)
  		inflated_addr += HPAGE_PMD_SIZE;
  
  	if (inflated_addr > TASK_SIZE - len)
  		return addr;
  	return inflated_addr;
  }
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2133
  #ifdef CONFIG_NUMA
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
2134
  static int shmem_set_policy(struct vm_area_struct *vma, struct mempolicy *mpol)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2135
  {
496ad9aa8   Al Viro   new helper: file_...
2136
  	struct inode *inode = file_inode(vma->vm_file);
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
2137
  	return mpol_set_shared_policy(&SHMEM_I(inode)->policy, vma, mpol);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2138
  }
d8dc74f21   Adrian Bunk   mm/shmem.c: make ...
2139
2140
  static struct mempolicy *shmem_get_policy(struct vm_area_struct *vma,
  					  unsigned long addr)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2141
  {
496ad9aa8   Al Viro   new helper: file_...
2142
  	struct inode *inode = file_inode(vma->vm_file);
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
2143
  	pgoff_t index;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2144

41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
2145
2146
  	index = ((addr - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff;
  	return mpol_shared_policy_lookup(&SHMEM_I(inode)->policy, index);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2147
2148
2149
2150
2151
  }
  #endif
  
  int shmem_lock(struct file *file, int lock, struct user_struct *user)
  {
496ad9aa8   Al Viro   new helper: file_...
2152
  	struct inode *inode = file_inode(file);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2153
2154
  	struct shmem_inode_info *info = SHMEM_I(inode);
  	int retval = -ENOMEM;
ea0dfeb42   Hugh Dickins   shmem: fix possib...
2155
2156
2157
2158
2159
  	/*
  	 * What serializes the accesses to info->flags?
  	 * ipc_lock_object() when called from shmctl_do_lock(),
  	 * no serialization needed when called from shm_destroy().
  	 */
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2160
2161
2162
2163
  	if (lock && !(info->flags & VM_LOCKED)) {
  		if (!user_shm_lock(inode->i_size, user))
  			goto out_nomem;
  		info->flags |= VM_LOCKED;
89e004ea5   Lee Schermerhorn   SHM_LOCKED pages ...
2164
  		mapping_set_unevictable(file->f_mapping);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2165
2166
2167
2168
  	}
  	if (!lock && (info->flags & VM_LOCKED) && user) {
  		user_shm_unlock(inode->i_size, user);
  		info->flags &= ~VM_LOCKED;
89e004ea5   Lee Schermerhorn   SHM_LOCKED pages ...
2169
  		mapping_clear_unevictable(file->f_mapping);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2170
2171
  	}
  	retval = 0;
89e004ea5   Lee Schermerhorn   SHM_LOCKED pages ...
2172

1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2173
  out_nomem:
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2174
2175
  	return retval;
  }
9b83a6a85   Adrian Bunk   [PATCH] mm/{,tiny...
2176
  static int shmem_mmap(struct file *file, struct vm_area_struct *vma)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2177
  {
ab3948f58   Joel Fernandes (Google)   mm/memfd: add an ...
2178
2179
2180
2181
2182
2183
2184
2185
2186
2187
2188
  	struct shmem_inode_info *info = SHMEM_I(file_inode(file));
  
  	if (info->seals & F_SEAL_FUTURE_WRITE) {
  		/*
  		 * New PROT_WRITE and MAP_SHARED mmaps are not allowed when
  		 * "future write" seal active.
  		 */
  		if ((vma->vm_flags & VM_SHARED) && (vma->vm_flags & VM_WRITE))
  			return -EPERM;
  
  		/*
05d351102   Nicolas Geoffray   mm, memfd: fix CO...
2189
2190
2191
2192
2193
  		 * Since an F_SEAL_FUTURE_WRITE sealed memfd can be mapped as
  		 * MAP_SHARED and read-only, take care to not allow mprotect to
  		 * revert protections on such mappings. Do this only for shared
  		 * mappings. For private mappings, don't need to mask
  		 * VM_MAYWRITE as we still want them to be COW-writable.
ab3948f58   Joel Fernandes (Google)   mm/memfd: add an ...
2194
  		 */
05d351102   Nicolas Geoffray   mm, memfd: fix CO...
2195
2196
  		if (vma->vm_flags & VM_SHARED)
  			vma->vm_flags &= ~(VM_MAYWRITE);
ab3948f58   Joel Fernandes (Google)   mm/memfd: add an ...
2197
  	}
51b0bff2f   Catalin Marinas   mm: Allow arm64 m...
2198
2199
  	/* arm64 - allow memory tagging on RAM-based files */
  	vma->vm_flags |= VM_MTE_ALLOWED;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2200
2201
  	file_accessed(file);
  	vma->vm_ops = &shmem_vm_ops;
396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
2202
  	if (IS_ENABLED(CONFIG_TRANSPARENT_HUGEPAGE) &&
f3f0e1d21   Kirill A. Shutemov   khugepaged: add s...
2203
2204
2205
2206
  			((vma->vm_start + ~HPAGE_PMD_MASK) & HPAGE_PMD_MASK) <
  			(vma->vm_end & HPAGE_PMD_MASK)) {
  		khugepaged_enter(vma, vma->vm_flags);
  	}
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2207
2208
  	return 0;
  }
454abafe9   Dmitry Monakhov   ramfs: replace in...
2209
  static struct inode *shmem_get_inode(struct super_block *sb, const struct inode *dir,
09208d150   Al Viro   shmem, ramfs: pro...
2210
  				     umode_t mode, dev_t dev, unsigned long flags)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2211
2212
2213
2214
  {
  	struct inode *inode;
  	struct shmem_inode_info *info;
  	struct shmem_sb_info *sbinfo = SHMEM_SB(sb);
e809d5f0b   Chris Down   tmpfs: per-superb...
2215
  	ino_t ino;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2216

e809d5f0b   Chris Down   tmpfs: per-superb...
2217
  	if (shmem_reserve_inode(sb, &ino))
5b04c6890   Pavel Emelyanov   shmem: factor out...
2218
  		return NULL;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2219
2220
2221
  
  	inode = new_inode(sb);
  	if (inode) {
e809d5f0b   Chris Down   tmpfs: per-superb...
2222
  		inode->i_ino = ino;
454abafe9   Dmitry Monakhov   ramfs: replace in...
2223
  		inode_init_owner(inode, dir, mode);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2224
  		inode->i_blocks = 0;
078cd8279   Deepa Dinamani   fs: Replace CURRE...
2225
  		inode->i_atime = inode->i_mtime = inode->i_ctime = current_time(inode);
46c9a946d   Arnd Bergmann   shmem: use monoto...
2226
  		inode->i_generation = prandom_u32();
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2227
2228
2229
  		info = SHMEM_I(inode);
  		memset(info, 0, (char *)inode - (char *)info);
  		spin_lock_init(&info->lock);
af53d3e9e   Hugh Dickins   mm: swapoff: shme...
2230
  		atomic_set(&info->stop_eviction, 0);
40e041a2c   David Herrmann   shm: add sealing API
2231
  		info->seals = F_SEAL_SEAL;
0b0a0806b   Hugh Dickins   shmem: fix shared...
2232
  		info->flags = flags & VM_NORESERVE;
779750d20   Kirill A. Shutemov   shmem: split huge...
2233
  		INIT_LIST_HEAD(&info->shrinklist);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2234
  		INIT_LIST_HEAD(&info->swaplist);
38f386574   Aristeu Rozanski   xattr: extract si...
2235
  		simple_xattrs_init(&info->xattrs);
72c04902d   Al Viro   Get "no acls for ...
2236
  		cache_no_acl(inode);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2237
2238
2239
  
  		switch (mode & S_IFMT) {
  		default:
39f0247d3   Andreas Gruenbacher   [PATCH] Access Co...
2240
  			inode->i_op = &shmem_special_inode_operations;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2241
2242
2243
  			init_special_inode(inode, mode, dev);
  			break;
  		case S_IFREG:
14fcc23fd   Hugh Dickins   tmpfs: fix kernel...
2244
  			inode->i_mapping->a_ops = &shmem_aops;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2245
2246
  			inode->i_op = &shmem_inode_operations;
  			inode->i_fop = &shmem_file_operations;
71fe804b6   Lee Schermerhorn   mempolicy: use st...
2247
2248
  			mpol_shared_policy_init(&info->policy,
  						 shmem_get_sbmpol(sbinfo));
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2249
2250
  			break;
  		case S_IFDIR:
d8c76e6f4   Dave Hansen   [PATCH] r/o bind ...
2251
  			inc_nlink(inode);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2252
2253
2254
2255
2256
2257
2258
2259
2260
2261
  			/* Some things misbehave if size == 0 on a directory */
  			inode->i_size = 2 * BOGO_DIRENT_SIZE;
  			inode->i_op = &shmem_dir_inode_operations;
  			inode->i_fop = &simple_dir_operations;
  			break;
  		case S_IFLNK:
  			/*
  			 * Must not load anything in the rbtree,
  			 * mpol_free_shared_policy will not be called.
  			 */
71fe804b6   Lee Schermerhorn   mempolicy: use st...
2262
  			mpol_shared_policy_init(&info->policy, NULL);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2263
2264
  			break;
  		}
b45d71fb8   Joel Fernandes (Google)   mm: shmem.c: Corr...
2265
2266
  
  		lockdep_annotate_inode_mutex_key(inode);
5b04c6890   Pavel Emelyanov   shmem: factor out...
2267
2268
  	} else
  		shmem_free_inode(sb);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2269
2270
  	return inode;
  }
0cd6144aa   Johannes Weiner   mm + fs: prepare ...
2271
2272
  bool shmem_mapping(struct address_space *mapping)
  {
f8005451d   Hugh Dickins   tmpfs: change shm...
2273
  	return mapping->a_ops == &shmem_aops;
0cd6144aa   Johannes Weiner   mm + fs: prepare ...
2274
  }
8d1039634   Mike Rapoport   userfaultfd: shme...
2275
2276
2277
2278
2279
2280
2281
  static int shmem_mfill_atomic_pte(struct mm_struct *dst_mm,
  				  pmd_t *dst_pmd,
  				  struct vm_area_struct *dst_vma,
  				  unsigned long dst_addr,
  				  unsigned long src_addr,
  				  bool zeropage,
  				  struct page **pagep)
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2282
2283
2284
  {
  	struct inode *inode = file_inode(dst_vma->vm_file);
  	struct shmem_inode_info *info = SHMEM_I(inode);
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2285
2286
2287
  	struct address_space *mapping = inode->i_mapping;
  	gfp_t gfp = mapping_gfp_mask(mapping);
  	pgoff_t pgoff = linear_page_index(dst_vma, dst_addr);
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2288
2289
2290
2291
2292
  	spinlock_t *ptl;
  	void *page_kaddr;
  	struct page *page;
  	pte_t _dst_pte, *dst_pte;
  	int ret;
e2a50c1f6   Andrea Arcangeli   userfaultfd: shme...
2293
  	pgoff_t offset, max_off;
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2294

cb658a453   Andrea Arcangeli   userfaultfd: shme...
2295
  	ret = -ENOMEM;
0f0796945   Mike Rapoport   shmem: introduce ...
2296
  	if (!shmem_inode_acct_block(inode, 1))
cb658a453   Andrea Arcangeli   userfaultfd: shme...
2297
  		goto out;
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2298

cb658a453   Andrea Arcangeli   userfaultfd: shme...
2299
  	if (!*pagep) {
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2300
2301
  		page = shmem_alloc_page(gfp, info, pgoff);
  		if (!page)
0f0796945   Mike Rapoport   shmem: introduce ...
2302
  			goto out_unacct_blocks;
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2303

8d1039634   Mike Rapoport   userfaultfd: shme...
2304
2305
2306
2307
2308
2309
  		if (!zeropage) {	/* mcopy_atomic */
  			page_kaddr = kmap_atomic(page);
  			ret = copy_from_user(page_kaddr,
  					     (const void __user *)src_addr,
  					     PAGE_SIZE);
  			kunmap_atomic(page_kaddr);
c1e8d7c6a   Michel Lespinasse   mmap locking API:...
2310
  			/* fallback to copy_from_user outside mmap_lock */
8d1039634   Mike Rapoport   userfaultfd: shme...
2311
2312
2313
2314
  			if (unlikely(ret)) {
  				*pagep = page;
  				shmem_inode_unacct_blocks(inode, 1);
  				/* don't free the page */
9e368259a   Andrea Arcangeli   userfaultfd: use ...
2315
  				return -ENOENT;
8d1039634   Mike Rapoport   userfaultfd: shme...
2316
2317
2318
  			}
  		} else {		/* mfill_zeropage_atomic */
  			clear_highpage(page);
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2319
2320
2321
2322
2323
  		}
  	} else {
  		page = *pagep;
  		*pagep = NULL;
  	}
9cc90c664   Andrea Arcangeli   userfaultfd: shme...
2324
2325
2326
  	VM_BUG_ON(PageLocked(page) || PageSwapBacked(page));
  	__SetPageLocked(page);
  	__SetPageSwapBacked(page);
a425d3584   Andrea Arcangeli   userfaultfd: shme...
2327
  	__SetPageUptodate(page);
9cc90c664   Andrea Arcangeli   userfaultfd: shme...
2328

e2a50c1f6   Andrea Arcangeli   userfaultfd: shme...
2329
2330
2331
2332
2333
  	ret = -EFAULT;
  	offset = linear_page_index(dst_vma, dst_addr);
  	max_off = DIV_ROUND_UP(i_size_read(inode), PAGE_SIZE);
  	if (unlikely(offset >= max_off))
  		goto out_release;
552446a41   Matthew Wilcox   shmem: Convert sh...
2334
  	ret = shmem_add_to_page_cache(page, mapping, pgoff, NULL,
3fea5a499   Johannes Weiner   mm: memcontrol: c...
2335
  				      gfp & GFP_RECLAIM_MASK, dst_mm);
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2336
  	if (ret)
3fea5a499   Johannes Weiner   mm: memcontrol: c...
2337
  		goto out_release;
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2338
2339
2340
2341
  
  	_dst_pte = mk_pte(page, dst_vma->vm_page_prot);
  	if (dst_vma->vm_flags & VM_WRITE)
  		_dst_pte = pte_mkwrite(pte_mkdirty(_dst_pte));
dcf7fe9d8   Andrea Arcangeli   userfaultfd: shme...
2342
2343
2344
2345
2346
2347
2348
2349
2350
2351
  	else {
  		/*
  		 * We don't set the pte dirty if the vma has no
  		 * VM_WRITE permission, so mark the page dirty or it
  		 * could be freed from under us. We could do it
  		 * unconditionally before unlock_page(), but doing it
  		 * only if VM_WRITE is not set is faster.
  		 */
  		set_page_dirty(page);
  	}
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2352

4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2353
  	dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl);
e2a50c1f6   Andrea Arcangeli   userfaultfd: shme...
2354
2355
2356
2357
  
  	ret = -EFAULT;
  	max_off = DIV_ROUND_UP(i_size_read(inode), PAGE_SIZE);
  	if (unlikely(offset >= max_off))
3fea5a499   Johannes Weiner   mm: memcontrol: c...
2358
  		goto out_release_unlock;
e2a50c1f6   Andrea Arcangeli   userfaultfd: shme...
2359
2360
  
  	ret = -EEXIST;
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2361
  	if (!pte_none(*dst_pte))
3fea5a499   Johannes Weiner   mm: memcontrol: c...
2362
  		goto out_release_unlock;
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2363

6058eaec8   Johannes Weiner   mm: fold and remo...
2364
  	lru_cache_add(page);
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2365

94b7cc01d   Yang Shi   mm: shmem: disabl...
2366
  	spin_lock_irq(&info->lock);
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2367
2368
2369
  	info->alloced++;
  	inode->i_blocks += BLOCKS_PER_PAGE;
  	shmem_recalc_inode(inode);
94b7cc01d   Yang Shi   mm: shmem: disabl...
2370
  	spin_unlock_irq(&info->lock);
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2371
2372
2373
2374
2375
2376
2377
  
  	inc_mm_counter(dst_mm, mm_counter_file(page));
  	page_add_file_rmap(page, false);
  	set_pte_at(dst_mm, dst_addr, dst_pte, _dst_pte);
  
  	/* No need to invalidate - it was non-present before */
  	update_mmu_cache(dst_vma, dst_addr, dst_pte);
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2378
  	pte_unmap_unlock(dst_pte, ptl);
e2a50c1f6   Andrea Arcangeli   userfaultfd: shme...
2379
  	unlock_page(page);
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2380
2381
2382
  	ret = 0;
  out:
  	return ret;
3fea5a499   Johannes Weiner   mm: memcontrol: c...
2383
  out_release_unlock:
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2384
  	pte_unmap_unlock(dst_pte, ptl);
dcf7fe9d8   Andrea Arcangeli   userfaultfd: shme...
2385
  	ClearPageDirty(page);
e2a50c1f6   Andrea Arcangeli   userfaultfd: shme...
2386
  	delete_from_page_cache(page);
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2387
  out_release:
9cc90c664   Andrea Arcangeli   userfaultfd: shme...
2388
  	unlock_page(page);
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2389
  	put_page(page);
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2390
  out_unacct_blocks:
0f0796945   Mike Rapoport   shmem: introduce ...
2391
  	shmem_inode_unacct_blocks(inode, 1);
4c27fe4c4   Mike Rapoport   userfaultfd: shme...
2392
2393
  	goto out;
  }
8d1039634   Mike Rapoport   userfaultfd: shme...
2394
2395
2396
2397
2398
2399
2400
2401
2402
2403
2404
2405
2406
2407
2408
2409
2410
2411
2412
2413
2414
  int shmem_mcopy_atomic_pte(struct mm_struct *dst_mm,
  			   pmd_t *dst_pmd,
  			   struct vm_area_struct *dst_vma,
  			   unsigned long dst_addr,
  			   unsigned long src_addr,
  			   struct page **pagep)
  {
  	return shmem_mfill_atomic_pte(dst_mm, dst_pmd, dst_vma,
  				      dst_addr, src_addr, false, pagep);
  }
  
  int shmem_mfill_zeropage_pte(struct mm_struct *dst_mm,
  			     pmd_t *dst_pmd,
  			     struct vm_area_struct *dst_vma,
  			     unsigned long dst_addr)
  {
  	struct page *page = NULL;
  
  	return shmem_mfill_atomic_pte(dst_mm, dst_pmd, dst_vma,
  				      dst_addr, 0, true, &page);
  }
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2415
  #ifdef CONFIG_TMPFS
92e1d5be9   Arjan van de Ven   [PATCH] mark stru...
2416
  static const struct inode_operations shmem_symlink_inode_operations;
69f07ec93   Hugh Dickins   tmpfs: use kmemdu...
2417
  static const struct inode_operations shmem_short_symlink_operations;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2418

6d9d88d07   Jarkko Sakkinen   tmpfs: security x...
2419
2420
2421
2422
2423
  #ifdef CONFIG_TMPFS_XATTR
  static int shmem_initxattrs(struct inode *, const struct xattr *, void *);
  #else
  #define shmem_initxattrs NULL
  #endif
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2424
  static int
800d15a53   Nick Piggin   implement simple ...
2425
2426
2427
  shmem_write_begin(struct file *file, struct address_space *mapping,
  			loff_t pos, unsigned len, unsigned flags,
  			struct page **pagep, void **fsdata)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2428
  {
800d15a53   Nick Piggin   implement simple ...
2429
  	struct inode *inode = mapping->host;
40e041a2c   David Herrmann   shm: add sealing API
2430
  	struct shmem_inode_info *info = SHMEM_I(inode);
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2431
  	pgoff_t index = pos >> PAGE_SHIFT;
40e041a2c   David Herrmann   shm: add sealing API
2432
2433
  
  	/* i_mutex is held by caller */
ab3948f58   Joel Fernandes (Google)   mm/memfd: add an ...
2434
2435
2436
  	if (unlikely(info->seals & (F_SEAL_GROW |
  				   F_SEAL_WRITE | F_SEAL_FUTURE_WRITE))) {
  		if (info->seals & (F_SEAL_WRITE | F_SEAL_FUTURE_WRITE))
40e041a2c   David Herrmann   shm: add sealing API
2437
2438
2439
2440
  			return -EPERM;
  		if ((info->seals & F_SEAL_GROW) && pos + len > inode->i_size)
  			return -EPERM;
  	}
9e18eb293   Andres Lagar-Cavilla   tmpfs: mem_cgroup...
2441
  	return shmem_getpage(inode, index, pagep, SGP_WRITE);
800d15a53   Nick Piggin   implement simple ...
2442
2443
2444
2445
2446
2447
2448
2449
  }
  
  static int
  shmem_write_end(struct file *file, struct address_space *mapping,
  			loff_t pos, unsigned len, unsigned copied,
  			struct page *page, void *fsdata)
  {
  	struct inode *inode = mapping->host;
d3602444e   Hugh Dickins   shmem_getpage ret...
2450
2451
  	if (pos + copied > inode->i_size)
  		i_size_write(inode, pos + copied);
ec9516fbc   Hugh Dickins   tmpfs: optimize c...
2452
  	if (!PageUptodate(page)) {
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
2453
2454
2455
2456
2457
2458
2459
2460
2461
2462
2463
  		struct page *head = compound_head(page);
  		if (PageTransCompound(page)) {
  			int i;
  
  			for (i = 0; i < HPAGE_PMD_NR; i++) {
  				if (head + i == page)
  					continue;
  				clear_highpage(head + i);
  				flush_dcache_page(head + i);
  			}
  		}
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2464
2465
  		if (copied < PAGE_SIZE) {
  			unsigned from = pos & (PAGE_SIZE - 1);
ec9516fbc   Hugh Dickins   tmpfs: optimize c...
2466
  			zero_user_segments(page, 0, from,
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2467
  					from + copied, PAGE_SIZE);
ec9516fbc   Hugh Dickins   tmpfs: optimize c...
2468
  		}
800d8c63b   Kirill A. Shutemov   shmem: add huge p...
2469
  		SetPageUptodate(head);
ec9516fbc   Hugh Dickins   tmpfs: optimize c...
2470
  	}
800d15a53   Nick Piggin   implement simple ...
2471
  	set_page_dirty(page);
6746aff74   Wu Fengguang   HWPOISON: shmem: ...
2472
  	unlock_page(page);
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2473
  	put_page(page);
800d15a53   Nick Piggin   implement simple ...
2474

800d15a53   Nick Piggin   implement simple ...
2475
  	return copied;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2476
  }
2ba5bbed0   Al Viro   shmem: switch to ...
2477
  static ssize_t shmem_file_read_iter(struct kiocb *iocb, struct iov_iter *to)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2478
  {
6e58e79db   Al Viro   introduce copy_pa...
2479
2480
  	struct file *file = iocb->ki_filp;
  	struct inode *inode = file_inode(file);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2481
  	struct address_space *mapping = inode->i_mapping;
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
2482
2483
  	pgoff_t index;
  	unsigned long offset;
a0ee5ec52   Hugh Dickins   tmpfs: allocate o...
2484
  	enum sgp_type sgp = SGP_READ;
f7c1d0742   Geert Uytterhoeven   mm: Initialize er...
2485
  	int error = 0;
cb66a7a1f   Al Viro   kill generic_segm...
2486
  	ssize_t retval = 0;
6e58e79db   Al Viro   introduce copy_pa...
2487
  	loff_t *ppos = &iocb->ki_pos;
a0ee5ec52   Hugh Dickins   tmpfs: allocate o...
2488
2489
2490
2491
2492
2493
  
  	/*
  	 * Might this read be for a stacking filesystem?  Then when reading
  	 * holes of a sparse file, we actually need to allocate those pages,
  	 * and even mark them dirty, so it cannot exceed the max_blocks limit.
  	 */
777eda2c5   Al Viro   new helper: iter_...
2494
  	if (!iter_is_iovec(to))
75edd345e   Hugh Dickins   tmpfs: preliminar...
2495
  		sgp = SGP_CACHE;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2496

09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2497
2498
  	index = *ppos >> PAGE_SHIFT;
  	offset = *ppos & ~PAGE_MASK;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2499
2500
2501
  
  	for (;;) {
  		struct page *page = NULL;
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
2502
2503
  		pgoff_t end_index;
  		unsigned long nr, ret;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2504
  		loff_t i_size = i_size_read(inode);
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2505
  		end_index = i_size >> PAGE_SHIFT;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2506
2507
2508
  		if (index > end_index)
  			break;
  		if (index == end_index) {
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2509
  			nr = i_size & ~PAGE_MASK;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2510
2511
2512
  			if (nr <= offset)
  				break;
  		}
9e18eb293   Andres Lagar-Cavilla   tmpfs: mem_cgroup...
2513
  		error = shmem_getpage(inode, index, &page, sgp);
6e58e79db   Al Viro   introduce copy_pa...
2514
2515
2516
  		if (error) {
  			if (error == -EINVAL)
  				error = 0;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2517
2518
  			break;
  		}
75edd345e   Hugh Dickins   tmpfs: preliminar...
2519
2520
2521
  		if (page) {
  			if (sgp == SGP_CACHE)
  				set_page_dirty(page);
d3602444e   Hugh Dickins   shmem_getpage ret...
2522
  			unlock_page(page);
75edd345e   Hugh Dickins   tmpfs: preliminar...
2523
  		}
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2524
2525
2526
  
  		/*
  		 * We must evaluate after, since reads (unlike writes)
1b1dcc1b5   Jes Sorensen   [PATCH] mutex sub...
2527
  		 * are called without i_mutex protection against truncate
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2528
  		 */
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2529
  		nr = PAGE_SIZE;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2530
  		i_size = i_size_read(inode);
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2531
  		end_index = i_size >> PAGE_SHIFT;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2532
  		if (index == end_index) {
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2533
  			nr = i_size & ~PAGE_MASK;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2534
2535
  			if (nr <= offset) {
  				if (page)
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2536
  					put_page(page);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2537
2538
2539
2540
2541
2542
2543
2544
2545
2546
2547
2548
2549
2550
2551
2552
2553
2554
  				break;
  			}
  		}
  		nr -= offset;
  
  		if (page) {
  			/*
  			 * If users can be writing to this page using arbitrary
  			 * virtual addresses, take care about potential aliasing
  			 * before reading the page on the kernel side.
  			 */
  			if (mapping_writably_mapped(mapping))
  				flush_dcache_page(page);
  			/*
  			 * Mark the page accessed if we read the beginning.
  			 */
  			if (!offset)
  				mark_page_accessed(page);
b5810039a   Nick Piggin   [PATCH] core remo...
2555
  		} else {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2556
  			page = ZERO_PAGE(0);
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2557
  			get_page(page);
b5810039a   Nick Piggin   [PATCH] core remo...
2558
  		}
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2559
2560
2561
2562
  
  		/*
  		 * Ok, we have the page, and it's up-to-date, so
  		 * now we can copy it to user space...
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2563
  		 */
2ba5bbed0   Al Viro   shmem: switch to ...
2564
  		ret = copy_page_to_iter(page, offset, nr, to);
6e58e79db   Al Viro   introduce copy_pa...
2565
  		retval += ret;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2566
  		offset += ret;
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2567
2568
  		index += offset >> PAGE_SHIFT;
  		offset &= ~PAGE_MASK;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2569

09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2570
  		put_page(page);
2ba5bbed0   Al Viro   shmem: switch to ...
2571
  		if (!iov_iter_count(to))
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2572
  			break;
6e58e79db   Al Viro   introduce copy_pa...
2573
2574
2575
2576
  		if (ret < nr) {
  			error = -EFAULT;
  			break;
  		}
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2577
2578
  		cond_resched();
  	}
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2579
  	*ppos = ((loff_t) index << PAGE_SHIFT) + offset;
6e58e79db   Al Viro   introduce copy_pa...
2580
2581
  	file_accessed(file);
  	return retval ? retval : error;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2582
  }
220f2ac91   Hugh Dickins   tmpfs: support SE...
2583
  /*
7f4446eef   Matthew Wilcox   shmem: Comment fi...
2584
   * llseek SEEK_DATA or SEEK_HOLE through the page cache.
220f2ac91   Hugh Dickins   tmpfs: support SE...
2585
2586
   */
  static pgoff_t shmem_seek_hole_data(struct address_space *mapping,
965c8e59c   Andrew Morton   lseek: the "whenc...
2587
  				    pgoff_t index, pgoff_t end, int whence)
220f2ac91   Hugh Dickins   tmpfs: support SE...
2588
2589
2590
2591
2592
2593
  {
  	struct page *page;
  	struct pagevec pvec;
  	pgoff_t indices[PAGEVEC_SIZE];
  	bool done = false;
  	int i;
866798201   Mel Gorman   mm, pagevec: remo...
2594
  	pagevec_init(&pvec);
220f2ac91   Hugh Dickins   tmpfs: support SE...
2595
2596
  	pvec.nr = 1;		/* start small: we may be there already */
  	while (!done) {
0cd6144aa   Johannes Weiner   mm + fs: prepare ...
2597
  		pvec.nr = find_get_entries(mapping, index,
220f2ac91   Hugh Dickins   tmpfs: support SE...
2598
2599
  					pvec.nr, pvec.pages, indices);
  		if (!pvec.nr) {
965c8e59c   Andrew Morton   lseek: the "whenc...
2600
  			if (whence == SEEK_DATA)
220f2ac91   Hugh Dickins   tmpfs: support SE...
2601
2602
2603
2604
2605
  				index = end;
  			break;
  		}
  		for (i = 0; i < pvec.nr; i++, index++) {
  			if (index < indices[i]) {
965c8e59c   Andrew Morton   lseek: the "whenc...
2606
  				if (whence == SEEK_HOLE) {
220f2ac91   Hugh Dickins   tmpfs: support SE...
2607
2608
2609
2610
2611
2612
  					done = true;
  					break;
  				}
  				index = indices[i];
  			}
  			page = pvec.pages[i];
3159f943a   Matthew Wilcox   xarray: Replace e...
2613
  			if (page && !xa_is_value(page)) {
220f2ac91   Hugh Dickins   tmpfs: support SE...
2614
2615
2616
2617
  				if (!PageUptodate(page))
  					page = NULL;
  			}
  			if (index >= end ||
965c8e59c   Andrew Morton   lseek: the "whenc...
2618
2619
  			    (page && whence == SEEK_DATA) ||
  			    (!page && whence == SEEK_HOLE)) {
220f2ac91   Hugh Dickins   tmpfs: support SE...
2620
2621
2622
2623
  				done = true;
  				break;
  			}
  		}
0cd6144aa   Johannes Weiner   mm + fs: prepare ...
2624
  		pagevec_remove_exceptionals(&pvec);
220f2ac91   Hugh Dickins   tmpfs: support SE...
2625
2626
2627
2628
2629
2630
  		pagevec_release(&pvec);
  		pvec.nr = PAGEVEC_SIZE;
  		cond_resched();
  	}
  	return index;
  }
965c8e59c   Andrew Morton   lseek: the "whenc...
2631
  static loff_t shmem_file_llseek(struct file *file, loff_t offset, int whence)
220f2ac91   Hugh Dickins   tmpfs: support SE...
2632
2633
2634
2635
2636
  {
  	struct address_space *mapping = file->f_mapping;
  	struct inode *inode = mapping->host;
  	pgoff_t start, end;
  	loff_t new_offset;
965c8e59c   Andrew Morton   lseek: the "whenc...
2637
2638
  	if (whence != SEEK_DATA && whence != SEEK_HOLE)
  		return generic_file_llseek_size(file, offset, whence,
220f2ac91   Hugh Dickins   tmpfs: support SE...
2639
  					MAX_LFS_FILESIZE, i_size_read(inode));
5955102c9   Al Viro   wrappers for ->i_...
2640
  	inode_lock(inode);
220f2ac91   Hugh Dickins   tmpfs: support SE...
2641
  	/* We're holding i_mutex so we can access i_size directly */
1a4136469   Yufen Yu   tmpfs: make lseek...
2642
  	if (offset < 0 || offset >= inode->i_size)
220f2ac91   Hugh Dickins   tmpfs: support SE...
2643
2644
  		offset = -ENXIO;
  	else {
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2645
2646
  		start = offset >> PAGE_SHIFT;
  		end = (inode->i_size + PAGE_SIZE - 1) >> PAGE_SHIFT;
965c8e59c   Andrew Morton   lseek: the "whenc...
2647
  		new_offset = shmem_seek_hole_data(mapping, start, end, whence);
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2648
  		new_offset <<= PAGE_SHIFT;
220f2ac91   Hugh Dickins   tmpfs: support SE...
2649
2650
2651
  		if (new_offset > offset) {
  			if (new_offset < inode->i_size)
  				offset = new_offset;
965c8e59c   Andrew Morton   lseek: the "whenc...
2652
  			else if (whence == SEEK_DATA)
220f2ac91   Hugh Dickins   tmpfs: support SE...
2653
2654
2655
2656
2657
  				offset = -ENXIO;
  			else
  				offset = inode->i_size;
  		}
  	}
387aae6fd   Hugh Dickins   tmpfs: fix SEEK_D...
2658
2659
  	if (offset >= 0)
  		offset = vfs_setpos(file, offset, MAX_LFS_FILESIZE);
5955102c9   Al Viro   wrappers for ->i_...
2660
  	inode_unlock(inode);
220f2ac91   Hugh Dickins   tmpfs: support SE...
2661
2662
  	return offset;
  }
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
2663
2664
2665
  static long shmem_fallocate(struct file *file, int mode, loff_t offset,
  							 loff_t len)
  {
496ad9aa8   Al Viro   new helper: file_...
2666
  	struct inode *inode = file_inode(file);
e2d12e22c   Hugh Dickins   tmpfs: support fa...
2667
  	struct shmem_sb_info *sbinfo = SHMEM_SB(inode->i_sb);
40e041a2c   David Herrmann   shm: add sealing API
2668
  	struct shmem_inode_info *info = SHMEM_I(inode);
1aac14003   Hugh Dickins   tmpfs: quit when ...
2669
  	struct shmem_falloc shmem_falloc;
e2d12e22c   Hugh Dickins   tmpfs: support fa...
2670
2671
  	pgoff_t start, index, end;
  	int error;
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
2672

13ace4d0d   Hugh Dickins   tmpfs: ZERO_RANGE...
2673
2674
  	if (mode & ~(FALLOC_FL_KEEP_SIZE | FALLOC_FL_PUNCH_HOLE))
  		return -EOPNOTSUPP;
5955102c9   Al Viro   wrappers for ->i_...
2675
  	inode_lock(inode);
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
2676
2677
2678
2679
2680
  
  	if (mode & FALLOC_FL_PUNCH_HOLE) {
  		struct address_space *mapping = file->f_mapping;
  		loff_t unmap_start = round_up(offset, PAGE_SIZE);
  		loff_t unmap_end = round_down(offset + len, PAGE_SIZE) - 1;
8e205f779   Hugh Dickins   shmem: fix faulti...
2681
  		DECLARE_WAIT_QUEUE_HEAD_ONSTACK(shmem_falloc_waitq);
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
2682

40e041a2c   David Herrmann   shm: add sealing API
2683
  		/* protected by i_mutex */
ab3948f58   Joel Fernandes (Google)   mm/memfd: add an ...
2684
  		if (info->seals & (F_SEAL_WRITE | F_SEAL_FUTURE_WRITE)) {
40e041a2c   David Herrmann   shm: add sealing API
2685
2686
2687
  			error = -EPERM;
  			goto out;
  		}
8e205f779   Hugh Dickins   shmem: fix faulti...
2688
  		shmem_falloc.waitq = &shmem_falloc_waitq;
aa71ecd8d   Chen Jun   mm/shmem.c: cast ...
2689
  		shmem_falloc.start = (u64)unmap_start >> PAGE_SHIFT;
f00cdc6df   Hugh Dickins   shmem: fix faulti...
2690
2691
2692
2693
  		shmem_falloc.next = (unmap_end + 1) >> PAGE_SHIFT;
  		spin_lock(&inode->i_lock);
  		inode->i_private = &shmem_falloc;
  		spin_unlock(&inode->i_lock);
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
2694
2695
2696
2697
2698
  		if ((u64)unmap_end > (u64)unmap_start)
  			unmap_mapping_range(mapping, unmap_start,
  					    1 + unmap_end - unmap_start, 0);
  		shmem_truncate_range(inode, offset, offset + len - 1);
  		/* No need to unmap again: hole-punching leaves COWed pages */
8e205f779   Hugh Dickins   shmem: fix faulti...
2699
2700
2701
2702
  
  		spin_lock(&inode->i_lock);
  		inode->i_private = NULL;
  		wake_up_all(&shmem_falloc_waitq);
2055da973   Ingo Molnar   sched/wait: Disam...
2703
  		WARN_ON_ONCE(!list_empty(&shmem_falloc_waitq.head));
8e205f779   Hugh Dickins   shmem: fix faulti...
2704
  		spin_unlock(&inode->i_lock);
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
2705
  		error = 0;
8e205f779   Hugh Dickins   shmem: fix faulti...
2706
  		goto out;
e2d12e22c   Hugh Dickins   tmpfs: support fa...
2707
2708
2709
2710
2711
2712
  	}
  
  	/* We need to check rlimit even when FALLOC_FL_KEEP_SIZE */
  	error = inode_newsize_ok(inode, offset + len);
  	if (error)
  		goto out;
40e041a2c   David Herrmann   shm: add sealing API
2713
2714
2715
2716
  	if ((info->seals & F_SEAL_GROW) && offset + len > inode->i_size) {
  		error = -EPERM;
  		goto out;
  	}
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2717
2718
  	start = offset >> PAGE_SHIFT;
  	end = (offset + len + PAGE_SIZE - 1) >> PAGE_SHIFT;
e2d12e22c   Hugh Dickins   tmpfs: support fa...
2719
2720
2721
2722
  	/* Try to avoid a swapstorm if len is impossible to satisfy */
  	if (sbinfo->max_blocks && end - start > sbinfo->max_blocks) {
  		error = -ENOSPC;
  		goto out;
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
2723
  	}
8e205f779   Hugh Dickins   shmem: fix faulti...
2724
  	shmem_falloc.waitq = NULL;
1aac14003   Hugh Dickins   tmpfs: quit when ...
2725
2726
2727
2728
2729
2730
2731
  	shmem_falloc.start = start;
  	shmem_falloc.next  = start;
  	shmem_falloc.nr_falloced = 0;
  	shmem_falloc.nr_unswapped = 0;
  	spin_lock(&inode->i_lock);
  	inode->i_private = &shmem_falloc;
  	spin_unlock(&inode->i_lock);
e2d12e22c   Hugh Dickins   tmpfs: support fa...
2732
2733
2734
2735
2736
2737
2738
2739
2740
  	for (index = start; index < end; index++) {
  		struct page *page;
  
  		/*
  		 * Good, the fallocate(2) manpage permits EINTR: we may have
  		 * been interrupted because we are using up too much memory.
  		 */
  		if (signal_pending(current))
  			error = -EINTR;
1aac14003   Hugh Dickins   tmpfs: quit when ...
2741
2742
  		else if (shmem_falloc.nr_unswapped > shmem_falloc.nr_falloced)
  			error = -ENOMEM;
e2d12e22c   Hugh Dickins   tmpfs: support fa...
2743
  		else
9e18eb293   Andres Lagar-Cavilla   tmpfs: mem_cgroup...
2744
  			error = shmem_getpage(inode, index, &page, SGP_FALLOC);
e2d12e22c   Hugh Dickins   tmpfs: support fa...
2745
  		if (error) {
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
2746
  			/* Remove the !PageUptodate pages we added */
7f5565670   Hugh Dickins   tmpfs: fix regres...
2747
2748
2749
2750
2751
  			if (index > start) {
  				shmem_undo_range(inode,
  				    (loff_t)start << PAGE_SHIFT,
  				    ((loff_t)index << PAGE_SHIFT) - 1, true);
  			}
1aac14003   Hugh Dickins   tmpfs: quit when ...
2752
  			goto undone;
e2d12e22c   Hugh Dickins   tmpfs: support fa...
2753
  		}
e2d12e22c   Hugh Dickins   tmpfs: support fa...
2754
  		/*
1aac14003   Hugh Dickins   tmpfs: quit when ...
2755
2756
2757
2758
2759
2760
2761
2762
  		 * Inform shmem_writepage() how far we have reached.
  		 * No need for lock or barrier: we have the page lock.
  		 */
  		shmem_falloc.next++;
  		if (!PageUptodate(page))
  			shmem_falloc.nr_falloced++;
  
  		/*
1635f6a74   Hugh Dickins   tmpfs: undo fallo...
2763
2764
2765
  		 * If !PageUptodate, leave it that way so that freeable pages
  		 * can be recognized if we need to rollback on error later.
  		 * But set_page_dirty so that memory pressure will swap rather
e2d12e22c   Hugh Dickins   tmpfs: support fa...
2766
2767
2768
2769
2770
  		 * than free the pages we are allocating (and SGP_CACHE pages
  		 * might still be clean: we now need to mark those dirty too).
  		 */
  		set_page_dirty(page);
  		unlock_page(page);
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2771
  		put_page(page);
e2d12e22c   Hugh Dickins   tmpfs: support fa...
2772
2773
2774
2775
2776
  		cond_resched();
  	}
  
  	if (!(mode & FALLOC_FL_KEEP_SIZE) && offset + len > inode->i_size)
  		i_size_write(inode, offset + len);
078cd8279   Deepa Dinamani   fs: Replace CURRE...
2777
  	inode->i_ctime = current_time(inode);
1aac14003   Hugh Dickins   tmpfs: quit when ...
2778
2779
2780
2781
  undone:
  	spin_lock(&inode->i_lock);
  	inode->i_private = NULL;
  	spin_unlock(&inode->i_lock);
e2d12e22c   Hugh Dickins   tmpfs: support fa...
2782
  out:
5955102c9   Al Viro   wrappers for ->i_...
2783
  	inode_unlock(inode);
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
2784
2785
  	return error;
  }
726c33422   David Howells   [PATCH] VFS: Perm...
2786
  static int shmem_statfs(struct dentry *dentry, struct kstatfs *buf)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2787
  {
726c33422   David Howells   [PATCH] VFS: Perm...
2788
  	struct shmem_sb_info *sbinfo = SHMEM_SB(dentry->d_sb);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2789
2790
  
  	buf->f_type = TMPFS_MAGIC;
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
2791
  	buf->f_bsize = PAGE_SIZE;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2792
  	buf->f_namelen = NAME_MAX;
0edd73b33   Hugh Dickins   [PATCH] shmem: re...
2793
  	if (sbinfo->max_blocks) {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2794
  		buf->f_blocks = sbinfo->max_blocks;
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
2795
2796
2797
  		buf->f_bavail =
  		buf->f_bfree  = sbinfo->max_blocks -
  				percpu_counter_sum(&sbinfo->used_blocks);
0edd73b33   Hugh Dickins   [PATCH] shmem: re...
2798
2799
  	}
  	if (sbinfo->max_inodes) {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2800
2801
  		buf->f_files = sbinfo->max_inodes;
  		buf->f_ffree = sbinfo->free_inodes;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2802
2803
2804
2805
2806
2807
2808
2809
2810
  	}
  	/* else leave those fields 0 like simple_statfs */
  	return 0;
  }
  
  /*
   * File creation. Allocate an inode, and we're done..
   */
  static int
1a67aafb5   Al Viro   switch ->mknod() ...
2811
  shmem_mknod(struct inode *dir, struct dentry *dentry, umode_t mode, dev_t dev)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2812
  {
0b0a0806b   Hugh Dickins   shmem: fix shared...
2813
  	struct inode *inode;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2814
  	int error = -ENOSPC;
454abafe9   Dmitry Monakhov   ramfs: replace in...
2815
  	inode = shmem_get_inode(dir->i_sb, dir, mode, dev, VM_NORESERVE);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2816
  	if (inode) {
feda821e7   Christoph Hellwig   fs: remove generi...
2817
2818
2819
  		error = simple_acl_create(dir, inode);
  		if (error)
  			goto out_iput;
2a7dba391   Eric Paris   fs/vfs/security: ...
2820
  		error = security_inode_init_security(inode, dir,
9d8f13ba3   Mimi Zohar   security: new sec...
2821
  						     &dentry->d_name,
6d9d88d07   Jarkko Sakkinen   tmpfs: security x...
2822
  						     shmem_initxattrs, NULL);
feda821e7   Christoph Hellwig   fs: remove generi...
2823
2824
  		if (error && error != -EOPNOTSUPP)
  			goto out_iput;
37ec43cdc   Mimi Zohar   evm: calculate HM...
2825

718deb6b6   Al Viro   Fix breakage in s...
2826
  		error = 0;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2827
  		dir->i_size += BOGO_DIRENT_SIZE;
078cd8279   Deepa Dinamani   fs: Replace CURRE...
2828
  		dir->i_ctime = dir->i_mtime = current_time(dir);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2829
2830
  		d_instantiate(dentry, inode);
  		dget(dentry); /* Extra count - pin the dentry in core */
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2831
2832
  	}
  	return error;
feda821e7   Christoph Hellwig   fs: remove generi...
2833
2834
2835
  out_iput:
  	iput(inode);
  	return error;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2836
  }
60545d0d4   Al Viro   [O_TMPFILE] it's ...
2837
2838
2839
2840
2841
2842
2843
2844
2845
2846
2847
  static int
  shmem_tmpfile(struct inode *dir, struct dentry *dentry, umode_t mode)
  {
  	struct inode *inode;
  	int error = -ENOSPC;
  
  	inode = shmem_get_inode(dir->i_sb, dir, mode, 0, VM_NORESERVE);
  	if (inode) {
  		error = security_inode_init_security(inode, dir,
  						     NULL,
  						     shmem_initxattrs, NULL);
feda821e7   Christoph Hellwig   fs: remove generi...
2848
2849
2850
2851
2852
  		if (error && error != -EOPNOTSUPP)
  			goto out_iput;
  		error = simple_acl_create(dir, inode);
  		if (error)
  			goto out_iput;
60545d0d4   Al Viro   [O_TMPFILE] it's ...
2853
2854
2855
  		d_tmpfile(dentry, inode);
  	}
  	return error;
feda821e7   Christoph Hellwig   fs: remove generi...
2856
2857
2858
  out_iput:
  	iput(inode);
  	return error;
60545d0d4   Al Viro   [O_TMPFILE] it's ...
2859
  }
18bb1db3e   Al Viro   switch vfs_mkdir(...
2860
  static int shmem_mkdir(struct inode *dir, struct dentry *dentry, umode_t mode)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2861
2862
2863
2864
2865
  {
  	int error;
  
  	if ((error = shmem_mknod(dir, dentry, mode | S_IFDIR, 0)))
  		return error;
d8c76e6f4   Dave Hansen   [PATCH] r/o bind ...
2866
  	inc_nlink(dir);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2867
2868
  	return 0;
  }
4acdaf27e   Al Viro   switch ->create()...
2869
  static int shmem_create(struct inode *dir, struct dentry *dentry, umode_t mode,
ebfc3b49a   Al Viro   don't pass nameid...
2870
  		bool excl)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2871
2872
2873
2874
2875
2876
2877
2878
2879
  {
  	return shmem_mknod(dir, dentry, mode | S_IFREG, 0);
  }
  
  /*
   * Link a file..
   */
  static int shmem_link(struct dentry *old_dentry, struct inode *dir, struct dentry *dentry)
  {
75c3cfa85   David Howells   VFS: assorted wei...
2880
  	struct inode *inode = d_inode(old_dentry);
29b00e609   Darrick J. Wong   tmpfs: fix uninit...
2881
  	int ret = 0;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2882
2883
2884
2885
2886
  
  	/*
  	 * No ordinary (disk based) filesystem counts links as inodes;
  	 * but each new link needs a new dentry, pinning lowmem, and
  	 * tmpfs dentries cannot be pruned until they are unlinked.
1062af920   Darrick J. Wong   tmpfs: fix link a...
2887
2888
  	 * But if an O_TMPFILE file is linked into the tmpfs, the
  	 * first link must skip that, to get the accounting right.
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2889
  	 */
1062af920   Darrick J. Wong   tmpfs: fix link a...
2890
  	if (inode->i_nlink) {
e809d5f0b   Chris Down   tmpfs: per-superb...
2891
  		ret = shmem_reserve_inode(inode->i_sb, NULL);
1062af920   Darrick J. Wong   tmpfs: fix link a...
2892
2893
2894
  		if (ret)
  			goto out;
  	}
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2895
2896
  
  	dir->i_size += BOGO_DIRENT_SIZE;
078cd8279   Deepa Dinamani   fs: Replace CURRE...
2897
  	inode->i_ctime = dir->i_ctime = dir->i_mtime = current_time(inode);
d8c76e6f4   Dave Hansen   [PATCH] r/o bind ...
2898
  	inc_nlink(inode);
7de9c6ee3   Al Viro   new helper: ihold()
2899
  	ihold(inode);	/* New dentry reference */
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2900
2901
  	dget(dentry);		/* Extra pinning count for the created dentry */
  	d_instantiate(dentry, inode);
5b04c6890   Pavel Emelyanov   shmem: factor out...
2902
2903
  out:
  	return ret;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2904
2905
2906
2907
  }
  
  static int shmem_unlink(struct inode *dir, struct dentry *dentry)
  {
75c3cfa85   David Howells   VFS: assorted wei...
2908
  	struct inode *inode = d_inode(dentry);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2909

5b04c6890   Pavel Emelyanov   shmem: factor out...
2910
2911
  	if (inode->i_nlink > 1 && !S_ISDIR(inode->i_mode))
  		shmem_free_inode(inode->i_sb);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2912
2913
  
  	dir->i_size -= BOGO_DIRENT_SIZE;
078cd8279   Deepa Dinamani   fs: Replace CURRE...
2914
  	inode->i_ctime = dir->i_ctime = dir->i_mtime = current_time(inode);
9a53c3a78   Dave Hansen   [PATCH] r/o bind ...
2915
  	drop_nlink(inode);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2916
2917
2918
2919
2920
2921
2922
2923
  	dput(dentry);	/* Undo the count from "create" - this does all the work */
  	return 0;
  }
  
  static int shmem_rmdir(struct inode *dir, struct dentry *dentry)
  {
  	if (!simple_empty(dentry))
  		return -ENOTEMPTY;
75c3cfa85   David Howells   VFS: assorted wei...
2924
  	drop_nlink(d_inode(dentry));
9a53c3a78   Dave Hansen   [PATCH] r/o bind ...
2925
  	drop_nlink(dir);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2926
2927
  	return shmem_unlink(dir, dentry);
  }
37456771c   Miklos Szeredi   shmem: support RE...
2928
2929
  static int shmem_exchange(struct inode *old_dir, struct dentry *old_dentry, struct inode *new_dir, struct dentry *new_dentry)
  {
e36cb0b89   David Howells   VFS: (Scripted) C...
2930
2931
  	bool old_is_dir = d_is_dir(old_dentry);
  	bool new_is_dir = d_is_dir(new_dentry);
37456771c   Miklos Szeredi   shmem: support RE...
2932
2933
2934
2935
2936
2937
2938
2939
2940
2941
2942
2943
  
  	if (old_dir != new_dir && old_is_dir != new_is_dir) {
  		if (old_is_dir) {
  			drop_nlink(old_dir);
  			inc_nlink(new_dir);
  		} else {
  			drop_nlink(new_dir);
  			inc_nlink(old_dir);
  		}
  	}
  	old_dir->i_ctime = old_dir->i_mtime =
  	new_dir->i_ctime = new_dir->i_mtime =
75c3cfa85   David Howells   VFS: assorted wei...
2944
  	d_inode(old_dentry)->i_ctime =
078cd8279   Deepa Dinamani   fs: Replace CURRE...
2945
  	d_inode(new_dentry)->i_ctime = current_time(old_dir);
37456771c   Miklos Szeredi   shmem: support RE...
2946
2947
2948
  
  	return 0;
  }
46fdb794e   Miklos Szeredi   shmem: support RE...
2949
2950
2951
2952
2953
2954
2955
2956
2957
2958
2959
2960
2961
2962
2963
2964
2965
2966
2967
2968
2969
2970
2971
2972
2973
  static int shmem_whiteout(struct inode *old_dir, struct dentry *old_dentry)
  {
  	struct dentry *whiteout;
  	int error;
  
  	whiteout = d_alloc(old_dentry->d_parent, &old_dentry->d_name);
  	if (!whiteout)
  		return -ENOMEM;
  
  	error = shmem_mknod(old_dir, whiteout,
  			    S_IFCHR | WHITEOUT_MODE, WHITEOUT_DEV);
  	dput(whiteout);
  	if (error)
  		return error;
  
  	/*
  	 * Cheat and hash the whiteout while the old dentry is still in
  	 * place, instead of playing games with FS_RENAME_DOES_D_MOVE.
  	 *
  	 * d_lookup() will consistently find one of them at this point,
  	 * not sure which one, but that isn't even important.
  	 */
  	d_rehash(whiteout);
  	return 0;
  }
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2974
2975
2976
2977
2978
2979
  /*
   * The VFS layer already does all the dentry stuff for rename,
   * we just have to decrement the usage count for the target if
   * it exists so that the VFS layer correctly free's it when it
   * gets overwritten.
   */
3b69ff51d   Miklos Szeredi   shmem: support RE...
2980
  static int shmem_rename2(struct inode *old_dir, struct dentry *old_dentry, struct inode *new_dir, struct dentry *new_dentry, unsigned int flags)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2981
  {
75c3cfa85   David Howells   VFS: assorted wei...
2982
  	struct inode *inode = d_inode(old_dentry);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2983
  	int they_are_dirs = S_ISDIR(inode->i_mode);
46fdb794e   Miklos Szeredi   shmem: support RE...
2984
  	if (flags & ~(RENAME_NOREPLACE | RENAME_EXCHANGE | RENAME_WHITEOUT))
3b69ff51d   Miklos Szeredi   shmem: support RE...
2985
  		return -EINVAL;
37456771c   Miklos Szeredi   shmem: support RE...
2986
2987
  	if (flags & RENAME_EXCHANGE)
  		return shmem_exchange(old_dir, old_dentry, new_dir, new_dentry);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2988
2989
  	if (!simple_empty(new_dentry))
  		return -ENOTEMPTY;
46fdb794e   Miklos Szeredi   shmem: support RE...
2990
2991
2992
2993
2994
2995
2996
  	if (flags & RENAME_WHITEOUT) {
  		int error;
  
  		error = shmem_whiteout(old_dir, old_dentry);
  		if (error)
  			return error;
  	}
75c3cfa85   David Howells   VFS: assorted wei...
2997
  	if (d_really_is_positive(new_dentry)) {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
2998
  		(void) shmem_unlink(new_dir, new_dentry);
b928095b0   Miklos Szeredi   shmem: fix nlink ...
2999
  		if (they_are_dirs) {
75c3cfa85   David Howells   VFS: assorted wei...
3000
  			drop_nlink(d_inode(new_dentry));
9a53c3a78   Dave Hansen   [PATCH] r/o bind ...
3001
  			drop_nlink(old_dir);
b928095b0   Miklos Szeredi   shmem: fix nlink ...
3002
  		}
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3003
  	} else if (they_are_dirs) {
9a53c3a78   Dave Hansen   [PATCH] r/o bind ...
3004
  		drop_nlink(old_dir);
d8c76e6f4   Dave Hansen   [PATCH] r/o bind ...
3005
  		inc_nlink(new_dir);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3006
3007
3008
3009
3010
3011
  	}
  
  	old_dir->i_size -= BOGO_DIRENT_SIZE;
  	new_dir->i_size += BOGO_DIRENT_SIZE;
  	old_dir->i_ctime = old_dir->i_mtime =
  	new_dir->i_ctime = new_dir->i_mtime =
078cd8279   Deepa Dinamani   fs: Replace CURRE...
3012
  	inode->i_ctime = current_time(old_dir);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3013
3014
3015
3016
3017
3018
3019
3020
  	return 0;
  }
  
  static int shmem_symlink(struct inode *dir, struct dentry *dentry, const char *symname)
  {
  	int error;
  	int len;
  	struct inode *inode;
9276aad6c   Hugh Dickins   tmpfs: remove_shm...
3021
  	struct page *page;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3022
3023
  
  	len = strlen(symname) + 1;
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
3024
  	if (len > PAGE_SIZE)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3025
  		return -ENAMETOOLONG;
0825a6f98   Joe Perches   mm: use octal not...
3026
3027
  	inode = shmem_get_inode(dir->i_sb, dir, S_IFLNK | 0777, 0,
  				VM_NORESERVE);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3028
3029
  	if (!inode)
  		return -ENOSPC;
9d8f13ba3   Mimi Zohar   security: new sec...
3030
  	error = security_inode_init_security(inode, dir, &dentry->d_name,
6d9d88d07   Jarkko Sakkinen   tmpfs: security x...
3031
  					     shmem_initxattrs, NULL);
343c3d7f0   Mateusz Nosek   mm/shmem.c: clean...
3032
3033
3034
  	if (error && error != -EOPNOTSUPP) {
  		iput(inode);
  		return error;
570bc1c2e   Stephen Smalley   [PATCH] tmpfs: En...
3035
  	}
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3036
  	inode->i_size = len-1;
69f07ec93   Hugh Dickins   tmpfs: use kmemdu...
3037
  	if (len <= SHORT_SYMLINK_LEN) {
3ed47db34   Al Viro   make sure that fr...
3038
3039
  		inode->i_link = kmemdup(symname, len, GFP_KERNEL);
  		if (!inode->i_link) {
69f07ec93   Hugh Dickins   tmpfs: use kmemdu...
3040
3041
3042
3043
  			iput(inode);
  			return -ENOMEM;
  		}
  		inode->i_op = &shmem_short_symlink_operations;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3044
  	} else {
e8ecde25f   Al Viro   Make sure that hi...
3045
  		inode_nohighmem(inode);
9e18eb293   Andres Lagar-Cavilla   tmpfs: mem_cgroup...
3046
  		error = shmem_getpage(inode, 0, &page, SGP_WRITE);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3047
3048
3049
3050
  		if (error) {
  			iput(inode);
  			return error;
  		}
14fcc23fd   Hugh Dickins   tmpfs: fix kernel...
3051
  		inode->i_mapping->a_ops = &shmem_aops;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3052
  		inode->i_op = &shmem_symlink_inode_operations;
21fc61c73   Al Viro   don't put symlink...
3053
  		memcpy(page_address(page), symname, len);
ec9516fbc   Hugh Dickins   tmpfs: optimize c...
3054
  		SetPageUptodate(page);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3055
  		set_page_dirty(page);
6746aff74   Wu Fengguang   HWPOISON: shmem: ...
3056
  		unlock_page(page);
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
3057
  		put_page(page);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3058
  	}
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3059
  	dir->i_size += BOGO_DIRENT_SIZE;
078cd8279   Deepa Dinamani   fs: Replace CURRE...
3060
  	dir->i_ctime = dir->i_mtime = current_time(dir);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3061
3062
3063
3064
  	d_instantiate(dentry, inode);
  	dget(dentry);
  	return 0;
  }
fceef393a   Al Viro   switch ->get_link...
3065
  static void shmem_put_link(void *arg)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3066
  {
fceef393a   Al Viro   switch ->get_link...
3067
3068
  	mark_page_accessed(arg);
  	put_page(arg);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3069
  }
6b2553918   Al Viro   replace ->follow_...
3070
  static const char *shmem_get_link(struct dentry *dentry,
fceef393a   Al Viro   switch ->get_link...
3071
3072
  				  struct inode *inode,
  				  struct delayed_call *done)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3073
  {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3074
  	struct page *page = NULL;
6b2553918   Al Viro   replace ->follow_...
3075
  	int error;
6a6c99049   Al Viro   teach shmem_get_l...
3076
3077
3078
3079
3080
3081
3082
3083
3084
  	if (!dentry) {
  		page = find_get_page(inode->i_mapping, 0);
  		if (!page)
  			return ERR_PTR(-ECHILD);
  		if (!PageUptodate(page)) {
  			put_page(page);
  			return ERR_PTR(-ECHILD);
  		}
  	} else {
9e18eb293   Andres Lagar-Cavilla   tmpfs: mem_cgroup...
3085
  		error = shmem_getpage(inode, 0, &page, SGP_READ);
6a6c99049   Al Viro   teach shmem_get_l...
3086
3087
3088
3089
  		if (error)
  			return ERR_PTR(error);
  		unlock_page(page);
  	}
fceef393a   Al Viro   switch ->get_link...
3090
  	set_delayed_call(done, shmem_put_link, page);
21fc61c73   Al Viro   don't put symlink...
3091
  	return page_address(page);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3092
  }
b09e0fa4b   Eric Paris   tmpfs: implement ...
3093
  #ifdef CONFIG_TMPFS_XATTR
467118102   Randy Dunlap   mm/shmem and tiny...
3094
  /*
b09e0fa4b   Eric Paris   tmpfs: implement ...
3095
3096
   * Superblocks without xattr inode operations may get some security.* xattr
   * support from the LSM "for free". As soon as we have any other xattrs
39f0247d3   Andreas Gruenbacher   [PATCH] Access Co...
3097
3098
3099
   * like ACLs, we also need to implement the security.* handlers at
   * filesystem level, though.
   */
6d9d88d07   Jarkko Sakkinen   tmpfs: security x...
3100
  /*
6d9d88d07   Jarkko Sakkinen   tmpfs: security x...
3101
3102
3103
3104
3105
3106
3107
3108
   * Callback for security_inode_init_security() for acquiring xattrs.
   */
  static int shmem_initxattrs(struct inode *inode,
  			    const struct xattr *xattr_array,
  			    void *fs_info)
  {
  	struct shmem_inode_info *info = SHMEM_I(inode);
  	const struct xattr *xattr;
38f386574   Aristeu Rozanski   xattr: extract si...
3109
  	struct simple_xattr *new_xattr;
6d9d88d07   Jarkko Sakkinen   tmpfs: security x...
3110
3111
3112
  	size_t len;
  
  	for (xattr = xattr_array; xattr->name != NULL; xattr++) {
38f386574   Aristeu Rozanski   xattr: extract si...
3113
  		new_xattr = simple_xattr_alloc(xattr->value, xattr->value_len);
6d9d88d07   Jarkko Sakkinen   tmpfs: security x...
3114
3115
3116
3117
3118
3119
3120
  		if (!new_xattr)
  			return -ENOMEM;
  
  		len = strlen(xattr->name) + 1;
  		new_xattr->name = kmalloc(XATTR_SECURITY_PREFIX_LEN + len,
  					  GFP_KERNEL);
  		if (!new_xattr->name) {
3bef735ad   Chengguang Xu   vfs/xattr: mm/shm...
3121
  			kvfree(new_xattr);
6d9d88d07   Jarkko Sakkinen   tmpfs: security x...
3122
3123
3124
3125
3126
3127
3128
  			return -ENOMEM;
  		}
  
  		memcpy(new_xattr->name, XATTR_SECURITY_PREFIX,
  		       XATTR_SECURITY_PREFIX_LEN);
  		memcpy(new_xattr->name + XATTR_SECURITY_PREFIX_LEN,
  		       xattr->name, len);
38f386574   Aristeu Rozanski   xattr: extract si...
3129
  		simple_xattr_list_add(&info->xattrs, new_xattr);
6d9d88d07   Jarkko Sakkinen   tmpfs: security x...
3130
3131
3132
3133
  	}
  
  	return 0;
  }
aa7c5241c   Andreas Gruenbacher   tmpfs: Use xattr ...
3134
  static int shmem_xattr_handler_get(const struct xattr_handler *handler,
b296821a7   Al Viro   xattr_handler: pa...
3135
  				   struct dentry *unused, struct inode *inode,
3484eba91   Mark Salyzyn   FROMLIST: Add fla...
3136
3137
  				   const char *name, void *buffer, size_t size,
  				   int flags)
b09e0fa4b   Eric Paris   tmpfs: implement ...
3138
  {
b296821a7   Al Viro   xattr_handler: pa...
3139
  	struct shmem_inode_info *info = SHMEM_I(inode);
b09e0fa4b   Eric Paris   tmpfs: implement ...
3140

aa7c5241c   Andreas Gruenbacher   tmpfs: Use xattr ...
3141
  	name = xattr_full_name(handler, name);
38f386574   Aristeu Rozanski   xattr: extract si...
3142
  	return simple_xattr_get(&info->xattrs, name, buffer, size);
b09e0fa4b   Eric Paris   tmpfs: implement ...
3143
  }
aa7c5241c   Andreas Gruenbacher   tmpfs: Use xattr ...
3144
  static int shmem_xattr_handler_set(const struct xattr_handler *handler,
593012268   Al Viro   switch xattr_hand...
3145
3146
3147
  				   struct dentry *unused, struct inode *inode,
  				   const char *name, const void *value,
  				   size_t size, int flags)
b09e0fa4b   Eric Paris   tmpfs: implement ...
3148
  {
593012268   Al Viro   switch xattr_hand...
3149
  	struct shmem_inode_info *info = SHMEM_I(inode);
b09e0fa4b   Eric Paris   tmpfs: implement ...
3150

aa7c5241c   Andreas Gruenbacher   tmpfs: Use xattr ...
3151
  	name = xattr_full_name(handler, name);
a46a22955   Daniel Xu   kernfs: Add remov...
3152
  	return simple_xattr_set(&info->xattrs, name, value, size, flags, NULL);
b09e0fa4b   Eric Paris   tmpfs: implement ...
3153
  }
aa7c5241c   Andreas Gruenbacher   tmpfs: Use xattr ...
3154
3155
3156
3157
3158
  static const struct xattr_handler shmem_security_xattr_handler = {
  	.prefix = XATTR_SECURITY_PREFIX,
  	.get = shmem_xattr_handler_get,
  	.set = shmem_xattr_handler_set,
  };
b09e0fa4b   Eric Paris   tmpfs: implement ...
3159

aa7c5241c   Andreas Gruenbacher   tmpfs: Use xattr ...
3160
3161
3162
3163
3164
  static const struct xattr_handler shmem_trusted_xattr_handler = {
  	.prefix = XATTR_TRUSTED_PREFIX,
  	.get = shmem_xattr_handler_get,
  	.set = shmem_xattr_handler_set,
  };
b09e0fa4b   Eric Paris   tmpfs: implement ...
3165

aa7c5241c   Andreas Gruenbacher   tmpfs: Use xattr ...
3166
3167
3168
3169
3170
3171
3172
3173
3174
  static const struct xattr_handler *shmem_xattr_handlers[] = {
  #ifdef CONFIG_TMPFS_POSIX_ACL
  	&posix_acl_access_xattr_handler,
  	&posix_acl_default_xattr_handler,
  #endif
  	&shmem_security_xattr_handler,
  	&shmem_trusted_xattr_handler,
  	NULL
  };
b09e0fa4b   Eric Paris   tmpfs: implement ...
3175
3176
3177
  
  static ssize_t shmem_listxattr(struct dentry *dentry, char *buffer, size_t size)
  {
75c3cfa85   David Howells   VFS: assorted wei...
3178
  	struct shmem_inode_info *info = SHMEM_I(d_inode(dentry));
786534b92   Andreas Gruenbacher   tmpfs: listxattr ...
3179
  	return simple_xattr_list(d_inode(dentry), &info->xattrs, buffer, size);
b09e0fa4b   Eric Paris   tmpfs: implement ...
3180
3181
  }
  #endif /* CONFIG_TMPFS_XATTR */
69f07ec93   Hugh Dickins   tmpfs: use kmemdu...
3182
  static const struct inode_operations shmem_short_symlink_operations = {
6b2553918   Al Viro   replace ->follow_...
3183
  	.get_link	= simple_get_link,
b09e0fa4b   Eric Paris   tmpfs: implement ...
3184
  #ifdef CONFIG_TMPFS_XATTR
b09e0fa4b   Eric Paris   tmpfs: implement ...
3185
  	.listxattr	= shmem_listxattr,
b09e0fa4b   Eric Paris   tmpfs: implement ...
3186
3187
3188
3189
  #endif
  };
  
  static const struct inode_operations shmem_symlink_inode_operations = {
6b2553918   Al Viro   replace ->follow_...
3190
  	.get_link	= shmem_get_link,
b09e0fa4b   Eric Paris   tmpfs: implement ...
3191
  #ifdef CONFIG_TMPFS_XATTR
b09e0fa4b   Eric Paris   tmpfs: implement ...
3192
  	.listxattr	= shmem_listxattr,
39f0247d3   Andreas Gruenbacher   [PATCH] Access Co...
3193
  #endif
b09e0fa4b   Eric Paris   tmpfs: implement ...
3194
  };
39f0247d3   Andreas Gruenbacher   [PATCH] Access Co...
3195

91828a405   David M. Grimes   [PATCH] knfsd: ad...
3196
3197
3198
3199
3200
3201
3202
3203
3204
3205
3206
3207
  static struct dentry *shmem_get_parent(struct dentry *child)
  {
  	return ERR_PTR(-ESTALE);
  }
  
  static int shmem_match(struct inode *ino, void *vfh)
  {
  	__u32 *fh = vfh;
  	__u64 inum = fh[2];
  	inum = (inum << 32) | fh[1];
  	return ino->i_ino == inum && fh[0] == ino->i_generation;
  }
12ba780d6   Amir Goldstein   tmpfs: allow deco...
3208
3209
3210
3211
3212
3213
3214
  /* Find any alias of inode, but prefer a hashed alias */
  static struct dentry *shmem_find_alias(struct inode *inode)
  {
  	struct dentry *alias = d_find_alias(inode);
  
  	return alias ?: d_find_any_alias(inode);
  }
480b116c9   Christoph Hellwig   shmem: new export...
3215
3216
  static struct dentry *shmem_fh_to_dentry(struct super_block *sb,
  		struct fid *fid, int fh_len, int fh_type)
91828a405   David M. Grimes   [PATCH] knfsd: ad...
3217
  {
91828a405   David M. Grimes   [PATCH] knfsd: ad...
3218
  	struct inode *inode;
480b116c9   Christoph Hellwig   shmem: new export...
3219
  	struct dentry *dentry = NULL;
35c2a7f49   Hugh Dickins   tmpfs,ceph,gfs2,i...
3220
  	u64 inum;
480b116c9   Christoph Hellwig   shmem: new export...
3221
3222
3223
  
  	if (fh_len < 3)
  		return NULL;
91828a405   David M. Grimes   [PATCH] knfsd: ad...
3224

35c2a7f49   Hugh Dickins   tmpfs,ceph,gfs2,i...
3225
3226
  	inum = fid->raw[2];
  	inum = (inum << 32) | fid->raw[1];
480b116c9   Christoph Hellwig   shmem: new export...
3227
3228
  	inode = ilookup5(sb, (unsigned long)(inum + fid->raw[0]),
  			shmem_match, fid->raw);
91828a405   David M. Grimes   [PATCH] knfsd: ad...
3229
  	if (inode) {
12ba780d6   Amir Goldstein   tmpfs: allow deco...
3230
  		dentry = shmem_find_alias(inode);
91828a405   David M. Grimes   [PATCH] knfsd: ad...
3231
3232
  		iput(inode);
  	}
480b116c9   Christoph Hellwig   shmem: new export...
3233
  	return dentry;
91828a405   David M. Grimes   [PATCH] knfsd: ad...
3234
  }
b0b0382bb   Al Viro   ->encode_fh() API...
3235
3236
  static int shmem_encode_fh(struct inode *inode, __u32 *fh, int *len,
  				struct inode *parent)
91828a405   David M. Grimes   [PATCH] knfsd: ad...
3237
  {
5fe0c2378   Aneesh Kumar K.V   exportfs: Return ...
3238
3239
  	if (*len < 3) {
  		*len = 3;
94e07a759   Namjae Jeon   fs: encode_fh: re...
3240
  		return FILEID_INVALID;
5fe0c2378   Aneesh Kumar K.V   exportfs: Return ...
3241
  	}
91828a405   David M. Grimes   [PATCH] knfsd: ad...
3242

1d3382cbf   Al Viro   new helper: inode...
3243
  	if (inode_unhashed(inode)) {
91828a405   David M. Grimes   [PATCH] knfsd: ad...
3244
3245
3246
3247
3248
3249
3250
  		/* Unfortunately insert_inode_hash is not idempotent,
  		 * so as we hash inodes here rather than at creation
  		 * time, we need a lock to ensure we only try
  		 * to do it once
  		 */
  		static DEFINE_SPINLOCK(lock);
  		spin_lock(&lock);
1d3382cbf   Al Viro   new helper: inode...
3251
  		if (inode_unhashed(inode))
91828a405   David M. Grimes   [PATCH] knfsd: ad...
3252
3253
3254
3255
3256
3257
3258
3259
3260
3261
3262
3263
  			__insert_inode_hash(inode,
  					    inode->i_ino + inode->i_generation);
  		spin_unlock(&lock);
  	}
  
  	fh[0] = inode->i_generation;
  	fh[1] = inode->i_ino;
  	fh[2] = ((__u64)inode->i_ino) >> 32;
  
  	*len = 3;
  	return 1;
  }
396551644   Christoph Hellwig   exportfs: make st...
3264
  static const struct export_operations shmem_export_ops = {
91828a405   David M. Grimes   [PATCH] knfsd: ad...
3265
  	.get_parent     = shmem_get_parent,
91828a405   David M. Grimes   [PATCH] knfsd: ad...
3266
  	.encode_fh      = shmem_encode_fh,
480b116c9   Christoph Hellwig   shmem: new export...
3267
  	.fh_to_dentry	= shmem_fh_to_dentry,
91828a405   David M. Grimes   [PATCH] knfsd: ad...
3268
  };
626c3920a   Al Viro   shmem_parse_one()...
3269
3270
3271
3272
3273
3274
3275
3276
3277
  enum shmem_param {
  	Opt_gid,
  	Opt_huge,
  	Opt_mode,
  	Opt_mpol,
  	Opt_nr_blocks,
  	Opt_nr_inodes,
  	Opt_size,
  	Opt_uid,
ea3271f71   Chris Down   tmpfs: support 64...
3278
3279
  	Opt_inode32,
  	Opt_inode64,
626c3920a   Al Viro   shmem_parse_one()...
3280
  };
5eede6252   Al Viro   fold struct fs_pa...
3281
  static const struct constant_table shmem_param_enums_huge[] = {
2710c957a   Al Viro   fs_parse: get rid...
3282
3283
3284
3285
  	{"never",	SHMEM_HUGE_NEVER },
  	{"always",	SHMEM_HUGE_ALWAYS },
  	{"within_size",	SHMEM_HUGE_WITHIN_SIZE },
  	{"advise",	SHMEM_HUGE_ADVISE },
2710c957a   Al Viro   fs_parse: get rid...
3286
3287
  	{}
  };
d7167b149   Al Viro   fs_parse: fold fs...
3288
  const struct fs_parameter_spec shmem_fs_parameters[] = {
626c3920a   Al Viro   shmem_parse_one()...
3289
  	fsparam_u32   ("gid",		Opt_gid),
2710c957a   Al Viro   fs_parse: get rid...
3290
  	fsparam_enum  ("huge",		Opt_huge,  shmem_param_enums_huge),
626c3920a   Al Viro   shmem_parse_one()...
3291
3292
3293
3294
3295
3296
  	fsparam_u32oct("mode",		Opt_mode),
  	fsparam_string("mpol",		Opt_mpol),
  	fsparam_string("nr_blocks",	Opt_nr_blocks),
  	fsparam_string("nr_inodes",	Opt_nr_inodes),
  	fsparam_string("size",		Opt_size),
  	fsparam_u32   ("uid",		Opt_uid),
ea3271f71   Chris Down   tmpfs: support 64...
3297
3298
  	fsparam_flag  ("inode32",	Opt_inode32),
  	fsparam_flag  ("inode64",	Opt_inode64),
626c3920a   Al Viro   shmem_parse_one()...
3299
3300
  	{}
  };
f32356261   David Howells   vfs: Convert ramf...
3301
  static int shmem_parse_one(struct fs_context *fc, struct fs_parameter *param)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3302
  {
f32356261   David Howells   vfs: Convert ramf...
3303
  	struct shmem_options *ctx = fc->fs_private;
626c3920a   Al Viro   shmem_parse_one()...
3304
3305
  	struct fs_parse_result result;
  	unsigned long long size;
e04dc423a   Al Viro   shmem_parse_optio...
3306
  	char *rest;
626c3920a   Al Viro   shmem_parse_one()...
3307
  	int opt;
d7167b149   Al Viro   fs_parse: fold fs...
3308
  	opt = fs_parse(fc, shmem_fs_parameters, param, &result);
f32356261   David Howells   vfs: Convert ramf...
3309
  	if (opt < 0)
626c3920a   Al Viro   shmem_parse_one()...
3310
  		return opt;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3311

626c3920a   Al Viro   shmem_parse_one()...
3312
3313
3314
  	switch (opt) {
  	case Opt_size:
  		size = memparse(param->string, &rest);
e04dc423a   Al Viro   shmem_parse_optio...
3315
3316
3317
3318
3319
3320
3321
  		if (*rest == '%') {
  			size <<= PAGE_SHIFT;
  			size *= totalram_pages();
  			do_div(size, 100);
  			rest++;
  		}
  		if (*rest)
626c3920a   Al Viro   shmem_parse_one()...
3322
  			goto bad_value;
e04dc423a   Al Viro   shmem_parse_optio...
3323
3324
  		ctx->blocks = DIV_ROUND_UP(size, PAGE_SIZE);
  		ctx->seen |= SHMEM_SEEN_BLOCKS;
626c3920a   Al Viro   shmem_parse_one()...
3325
3326
3327
  		break;
  	case Opt_nr_blocks:
  		ctx->blocks = memparse(param->string, &rest);
e04dc423a   Al Viro   shmem_parse_optio...
3328
  		if (*rest)
626c3920a   Al Viro   shmem_parse_one()...
3329
  			goto bad_value;
e04dc423a   Al Viro   shmem_parse_optio...
3330
  		ctx->seen |= SHMEM_SEEN_BLOCKS;
626c3920a   Al Viro   shmem_parse_one()...
3331
3332
3333
  		break;
  	case Opt_nr_inodes:
  		ctx->inodes = memparse(param->string, &rest);
e04dc423a   Al Viro   shmem_parse_optio...
3334
  		if (*rest)
626c3920a   Al Viro   shmem_parse_one()...
3335
  			goto bad_value;
e04dc423a   Al Viro   shmem_parse_optio...
3336
  		ctx->seen |= SHMEM_SEEN_INODES;
626c3920a   Al Viro   shmem_parse_one()...
3337
3338
3339
3340
3341
3342
  		break;
  	case Opt_mode:
  		ctx->mode = result.uint_32 & 07777;
  		break;
  	case Opt_uid:
  		ctx->uid = make_kuid(current_user_ns(), result.uint_32);
e04dc423a   Al Viro   shmem_parse_optio...
3343
  		if (!uid_valid(ctx->uid))
626c3920a   Al Viro   shmem_parse_one()...
3344
3345
3346
3347
  			goto bad_value;
  		break;
  	case Opt_gid:
  		ctx->gid = make_kgid(current_user_ns(), result.uint_32);
e04dc423a   Al Viro   shmem_parse_optio...
3348
  		if (!gid_valid(ctx->gid))
626c3920a   Al Viro   shmem_parse_one()...
3349
3350
3351
3352
3353
  			goto bad_value;
  		break;
  	case Opt_huge:
  		ctx->huge = result.uint_32;
  		if (ctx->huge != SHMEM_HUGE_NEVER &&
396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
3354
  		    !(IS_ENABLED(CONFIG_TRANSPARENT_HUGEPAGE) &&
626c3920a   Al Viro   shmem_parse_one()...
3355
3356
  		      has_transparent_hugepage()))
  			goto unsupported_parameter;
e04dc423a   Al Viro   shmem_parse_optio...
3357
  		ctx->seen |= SHMEM_SEEN_HUGE;
626c3920a   Al Viro   shmem_parse_one()...
3358
3359
3360
3361
3362
3363
3364
3365
3366
3367
  		break;
  	case Opt_mpol:
  		if (IS_ENABLED(CONFIG_NUMA)) {
  			mpol_put(ctx->mpol);
  			ctx->mpol = NULL;
  			if (mpol_parse_str(param->string, &ctx->mpol))
  				goto bad_value;
  			break;
  		}
  		goto unsupported_parameter;
ea3271f71   Chris Down   tmpfs: support 64...
3368
3369
3370
3371
3372
3373
3374
3375
3376
3377
3378
3379
3380
  	case Opt_inode32:
  		ctx->full_inums = false;
  		ctx->seen |= SHMEM_SEEN_INUMS;
  		break;
  	case Opt_inode64:
  		if (sizeof(ino_t) < 8) {
  			return invalfc(fc,
  				       "Cannot use inode64 with <64bit inums in kernel
  ");
  		}
  		ctx->full_inums = true;
  		ctx->seen |= SHMEM_SEEN_INUMS;
  		break;
e04dc423a   Al Viro   shmem_parse_optio...
3381
3382
  	}
  	return 0;
626c3920a   Al Viro   shmem_parse_one()...
3383
  unsupported_parameter:
f35aa2bc8   Al Viro   tmpfs: switch to ...
3384
  	return invalfc(fc, "Unsupported parameter '%s'", param->key);
626c3920a   Al Viro   shmem_parse_one()...
3385
  bad_value:
f35aa2bc8   Al Viro   tmpfs: switch to ...
3386
  	return invalfc(fc, "Bad value for '%s'", param->key);
e04dc423a   Al Viro   shmem_parse_optio...
3387
  }
f32356261   David Howells   vfs: Convert ramf...
3388
  static int shmem_parse_options(struct fs_context *fc, void *data)
e04dc423a   Al Viro   shmem_parse_optio...
3389
  {
f32356261   David Howells   vfs: Convert ramf...
3390
  	char *options = data;
33f37c648   Al Viro   shmem: fix LSM op...
3391
3392
3393
3394
3395
  	if (options) {
  		int err = security_sb_eat_lsm_opts(options, &fc->security);
  		if (err)
  			return err;
  	}
b00dc3ad7   Hugh Dickins   [PATCH] tmpfs: fi...
3396
  	while (options != NULL) {
626c3920a   Al Viro   shmem_parse_one()...
3397
  		char *this_char = options;
b00dc3ad7   Hugh Dickins   [PATCH] tmpfs: fi...
3398
3399
3400
3401
3402
3403
3404
3405
3406
3407
3408
3409
3410
3411
3412
  		for (;;) {
  			/*
  			 * NUL-terminate this option: unfortunately,
  			 * mount options form a comma-separated list,
  			 * but mpol's nodelist may also contain commas.
  			 */
  			options = strchr(options, ',');
  			if (options == NULL)
  				break;
  			options++;
  			if (!isdigit(*options)) {
  				options[-1] = '\0';
  				break;
  			}
  		}
626c3920a   Al Viro   shmem_parse_one()...
3413
3414
  		if (*this_char) {
  			char *value = strchr(this_char,'=');
f32356261   David Howells   vfs: Convert ramf...
3415
  			size_t len = 0;
626c3920a   Al Viro   shmem_parse_one()...
3416
3417
3418
3419
  			int err;
  
  			if (value) {
  				*value++ = '\0';
f32356261   David Howells   vfs: Convert ramf...
3420
  				len = strlen(value);
626c3920a   Al Viro   shmem_parse_one()...
3421
  			}
f32356261   David Howells   vfs: Convert ramf...
3422
3423
3424
  			err = vfs_parse_fs_string(fc, this_char, value, len);
  			if (err < 0)
  				return err;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3425
  		}
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3426
3427
  	}
  	return 0;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3428
  }
f32356261   David Howells   vfs: Convert ramf...
3429
3430
3431
3432
3433
3434
3435
3436
  /*
   * Reconfigure a shmem filesystem.
   *
   * Note that we disallow change from limited->unlimited blocks/inodes while any
   * are in use; but we must separately disallow unlimited->limited, because in
   * that case we have no record of how much is already in use.
   */
  static int shmem_reconfigure(struct fs_context *fc)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3437
  {
f32356261   David Howells   vfs: Convert ramf...
3438
3439
  	struct shmem_options *ctx = fc->fs_private;
  	struct shmem_sb_info *sbinfo = SHMEM_SB(fc->root->d_sb);
0edd73b33   Hugh Dickins   [PATCH] shmem: re...
3440
  	unsigned long inodes;
f32356261   David Howells   vfs: Convert ramf...
3441
  	const char *err;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3442

0edd73b33   Hugh Dickins   [PATCH] shmem: re...
3443
  	spin_lock(&sbinfo->stat_lock);
0edd73b33   Hugh Dickins   [PATCH] shmem: re...
3444
  	inodes = sbinfo->max_inodes - sbinfo->free_inodes;
f32356261   David Howells   vfs: Convert ramf...
3445
3446
3447
  	if ((ctx->seen & SHMEM_SEEN_BLOCKS) && ctx->blocks) {
  		if (!sbinfo->max_blocks) {
  			err = "Cannot retroactively limit size";
0b5071dd3   Al Viro   shmem_parse_optio...
3448
  			goto out;
f32356261   David Howells   vfs: Convert ramf...
3449
  		}
0b5071dd3   Al Viro   shmem_parse_optio...
3450
  		if (percpu_counter_compare(&sbinfo->used_blocks,
f32356261   David Howells   vfs: Convert ramf...
3451
3452
  					   ctx->blocks) > 0) {
  			err = "Too small a size for current use";
0b5071dd3   Al Viro   shmem_parse_optio...
3453
  			goto out;
f32356261   David Howells   vfs: Convert ramf...
3454
  		}
0b5071dd3   Al Viro   shmem_parse_optio...
3455
  	}
f32356261   David Howells   vfs: Convert ramf...
3456
3457
3458
  	if ((ctx->seen & SHMEM_SEEN_INODES) && ctx->inodes) {
  		if (!sbinfo->max_inodes) {
  			err = "Cannot retroactively limit inodes";
0b5071dd3   Al Viro   shmem_parse_optio...
3459
  			goto out;
f32356261   David Howells   vfs: Convert ramf...
3460
3461
3462
  		}
  		if (ctx->inodes < inodes) {
  			err = "Too few inodes for current use";
0b5071dd3   Al Viro   shmem_parse_optio...
3463
  			goto out;
f32356261   David Howells   vfs: Convert ramf...
3464
  		}
0b5071dd3   Al Viro   shmem_parse_optio...
3465
  	}
0edd73b33   Hugh Dickins   [PATCH] shmem: re...
3466

ea3271f71   Chris Down   tmpfs: support 64...
3467
3468
3469
3470
3471
  	if ((ctx->seen & SHMEM_SEEN_INUMS) && !ctx->full_inums &&
  	    sbinfo->next_ino > UINT_MAX) {
  		err = "Current inum too high to switch to 32-bit inums";
  		goto out;
  	}
f32356261   David Howells   vfs: Convert ramf...
3472
3473
  	if (ctx->seen & SHMEM_SEEN_HUGE)
  		sbinfo->huge = ctx->huge;
ea3271f71   Chris Down   tmpfs: support 64...
3474
3475
  	if (ctx->seen & SHMEM_SEEN_INUMS)
  		sbinfo->full_inums = ctx->full_inums;
f32356261   David Howells   vfs: Convert ramf...
3476
3477
3478
3479
3480
  	if (ctx->seen & SHMEM_SEEN_BLOCKS)
  		sbinfo->max_blocks  = ctx->blocks;
  	if (ctx->seen & SHMEM_SEEN_INODES) {
  		sbinfo->max_inodes  = ctx->inodes;
  		sbinfo->free_inodes = ctx->inodes - inodes;
0b5071dd3   Al Viro   shmem_parse_optio...
3481
  	}
71fe804b6   Lee Schermerhorn   mempolicy: use st...
3482

5f00110f7   Greg Thelen   tmpfs: fix use-af...
3483
3484
3485
  	/*
  	 * Preserve previous mempolicy unless mpol remount option was specified.
  	 */
f32356261   David Howells   vfs: Convert ramf...
3486
  	if (ctx->mpol) {
5f00110f7   Greg Thelen   tmpfs: fix use-af...
3487
  		mpol_put(sbinfo->mpol);
f32356261   David Howells   vfs: Convert ramf...
3488
3489
  		sbinfo->mpol = ctx->mpol;	/* transfers initial ref */
  		ctx->mpol = NULL;
5f00110f7   Greg Thelen   tmpfs: fix use-af...
3490
  	}
f32356261   David Howells   vfs: Convert ramf...
3491
3492
  	spin_unlock(&sbinfo->stat_lock);
  	return 0;
0edd73b33   Hugh Dickins   [PATCH] shmem: re...
3493
3494
  out:
  	spin_unlock(&sbinfo->stat_lock);
f35aa2bc8   Al Viro   tmpfs: switch to ...
3495
  	return invalfc(fc, "%s", err);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3496
  }
680d794ba   akpm@linux-foundation.org   mount options: fi...
3497

34c80b1d9   Al Viro   vfs: switch ->sho...
3498
  static int shmem_show_options(struct seq_file *seq, struct dentry *root)
680d794ba   akpm@linux-foundation.org   mount options: fi...
3499
  {
34c80b1d9   Al Viro   vfs: switch ->sho...
3500
  	struct shmem_sb_info *sbinfo = SHMEM_SB(root->d_sb);
680d794ba   akpm@linux-foundation.org   mount options: fi...
3501
3502
3503
  
  	if (sbinfo->max_blocks != shmem_default_max_blocks())
  		seq_printf(seq, ",size=%luk",
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
3504
  			sbinfo->max_blocks << (PAGE_SHIFT - 10));
680d794ba   akpm@linux-foundation.org   mount options: fi...
3505
3506
  	if (sbinfo->max_inodes != shmem_default_max_inodes())
  		seq_printf(seq, ",nr_inodes=%lu", sbinfo->max_inodes);
0825a6f98   Joe Perches   mm: use octal not...
3507
  	if (sbinfo->mode != (0777 | S_ISVTX))
09208d150   Al Viro   shmem, ramfs: pro...
3508
  		seq_printf(seq, ",mode=%03ho", sbinfo->mode);
8751e0395   Eric W. Biederman   userns: Convert t...
3509
3510
3511
3512
3513
3514
  	if (!uid_eq(sbinfo->uid, GLOBAL_ROOT_UID))
  		seq_printf(seq, ",uid=%u",
  				from_kuid_munged(&init_user_ns, sbinfo->uid));
  	if (!gid_eq(sbinfo->gid, GLOBAL_ROOT_GID))
  		seq_printf(seq, ",gid=%u",
  				from_kgid_munged(&init_user_ns, sbinfo->gid));
ea3271f71   Chris Down   tmpfs: support 64...
3515
3516
3517
3518
3519
3520
3521
3522
3523
3524
3525
3526
3527
3528
3529
3530
3531
3532
3533
3534
3535
3536
3537
  
  	/*
  	 * Showing inode{64,32} might be useful even if it's the system default,
  	 * since then people don't have to resort to checking both here and
  	 * /proc/config.gz to confirm 64-bit inums were successfully applied
  	 * (which may not even exist if IKCONFIG_PROC isn't enabled).
  	 *
  	 * We hide it when inode64 isn't the default and we are using 32-bit
  	 * inodes, since that probably just means the feature isn't even under
  	 * consideration.
  	 *
  	 * As such:
  	 *
  	 *                     +-----------------+-----------------+
  	 *                     | TMPFS_INODE64=y | TMPFS_INODE64=n |
  	 *  +------------------+-----------------+-----------------+
  	 *  | full_inums=true  | show            | show            |
  	 *  | full_inums=false | show            | hide            |
  	 *  +------------------+-----------------+-----------------+
  	 *
  	 */
  	if (IS_ENABLED(CONFIG_TMPFS_INODE64) || sbinfo->full_inums)
  		seq_printf(seq, ",inode%d", (sbinfo->full_inums ? 64 : 32));
396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
3538
  #ifdef CONFIG_TRANSPARENT_HUGEPAGE
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
3539
3540
3541
3542
  	/* Rightly or wrongly, show huge mount option unmasked by shmem_huge */
  	if (sbinfo->huge)
  		seq_printf(seq, ",huge=%s", shmem_format_huge(sbinfo->huge));
  #endif
71fe804b6   Lee Schermerhorn   mempolicy: use st...
3543
  	shmem_show_mpol(seq, sbinfo->mpol);
680d794ba   akpm@linux-foundation.org   mount options: fi...
3544
3545
  	return 0;
  }
9183df25f   David Herrmann   shm: add memfd_cr...
3546

680d794ba   akpm@linux-foundation.org   mount options: fi...
3547
  #endif /* CONFIG_TMPFS */
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3548
3549
3550
  
  static void shmem_put_super(struct super_block *sb)
  {
602586a83   Hugh Dickins   shmem: put_super ...
3551
  	struct shmem_sb_info *sbinfo = SHMEM_SB(sb);
e809d5f0b   Chris Down   tmpfs: per-superb...
3552
  	free_percpu(sbinfo->ino_batch);
602586a83   Hugh Dickins   shmem: put_super ...
3553
  	percpu_counter_destroy(&sbinfo->used_blocks);
49cd0a5c2   Greg Thelen   tmpfs: fix mempol...
3554
  	mpol_put(sbinfo->mpol);
602586a83   Hugh Dickins   shmem: put_super ...
3555
  	kfree(sbinfo);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3556
3557
  	sb->s_fs_info = NULL;
  }
f32356261   David Howells   vfs: Convert ramf...
3558
  static int shmem_fill_super(struct super_block *sb, struct fs_context *fc)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3559
  {
f32356261   David Howells   vfs: Convert ramf...
3560
  	struct shmem_options *ctx = fc->fs_private;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3561
  	struct inode *inode;
0edd73b33   Hugh Dickins   [PATCH] shmem: re...
3562
  	struct shmem_sb_info *sbinfo;
680d794ba   akpm@linux-foundation.org   mount options: fi...
3563
3564
3565
  	int err = -ENOMEM;
  
  	/* Round up to L1_CACHE_BYTES to resist false sharing */
425fbf047   Pekka Enberg   shmem: initialize...
3566
  	sbinfo = kzalloc(max((int)sizeof(struct shmem_sb_info),
680d794ba   akpm@linux-foundation.org   mount options: fi...
3567
3568
3569
  				L1_CACHE_BYTES), GFP_KERNEL);
  	if (!sbinfo)
  		return -ENOMEM;
680d794ba   akpm@linux-foundation.org   mount options: fi...
3570
  	sb->s_fs_info = sbinfo;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3571

0edd73b33   Hugh Dickins   [PATCH] shmem: re...
3572
  #ifdef CONFIG_TMPFS
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3573
3574
3575
3576
3577
  	/*
  	 * Per default we only allow half of the physical ram per
  	 * tmpfs instance, limiting inodes to one per page of lowmem;
  	 * but the internal instance is left unlimited.
  	 */
1751e8a6c   Linus Torvalds   Rename superblock...
3578
  	if (!(sb->s_flags & SB_KERNMOUNT)) {
f32356261   David Howells   vfs: Convert ramf...
3579
3580
3581
3582
  		if (!(ctx->seen & SHMEM_SEEN_BLOCKS))
  			ctx->blocks = shmem_default_max_blocks();
  		if (!(ctx->seen & SHMEM_SEEN_INODES))
  			ctx->inodes = shmem_default_max_inodes();
ea3271f71   Chris Down   tmpfs: support 64...
3583
3584
  		if (!(ctx->seen & SHMEM_SEEN_INUMS))
  			ctx->full_inums = IS_ENABLED(CONFIG_TMPFS_INODE64);
ca4e05195   Al Viro   shm_mnt is as lon...
3585
  	} else {
1751e8a6c   Linus Torvalds   Rename superblock...
3586
  		sb->s_flags |= SB_NOUSER;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3587
  	}
91828a405   David M. Grimes   [PATCH] knfsd: ad...
3588
  	sb->s_export_op = &shmem_export_ops;
1751e8a6c   Linus Torvalds   Rename superblock...
3589
  	sb->s_flags |= SB_NOSEC;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3590
  #else
1751e8a6c   Linus Torvalds   Rename superblock...
3591
  	sb->s_flags |= SB_NOUSER;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3592
  #endif
f32356261   David Howells   vfs: Convert ramf...
3593
3594
  	sbinfo->max_blocks = ctx->blocks;
  	sbinfo->free_inodes = sbinfo->max_inodes = ctx->inodes;
e809d5f0b   Chris Down   tmpfs: per-superb...
3595
3596
3597
3598
3599
  	if (sb->s_flags & SB_KERNMOUNT) {
  		sbinfo->ino_batch = alloc_percpu(ino_t);
  		if (!sbinfo->ino_batch)
  			goto failed;
  	}
f32356261   David Howells   vfs: Convert ramf...
3600
3601
  	sbinfo->uid = ctx->uid;
  	sbinfo->gid = ctx->gid;
ea3271f71   Chris Down   tmpfs: support 64...
3602
  	sbinfo->full_inums = ctx->full_inums;
f32356261   David Howells   vfs: Convert ramf...
3603
3604
3605
3606
  	sbinfo->mode = ctx->mode;
  	sbinfo->huge = ctx->huge;
  	sbinfo->mpol = ctx->mpol;
  	ctx->mpol = NULL;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3607

0edd73b33   Hugh Dickins   [PATCH] shmem: re...
3608
  	spin_lock_init(&sbinfo->stat_lock);
908c7f194   Tejun Heo   percpu_counter: a...
3609
  	if (percpu_counter_init(&sbinfo->used_blocks, 0, GFP_KERNEL))
602586a83   Hugh Dickins   shmem: put_super ...
3610
  		goto failed;
779750d20   Kirill A. Shutemov   shmem: split huge...
3611
3612
  	spin_lock_init(&sbinfo->shrinklist_lock);
  	INIT_LIST_HEAD(&sbinfo->shrinklist);
0edd73b33   Hugh Dickins   [PATCH] shmem: re...
3613

285b2c4fd   Hugh Dickins   tmpfs: demolish o...
3614
  	sb->s_maxbytes = MAX_LFS_FILESIZE;
09cbfeaf1   Kirill A. Shutemov   mm, fs: get rid o...
3615
3616
  	sb->s_blocksize = PAGE_SIZE;
  	sb->s_blocksize_bits = PAGE_SHIFT;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3617
3618
  	sb->s_magic = TMPFS_MAGIC;
  	sb->s_op = &shmem_ops;
cfd95a9cf   Robin H. Johnson   [PATCH] tmpfs: ti...
3619
  	sb->s_time_gran = 1;
b09e0fa4b   Eric Paris   tmpfs: implement ...
3620
  #ifdef CONFIG_TMPFS_XATTR
39f0247d3   Andreas Gruenbacher   [PATCH] Access Co...
3621
  	sb->s_xattr = shmem_xattr_handlers;
b09e0fa4b   Eric Paris   tmpfs: implement ...
3622
3623
  #endif
  #ifdef CONFIG_TMPFS_POSIX_ACL
1751e8a6c   Linus Torvalds   Rename superblock...
3624
  	sb->s_flags |= SB_POSIXACL;
39f0247d3   Andreas Gruenbacher   [PATCH] Access Co...
3625
  #endif
2b4db7961   Amir Goldstein   tmpfs: generate r...
3626
  	uuid_gen(&sb->s_uuid);
0edd73b33   Hugh Dickins   [PATCH] shmem: re...
3627

454abafe9   Dmitry Monakhov   ramfs: replace in...
3628
  	inode = shmem_get_inode(sb, NULL, S_IFDIR | sbinfo->mode, 0, VM_NORESERVE);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3629
3630
  	if (!inode)
  		goto failed;
680d794ba   akpm@linux-foundation.org   mount options: fi...
3631
3632
  	inode->i_uid = sbinfo->uid;
  	inode->i_gid = sbinfo->gid;
318ceed08   Al Viro   tidy up after d_m...
3633
3634
  	sb->s_root = d_make_root(inode);
  	if (!sb->s_root)
48fde701a   Al Viro   switch open-coded...
3635
  		goto failed;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3636
  	return 0;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3637
3638
3639
3640
  failed:
  	shmem_put_super(sb);
  	return err;
  }
f32356261   David Howells   vfs: Convert ramf...
3641
3642
3643
3644
3645
3646
3647
3648
3649
3650
3651
3652
3653
3654
3655
3656
3657
3658
3659
3660
3661
3662
3663
3664
  static int shmem_get_tree(struct fs_context *fc)
  {
  	return get_tree_nodev(fc, shmem_fill_super);
  }
  
  static void shmem_free_fc(struct fs_context *fc)
  {
  	struct shmem_options *ctx = fc->fs_private;
  
  	if (ctx) {
  		mpol_put(ctx->mpol);
  		kfree(ctx);
  	}
  }
  
  static const struct fs_context_operations shmem_fs_context_ops = {
  	.free			= shmem_free_fc,
  	.get_tree		= shmem_get_tree,
  #ifdef CONFIG_TMPFS
  	.parse_monolithic	= shmem_parse_options,
  	.parse_param		= shmem_parse_one,
  	.reconfigure		= shmem_reconfigure,
  #endif
  };
fcc234f88   Pekka Enberg   [PATCH] mm: kill ...
3665
  static struct kmem_cache *shmem_inode_cachep;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3666
3667
3668
  
  static struct inode *shmem_alloc_inode(struct super_block *sb)
  {
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
3669
3670
3671
  	struct shmem_inode_info *info;
  	info = kmem_cache_alloc(shmem_inode_cachep, GFP_KERNEL);
  	if (!info)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3672
  		return NULL;
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
3673
  	return &info->vfs_inode;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3674
  }
74b1da564   Al Viro   shmem: make use o...
3675
  static void shmem_free_in_core_inode(struct inode *inode)
fa0d7e3de   Nick Piggin   fs: icache RCU fr...
3676
  {
84e710da2   Al Viro   parallel lookups ...
3677
3678
  	if (S_ISLNK(inode->i_mode))
  		kfree(inode->i_link);
fa0d7e3de   Nick Piggin   fs: icache RCU fr...
3679
3680
  	kmem_cache_free(shmem_inode_cachep, SHMEM_I(inode));
  }
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3681
3682
  static void shmem_destroy_inode(struct inode *inode)
  {
09208d150   Al Viro   shmem, ramfs: pro...
3683
  	if (S_ISREG(inode->i_mode))
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3684
  		mpol_free_shared_policy(&SHMEM_I(inode)->policy);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3685
  }
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
3686
  static void shmem_init_inode(void *foo)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3687
  {
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
3688
3689
  	struct shmem_inode_info *info = foo;
  	inode_init_once(&info->vfs_inode);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3690
  }
9a8ec03ed   weiping zhang   shmem: convert sh...
3691
  static void shmem_init_inodecache(void)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3692
3693
3694
  {
  	shmem_inode_cachep = kmem_cache_create("shmem_inode_cache",
  				sizeof(struct shmem_inode_info),
5d097056c   Vladimir Davydov   kmemcg: account c...
3695
  				0, SLAB_PANIC|SLAB_ACCOUNT, shmem_init_inode);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3696
  }
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
3697
  static void shmem_destroy_inodecache(void)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3698
  {
1a1d92c10   Alexey Dobriyan   [PATCH] Really ig...
3699
  	kmem_cache_destroy(shmem_inode_cachep);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3700
  }
f5e54d6e5   Christoph Hellwig   [PATCH] mark addr...
3701
  static const struct address_space_operations shmem_aops = {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3702
  	.writepage	= shmem_writepage,
767193253   Ken Chen   [PATCH] simplify ...
3703
  	.set_page_dirty	= __set_page_dirty_no_writeback,
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3704
  #ifdef CONFIG_TMPFS
800d15a53   Nick Piggin   implement simple ...
3705
3706
  	.write_begin	= shmem_write_begin,
  	.write_end	= shmem_write_end,
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3707
  #endif
1c93923cc   Andrew Morton   include/linux/mig...
3708
  #ifdef CONFIG_MIGRATION
304dbdb7a   Lee Schermerhorn   [PATCH] add migra...
3709
  	.migratepage	= migrate_page,
1c93923cc   Andrew Morton   include/linux/mig...
3710
  #endif
aa261f549   Andi Kleen   HWPOISON: Enable ...
3711
  	.error_remove_page = generic_error_remove_page,
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3712
  };
15ad7cdcf   Helge Deller   [PATCH] struct se...
3713
  static const struct file_operations shmem_file_operations = {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3714
  	.mmap		= shmem_mmap,
c01d5b300   Hugh Dickins   shmem: get_unmapp...
3715
  	.get_unmapped_area = shmem_get_unmapped_area,
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3716
  #ifdef CONFIG_TMPFS
220f2ac91   Hugh Dickins   tmpfs: support SE...
3717
  	.llseek		= shmem_file_llseek,
2ba5bbed0   Al Viro   shmem: switch to ...
3718
  	.read_iter	= shmem_file_read_iter,
8174202b3   Al Viro   write_iter varian...
3719
  	.write_iter	= generic_file_write_iter,
1b061d924   Christoph Hellwig   rename the generi...
3720
  	.fsync		= noop_fsync,
82c156f85   Al Viro   switch generic_fi...
3721
  	.splice_read	= generic_file_splice_read,
f6cb85d00   Al Viro   shmem: switch to ...
3722
  	.splice_write	= iter_file_splice_write,
83e4fa9c1   Hugh Dickins   tmpfs: support fa...
3723
  	.fallocate	= shmem_fallocate,
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3724
3725
  #endif
  };
92e1d5be9   Arjan van de Ven   [PATCH] mark stru...
3726
  static const struct inode_operations shmem_inode_operations = {
44a30220b   Yu Zhao   shmem: recalculat...
3727
  	.getattr	= shmem_getattr,
94c1e62df   Hugh Dickins   tmpfs: take contr...
3728
  	.setattr	= shmem_setattr,
b09e0fa4b   Eric Paris   tmpfs: implement ...
3729
  #ifdef CONFIG_TMPFS_XATTR
b09e0fa4b   Eric Paris   tmpfs: implement ...
3730
  	.listxattr	= shmem_listxattr,
feda821e7   Christoph Hellwig   fs: remove generi...
3731
  	.set_acl	= simple_set_acl,
b09e0fa4b   Eric Paris   tmpfs: implement ...
3732
  #endif
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3733
  };
92e1d5be9   Arjan van de Ven   [PATCH] mark stru...
3734
  static const struct inode_operations shmem_dir_inode_operations = {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3735
3736
3737
3738
3739
3740
3741
3742
3743
  #ifdef CONFIG_TMPFS
  	.create		= shmem_create,
  	.lookup		= simple_lookup,
  	.link		= shmem_link,
  	.unlink		= shmem_unlink,
  	.symlink	= shmem_symlink,
  	.mkdir		= shmem_mkdir,
  	.rmdir		= shmem_rmdir,
  	.mknod		= shmem_mknod,
2773bf00a   Miklos Szeredi   fs: rename "renam...
3744
  	.rename		= shmem_rename2,
60545d0d4   Al Viro   [O_TMPFILE] it's ...
3745
  	.tmpfile	= shmem_tmpfile,
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3746
  #endif
b09e0fa4b   Eric Paris   tmpfs: implement ...
3747
  #ifdef CONFIG_TMPFS_XATTR
b09e0fa4b   Eric Paris   tmpfs: implement ...
3748
  	.listxattr	= shmem_listxattr,
b09e0fa4b   Eric Paris   tmpfs: implement ...
3749
  #endif
39f0247d3   Andreas Gruenbacher   [PATCH] Access Co...
3750
  #ifdef CONFIG_TMPFS_POSIX_ACL
94c1e62df   Hugh Dickins   tmpfs: take contr...
3751
  	.setattr	= shmem_setattr,
feda821e7   Christoph Hellwig   fs: remove generi...
3752
  	.set_acl	= simple_set_acl,
39f0247d3   Andreas Gruenbacher   [PATCH] Access Co...
3753
3754
  #endif
  };
92e1d5be9   Arjan van de Ven   [PATCH] mark stru...
3755
  static const struct inode_operations shmem_special_inode_operations = {
b09e0fa4b   Eric Paris   tmpfs: implement ...
3756
  #ifdef CONFIG_TMPFS_XATTR
b09e0fa4b   Eric Paris   tmpfs: implement ...
3757
  	.listxattr	= shmem_listxattr,
b09e0fa4b   Eric Paris   tmpfs: implement ...
3758
  #endif
39f0247d3   Andreas Gruenbacher   [PATCH] Access Co...
3759
  #ifdef CONFIG_TMPFS_POSIX_ACL
94c1e62df   Hugh Dickins   tmpfs: take contr...
3760
  	.setattr	= shmem_setattr,
feda821e7   Christoph Hellwig   fs: remove generi...
3761
  	.set_acl	= simple_set_acl,
39f0247d3   Andreas Gruenbacher   [PATCH] Access Co...
3762
  #endif
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3763
  };
759b9775c   Hugh Dickins   [PATCH] shmem and...
3764
  static const struct super_operations shmem_ops = {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3765
  	.alloc_inode	= shmem_alloc_inode,
74b1da564   Al Viro   shmem: make use o...
3766
  	.free_inode	= shmem_free_in_core_inode,
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3767
3768
3769
  	.destroy_inode	= shmem_destroy_inode,
  #ifdef CONFIG_TMPFS
  	.statfs		= shmem_statfs,
680d794ba   akpm@linux-foundation.org   mount options: fi...
3770
  	.show_options	= shmem_show_options,
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3771
  #endif
1f895f75d   Al Viro   switch shmem.c to...
3772
  	.evict_inode	= shmem_evict_inode,
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3773
3774
  	.drop_inode	= generic_delete_inode,
  	.put_super	= shmem_put_super,
396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
3775
  #ifdef CONFIG_TRANSPARENT_HUGEPAGE
779750d20   Kirill A. Shutemov   shmem: split huge...
3776
3777
3778
  	.nr_cached_objects	= shmem_unused_huge_count,
  	.free_cached_objects	= shmem_unused_huge_scan,
  #endif
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3779
  };
f0f37e2f7   Alexey Dobriyan   const: mark struc...
3780
  static const struct vm_operations_struct shmem_vm_ops = {
54cb8821d   Nick Piggin   mm: merge populat...
3781
  	.fault		= shmem_fault,
d7c175517   Ning Qu   mm: implement ->m...
3782
  	.map_pages	= filemap_map_pages,
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3783
3784
3785
3786
3787
  #ifdef CONFIG_NUMA
  	.set_policy     = shmem_set_policy,
  	.get_policy     = shmem_get_policy,
  #endif
  };
f32356261   David Howells   vfs: Convert ramf...
3788
  int shmem_init_fs_context(struct fs_context *fc)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3789
  {
f32356261   David Howells   vfs: Convert ramf...
3790
3791
3792
3793
3794
3795
3796
3797
3798
3799
3800
3801
3802
  	struct shmem_options *ctx;
  
  	ctx = kzalloc(sizeof(struct shmem_options), GFP_KERNEL);
  	if (!ctx)
  		return -ENOMEM;
  
  	ctx->mode = 0777 | S_ISVTX;
  	ctx->uid = current_fsuid();
  	ctx->gid = current_fsgid();
  
  	fc->fs_private = ctx;
  	fc->ops = &shmem_fs_context_ops;
  	return 0;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3803
  }
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
3804
  static struct file_system_type shmem_fs_type = {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3805
3806
  	.owner		= THIS_MODULE,
  	.name		= "tmpfs",
f32356261   David Howells   vfs: Convert ramf...
3807
3808
  	.init_fs_context = shmem_init_fs_context,
  #ifdef CONFIG_TMPFS
d7167b149   Al Viro   fs_parse: fold fs...
3809
  	.parameters	= shmem_fs_parameters,
f32356261   David Howells   vfs: Convert ramf...
3810
  #endif
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3811
  	.kill_sb	= kill_litter_super,
01c702670   Matthew Wilcox (Oracle)   fs: add a filesys...
3812
  	.fs_flags	= FS_USERNS_MOUNT | FS_THP_SUPPORT,
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3813
  };
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3814

41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
3815
  int __init shmem_init(void)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3816
3817
  {
  	int error;
9a8ec03ed   weiping zhang   shmem: convert sh...
3818
  	shmem_init_inodecache();
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3819

41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
3820
  	error = register_filesystem(&shmem_fs_type);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3821
  	if (error) {
1170532bb   Joe Perches   mm: convert print...
3822
3823
  		pr_err("Could not register tmpfs
  ");
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3824
3825
  		goto out2;
  	}
95dc112a5   Greg Kroah-Hartman   [PATCH] devfs: Re...
3826

ca4e05195   Al Viro   shm_mnt is as lon...
3827
  	shm_mnt = kern_mount(&shmem_fs_type);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3828
3829
  	if (IS_ERR(shm_mnt)) {
  		error = PTR_ERR(shm_mnt);
1170532bb   Joe Perches   mm: convert print...
3830
3831
  		pr_err("Could not kern_mount tmpfs
  ");
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3832
3833
  		goto out1;
  	}
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
3834

396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
3835
  #ifdef CONFIG_TRANSPARENT_HUGEPAGE
435c0b87d   Kirill A. Shutemov   mm, shmem: fix ha...
3836
  	if (has_transparent_hugepage() && shmem_huge > SHMEM_HUGE_DENY)
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
3837
3838
3839
3840
  		SHMEM_SB(shm_mnt->mnt_sb)->huge = shmem_huge;
  	else
  		shmem_huge = 0; /* just in case it was patched */
  #endif
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3841
3842
3843
  	return 0;
  
  out1:
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
3844
  	unregister_filesystem(&shmem_fs_type);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3845
  out2:
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
3846
  	shmem_destroy_inodecache();
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
3847
3848
3849
  	shm_mnt = ERR_PTR(error);
  	return error;
  }
853ac43ab   Matt Mackall   shmem: unify regu...
3850

396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
3851
  #if defined(CONFIG_TRANSPARENT_HUGEPAGE) && defined(CONFIG_SYSFS)
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
3852
3853
3854
  static ssize_t shmem_enabled_show(struct kobject *kobj,
  		struct kobj_attribute *attr, char *buf)
  {
26083eb6b   Colin Ian King   mm/shmem.c: make ...
3855
  	static const int values[] = {
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
3856
3857
3858
3859
3860
3861
3862
3863
3864
3865
3866
3867
3868
3869
3870
3871
3872
3873
3874
3875
3876
3877
3878
3879
3880
3881
3882
3883
3884
3885
3886
3887
3888
3889
3890
3891
3892
3893
3894
3895
3896
3897
  		SHMEM_HUGE_ALWAYS,
  		SHMEM_HUGE_WITHIN_SIZE,
  		SHMEM_HUGE_ADVISE,
  		SHMEM_HUGE_NEVER,
  		SHMEM_HUGE_DENY,
  		SHMEM_HUGE_FORCE,
  	};
  	int i, count;
  
  	for (i = 0, count = 0; i < ARRAY_SIZE(values); i++) {
  		const char *fmt = shmem_huge == values[i] ? "[%s] " : "%s ";
  
  		count += sprintf(buf + count, fmt,
  				shmem_format_huge(values[i]));
  	}
  	buf[count - 1] = '
  ';
  	return count;
  }
  
  static ssize_t shmem_enabled_store(struct kobject *kobj,
  		struct kobj_attribute *attr, const char *buf, size_t count)
  {
  	char tmp[16];
  	int huge;
  
  	if (count + 1 > sizeof(tmp))
  		return -EINVAL;
  	memcpy(tmp, buf, count);
  	tmp[count] = '\0';
  	if (count && tmp[count - 1] == '
  ')
  		tmp[count - 1] = '\0';
  
  	huge = shmem_parse_huge(tmp);
  	if (huge == -EINVAL)
  		return -EINVAL;
  	if (!has_transparent_hugepage() &&
  			huge != SHMEM_HUGE_NEVER && huge != SHMEM_HUGE_DENY)
  		return -EINVAL;
  
  	shmem_huge = huge;
435c0b87d   Kirill A. Shutemov   mm, shmem: fix ha...
3898
  	if (shmem_huge > SHMEM_HUGE_DENY)
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
3899
3900
3901
3902
3903
3904
  		SHMEM_SB(shm_mnt->mnt_sb)->huge = shmem_huge;
  	return count;
  }
  
  struct kobj_attribute shmem_enabled_attr =
  	__ATTR(shmem_enabled, 0644, shmem_enabled_show, shmem_enabled_store);
396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
3905
  #endif /* CONFIG_TRANSPARENT_HUGEPAGE && CONFIG_SYSFS */
f3f0e1d21   Kirill A. Shutemov   khugepaged: add s...
3906

396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
3907
  #ifdef CONFIG_TRANSPARENT_HUGEPAGE
f3f0e1d21   Kirill A. Shutemov   khugepaged: add s...
3908
3909
3910
3911
3912
3913
  bool shmem_huge_enabled(struct vm_area_struct *vma)
  {
  	struct inode *inode = file_inode(vma->vm_file);
  	struct shmem_sb_info *sbinfo = SHMEM_SB(inode->i_sb);
  	loff_t i_size;
  	pgoff_t off;
c06306696   Yang Shi   mm: thp: fix fals...
3914
3915
3916
  	if ((vma->vm_flags & VM_NOHUGEPAGE) ||
  	    test_bit(MMF_DISABLE_THP, &vma->vm_mm->flags))
  		return false;
f3f0e1d21   Kirill A. Shutemov   khugepaged: add s...
3917
3918
3919
3920
3921
3922
3923
3924
3925
3926
3927
3928
3929
3930
3931
  	if (shmem_huge == SHMEM_HUGE_FORCE)
  		return true;
  	if (shmem_huge == SHMEM_HUGE_DENY)
  		return false;
  	switch (sbinfo->huge) {
  		case SHMEM_HUGE_NEVER:
  			return false;
  		case SHMEM_HUGE_ALWAYS:
  			return true;
  		case SHMEM_HUGE_WITHIN_SIZE:
  			off = round_up(vma->vm_pgoff, HPAGE_PMD_NR);
  			i_size = round_up(i_size_read(inode), PAGE_SIZE);
  			if (i_size >= HPAGE_PMD_SIZE &&
  					i_size >> PAGE_SHIFT >= off)
  				return true;
e4a9bc589   Joe Perches   mm: use fallthrough;
3932
  			fallthrough;
f3f0e1d21   Kirill A. Shutemov   khugepaged: add s...
3933
3934
3935
3936
3937
3938
3939
3940
  		case SHMEM_HUGE_ADVISE:
  			/* TODO: implement fadvise() hints */
  			return (vma->vm_flags & VM_HUGEPAGE);
  		default:
  			VM_BUG_ON(1);
  			return false;
  	}
  }
396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
3941
  #endif /* CONFIG_TRANSPARENT_HUGEPAGE */
5a6e75f81   Kirill A. Shutemov   shmem: prepare hu...
3942

853ac43ab   Matt Mackall   shmem: unify regu...
3943
3944
3945
3946
3947
3948
3949
3950
3951
3952
  #else /* !CONFIG_SHMEM */
  
  /*
   * tiny-shmem: simple shmemfs and tmpfs using ramfs code
   *
   * This is intended for small system where the benefits of the full
   * shmem code (swap-backed and resource-limited) are outweighed by
   * their complexity. On systems without swap this code should be
   * effectively equivalent, but much lighter weight.
   */
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
3953
  static struct file_system_type shmem_fs_type = {
853ac43ab   Matt Mackall   shmem: unify regu...
3954
  	.name		= "tmpfs",
f32356261   David Howells   vfs: Convert ramf...
3955
  	.init_fs_context = ramfs_init_fs_context,
d7167b149   Al Viro   fs_parse: fold fs...
3956
  	.parameters	= ramfs_fs_parameters,
853ac43ab   Matt Mackall   shmem: unify regu...
3957
  	.kill_sb	= kill_litter_super,
2b8576cb0   Eric W. Biederman   userns: Allow the...
3958
  	.fs_flags	= FS_USERNS_MOUNT,
853ac43ab   Matt Mackall   shmem: unify regu...
3959
  };
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
3960
  int __init shmem_init(void)
853ac43ab   Matt Mackall   shmem: unify regu...
3961
  {
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
3962
  	BUG_ON(register_filesystem(&shmem_fs_type) != 0);
853ac43ab   Matt Mackall   shmem: unify regu...
3963

41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
3964
  	shm_mnt = kern_mount(&shmem_fs_type);
853ac43ab   Matt Mackall   shmem: unify regu...
3965
3966
3967
3968
  	BUG_ON(IS_ERR(shm_mnt));
  
  	return 0;
  }
b56a2d8af   Vineeth Remanan Pillai   mm: rid swapoff o...
3969
3970
  int shmem_unuse(unsigned int type, bool frontswap,
  		unsigned long *fs_pages_to_unuse)
853ac43ab   Matt Mackall   shmem: unify regu...
3971
3972
3973
  {
  	return 0;
  }
3f96b79ad   Hugh Dickins   tmpfs: depend on ...
3974
3975
3976
3977
  int shmem_lock(struct file *file, int lock, struct user_struct *user)
  {
  	return 0;
  }
245132643   Hugh Dickins   SHM_UNLOCK: fix U...
3978
3979
3980
  void shmem_unlock_mapping(struct address_space *mapping)
  {
  }
c01d5b300   Hugh Dickins   shmem: get_unmapp...
3981
3982
3983
3984
3985
3986
3987
3988
  #ifdef CONFIG_MMU
  unsigned long shmem_get_unmapped_area(struct file *file,
  				      unsigned long addr, unsigned long len,
  				      unsigned long pgoff, unsigned long flags)
  {
  	return current->mm->get_unmapped_area(file, addr, len, pgoff, flags);
  }
  #endif
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
3989
  void shmem_truncate_range(struct inode *inode, loff_t lstart, loff_t lend)
94c1e62df   Hugh Dickins   tmpfs: take contr...
3990
  {
41ffe5d5c   Hugh Dickins   tmpfs: miscellane...
3991
  	truncate_inode_pages_range(inode->i_mapping, lstart, lend);
94c1e62df   Hugh Dickins   tmpfs: take contr...
3992
3993
  }
  EXPORT_SYMBOL_GPL(shmem_truncate_range);
0b0a0806b   Hugh Dickins   shmem: fix shared...
3994
3995
  #define shmem_vm_ops				generic_file_vm_ops
  #define shmem_file_operations			ramfs_file_operations
454abafe9   Dmitry Monakhov   ramfs: replace in...
3996
  #define shmem_get_inode(sb, dir, mode, dev, flags)	ramfs_get_inode(sb, dir, mode, dev)
0b0a0806b   Hugh Dickins   shmem: fix shared...
3997
3998
  #define shmem_acct_size(flags, size)		0
  #define shmem_unacct_size(flags, size)		do {} while (0)
853ac43ab   Matt Mackall   shmem: unify regu...
3999
4000
4001
4002
  
  #endif /* CONFIG_SHMEM */
  
  /* common code */
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
4003

703321b60   Matthew Auld   mm/shmem: introdu...
4004
  static struct file *__shmem_file_setup(struct vfsmount *mnt, const char *name, loff_t size,
c72770909   Eric Paris   security: shmem: ...
4005
  				       unsigned long flags, unsigned int i_flags)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
4006
  {
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
4007
  	struct inode *inode;
93dec2da7   Al Viro   ... and switch sh...
4008
  	struct file *res;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
4009

703321b60   Matthew Auld   mm/shmem: introdu...
4010
4011
  	if (IS_ERR(mnt))
  		return ERR_CAST(mnt);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
4012

285b2c4fd   Hugh Dickins   tmpfs: demolish o...
4013
  	if (size < 0 || size > MAX_LFS_FILESIZE)
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
4014
4015
4016
4017
  		return ERR_PTR(-EINVAL);
  
  	if (shmem_acct_size(flags, size))
  		return ERR_PTR(-ENOMEM);
93dec2da7   Al Viro   ... and switch sh...
4018
4019
  	inode = shmem_get_inode(mnt->mnt_sb, NULL, S_IFREG | S_IRWXUGO, 0,
  				flags);
dac2d1f6c   Al Viro   __shmem_file_setu...
4020
4021
4022
4023
  	if (unlikely(!inode)) {
  		shmem_unacct_size(flags, size);
  		return ERR_PTR(-ENOSPC);
  	}
c72770909   Eric Paris   security: shmem: ...
4024
  	inode->i_flags |= i_flags;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
4025
  	inode->i_size = size;
6d6b77f16   Miklos Szeredi   filesystems: add ...
4026
  	clear_nlink(inode);	/* It is unlinked */
26567cdbb   Al Viro   fix nommu breakag...
4027
  	res = ERR_PTR(ramfs_nommu_expand_for_mapping(inode, size));
93dec2da7   Al Viro   ... and switch sh...
4028
4029
4030
  	if (!IS_ERR(res))
  		res = alloc_file_pseudo(inode, mnt, name, O_RDWR,
  				&shmem_file_operations);
26567cdbb   Al Viro   fix nommu breakag...
4031
  	if (IS_ERR(res))
93dec2da7   Al Viro   ... and switch sh...
4032
  		iput(inode);
6b4d0b279   Al Viro   clean shmem_file_...
4033
  	return res;
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
4034
  }
c72770909   Eric Paris   security: shmem: ...
4035
4036
4037
4038
4039
  
  /**
   * shmem_kernel_file_setup - get an unlinked file living in tmpfs which must be
   * 	kernel internal.  There will be NO LSM permission checks against the
   * 	underlying inode.  So users of this interface must do LSM checks at a
e1832f292   Stephen Smalley   ipc: use private ...
4040
4041
   *	higher layer.  The users are the big_key and shm implementations.  LSM
   *	checks are provided at the key or shm level rather than the inode.
c72770909   Eric Paris   security: shmem: ...
4042
4043
4044
4045
4046
4047
   * @name: name for dentry (to be seen in /proc/<pid>/maps
   * @size: size to be set for the file
   * @flags: VM_NORESERVE suppresses pre-accounting of the entire object size
   */
  struct file *shmem_kernel_file_setup(const char *name, loff_t size, unsigned long flags)
  {
703321b60   Matthew Auld   mm/shmem: introdu...
4048
  	return __shmem_file_setup(shm_mnt, name, size, flags, S_PRIVATE);
c72770909   Eric Paris   security: shmem: ...
4049
4050
4051
4052
4053
4054
4055
4056
4057
4058
  }
  
  /**
   * shmem_file_setup - get an unlinked file living in tmpfs
   * @name: name for dentry (to be seen in /proc/<pid>/maps
   * @size: size to be set for the file
   * @flags: VM_NORESERVE suppresses pre-accounting of the entire object size
   */
  struct file *shmem_file_setup(const char *name, loff_t size, unsigned long flags)
  {
703321b60   Matthew Auld   mm/shmem: introdu...
4059
  	return __shmem_file_setup(shm_mnt, name, size, flags, 0);
c72770909   Eric Paris   security: shmem: ...
4060
  }
395e0ddc4   Keith Packard   Export shmem_file...
4061
  EXPORT_SYMBOL_GPL(shmem_file_setup);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
4062

467118102   Randy Dunlap   mm/shmem and tiny...
4063
  /**
703321b60   Matthew Auld   mm/shmem: introdu...
4064
4065
4066
4067
4068
4069
4070
4071
4072
4073
4074
4075
4076
4077
   * shmem_file_setup_with_mnt - get an unlinked file living in tmpfs
   * @mnt: the tmpfs mount where the file will be created
   * @name: name for dentry (to be seen in /proc/<pid>/maps
   * @size: size to be set for the file
   * @flags: VM_NORESERVE suppresses pre-accounting of the entire object size
   */
  struct file *shmem_file_setup_with_mnt(struct vfsmount *mnt, const char *name,
  				       loff_t size, unsigned long flags)
  {
  	return __shmem_file_setup(mnt, name, size, flags, 0);
  }
  EXPORT_SYMBOL_GPL(shmem_file_setup_with_mnt);
  
  /**
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
4078
   * shmem_zero_setup - setup a shared anonymous mapping
45e55300f   Peter Collingbourne   mm: remove unnece...
4079
   * @vma: the vma to be mmapped is prepared by do_mmap
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
4080
4081
4082
4083
4084
   */
  int shmem_zero_setup(struct vm_area_struct *vma)
  {
  	struct file *file;
  	loff_t size = vma->vm_end - vma->vm_start;
66fc13039   Hugh Dickins   mm: shmem_zero_se...
4085
  	/*
c1e8d7c6a   Michel Lespinasse   mmap locking API:...
4086
  	 * Cloning a new file under mmap_lock leads to a lock ordering conflict
66fc13039   Hugh Dickins   mm: shmem_zero_se...
4087
4088
4089
4090
  	 * between XFS directory reading and selinux: since this file is only
  	 * accessible to the user through its mapping, use S_PRIVATE flag to
  	 * bypass file security, in the same way as shmem_kernel_file_setup().
  	 */
703321b60   Matthew Auld   mm/shmem: introdu...
4091
  	file = shmem_kernel_file_setup("dev/zero", size, vma->vm_flags);
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
4092
4093
4094
4095
4096
4097
4098
  	if (IS_ERR(file))
  		return PTR_ERR(file);
  
  	if (vma->vm_file)
  		fput(vma->vm_file);
  	vma->vm_file = file;
  	vma->vm_ops = &shmem_vm_ops;
f3f0e1d21   Kirill A. Shutemov   khugepaged: add s...
4099

396bcc529   Matthew Wilcox (Oracle)   mm: remove CONFIG...
4100
  	if (IS_ENABLED(CONFIG_TRANSPARENT_HUGEPAGE) &&
f3f0e1d21   Kirill A. Shutemov   khugepaged: add s...
4101
4102
4103
4104
  			((vma->vm_start + ~HPAGE_PMD_MASK) & HPAGE_PMD_MASK) <
  			(vma->vm_end & HPAGE_PMD_MASK)) {
  		khugepaged_enter(vma, vma->vm_flags);
  	}
1da177e4c   Linus Torvalds   Linux-2.6.12-rc2
4105
4106
  	return 0;
  }
d9d90e5eb   Hugh Dickins   tmpfs: add shmem_...
4107
4108
4109
4110
4111
4112
4113
4114
4115
4116
4117
4118
4119
  
  /**
   * shmem_read_mapping_page_gfp - read into page cache, using specified page allocation flags.
   * @mapping:	the page's address_space
   * @index:	the page index
   * @gfp:	the page allocator flags to use if allocating
   *
   * This behaves as a tmpfs "read_cache_page_gfp(mapping, index, gfp)",
   * with any new page allocations done using the specified allocation flags.
   * But read_cache_page_gfp() uses the ->readpage() method: which does not
   * suit tmpfs, since it may have pages in swapcache, and needs to find those
   * for itself; although drivers/gpu/drm i915 and ttm rely upon this support.
   *
68da9f055   Hugh Dickins   tmpfs: pass gfp t...
4120
4121
   * i915_gem_object_get_pages_gtt() mixes __GFP_NORETRY | __GFP_NOWARN in
   * with the mapping_gfp_mask(), to avoid OOMing the machine unnecessarily.
d9d90e5eb   Hugh Dickins   tmpfs: add shmem_...
4122
4123
4124
4125
   */
  struct page *shmem_read_mapping_page_gfp(struct address_space *mapping,
  					 pgoff_t index, gfp_t gfp)
  {
68da9f055   Hugh Dickins   tmpfs: pass gfp t...
4126
4127
  #ifdef CONFIG_SHMEM
  	struct inode *inode = mapping->host;
9276aad6c   Hugh Dickins   tmpfs: remove_shm...
4128
  	struct page *page;
68da9f055   Hugh Dickins   tmpfs: pass gfp t...
4129
4130
4131
  	int error;
  
  	BUG_ON(mapping->a_ops != &shmem_aops);
9e18eb293   Andres Lagar-Cavilla   tmpfs: mem_cgroup...
4132
  	error = shmem_getpage_gfp(inode, index, &page, SGP_CACHE,
cfda05267   Mike Rapoport   userfaultfd: shme...
4133
  				  gfp, NULL, NULL, NULL);
68da9f055   Hugh Dickins   tmpfs: pass gfp t...
4134
4135
4136
4137
4138
4139
4140
4141
4142
  	if (error)
  		page = ERR_PTR(error);
  	else
  		unlock_page(page);
  	return page;
  #else
  	/*
  	 * The tiny !SHMEM case uses ramfs without swap
  	 */
d9d90e5eb   Hugh Dickins   tmpfs: add shmem_...
4143
  	return read_cache_page_gfp(mapping, index, gfp);
68da9f055   Hugh Dickins   tmpfs: pass gfp t...
4144
  #endif
d9d90e5eb   Hugh Dickins   tmpfs: add shmem_...
4145
4146
  }
  EXPORT_SYMBOL_GPL(shmem_read_mapping_page_gfp);