Blame view

fs/eventfd.c 12 KB
457c89965   Thomas Gleixner   treewide: Add SPD...
1
  // SPDX-License-Identifier: GPL-2.0-only
e1ad7468c   Davide Libenzi   signal/timer/even...
2
3
4
5
6
7
8
9
10
11
12
  /*
   *  fs/eventfd.c
   *
   *  Copyright (C) 2007  Davide Libenzi <davidel@xmailserver.org>
   *
   */
  
  #include <linux/file.h>
  #include <linux/poll.h>
  #include <linux/init.h>
  #include <linux/fs.h>
174cd4b1e   Ingo Molnar   sched/headers: Pr...
13
  #include <linux/sched/signal.h>
e1ad7468c   Davide Libenzi   signal/timer/even...
14
  #include <linux/kernel.h>
5a0e3ad6a   Tejun Heo   include cleanup: ...
15
  #include <linux/slab.h>
e1ad7468c   Davide Libenzi   signal/timer/even...
16
17
18
  #include <linux/list.h>
  #include <linux/spinlock.h>
  #include <linux/anon_inodes.h>
7747cdb2f   Adrian Bunk   fs/eventfd.c shou...
19
  #include <linux/syscalls.h>
630d9c472   Paul Gortmaker   fs: reduce the us...
20
  #include <linux/export.h>
133890103   Davide Libenzi   eventfd: revised ...
21
22
  #include <linux/kref.h>
  #include <linux/eventfd.h>
cbac5542d   Cyrill Gorcunov   fs, eventfd: add ...
23
24
  #include <linux/proc_fs.h>
  #include <linux/seq_file.h>
b556db17b   Masatake YAMATO   eventfd: present ...
25
  #include <linux/idr.h>
12aceb89b   Jens Axboe   eventfd: convert ...
26
  #include <linux/uio.h>
b556db17b   Masatake YAMATO   eventfd: present ...
27

b5e683d5c   Jens Axboe   eventfd: track ev...
28
  DEFINE_PER_CPU(int, eventfd_wake_count);
ce528c4c2   YueHaibing   fs/eventfd.c: mak...
29
  static DEFINE_IDA(eventfd_ida);
e1ad7468c   Davide Libenzi   signal/timer/even...
30
31
  
  struct eventfd_ctx {
133890103   Davide Libenzi   eventfd: revised ...
32
  	struct kref kref;
e1ad7468c   Davide Libenzi   signal/timer/even...
33
34
35
36
37
38
  	wait_queue_head_t wqh;
  	/*
  	 * Every time that a write(2) is performed on an eventfd, the
  	 * value of the __u64 being written is added to "count" and a
  	 * wakeup is performed on "wqh". A read(2) will return the "count"
  	 * value to userspace, and will reset "count" to zero. The kernel
133890103   Davide Libenzi   eventfd: revised ...
39
  	 * side eventfd_signal() also, adds to the "count" counter and
e1ad7468c   Davide Libenzi   signal/timer/even...
40
41
42
  	 * issue a wakeup.
  	 */
  	__u64 count;
bcd0b235b   Davide Libenzi   eventfd: improve ...
43
  	unsigned int flags;
b556db17b   Masatake YAMATO   eventfd: present ...
44
  	int id;
e1ad7468c   Davide Libenzi   signal/timer/even...
45
  };
133890103   Davide Libenzi   eventfd: revised ...
46
47
48
49
50
51
52
53
  /**
   * eventfd_signal - Adds @n to the eventfd counter.
   * @ctx: [in] Pointer to the eventfd context.
   * @n: [in] Value of the counter to be added to the eventfd internal counter.
   *          The value cannot be negative.
   *
   * This function is supposed to be called by the kernel in paths that do not
   * allow sleeping. In this function we allow the counter to reach the ULLONG_MAX
a9a08845e   Linus Torvalds   vfs: do bulk POLL...
54
   * value, and we signal this as overflow condition by returning a EPOLLERR
133890103   Davide Libenzi   eventfd: revised ...
55
56
   * to poll(2).
   *
20d5a865e   Masanari Iida   Documentation: fi...
57
   * Returns the amount by which the counter was incremented.  This will be less
ee62c6b2d   Sha Zhengju   eventfd: change i...
58
   * than @n if the counter has overflowed.
e1ad7468c   Davide Libenzi   signal/timer/even...
59
   */
ee62c6b2d   Sha Zhengju   eventfd: change i...
60
  __u64 eventfd_signal(struct eventfd_ctx *ctx, __u64 n)
e1ad7468c   Davide Libenzi   signal/timer/even...
61
  {
e1ad7468c   Davide Libenzi   signal/timer/even...
62
  	unsigned long flags;
b5e683d5c   Jens Axboe   eventfd: track ev...
63
64
65
66
67
68
69
70
71
72
  	/*
  	 * Deadlock or stack overflow issues can happen if we recurse here
  	 * through waitqueue wakeup handlers. If the caller users potentially
  	 * nested waitqueues with custom wakeup handlers, then it should
  	 * check eventfd_signal_count() before calling this function. If
  	 * it returns true, the eventfd_signal() call should be deferred to a
  	 * safe context.
  	 */
  	if (WARN_ON_ONCE(this_cpu_read(eventfd_wake_count)))
  		return 0;
d48eb2331   Davide Libenzi   eventfd use waitq...
73
  	spin_lock_irqsave(&ctx->wqh.lock, flags);
b5e683d5c   Jens Axboe   eventfd: track ev...
74
  	this_cpu_inc(eventfd_wake_count);
e1ad7468c   Davide Libenzi   signal/timer/even...
75
  	if (ULLONG_MAX - ctx->count < n)
ee62c6b2d   Sha Zhengju   eventfd: change i...
76
  		n = ULLONG_MAX - ctx->count;
e1ad7468c   Davide Libenzi   signal/timer/even...
77
78
  	ctx->count += n;
  	if (waitqueue_active(&ctx->wqh))
a9a08845e   Linus Torvalds   vfs: do bulk POLL...
79
  		wake_up_locked_poll(&ctx->wqh, EPOLLIN);
b5e683d5c   Jens Axboe   eventfd: track ev...
80
  	this_cpu_dec(eventfd_wake_count);
d48eb2331   Davide Libenzi   eventfd use waitq...
81
  	spin_unlock_irqrestore(&ctx->wqh.lock, flags);
e1ad7468c   Davide Libenzi   signal/timer/even...
82
83
84
  
  	return n;
  }
5718607bb   Rusty Russell   eventfd: export e...
85
  EXPORT_SYMBOL_GPL(eventfd_signal);
e1ad7468c   Davide Libenzi   signal/timer/even...
86

562787a5c   Davide Libenzi   anonfd: split int...
87
88
  static void eventfd_free_ctx(struct eventfd_ctx *ctx)
  {
b556db17b   Masatake YAMATO   eventfd: present ...
89
90
  	if (ctx->id >= 0)
  		ida_simple_remove(&eventfd_ida, ctx->id);
562787a5c   Davide Libenzi   anonfd: split int...
91
92
  	kfree(ctx);
  }
133890103   Davide Libenzi   eventfd: revised ...
93
94
95
  static void eventfd_free(struct kref *kref)
  {
  	struct eventfd_ctx *ctx = container_of(kref, struct eventfd_ctx, kref);
562787a5c   Davide Libenzi   anonfd: split int...
96
  	eventfd_free_ctx(ctx);
133890103   Davide Libenzi   eventfd: revised ...
97
98
99
  }
  
  /**
133890103   Davide Libenzi   eventfd: revised ...
100
101
102
103
   * eventfd_ctx_put - Releases a reference to the internal eventfd context.
   * @ctx: [in] Pointer to eventfd context.
   *
   * The eventfd context reference must have been previously acquired either
105f2b709   Eric Biggers   eventfd: fold eve...
104
   * with eventfd_ctx_fdget() or eventfd_ctx_fileget().
133890103   Davide Libenzi   eventfd: revised ...
105
106
107
108
109
110
   */
  void eventfd_ctx_put(struct eventfd_ctx *ctx)
  {
  	kref_put(&ctx->kref, eventfd_free);
  }
  EXPORT_SYMBOL_GPL(eventfd_ctx_put);
e1ad7468c   Davide Libenzi   signal/timer/even...
111
112
  static int eventfd_release(struct inode *inode, struct file *file)
  {
133890103   Davide Libenzi   eventfd: revised ...
113
  	struct eventfd_ctx *ctx = file->private_data;
a9a08845e   Linus Torvalds   vfs: do bulk POLL...
114
  	wake_up_poll(&ctx->wqh, EPOLLHUP);
133890103   Davide Libenzi   eventfd: revised ...
115
  	eventfd_ctx_put(ctx);
e1ad7468c   Davide Libenzi   signal/timer/even...
116
117
  	return 0;
  }
a11e1d432   Linus Torvalds   Revert changes to...
118
  static __poll_t eventfd_poll(struct file *file, poll_table *wait)
e1ad7468c   Davide Libenzi   signal/timer/even...
119
120
  {
  	struct eventfd_ctx *ctx = file->private_data;
076ccb76e   Al Viro   fs: annotate ->po...
121
  	__poll_t events = 0;
e22553e2a   Chris Mason   eventfd: don't ta...
122
  	u64 count;
e1ad7468c   Davide Libenzi   signal/timer/even...
123

a11e1d432   Linus Torvalds   Revert changes to...
124
  	poll_wait(file, &ctx->wqh, wait);
a484c3dd9   Paolo Bonzini   eventfd: document...
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
  	/*
  	 * All writes to ctx->count occur within ctx->wqh.lock.  This read
  	 * can be done outside ctx->wqh.lock because we know that poll_wait
  	 * takes that lock (through add_wait_queue) if our caller will sleep.
  	 *
  	 * The read _can_ therefore seep into add_wait_queue's critical
  	 * section, but cannot move above it!  add_wait_queue's spin_lock acts
  	 * as an acquire barrier and ensures that the read be ordered properly
  	 * against the writes.  The following CAN happen and is safe:
  	 *
  	 *     poll                               write
  	 *     -----------------                  ------------
  	 *     lock ctx->wqh.lock (in poll_wait)
  	 *     count = ctx->count
  	 *     __add_wait_queue
  	 *     unlock ctx->wqh.lock
  	 *                                        lock ctx->qwh.lock
  	 *                                        ctx->count += n
  	 *                                        if (waitqueue_active)
  	 *                                          wake_up_locked_poll
  	 *                                        unlock ctx->qwh.lock
  	 *     eventfd_poll returns 0
  	 *
  	 * but the following, which would miss a wakeup, cannot happen:
  	 *
  	 *     poll                               write
  	 *     -----------------                  ------------
  	 *     count = ctx->count (INVALID!)
  	 *                                        lock ctx->qwh.lock
  	 *                                        ctx->count += n
  	 *                                        **waitqueue_active is false**
  	 *                                        **no wake_up_locked_poll!**
  	 *                                        unlock ctx->qwh.lock
  	 *     lock ctx->wqh.lock (in poll_wait)
  	 *     __add_wait_queue
  	 *     unlock ctx->wqh.lock
  	 *     eventfd_poll returns 0
  	 */
  	count = READ_ONCE(ctx->count);
e1ad7468c   Davide Libenzi   signal/timer/even...
164

e22553e2a   Chris Mason   eventfd: don't ta...
165
  	if (count > 0)
a11e1d432   Linus Torvalds   Revert changes to...
166
  		events |= EPOLLIN;
e22553e2a   Chris Mason   eventfd: don't ta...
167
  	if (count == ULLONG_MAX)
a9a08845e   Linus Torvalds   vfs: do bulk POLL...
168
  		events |= EPOLLERR;
e22553e2a   Chris Mason   eventfd: don't ta...
169
  	if (ULLONG_MAX - 1 > count)
a11e1d432   Linus Torvalds   Revert changes to...
170
  		events |= EPOLLOUT;
e1ad7468c   Davide Libenzi   signal/timer/even...
171
172
173
  
  	return events;
  }
cb289d624   Davide Libenzi   eventfd - allow a...
174
175
176
177
178
179
180
181
182
183
  static void eventfd_ctx_do_read(struct eventfd_ctx *ctx, __u64 *cnt)
  {
  	*cnt = (ctx->flags & EFD_SEMAPHORE) ? 1 : ctx->count;
  	ctx->count -= *cnt;
  }
  
  /**
   * eventfd_ctx_remove_wait_queue - Read the current counter and removes wait queue.
   * @ctx: [in] Pointer to eventfd context.
   * @wait: [in] Wait queue to be removed.
361821854   Randy Dunlap   Docbook: add fs/e...
184
   * @cnt: [out] Pointer to the 64-bit counter value.
cb289d624   Davide Libenzi   eventfd - allow a...
185
   *
361821854   Randy Dunlap   Docbook: add fs/e...
186
   * Returns %0 if successful, or the following error codes:
cb289d624   Davide Libenzi   eventfd - allow a...
187
188
189
190
191
192
   *
   * -EAGAIN      : The operation would have blocked.
   *
   * This is used to atomically remove a wait queue entry from the eventfd wait
   * queue head, and read/reset the counter value.
   */
ac6424b98   Ingo Molnar   sched/wait: Renam...
193
  int eventfd_ctx_remove_wait_queue(struct eventfd_ctx *ctx, wait_queue_entry_t *wait,
cb289d624   Davide Libenzi   eventfd - allow a...
194
195
196
197
198
199
200
201
  				  __u64 *cnt)
  {
  	unsigned long flags;
  
  	spin_lock_irqsave(&ctx->wqh.lock, flags);
  	eventfd_ctx_do_read(ctx, cnt);
  	__remove_wait_queue(&ctx->wqh, wait);
  	if (*cnt != 0 && waitqueue_active(&ctx->wqh))
a9a08845e   Linus Torvalds   vfs: do bulk POLL...
202
  		wake_up_locked_poll(&ctx->wqh, EPOLLOUT);
cb289d624   Davide Libenzi   eventfd - allow a...
203
204
205
206
207
  	spin_unlock_irqrestore(&ctx->wqh.lock, flags);
  
  	return *cnt != 0 ? 0 : -EAGAIN;
  }
  EXPORT_SYMBOL_GPL(eventfd_ctx_remove_wait_queue);
12aceb89b   Jens Axboe   eventfd: convert ...
208
  static ssize_t eventfd_read(struct kiocb *iocb, struct iov_iter *to)
e1ad7468c   Davide Libenzi   signal/timer/even...
209
  {
12aceb89b   Jens Axboe   eventfd: convert ...
210
  	struct file *file = iocb->ki_filp;
b6364572d   Eric Biggers   eventfd: fold eve...
211
  	struct eventfd_ctx *ctx = file->private_data;
b6364572d   Eric Biggers   eventfd: fold eve...
212
  	__u64 ucnt = 0;
e1ad7468c   Davide Libenzi   signal/timer/even...
213
  	DECLARE_WAITQUEUE(wait, current);
12aceb89b   Jens Axboe   eventfd: convert ...
214
  	if (iov_iter_count(to) < sizeof(ucnt))
b6364572d   Eric Biggers   eventfd: fold eve...
215
  		return -EINVAL;
d48eb2331   Davide Libenzi   eventfd use waitq...
216
  	spin_lock_irq(&ctx->wqh.lock);
12aceb89b   Jens Axboe   eventfd: convert ...
217
218
219
220
221
222
  	if (!ctx->count) {
  		if ((file->f_flags & O_NONBLOCK) ||
  		    (iocb->ki_flags & IOCB_NOWAIT)) {
  			spin_unlock_irq(&ctx->wqh.lock);
  			return -EAGAIN;
  		}
e1ad7468c   Davide Libenzi   signal/timer/even...
223
  		__add_wait_queue(&ctx->wqh, &wait);
cb289d624   Davide Libenzi   eventfd - allow a...
224
  		for (;;) {
e1ad7468c   Davide Libenzi   signal/timer/even...
225
  			set_current_state(TASK_INTERRUPTIBLE);
12aceb89b   Jens Axboe   eventfd: convert ...
226
  			if (ctx->count)
e1ad7468c   Davide Libenzi   signal/timer/even...
227
  				break;
e1ad7468c   Davide Libenzi   signal/timer/even...
228
  			if (signal_pending(current)) {
12aceb89b   Jens Axboe   eventfd: convert ...
229
230
231
232
  				__remove_wait_queue(&ctx->wqh, &wait);
  				__set_current_state(TASK_RUNNING);
  				spin_unlock_irq(&ctx->wqh.lock);
  				return -ERESTARTSYS;
e1ad7468c   Davide Libenzi   signal/timer/even...
233
  			}
d48eb2331   Davide Libenzi   eventfd use waitq...
234
  			spin_unlock_irq(&ctx->wqh.lock);
e1ad7468c   Davide Libenzi   signal/timer/even...
235
  			schedule();
d48eb2331   Davide Libenzi   eventfd use waitq...
236
  			spin_lock_irq(&ctx->wqh.lock);
e1ad7468c   Davide Libenzi   signal/timer/even...
237
238
239
240
  		}
  		__remove_wait_queue(&ctx->wqh, &wait);
  		__set_current_state(TASK_RUNNING);
  	}
12aceb89b   Jens Axboe   eventfd: convert ...
241
242
243
  	eventfd_ctx_do_read(ctx, &ucnt);
  	if (waitqueue_active(&ctx->wqh))
  		wake_up_locked_poll(&ctx->wqh, EPOLLOUT);
d48eb2331   Davide Libenzi   eventfd use waitq...
244
  	spin_unlock_irq(&ctx->wqh.lock);
12aceb89b   Jens Axboe   eventfd: convert ...
245
  	if (unlikely(copy_to_iter(&ucnt, sizeof(ucnt), to) != sizeof(ucnt)))
b6364572d   Eric Biggers   eventfd: fold eve...
246
  		return -EFAULT;
cb289d624   Davide Libenzi   eventfd - allow a...
247

12aceb89b   Jens Axboe   eventfd: convert ...
248
  	return sizeof(ucnt);
cb289d624   Davide Libenzi   eventfd - allow a...
249
  }
e1ad7468c   Davide Libenzi   signal/timer/even...
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
  
  static ssize_t eventfd_write(struct file *file, const char __user *buf, size_t count,
  			     loff_t *ppos)
  {
  	struct eventfd_ctx *ctx = file->private_data;
  	ssize_t res;
  	__u64 ucnt;
  	DECLARE_WAITQUEUE(wait, current);
  
  	if (count < sizeof(ucnt))
  		return -EINVAL;
  	if (copy_from_user(&ucnt, buf, sizeof(ucnt)))
  		return -EFAULT;
  	if (ucnt == ULLONG_MAX)
  		return -EINVAL;
d48eb2331   Davide Libenzi   eventfd use waitq...
265
  	spin_lock_irq(&ctx->wqh.lock);
e1ad7468c   Davide Libenzi   signal/timer/even...
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
  	res = -EAGAIN;
  	if (ULLONG_MAX - ctx->count > ucnt)
  		res = sizeof(ucnt);
  	else if (!(file->f_flags & O_NONBLOCK)) {
  		__add_wait_queue(&ctx->wqh, &wait);
  		for (res = 0;;) {
  			set_current_state(TASK_INTERRUPTIBLE);
  			if (ULLONG_MAX - ctx->count > ucnt) {
  				res = sizeof(ucnt);
  				break;
  			}
  			if (signal_pending(current)) {
  				res = -ERESTARTSYS;
  				break;
  			}
d48eb2331   Davide Libenzi   eventfd use waitq...
281
  			spin_unlock_irq(&ctx->wqh.lock);
e1ad7468c   Davide Libenzi   signal/timer/even...
282
  			schedule();
d48eb2331   Davide Libenzi   eventfd use waitq...
283
  			spin_lock_irq(&ctx->wqh.lock);
e1ad7468c   Davide Libenzi   signal/timer/even...
284
285
286
287
  		}
  		__remove_wait_queue(&ctx->wqh, &wait);
  		__set_current_state(TASK_RUNNING);
  	}
bcd0b235b   Davide Libenzi   eventfd: improve ...
288
  	if (likely(res > 0)) {
e1ad7468c   Davide Libenzi   signal/timer/even...
289
290
  		ctx->count += ucnt;
  		if (waitqueue_active(&ctx->wqh))
a9a08845e   Linus Torvalds   vfs: do bulk POLL...
291
  			wake_up_locked_poll(&ctx->wqh, EPOLLIN);
e1ad7468c   Davide Libenzi   signal/timer/even...
292
  	}
d48eb2331   Davide Libenzi   eventfd use waitq...
293
  	spin_unlock_irq(&ctx->wqh.lock);
e1ad7468c   Davide Libenzi   signal/timer/even...
294
295
296
  
  	return res;
  }
cbac5542d   Cyrill Gorcunov   fs, eventfd: add ...
297
  #ifdef CONFIG_PROC_FS
a3816ab0e   Joe Perches   fs: Convert show_...
298
  static void eventfd_show_fdinfo(struct seq_file *m, struct file *f)
cbac5542d   Cyrill Gorcunov   fs, eventfd: add ...
299
300
  {
  	struct eventfd_ctx *ctx = f->private_data;
cbac5542d   Cyrill Gorcunov   fs, eventfd: add ...
301
302
  
  	spin_lock_irq(&ctx->wqh.lock);
a3816ab0e   Joe Perches   fs: Convert show_...
303
304
305
  	seq_printf(m, "eventfd-count: %16llx
  ",
  		   (unsigned long long)ctx->count);
cbac5542d   Cyrill Gorcunov   fs, eventfd: add ...
306
  	spin_unlock_irq(&ctx->wqh.lock);
b556db17b   Masatake YAMATO   eventfd: present ...
307
308
  	seq_printf(m, "eventfd-id: %d
  ", ctx->id);
cbac5542d   Cyrill Gorcunov   fs, eventfd: add ...
309
310
  }
  #endif
e1ad7468c   Davide Libenzi   signal/timer/even...
311
  static const struct file_operations eventfd_fops = {
cbac5542d   Cyrill Gorcunov   fs, eventfd: add ...
312
313
314
  #ifdef CONFIG_PROC_FS
  	.show_fdinfo	= eventfd_show_fdinfo,
  #endif
e1ad7468c   Davide Libenzi   signal/timer/even...
315
  	.release	= eventfd_release,
a11e1d432   Linus Torvalds   Revert changes to...
316
  	.poll		= eventfd_poll,
12aceb89b   Jens Axboe   eventfd: convert ...
317
  	.read_iter	= eventfd_read,
e1ad7468c   Davide Libenzi   signal/timer/even...
318
  	.write		= eventfd_write,
6038f373a   Arnd Bergmann   llseek: automatic...
319
  	.llseek		= noop_llseek,
e1ad7468c   Davide Libenzi   signal/timer/even...
320
  };
133890103   Davide Libenzi   eventfd: revised ...
321
322
323
324
325
326
327
328
329
330
  /**
   * eventfd_fget - Acquire a reference of an eventfd file descriptor.
   * @fd: [in] Eventfd file descriptor.
   *
   * Returns a pointer to the eventfd file structure in case of success, or the
   * following error pointer:
   *
   * -EBADF    : Invalid @fd file descriptor.
   * -EINVAL   : The @fd file descriptor is not an eventfd file.
   */
e1ad7468c   Davide Libenzi   signal/timer/even...
331
332
333
334
335
336
337
338
339
340
341
342
343
344
  struct file *eventfd_fget(int fd)
  {
  	struct file *file;
  
  	file = fget(fd);
  	if (!file)
  		return ERR_PTR(-EBADF);
  	if (file->f_op != &eventfd_fops) {
  		fput(file);
  		return ERR_PTR(-EINVAL);
  	}
  
  	return file;
  }
5718607bb   Rusty Russell   eventfd: export e...
345
  EXPORT_SYMBOL_GPL(eventfd_fget);
e1ad7468c   Davide Libenzi   signal/timer/even...
346

133890103   Davide Libenzi   eventfd: revised ...
347
348
349
350
351
352
353
354
355
356
357
  /**
   * eventfd_ctx_fdget - Acquires a reference to the internal eventfd context.
   * @fd: [in] Eventfd file descriptor.
   *
   * Returns a pointer to the internal eventfd context, otherwise the error
   * pointers returned by the following functions:
   *
   * eventfd_fget
   */
  struct eventfd_ctx *eventfd_ctx_fdget(int fd)
  {
133890103   Davide Libenzi   eventfd: revised ...
358
  	struct eventfd_ctx *ctx;
36a741172   Al Viro   eventfd_ctx_fdget...
359
360
361
362
363
  	struct fd f = fdget(fd);
  	if (!f.file)
  		return ERR_PTR(-EBADF);
  	ctx = eventfd_ctx_fileget(f.file);
  	fdput(f);
133890103   Davide Libenzi   eventfd: revised ...
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
  	return ctx;
  }
  EXPORT_SYMBOL_GPL(eventfd_ctx_fdget);
  
  /**
   * eventfd_ctx_fileget - Acquires a reference to the internal eventfd context.
   * @file: [in] Eventfd file pointer.
   *
   * Returns a pointer to the internal eventfd context, otherwise the error
   * pointer:
   *
   * -EINVAL   : The @fd file descriptor is not an eventfd file.
   */
  struct eventfd_ctx *eventfd_ctx_fileget(struct file *file)
  {
105f2b709   Eric Biggers   eventfd: fold eve...
379
  	struct eventfd_ctx *ctx;
133890103   Davide Libenzi   eventfd: revised ...
380
381
  	if (file->f_op != &eventfd_fops)
  		return ERR_PTR(-EINVAL);
105f2b709   Eric Biggers   eventfd: fold eve...
382
383
384
  	ctx = file->private_data;
  	kref_get(&ctx->kref);
  	return ctx;
133890103   Davide Libenzi   eventfd: revised ...
385
386
  }
  EXPORT_SYMBOL_GPL(eventfd_ctx_fileget);
2fc96f833   Dominik Brodowski   fs: add do_eventf...
387
  static int do_eventfd(unsigned int count, int flags)
e1ad7468c   Davide Libenzi   signal/timer/even...
388
  {
e1ad7468c   Davide Libenzi   signal/timer/even...
389
  	struct eventfd_ctx *ctx;
12aceb89b   Jens Axboe   eventfd: convert ...
390
  	struct file *file;
7d815165c   Eric Biggers   eventfd: convert ...
391
  	int fd;
e1ad7468c   Davide Libenzi   signal/timer/even...
392

e38b36f32   Ulrich Drepper   flag parameters: ...
393
394
395
  	/* Check the EFD_* constants for consistency.  */
  	BUILD_BUG_ON(EFD_CLOEXEC != O_CLOEXEC);
  	BUILD_BUG_ON(EFD_NONBLOCK != O_NONBLOCK);
bcd0b235b   Davide Libenzi   eventfd: improve ...
396
  	if (flags & ~EFD_FLAGS_SET)
7d815165c   Eric Biggers   eventfd: convert ...
397
  		return -EINVAL;
b087498eb   Ulrich Drepper   flag parameters: ...
398

e1ad7468c   Davide Libenzi   signal/timer/even...
399
400
  	ctx = kmalloc(sizeof(*ctx), GFP_KERNEL);
  	if (!ctx)
7d815165c   Eric Biggers   eventfd: convert ...
401
  		return -ENOMEM;
e1ad7468c   Davide Libenzi   signal/timer/even...
402

133890103   Davide Libenzi   eventfd: revised ...
403
  	kref_init(&ctx->kref);
e1ad7468c   Davide Libenzi   signal/timer/even...
404
  	init_waitqueue_head(&ctx->wqh);
e1ad7468c   Davide Libenzi   signal/timer/even...
405
  	ctx->count = count;
bcd0b235b   Davide Libenzi   eventfd: improve ...
406
  	ctx->flags = flags;
b556db17b   Masatake YAMATO   eventfd: present ...
407
  	ctx->id = ida_simple_get(&eventfd_ida, 0, 0, GFP_KERNEL);
e1ad7468c   Davide Libenzi   signal/timer/even...
408

12aceb89b   Jens Axboe   eventfd: convert ...
409
410
411
  	flags &= EFD_SHARED_FCNTL_FLAGS;
  	flags |= O_RDWR;
  	fd = get_unused_fd_flags(flags);
7d815165c   Eric Biggers   eventfd: convert ...
412
  	if (fd < 0)
12aceb89b   Jens Axboe   eventfd: convert ...
413
414
415
416
417
418
419
420
  		goto err;
  
  	file = anon_inode_getfile("[eventfd]", &eventfd_fops, ctx, flags);
  	if (IS_ERR(file)) {
  		put_unused_fd(fd);
  		fd = PTR_ERR(file);
  		goto err;
  	}
562787a5c   Davide Libenzi   anonfd: split int...
421

12aceb89b   Jens Axboe   eventfd: convert ...
422
423
424
425
426
  	file->f_mode |= FMODE_NOWAIT;
  	fd_install(fd, file);
  	return fd;
  err:
  	eventfd_free_ctx(ctx);
2030a42ce   Al Viro   [PATCH] sanitize ...
427
  	return fd;
e1ad7468c   Davide Libenzi   signal/timer/even...
428
  }
2fc96f833   Dominik Brodowski   fs: add do_eventf...
429
430
431
432
  SYSCALL_DEFINE2(eventfd2, unsigned int, count, int, flags)
  {
  	return do_eventfd(count, flags);
  }
d4e82042c   Heiko Carstens   [CVE-2009-0029] S...
433
  SYSCALL_DEFINE1(eventfd, unsigned int, count)
b087498eb   Ulrich Drepper   flag parameters: ...
434
  {
2fc96f833   Dominik Brodowski   fs: add do_eventf...
435
  	return do_eventfd(count, 0);
b087498eb   Ulrich Drepper   flag parameters: ...
436
  }
bcd0b235b   Davide Libenzi   eventfd: improve ...
437