Blame view

kernel/locking/rwsem.c 48.5 KB
b24413180   Greg Kroah-Hartman   License cleanup: ...
1
  // SPDX-License-Identifier: GPL-2.0
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
2
3
4
5
  /* kernel/rwsem.c: R/W semaphores, public implementation
   *
   * Written by David Howells (dhowells@redhat.com).
   * Derived from asm-i386/semaphore.h
5dec94d49   Waiman Long   locking/rwsem: Me...
6
7
8
9
10
11
12
   *
   * Writer lock-stealing by Alex Shi <alex.shi@intel.com>
   * and Michel Lespinasse <walken@google.com>
   *
   * Optimistic spinning by Tim Chen <tim.c.chen@intel.com>
   * and Davidlohr Bueso <davidlohr@hp.com>. Based on mutexes.
   *
4f23dbc1e   Waiman Long   locking/rwsem: Im...
13
14
15
   * Rwsem count bit fields re-definition and rwsem rearchitecture by
   * Waiman Long <longman@redhat.com> and
   * Peter Zijlstra <peterz@infradead.org>.
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
16
17
18
19
   */
  
  #include <linux/types.h>
  #include <linux/kernel.h>
c7af77b58   Livio Soares   sched: mark rwsem...
20
  #include <linux/sched.h>
5dec94d49   Waiman Long   locking/rwsem: Me...
21
22
  #include <linux/sched/rt.h>
  #include <linux/sched/task.h>
b17b01533   Ingo Molnar   sched/headers: Pr...
23
  #include <linux/sched/debug.h>
5dec94d49   Waiman Long   locking/rwsem: Me...
24
25
  #include <linux/sched/wake_q.h>
  #include <linux/sched/signal.h>
7d43f1ce9   Waiman Long   locking/rwsem: En...
26
  #include <linux/sched/clock.h>
9984de1a5   Paul Gortmaker   kernel: Map most ...
27
  #include <linux/export.h>
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
28
  #include <linux/rwsem.h>
60063497a   Arun Sharma   atomic: use <linu...
29
  #include <linux/atomic.h>
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
30

5dec94d49   Waiman Long   locking/rwsem: Me...
31
  #include "lock_events.h"
df18d99da   JianMin Liu   ANDROID: rwsem: A...
32
  #include <trace/hooks/rwsem.h>
9ad8ff902   Sangmoon Kim   ANDROID: vendor_h...
33
  #include <trace/hooks/dtask.h>
5dec94d49   Waiman Long   locking/rwsem: Me...
34
35
  
  /*
7d43f1ce9   Waiman Long   locking/rwsem: En...
36
   * The least significant 3 bits of the owner value has the following
5dec94d49   Waiman Long   locking/rwsem: Me...
37
   * meanings when set.
02f1082b0   Waiman Long   locking/rwsem: Cl...
38
   *  - Bit 0: RWSEM_READER_OWNED - The rwsem is owned by readers
7d43f1ce9   Waiman Long   locking/rwsem: En...
39
40
   *  - Bit 1: RWSEM_RD_NONSPINNABLE - Readers cannot spin on this lock.
   *  - Bit 2: RWSEM_WR_NONSPINNABLE - Writers cannot spin on this lock.
5dec94d49   Waiman Long   locking/rwsem: Me...
41
   *
7d43f1ce9   Waiman Long   locking/rwsem: En...
42
43
44
45
46
47
48
   * When the rwsem is either owned by an anonymous writer, or it is
   * reader-owned, but a spinning writer has timed out, both nonspinnable
   * bits will be set to disable optimistic spinning by readers and writers.
   * In the later case, the last unlocking reader should then check the
   * writer nonspinnable bit and clear it only to give writers preference
   * to acquire the lock via optimistic spinning, but not readers. Similar
   * action is also done in the reader slowpath.
5dec94d49   Waiman Long   locking/rwsem: Me...
49
50
51
52
   * When a writer acquires a rwsem, it puts its task_struct pointer
   * into the owner field. It is cleared after an unlock.
   *
   * When a reader acquires a rwsem, it will also puts its task_struct
7d43f1ce9   Waiman Long   locking/rwsem: En...
53
54
55
56
   * pointer into the owner field with the RWSEM_READER_OWNED bit set.
   * On unlock, the owner field will largely be left untouched. So
   * for a free or reader-owned rwsem, the owner value may contain
   * information about the last reader that acquires the rwsem.
5dec94d49   Waiman Long   locking/rwsem: Me...
57
58
59
60
61
   *
   * That information may be helpful in debugging cases where the system
   * seems to hang on a reader owned rwsem especially if only one reader
   * is involved. Ideally we would like to track all the readers that own
   * a rwsem, but the overhead is simply too big.
5cfd92e12   Waiman Long   locking/rwsem: Ad...
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
   *
   * Reader optimistic spinning is helpful when the reader critical section
   * is short and there aren't that many readers around. It makes readers
   * relatively more preferred than writers. When a writer times out spinning
   * on a reader-owned lock and set the nospinnable bits, there are two main
   * reasons for that.
   *
   *  1) The reader critical section is long, perhaps the task sleeps after
   *     acquiring the read lock.
   *  2) There are just too many readers contending the lock causing it to
   *     take a while to service all of them.
   *
   * In the former case, long reader critical section will impede the progress
   * of writers which is usually more important for system performance. In
   * the later case, reader optimistic spinning tends to make the reader
   * groups that contain readers that acquire the lock together smaller
   * leading to more of them. That may hurt performance in some cases. In
   * other words, the setting of nonspinnable bits indicates that reader
   * optimistic spinning may not be helpful for those workloads that cause
   * it.
   *
   * Therefore, any writers that had observed the setting of the writer
   * nonspinnable bit for a given rwsem after they fail to acquire the lock
   * via optimistic spinning will set the reader nonspinnable bit once they
   * acquire the write lock. Similarly, readers that observe the setting
   * of reader nonspinnable bit at slowpath entry will set the reader
   * nonspinnable bits when they acquire the read lock via the wakeup path.
   *
   * Once the reader nonspinnable bit is on, it will only be reset when
   * a writer is able to acquire the rwsem in the fast path or somehow a
   * reader or writer in the slowpath doesn't observe the nonspinable bit.
   *
   * This is to discourage reader optmistic spinning on that particular
   * rwsem and make writers more preferred. This adaptive disabling of reader
   * optimistic spinning will alleviate the negative side effect of this
   * feature.
5dec94d49   Waiman Long   locking/rwsem: Me...
98
99
   */
  #define RWSEM_READER_OWNED	(1UL << 0)
7d43f1ce9   Waiman Long   locking/rwsem: En...
100
101
102
  #define RWSEM_RD_NONSPINNABLE	(1UL << 1)
  #define RWSEM_WR_NONSPINNABLE	(1UL << 2)
  #define RWSEM_NONSPINNABLE	(RWSEM_RD_NONSPINNABLE | RWSEM_WR_NONSPINNABLE)
02f1082b0   Waiman Long   locking/rwsem: Cl...
103
  #define RWSEM_OWNER_FLAGS_MASK	(RWSEM_READER_OWNED | RWSEM_NONSPINNABLE)
5dec94d49   Waiman Long   locking/rwsem: Me...
104
105
106
107
  
  #ifdef CONFIG_DEBUG_RWSEMS
  # define DEBUG_RWSEMS_WARN_ON(c, sem)	do {			\
  	if (!debug_locks_silent &&				\
fce45cd41   Davidlohr Bueso   locking/rwsem: Ch...
108
109
  	    WARN_ONCE(c, "DEBUG_RWSEMS_WARN_ON(%s): count = 0x%lx, magic = 0x%lx, owner = 0x%lx, curr 0x%lx, list %sempty
  ",\
5dec94d49   Waiman Long   locking/rwsem: Me...
110
  		#c, atomic_long_read(&(sem)->count),		\
fce45cd41   Davidlohr Bueso   locking/rwsem: Ch...
111
  		(unsigned long) sem->magic,			\
94a9717b3   Waiman Long   locking/rwsem: Ma...
112
  		atomic_long_read(&(sem)->owner), (long)current,	\
5dec94d49   Waiman Long   locking/rwsem: Me...
113
114
115
116
117
118
119
120
  		list_empty(&(sem)->wait_list) ? "" : "not "))	\
  			debug_locks_off();			\
  	} while (0)
  #else
  # define DEBUG_RWSEMS_WARN_ON(c, sem)
  #endif
  
  /*
a15ea1a35   Waiman Long   locking/rwsem: Gu...
121
   * On 64-bit architectures, the bit definitions of the count are:
5dec94d49   Waiman Long   locking/rwsem: Me...
122
   *
a15ea1a35   Waiman Long   locking/rwsem: Gu...
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
   * Bit  0    - writer locked bit
   * Bit  1    - waiters present bit
   * Bit  2    - lock handoff bit
   * Bits 3-7  - reserved
   * Bits 8-62 - 55-bit reader count
   * Bit  63   - read fail bit
   *
   * On 32-bit architectures, the bit definitions of the count are:
   *
   * Bit  0    - writer locked bit
   * Bit  1    - waiters present bit
   * Bit  2    - lock handoff bit
   * Bits 3-7  - reserved
   * Bits 8-30 - 23-bit reader count
   * Bit  31   - read fail bit
   *
   * It is not likely that the most significant bit (read fail bit) will ever
   * be set. This guard bit is still checked anyway in the down_read() fastpath
   * just in case we need to use up more of the reader bits for other purpose
   * in the future.
5dec94d49   Waiman Long   locking/rwsem: Me...
143
144
145
   *
   * atomic_long_fetch_add() is used to obtain reader lock, whereas
   * atomic_long_cmpxchg() will be used to obtain writer lock.
4f23dbc1e   Waiman Long   locking/rwsem: Im...
146
147
148
149
150
151
152
153
154
   *
   * There are three places where the lock handoff bit may be set or cleared.
   * 1) rwsem_mark_wake() for readers.
   * 2) rwsem_try_write_lock() for writers.
   * 3) Error path of rwsem_down_write_slowpath().
   *
   * For all the above cases, wait_lock will be held. A writer must also
   * be the first one in the wait_list to be eligible for setting the handoff
   * bit. So concurrent setting/clearing of handoff bit is not possible.
5dec94d49   Waiman Long   locking/rwsem: Me...
155
156
157
   */
  #define RWSEM_WRITER_LOCKED	(1UL << 0)
  #define RWSEM_FLAG_WAITERS	(1UL << 1)
4f23dbc1e   Waiman Long   locking/rwsem: Im...
158
  #define RWSEM_FLAG_HANDOFF	(1UL << 2)
a15ea1a35   Waiman Long   locking/rwsem: Gu...
159
  #define RWSEM_FLAG_READFAIL	(1UL << (BITS_PER_LONG - 1))
4f23dbc1e   Waiman Long   locking/rwsem: Im...
160

5dec94d49   Waiman Long   locking/rwsem: Me...
161
162
163
164
165
  #define RWSEM_READER_SHIFT	8
  #define RWSEM_READER_BIAS	(1UL << RWSEM_READER_SHIFT)
  #define RWSEM_READER_MASK	(~(RWSEM_READER_BIAS - 1))
  #define RWSEM_WRITER_MASK	RWSEM_WRITER_LOCKED
  #define RWSEM_LOCK_MASK		(RWSEM_WRITER_MASK|RWSEM_READER_MASK)
4f23dbc1e   Waiman Long   locking/rwsem: Im...
166
  #define RWSEM_READ_FAILED_MASK	(RWSEM_WRITER_MASK|RWSEM_FLAG_WAITERS|\
a15ea1a35   Waiman Long   locking/rwsem: Gu...
167
  				 RWSEM_FLAG_HANDOFF|RWSEM_FLAG_READFAIL)
5dec94d49   Waiman Long   locking/rwsem: Me...
168
169
170
171
172
173
174
175
176
177
  
  /*
   * All writes to owner are protected by WRITE_ONCE() to make sure that
   * store tearing can't happen as optimistic spinners may read and use
   * the owner value concurrently without lock. Read from owner, however,
   * may not need READ_ONCE() as long as the pointer value is only used
   * for comparison and isn't being dereferenced.
   */
  static inline void rwsem_set_owner(struct rw_semaphore *sem)
  {
94a9717b3   Waiman Long   locking/rwsem: Ma...
178
  	atomic_long_set(&sem->owner, (long)current);
5dec94d49   Waiman Long   locking/rwsem: Me...
179
180
181
182
  }
  
  static inline void rwsem_clear_owner(struct rw_semaphore *sem)
  {
94a9717b3   Waiman Long   locking/rwsem: Ma...
183
184
185
186
187
188
189
190
191
  	atomic_long_set(&sem->owner, 0);
  }
  
  /*
   * Test the flags in the owner field.
   */
  static inline bool rwsem_test_oflags(struct rw_semaphore *sem, long flags)
  {
  	return atomic_long_read(&sem->owner) & flags;
5dec94d49   Waiman Long   locking/rwsem: Me...
192
193
194
195
196
197
198
199
200
  }
  
  /*
   * The task_struct pointer of the last owning reader will be left in
   * the owner field.
   *
   * Note that the owner value just indicates the task has owned the rwsem
   * previously, it may not be the real owner or one of the real owners
   * anymore when that field is examined, so take it with a grain of salt.
5cfd92e12   Waiman Long   locking/rwsem: Ad...
201
202
   *
   * The reader non-spinnable bit is preserved.
5dec94d49   Waiman Long   locking/rwsem: Me...
203
204
205
206
   */
  static inline void __rwsem_set_reader_owned(struct rw_semaphore *sem,
  					    struct task_struct *owner)
  {
5cfd92e12   Waiman Long   locking/rwsem: Ad...
207
208
  	unsigned long val = (unsigned long)owner | RWSEM_READER_OWNED |
  		(atomic_long_read(&sem->owner) & RWSEM_RD_NONSPINNABLE);
5dec94d49   Waiman Long   locking/rwsem: Me...
209

94a9717b3   Waiman Long   locking/rwsem: Ma...
210
  	atomic_long_set(&sem->owner, val);
5dec94d49   Waiman Long   locking/rwsem: Me...
211
212
213
214
215
216
217
218
  }
  
  static inline void rwsem_set_reader_owned(struct rw_semaphore *sem)
  {
  	__rwsem_set_reader_owned(sem, current);
  }
  
  /*
94a9717b3   Waiman Long   locking/rwsem: Ma...
219
   * Return true if the rwsem is owned by a reader.
5dec94d49   Waiman Long   locking/rwsem: Me...
220
   */
94a9717b3   Waiman Long   locking/rwsem: Ma...
221
  static inline bool is_rwsem_reader_owned(struct rw_semaphore *sem)
5dec94d49   Waiman Long   locking/rwsem: Me...
222
  {
94a9717b3   Waiman Long   locking/rwsem: Ma...
223
224
225
226
227
228
229
230
231
232
  #ifdef CONFIG_DEBUG_RWSEMS
  	/*
  	 * Check the count to see if it is write-locked.
  	 */
  	long count = atomic_long_read(&sem->count);
  
  	if (count & RWSEM_WRITER_MASK)
  		return false;
  #endif
  	return rwsem_test_oflags(sem, RWSEM_READER_OWNED);
5dec94d49   Waiman Long   locking/rwsem: Me...
233
234
235
236
237
238
239
240
241
242
243
  }
  
  #ifdef CONFIG_DEBUG_RWSEMS
  /*
   * With CONFIG_DEBUG_RWSEMS configured, it will make sure that if there
   * is a task pointer in owner of a reader-owned rwsem, it will be the
   * real owner or one of the real owners. The only exception is when the
   * unlock is done by up_read_non_owner().
   */
  static inline void rwsem_clear_reader_owned(struct rw_semaphore *sem)
  {
94a9717b3   Waiman Long   locking/rwsem: Ma...
244
245
246
247
248
249
250
  	unsigned long val = atomic_long_read(&sem->owner);
  
  	while ((val & ~RWSEM_OWNER_FLAGS_MASK) == (unsigned long)current) {
  		if (atomic_long_try_cmpxchg(&sem->owner, &val,
  					    val & RWSEM_OWNER_FLAGS_MASK))
  			return;
  	}
5dec94d49   Waiman Long   locking/rwsem: Me...
251
252
253
254
255
256
257
258
  }
  #else
  static inline void rwsem_clear_reader_owned(struct rw_semaphore *sem)
  {
  }
  #endif
  
  /*
7d43f1ce9   Waiman Long   locking/rwsem: En...
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
   * Set the RWSEM_NONSPINNABLE bits if the RWSEM_READER_OWNED flag
   * remains set. Otherwise, the operation will be aborted.
   */
  static inline void rwsem_set_nonspinnable(struct rw_semaphore *sem)
  {
  	unsigned long owner = atomic_long_read(&sem->owner);
  
  	do {
  		if (!(owner & RWSEM_READER_OWNED))
  			break;
  		if (owner & RWSEM_NONSPINNABLE)
  			break;
  	} while (!atomic_long_try_cmpxchg(&sem->owner, &owner,
  					  owner | RWSEM_NONSPINNABLE));
  }
a15ea1a35   Waiman Long   locking/rwsem: Gu...
274
275
276
277
278
279
280
  static inline bool rwsem_read_trylock(struct rw_semaphore *sem)
  {
  	long cnt = atomic_long_add_return_acquire(RWSEM_READER_BIAS, &sem->count);
  	if (WARN_ON_ONCE(cnt < 0))
  		rwsem_set_nonspinnable(sem);
  	return !(cnt & RWSEM_READ_FAILED_MASK);
  }
7d43f1ce9   Waiman Long   locking/rwsem: En...
281
  /*
94a9717b3   Waiman Long   locking/rwsem: Ma...
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
   * Return just the real task structure pointer of the owner
   */
  static inline struct task_struct *rwsem_owner(struct rw_semaphore *sem)
  {
  	return (struct task_struct *)
  		(atomic_long_read(&sem->owner) & ~RWSEM_OWNER_FLAGS_MASK);
  }
  
  /*
   * Return the real task structure pointer of the owner and the embedded
   * flags in the owner. pflags must be non-NULL.
   */
  static inline struct task_struct *
  rwsem_owner_flags(struct rw_semaphore *sem, unsigned long *pflags)
  {
  	unsigned long owner = atomic_long_read(&sem->owner);
  
  	*pflags = owner & RWSEM_OWNER_FLAGS_MASK;
  	return (struct task_struct *)(owner & ~RWSEM_OWNER_FLAGS_MASK);
  }
  
  /*
5dec94d49   Waiman Long   locking/rwsem: Me...
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
   * Guide to the rw_semaphore's count field.
   *
   * When the RWSEM_WRITER_LOCKED bit in count is set, the lock is owned
   * by a writer.
   *
   * The lock is owned by readers when
   * (1) the RWSEM_WRITER_LOCKED isn't set in count,
   * (2) some of the reader bits are set in count, and
   * (3) the owner field has RWSEM_READ_OWNED bit set.
   *
   * Having some reader bits set is not enough to guarantee a readers owned
   * lock as the readers may be in the process of backing out from the count
   * and a writer has just released the lock. So another writer may steal
   * the lock immediately after that.
   */
  
  /*
   * Initialize an rwsem:
   */
  void __init_rwsem(struct rw_semaphore *sem, const char *name,
  		  struct lock_class_key *key)
  {
  #ifdef CONFIG_DEBUG_LOCK_ALLOC
  	/*
  	 * Make sure we are not reinitializing a held semaphore:
  	 */
  	debug_check_no_locks_freed((void *)sem, sizeof(*sem));
de8f5e4f2   Peter Zijlstra   lockdep: Introduc...
331
  	lockdep_init_map_wait(&sem->dep_map, name, key, 0, LD_WAIT_SLEEP);
5dec94d49   Waiman Long   locking/rwsem: Me...
332
  #endif
fce45cd41   Davidlohr Bueso   locking/rwsem: Ch...
333
334
335
  #ifdef CONFIG_DEBUG_RWSEMS
  	sem->magic = sem;
  #endif
5dec94d49   Waiman Long   locking/rwsem: Me...
336
337
338
  	atomic_long_set(&sem->count, RWSEM_UNLOCKED_VALUE);
  	raw_spin_lock_init(&sem->wait_lock);
  	INIT_LIST_HEAD(&sem->wait_list);
94a9717b3   Waiman Long   locking/rwsem: Ma...
339
  	atomic_long_set(&sem->owner, 0L);
5dec94d49   Waiman Long   locking/rwsem: Me...
340
341
342
  #ifdef CONFIG_RWSEM_SPIN_ON_OWNER
  	osq_lock_init(&sem->osq);
  #endif
df18d99da   JianMin Liu   ANDROID: rwsem: A...
343
  	trace_android_vh_rwsem_init(sem);
5dec94d49   Waiman Long   locking/rwsem: Me...
344
  }
5dec94d49   Waiman Long   locking/rwsem: Me...
345
  EXPORT_SYMBOL(__init_rwsem);
4f23dbc1e   Waiman Long   locking/rwsem: Im...
346
347
  #define rwsem_first_waiter(sem) \
  	list_first_entry(&sem->wait_list, struct rwsem_waiter, list)
5dec94d49   Waiman Long   locking/rwsem: Me...
348
349
350
351
352
353
  
  enum rwsem_wake_type {
  	RWSEM_WAKE_ANY,		/* Wake whatever's at head of wait list */
  	RWSEM_WAKE_READERS,	/* Wake readers only */
  	RWSEM_WAKE_READ_OWNED	/* Waker thread holds the read lock */
  };
4f23dbc1e   Waiman Long   locking/rwsem: Im...
354
355
356
357
358
359
360
361
362
363
364
365
  enum writer_wait_state {
  	WRITER_NOT_FIRST,	/* Writer is not first in wait list */
  	WRITER_FIRST,		/* Writer is first in wait list     */
  	WRITER_HANDOFF		/* Writer is first & handoff needed */
  };
  
  /*
   * The typical HZ value is either 250 or 1000. So set the minimum waiting
   * time to at least 4ms or 1 jiffy (if it is higher than 4ms) in the wait
   * queue before initiating the handoff protocol.
   */
  #define RWSEM_WAIT_TIMEOUT	DIV_ROUND_UP(HZ, 250)
5dec94d49   Waiman Long   locking/rwsem: Me...
366
  /*
d3681e269   Waiman Long   locking/rwsem: Wa...
367
368
369
370
371
372
373
374
   * Magic number to batch-wakeup waiting readers, even when writers are
   * also present in the queue. This both limits the amount of work the
   * waking thread must do and also prevents any potential counter overflow,
   * however unlikely.
   */
  #define MAX_READERS_WAKEUP	0x100
  
  /*
5dec94d49   Waiman Long   locking/rwsem: Me...
375
376
377
378
379
380
381
382
383
384
385
   * handle the lock release when processes blocked on it that can now run
   * - if we come here from up_xxxx(), then the RWSEM_FLAG_WAITERS bit must
   *   have been set.
   * - there must be someone on the queue
   * - the wait_lock must be held by the caller
   * - tasks are marked for wakeup, the caller must later invoke wake_up_q()
   *   to actually wakeup the blocked task(s) and drop the reference count,
   *   preferably when the wait_lock is released
   * - woken process blocks are discarded from the list after having task zeroed
   * - writers are only marked woken if downgrading is false
   */
6cef7ff6e   Waiman Long   locking/rwsem: Co...
386
387
388
  static void rwsem_mark_wake(struct rw_semaphore *sem,
  			    enum rwsem_wake_type wake_type,
  			    struct wake_q_head *wake_q)
5dec94d49   Waiman Long   locking/rwsem: Me...
389
390
391
392
  {
  	struct rwsem_waiter *waiter, *tmp;
  	long oldcount, woken = 0, adjustment = 0;
  	struct list_head wlist;
4f23dbc1e   Waiman Long   locking/rwsem: Im...
393
  	lockdep_assert_held(&sem->wait_lock);
5dec94d49   Waiman Long   locking/rwsem: Me...
394
395
396
397
  	/*
  	 * Take a peek at the queue head waiter such that we can determine
  	 * the wakeup(s) to perform.
  	 */
4f23dbc1e   Waiman Long   locking/rwsem: Im...
398
  	waiter = rwsem_first_waiter(sem);
5dec94d49   Waiman Long   locking/rwsem: Me...
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
  
  	if (waiter->type == RWSEM_WAITING_FOR_WRITE) {
  		if (wake_type == RWSEM_WAKE_ANY) {
  			/*
  			 * Mark writer at the front of the queue for wakeup.
  			 * Until the task is actually later awoken later by
  			 * the caller, other writers are able to steal it.
  			 * Readers, on the other hand, will block as they
  			 * will notice the queued writer.
  			 */
  			wake_q_add(wake_q, waiter->task);
  			lockevent_inc(rwsem_wake_writer);
  		}
  
  		return;
  	}
  
  	/*
a15ea1a35   Waiman Long   locking/rwsem: Gu...
417
418
419
420
421
422
  	 * No reader wakeup if there are too many of them already.
  	 */
  	if (unlikely(atomic_long_read(&sem->count) < 0))
  		return;
  
  	/*
5dec94d49   Waiman Long   locking/rwsem: Me...
423
424
425
426
427
  	 * Writers might steal the lock before we grant it to the next reader.
  	 * We prefer to do the first reader grant before counting readers
  	 * so we can bail out early if a writer stole the lock.
  	 */
  	if (wake_type != RWSEM_WAKE_READ_OWNED) {
5cfd92e12   Waiman Long   locking/rwsem: Ad...
428
  		struct task_struct *owner;
5dec94d49   Waiman Long   locking/rwsem: Me...
429
430
431
  		adjustment = RWSEM_READER_BIAS;
  		oldcount = atomic_long_fetch_add(adjustment, &sem->count);
  		if (unlikely(oldcount & RWSEM_WRITER_MASK)) {
4f23dbc1e   Waiman Long   locking/rwsem: Im...
432
433
434
435
436
437
438
439
440
441
442
443
  			/*
  			 * When we've been waiting "too" long (for writers
  			 * to give up the lock), request a HANDOFF to
  			 * force the issue.
  			 */
  			if (!(oldcount & RWSEM_FLAG_HANDOFF) &&
  			    time_after(jiffies, waiter->timeout)) {
  				adjustment -= RWSEM_FLAG_HANDOFF;
  				lockevent_inc(rwsem_rlock_handoff);
  			}
  
  			atomic_long_add(-adjustment, &sem->count);
5dec94d49   Waiman Long   locking/rwsem: Me...
444
445
446
447
448
  			return;
  		}
  		/*
  		 * Set it to reader-owned to give spinners an early
  		 * indication that readers now have the lock.
5cfd92e12   Waiman Long   locking/rwsem: Ad...
449
450
  		 * The reader nonspinnable bit seen at slowpath entry of
  		 * the reader is copied over.
5dec94d49   Waiman Long   locking/rwsem: Me...
451
  		 */
5cfd92e12   Waiman Long   locking/rwsem: Ad...
452
453
454
455
456
457
  		owner = waiter->task;
  		if (waiter->last_rowner & RWSEM_RD_NONSPINNABLE) {
  			owner = (void *)((unsigned long)owner | RWSEM_RD_NONSPINNABLE);
  			lockevent_inc(rwsem_opt_norspin);
  		}
  		__rwsem_set_reader_owned(sem, owner);
5dec94d49   Waiman Long   locking/rwsem: Me...
458
459
460
  	}
  
  	/*
d3681e269   Waiman Long   locking/rwsem: Wa...
461
462
  	 * Grant up to MAX_READERS_WAKEUP read locks to all the readers in the
  	 * queue. We know that the woken will be at least 1 as we accounted
5dec94d49   Waiman Long   locking/rwsem: Me...
463
464
465
  	 * for above. Note we increment the 'active part' of the count by the
  	 * number of readers before waking any processes up.
  	 *
d3681e269   Waiman Long   locking/rwsem: Wa...
466
467
468
469
470
471
  	 * This is an adaptation of the phase-fair R/W locks where at the
  	 * reader phase (first waiter is a reader), all readers are eligible
  	 * to acquire the lock at the same time irrespective of their order
  	 * in the queue. The writers acquire the lock according to their
  	 * order in the queue.
  	 *
5dec94d49   Waiman Long   locking/rwsem: Me...
472
473
474
475
476
477
478
479
480
481
482
  	 * We have to do wakeup in 2 passes to prevent the possibility that
  	 * the reader count may be decremented before it is incremented. It
  	 * is because the to-be-woken waiter may not have slept yet. So it
  	 * may see waiter->task got cleared, finish its critical section and
  	 * do an unlock before the reader count increment.
  	 *
  	 * 1) Collect the read-waiters in a separate list, count them and
  	 *    fully increment the reader count in rwsem.
  	 * 2) For each waiters in the new list, clear waiter->task and
  	 *    put them into wake_q to be woken up later.
  	 */
d3681e269   Waiman Long   locking/rwsem: Wa...
483
484
  	INIT_LIST_HEAD(&wlist);
  	list_for_each_entry_safe(waiter, tmp, &sem->wait_list, list) {
5dec94d49   Waiman Long   locking/rwsem: Me...
485
  		if (waiter->type == RWSEM_WAITING_FOR_WRITE)
d3681e269   Waiman Long   locking/rwsem: Wa...
486
  			continue;
5dec94d49   Waiman Long   locking/rwsem: Me...
487
488
  
  		woken++;
d3681e269   Waiman Long   locking/rwsem: Wa...
489
490
491
492
493
494
495
  		list_move_tail(&waiter->list, &wlist);
  
  		/*
  		 * Limit # of readers that can be woken up per wakeup call.
  		 */
  		if (woken >= MAX_READERS_WAKEUP)
  			break;
5dec94d49   Waiman Long   locking/rwsem: Me...
496
  	}
5dec94d49   Waiman Long   locking/rwsem: Me...
497
498
499
500
501
502
503
  
  	adjustment = woken * RWSEM_READER_BIAS - adjustment;
  	lockevent_cond_inc(rwsem_wake_reader, woken);
  	if (list_empty(&sem->wait_list)) {
  		/* hit end of list above */
  		adjustment -= RWSEM_FLAG_WAITERS;
  	}
4f23dbc1e   Waiman Long   locking/rwsem: Im...
504
505
506
507
508
509
  	/*
  	 * When we've woken a reader, we no longer need to force writers
  	 * to give up the lock and we can clear HANDOFF.
  	 */
  	if (woken && (atomic_long_read(&sem->count) & RWSEM_FLAG_HANDOFF))
  		adjustment -= RWSEM_FLAG_HANDOFF;
5dec94d49   Waiman Long   locking/rwsem: Me...
510
511
512
513
514
515
516
517
518
519
520
521
  	if (adjustment)
  		atomic_long_add(adjustment, &sem->count);
  
  	/* 2nd pass */
  	list_for_each_entry_safe(waiter, tmp, &wlist, list) {
  		struct task_struct *tsk;
  
  		tsk = waiter->task;
  		get_task_struct(tsk);
  
  		/*
  		 * Ensure calling get_task_struct() before setting the reader
6cef7ff6e   Waiman Long   locking/rwsem: Co...
522
  		 * waiter to nil such that rwsem_down_read_slowpath() cannot
5dec94d49   Waiman Long   locking/rwsem: Me...
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
  		 * race with do_exit() by always holding a reference count
  		 * to the task to wakeup.
  		 */
  		smp_store_release(&waiter->task, NULL);
  		/*
  		 * Ensure issuing the wakeup (either by us or someone else)
  		 * after setting the reader waiter to nil.
  		 */
  		wake_q_add_safe(wake_q, tsk);
  	}
  }
  
  /*
   * This function must be called with the sem->wait_lock held to prevent
   * race conditions between checking the rwsem wait list and setting the
   * sem->count accordingly.
4f23dbc1e   Waiman Long   locking/rwsem: Im...
539
540
541
   *
   * If wstate is WRITER_HANDOFF, it will make sure that either the handoff
   * bit is set or the lock is acquired with handoff bit cleared.
5dec94d49   Waiman Long   locking/rwsem: Me...
542
   */
00f3c5a3d   Waiman Long   locking/rwsem: Al...
543
  static inline bool rwsem_try_write_lock(struct rw_semaphore *sem,
4f23dbc1e   Waiman Long   locking/rwsem: Im...
544
  					enum writer_wait_state wstate)
5dec94d49   Waiman Long   locking/rwsem: Me...
545
  {
00f3c5a3d   Waiman Long   locking/rwsem: Al...
546
  	long count, new;
5dec94d49   Waiman Long   locking/rwsem: Me...
547

4f23dbc1e   Waiman Long   locking/rwsem: Im...
548
  	lockdep_assert_held(&sem->wait_lock);
5dec94d49   Waiman Long   locking/rwsem: Me...
549

00f3c5a3d   Waiman Long   locking/rwsem: Al...
550
  	count = atomic_long_read(&sem->count);
4f23dbc1e   Waiman Long   locking/rwsem: Im...
551
552
  	do {
  		bool has_handoff = !!(count & RWSEM_FLAG_HANDOFF);
5dec94d49   Waiman Long   locking/rwsem: Me...
553

4f23dbc1e   Waiman Long   locking/rwsem: Im...
554
555
  		if (has_handoff && wstate == WRITER_NOT_FIRST)
  			return false;
5dec94d49   Waiman Long   locking/rwsem: Me...
556

4f23dbc1e   Waiman Long   locking/rwsem: Im...
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
  		new = count;
  
  		if (count & RWSEM_LOCK_MASK) {
  			if (has_handoff || (wstate != WRITER_HANDOFF))
  				return false;
  
  			new |= RWSEM_FLAG_HANDOFF;
  		} else {
  			new |= RWSEM_WRITER_LOCKED;
  			new &= ~RWSEM_FLAG_HANDOFF;
  
  			if (list_is_singular(&sem->wait_list))
  				new &= ~RWSEM_FLAG_WAITERS;
  		}
  	} while (!atomic_long_try_cmpxchg_acquire(&sem->count, &count, new));
  
  	/*
  	 * We have either acquired the lock with handoff bit cleared or
  	 * set the handoff bit.
  	 */
  	if (new & RWSEM_FLAG_HANDOFF)
  		return false;
  
  	rwsem_set_owner(sem);
  	return true;
5dec94d49   Waiman Long   locking/rwsem: Me...
582
583
584
585
  }
  
  #ifdef CONFIG_RWSEM_SPIN_ON_OWNER
  /*
cf69482d6   Waiman Long   locking/rwsem: En...
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
   * Try to acquire read lock before the reader is put on wait queue.
   * Lock acquisition isn't allowed if the rwsem is locked or a writer handoff
   * is ongoing.
   */
  static inline bool rwsem_try_read_lock_unqueued(struct rw_semaphore *sem)
  {
  	long count = atomic_long_read(&sem->count);
  
  	if (count & (RWSEM_WRITER_MASK | RWSEM_FLAG_HANDOFF))
  		return false;
  
  	count = atomic_long_fetch_add_acquire(RWSEM_READER_BIAS, &sem->count);
  	if (!(count & (RWSEM_WRITER_MASK | RWSEM_FLAG_HANDOFF))) {
  		rwsem_set_reader_owned(sem);
  		lockevent_inc(rwsem_opt_rlock);
  		return true;
  	}
  
  	/* Back out the change */
  	atomic_long_add(-RWSEM_READER_BIAS, &sem->count);
  	return false;
  }
  
  /*
5dec94d49   Waiman Long   locking/rwsem: Me...
610
611
612
613
614
   * Try to acquire write lock before the writer has been put on wait queue.
   */
  static inline bool rwsem_try_write_lock_unqueued(struct rw_semaphore *sem)
  {
  	long count = atomic_long_read(&sem->count);
4f23dbc1e   Waiman Long   locking/rwsem: Im...
615
  	while (!(count & (RWSEM_LOCK_MASK|RWSEM_FLAG_HANDOFF))) {
5dec94d49   Waiman Long   locking/rwsem: Me...
616
  		if (atomic_long_try_cmpxchg_acquire(&sem->count, &count,
4f23dbc1e   Waiman Long   locking/rwsem: Im...
617
  					count | RWSEM_WRITER_LOCKED)) {
5dec94d49   Waiman Long   locking/rwsem: Me...
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
  			rwsem_set_owner(sem);
  			lockevent_inc(rwsem_opt_wlock);
  			return true;
  		}
  	}
  	return false;
  }
  
  static inline bool owner_on_cpu(struct task_struct *owner)
  {
  	/*
  	 * As lock holder preemption issue, we both skip spinning if
  	 * task is not on cpu or its cpu is preempted
  	 */
  	return owner->on_cpu && !vcpu_is_preempted(task_cpu(owner));
  }
7d43f1ce9   Waiman Long   locking/rwsem: En...
634
635
  static inline bool rwsem_can_spin_on_owner(struct rw_semaphore *sem,
  					   unsigned long nonspinnable)
5dec94d49   Waiman Long   locking/rwsem: Me...
636
637
  {
  	struct task_struct *owner;
94a9717b3   Waiman Long   locking/rwsem: Ma...
638
  	unsigned long flags;
5dec94d49   Waiman Long   locking/rwsem: Me...
639
  	bool ret = true;
cf69482d6   Waiman Long   locking/rwsem: En...
640
641
  	if (need_resched()) {
  		lockevent_inc(rwsem_opt_fail);
5dec94d49   Waiman Long   locking/rwsem: Me...
642
  		return false;
cf69482d6   Waiman Long   locking/rwsem: En...
643
  	}
5dec94d49   Waiman Long   locking/rwsem: Me...
644

cf69482d6   Waiman Long   locking/rwsem: En...
645
  	preempt_disable();
5dec94d49   Waiman Long   locking/rwsem: Me...
646
  	rcu_read_lock();
94a9717b3   Waiman Long   locking/rwsem: Ma...
647
  	owner = rwsem_owner_flags(sem, &flags);
781343005   Waiman Long   locking/rwsem: Do...
648
649
650
651
652
  	/*
  	 * Don't check the read-owner as the entry may be stale.
  	 */
  	if ((flags & nonspinnable) ||
  	    (owner && !(flags & RWSEM_READER_OWNED) && !owner_on_cpu(owner)))
94a9717b3   Waiman Long   locking/rwsem: Ma...
653
  		ret = false;
5dec94d49   Waiman Long   locking/rwsem: Me...
654
  	rcu_read_unlock();
cf69482d6   Waiman Long   locking/rwsem: En...
655
656
657
  	preempt_enable();
  
  	lockevent_cond_inc(rwsem_opt_fail, !ret);
5dec94d49   Waiman Long   locking/rwsem: Me...
658
659
660
661
  	return ret;
  }
  
  /*
3f6d517a3   Waiman Long   locking/rwsem: Ma...
662
663
664
665
666
667
668
669
670
   * The rwsem_spin_on_owner() function returns the folowing 4 values
   * depending on the lock owner state.
   *   OWNER_NULL  : owner is currently NULL
   *   OWNER_WRITER: when owner changes and is a writer
   *   OWNER_READER: when owner changes and the new owner may be a reader.
   *   OWNER_NONSPINNABLE:
   *		   when optimistic spinning has to stop because either the
   *		   owner stops running, is unknown, or its timeslice has
   *		   been used up.
5dec94d49   Waiman Long   locking/rwsem: Me...
671
   */
3f6d517a3   Waiman Long   locking/rwsem: Ma...
672
673
674
675
676
677
  enum owner_state {
  	OWNER_NULL		= 1 << 0,
  	OWNER_WRITER		= 1 << 1,
  	OWNER_READER		= 1 << 2,
  	OWNER_NONSPINNABLE	= 1 << 3,
  };
7d43f1ce9   Waiman Long   locking/rwsem: En...
678
  #define OWNER_SPINNABLE		(OWNER_NULL | OWNER_WRITER | OWNER_READER)
3f6d517a3   Waiman Long   locking/rwsem: Ma...
679

94a9717b3   Waiman Long   locking/rwsem: Ma...
680
  static inline enum owner_state
7d43f1ce9   Waiman Long   locking/rwsem: En...
681
  rwsem_owner_state(struct task_struct *owner, unsigned long flags, unsigned long nonspinnable)
5dec94d49   Waiman Long   locking/rwsem: Me...
682
  {
7d43f1ce9   Waiman Long   locking/rwsem: En...
683
  	if (flags & nonspinnable)
3f6d517a3   Waiman Long   locking/rwsem: Ma...
684
  		return OWNER_NONSPINNABLE;
94a9717b3   Waiman Long   locking/rwsem: Ma...
685
  	if (flags & RWSEM_READER_OWNED)
3f6d517a3   Waiman Long   locking/rwsem: Ma...
686
  		return OWNER_READER;
94a9717b3   Waiman Long   locking/rwsem: Ma...
687
  	return owner ? OWNER_WRITER : OWNER_NULL;
3f6d517a3   Waiman Long   locking/rwsem: Ma...
688
  }
7d43f1ce9   Waiman Long   locking/rwsem: En...
689
690
  static noinline enum owner_state
  rwsem_spin_on_owner(struct rw_semaphore *sem, unsigned long nonspinnable)
3f6d517a3   Waiman Long   locking/rwsem: Ma...
691
  {
94a9717b3   Waiman Long   locking/rwsem: Ma...
692
693
694
  	struct task_struct *new, *owner;
  	unsigned long flags, new_flags;
  	enum owner_state state;
3f6d517a3   Waiman Long   locking/rwsem: Ma...
695

94a9717b3   Waiman Long   locking/rwsem: Ma...
696
  	owner = rwsem_owner_flags(sem, &flags);
7d43f1ce9   Waiman Long   locking/rwsem: En...
697
  	state = rwsem_owner_state(owner, flags, nonspinnable);
3f6d517a3   Waiman Long   locking/rwsem: Ma...
698
699
  	if (state != OWNER_WRITER)
  		return state;
5dec94d49   Waiman Long   locking/rwsem: Me...
700
701
  
  	rcu_read_lock();
3f6d517a3   Waiman Long   locking/rwsem: Ma...
702
  	for (;;) {
91d2a812d   Waiman Long   locking/rwsem: Ma...
703
704
705
706
707
708
  		/*
  		 * When a waiting writer set the handoff flag, it may spin
  		 * on the owner as well. Once that writer acquires the lock,
  		 * we can spin on it. So we don't need to quit even when the
  		 * handoff bit is set.
  		 */
94a9717b3   Waiman Long   locking/rwsem: Ma...
709
710
  		new = rwsem_owner_flags(sem, &new_flags);
  		if ((new != owner) || (new_flags != flags)) {
7d43f1ce9   Waiman Long   locking/rwsem: En...
711
  			state = rwsem_owner_state(new, new_flags, nonspinnable);
3f6d517a3   Waiman Long   locking/rwsem: Ma...
712
713
  			break;
  		}
5dec94d49   Waiman Long   locking/rwsem: Me...
714
715
716
717
718
719
720
  		/*
  		 * Ensure we emit the owner->on_cpu, dereference _after_
  		 * checking sem->owner still matches owner, if that fails,
  		 * owner might point to free()d memory, if it still matches,
  		 * the rcu_read_lock() ensures the memory stays valid.
  		 */
  		barrier();
5dec94d49   Waiman Long   locking/rwsem: Me...
721
  		if (need_resched() || !owner_on_cpu(owner)) {
3f6d517a3   Waiman Long   locking/rwsem: Ma...
722
723
  			state = OWNER_NONSPINNABLE;
  			break;
5dec94d49   Waiman Long   locking/rwsem: Me...
724
725
726
727
728
  		}
  
  		cpu_relax();
  	}
  	rcu_read_unlock();
3f6d517a3   Waiman Long   locking/rwsem: Ma...
729
  	return state;
5dec94d49   Waiman Long   locking/rwsem: Me...
730
  }
7d43f1ce9   Waiman Long   locking/rwsem: En...
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
  /*
   * Calculate reader-owned rwsem spinning threshold for writer
   *
   * The more readers own the rwsem, the longer it will take for them to
   * wind down and free the rwsem. So the empirical formula used to
   * determine the actual spinning time limit here is:
   *
   *   Spinning threshold = (10 + nr_readers/2)us
   *
   * The limit is capped to a maximum of 25us (30 readers). This is just
   * a heuristic and is subjected to change in the future.
   */
  static inline u64 rwsem_rspin_threshold(struct rw_semaphore *sem)
  {
  	long count = atomic_long_read(&sem->count);
  	int readers = count >> RWSEM_READER_SHIFT;
  	u64 delta;
  
  	if (readers > 30)
  		readers = 30;
  	delta = (20 + readers) * NSEC_PER_USEC / 2;
  
  	return sched_clock() + delta;
  }
cf69482d6   Waiman Long   locking/rwsem: En...
755
  static bool rwsem_optimistic_spin(struct rw_semaphore *sem, bool wlock)
5dec94d49   Waiman Long   locking/rwsem: Me...
756
757
  {
  	bool taken = false;
990fa7384   Waiman Long   locking/rwsem: Mo...
758
  	int prev_owner_state = OWNER_NULL;
7d43f1ce9   Waiman Long   locking/rwsem: En...
759
760
761
762
  	int loop = 0;
  	u64 rspin_threshold = 0;
  	unsigned long nonspinnable = wlock ? RWSEM_WR_NONSPINNABLE
  					   : RWSEM_RD_NONSPINNABLE;
5dec94d49   Waiman Long   locking/rwsem: Me...
763
764
765
766
  
  	preempt_disable();
  
  	/* sem->wait_lock should not be held when doing optimistic spinning */
5dec94d49   Waiman Long   locking/rwsem: Me...
767
768
769
770
771
772
773
  	if (!osq_lock(&sem->osq))
  		goto done;
  
  	/*
  	 * Optimistically spin on the owner field and attempt to acquire the
  	 * lock whenever the owner changes. Spinning will be stopped when:
  	 *  1) the owning writer isn't running; or
7d43f1ce9   Waiman Long   locking/rwsem: En...
774
  	 *  2) readers own the lock and spinning time has exceeded limit.
5dec94d49   Waiman Long   locking/rwsem: Me...
775
  	 */
990fa7384   Waiman Long   locking/rwsem: Mo...
776
  	for (;;) {
7d43f1ce9   Waiman Long   locking/rwsem: En...
777
  		enum owner_state owner_state;
990fa7384   Waiman Long   locking/rwsem: Mo...
778

7d43f1ce9   Waiman Long   locking/rwsem: En...
779
  		owner_state = rwsem_spin_on_owner(sem, nonspinnable);
990fa7384   Waiman Long   locking/rwsem: Mo...
780
781
  		if (!(owner_state & OWNER_SPINNABLE))
  			break;
5dec94d49   Waiman Long   locking/rwsem: Me...
782
783
784
  		/*
  		 * Try to acquire the lock
  		 */
cf69482d6   Waiman Long   locking/rwsem: En...
785
786
787
788
  		taken = wlock ? rwsem_try_write_lock_unqueued(sem)
  			      : rwsem_try_read_lock_unqueued(sem);
  
  		if (taken)
5dec94d49   Waiman Long   locking/rwsem: Me...
789
  			break;
5dec94d49   Waiman Long   locking/rwsem: Me...
790
791
  
  		/*
7d43f1ce9   Waiman Long   locking/rwsem: En...
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
  		 * Time-based reader-owned rwsem optimistic spinning
  		 */
  		if (wlock && (owner_state == OWNER_READER)) {
  			/*
  			 * Re-initialize rspin_threshold every time when
  			 * the owner state changes from non-reader to reader.
  			 * This allows a writer to steal the lock in between
  			 * 2 reader phases and have the threshold reset at
  			 * the beginning of the 2nd reader phase.
  			 */
  			if (prev_owner_state != OWNER_READER) {
  				if (rwsem_test_oflags(sem, nonspinnable))
  					break;
  				rspin_threshold = rwsem_rspin_threshold(sem);
  				loop = 0;
  			}
  
  			/*
  			 * Check time threshold once every 16 iterations to
  			 * avoid calling sched_clock() too frequently so
  			 * as to reduce the average latency between the times
  			 * when the lock becomes free and when the spinner
  			 * is ready to do a trylock.
  			 */
  			else if (!(++loop & 0xf) && (sched_clock() > rspin_threshold)) {
  				rwsem_set_nonspinnable(sem);
  				lockevent_inc(rwsem_opt_nospin);
  				break;
  			}
  		}
  
  		/*
990fa7384   Waiman Long   locking/rwsem: Mo...
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
  		 * An RT task cannot do optimistic spinning if it cannot
  		 * be sure the lock holder is running or live-lock may
  		 * happen if the current task and the lock holder happen
  		 * to run in the same CPU. However, aborting optimistic
  		 * spinning while a NULL owner is detected may miss some
  		 * opportunity where spinning can continue without causing
  		 * problem.
  		 *
  		 * There are 2 possible cases where an RT task may be able
  		 * to continue spinning.
  		 *
  		 * 1) The lock owner is in the process of releasing the
  		 *    lock, sem->owner is cleared but the lock has not
  		 *    been released yet.
  		 * 2) The lock was free and owner cleared, but another
  		 *    task just comes in and acquire the lock before
  		 *    we try to get it. The new owner may be a spinnable
  		 *    writer.
  		 *
  		 * To take advantage of two scenarios listed agove, the RT
  		 * task is made to retry one more time to see if it can
  		 * acquire the lock or continue spinning on the new owning
  		 * writer. Of course, if the time lag is long enough or the
  		 * new owner is not a writer or spinnable, the RT task will
  		 * quit spinning.
  		 *
  		 * If the owner is a writer, the need_resched() check is
  		 * done inside rwsem_spin_on_owner(). If the owner is not
  		 * a writer, need_resched() check needs to be done here.
5dec94d49   Waiman Long   locking/rwsem: Me...
853
  		 */
990fa7384   Waiman Long   locking/rwsem: Mo...
854
855
856
857
858
859
860
861
  		if (owner_state != OWNER_WRITER) {
  			if (need_resched())
  				break;
  			if (rt_task(current) &&
  			   (prev_owner_state != OWNER_WRITER))
  				break;
  		}
  		prev_owner_state = owner_state;
5dec94d49   Waiman Long   locking/rwsem: Me...
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
  
  		/*
  		 * The cpu_relax() call is a compiler barrier which forces
  		 * everything in this loop to be re-loaded. We don't need
  		 * memory barriers as we'll eventually observe the right
  		 * values at the cost of a few extra spins.
  		 */
  		cpu_relax();
  	}
  	osq_unlock(&sem->osq);
  done:
  	preempt_enable();
  	lockevent_cond_inc(rwsem_opt_fail, !taken);
  	return taken;
  }
7d43f1ce9   Waiman Long   locking/rwsem: En...
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
  
  /*
   * Clear the owner's RWSEM_WR_NONSPINNABLE bit if it is set. This should
   * only be called when the reader count reaches 0.
   *
   * This give writers better chance to acquire the rwsem first before
   * readers when the rwsem was being held by readers for a relatively long
   * period of time. Race can happen that an optimistic spinner may have
   * just stolen the rwsem and set the owner, but just clearing the
   * RWSEM_WR_NONSPINNABLE bit will do no harm anyway.
   */
  static inline void clear_wr_nonspinnable(struct rw_semaphore *sem)
  {
  	if (rwsem_test_oflags(sem, RWSEM_WR_NONSPINNABLE))
  		atomic_long_andnot(RWSEM_WR_NONSPINNABLE, &sem->owner);
  }
5cfd92e12   Waiman Long   locking/rwsem: Ad...
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
  
  /*
   * This function is called when the reader fails to acquire the lock via
   * optimistic spinning. In this case we will still attempt to do a trylock
   * when comparing the rwsem state right now with the state when entering
   * the slowpath indicates that the reader is still in a valid reader phase.
   * This happens when the following conditions are true:
   *
   * 1) The lock is currently reader owned, and
   * 2) The lock is previously not reader-owned or the last read owner changes.
   *
   * In the former case, we have transitioned from a writer phase to a
   * reader-phase while spinning. In the latter case, it means the reader
   * phase hasn't ended when we entered the optimistic spinning loop. In
   * both cases, the reader is eligible to acquire the lock. This is the
   * secondary path where a read lock is acquired optimistically.
   *
   * The reader non-spinnable bit wasn't set at time of entry or it will
   * not be here at all.
   */
  static inline bool rwsem_reader_phase_trylock(struct rw_semaphore *sem,
  					      unsigned long last_rowner)
  {
  	unsigned long owner = atomic_long_read(&sem->owner);
  
  	if (!(owner & RWSEM_READER_OWNED))
  		return false;
  
  	if (((owner ^ last_rowner) & ~RWSEM_OWNER_FLAGS_MASK) &&
  	    rwsem_try_read_lock_unqueued(sem)) {
  		lockevent_inc(rwsem_opt_rlock2);
  		lockevent_add(rwsem_opt_fail, -1);
  		return true;
  	}
  	return false;
  }
5dec94d49   Waiman Long   locking/rwsem: Me...
929
  #else
7d43f1ce9   Waiman Long   locking/rwsem: En...
930
931
  static inline bool rwsem_can_spin_on_owner(struct rw_semaphore *sem,
  					   unsigned long nonspinnable)
cf69482d6   Waiman Long   locking/rwsem: En...
932
933
934
935
936
  {
  	return false;
  }
  
  static inline bool rwsem_optimistic_spin(struct rw_semaphore *sem, bool wlock)
5dec94d49   Waiman Long   locking/rwsem: Me...
937
938
939
  {
  	return false;
  }
7d43f1ce9   Waiman Long   locking/rwsem: En...
940
941
  
  static inline void clear_wr_nonspinnable(struct rw_semaphore *sem) { }
5cfd92e12   Waiman Long   locking/rwsem: Ad...
942
943
944
945
946
947
  
  static inline bool rwsem_reader_phase_trylock(struct rw_semaphore *sem,
  					      unsigned long last_rowner)
  {
  	return false;
  }
91d2a812d   Waiman Long   locking/rwsem: Ma...
948
949
950
951
952
953
954
  
  static inline int
  rwsem_spin_on_owner(struct rw_semaphore *sem, unsigned long nonspinnable)
  {
  	return 0;
  }
  #define OWNER_NULL	1
5dec94d49   Waiman Long   locking/rwsem: Me...
955
956
957
958
959
  #endif
  
  /*
   * Wait for the read lock to be granted
   */
6cef7ff6e   Waiman Long   locking/rwsem: Co...
960
961
  static struct rw_semaphore __sched *
  rwsem_down_read_slowpath(struct rw_semaphore *sem, int state)
5dec94d49   Waiman Long   locking/rwsem: Me...
962
963
964
965
  {
  	long count, adjustment = -RWSEM_READER_BIAS;
  	struct rwsem_waiter waiter;
  	DEFINE_WAKE_Q(wake_q);
a15ea1a35   Waiman Long   locking/rwsem: Gu...
966
  	bool wake = false;
df18d99da   JianMin Liu   ANDROID: rwsem: A...
967
  	bool already_on_list = false;
5dec94d49   Waiman Long   locking/rwsem: Me...
968

5cfd92e12   Waiman Long   locking/rwsem: Ad...
969
970
971
972
973
974
975
  	/*
  	 * Save the current read-owner of rwsem, if available, and the
  	 * reader nonspinnable bit.
  	 */
  	waiter.last_rowner = atomic_long_read(&sem->owner);
  	if (!(waiter.last_rowner & RWSEM_READER_OWNED))
  		waiter.last_rowner &= RWSEM_RD_NONSPINNABLE;
7d43f1ce9   Waiman Long   locking/rwsem: En...
976
  	if (!rwsem_can_spin_on_owner(sem, RWSEM_RD_NONSPINNABLE))
cf69482d6   Waiman Long   locking/rwsem: En...
977
978
979
980
981
982
983
984
  		goto queue;
  
  	/*
  	 * Undo read bias from down_read() and do optimistic spinning.
  	 */
  	atomic_long_add(-RWSEM_READER_BIAS, &sem->count);
  	adjustment = 0;
  	if (rwsem_optimistic_spin(sem, false)) {
6ffddfb9e   Peter Zijlstra   locking/rwsem: Ad...
985
  		/* rwsem_optimistic_spin() implies ACQUIRE on success */
cf69482d6   Waiman Long   locking/rwsem: En...
986
987
988
989
990
991
992
993
994
995
996
997
998
  		/*
  		 * Wake up other readers in the wait list if the front
  		 * waiter is a reader.
  		 */
  		if ((atomic_long_read(&sem->count) & RWSEM_FLAG_WAITERS)) {
  			raw_spin_lock_irq(&sem->wait_lock);
  			if (!list_empty(&sem->wait_list))
  				rwsem_mark_wake(sem, RWSEM_WAKE_READ_OWNED,
  						&wake_q);
  			raw_spin_unlock_irq(&sem->wait_lock);
  			wake_up_q(&wake_q);
  		}
  		return sem;
5cfd92e12   Waiman Long   locking/rwsem: Ad...
999
  	} else if (rwsem_reader_phase_trylock(sem, waiter.last_rowner)) {
6ffddfb9e   Peter Zijlstra   locking/rwsem: Ad...
1000
  		/* rwsem_reader_phase_trylock() implies ACQUIRE on success */
5cfd92e12   Waiman Long   locking/rwsem: Ad...
1001
  		return sem;
cf69482d6   Waiman Long   locking/rwsem: En...
1002
1003
1004
  	}
  
  queue:
5dec94d49   Waiman Long   locking/rwsem: Me...
1005
1006
  	waiter.task = current;
  	waiter.type = RWSEM_WAITING_FOR_READ;
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1007
  	waiter.timeout = jiffies + RWSEM_WAIT_TIMEOUT;
5dec94d49   Waiman Long   locking/rwsem: Me...
1008
1009
1010
1011
1012
  
  	raw_spin_lock_irq(&sem->wait_lock);
  	if (list_empty(&sem->wait_list)) {
  		/*
  		 * In case the wait queue is empty and the lock isn't owned
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1013
1014
1015
  		 * by a writer or has the handoff bit set, this reader can
  		 * exit the slowpath and return immediately as its
  		 * RWSEM_READER_BIAS has already been set in the count.
5dec94d49   Waiman Long   locking/rwsem: Me...
1016
  		 */
cf69482d6   Waiman Long   locking/rwsem: En...
1017
  		if (adjustment && !(atomic_long_read(&sem->count) &
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1018
  		     (RWSEM_WRITER_MASK | RWSEM_FLAG_HANDOFF))) {
e1b98fa31   Jan Stancek   locking/rwsem: Ad...
1019
1020
  			/* Provide lock ACQUIRE */
  			smp_acquire__after_ctrl_dep();
5dec94d49   Waiman Long   locking/rwsem: Me...
1021
1022
1023
1024
1025
1026
1027
  			raw_spin_unlock_irq(&sem->wait_lock);
  			rwsem_set_reader_owned(sem);
  			lockevent_inc(rwsem_rlock_fast);
  			return sem;
  		}
  		adjustment += RWSEM_FLAG_WAITERS;
  	}
df18d99da   JianMin Liu   ANDROID: rwsem: A...
1028
1029
1030
1031
1032
  	trace_android_vh_alter_rwsem_list_add(
  					&waiter,
  					sem, &already_on_list);
  	if (!already_on_list)
  		list_add_tail(&waiter.list, &sem->wait_list);
5dec94d49   Waiman Long   locking/rwsem: Me...
1033
1034
  
  	/* we're now waiting on the lock, but no longer actively locking */
cf69482d6   Waiman Long   locking/rwsem: En...
1035
1036
1037
1038
  	if (adjustment)
  		count = atomic_long_add_return(adjustment, &sem->count);
  	else
  		count = atomic_long_read(&sem->count);
5dec94d49   Waiman Long   locking/rwsem: Me...
1039
1040
1041
1042
1043
1044
1045
  
  	/*
  	 * If there are no active locks, wake the front queued process(es).
  	 *
  	 * If there are no writers and we are first in the queue,
  	 * wake our own waiter to join the existing active readers !
  	 */
7d43f1ce9   Waiman Long   locking/rwsem: En...
1046
1047
1048
1049
1050
1051
  	if (!(count & RWSEM_LOCK_MASK)) {
  		clear_wr_nonspinnable(sem);
  		wake = true;
  	}
  	if (wake || (!(count & RWSEM_WRITER_MASK) &&
  		    (adjustment & RWSEM_FLAG_WAITERS)))
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1052
  		rwsem_mark_wake(sem, RWSEM_WAKE_ANY, &wake_q);
5dec94d49   Waiman Long   locking/rwsem: Me...
1053

df18d99da   JianMin Liu   ANDROID: rwsem: A...
1054
  	trace_android_vh_rwsem_wake(sem);
5dec94d49   Waiman Long   locking/rwsem: Me...
1055
1056
1057
1058
  	raw_spin_unlock_irq(&sem->wait_lock);
  	wake_up_q(&wake_q);
  
  	/* wait to be given the lock */
9ad8ff902   Sangmoon Kim   ANDROID: vendor_h...
1059
  	trace_android_vh_rwsem_read_wait_start(sem);
6ffddfb9e   Peter Zijlstra   locking/rwsem: Ad...
1060
  	for (;;) {
5dec94d49   Waiman Long   locking/rwsem: Me...
1061
  		set_current_state(state);
99143f82a   Peter Zijlstra   lcoking/rwsem: Ad...
1062
  		if (!smp_load_acquire(&waiter.task)) {
6ffddfb9e   Peter Zijlstra   locking/rwsem: Ad...
1063
  			/* Matches rwsem_mark_wake()'s smp_store_release(). */
5dec94d49   Waiman Long   locking/rwsem: Me...
1064
  			break;
99143f82a   Peter Zijlstra   lcoking/rwsem: Ad...
1065
  		}
5dec94d49   Waiman Long   locking/rwsem: Me...
1066
1067
1068
1069
1070
  		if (signal_pending_state(state, current)) {
  			raw_spin_lock_irq(&sem->wait_lock);
  			if (waiter.task)
  				goto out_nolock;
  			raw_spin_unlock_irq(&sem->wait_lock);
6ffddfb9e   Peter Zijlstra   locking/rwsem: Ad...
1071
  			/* Ordered by sem->wait_lock against rwsem_mark_wake(). */
5dec94d49   Waiman Long   locking/rwsem: Me...
1072
1073
1074
1075
1076
1077
1078
  			break;
  		}
  		schedule();
  		lockevent_inc(rwsem_sleep_reader);
  	}
  
  	__set_current_state(TASK_RUNNING);
9ad8ff902   Sangmoon Kim   ANDROID: vendor_h...
1079
  	trace_android_vh_rwsem_read_wait_finish(sem);
5dec94d49   Waiman Long   locking/rwsem: Me...
1080
1081
  	lockevent_inc(rwsem_rlock);
  	return sem;
6ffddfb9e   Peter Zijlstra   locking/rwsem: Ad...
1082

5dec94d49   Waiman Long   locking/rwsem: Me...
1083
1084
  out_nolock:
  	list_del(&waiter.list);
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1085
1086
1087
1088
  	if (list_empty(&sem->wait_list)) {
  		atomic_long_andnot(RWSEM_FLAG_WAITERS|RWSEM_FLAG_HANDOFF,
  				   &sem->count);
  	}
5dec94d49   Waiman Long   locking/rwsem: Me...
1089
1090
  	raw_spin_unlock_irq(&sem->wait_lock);
  	__set_current_state(TASK_RUNNING);
9ad8ff902   Sangmoon Kim   ANDROID: vendor_h...
1091
  	trace_android_vh_rwsem_read_wait_finish(sem);
5dec94d49   Waiman Long   locking/rwsem: Me...
1092
1093
1094
  	lockevent_inc(rwsem_rlock_fail);
  	return ERR_PTR(-EINTR);
  }
5dec94d49   Waiman Long   locking/rwsem: Me...
1095
  /*
5cfd92e12   Waiman Long   locking/rwsem: Ad...
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
   * This function is called by the a write lock owner. So the owner value
   * won't get changed by others.
   */
  static inline void rwsem_disable_reader_optspin(struct rw_semaphore *sem,
  						bool disable)
  {
  	if (unlikely(disable)) {
  		atomic_long_or(RWSEM_RD_NONSPINNABLE, &sem->owner);
  		lockevent_inc(rwsem_opt_norspin);
  	}
  }
  
  /*
5dec94d49   Waiman Long   locking/rwsem: Me...
1109
1110
   * Wait until we successfully acquire the write lock
   */
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1111
1112
  static struct rw_semaphore *
  rwsem_down_write_slowpath(struct rw_semaphore *sem, int state)
5dec94d49   Waiman Long   locking/rwsem: Me...
1113
1114
  {
  	long count;
5cfd92e12   Waiman Long   locking/rwsem: Ad...
1115
  	bool disable_rspin;
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1116
  	enum writer_wait_state wstate;
5dec94d49   Waiman Long   locking/rwsem: Me...
1117
1118
1119
  	struct rwsem_waiter waiter;
  	struct rw_semaphore *ret = sem;
  	DEFINE_WAKE_Q(wake_q);
df18d99da   JianMin Liu   ANDROID: rwsem: A...
1120
  	bool already_on_list = false;
5dec94d49   Waiman Long   locking/rwsem: Me...
1121
1122
  
  	/* do optimistic spinning and steal lock if possible */
7d43f1ce9   Waiman Long   locking/rwsem: En...
1123
  	if (rwsem_can_spin_on_owner(sem, RWSEM_WR_NONSPINNABLE) &&
6ffddfb9e   Peter Zijlstra   locking/rwsem: Ad...
1124
1125
  	    rwsem_optimistic_spin(sem, true)) {
  		/* rwsem_optimistic_spin() implies ACQUIRE on success */
5dec94d49   Waiman Long   locking/rwsem: Me...
1126
  		return sem;
6ffddfb9e   Peter Zijlstra   locking/rwsem: Ad...
1127
  	}
5dec94d49   Waiman Long   locking/rwsem: Me...
1128
1129
  
  	/*
5cfd92e12   Waiman Long   locking/rwsem: Ad...
1130
1131
1132
1133
1134
1135
1136
  	 * Disable reader optimistic spinning for this rwsem after
  	 * acquiring the write lock when the setting of the nonspinnable
  	 * bits are observed.
  	 */
  	disable_rspin = atomic_long_read(&sem->owner) & RWSEM_NONSPINNABLE;
  
  	/*
5dec94d49   Waiman Long   locking/rwsem: Me...
1137
1138
1139
1140
1141
  	 * Optimistic spinning failed, proceed to the slowpath
  	 * and block until we can acquire the sem.
  	 */
  	waiter.task = current;
  	waiter.type = RWSEM_WAITING_FOR_WRITE;
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1142
  	waiter.timeout = jiffies + RWSEM_WAIT_TIMEOUT;
5dec94d49   Waiman Long   locking/rwsem: Me...
1143
1144
1145
1146
  
  	raw_spin_lock_irq(&sem->wait_lock);
  
  	/* account for this before adding a new element to the list */
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1147
  	wstate = list_empty(&sem->wait_list) ? WRITER_FIRST : WRITER_NOT_FIRST;
5dec94d49   Waiman Long   locking/rwsem: Me...
1148

df18d99da   JianMin Liu   ANDROID: rwsem: A...
1149
1150
1151
1152
1153
  	trace_android_vh_alter_rwsem_list_add(
  					&waiter,
  					sem, &already_on_list);
  	if (!already_on_list)
  		list_add_tail(&waiter.list, &sem->wait_list);
5dec94d49   Waiman Long   locking/rwsem: Me...
1154
1155
  
  	/* we're now waiting on the lock */
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1156
  	if (wstate == WRITER_NOT_FIRST) {
5dec94d49   Waiman Long   locking/rwsem: Me...
1157
1158
1159
  		count = atomic_long_read(&sem->count);
  
  		/*
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1160
1161
1162
1163
1164
1165
  		 * If there were already threads queued before us and:
  		 *  1) there are no no active locks, wake the front
  		 *     queued process(es) as the handoff bit might be set.
  		 *  2) there are no active writers and some readers, the lock
  		 *     must be read owned; so we try to wake any read lock
  		 *     waiters that were queued ahead of us.
5dec94d49   Waiman Long   locking/rwsem: Me...
1166
  		 */
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1167
1168
  		if (count & RWSEM_WRITER_MASK)
  			goto wait;
5dec94d49   Waiman Long   locking/rwsem: Me...
1169

4f23dbc1e   Waiman Long   locking/rwsem: Im...
1170
1171
1172
  		rwsem_mark_wake(sem, (count & RWSEM_READER_MASK)
  					? RWSEM_WAKE_READERS
  					: RWSEM_WAKE_ANY, &wake_q);
5dec94d49   Waiman Long   locking/rwsem: Me...
1173

00f3c5a3d   Waiman Long   locking/rwsem: Al...
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
  		if (!wake_q_empty(&wake_q)) {
  			/*
  			 * We want to minimize wait_lock hold time especially
  			 * when a large number of readers are to be woken up.
  			 */
  			raw_spin_unlock_irq(&sem->wait_lock);
  			wake_up_q(&wake_q);
  			wake_q_init(&wake_q);	/* Used again, reinit */
  			raw_spin_lock_irq(&sem->wait_lock);
  		}
5dec94d49   Waiman Long   locking/rwsem: Me...
1184
  	} else {
00f3c5a3d   Waiman Long   locking/rwsem: Al...
1185
  		atomic_long_or(RWSEM_FLAG_WAITERS, &sem->count);
5dec94d49   Waiman Long   locking/rwsem: Me...
1186
  	}
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1187
  wait:
df18d99da   JianMin Liu   ANDROID: rwsem: A...
1188
  	trace_android_vh_rwsem_wake(sem);
5dec94d49   Waiman Long   locking/rwsem: Me...
1189
  	/* wait until we successfully acquire the lock */
9ad8ff902   Sangmoon Kim   ANDROID: vendor_h...
1190
  	trace_android_vh_rwsem_write_wait_start(sem);
5dec94d49   Waiman Long   locking/rwsem: Me...
1191
  	set_current_state(state);
6ffddfb9e   Peter Zijlstra   locking/rwsem: Ad...
1192
1193
1194
  	for (;;) {
  		if (rwsem_try_write_lock(sem, wstate)) {
  			/* rwsem_try_write_lock() implies ACQUIRE on success */
5dec94d49   Waiman Long   locking/rwsem: Me...
1195
  			break;
6ffddfb9e   Peter Zijlstra   locking/rwsem: Ad...
1196
  		}
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1197

5dec94d49   Waiman Long   locking/rwsem: Me...
1198
  		raw_spin_unlock_irq(&sem->wait_lock);
91d2a812d   Waiman Long   locking/rwsem: Ma...
1199
1200
1201
1202
1203
1204
1205
1206
  		/*
  		 * After setting the handoff bit and failing to acquire
  		 * the lock, attempt to spin on owner to accelerate lock
  		 * transfer. If the previous owner is a on-cpu writer and it
  		 * has just released the lock, OWNER_NULL will be returned.
  		 * In this case, we attempt to acquire the lock again
  		 * without sleeping.
  		 */
39e7234f0   Waiman Long   locking/rwsem: Fi...
1207
1208
  		if (wstate == WRITER_HANDOFF &&
  		    rwsem_spin_on_owner(sem, RWSEM_NONSPINNABLE) == OWNER_NULL)
91d2a812d   Waiman Long   locking/rwsem: Ma...
1209
  			goto trylock_again;
5dec94d49   Waiman Long   locking/rwsem: Me...
1210
  		/* Block until there are no active lockers. */
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1211
  		for (;;) {
5dec94d49   Waiman Long   locking/rwsem: Me...
1212
1213
1214
1215
1216
1217
  			if (signal_pending_state(state, current))
  				goto out_nolock;
  
  			schedule();
  			lockevent_inc(rwsem_sleep_writer);
  			set_current_state(state);
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
  			/*
  			 * If HANDOFF bit is set, unconditionally do
  			 * a trylock.
  			 */
  			if (wstate == WRITER_HANDOFF)
  				break;
  
  			if ((wstate == WRITER_NOT_FIRST) &&
  			    (rwsem_first_waiter(sem) == &waiter))
  				wstate = WRITER_FIRST;
5dec94d49   Waiman Long   locking/rwsem: Me...
1228
  			count = atomic_long_read(&sem->count);
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
  			if (!(count & RWSEM_LOCK_MASK))
  				break;
  
  			/*
  			 * The setting of the handoff bit is deferred
  			 * until rwsem_try_write_lock() is called.
  			 */
  			if ((wstate == WRITER_FIRST) && (rt_task(current) ||
  			    time_after(jiffies, waiter.timeout))) {
  				wstate = WRITER_HANDOFF;
  				lockevent_inc(rwsem_wlock_handoff);
  				break;
  			}
  		}
91d2a812d   Waiman Long   locking/rwsem: Ma...
1243
  trylock_again:
5dec94d49   Waiman Long   locking/rwsem: Me...
1244
1245
1246
  		raw_spin_lock_irq(&sem->wait_lock);
  	}
  	__set_current_state(TASK_RUNNING);
9ad8ff902   Sangmoon Kim   ANDROID: vendor_h...
1247
  	trace_android_vh_rwsem_write_wait_finish(sem);
5dec94d49   Waiman Long   locking/rwsem: Me...
1248
  	list_del(&waiter.list);
5cfd92e12   Waiman Long   locking/rwsem: Ad...
1249
  	rwsem_disable_reader_optspin(sem, disable_rspin);
5dec94d49   Waiman Long   locking/rwsem: Me...
1250
1251
1252
1253
1254
1255
1256
  	raw_spin_unlock_irq(&sem->wait_lock);
  	lockevent_inc(rwsem_wlock);
  
  	return ret;
  
  out_nolock:
  	__set_current_state(TASK_RUNNING);
9ad8ff902   Sangmoon Kim   ANDROID: vendor_h...
1257
  	trace_android_vh_rwsem_write_wait_finish(sem);
5dec94d49   Waiman Long   locking/rwsem: Me...
1258
1259
  	raw_spin_lock_irq(&sem->wait_lock);
  	list_del(&waiter.list);
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1260
1261
1262
  
  	if (unlikely(wstate == WRITER_HANDOFF))
  		atomic_long_add(-RWSEM_FLAG_HANDOFF,  &sem->count);
5dec94d49   Waiman Long   locking/rwsem: Me...
1263
1264
1265
  	if (list_empty(&sem->wait_list))
  		atomic_long_andnot(RWSEM_FLAG_WAITERS, &sem->count);
  	else
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1266
  		rwsem_mark_wake(sem, RWSEM_WAKE_ANY, &wake_q);
5dec94d49   Waiman Long   locking/rwsem: Me...
1267
1268
1269
1270
1271
1272
  	raw_spin_unlock_irq(&sem->wait_lock);
  	wake_up_q(&wake_q);
  	lockevent_inc(rwsem_wlock_fail);
  
  	return ERR_PTR(-EINTR);
  }
5dec94d49   Waiman Long   locking/rwsem: Me...
1273
1274
1275
1276
  /*
   * handle waking up a waiter on the semaphore
   * - up_read/up_write has decremented the active part of count if we come here
   */
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1277
  static struct rw_semaphore *rwsem_wake(struct rw_semaphore *sem, long count)
5dec94d49   Waiman Long   locking/rwsem: Me...
1278
1279
1280
1281
1282
1283
1284
  {
  	unsigned long flags;
  	DEFINE_WAKE_Q(wake_q);
  
  	raw_spin_lock_irqsave(&sem->wait_lock, flags);
  
  	if (!list_empty(&sem->wait_list))
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1285
  		rwsem_mark_wake(sem, RWSEM_WAKE_ANY, &wake_q);
5dec94d49   Waiman Long   locking/rwsem: Me...
1286
1287
1288
1289
1290
1291
  
  	raw_spin_unlock_irqrestore(&sem->wait_lock, flags);
  	wake_up_q(&wake_q);
  
  	return sem;
  }
5dec94d49   Waiman Long   locking/rwsem: Me...
1292
1293
1294
1295
1296
1297
  
  /*
   * downgrade a write lock into a read lock
   * - caller incremented waiting part of count and discovered it still negative
   * - just wake up any readers at the front of the queue
   */
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1298
  static struct rw_semaphore *rwsem_downgrade_wake(struct rw_semaphore *sem)
5dec94d49   Waiman Long   locking/rwsem: Me...
1299
1300
1301
1302
1303
1304
1305
  {
  	unsigned long flags;
  	DEFINE_WAKE_Q(wake_q);
  
  	raw_spin_lock_irqsave(&sem->wait_lock, flags);
  
  	if (!list_empty(&sem->wait_list))
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1306
  		rwsem_mark_wake(sem, RWSEM_WAKE_READ_OWNED, &wake_q);
5dec94d49   Waiman Long   locking/rwsem: Me...
1307
1308
1309
1310
1311
1312
  
  	raw_spin_unlock_irqrestore(&sem->wait_lock, flags);
  	wake_up_q(&wake_q);
  
  	return sem;
  }
5dec94d49   Waiman Long   locking/rwsem: Me...
1313
1314
1315
1316
  
  /*
   * lock for reading
   */
7f26482a8   Peter Zijlstra   locking/percpu-rw...
1317
  static inline void __down_read(struct rw_semaphore *sem)
5dec94d49   Waiman Long   locking/rwsem: Me...
1318
  {
a15ea1a35   Waiman Long   locking/rwsem: Gu...
1319
  	if (!rwsem_read_trylock(sem)) {
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1320
  		rwsem_down_read_slowpath(sem, TASK_UNINTERRUPTIBLE);
94a9717b3   Waiman Long   locking/rwsem: Ma...
1321
  		DEBUG_RWSEMS_WARN_ON(!is_rwsem_reader_owned(sem), sem);
5dec94d49   Waiman Long   locking/rwsem: Me...
1322
1323
1324
1325
  	} else {
  		rwsem_set_reader_owned(sem);
  	}
  }
933b7cc86   Eric W. Biederman   rwsem: Implement ...
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
  static inline int __down_read_interruptible(struct rw_semaphore *sem)
  {
  	if (!rwsem_read_trylock(sem)) {
  		if (IS_ERR(rwsem_down_read_slowpath(sem, TASK_INTERRUPTIBLE)))
  			return -EINTR;
  		DEBUG_RWSEMS_WARN_ON(!is_rwsem_reader_owned(sem), sem);
  	} else {
  		rwsem_set_reader_owned(sem);
  	}
  	return 0;
  }
5dec94d49   Waiman Long   locking/rwsem: Me...
1337
1338
  static inline int __down_read_killable(struct rw_semaphore *sem)
  {
a15ea1a35   Waiman Long   locking/rwsem: Gu...
1339
  	if (!rwsem_read_trylock(sem)) {
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1340
  		if (IS_ERR(rwsem_down_read_slowpath(sem, TASK_KILLABLE)))
5dec94d49   Waiman Long   locking/rwsem: Me...
1341
  			return -EINTR;
94a9717b3   Waiman Long   locking/rwsem: Ma...
1342
  		DEBUG_RWSEMS_WARN_ON(!is_rwsem_reader_owned(sem), sem);
5dec94d49   Waiman Long   locking/rwsem: Me...
1343
1344
1345
1346
1347
1348
1349
1350
  	} else {
  		rwsem_set_reader_owned(sem);
  	}
  	return 0;
  }
  
  static inline int __down_read_trylock(struct rw_semaphore *sem)
  {
fce45cd41   Davidlohr Bueso   locking/rwsem: Ch...
1351
1352
1353
  	long tmp;
  
  	DEBUG_RWSEMS_WARN_ON(sem->magic != sem, sem);
5dec94d49   Waiman Long   locking/rwsem: Me...
1354
1355
1356
  	/*
  	 * Optimize for the case when the rwsem is not locked at all.
  	 */
fce45cd41   Davidlohr Bueso   locking/rwsem: Ch...
1357
  	tmp = RWSEM_UNLOCKED_VALUE;
5dec94d49   Waiman Long   locking/rwsem: Me...
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
  	do {
  		if (atomic_long_try_cmpxchg_acquire(&sem->count, &tmp,
  					tmp + RWSEM_READER_BIAS)) {
  			rwsem_set_reader_owned(sem);
  			return 1;
  		}
  	} while (!(tmp & RWSEM_READ_FAILED_MASK));
  	return 0;
  }
  
  /*
   * lock for writing
   */
7f26482a8   Peter Zijlstra   locking/percpu-rw...
1371
  static inline void __down_write(struct rw_semaphore *sem)
5dec94d49   Waiman Long   locking/rwsem: Me...
1372
  {
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1373
1374
1375
1376
1377
  	long tmp = RWSEM_UNLOCKED_VALUE;
  
  	if (unlikely(!atomic_long_try_cmpxchg_acquire(&sem->count, &tmp,
  						      RWSEM_WRITER_LOCKED)))
  		rwsem_down_write_slowpath(sem, TASK_UNINTERRUPTIBLE);
5cfd92e12   Waiman Long   locking/rwsem: Ad...
1378
1379
  	else
  		rwsem_set_owner(sem);
5dec94d49   Waiman Long   locking/rwsem: Me...
1380
1381
1382
1383
  }
  
  static inline int __down_write_killable(struct rw_semaphore *sem)
  {
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1384
1385
1386
1387
1388
  	long tmp = RWSEM_UNLOCKED_VALUE;
  
  	if (unlikely(!atomic_long_try_cmpxchg_acquire(&sem->count, &tmp,
  						      RWSEM_WRITER_LOCKED))) {
  		if (IS_ERR(rwsem_down_write_slowpath(sem, TASK_KILLABLE)))
5dec94d49   Waiman Long   locking/rwsem: Me...
1389
  			return -EINTR;
5cfd92e12   Waiman Long   locking/rwsem: Ad...
1390
1391
  	} else {
  		rwsem_set_owner(sem);
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1392
  	}
5dec94d49   Waiman Long   locking/rwsem: Me...
1393
1394
1395
1396
1397
  	return 0;
  }
  
  static inline int __down_write_trylock(struct rw_semaphore *sem)
  {
fce45cd41   Davidlohr Bueso   locking/rwsem: Ch...
1398
  	long tmp;
5dec94d49   Waiman Long   locking/rwsem: Me...
1399

fce45cd41   Davidlohr Bueso   locking/rwsem: Ch...
1400
1401
1402
  	DEBUG_RWSEMS_WARN_ON(sem->magic != sem, sem);
  
  	tmp  = RWSEM_UNLOCKED_VALUE;
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1403
1404
  	if (atomic_long_try_cmpxchg_acquire(&sem->count, &tmp,
  					    RWSEM_WRITER_LOCKED)) {
5dec94d49   Waiman Long   locking/rwsem: Me...
1405
1406
1407
1408
1409
1410
1411
1412
1413
  		rwsem_set_owner(sem);
  		return true;
  	}
  	return false;
  }
  
  /*
   * unlock after reading
   */
7f26482a8   Peter Zijlstra   locking/percpu-rw...
1414
  static inline void __up_read(struct rw_semaphore *sem)
5dec94d49   Waiman Long   locking/rwsem: Me...
1415
1416
  {
  	long tmp;
fce45cd41   Davidlohr Bueso   locking/rwsem: Ch...
1417
  	DEBUG_RWSEMS_WARN_ON(sem->magic != sem, sem);
94a9717b3   Waiman Long   locking/rwsem: Ma...
1418
  	DEBUG_RWSEMS_WARN_ON(!is_rwsem_reader_owned(sem), sem);
fce45cd41   Davidlohr Bueso   locking/rwsem: Ch...
1419

5dec94d49   Waiman Long   locking/rwsem: Me...
1420
1421
  	rwsem_clear_reader_owned(sem);
  	tmp = atomic_long_add_return_release(-RWSEM_READER_BIAS, &sem->count);
a15ea1a35   Waiman Long   locking/rwsem: Gu...
1422
  	DEBUG_RWSEMS_WARN_ON(tmp < 0, sem);
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1423
  	if (unlikely((tmp & (RWSEM_LOCK_MASK|RWSEM_FLAG_WAITERS)) ==
7d43f1ce9   Waiman Long   locking/rwsem: En...
1424
1425
  		      RWSEM_FLAG_WAITERS)) {
  		clear_wr_nonspinnable(sem);
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1426
  		rwsem_wake(sem, tmp);
7d43f1ce9   Waiman Long   locking/rwsem: En...
1427
  	}
5dec94d49   Waiman Long   locking/rwsem: Me...
1428
1429
1430
1431
1432
  }
  
  /*
   * unlock after writing
   */
7f26482a8   Peter Zijlstra   locking/percpu-rw...
1433
  static inline void __up_write(struct rw_semaphore *sem)
5dec94d49   Waiman Long   locking/rwsem: Me...
1434
  {
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1435
  	long tmp;
fce45cd41   Davidlohr Bueso   locking/rwsem: Ch...
1436
  	DEBUG_RWSEMS_WARN_ON(sem->magic != sem, sem);
02f1082b0   Waiman Long   locking/rwsem: Cl...
1437
1438
1439
1440
  	/*
  	 * sem->owner may differ from current if the ownership is transferred
  	 * to an anonymous writer by setting the RWSEM_NONSPINNABLE bits.
  	 */
94a9717b3   Waiman Long   locking/rwsem: Ma...
1441
1442
  	DEBUG_RWSEMS_WARN_ON((rwsem_owner(sem) != current) &&
  			    !rwsem_test_oflags(sem, RWSEM_NONSPINNABLE), sem);
fce45cd41   Davidlohr Bueso   locking/rwsem: Ch...
1443

5dec94d49   Waiman Long   locking/rwsem: Me...
1444
  	rwsem_clear_owner(sem);
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1445
1446
  	tmp = atomic_long_fetch_add_release(-RWSEM_WRITER_LOCKED, &sem->count);
  	if (unlikely(tmp & RWSEM_FLAG_WAITERS))
4f23dbc1e   Waiman Long   locking/rwsem: Im...
1447
  		rwsem_wake(sem, tmp);
5dec94d49   Waiman Long   locking/rwsem: Me...
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
  }
  
  /*
   * downgrade write lock to read lock
   */
  static inline void __downgrade_write(struct rw_semaphore *sem)
  {
  	long tmp;
  
  	/*
  	 * When downgrading from exclusive to shared ownership,
  	 * anything inside the write-locked region cannot leak
  	 * into the read side. In contrast, anything in the
  	 * read-locked region is ok to be re-ordered into the
  	 * write side. As such, rely on RELEASE semantics.
  	 */
94a9717b3   Waiman Long   locking/rwsem: Ma...
1464
  	DEBUG_RWSEMS_WARN_ON(rwsem_owner(sem) != current, sem);
5dec94d49   Waiman Long   locking/rwsem: Me...
1465
1466
1467
1468
1469
1470
  	tmp = atomic_long_fetch_add_release(
  		-RWSEM_WRITER_LOCKED+RWSEM_READER_BIAS, &sem->count);
  	rwsem_set_reader_owned(sem);
  	if (tmp & RWSEM_FLAG_WAITERS)
  		rwsem_downgrade_wake(sem);
  }
4fc828e24   Davidlohr Bueso   locking/rwsem: Su...
1471

c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1472
1473
1474
  /*
   * lock for reading
   */
c7af77b58   Livio Soares   sched: mark rwsem...
1475
  void __sched down_read(struct rw_semaphore *sem)
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1476
1477
1478
  {
  	might_sleep();
  	rwsem_acquire_read(&sem->dep_map, 0, 0, _RET_IP_);
4fe87745a   Peter Zijlstra   lockstat: hook in...
1479
  	LOCK_CONTENDED(sem, __down_read_trylock, __down_read);
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1480
  }
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1481
  EXPORT_SYMBOL(down_read);
933b7cc86   Eric W. Biederman   rwsem: Implement ...
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
  int __sched down_read_interruptible(struct rw_semaphore *sem)
  {
  	might_sleep();
  	rwsem_acquire_read(&sem->dep_map, 0, 0, _RET_IP_);
  
  	if (LOCK_CONTENDED_RETURN(sem, __down_read_trylock, __down_read_interruptible)) {
  		rwsem_release(&sem->dep_map, _RET_IP_);
  		return -EINTR;
  	}
  
  	return 0;
  }
  EXPORT_SYMBOL(down_read_interruptible);
76f8507f7   Kirill Tkhai   locking/rwsem: Ad...
1495
1496
1497
1498
1499
1500
  int __sched down_read_killable(struct rw_semaphore *sem)
  {
  	might_sleep();
  	rwsem_acquire_read(&sem->dep_map, 0, 0, _RET_IP_);
  
  	if (LOCK_CONTENDED_RETURN(sem, __down_read_trylock, __down_read_killable)) {
5facae4f3   Qian Cai   locking/lockdep: ...
1501
  		rwsem_release(&sem->dep_map, _RET_IP_);
76f8507f7   Kirill Tkhai   locking/rwsem: Ad...
1502
1503
  		return -EINTR;
  	}
76f8507f7   Kirill Tkhai   locking/rwsem: Ad...
1504
1505
  	return 0;
  }
76f8507f7   Kirill Tkhai   locking/rwsem: Ad...
1506
  EXPORT_SYMBOL(down_read_killable);
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1507
1508
1509
1510
1511
1512
  /*
   * trylock for reading -- returns 1 if successful, 0 if contention
   */
  int down_read_trylock(struct rw_semaphore *sem)
  {
  	int ret = __down_read_trylock(sem);
c7580c1e8   Waiman Long   locking/rwsem: Mo...
1513
  	if (ret == 1)
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1514
1515
1516
  		rwsem_acquire_read(&sem->dep_map, 0, 1, _RET_IP_);
  	return ret;
  }
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1517
1518
1519
1520
1521
  EXPORT_SYMBOL(down_read_trylock);
  
  /*
   * lock for writing
   */
c7af77b58   Livio Soares   sched: mark rwsem...
1522
  void __sched down_write(struct rw_semaphore *sem)
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1523
1524
1525
  {
  	might_sleep();
  	rwsem_acquire(&sem->dep_map, 0, 0, _RET_IP_);
4fe87745a   Peter Zijlstra   lockstat: hook in...
1526
  	LOCK_CONTENDED(sem, __down_write_trylock, __down_write);
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1527
  }
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1528
1529
1530
  EXPORT_SYMBOL(down_write);
  
  /*
916633a40   Michal Hocko   locking/rwsem: Pr...
1531
1532
1533
1534
1535
1536
   * lock for writing
   */
  int __sched down_write_killable(struct rw_semaphore *sem)
  {
  	might_sleep();
  	rwsem_acquire(&sem->dep_map, 0, 0, _RET_IP_);
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1537
1538
  	if (LOCK_CONTENDED_RETURN(sem, __down_write_trylock,
  				  __down_write_killable)) {
5facae4f3   Qian Cai   locking/lockdep: ...
1539
  		rwsem_release(&sem->dep_map, _RET_IP_);
916633a40   Michal Hocko   locking/rwsem: Pr...
1540
1541
  		return -EINTR;
  	}
916633a40   Michal Hocko   locking/rwsem: Pr...
1542
1543
  	return 0;
  }
916633a40   Michal Hocko   locking/rwsem: Pr...
1544
1545
1546
  EXPORT_SYMBOL(down_write_killable);
  
  /*
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1547
1548
1549
1550
1551
   * trylock for writing -- returns 1 if successful, 0 if contention
   */
  int down_write_trylock(struct rw_semaphore *sem)
  {
  	int ret = __down_write_trylock(sem);
c7580c1e8   Waiman Long   locking/rwsem: Mo...
1552
  	if (ret == 1)
428e6ce02   Pavel Emelianov   Lockdep treats do...
1553
  		rwsem_acquire(&sem->dep_map, 0, 1, _RET_IP_);
4fc828e24   Davidlohr Bueso   locking/rwsem: Su...
1554

c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1555
1556
  	return ret;
  }
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1557
1558
1559
1560
1561
1562
1563
  EXPORT_SYMBOL(down_write_trylock);
  
  /*
   * release a read lock
   */
  void up_read(struct rw_semaphore *sem)
  {
5facae4f3   Qian Cai   locking/lockdep: ...
1564
  	rwsem_release(&sem->dep_map, _RET_IP_);
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1565
1566
  	__up_read(sem);
  }
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1567
1568
1569
1570
1571
1572
1573
  EXPORT_SYMBOL(up_read);
  
  /*
   * release a write lock
   */
  void up_write(struct rw_semaphore *sem)
  {
5facae4f3   Qian Cai   locking/lockdep: ...
1574
  	rwsem_release(&sem->dep_map, _RET_IP_);
df18d99da   JianMin Liu   ANDROID: rwsem: A...
1575
  	trace_android_vh_rwsem_write_finished(sem);
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1576
1577
  	__up_write(sem);
  }
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1578
1579
1580
1581
1582
1583
1584
  EXPORT_SYMBOL(up_write);
  
  /*
   * downgrade write lock to read lock
   */
  void downgrade_write(struct rw_semaphore *sem)
  {
6419c4af7   J. R. Okajima   locking/lockdep: ...
1585
  	lock_downgrade(&sem->dep_map, _RET_IP_);
df18d99da   JianMin Liu   ANDROID: rwsem: A...
1586
  	trace_android_vh_rwsem_write_finished(sem);
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1587
1588
  	__downgrade_write(sem);
  }
c4e05116a   Ingo Molnar   [PATCH] lockdep: ...
1589
  EXPORT_SYMBOL(downgrade_write);
4ea2176df   Ingo Molnar   [PATCH] lockdep: ...
1590
1591
1592
1593
1594
1595
1596
  
  #ifdef CONFIG_DEBUG_LOCK_ALLOC
  
  void down_read_nested(struct rw_semaphore *sem, int subclass)
  {
  	might_sleep();
  	rwsem_acquire_read(&sem->dep_map, subclass, 0, _RET_IP_);
4fe87745a   Peter Zijlstra   lockstat: hook in...
1597
  	LOCK_CONTENDED(sem, __down_read_trylock, __down_read);
4ea2176df   Ingo Molnar   [PATCH] lockdep: ...
1598
  }
4ea2176df   Ingo Molnar   [PATCH] lockdep: ...
1599
  EXPORT_SYMBOL(down_read_nested);
27bae39e4   Eric W. Biederman   rwsem: Implement ...
1600
1601
1602
1603
1604
1605
1606
1607
1608
1609
1610
1611
1612
  int down_read_killable_nested(struct rw_semaphore *sem, int subclass)
  {
  	might_sleep();
  	rwsem_acquire_read(&sem->dep_map, subclass, 0, _RET_IP_);
  
  	if (LOCK_CONTENDED_RETURN(sem, __down_read_trylock, __down_read_killable)) {
  		rwsem_release(&sem->dep_map, _RET_IP_);
  		return -EINTR;
  	}
  
  	return 0;
  }
  EXPORT_SYMBOL(down_read_killable_nested);
1b963c81b   Jiri Kosina   lockdep, rwsem: p...
1613
1614
1615
1616
  void _down_write_nest_lock(struct rw_semaphore *sem, struct lockdep_map *nest)
  {
  	might_sleep();
  	rwsem_acquire_nest(&sem->dep_map, 0, 0, nest, _RET_IP_);
1b963c81b   Jiri Kosina   lockdep, rwsem: p...
1617
1618
  	LOCK_CONTENDED(sem, __down_write_trylock, __down_write);
  }
1b963c81b   Jiri Kosina   lockdep, rwsem: p...
1619
  EXPORT_SYMBOL(_down_write_nest_lock);
84759c6d1   Kent Overstreet   Revert "rw_semaph...
1620
1621
1622
  void down_read_non_owner(struct rw_semaphore *sem)
  {
  	might_sleep();
84759c6d1   Kent Overstreet   Revert "rw_semaph...
1623
  	__down_read(sem);
925b9cd1b   Waiman Long   locking/rwsem: Ma...
1624
  	__rwsem_set_reader_owned(sem, NULL);
84759c6d1   Kent Overstreet   Revert "rw_semaph...
1625
  }
84759c6d1   Kent Overstreet   Revert "rw_semaph...
1626
  EXPORT_SYMBOL(down_read_non_owner);
4ea2176df   Ingo Molnar   [PATCH] lockdep: ...
1627
1628
1629
1630
  void down_write_nested(struct rw_semaphore *sem, int subclass)
  {
  	might_sleep();
  	rwsem_acquire(&sem->dep_map, subclass, 0, _RET_IP_);
4fe87745a   Peter Zijlstra   lockstat: hook in...
1631
  	LOCK_CONTENDED(sem, __down_write_trylock, __down_write);
4ea2176df   Ingo Molnar   [PATCH] lockdep: ...
1632
  }
4ea2176df   Ingo Molnar   [PATCH] lockdep: ...
1633
  EXPORT_SYMBOL(down_write_nested);
887bddfa9   Al Viro   add down_write_ki...
1634
1635
1636
1637
  int __sched down_write_killable_nested(struct rw_semaphore *sem, int subclass)
  {
  	might_sleep();
  	rwsem_acquire(&sem->dep_map, subclass, 0, _RET_IP_);
6cef7ff6e   Waiman Long   locking/rwsem: Co...
1638
1639
  	if (LOCK_CONTENDED_RETURN(sem, __down_write_trylock,
  				  __down_write_killable)) {
5facae4f3   Qian Cai   locking/lockdep: ...
1640
  		rwsem_release(&sem->dep_map, _RET_IP_);
887bddfa9   Al Viro   add down_write_ki...
1641
1642
  		return -EINTR;
  	}
887bddfa9   Al Viro   add down_write_ki...
1643
1644
  	return 0;
  }
887bddfa9   Al Viro   add down_write_ki...
1645
  EXPORT_SYMBOL(down_write_killable_nested);
84759c6d1   Kent Overstreet   Revert "rw_semaph...
1646
1647
  void up_read_non_owner(struct rw_semaphore *sem)
  {
94a9717b3   Waiman Long   locking/rwsem: Ma...
1648
  	DEBUG_RWSEMS_WARN_ON(!is_rwsem_reader_owned(sem), sem);
84759c6d1   Kent Overstreet   Revert "rw_semaph...
1649
1650
  	__up_read(sem);
  }
84759c6d1   Kent Overstreet   Revert "rw_semaph...
1651
  EXPORT_SYMBOL(up_read_non_owner);
4ea2176df   Ingo Molnar   [PATCH] lockdep: ...
1652
  #endif