Blame view

Documentation/futex-requeue-pi.txt 5.03 KB
b30505c81   Darren Hart   futex: add requeu...
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
  Futex Requeue PI
  ----------------
  
  Requeueing of tasks from a non-PI futex to a PI futex requires
  special handling in order to ensure the underlying rt_mutex is never
  left without an owner if it has waiters; doing so would break the PI
  boosting logic [see rt-mutex-desgin.txt] For the purposes of
  brevity, this action will be referred to as "requeue_pi" throughout
  this document.  Priority inheritance is abbreviated throughout as
  "PI".
  
  Motivation
  ----------
  
  Without requeue_pi, the glibc implementation of
  pthread_cond_broadcast() must resort to waking all the tasks waiting
  on a pthread_condvar and letting them try to sort out which task
  gets to run first in classic thundering-herd formation.  An ideal
  implementation would wake the highest-priority waiter, and leave the
  rest to the natural wakeup inherent in unlocking the mutex
  associated with the condvar.
  
  Consider the simplified glibc calls:
  
  /* caller must lock mutex */
  pthread_cond_wait(cond, mutex)
  {
  	lock(cond->__data.__lock);
  	unlock(mutex);
  	do {
  	   unlock(cond->__data.__lock);
  	   futex_wait(cond->__data.__futex);
  	   lock(cond->__data.__lock);
  	} while(...)
  	unlock(cond->__data.__lock);
  	lock(mutex);
  }
  
  pthread_cond_broadcast(cond)
  {
  	lock(cond->__data.__lock);
  	unlock(cond->__data.__lock);
  	futex_requeue(cond->data.__futex, cond->mutex);
  }
  
  Once pthread_cond_broadcast() requeues the tasks, the cond->mutex
  has waiters. Note that pthread_cond_wait() attempts to lock the
  mutex only after it has returned to user space.  This will leave the
  underlying rt_mutex with waiters, and no owner, breaking the
  previously mentioned PI-boosting algorithms.
  
  In order to support PI-aware pthread_condvar's, the kernel needs to
  be able to requeue tasks to PI futexes.  This support implies that
  upon a successful futex_wait system call, the caller would return to
  user space already holding the PI futex.  The glibc implementation
  would be modified as follows:
  
  
  /* caller must lock mutex */
  pthread_cond_wait_pi(cond, mutex)
  {
  	lock(cond->__data.__lock);
  	unlock(mutex);
  	do {
  	   unlock(cond->__data.__lock);
  	   futex_wait_requeue_pi(cond->__data.__futex);
  	   lock(cond->__data.__lock);
  	} while(...)
  	unlock(cond->__data.__lock);
          /* the kernel acquired the the mutex for us */
  }
  
  pthread_cond_broadcast_pi(cond)
  {
  	lock(cond->__data.__lock);
  	unlock(cond->__data.__lock);
  	futex_requeue_pi(cond->data.__futex, cond->mutex);
  }
  
  The actual glibc implementation will likely test for PI and make the
  necessary changes inside the existing calls rather than creating new
  calls for the PI cases.  Similar changes are needed for
  pthread_cond_timedwait() and pthread_cond_signal().
  
  Implementation
  --------------
  
  In order to ensure the rt_mutex has an owner if it has waiters, it
  is necessary for both the requeue code, as well as the waiting code,
  to be able to acquire the rt_mutex before returning to user space.
  The requeue code cannot simply wake the waiter and leave it to
  acquire the rt_mutex as it would open a race window between the
  requeue call returning to user space and the waiter waking and
  starting to run.  This is especially true in the uncontended case.
  
  The solution involves two new rt_mutex helper routines,
  rt_mutex_start_proxy_lock() and rt_mutex_finish_proxy_lock(), which
  allow the requeue code to acquire an uncontended rt_mutex on behalf
  of the waiter and to enqueue the waiter on a contended rt_mutex.
  Two new system calls provide the kernel<->user interface to
  requeue_pi: FUTEX_WAIT_REQUEUE_PI and FUTEX_REQUEUE_CMP_PI.
  
  FUTEX_WAIT_REQUEUE_PI is called by the waiter (pthread_cond_wait()
  and pthread_cond_timedwait()) to block on the initial futex and wait
  to be requeued to a PI-aware futex.  The implementation is the
  result of a high-speed collision between futex_wait() and
  futex_lock_pi(), with some extra logic to check for the additional
  wake-up scenarios.
  
  FUTEX_REQUEUE_CMP_PI is called by the waker
  (pthread_cond_broadcast() and pthread_cond_signal()) to requeue and
  possibly wake the waiting tasks. Internally, this system call is
  still handled by futex_requeue (by passing requeue_pi=1).  Before
  requeueing, futex_requeue() attempts to acquire the requeue target
  PI futex on behalf of the top waiter.  If it can, this waiter is
  woken.  futex_requeue() then proceeds to requeue the remaining
  nr_wake+nr_requeue tasks to the PI futex, calling
  rt_mutex_start_proxy_lock() prior to each requeue to prepare the
  task as a waiter on the underlying rt_mutex.  It is possible that
  the lock can be acquired at this stage as well, if so, the next
  waiter is woken to finish the acquisition of the lock.
  
  FUTEX_REQUEUE_PI accepts nr_wake and nr_requeue as arguments, but
  their sum is all that really matters.  futex_requeue() will wake or
  requeue up to nr_wake + nr_requeue tasks.  It will wake only as many
  tasks as it can acquire the lock for, which in the majority of cases
  should be 0 as good programming practice dictates that the caller of
  either pthread_cond_broadcast() or pthread_cond_signal() acquire the
  mutex prior to making the call. FUTEX_REQUEUE_PI requires that
  nr_wake=1.  nr_requeue should be INT_MAX for broadcast and 0 for
  signal.