Blame view

kernel/sched/features.h 1.85 KB
e26af0e8b   Peter Zijlstra   sched: Add come c...
1
  /*
51e0304ce   Ingo Molnar   sched: Implement ...
2
3
4
5
   * Only give sleepers 50% of their service deficit. This allows
   * them to run sooner, but does not allow tons of sleepers to
   * rip the spread apart.
   */
f8b6d1cc7   Peter Zijlstra   sched: Use jump_l...
6
  SCHED_FEAT(GENTLE_FAIR_SLEEPERS, true)
e26af0e8b   Peter Zijlstra   sched: Add come c...
7
8
  
  /*
e26af0e8b   Peter Zijlstra   sched: Add come c...
9
10
11
   * Place new tasks ahead so that they do not starve already running
   * tasks
   */
f8b6d1cc7   Peter Zijlstra   sched: Use jump_l...
12
  SCHED_FEAT(START_DEBIT, true)
e26af0e8b   Peter Zijlstra   sched: Add come c...
13
14
  
  /*
e26af0e8b   Peter Zijlstra   sched: Add come c...
15
16
17
18
19
   * Based on load and program behaviour, see if it makes sense to place
   * a newly woken task on the same cpu as the task that woke it --
   * improve cache locality. Typically used with SYNC wakeups as
   * generated by pipes and the like, see also SYNC_WAKEUPS.
   */
f8b6d1cc7   Peter Zijlstra   sched: Use jump_l...
20
  SCHED_FEAT(AFFINE_WAKEUPS, true)
e26af0e8b   Peter Zijlstra   sched: Add come c...
21
22
23
24
25
26
  
  /*
   * Prefer to schedule the task we woke last (assuming it failed
   * wakeup-preemption), since its likely going to consume data we
   * touched, increases cache locality.
   */
f8b6d1cc7   Peter Zijlstra   sched: Use jump_l...
27
  SCHED_FEAT(NEXT_BUDDY, false)
e26af0e8b   Peter Zijlstra   sched: Add come c...
28
29
30
31
32
33
  
  /*
   * Prefer to schedule the task that ran last (when we did
   * wake-preempt) as that likely will touch the same data, increases
   * cache locality.
   */
f8b6d1cc7   Peter Zijlstra   sched: Use jump_l...
34
  SCHED_FEAT(LAST_BUDDY, true)
e26af0e8b   Peter Zijlstra   sched: Add come c...
35
36
37
38
39
  
  /*
   * Consider buddies to be cache hot, decreases the likelyness of a
   * cache buddy being migrated away, increases cache locality.
   */
f8b6d1cc7   Peter Zijlstra   sched: Use jump_l...
40
  SCHED_FEAT(CACHE_HOT_BUDDY, true)
e26af0e8b   Peter Zijlstra   sched: Add come c...
41

8e6598af3   Peter Zijlstra   sched: Feature to...
42
43
44
  /*
   * Use arch dependent cpu power functions
   */
f8b6d1cc7   Peter Zijlstra   sched: Use jump_l...
45
  SCHED_FEAT(ARCH_POWER, false)
8e6598af3   Peter Zijlstra   sched: Feature to...
46

f8b6d1cc7   Peter Zijlstra   sched: Use jump_l...
47
48
49
  SCHED_FEAT(HRTICK, false)
  SCHED_FEAT(DOUBLE_TICK, false)
  SCHED_FEAT(LB_BIAS, true)
e26af0e8b   Peter Zijlstra   sched: Add come c...
50
51
52
53
54
55
  
  /*
   * Spin-wait on mutex acquisition when the mutex owner is running on
   * another cpu -- assumes that when the owner is running, it will soon
   * release the lock. Decreases scheduling overhead.
   */
f8b6d1cc7   Peter Zijlstra   sched: Use jump_l...
56
  SCHED_FEAT(OWNER_SPIN, true)
aa4838085   Venkatesh Pallipadi   sched: Remove irq...
57
58
  
  /*
095c0aa83   Glauber Costa   sched: adjust sch...
59
   * Decrement CPU power based on time not spent running tasks
aa4838085   Venkatesh Pallipadi   sched: Remove irq...
60
   */
f8b6d1cc7   Peter Zijlstra   sched: Use jump_l...
61
  SCHED_FEAT(NONTASK_POWER, true)
317f39416   Peter Zijlstra   sched: Move the s...
62
63
64
65
66
  
  /*
   * Queue remote wakeups on the target CPU and process them
   * using the scheduler IPI. Reduces rq->lock contention/bounces.
   */
f8b6d1cc7   Peter Zijlstra   sched: Use jump_l...
67
  SCHED_FEAT(TTWU_QUEUE, true)
e3589f6c8   Peter Zijlstra   sched: Allow for ...
68

f8b6d1cc7   Peter Zijlstra   sched: Use jump_l...
69
70
  SCHED_FEAT(FORCE_SD_OVERLAP, false)
  SCHED_FEAT(RT_RUNTIME_SHARE, true)