Blame view

kernel/smpboot.c 13.3 KB
38498a67a   Thomas Gleixner   smp: Add generic ...
1
2
3
  /*
   * Common SMP CPU bringup/teardown functions
   */
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
4
  #include <linux/cpu.h>
29d5e0476   Thomas Gleixner   smp: Provide gene...
5
6
  #include <linux/err.h>
  #include <linux/smp.h>
8038dad7e   Paul E. McKenney   smpboot: Add comm...
7
  #include <linux/delay.h>
38498a67a   Thomas Gleixner   smp: Add generic ...
8
  #include <linux/init.h>
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
9
10
  #include <linux/list.h>
  #include <linux/slab.h>
29d5e0476   Thomas Gleixner   smp: Provide gene...
11
  #include <linux/sched.h>
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
12
  #include <linux/export.h>
29d5e0476   Thomas Gleixner   smp: Provide gene...
13
  #include <linux/percpu.h>
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
14
15
  #include <linux/kthread.h>
  #include <linux/smpboot.h>
38498a67a   Thomas Gleixner   smp: Add generic ...
16
17
  
  #include "smpboot.h"
3180d89b4   Paul E. McKenney   hotplug: Fix UP b...
18
  #ifdef CONFIG_SMP
29d5e0476   Thomas Gleixner   smp: Provide gene...
19
  #ifdef CONFIG_GENERIC_SMP_IDLE_THREAD
29d5e0476   Thomas Gleixner   smp: Provide gene...
20
21
22
23
24
  /*
   * For the hotplug case we keep the task structs around and reuse
   * them.
   */
  static DEFINE_PER_CPU(struct task_struct *, idle_threads);
0db0628d9   Paul Gortmaker   kernel: delete __...
25
  struct task_struct *idle_thread_get(unsigned int cpu)
29d5e0476   Thomas Gleixner   smp: Provide gene...
26
27
28
29
  {
  	struct task_struct *tsk = per_cpu(idle_threads, cpu);
  
  	if (!tsk)
3bb5d2ee3   Suresh Siddha   smp, idle: Alloca...
30
  		return ERR_PTR(-ENOMEM);
29d5e0476   Thomas Gleixner   smp: Provide gene...
31
32
33
  	init_idle(tsk, cpu);
  	return tsk;
  }
3bb5d2ee3   Suresh Siddha   smp, idle: Alloca...
34
  void __init idle_thread_set_boot_cpu(void)
29d5e0476   Thomas Gleixner   smp: Provide gene...
35
  {
3bb5d2ee3   Suresh Siddha   smp, idle: Alloca...
36
  	per_cpu(idle_threads, smp_processor_id()) = current;
29d5e0476   Thomas Gleixner   smp: Provide gene...
37
  }
4a70d2d99   Srivatsa S. Bhat   smpboot, idle: Fi...
38
39
40
41
42
43
  /**
   * idle_init - Initialize the idle thread for a cpu
   * @cpu:	The cpu for which the idle thread should be initialized
   *
   * Creates the thread if it does not exist.
   */
3bb5d2ee3   Suresh Siddha   smp, idle: Alloca...
44
  static inline void idle_init(unsigned int cpu)
29d5e0476   Thomas Gleixner   smp: Provide gene...
45
  {
3bb5d2ee3   Suresh Siddha   smp, idle: Alloca...
46
47
48
49
50
51
52
53
54
55
  	struct task_struct *tsk = per_cpu(idle_threads, cpu);
  
  	if (!tsk) {
  		tsk = fork_idle(cpu);
  		if (IS_ERR(tsk))
  			pr_err("SMP: fork_idle() failed for CPU %u
  ", cpu);
  		else
  			per_cpu(idle_threads, cpu) = tsk;
  	}
29d5e0476   Thomas Gleixner   smp: Provide gene...
56
57
58
  }
  
  /**
4a70d2d99   Srivatsa S. Bhat   smpboot, idle: Fi...
59
   * idle_threads_init - Initialize idle threads for all cpus
29d5e0476   Thomas Gleixner   smp: Provide gene...
60
   */
3bb5d2ee3   Suresh Siddha   smp, idle: Alloca...
61
  void __init idle_threads_init(void)
29d5e0476   Thomas Gleixner   smp: Provide gene...
62
  {
ee74d1322   Srivatsa S. Bhat   smpboot, idle: Op...
63
64
65
  	unsigned int cpu, boot_cpu;
  
  	boot_cpu = smp_processor_id();
29d5e0476   Thomas Gleixner   smp: Provide gene...
66

3bb5d2ee3   Suresh Siddha   smp, idle: Alloca...
67
  	for_each_possible_cpu(cpu) {
ee74d1322   Srivatsa S. Bhat   smpboot, idle: Op...
68
  		if (cpu != boot_cpu)
3bb5d2ee3   Suresh Siddha   smp, idle: Alloca...
69
  			idle_init(cpu);
29d5e0476   Thomas Gleixner   smp: Provide gene...
70
  	}
29d5e0476   Thomas Gleixner   smp: Provide gene...
71
  }
29d5e0476   Thomas Gleixner   smp: Provide gene...
72
  #endif
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
73

3180d89b4   Paul E. McKenney   hotplug: Fix UP b...
74
  #endif /* #ifdef CONFIG_SMP */
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
  static LIST_HEAD(hotplug_threads);
  static DEFINE_MUTEX(smpboot_threads_lock);
  
  struct smpboot_thread_data {
  	unsigned int			cpu;
  	unsigned int			status;
  	struct smp_hotplug_thread	*ht;
  };
  
  enum {
  	HP_THREAD_NONE = 0,
  	HP_THREAD_ACTIVE,
  	HP_THREAD_PARKED,
  };
  
  /**
   * smpboot_thread_fn - percpu hotplug thread loop function
   * @data:	thread data pointer
   *
   * Checks for thread stop and park conditions. Calls the necessary
   * setup, cleanup, park and unpark functions for the registered
   * thread.
   *
   * Returns 1 when the thread should exit, 0 otherwise.
   */
  static int smpboot_thread_fn(void *data)
  {
  	struct smpboot_thread_data *td = data;
  	struct smp_hotplug_thread *ht = td->ht;
  
  	while (1) {
  		set_current_state(TASK_INTERRUPTIBLE);
  		preempt_disable();
  		if (kthread_should_stop()) {
7d4d26966   Peter Zijlstra   sched, smp: Corre...
109
  			__set_current_state(TASK_RUNNING);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
110
  			preempt_enable();
3dd08c0c9   Frederic Weisbecker   smpboot: make cle...
111
112
  			/* cleanup must mirror setup */
  			if (ht->cleanup && td->status != HP_THREAD_NONE)
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
113
114
115
116
117
118
119
  				ht->cleanup(td->cpu, cpu_online(td->cpu));
  			kfree(td);
  			return 0;
  		}
  
  		if (kthread_should_park()) {
  			__set_current_state(TASK_RUNNING);
be6a2e4c4   Ingo Molnar   Revert "sched/cor...
120
  			preempt_enable();
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
121
122
123
124
125
126
127
128
129
  			if (ht->park && td->status == HP_THREAD_ACTIVE) {
  				BUG_ON(td->cpu != smp_processor_id());
  				ht->park(td->cpu);
  				td->status = HP_THREAD_PARKED;
  			}
  			kthread_parkme();
  			/* We might have been woken for stop */
  			continue;
  		}
dc893e19b   Arnd Bergmann   Revert parts of "...
130
  		BUG_ON(td->cpu != smp_processor_id());
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
131
132
133
134
  
  		/* Check for state change setup */
  		switch (td->status) {
  		case HP_THREAD_NONE:
7d4d26966   Peter Zijlstra   sched, smp: Corre...
135
  			__set_current_state(TASK_RUNNING);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
136
137
138
139
  			preempt_enable();
  			if (ht->setup)
  				ht->setup(td->cpu);
  			td->status = HP_THREAD_ACTIVE;
7d4d26966   Peter Zijlstra   sched, smp: Corre...
140
  			continue;
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
141
  		case HP_THREAD_PARKED:
7d4d26966   Peter Zijlstra   sched, smp: Corre...
142
  			__set_current_state(TASK_RUNNING);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
143
144
145
146
  			preempt_enable();
  			if (ht->unpark)
  				ht->unpark(td->cpu);
  			td->status = HP_THREAD_ACTIVE;
7d4d26966   Peter Zijlstra   sched, smp: Corre...
147
  			continue;
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
148
149
150
  		}
  
  		if (!ht->thread_should_run(td->cpu)) {
7d4d26966   Peter Zijlstra   sched, smp: Corre...
151
  			preempt_enable_no_resched();
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
152
153
  			schedule();
  		} else {
7d4d26966   Peter Zijlstra   sched, smp: Corre...
154
  			__set_current_state(TASK_RUNNING);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
  			preempt_enable();
  			ht->thread_fn(td->cpu);
  		}
  	}
  }
  
  static int
  __smpboot_create_thread(struct smp_hotplug_thread *ht, unsigned int cpu)
  {
  	struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu);
  	struct smpboot_thread_data *td;
  
  	if (tsk)
  		return 0;
  
  	td = kzalloc_node(sizeof(*td), GFP_KERNEL, cpu_to_node(cpu));
  	if (!td)
  		return -ENOMEM;
  	td->cpu = cpu;
  	td->ht = ht;
  
  	tsk = kthread_create_on_cpu(smpboot_thread_fn, td, cpu,
  				    ht->thread_comm);
  	if (IS_ERR(tsk)) {
  		kfree(td);
  		return PTR_ERR(tsk);
  	}
a65d40961   Petr Mladek   kthread/smpboot: ...
182
183
184
185
186
  	/*
  	 * Park the thread so that it could start right on the CPU
  	 * when it is available.
  	 */
  	kthread_park(tsk);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
187
188
  	get_task_struct(tsk);
  	*per_cpu_ptr(ht->store, cpu) = tsk;
f2530dc71   Thomas Gleixner   kthread: Prevent ...
189
190
191
192
193
194
195
196
197
198
199
200
  	if (ht->create) {
  		/*
  		 * Make sure that the task has actually scheduled out
  		 * into park position, before calling the create
  		 * callback. At least the migration thread callback
  		 * requires that the task is off the runqueue.
  		 */
  		if (!wait_task_inactive(tsk, TASK_PARKED))
  			WARN_ON(1);
  		else
  			ht->create(cpu);
  	}
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
  	return 0;
  }
  
  int smpboot_create_threads(unsigned int cpu)
  {
  	struct smp_hotplug_thread *cur;
  	int ret = 0;
  
  	mutex_lock(&smpboot_threads_lock);
  	list_for_each_entry(cur, &hotplug_threads, list) {
  		ret = __smpboot_create_thread(cur, cpu);
  		if (ret)
  			break;
  	}
  	mutex_unlock(&smpboot_threads_lock);
  	return ret;
  }
  
  static void smpboot_unpark_thread(struct smp_hotplug_thread *ht, unsigned int cpu)
  {
  	struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu);
c00166d87   Oleg Nesterov   stop_machine: Kil...
222
223
  	if (!ht->selfparking)
  		kthread_unpark(tsk);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
224
  }
931ef1633   Thomas Gleixner   cpu/hotplug: Unpa...
225
  int smpboot_unpark_threads(unsigned int cpu)
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
226
227
228
229
230
  {
  	struct smp_hotplug_thread *cur;
  
  	mutex_lock(&smpboot_threads_lock);
  	list_for_each_entry(cur, &hotplug_threads, list)
b5242e98c   Chris Metcalf   smpboot: allow ex...
231
232
  		if (cpumask_test_cpu(cpu, cur->cpumask))
  			smpboot_unpark_thread(cur, cpu);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
233
  	mutex_unlock(&smpboot_threads_lock);
931ef1633   Thomas Gleixner   cpu/hotplug: Unpa...
234
  	return 0;
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
235
236
237
238
239
  }
  
  static void smpboot_park_thread(struct smp_hotplug_thread *ht, unsigned int cpu)
  {
  	struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu);
7d7e499f7   Thomas Gleixner   smpboot: Allow se...
240
  	if (tsk && !ht->selfparking)
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
241
242
  		kthread_park(tsk);
  }
931ef1633   Thomas Gleixner   cpu/hotplug: Unpa...
243
  int smpboot_park_threads(unsigned int cpu)
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
244
245
246
247
248
249
250
  {
  	struct smp_hotplug_thread *cur;
  
  	mutex_lock(&smpboot_threads_lock);
  	list_for_each_entry_reverse(cur, &hotplug_threads, list)
  		smpboot_park_thread(cur, cpu);
  	mutex_unlock(&smpboot_threads_lock);
931ef1633   Thomas Gleixner   cpu/hotplug: Unpa...
251
  	return 0;
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
  }
  
  static void smpboot_destroy_threads(struct smp_hotplug_thread *ht)
  {
  	unsigned int cpu;
  
  	/* We need to destroy also the parked threads of offline cpus */
  	for_each_possible_cpu(cpu) {
  		struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu);
  
  		if (tsk) {
  			kthread_stop(tsk);
  			put_task_struct(tsk);
  			*per_cpu_ptr(ht->store, cpu) = NULL;
  		}
  	}
  }
  
  /**
230ec9390   Frederic Weisbecker   smpboot: allow pa...
271
272
   * smpboot_register_percpu_thread_cpumask - Register a per_cpu thread related
   * 					    to hotplug
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
273
   * @plug_thread:	Hotplug thread descriptor
230ec9390   Frederic Weisbecker   smpboot: allow pa...
274
   * @cpumask:		The cpumask where threads run
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
275
276
277
   *
   * Creates and starts the threads on all online cpus.
   */
230ec9390   Frederic Weisbecker   smpboot: allow pa...
278
279
  int smpboot_register_percpu_thread_cpumask(struct smp_hotplug_thread *plug_thread,
  					   const struct cpumask *cpumask)
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
280
281
282
  {
  	unsigned int cpu;
  	int ret = 0;
b5242e98c   Chris Metcalf   smpboot: allow ex...
283
284
  	if (!alloc_cpumask_var(&plug_thread->cpumask, GFP_KERNEL))
  		return -ENOMEM;
230ec9390   Frederic Weisbecker   smpboot: allow pa...
285
  	cpumask_copy(plug_thread->cpumask, cpumask);
b5242e98c   Chris Metcalf   smpboot: allow ex...
286

4bee96860   Lai Jiangshan   smpboot: Add miss...
287
  	get_online_cpus();
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
288
289
290
291
292
  	mutex_lock(&smpboot_threads_lock);
  	for_each_online_cpu(cpu) {
  		ret = __smpboot_create_thread(plug_thread, cpu);
  		if (ret) {
  			smpboot_destroy_threads(plug_thread);
5869b5064   Frederic Weisbecker   smpboot: fix memo...
293
  			free_cpumask_var(plug_thread->cpumask);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
294
295
  			goto out;
  		}
230ec9390   Frederic Weisbecker   smpboot: allow pa...
296
297
  		if (cpumask_test_cpu(cpu, cpumask))
  			smpboot_unpark_thread(plug_thread, cpu);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
298
299
300
301
  	}
  	list_add(&plug_thread->list, &hotplug_threads);
  out:
  	mutex_unlock(&smpboot_threads_lock);
4bee96860   Lai Jiangshan   smpboot: Add miss...
302
  	put_online_cpus();
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
303
304
  	return ret;
  }
230ec9390   Frederic Weisbecker   smpboot: allow pa...
305
  EXPORT_SYMBOL_GPL(smpboot_register_percpu_thread_cpumask);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
  
  /**
   * smpboot_unregister_percpu_thread - Unregister a per_cpu thread related to hotplug
   * @plug_thread:	Hotplug thread descriptor
   *
   * Stops all threads on all possible cpus.
   */
  void smpboot_unregister_percpu_thread(struct smp_hotplug_thread *plug_thread)
  {
  	get_online_cpus();
  	mutex_lock(&smpboot_threads_lock);
  	list_del(&plug_thread->list);
  	smpboot_destroy_threads(plug_thread);
  	mutex_unlock(&smpboot_threads_lock);
  	put_online_cpus();
b5242e98c   Chris Metcalf   smpboot: allow ex...
321
  	free_cpumask_var(plug_thread->cpumask);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
322
323
  }
  EXPORT_SYMBOL_GPL(smpboot_unregister_percpu_thread);
8038dad7e   Paul E. McKenney   smpboot: Add comm...
324

b5242e98c   Chris Metcalf   smpboot: allow ex...
325
326
327
328
329
330
331
  /**
   * smpboot_update_cpumask_percpu_thread - Adjust which per_cpu hotplug threads stay parked
   * @plug_thread:	Hotplug thread descriptor
   * @new:		Revised mask to use
   *
   * The cpumask field in the smp_hotplug_thread must not be updated directly
   * by the client, but only by calling this function.
fe4ba3c34   Chris Metcalf   watchdog: add wat...
332
   * This function can only be called on a registered smp_hotplug_thread.
b5242e98c   Chris Metcalf   smpboot: allow ex...
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
   */
  int smpboot_update_cpumask_percpu_thread(struct smp_hotplug_thread *plug_thread,
  					 const struct cpumask *new)
  {
  	struct cpumask *old = plug_thread->cpumask;
  	cpumask_var_t tmp;
  	unsigned int cpu;
  
  	if (!alloc_cpumask_var(&tmp, GFP_KERNEL))
  		return -ENOMEM;
  
  	get_online_cpus();
  	mutex_lock(&smpboot_threads_lock);
  
  	/* Park threads that were exclusively enabled on the old mask. */
  	cpumask_andnot(tmp, old, new);
  	for_each_cpu_and(cpu, tmp, cpu_online_mask)
  		smpboot_park_thread(plug_thread, cpu);
  
  	/* Unpark threads that are exclusively enabled on the new mask. */
  	cpumask_andnot(tmp, new, old);
  	for_each_cpu_and(cpu, tmp, cpu_online_mask)
  		smpboot_unpark_thread(plug_thread, cpu);
  
  	cpumask_copy(old, new);
  
  	mutex_unlock(&smpboot_threads_lock);
  	put_online_cpus();
  
  	free_cpumask_var(tmp);
  
  	return 0;
  }
  EXPORT_SYMBOL_GPL(smpboot_update_cpumask_percpu_thread);
8038dad7e   Paul E. McKenney   smpboot: Add comm...
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
  static DEFINE_PER_CPU(atomic_t, cpu_hotplug_state) = ATOMIC_INIT(CPU_POST_DEAD);
  
  /*
   * Called to poll specified CPU's state, for example, when waiting for
   * a CPU to come online.
   */
  int cpu_report_state(int cpu)
  {
  	return atomic_read(&per_cpu(cpu_hotplug_state, cpu));
  }
  
  /*
   * If CPU has died properly, set its state to CPU_UP_PREPARE and
   * return success.  Otherwise, return -EBUSY if the CPU died after
   * cpu_wait_death() timed out.  And yet otherwise again, return -EAGAIN
   * if cpu_wait_death() timed out and the CPU still hasn't gotten around
   * to dying.  In the latter two cases, the CPU might not be set up
   * properly, but it is up to the arch-specific code to decide.
   * Finally, -EIO indicates an unanticipated problem.
   *
   * Note that it is permissible to omit this call entirely, as is
   * done in architectures that do no CPU-hotplug error checking.
   */
  int cpu_check_up_prepare(int cpu)
  {
  	if (!IS_ENABLED(CONFIG_HOTPLUG_CPU)) {
  		atomic_set(&per_cpu(cpu_hotplug_state, cpu), CPU_UP_PREPARE);
  		return 0;
  	}
  
  	switch (atomic_read(&per_cpu(cpu_hotplug_state, cpu))) {
  
  	case CPU_POST_DEAD:
  
  		/* The CPU died properly, so just start it up again. */
  		atomic_set(&per_cpu(cpu_hotplug_state, cpu), CPU_UP_PREPARE);
  		return 0;
  
  	case CPU_DEAD_FROZEN:
  
  		/*
  		 * Timeout during CPU death, so let caller know.
  		 * The outgoing CPU completed its processing, but after
  		 * cpu_wait_death() timed out and reported the error. The
  		 * caller is free to proceed, in which case the state
  		 * will be reset properly by cpu_set_state_online().
  		 * Proceeding despite this -EBUSY return makes sense
  		 * for systems where the outgoing CPUs take themselves
  		 * offline, with no post-death manipulation required from
  		 * a surviving CPU.
  		 */
  		return -EBUSY;
  
  	case CPU_BROKEN:
  
  		/*
  		 * The most likely reason we got here is that there was
  		 * a timeout during CPU death, and the outgoing CPU never
  		 * did complete its processing.  This could happen on
  		 * a virtualized system if the outgoing VCPU gets preempted
  		 * for more than five seconds, and the user attempts to
  		 * immediately online that same CPU.  Trying again later
  		 * might return -EBUSY above, hence -EAGAIN.
  		 */
  		return -EAGAIN;
  
  	default:
  
  		/* Should not happen.  Famous last words. */
  		return -EIO;
  	}
  }
  
  /*
   * Mark the specified CPU online.
   *
   * Note that it is permissible to omit this call entirely, as is
   * done in architectures that do no CPU-hotplug error checking.
   */
  void cpu_set_state_online(int cpu)
  {
  	(void)atomic_xchg(&per_cpu(cpu_hotplug_state, cpu), CPU_ONLINE);
  }
  
  #ifdef CONFIG_HOTPLUG_CPU
  
  /*
   * Wait for the specified CPU to exit the idle loop and die.
   */
  bool cpu_wait_death(unsigned int cpu, int seconds)
  {
  	int jf_left = seconds * HZ;
  	int oldstate;
  	bool ret = true;
  	int sleep_jf = 1;
  
  	might_sleep();
  
  	/* The outgoing CPU will normally get done quite quickly. */
  	if (atomic_read(&per_cpu(cpu_hotplug_state, cpu)) == CPU_DEAD)
  		goto update_state;
  	udelay(5);
  
  	/* But if the outgoing CPU dawdles, wait increasingly long times. */
  	while (atomic_read(&per_cpu(cpu_hotplug_state, cpu)) != CPU_DEAD) {
  		schedule_timeout_uninterruptible(sleep_jf);
  		jf_left -= sleep_jf;
  		if (jf_left <= 0)
  			break;
  		sleep_jf = DIV_ROUND_UP(sleep_jf * 11, 10);
  	}
  update_state:
  	oldstate = atomic_read(&per_cpu(cpu_hotplug_state, cpu));
  	if (oldstate == CPU_DEAD) {
  		/* Outgoing CPU died normally, update state. */
  		smp_mb(); /* atomic_read() before update. */
  		atomic_set(&per_cpu(cpu_hotplug_state, cpu), CPU_POST_DEAD);
  	} else {
  		/* Outgoing CPU still hasn't died, set state accordingly. */
  		if (atomic_cmpxchg(&per_cpu(cpu_hotplug_state, cpu),
  				   oldstate, CPU_BROKEN) != oldstate)
  			goto update_state;
  		ret = false;
  	}
  	return ret;
  }
  
  /*
   * Called by the outgoing CPU to report its successful death.  Return
   * false if this report follows the surviving CPU's timing out.
   *
   * A separate "CPU_DEAD_FROZEN" is used when the surviving CPU
   * timed out.  This approach allows architectures to omit calls to
   * cpu_check_up_prepare() and cpu_set_state_online() without defeating
   * the next cpu_wait_death()'s polling loop.
   */
  bool cpu_report_death(void)
  {
  	int oldstate;
  	int newstate;
  	int cpu = smp_processor_id();
  
  	do {
  		oldstate = atomic_read(&per_cpu(cpu_hotplug_state, cpu));
  		if (oldstate != CPU_BROKEN)
  			newstate = CPU_DEAD;
  		else
  			newstate = CPU_DEAD_FROZEN;
  	} while (atomic_cmpxchg(&per_cpu(cpu_hotplug_state, cpu),
  				oldstate, newstate) != oldstate);
  	return newstate == CPU_DEAD;
  }
  
  #endif /* #ifdef CONFIG_HOTPLUG_CPU */