Blame view

kernel/smpboot.c 11.7 KB
457c89965   Thomas Gleixner   treewide: Add SPD...
1
  // SPDX-License-Identifier: GPL-2.0-only
38498a67a   Thomas Gleixner   smp: Add generic ...
2
3
4
  /*
   * Common SMP CPU bringup/teardown functions
   */
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
5
  #include <linux/cpu.h>
29d5e0476   Thomas Gleixner   smp: Provide gene...
6
7
  #include <linux/err.h>
  #include <linux/smp.h>
8038dad7e   Paul E. McKenney   smpboot: Add comm...
8
  #include <linux/delay.h>
38498a67a   Thomas Gleixner   smp: Add generic ...
9
  #include <linux/init.h>
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
10
11
  #include <linux/list.h>
  #include <linux/slab.h>
29d5e0476   Thomas Gleixner   smp: Provide gene...
12
  #include <linux/sched.h>
299300258   Ingo Molnar   sched/headers: Pr...
13
  #include <linux/sched/task.h>
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
14
  #include <linux/export.h>
29d5e0476   Thomas Gleixner   smp: Provide gene...
15
  #include <linux/percpu.h>
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
16
17
  #include <linux/kthread.h>
  #include <linux/smpboot.h>
38498a67a   Thomas Gleixner   smp: Add generic ...
18
19
  
  #include "smpboot.h"
3180d89b4   Paul E. McKenney   hotplug: Fix UP b...
20
  #ifdef CONFIG_SMP
29d5e0476   Thomas Gleixner   smp: Provide gene...
21
  #ifdef CONFIG_GENERIC_SMP_IDLE_THREAD
29d5e0476   Thomas Gleixner   smp: Provide gene...
22
23
24
25
26
  /*
   * For the hotplug case we keep the task structs around and reuse
   * them.
   */
  static DEFINE_PER_CPU(struct task_struct *, idle_threads);
0db0628d9   Paul Gortmaker   kernel: delete __...
27
  struct task_struct *idle_thread_get(unsigned int cpu)
29d5e0476   Thomas Gleixner   smp: Provide gene...
28
29
30
31
  {
  	struct task_struct *tsk = per_cpu(idle_threads, cpu);
  
  	if (!tsk)
3bb5d2ee3   Suresh Siddha   smp, idle: Alloca...
32
  		return ERR_PTR(-ENOMEM);
29d5e0476   Thomas Gleixner   smp: Provide gene...
33
34
35
  	init_idle(tsk, cpu);
  	return tsk;
  }
3bb5d2ee3   Suresh Siddha   smp, idle: Alloca...
36
  void __init idle_thread_set_boot_cpu(void)
29d5e0476   Thomas Gleixner   smp: Provide gene...
37
  {
3bb5d2ee3   Suresh Siddha   smp, idle: Alloca...
38
  	per_cpu(idle_threads, smp_processor_id()) = current;
29d5e0476   Thomas Gleixner   smp: Provide gene...
39
  }
4a70d2d99   Srivatsa S. Bhat   smpboot, idle: Fi...
40
41
42
43
44
45
  /**
   * idle_init - Initialize the idle thread for a cpu
   * @cpu:	The cpu for which the idle thread should be initialized
   *
   * Creates the thread if it does not exist.
   */
3bb5d2ee3   Suresh Siddha   smp, idle: Alloca...
46
  static inline void idle_init(unsigned int cpu)
29d5e0476   Thomas Gleixner   smp: Provide gene...
47
  {
3bb5d2ee3   Suresh Siddha   smp, idle: Alloca...
48
49
50
51
52
53
54
55
56
57
  	struct task_struct *tsk = per_cpu(idle_threads, cpu);
  
  	if (!tsk) {
  		tsk = fork_idle(cpu);
  		if (IS_ERR(tsk))
  			pr_err("SMP: fork_idle() failed for CPU %u
  ", cpu);
  		else
  			per_cpu(idle_threads, cpu) = tsk;
  	}
29d5e0476   Thomas Gleixner   smp: Provide gene...
58
59
60
  }
  
  /**
4a70d2d99   Srivatsa S. Bhat   smpboot, idle: Fi...
61
   * idle_threads_init - Initialize idle threads for all cpus
29d5e0476   Thomas Gleixner   smp: Provide gene...
62
   */
3bb5d2ee3   Suresh Siddha   smp, idle: Alloca...
63
  void __init idle_threads_init(void)
29d5e0476   Thomas Gleixner   smp: Provide gene...
64
  {
ee74d1322   Srivatsa S. Bhat   smpboot, idle: Op...
65
66
67
  	unsigned int cpu, boot_cpu;
  
  	boot_cpu = smp_processor_id();
29d5e0476   Thomas Gleixner   smp: Provide gene...
68

3bb5d2ee3   Suresh Siddha   smp, idle: Alloca...
69
  	for_each_possible_cpu(cpu) {
ee74d1322   Srivatsa S. Bhat   smpboot, idle: Op...
70
  		if (cpu != boot_cpu)
3bb5d2ee3   Suresh Siddha   smp, idle: Alloca...
71
  			idle_init(cpu);
29d5e0476   Thomas Gleixner   smp: Provide gene...
72
  	}
29d5e0476   Thomas Gleixner   smp: Provide gene...
73
  }
29d5e0476   Thomas Gleixner   smp: Provide gene...
74
  #endif
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
75

3180d89b4   Paul E. McKenney   hotplug: Fix UP b...
76
  #endif /* #ifdef CONFIG_SMP */
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
  static LIST_HEAD(hotplug_threads);
  static DEFINE_MUTEX(smpboot_threads_lock);
  
  struct smpboot_thread_data {
  	unsigned int			cpu;
  	unsigned int			status;
  	struct smp_hotplug_thread	*ht;
  };
  
  enum {
  	HP_THREAD_NONE = 0,
  	HP_THREAD_ACTIVE,
  	HP_THREAD_PARKED,
  };
  
  /**
   * smpboot_thread_fn - percpu hotplug thread loop function
   * @data:	thread data pointer
   *
   * Checks for thread stop and park conditions. Calls the necessary
   * setup, cleanup, park and unpark functions for the registered
   * thread.
   *
   * Returns 1 when the thread should exit, 0 otherwise.
   */
  static int smpboot_thread_fn(void *data)
  {
  	struct smpboot_thread_data *td = data;
  	struct smp_hotplug_thread *ht = td->ht;
  
  	while (1) {
  		set_current_state(TASK_INTERRUPTIBLE);
  		preempt_disable();
  		if (kthread_should_stop()) {
7d4d26966   Peter Zijlstra   sched, smp: Corre...
111
  			__set_current_state(TASK_RUNNING);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
112
  			preempt_enable();
3dd08c0c9   Frederic Weisbecker   smpboot: make cle...
113
114
  			/* cleanup must mirror setup */
  			if (ht->cleanup && td->status != HP_THREAD_NONE)
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
115
116
117
118
119
120
121
  				ht->cleanup(td->cpu, cpu_online(td->cpu));
  			kfree(td);
  			return 0;
  		}
  
  		if (kthread_should_park()) {
  			__set_current_state(TASK_RUNNING);
be6a2e4c4   Ingo Molnar   Revert "sched/cor...
122
  			preempt_enable();
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
123
124
125
126
127
128
129
130
131
  			if (ht->park && td->status == HP_THREAD_ACTIVE) {
  				BUG_ON(td->cpu != smp_processor_id());
  				ht->park(td->cpu);
  				td->status = HP_THREAD_PARKED;
  			}
  			kthread_parkme();
  			/* We might have been woken for stop */
  			continue;
  		}
dc893e19b   Arnd Bergmann   Revert parts of "...
132
  		BUG_ON(td->cpu != smp_processor_id());
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
133
134
135
136
  
  		/* Check for state change setup */
  		switch (td->status) {
  		case HP_THREAD_NONE:
7d4d26966   Peter Zijlstra   sched, smp: Corre...
137
  			__set_current_state(TASK_RUNNING);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
138
139
140
141
  			preempt_enable();
  			if (ht->setup)
  				ht->setup(td->cpu);
  			td->status = HP_THREAD_ACTIVE;
7d4d26966   Peter Zijlstra   sched, smp: Corre...
142
  			continue;
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
143
  		case HP_THREAD_PARKED:
7d4d26966   Peter Zijlstra   sched, smp: Corre...
144
  			__set_current_state(TASK_RUNNING);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
145
146
147
148
  			preempt_enable();
  			if (ht->unpark)
  				ht->unpark(td->cpu);
  			td->status = HP_THREAD_ACTIVE;
7d4d26966   Peter Zijlstra   sched, smp: Corre...
149
  			continue;
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
150
151
152
  		}
  
  		if (!ht->thread_should_run(td->cpu)) {
7d4d26966   Peter Zijlstra   sched, smp: Corre...
153
  			preempt_enable_no_resched();
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
154
155
  			schedule();
  		} else {
7d4d26966   Peter Zijlstra   sched, smp: Corre...
156
  			__set_current_state(TASK_RUNNING);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
  			preempt_enable();
  			ht->thread_fn(td->cpu);
  		}
  	}
  }
  
  static int
  __smpboot_create_thread(struct smp_hotplug_thread *ht, unsigned int cpu)
  {
  	struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu);
  	struct smpboot_thread_data *td;
  
  	if (tsk)
  		return 0;
  
  	td = kzalloc_node(sizeof(*td), GFP_KERNEL, cpu_to_node(cpu));
  	if (!td)
  		return -ENOMEM;
  	td->cpu = cpu;
  	td->ht = ht;
  
  	tsk = kthread_create_on_cpu(smpboot_thread_fn, td, cpu,
  				    ht->thread_comm);
  	if (IS_ERR(tsk)) {
  		kfree(td);
  		return PTR_ERR(tsk);
  	}
a65d40961   Petr Mladek   kthread/smpboot: ...
184
185
186
187
188
  	/*
  	 * Park the thread so that it could start right on the CPU
  	 * when it is available.
  	 */
  	kthread_park(tsk);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
189
190
  	get_task_struct(tsk);
  	*per_cpu_ptr(ht->store, cpu) = tsk;
f2530dc71   Thomas Gleixner   kthread: Prevent ...
191
192
193
194
195
196
197
198
199
200
201
202
  	if (ht->create) {
  		/*
  		 * Make sure that the task has actually scheduled out
  		 * into park position, before calling the create
  		 * callback. At least the migration thread callback
  		 * requires that the task is off the runqueue.
  		 */
  		if (!wait_task_inactive(tsk, TASK_PARKED))
  			WARN_ON(1);
  		else
  			ht->create(cpu);
  	}
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
  	return 0;
  }
  
  int smpboot_create_threads(unsigned int cpu)
  {
  	struct smp_hotplug_thread *cur;
  	int ret = 0;
  
  	mutex_lock(&smpboot_threads_lock);
  	list_for_each_entry(cur, &hotplug_threads, list) {
  		ret = __smpboot_create_thread(cur, cpu);
  		if (ret)
  			break;
  	}
  	mutex_unlock(&smpboot_threads_lock);
  	return ret;
  }
  
  static void smpboot_unpark_thread(struct smp_hotplug_thread *ht, unsigned int cpu)
  {
  	struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu);
c00166d87   Oleg Nesterov   stop_machine: Kil...
224
225
  	if (!ht->selfparking)
  		kthread_unpark(tsk);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
226
  }
931ef1633   Thomas Gleixner   cpu/hotplug: Unpa...
227
  int smpboot_unpark_threads(unsigned int cpu)
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
228
229
230
231
232
  {
  	struct smp_hotplug_thread *cur;
  
  	mutex_lock(&smpboot_threads_lock);
  	list_for_each_entry(cur, &hotplug_threads, list)
167a88677   Peter Zijlstra   smpboot: Remove c...
233
  		smpboot_unpark_thread(cur, cpu);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
234
  	mutex_unlock(&smpboot_threads_lock);
931ef1633   Thomas Gleixner   cpu/hotplug: Unpa...
235
  	return 0;
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
236
237
238
239
240
  }
  
  static void smpboot_park_thread(struct smp_hotplug_thread *ht, unsigned int cpu)
  {
  	struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu);
7d7e499f7   Thomas Gleixner   smpboot: Allow se...
241
  	if (tsk && !ht->selfparking)
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
242
243
  		kthread_park(tsk);
  }
931ef1633   Thomas Gleixner   cpu/hotplug: Unpa...
244
  int smpboot_park_threads(unsigned int cpu)
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
245
246
247
248
249
250
251
  {
  	struct smp_hotplug_thread *cur;
  
  	mutex_lock(&smpboot_threads_lock);
  	list_for_each_entry_reverse(cur, &hotplug_threads, list)
  		smpboot_park_thread(cur, cpu);
  	mutex_unlock(&smpboot_threads_lock);
931ef1633   Thomas Gleixner   cpu/hotplug: Unpa...
252
  	return 0;
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
  }
  
  static void smpboot_destroy_threads(struct smp_hotplug_thread *ht)
  {
  	unsigned int cpu;
  
  	/* We need to destroy also the parked threads of offline cpus */
  	for_each_possible_cpu(cpu) {
  		struct task_struct *tsk = *per_cpu_ptr(ht->store, cpu);
  
  		if (tsk) {
  			kthread_stop(tsk);
  			put_task_struct(tsk);
  			*per_cpu_ptr(ht->store, cpu) = NULL;
  		}
  	}
  }
  
  /**
167a88677   Peter Zijlstra   smpboot: Remove c...
272
   * smpboot_register_percpu_thread - Register a per_cpu thread related
230ec9390   Frederic Weisbecker   smpboot: allow pa...
273
   * 					    to hotplug
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
274
275
276
277
   * @plug_thread:	Hotplug thread descriptor
   *
   * Creates and starts the threads on all online cpus.
   */
167a88677   Peter Zijlstra   smpboot: Remove c...
278
  int smpboot_register_percpu_thread(struct smp_hotplug_thread *plug_thread)
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
279
280
281
  {
  	unsigned int cpu;
  	int ret = 0;
4bee96860   Lai Jiangshan   smpboot: Add miss...
282
  	get_online_cpus();
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
283
284
285
286
287
288
289
  	mutex_lock(&smpboot_threads_lock);
  	for_each_online_cpu(cpu) {
  		ret = __smpboot_create_thread(plug_thread, cpu);
  		if (ret) {
  			smpboot_destroy_threads(plug_thread);
  			goto out;
  		}
167a88677   Peter Zijlstra   smpboot: Remove c...
290
  		smpboot_unpark_thread(plug_thread, cpu);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
291
292
293
294
  	}
  	list_add(&plug_thread->list, &hotplug_threads);
  out:
  	mutex_unlock(&smpboot_threads_lock);
4bee96860   Lai Jiangshan   smpboot: Add miss...
295
  	put_online_cpus();
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
296
297
  	return ret;
  }
167a88677   Peter Zijlstra   smpboot: Remove c...
298
  EXPORT_SYMBOL_GPL(smpboot_register_percpu_thread);
f97f8f06a   Thomas Gleixner   smpboot: Provide ...
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
  
  /**
   * smpboot_unregister_percpu_thread - Unregister a per_cpu thread related to hotplug
   * @plug_thread:	Hotplug thread descriptor
   *
   * Stops all threads on all possible cpus.
   */
  void smpboot_unregister_percpu_thread(struct smp_hotplug_thread *plug_thread)
  {
  	get_online_cpus();
  	mutex_lock(&smpboot_threads_lock);
  	list_del(&plug_thread->list);
  	smpboot_destroy_threads(plug_thread);
  	mutex_unlock(&smpboot_threads_lock);
  	put_online_cpus();
  }
  EXPORT_SYMBOL_GPL(smpboot_unregister_percpu_thread);
8038dad7e   Paul E. McKenney   smpboot: Add comm...
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
  
  static DEFINE_PER_CPU(atomic_t, cpu_hotplug_state) = ATOMIC_INIT(CPU_POST_DEAD);
  
  /*
   * Called to poll specified CPU's state, for example, when waiting for
   * a CPU to come online.
   */
  int cpu_report_state(int cpu)
  {
  	return atomic_read(&per_cpu(cpu_hotplug_state, cpu));
  }
  
  /*
   * If CPU has died properly, set its state to CPU_UP_PREPARE and
   * return success.  Otherwise, return -EBUSY if the CPU died after
   * cpu_wait_death() timed out.  And yet otherwise again, return -EAGAIN
   * if cpu_wait_death() timed out and the CPU still hasn't gotten around
   * to dying.  In the latter two cases, the CPU might not be set up
   * properly, but it is up to the arch-specific code to decide.
   * Finally, -EIO indicates an unanticipated problem.
   *
   * Note that it is permissible to omit this call entirely, as is
   * done in architectures that do no CPU-hotplug error checking.
   */
  int cpu_check_up_prepare(int cpu)
  {
  	if (!IS_ENABLED(CONFIG_HOTPLUG_CPU)) {
  		atomic_set(&per_cpu(cpu_hotplug_state, cpu), CPU_UP_PREPARE);
  		return 0;
  	}
  
  	switch (atomic_read(&per_cpu(cpu_hotplug_state, cpu))) {
  
  	case CPU_POST_DEAD:
  
  		/* The CPU died properly, so just start it up again. */
  		atomic_set(&per_cpu(cpu_hotplug_state, cpu), CPU_UP_PREPARE);
  		return 0;
  
  	case CPU_DEAD_FROZEN:
  
  		/*
  		 * Timeout during CPU death, so let caller know.
  		 * The outgoing CPU completed its processing, but after
  		 * cpu_wait_death() timed out and reported the error. The
  		 * caller is free to proceed, in which case the state
  		 * will be reset properly by cpu_set_state_online().
  		 * Proceeding despite this -EBUSY return makes sense
  		 * for systems where the outgoing CPUs take themselves
  		 * offline, with no post-death manipulation required from
  		 * a surviving CPU.
  		 */
  		return -EBUSY;
  
  	case CPU_BROKEN:
  
  		/*
  		 * The most likely reason we got here is that there was
  		 * a timeout during CPU death, and the outgoing CPU never
  		 * did complete its processing.  This could happen on
  		 * a virtualized system if the outgoing VCPU gets preempted
  		 * for more than five seconds, and the user attempts to
  		 * immediately online that same CPU.  Trying again later
  		 * might return -EBUSY above, hence -EAGAIN.
  		 */
  		return -EAGAIN;
  
  	default:
  
  		/* Should not happen.  Famous last words. */
  		return -EIO;
  	}
  }
  
  /*
   * Mark the specified CPU online.
   *
   * Note that it is permissible to omit this call entirely, as is
   * done in architectures that do no CPU-hotplug error checking.
   */
  void cpu_set_state_online(int cpu)
  {
  	(void)atomic_xchg(&per_cpu(cpu_hotplug_state, cpu), CPU_ONLINE);
  }
  
  #ifdef CONFIG_HOTPLUG_CPU
  
  /*
   * Wait for the specified CPU to exit the idle loop and die.
   */
  bool cpu_wait_death(unsigned int cpu, int seconds)
  {
  	int jf_left = seconds * HZ;
  	int oldstate;
  	bool ret = true;
  	int sleep_jf = 1;
  
  	might_sleep();
  
  	/* The outgoing CPU will normally get done quite quickly. */
  	if (atomic_read(&per_cpu(cpu_hotplug_state, cpu)) == CPU_DEAD)
  		goto update_state;
  	udelay(5);
  
  	/* But if the outgoing CPU dawdles, wait increasingly long times. */
  	while (atomic_read(&per_cpu(cpu_hotplug_state, cpu)) != CPU_DEAD) {
  		schedule_timeout_uninterruptible(sleep_jf);
  		jf_left -= sleep_jf;
  		if (jf_left <= 0)
  			break;
  		sleep_jf = DIV_ROUND_UP(sleep_jf * 11, 10);
  	}
  update_state:
  	oldstate = atomic_read(&per_cpu(cpu_hotplug_state, cpu));
  	if (oldstate == CPU_DEAD) {
  		/* Outgoing CPU died normally, update state. */
  		smp_mb(); /* atomic_read() before update. */
  		atomic_set(&per_cpu(cpu_hotplug_state, cpu), CPU_POST_DEAD);
  	} else {
  		/* Outgoing CPU still hasn't died, set state accordingly. */
  		if (atomic_cmpxchg(&per_cpu(cpu_hotplug_state, cpu),
  				   oldstate, CPU_BROKEN) != oldstate)
  			goto update_state;
  		ret = false;
  	}
  	return ret;
  }
  
  /*
   * Called by the outgoing CPU to report its successful death.  Return
   * false if this report follows the surviving CPU's timing out.
   *
   * A separate "CPU_DEAD_FROZEN" is used when the surviving CPU
   * timed out.  This approach allows architectures to omit calls to
   * cpu_check_up_prepare() and cpu_set_state_online() without defeating
   * the next cpu_wait_death()'s polling loop.
   */
  bool cpu_report_death(void)
  {
  	int oldstate;
  	int newstate;
  	int cpu = smp_processor_id();
  
  	do {
  		oldstate = atomic_read(&per_cpu(cpu_hotplug_state, cpu));
  		if (oldstate != CPU_BROKEN)
  			newstate = CPU_DEAD;
  		else
  			newstate = CPU_DEAD_FROZEN;
  	} while (atomic_cmpxchg(&per_cpu(cpu_hotplug_state, cpu),
  				oldstate, newstate) != oldstate);
  	return newstate == CPU_DEAD;
  }
  
  #endif /* #ifdef CONFIG_HOTPLUG_CPU */