Blame view

lib/cpu_rmap.c 7.58 KB
d2912cb15   Thomas Gleixner   treewide: Replace...
1
  // SPDX-License-Identifier: GPL-2.0-only
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
2
3
4
  /*
   * cpu_rmap.c: CPU affinity reverse-map support
   * Copyright 2011 Solarflare Communications Inc.
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
5
6
7
   */
  
  #include <linux/cpu_rmap.h>
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
8
  #include <linux/interrupt.h>
8bc3bcc93   Paul Gortmaker   lib: reduce the u...
9
  #include <linux/export.h>
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
  
  /*
   * These functions maintain a mapping from CPUs to some ordered set of
   * objects with CPU affinities.  This can be seen as a reverse-map of
   * CPU affinity.  However, we do not assume that the object affinities
   * cover all CPUs in the system.  For those CPUs not directly covered
   * by object affinities, we attempt to find a nearest object based on
   * CPU topology.
   */
  
  /**
   * alloc_cpu_rmap - allocate CPU affinity reverse-map
   * @size: Number of objects to be mapped
   * @flags: Allocation flags e.g. %GFP_KERNEL
   */
  struct cpu_rmap *alloc_cpu_rmap(unsigned int size, gfp_t flags)
  {
  	struct cpu_rmap *rmap;
  	unsigned int cpu;
  	size_t obj_offset;
  
  	/* This is a silly number of objects, and we use u16 indices. */
  	if (size > 0xffff)
  		return NULL;
  
  	/* Offset of object pointer array from base structure */
  	obj_offset = ALIGN(offsetof(struct cpu_rmap, near[nr_cpu_ids]),
  			   sizeof(void *));
  
  	rmap = kzalloc(obj_offset + size * sizeof(rmap->obj[0]), flags);
  	if (!rmap)
  		return NULL;
896f97ea9   David Decotigny   lib: cpu_rmap: av...
42
  	kref_init(&rmap->refcount);
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
  	rmap->obj = (void **)((char *)rmap + obj_offset);
  
  	/* Initially assign CPUs to objects on a rota, since we have
  	 * no idea where the objects are.  Use infinite distance, so
  	 * any object with known distance is preferable.  Include the
  	 * CPUs that are not present/online, since we definitely want
  	 * any newly-hotplugged CPUs to have some object assigned.
  	 */
  	for_each_possible_cpu(cpu) {
  		rmap->near[cpu].index = cpu % size;
  		rmap->near[cpu].dist = CPU_RMAP_DIST_INF;
  	}
  
  	rmap->size = size;
  	return rmap;
  }
  EXPORT_SYMBOL(alloc_cpu_rmap);
896f97ea9   David Decotigny   lib: cpu_rmap: av...
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
  /**
   * cpu_rmap_release - internal reclaiming helper called from kref_put
   * @ref: kref to struct cpu_rmap
   */
  static void cpu_rmap_release(struct kref *ref)
  {
  	struct cpu_rmap *rmap = container_of(ref, struct cpu_rmap, refcount);
  	kfree(rmap);
  }
  
  /**
   * cpu_rmap_get - internal helper to get new ref on a cpu_rmap
   * @rmap: reverse-map allocated with alloc_cpu_rmap()
   */
  static inline void cpu_rmap_get(struct cpu_rmap *rmap)
  {
  	kref_get(&rmap->refcount);
  }
  
  /**
   * cpu_rmap_put - release ref on a cpu_rmap
   * @rmap: reverse-map allocated with alloc_cpu_rmap()
   */
  int cpu_rmap_put(struct cpu_rmap *rmap)
  {
  	return kref_put(&rmap->refcount, cpu_rmap_release);
  }
  EXPORT_SYMBOL(cpu_rmap_put);
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
  /* Reevaluate nearest object for given CPU, comparing with the given
   * neighbours at the given distance.
   */
  static bool cpu_rmap_copy_neigh(struct cpu_rmap *rmap, unsigned int cpu,
  				const struct cpumask *mask, u16 dist)
  {
  	int neigh;
  
  	for_each_cpu(neigh, mask) {
  		if (rmap->near[cpu].dist > dist &&
  		    rmap->near[neigh].dist <= dist) {
  			rmap->near[cpu].index = rmap->near[neigh].index;
  			rmap->near[cpu].dist = dist;
  			return true;
  		}
  	}
  	return false;
  }
  
  #ifdef DEBUG
  static void debug_print_rmap(const struct cpu_rmap *rmap, const char *prefix)
  {
  	unsigned index;
  	unsigned int cpu;
  
  	pr_info("cpu_rmap %p, %s:
  ", rmap, prefix);
  
  	for_each_possible_cpu(cpu) {
  		index = rmap->near[cpu].index;
  		pr_info("cpu %d -> obj %u (distance %u)
  ",
  			cpu, index, rmap->near[cpu].dist);
  	}
  }
  #else
  static inline void
  debug_print_rmap(const struct cpu_rmap *rmap, const char *prefix)
  {
  }
  #endif
  
  /**
   * cpu_rmap_add - add object to a rmap
   * @rmap: CPU rmap allocated with alloc_cpu_rmap()
   * @obj: Object to add to rmap
   *
   * Return index of object.
   */
  int cpu_rmap_add(struct cpu_rmap *rmap, void *obj)
  {
  	u16 index;
  
  	BUG_ON(rmap->used >= rmap->size);
  	index = rmap->used++;
  	rmap->obj[index] = obj;
  	return index;
  }
  EXPORT_SYMBOL(cpu_rmap_add);
  
  /**
   * cpu_rmap_update - update CPU rmap following a change of object affinity
   * @rmap: CPU rmap to update
   * @index: Index of object whose affinity changed
   * @affinity: New CPU affinity of object
   */
  int cpu_rmap_update(struct cpu_rmap *rmap, u16 index,
  		    const struct cpumask *affinity)
  {
  	cpumask_var_t update_mask;
  	unsigned int cpu;
  
  	if (unlikely(!zalloc_cpumask_var(&update_mask, GFP_KERNEL)))
  		return -ENOMEM;
  
  	/* Invalidate distance for all CPUs for which this used to be
  	 * the nearest object.  Mark those CPUs for update.
  	 */
  	for_each_online_cpu(cpu) {
  		if (rmap->near[cpu].index == index) {
  			rmap->near[cpu].dist = CPU_RMAP_DIST_INF;
  			cpumask_set_cpu(cpu, update_mask);
  		}
  	}
  
  	debug_print_rmap(rmap, "after invalidating old distances");
  
  	/* Set distance to 0 for all CPUs in the new affinity mask.
  	 * Mark all CPUs within their NUMA nodes for update.
  	 */
  	for_each_cpu(cpu, affinity) {
  		rmap->near[cpu].index = index;
  		rmap->near[cpu].dist = 0;
  		cpumask_or(update_mask, update_mask,
  			   cpumask_of_node(cpu_to_node(cpu)));
  	}
  
  	debug_print_rmap(rmap, "after updating neighbours");
  
  	/* Update distances based on topology */
  	for_each_cpu(cpu, update_mask) {
  		if (cpu_rmap_copy_neigh(rmap, cpu,
06931e622   Bartosz Golaszewski   sched/topology: R...
190
  					topology_sibling_cpumask(cpu), 1))
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
  			continue;
  		if (cpu_rmap_copy_neigh(rmap, cpu,
  					topology_core_cpumask(cpu), 2))
  			continue;
  		if (cpu_rmap_copy_neigh(rmap, cpu,
  					cpumask_of_node(cpu_to_node(cpu)), 3))
  			continue;
  		/* We could continue into NUMA node distances, but for now
  		 * we give up.
  		 */
  	}
  
  	debug_print_rmap(rmap, "after copying neighbours");
  
  	free_cpumask_var(update_mask);
  	return 0;
  }
  EXPORT_SYMBOL(cpu_rmap_update);
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
209
210
211
212
213
214
215
216
217
218
219
220
  /* Glue between IRQ affinity notifiers and CPU rmaps */
  
  struct irq_glue {
  	struct irq_affinity_notify notify;
  	struct cpu_rmap *rmap;
  	u16 index;
  };
  
  /**
   * free_irq_cpu_rmap - free a CPU affinity reverse-map used for IRQs
   * @rmap: Reverse-map allocated with alloc_irq_cpu_map(), or %NULL
   *
896f97ea9   David Decotigny   lib: cpu_rmap: av...
221
   * Must be called in process context, before freeing the IRQs.
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
222
223
224
225
226
227
228
229
230
231
232
233
234
   */
  void free_irq_cpu_rmap(struct cpu_rmap *rmap)
  {
  	struct irq_glue *glue;
  	u16 index;
  
  	if (!rmap)
  		return;
  
  	for (index = 0; index < rmap->used; index++) {
  		glue = rmap->obj[index];
  		irq_set_affinity_notifier(glue->notify.irq, NULL);
  	}
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
235

896f97ea9   David Decotigny   lib: cpu_rmap: av...
236
  	cpu_rmap_put(rmap);
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
237
238
  }
  EXPORT_SYMBOL(free_irq_cpu_rmap);
896f97ea9   David Decotigny   lib: cpu_rmap: av...
239
240
241
242
243
244
245
  /**
   * irq_cpu_rmap_notify - callback for IRQ subsystem when IRQ affinity updated
   * @notify: struct irq_affinity_notify passed by irq/manage.c
   * @mask: cpu mask for new SMP affinity
   *
   * This is executed in workqueue context.
   */
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
246
247
248
249
250
251
252
253
254
  static void
  irq_cpu_rmap_notify(struct irq_affinity_notify *notify, const cpumask_t *mask)
  {
  	struct irq_glue *glue =
  		container_of(notify, struct irq_glue, notify);
  	int rc;
  
  	rc = cpu_rmap_update(glue->rmap, glue->index, mask);
  	if (rc)
256339d60   Kefeng Wang   lib: cpu_rmap: Us...
255
256
  		pr_warn("irq_cpu_rmap_notify: update failed: %d
  ", rc);
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
257
  }
896f97ea9   David Decotigny   lib: cpu_rmap: av...
258
259
260
261
  /**
   * irq_cpu_rmap_release - reclaiming callback for IRQ subsystem
   * @ref: kref to struct irq_affinity_notify passed by irq/manage.c
   */
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
262
263
264
265
  static void irq_cpu_rmap_release(struct kref *ref)
  {
  	struct irq_glue *glue =
  		container_of(ref, struct irq_glue, notify.kref);
896f97ea9   David Decotigny   lib: cpu_rmap: av...
266
267
  
  	cpu_rmap_put(glue->rmap);
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
  	kfree(glue);
  }
  
  /**
   * irq_cpu_rmap_add - add an IRQ to a CPU affinity reverse-map
   * @rmap: The reverse-map
   * @irq: The IRQ number
   *
   * This adds an IRQ affinity notifier that will update the reverse-map
   * automatically.
   *
   * Must be called in process context, after the IRQ is allocated but
   * before it is bound with request_irq().
   */
  int irq_cpu_rmap_add(struct cpu_rmap *rmap, int irq)
  {
  	struct irq_glue *glue = kzalloc(sizeof(*glue), GFP_KERNEL);
  	int rc;
  
  	if (!glue)
  		return -ENOMEM;
  	glue->notify.notify = irq_cpu_rmap_notify;
  	glue->notify.release = irq_cpu_rmap_release;
  	glue->rmap = rmap;
896f97ea9   David Decotigny   lib: cpu_rmap: av...
292
  	cpu_rmap_get(rmap);
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
293
294
  	glue->index = cpu_rmap_add(rmap, glue);
  	rc = irq_set_affinity_notifier(irq, &glue->notify);
896f97ea9   David Decotigny   lib: cpu_rmap: av...
295
296
  	if (rc) {
  		cpu_rmap_put(glue->rmap);
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
297
  		kfree(glue);
896f97ea9   David Decotigny   lib: cpu_rmap: av...
298
  	}
c39649c33   Ben Hutchings   lib: cpu_rmap: CP...
299
300
301
  	return rc;
  }
  EXPORT_SYMBOL(irq_cpu_rmap_add);