Blame view

security/device_cgroup.c 12.6 KB
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
1
2
3
4
5
6
7
8
9
10
11
  /*
   * dev_cgroup.c - device cgroup subsystem
   *
   * Copyright 2007 IBM Corp
   */
  
  #include <linux/device_cgroup.h>
  #include <linux/cgroup.h>
  #include <linux/ctype.h>
  #include <linux/list.h>
  #include <linux/uaccess.h>
29486df32   Serge E. Hallyn   cgroups: introduc...
12
  #include <linux/seq_file.h>
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
  
  #define ACC_MKNOD 1
  #define ACC_READ  2
  #define ACC_WRITE 4
  #define ACC_MASK (ACC_MKNOD | ACC_READ | ACC_WRITE)
  
  #define DEV_BLOCK 1
  #define DEV_CHAR  2
  #define DEV_ALL   4  /* this represents all devices */
  
  /*
   * whitelist locking rules:
   * cgroup_lock() cannot be taken under dev_cgroup->lock.
   * dev_cgroup->lock can be taken with or without cgroup_lock().
   *
   * modifications always require cgroup_lock
   * modifications to a list which is visible require the
   *   dev_cgroup->lock *and* cgroup_lock()
   * walking the list requires dev_cgroup->lock or cgroup_lock().
   *
   * reasoning: dev_whitelist_copy() needs to kmalloc, so needs
   *   a mutex, which the cgroup_lock() is.  Since modifying
   *   a visible list requires both locks, either lock can be
   *   taken for walking the list.
   */
  
  struct dev_whitelist_item {
  	u32 major, minor;
  	short type;
  	short access;
  	struct list_head list;
4efd1a1b2   Pavel Emelyanov   devcgroup: relax ...
44
  	struct rcu_head rcu;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
45
46
47
48
49
50
51
  };
  
  struct dev_cgroup {
  	struct cgroup_subsys_state css;
  	struct list_head whitelist;
  	spinlock_t lock;
  };
b66862f76   Pavel Emelyanov   devcgroup: make a...
52
53
54
55
  static inline struct dev_cgroup *css_to_devcgroup(struct cgroup_subsys_state *s)
  {
  	return container_of(s, struct dev_cgroup, css);
  }
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
56
57
  static inline struct dev_cgroup *cgroup_to_devcgroup(struct cgroup *cgroup)
  {
b66862f76   Pavel Emelyanov   devcgroup: make a...
58
  	return css_to_devcgroup(cgroup_subsys_state(cgroup, devices_subsys_id));
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
59
  }
f92523e3a   Paul Menage   cgroup files: con...
60
61
62
63
  static inline struct dev_cgroup *task_devcgroup(struct task_struct *task)
  {
  	return css_to_devcgroup(task_subsys_state(task, devices_subsys_id));
  }
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
  struct cgroup_subsys devices_subsys;
  
  static int devcgroup_can_attach(struct cgroup_subsys *ss,
  		struct cgroup *new_cgroup, struct task_struct *task)
  {
  	if (current != task && !capable(CAP_SYS_ADMIN))
  			return -EPERM;
  
  	return 0;
  }
  
  /*
   * called under cgroup_lock()
   */
  static int dev_whitelist_copy(struct list_head *dest, struct list_head *orig)
  {
  	struct dev_whitelist_item *wh, *tmp, *new;
  
  	list_for_each_entry(wh, orig, list) {
  		new = kmalloc(sizeof(*wh), GFP_KERNEL);
  		if (!new)
  			goto free_and_exit;
  		new->major = wh->major;
  		new->minor = wh->minor;
  		new->type = wh->type;
  		new->access = wh->access;
  		list_add_tail(&new->list, dest);
  	}
  
  	return 0;
  
  free_and_exit:
  	list_for_each_entry_safe(wh, tmp, dest, list) {
  		list_del(&wh->list);
  		kfree(wh);
  	}
  	return -ENOMEM;
  }
  
  /* Stupid prototype - don't bother combining existing entries */
  /*
   * called under cgroup_lock()
   * since the list is visible to other tasks, we need the spinlock also
   */
  static int dev_whitelist_add(struct dev_cgroup *dev_cgroup,
  			struct dev_whitelist_item *wh)
  {
d1ee2971f   Pavel Emelyanov   devscgroup: make ...
111
  	struct dev_whitelist_item *whcopy, *walk;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
112
113
114
115
116
117
118
  
  	whcopy = kmalloc(sizeof(*whcopy), GFP_KERNEL);
  	if (!whcopy)
  		return -ENOMEM;
  
  	memcpy(whcopy, wh, sizeof(*whcopy));
  	spin_lock(&dev_cgroup->lock);
d1ee2971f   Pavel Emelyanov   devscgroup: make ...
119
120
121
122
123
124
125
126
127
128
129
130
131
132
  	list_for_each_entry(walk, &dev_cgroup->whitelist, list) {
  		if (walk->type != wh->type)
  			continue;
  		if (walk->major != wh->major)
  			continue;
  		if (walk->minor != wh->minor)
  			continue;
  
  		walk->access |= wh->access;
  		kfree(whcopy);
  		whcopy = NULL;
  	}
  
  	if (whcopy != NULL)
4efd1a1b2   Pavel Emelyanov   devcgroup: relax ...
133
  		list_add_tail_rcu(&whcopy->list, &dev_cgroup->whitelist);
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
134
135
136
  	spin_unlock(&dev_cgroup->lock);
  	return 0;
  }
4efd1a1b2   Pavel Emelyanov   devcgroup: relax ...
137
138
139
140
141
142
143
  static void whitelist_item_free(struct rcu_head *rcu)
  {
  	struct dev_whitelist_item *item;
  
  	item = container_of(rcu, struct dev_whitelist_item, rcu);
  	kfree(item);
  }
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
  /*
   * called under cgroup_lock()
   * since the list is visible to other tasks, we need the spinlock also
   */
  static void dev_whitelist_rm(struct dev_cgroup *dev_cgroup,
  			struct dev_whitelist_item *wh)
  {
  	struct dev_whitelist_item *walk, *tmp;
  
  	spin_lock(&dev_cgroup->lock);
  	list_for_each_entry_safe(walk, tmp, &dev_cgroup->whitelist, list) {
  		if (walk->type == DEV_ALL)
  			goto remove;
  		if (walk->type != wh->type)
  			continue;
  		if (walk->major != ~0 && walk->major != wh->major)
  			continue;
  		if (walk->minor != ~0 && walk->minor != wh->minor)
  			continue;
  
  remove:
  		walk->access &= ~wh->access;
  		if (!walk->access) {
4efd1a1b2   Pavel Emelyanov   devcgroup: relax ...
167
168
  			list_del_rcu(&walk->list);
  			call_rcu(&walk->rcu, whitelist_item_free);
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
  		}
  	}
  	spin_unlock(&dev_cgroup->lock);
  }
  
  /*
   * called from kernel/cgroup.c with cgroup_lock() held.
   */
  static struct cgroup_subsys_state *devcgroup_create(struct cgroup_subsys *ss,
  						struct cgroup *cgroup)
  {
  	struct dev_cgroup *dev_cgroup, *parent_dev_cgroup;
  	struct cgroup *parent_cgroup;
  	int ret;
  
  	dev_cgroup = kzalloc(sizeof(*dev_cgroup), GFP_KERNEL);
  	if (!dev_cgroup)
  		return ERR_PTR(-ENOMEM);
  	INIT_LIST_HEAD(&dev_cgroup->whitelist);
  	parent_cgroup = cgroup->parent;
  
  	if (parent_cgroup == NULL) {
  		struct dev_whitelist_item *wh;
  		wh = kmalloc(sizeof(*wh), GFP_KERNEL);
  		if (!wh) {
  			kfree(dev_cgroup);
  			return ERR_PTR(-ENOMEM);
  		}
  		wh->minor = wh->major = ~0;
  		wh->type = DEV_ALL;
7759fc9d1   Li Zefan   devcgroup: code c...
199
  		wh->access = ACC_MASK;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
  		list_add(&wh->list, &dev_cgroup->whitelist);
  	} else {
  		parent_dev_cgroup = cgroup_to_devcgroup(parent_cgroup);
  		ret = dev_whitelist_copy(&dev_cgroup->whitelist,
  				&parent_dev_cgroup->whitelist);
  		if (ret) {
  			kfree(dev_cgroup);
  			return ERR_PTR(ret);
  		}
  	}
  
  	spin_lock_init(&dev_cgroup->lock);
  	return &dev_cgroup->css;
  }
  
  static void devcgroup_destroy(struct cgroup_subsys *ss,
  			struct cgroup *cgroup)
  {
  	struct dev_cgroup *dev_cgroup;
  	struct dev_whitelist_item *wh, *tmp;
  
  	dev_cgroup = cgroup_to_devcgroup(cgroup);
  	list_for_each_entry_safe(wh, tmp, &dev_cgroup->whitelist, list) {
  		list_del(&wh->list);
  		kfree(wh);
  	}
  	kfree(dev_cgroup);
  }
  
  #define DEVCG_ALLOW 1
  #define DEVCG_DENY 2
29486df32   Serge E. Hallyn   cgroups: introduc...
231
  #define DEVCG_LIST 3
17d213f80   Li Zefan   devcgroup: always...
232
  #define MAJMINLEN 13
29486df32   Serge E. Hallyn   cgroups: introduc...
233
  #define ACCLEN 4
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
234
235
236
237
  
  static void set_access(char *acc, short access)
  {
  	int idx = 0;
29486df32   Serge E. Hallyn   cgroups: introduc...
238
  	memset(acc, 0, ACCLEN);
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
  	if (access & ACC_READ)
  		acc[idx++] = 'r';
  	if (access & ACC_WRITE)
  		acc[idx++] = 'w';
  	if (access & ACC_MKNOD)
  		acc[idx++] = 'm';
  }
  
  static char type_to_char(short type)
  {
  	if (type == DEV_ALL)
  		return 'a';
  	if (type == DEV_CHAR)
  		return 'c';
  	if (type == DEV_BLOCK)
  		return 'b';
  	return 'X';
  }
29486df32   Serge E. Hallyn   cgroups: introduc...
257
  static void set_majmin(char *str, unsigned m)
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
258
  {
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
259
  	if (m == ~0)
7759fc9d1   Li Zefan   devcgroup: code c...
260
  		strcpy(str, "*");
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
261
  	else
7759fc9d1   Li Zefan   devcgroup: code c...
262
  		sprintf(str, "%u", m);
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
263
  }
29486df32   Serge E. Hallyn   cgroups: introduc...
264
265
  static int devcgroup_seq_read(struct cgroup *cgroup, struct cftype *cft,
  				struct seq_file *m)
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
266
  {
29486df32   Serge E. Hallyn   cgroups: introduc...
267
  	struct dev_cgroup *devcgroup = cgroup_to_devcgroup(cgroup);
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
268
  	struct dev_whitelist_item *wh;
29486df32   Serge E. Hallyn   cgroups: introduc...
269
  	char maj[MAJMINLEN], min[MAJMINLEN], acc[ACCLEN];
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
270

4efd1a1b2   Pavel Emelyanov   devcgroup: relax ...
271
272
  	rcu_read_lock();
  	list_for_each_entry_rcu(wh, &devcgroup->whitelist, list) {
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
273
  		set_access(acc, wh->access);
29486df32   Serge E. Hallyn   cgroups: introduc...
274
275
276
277
278
  		set_majmin(maj, wh->major);
  		set_majmin(min, wh->minor);
  		seq_printf(m, "%c %s:%s %s
  ", type_to_char(wh->type),
  			   maj, min, acc);
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
279
  	}
4efd1a1b2   Pavel Emelyanov   devcgroup: relax ...
280
  	rcu_read_unlock();
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
281

29486df32   Serge E. Hallyn   cgroups: introduc...
282
  	return 0;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
  }
  
  /*
   * may_access_whitelist:
   * does the access granted to dev_cgroup c contain the access
   * requested in whitelist item refwh.
   * return 1 if yes, 0 if no.
   * call with c->lock held
   */
  static int may_access_whitelist(struct dev_cgroup *c,
  				       struct dev_whitelist_item *refwh)
  {
  	struct dev_whitelist_item *whitem;
  
  	list_for_each_entry(whitem, &c->whitelist, list) {
  		if (whitem->type & DEV_ALL)
  			return 1;
  		if ((refwh->type & DEV_BLOCK) && !(whitem->type & DEV_BLOCK))
  			continue;
  		if ((refwh->type & DEV_CHAR) && !(whitem->type & DEV_CHAR))
  			continue;
  		if (whitem->major != ~0 && whitem->major != refwh->major)
  			continue;
  		if (whitem->minor != ~0 && whitem->minor != refwh->minor)
  			continue;
ec229e830   Li Zefan   devcgroup: fix pe...
308
  		if (refwh->access & (~whitem->access))
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
309
310
311
312
313
314
315
316
317
318
319
  			continue;
  		return 1;
  	}
  	return 0;
  }
  
  /*
   * parent_has_perm:
   * when adding a new allow rule to a device whitelist, the rule
   * must be allowed in the parent device
   */
f92523e3a   Paul Menage   cgroup files: con...
320
  static int parent_has_perm(struct dev_cgroup *childcg,
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
321
322
  				  struct dev_whitelist_item *wh)
  {
f92523e3a   Paul Menage   cgroup files: con...
323
  	struct cgroup *pcg = childcg->css.cgroup->parent;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
  	struct dev_cgroup *parent;
  	int ret;
  
  	if (!pcg)
  		return 1;
  	parent = cgroup_to_devcgroup(pcg);
  	spin_lock(&parent->lock);
  	ret = may_access_whitelist(parent, wh);
  	spin_unlock(&parent->lock);
  	return ret;
  }
  
  /*
   * Modify the whitelist using allow/deny rules.
   * CAP_SYS_ADMIN is needed for this.  It's at least separate from CAP_MKNOD
   * so we can give a container CAP_MKNOD to let it create devices but not
   * modify the whitelist.
   * It seems likely we'll want to add a CAP_CONTAINER capability to allow
   * us to also grant CAP_SYS_ADMIN to containers without giving away the
   * device whitelist controls, but for now we'll stick with CAP_SYS_ADMIN
   *
   * Taking rules away is always allowed (given CAP_SYS_ADMIN).  Granting
   * new access is only allowed if you're in the top-level cgroup, or your
   * parent cgroup has the access you're asking for.
   */
f92523e3a   Paul Menage   cgroup files: con...
349
350
  static int devcgroup_update_access(struct dev_cgroup *devcgroup,
  				   int filetype, const char *buffer)
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
351
  {
f92523e3a   Paul Menage   cgroup files: con...
352
353
  	struct dev_cgroup *cur_devcgroup;
  	const char *b;
7759fc9d1   Li Zefan   devcgroup: code c...
354
  	char *endp;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
355
356
357
358
359
  	int retval = 0, count;
  	struct dev_whitelist_item wh;
  
  	if (!capable(CAP_SYS_ADMIN))
  		return -EPERM;
f92523e3a   Paul Menage   cgroup files: con...
360
  	cur_devcgroup = task_devcgroup(current);
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
361
362
363
364
365
366
367
368
  
  	memset(&wh, 0, sizeof(wh));
  	b = buffer;
  
  	switch (*b) {
  	case 'a':
  		wh.type = DEV_ALL;
  		wh.access = ACC_MASK;
d823f6bfe   Li Zefan   devcgroup: fix od...
369
370
  		wh.major = ~0;
  		wh.minor = ~0;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
371
372
373
374
375
376
377
378
  		goto handle;
  	case 'b':
  		wh.type = DEV_BLOCK;
  		break;
  	case 'c':
  		wh.type = DEV_CHAR;
  		break;
  	default:
f92523e3a   Paul Menage   cgroup files: con...
379
  		return -EINVAL;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
380
381
  	}
  	b++;
f92523e3a   Paul Menage   cgroup files: con...
382
383
  	if (!isspace(*b))
  		return -EINVAL;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
384
385
386
387
388
  	b++;
  	if (*b == '*') {
  		wh.major = ~0;
  		b++;
  	} else if (isdigit(*b)) {
7759fc9d1   Li Zefan   devcgroup: code c...
389
390
  		wh.major = simple_strtoul(b, &endp, 10);
  		b = endp;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
391
  	} else {
f92523e3a   Paul Menage   cgroup files: con...
392
  		return -EINVAL;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
393
  	}
f92523e3a   Paul Menage   cgroup files: con...
394
395
  	if (*b != ':')
  		return -EINVAL;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
396
397
398
399
400
401
402
  	b++;
  
  	/* read minor */
  	if (*b == '*') {
  		wh.minor = ~0;
  		b++;
  	} else if (isdigit(*b)) {
7759fc9d1   Li Zefan   devcgroup: code c...
403
404
  		wh.minor = simple_strtoul(b, &endp, 10);
  		b = endp;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
405
  	} else {
f92523e3a   Paul Menage   cgroup files: con...
406
  		return -EINVAL;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
407
  	}
f92523e3a   Paul Menage   cgroup files: con...
408
409
  	if (!isspace(*b))
  		return -EINVAL;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
  	for (b++, count = 0; count < 3; count++, b++) {
  		switch (*b) {
  		case 'r':
  			wh.access |= ACC_READ;
  			break;
  		case 'w':
  			wh.access |= ACC_WRITE;
  			break;
  		case 'm':
  			wh.access |= ACC_MKNOD;
  			break;
  		case '
  ':
  		case '\0':
  			count = 3;
  			break;
  		default:
f92523e3a   Paul Menage   cgroup files: con...
427
  			return -EINVAL;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
428
429
430
431
432
433
434
  		}
  	}
  
  handle:
  	retval = 0;
  	switch (filetype) {
  	case DEVCG_ALLOW:
f92523e3a   Paul Menage   cgroup files: con...
435
436
437
  		if (!parent_has_perm(devcgroup, &wh))
  			return -EPERM;
  		return dev_whitelist_add(devcgroup, &wh);
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
438
439
440
441
  	case DEVCG_DENY:
  		dev_whitelist_rm(devcgroup, &wh);
  		break;
  	default:
f92523e3a   Paul Menage   cgroup files: con...
442
  		return -EINVAL;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
443
  	}
f92523e3a   Paul Menage   cgroup files: con...
444
445
  	return 0;
  }
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
446

f92523e3a   Paul Menage   cgroup files: con...
447
448
449
450
451
452
453
454
  static int devcgroup_access_write(struct cgroup *cgrp, struct cftype *cft,
  				  const char *buffer)
  {
  	int retval;
  	if (!cgroup_lock_live_group(cgrp))
  		return -ENODEV;
  	retval = devcgroup_update_access(cgroup_to_devcgroup(cgrp),
  					 cft->private, buffer);
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
455
  	cgroup_unlock();
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
456
457
458
459
460
461
  	return retval;
  }
  
  static struct cftype dev_cgroup_files[] = {
  	{
  		.name = "allow",
f92523e3a   Paul Menage   cgroup files: con...
462
  		.write_string  = devcgroup_access_write,
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
463
464
465
466
  		.private = DEVCG_ALLOW,
  	},
  	{
  		.name = "deny",
f92523e3a   Paul Menage   cgroup files: con...
467
  		.write_string = devcgroup_access_write,
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
468
469
  		.private = DEVCG_DENY,
  	},
29486df32   Serge E. Hallyn   cgroups: introduc...
470
471
472
473
474
  	{
  		.name = "list",
  		.read_seq_string = devcgroup_seq_read,
  		.private = DEVCG_LIST,
  	},
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
  };
  
  static int devcgroup_populate(struct cgroup_subsys *ss,
  				struct cgroup *cgroup)
  {
  	return cgroup_add_files(cgroup, ss, dev_cgroup_files,
  					ARRAY_SIZE(dev_cgroup_files));
  }
  
  struct cgroup_subsys devices_subsys = {
  	.name = "devices",
  	.can_attach = devcgroup_can_attach,
  	.create = devcgroup_create,
  	.destroy  = devcgroup_destroy,
  	.populate = devcgroup_populate,
  	.subsys_id = devices_subsys_id,
  };
  
  int devcgroup_inode_permission(struct inode *inode, int mask)
  {
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
495
496
497
498
499
500
501
502
  	struct dev_cgroup *dev_cgroup;
  	struct dev_whitelist_item *wh;
  
  	dev_t device = inode->i_rdev;
  	if (!device)
  		return 0;
  	if (!S_ISBLK(inode->i_mode) && !S_ISCHR(inode->i_mode))
  		return 0;
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
503

4efd1a1b2   Pavel Emelyanov   devcgroup: relax ...
504
  	rcu_read_lock();
36fd71d29   Li Zefan   devcgroup: fix ra...
505
506
  
  	dev_cgroup = task_devcgroup(current);
4efd1a1b2   Pavel Emelyanov   devcgroup: relax ...
507
  	list_for_each_entry_rcu(wh, &dev_cgroup->whitelist, list) {
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
  		if (wh->type & DEV_ALL)
  			goto acc_check;
  		if ((wh->type & DEV_BLOCK) && !S_ISBLK(inode->i_mode))
  			continue;
  		if ((wh->type & DEV_CHAR) && !S_ISCHR(inode->i_mode))
  			continue;
  		if (wh->major != ~0 && wh->major != imajor(inode))
  			continue;
  		if (wh->minor != ~0 && wh->minor != iminor(inode))
  			continue;
  acc_check:
  		if ((mask & MAY_WRITE) && !(wh->access & ACC_WRITE))
  			continue;
  		if ((mask & MAY_READ) && !(wh->access & ACC_READ))
  			continue;
4efd1a1b2   Pavel Emelyanov   devcgroup: relax ...
523
  		rcu_read_unlock();
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
524
525
  		return 0;
  	}
36fd71d29   Li Zefan   devcgroup: fix ra...
526

4efd1a1b2   Pavel Emelyanov   devcgroup: relax ...
527
  	rcu_read_unlock();
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
528
529
530
531
532
533
  
  	return -EPERM;
  }
  
  int devcgroup_inode_mknod(int mode, dev_t dev)
  {
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
534
535
  	struct dev_cgroup *dev_cgroup;
  	struct dev_whitelist_item *wh;
4efd1a1b2   Pavel Emelyanov   devcgroup: relax ...
536
  	rcu_read_lock();
36fd71d29   Li Zefan   devcgroup: fix ra...
537
538
  
  	dev_cgroup = task_devcgroup(current);
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
539
540
541
542
543
544
545
546
547
548
549
550
551
552
  	list_for_each_entry(wh, &dev_cgroup->whitelist, list) {
  		if (wh->type & DEV_ALL)
  			goto acc_check;
  		if ((wh->type & DEV_BLOCK) && !S_ISBLK(mode))
  			continue;
  		if ((wh->type & DEV_CHAR) && !S_ISCHR(mode))
  			continue;
  		if (wh->major != ~0 && wh->major != MAJOR(dev))
  			continue;
  		if (wh->minor != ~0 && wh->minor != MINOR(dev))
  			continue;
  acc_check:
  		if (!(wh->access & ACC_MKNOD))
  			continue;
4efd1a1b2   Pavel Emelyanov   devcgroup: relax ...
553
  		rcu_read_unlock();
08ce5f16e   Serge E. Hallyn   cgroups: implemen...
554
555
  		return 0;
  	}
36fd71d29   Li Zefan   devcgroup: fix ra...
556

4efd1a1b2   Pavel Emelyanov   devcgroup: relax ...
557
  	rcu_read_unlock();
36fd71d29   Li Zefan   devcgroup: fix ra...
558

08ce5f16e   Serge E. Hallyn   cgroups: implemen...
559
560
  	return -EPERM;
  }