Blame view

crypto/cryptd.c 36.4 KB
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
1
2
3
4
5
  /*
   * Software async crypto daemon.
   *
   * Copyright (c) 2006 Herbert Xu <herbert@gondor.apana.org.au>
   *
298c926c6   Adrian Hoban   crypto: cryptd - ...
6
7
8
9
10
11
12
   * Added AEAD support to cryptd.
   *    Authors: Tadeusz Struk (tadeusz.struk@intel.com)
   *             Adrian Hoban <adrian.hoban@intel.com>
   *             Gabriele Paoloni <gabriele.paoloni@intel.com>
   *             Aidan O'Mahony (aidan.o.mahony@intel.com)
   *    Copyright (c) 2010, Intel Corporation.
   *
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
13
14
15
16
17
18
   * This program is free software; you can redistribute it and/or modify it
   * under the terms of the GNU General Public License as published by the Free
   * Software Foundation; either version 2 of the License, or (at your option)
   * any later version.
   *
   */
18e33e6d5   Herbert Xu   crypto: hash - Mo...
19
  #include <crypto/internal/hash.h>
298c926c6   Adrian Hoban   crypto: cryptd - ...
20
  #include <crypto/internal/aead.h>
4e0958d19   Herbert Xu   crypto: cryptd - ...
21
  #include <crypto/internal/skcipher.h>
1cac2cbc7   Huang Ying   crypto: cryptd - ...
22
  #include <crypto/cryptd.h>
254eff771   Huang Ying   crypto: cryptd - ...
23
  #include <crypto/crypto_wq.h>
81760ea6a   Herbert Xu   crypto: cryptd - ...
24
  #include <linux/atomic.h>
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
25
26
27
  #include <linux/err.h>
  #include <linux/init.h>
  #include <linux/kernel.h>
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
28
29
  #include <linux/list.h>
  #include <linux/module.h>
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
30
31
32
  #include <linux/scatterlist.h>
  #include <linux/sched.h>
  #include <linux/slab.h>
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
33

81760ea6a   Herbert Xu   crypto: cryptd - ...
34
  #define CRYPTD_MAX_CPU_QLEN 1000
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
35

254eff771   Huang Ying   crypto: cryptd - ...
36
  struct cryptd_cpu_queue {
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
37
  	struct crypto_queue queue;
254eff771   Huang Ying   crypto: cryptd - ...
38
39
40
41
  	struct work_struct work;
  };
  
  struct cryptd_queue {
a29d8b8e2   Tejun Heo   percpu: add __per...
42
  	struct cryptd_cpu_queue __percpu *cpu_queue;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
43
44
45
46
  };
  
  struct cryptd_instance_ctx {
  	struct crypto_spawn spawn;
254eff771   Huang Ying   crypto: cryptd - ...
47
  	struct cryptd_queue *queue;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
48
  };
4e0958d19   Herbert Xu   crypto: cryptd - ...
49
50
51
52
  struct skcipherd_instance_ctx {
  	struct crypto_skcipher_spawn spawn;
  	struct cryptd_queue *queue;
  };
46309d893   Herbert Xu   crypto: cryptd - ...
53
54
55
56
  struct hashd_instance_ctx {
  	struct crypto_shash_spawn spawn;
  	struct cryptd_queue *queue;
  };
298c926c6   Adrian Hoban   crypto: cryptd - ...
57
58
59
60
  struct aead_instance_ctx {
  	struct crypto_aead_spawn aead_spawn;
  	struct cryptd_queue *queue;
  };
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
61
  struct cryptd_blkcipher_ctx {
81760ea6a   Herbert Xu   crypto: cryptd - ...
62
  	atomic_t refcnt;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
63
64
65
66
67
68
  	struct crypto_blkcipher *child;
  };
  
  struct cryptd_blkcipher_request_ctx {
  	crypto_completion_t complete;
  };
4e0958d19   Herbert Xu   crypto: cryptd - ...
69
70
71
72
73
74
75
76
  struct cryptd_skcipher_ctx {
  	atomic_t refcnt;
  	struct crypto_skcipher *child;
  };
  
  struct cryptd_skcipher_request_ctx {
  	crypto_completion_t complete;
  };
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
77
  struct cryptd_hash_ctx {
81760ea6a   Herbert Xu   crypto: cryptd - ...
78
  	atomic_t refcnt;
46309d893   Herbert Xu   crypto: cryptd - ...
79
  	struct crypto_shash *child;
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
80
81
82
83
  };
  
  struct cryptd_hash_request_ctx {
  	crypto_completion_t complete;
46309d893   Herbert Xu   crypto: cryptd - ...
84
  	struct shash_desc desc;
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
85
  };
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
86

298c926c6   Adrian Hoban   crypto: cryptd - ...
87
  struct cryptd_aead_ctx {
81760ea6a   Herbert Xu   crypto: cryptd - ...
88
  	atomic_t refcnt;
298c926c6   Adrian Hoban   crypto: cryptd - ...
89
90
91
92
93
94
  	struct crypto_aead *child;
  };
  
  struct cryptd_aead_request_ctx {
  	crypto_completion_t complete;
  };
254eff771   Huang Ying   crypto: cryptd - ...
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
  static void cryptd_queue_worker(struct work_struct *work);
  
  static int cryptd_init_queue(struct cryptd_queue *queue,
  			     unsigned int max_cpu_qlen)
  {
  	int cpu;
  	struct cryptd_cpu_queue *cpu_queue;
  
  	queue->cpu_queue = alloc_percpu(struct cryptd_cpu_queue);
  	if (!queue->cpu_queue)
  		return -ENOMEM;
  	for_each_possible_cpu(cpu) {
  		cpu_queue = per_cpu_ptr(queue->cpu_queue, cpu);
  		crypto_init_queue(&cpu_queue->queue, max_cpu_qlen);
  		INIT_WORK(&cpu_queue->work, cryptd_queue_worker);
  	}
  	return 0;
  }
  
  static void cryptd_fini_queue(struct cryptd_queue *queue)
  {
  	int cpu;
  	struct cryptd_cpu_queue *cpu_queue;
  
  	for_each_possible_cpu(cpu) {
  		cpu_queue = per_cpu_ptr(queue->cpu_queue, cpu);
  		BUG_ON(cpu_queue->queue.qlen);
  	}
  	free_percpu(queue->cpu_queue);
  }
  
  static int cryptd_enqueue_request(struct cryptd_queue *queue,
  				  struct crypto_async_request *request)
  {
  	int cpu, err;
  	struct cryptd_cpu_queue *cpu_queue;
81760ea6a   Herbert Xu   crypto: cryptd - ...
131
132
  	atomic_t *refcnt;
  	bool may_backlog;
254eff771   Huang Ying   crypto: cryptd - ...
133
134
  
  	cpu = get_cpu();
0b44f4861   Christoph Lameter   this_cpu: Use thi...
135
  	cpu_queue = this_cpu_ptr(queue->cpu_queue);
254eff771   Huang Ying   crypto: cryptd - ...
136
  	err = crypto_enqueue_request(&cpu_queue->queue, request);
81760ea6a   Herbert Xu   crypto: cryptd - ...
137
138
139
140
141
142
  
  	refcnt = crypto_tfm_ctx(request->tfm);
  	may_backlog = request->flags & CRYPTO_TFM_REQ_MAY_BACKLOG;
  
  	if (err == -EBUSY && !may_backlog)
  		goto out_put_cpu;
254eff771   Huang Ying   crypto: cryptd - ...
143
  	queue_work_on(cpu, kcrypto_wq, &cpu_queue->work);
81760ea6a   Herbert Xu   crypto: cryptd - ...
144
145
146
  
  	if (!atomic_read(refcnt))
  		goto out_put_cpu;
81760ea6a   Herbert Xu   crypto: cryptd - ...
147
148
149
  	atomic_inc(refcnt);
  
  out_put_cpu:
254eff771   Huang Ying   crypto: cryptd - ...
150
151
152
153
154
155
156
157
158
159
160
161
162
163
  	put_cpu();
  
  	return err;
  }
  
  /* Called in workqueue context, do one real cryption work (via
   * req->complete) and reschedule itself if there are more work to
   * do. */
  static void cryptd_queue_worker(struct work_struct *work)
  {
  	struct cryptd_cpu_queue *cpu_queue;
  	struct crypto_async_request *req, *backlog;
  
  	cpu_queue = container_of(work, struct cryptd_cpu_queue, work);
9efade1b3   Jussi Kivilinna   crypto: cryptd - ...
164
165
166
167
168
169
170
  	/*
  	 * Only handle one request at a time to avoid hogging crypto workqueue.
  	 * preempt_disable/enable is used to prevent being preempted by
  	 * cryptd_enqueue_request(). local_bh_disable/enable is used to prevent
  	 * cryptd_enqueue_request() being accessed from software interrupts.
  	 */
  	local_bh_disable();
254eff771   Huang Ying   crypto: cryptd - ...
171
172
173
174
  	preempt_disable();
  	backlog = crypto_get_backlog(&cpu_queue->queue);
  	req = crypto_dequeue_request(&cpu_queue->queue);
  	preempt_enable();
9efade1b3   Jussi Kivilinna   crypto: cryptd - ...
175
  	local_bh_enable();
254eff771   Huang Ying   crypto: cryptd - ...
176
177
178
179
180
181
182
183
184
185
186
187
188
  
  	if (!req)
  		return;
  
  	if (backlog)
  		backlog->complete(backlog, -EINPROGRESS);
  	req->complete(req, 0);
  
  	if (cpu_queue->queue.qlen)
  		queue_work(kcrypto_wq, &cpu_queue->work);
  }
  
  static inline struct cryptd_queue *cryptd_get_queue(struct crypto_tfm *tfm)
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
189
190
191
  {
  	struct crypto_instance *inst = crypto_tfm_alg_instance(tfm);
  	struct cryptd_instance_ctx *ictx = crypto_instance_ctx(inst);
254eff771   Huang Ying   crypto: cryptd - ...
192
  	return ictx->queue;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
193
  }
466a7b9e3   Stephan Mueller   crypto: cryptd - ...
194
195
196
197
198
199
200
201
  static inline void cryptd_check_internal(struct rtattr **tb, u32 *type,
  					 u32 *mask)
  {
  	struct crypto_attr_type *algt;
  
  	algt = crypto_get_attr_type(tb);
  	if (IS_ERR(algt))
  		return;
f6da32059   Herbert Xu   crypto: cryptd - ...
202

5e4b8c1fc   Herbert Xu   crypto: aead - Re...
203
204
  	*type |= algt->type & CRYPTO_ALG_INTERNAL;
  	*mask |= algt->mask & CRYPTO_ALG_INTERNAL;
466a7b9e3   Stephan Mueller   crypto: cryptd - ...
205
  }
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
  static int cryptd_blkcipher_setkey(struct crypto_ablkcipher *parent,
  				   const u8 *key, unsigned int keylen)
  {
  	struct cryptd_blkcipher_ctx *ctx = crypto_ablkcipher_ctx(parent);
  	struct crypto_blkcipher *child = ctx->child;
  	int err;
  
  	crypto_blkcipher_clear_flags(child, CRYPTO_TFM_REQ_MASK);
  	crypto_blkcipher_set_flags(child, crypto_ablkcipher_get_flags(parent) &
  					  CRYPTO_TFM_REQ_MASK);
  	err = crypto_blkcipher_setkey(child, key, keylen);
  	crypto_ablkcipher_set_flags(parent, crypto_blkcipher_get_flags(child) &
  					    CRYPTO_TFM_RES_MASK);
  	return err;
  }
  
  static void cryptd_blkcipher_crypt(struct ablkcipher_request *req,
  				   struct crypto_blkcipher *child,
  				   int err,
  				   int (*crypt)(struct blkcipher_desc *desc,
  						struct scatterlist *dst,
  						struct scatterlist *src,
  						unsigned int len))
  {
  	struct cryptd_blkcipher_request_ctx *rctx;
81760ea6a   Herbert Xu   crypto: cryptd - ...
231
232
  	struct cryptd_blkcipher_ctx *ctx;
  	struct crypto_ablkcipher *tfm;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
233
  	struct blkcipher_desc desc;
81760ea6a   Herbert Xu   crypto: cryptd - ...
234
  	int refcnt;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
235
236
  
  	rctx = ablkcipher_request_ctx(req);
93aa7f8a1   Herbert Xu   [CRYPTO] cryptd: ...
237
238
  	if (unlikely(err == -EINPROGRESS))
  		goto out;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
239
240
241
242
243
244
245
246
  
  	desc.tfm = child;
  	desc.info = req->info;
  	desc.flags = CRYPTO_TFM_REQ_MAY_SLEEP;
  
  	err = crypt(&desc, req->dst, req->src, req->nbytes);
  
  	req->base.complete = rctx->complete;
93aa7f8a1   Herbert Xu   [CRYPTO] cryptd: ...
247
  out:
81760ea6a   Herbert Xu   crypto: cryptd - ...
248
249
250
  	tfm = crypto_ablkcipher_reqtfm(req);
  	ctx = crypto_ablkcipher_ctx(tfm);
  	refcnt = atomic_read(&ctx->refcnt);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
251
  	local_bh_disable();
93aa7f8a1   Herbert Xu   [CRYPTO] cryptd: ...
252
  	rctx->complete(&req->base, err);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
253
  	local_bh_enable();
81760ea6a   Herbert Xu   crypto: cryptd - ...
254
255
256
  
  	if (err != -EINPROGRESS && refcnt && atomic_dec_and_test(&ctx->refcnt))
  		crypto_free_ablkcipher(tfm);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
  }
  
  static void cryptd_blkcipher_encrypt(struct crypto_async_request *req, int err)
  {
  	struct cryptd_blkcipher_ctx *ctx = crypto_tfm_ctx(req->tfm);
  	struct crypto_blkcipher *child = ctx->child;
  
  	cryptd_blkcipher_crypt(ablkcipher_request_cast(req), child, err,
  			       crypto_blkcipher_crt(child)->encrypt);
  }
  
  static void cryptd_blkcipher_decrypt(struct crypto_async_request *req, int err)
  {
  	struct cryptd_blkcipher_ctx *ctx = crypto_tfm_ctx(req->tfm);
  	struct crypto_blkcipher *child = ctx->child;
  
  	cryptd_blkcipher_crypt(ablkcipher_request_cast(req), child, err,
  			       crypto_blkcipher_crt(child)->decrypt);
  }
  
  static int cryptd_blkcipher_enqueue(struct ablkcipher_request *req,
3e3dc25fe   Mark Rustad   crypto: Resolve s...
278
  				    crypto_completion_t compl)
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
279
280
281
  {
  	struct cryptd_blkcipher_request_ctx *rctx = ablkcipher_request_ctx(req);
  	struct crypto_ablkcipher *tfm = crypto_ablkcipher_reqtfm(req);
254eff771   Huang Ying   crypto: cryptd - ...
282
  	struct cryptd_queue *queue;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
283

254eff771   Huang Ying   crypto: cryptd - ...
284
  	queue = cryptd_get_queue(crypto_ablkcipher_tfm(tfm));
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
285
  	rctx->complete = req->base.complete;
3e3dc25fe   Mark Rustad   crypto: Resolve s...
286
  	req->base.complete = compl;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
287

254eff771   Huang Ying   crypto: cryptd - ...
288
  	return cryptd_enqueue_request(queue, &req->base);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
  }
  
  static int cryptd_blkcipher_encrypt_enqueue(struct ablkcipher_request *req)
  {
  	return cryptd_blkcipher_enqueue(req, cryptd_blkcipher_encrypt);
  }
  
  static int cryptd_blkcipher_decrypt_enqueue(struct ablkcipher_request *req)
  {
  	return cryptd_blkcipher_enqueue(req, cryptd_blkcipher_decrypt);
  }
  
  static int cryptd_blkcipher_init_tfm(struct crypto_tfm *tfm)
  {
  	struct crypto_instance *inst = crypto_tfm_alg_instance(tfm);
  	struct cryptd_instance_ctx *ictx = crypto_instance_ctx(inst);
  	struct crypto_spawn *spawn = &ictx->spawn;
  	struct cryptd_blkcipher_ctx *ctx = crypto_tfm_ctx(tfm);
  	struct crypto_blkcipher *cipher;
  
  	cipher = crypto_spawn_blkcipher(spawn);
  	if (IS_ERR(cipher))
  		return PTR_ERR(cipher);
  
  	ctx->child = cipher;
  	tfm->crt_ablkcipher.reqsize =
  		sizeof(struct cryptd_blkcipher_request_ctx);
  	return 0;
  }
  
  static void cryptd_blkcipher_exit_tfm(struct crypto_tfm *tfm)
  {
  	struct cryptd_blkcipher_ctx *ctx = crypto_tfm_ctx(tfm);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
322
323
324
  
  	crypto_free_blkcipher(ctx->child);
  }
9b8c456e0   Herbert Xu   crypto: cryptd - ...
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
  static int cryptd_init_instance(struct crypto_instance *inst,
  				struct crypto_alg *alg)
  {
  	if (snprintf(inst->alg.cra_driver_name, CRYPTO_MAX_ALG_NAME,
  		     "cryptd(%s)",
  		     alg->cra_driver_name) >= CRYPTO_MAX_ALG_NAME)
  		return -ENAMETOOLONG;
  
  	memcpy(inst->alg.cra_name, alg->cra_name, CRYPTO_MAX_ALG_NAME);
  
  	inst->alg.cra_priority = alg->cra_priority + 50;
  	inst->alg.cra_blocksize = alg->cra_blocksize;
  	inst->alg.cra_alignmask = alg->cra_alignmask;
  
  	return 0;
  }
0b535adfb   Herbert Xu   crypto: cryptd - ...
341
342
  static void *cryptd_alloc_instance(struct crypto_alg *alg, unsigned int head,
  				   unsigned int tail)
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
343
  {
0b535adfb   Herbert Xu   crypto: cryptd - ...
344
  	char *p;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
345
  	struct crypto_instance *inst;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
346
  	int err;
0b535adfb   Herbert Xu   crypto: cryptd - ...
347
348
349
350
351
  	p = kzalloc(head + sizeof(*inst) + tail, GFP_KERNEL);
  	if (!p)
  		return ERR_PTR(-ENOMEM);
  
  	inst = (void *)(p + head);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
352

9b8c456e0   Herbert Xu   crypto: cryptd - ...
353
354
  	err = cryptd_init_instance(inst, alg);
  	if (err)
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
355
  		goto out_free_inst;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
356
  out:
0b535adfb   Herbert Xu   crypto: cryptd - ...
357
  	return p;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
358
359
  
  out_free_inst:
0b535adfb   Herbert Xu   crypto: cryptd - ...
360
361
  	kfree(p);
  	p = ERR_PTR(err);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
362
363
  	goto out;
  }
9cd899a32   Herbert Xu   crypto: cryptd - ...
364
365
366
  static int cryptd_create_blkcipher(struct crypto_template *tmpl,
  				   struct rtattr **tb,
  				   struct cryptd_queue *queue)
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
367
  {
46309d893   Herbert Xu   crypto: cryptd - ...
368
  	struct cryptd_instance_ctx *ctx;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
369
370
  	struct crypto_instance *inst;
  	struct crypto_alg *alg;
466a7b9e3   Stephan Mueller   crypto: cryptd - ...
371
372
  	u32 type = CRYPTO_ALG_TYPE_BLKCIPHER;
  	u32 mask = CRYPTO_ALG_TYPE_MASK;
46309d893   Herbert Xu   crypto: cryptd - ...
373
  	int err;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
374

466a7b9e3   Stephan Mueller   crypto: cryptd - ...
375
376
377
  	cryptd_check_internal(tb, &type, &mask);
  
  	alg = crypto_get_attr_alg(tb, type, mask);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
378
  	if (IS_ERR(alg))
9cd899a32   Herbert Xu   crypto: cryptd - ...
379
  		return PTR_ERR(alg);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
380

0b535adfb   Herbert Xu   crypto: cryptd - ...
381
  	inst = cryptd_alloc_instance(alg, 0, sizeof(*ctx));
05ed8758f   Steffen Klassert   crypto: cryptd - ...
382
  	err = PTR_ERR(inst);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
383
384
  	if (IS_ERR(inst))
  		goto out_put_alg;
46309d893   Herbert Xu   crypto: cryptd - ...
385
386
387
388
389
390
391
  	ctx = crypto_instance_ctx(inst);
  	ctx->queue = queue;
  
  	err = crypto_init_spawn(&ctx->spawn, alg, inst,
  				CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_ASYNC);
  	if (err)
  		goto out_free_inst;
466a7b9e3   Stephan Mueller   crypto: cryptd - ...
392
393
394
395
  	type = CRYPTO_ALG_TYPE_ABLKCIPHER | CRYPTO_ALG_ASYNC;
  	if (alg->cra_flags & CRYPTO_ALG_INTERNAL)
  		type |= CRYPTO_ALG_INTERNAL;
  	inst->alg.cra_flags = type;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
396
397
398
399
400
  	inst->alg.cra_type = &crypto_ablkcipher_type;
  
  	inst->alg.cra_ablkcipher.ivsize = alg->cra_blkcipher.ivsize;
  	inst->alg.cra_ablkcipher.min_keysize = alg->cra_blkcipher.min_keysize;
  	inst->alg.cra_ablkcipher.max_keysize = alg->cra_blkcipher.max_keysize;
927eead52   Herbert Xu   [CRYPTO] cryptd: ...
401
  	inst->alg.cra_ablkcipher.geniv = alg->cra_blkcipher.geniv;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
402
403
404
405
406
407
408
409
  	inst->alg.cra_ctxsize = sizeof(struct cryptd_blkcipher_ctx);
  
  	inst->alg.cra_init = cryptd_blkcipher_init_tfm;
  	inst->alg.cra_exit = cryptd_blkcipher_exit_tfm;
  
  	inst->alg.cra_ablkcipher.setkey = cryptd_blkcipher_setkey;
  	inst->alg.cra_ablkcipher.encrypt = cryptd_blkcipher_encrypt_enqueue;
  	inst->alg.cra_ablkcipher.decrypt = cryptd_blkcipher_decrypt_enqueue;
9cd899a32   Herbert Xu   crypto: cryptd - ...
410
411
412
413
414
415
  	err = crypto_register_instance(tmpl, inst);
  	if (err) {
  		crypto_drop_spawn(&ctx->spawn);
  out_free_inst:
  		kfree(inst);
  	}
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
416
417
  out_put_alg:
  	crypto_mod_put(alg);
9cd899a32   Herbert Xu   crypto: cryptd - ...
418
  	return err;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
419
  }
4e0958d19   Herbert Xu   crypto: cryptd - ...
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
  static int cryptd_skcipher_setkey(struct crypto_skcipher *parent,
  				  const u8 *key, unsigned int keylen)
  {
  	struct cryptd_skcipher_ctx *ctx = crypto_skcipher_ctx(parent);
  	struct crypto_skcipher *child = ctx->child;
  	int err;
  
  	crypto_skcipher_clear_flags(child, CRYPTO_TFM_REQ_MASK);
  	crypto_skcipher_set_flags(child, crypto_skcipher_get_flags(parent) &
  					 CRYPTO_TFM_REQ_MASK);
  	err = crypto_skcipher_setkey(child, key, keylen);
  	crypto_skcipher_set_flags(parent, crypto_skcipher_get_flags(child) &
  					  CRYPTO_TFM_RES_MASK);
  	return err;
  }
  
  static void cryptd_skcipher_complete(struct skcipher_request *req, int err)
  {
  	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
  	struct cryptd_skcipher_ctx *ctx = crypto_skcipher_ctx(tfm);
  	struct cryptd_skcipher_request_ctx *rctx = skcipher_request_ctx(req);
  	int refcnt = atomic_read(&ctx->refcnt);
  
  	local_bh_disable();
  	rctx->complete(&req->base, err);
  	local_bh_enable();
  
  	if (err != -EINPROGRESS && refcnt && atomic_dec_and_test(&ctx->refcnt))
  		crypto_free_skcipher(tfm);
  }
  
  static void cryptd_skcipher_encrypt(struct crypto_async_request *base,
  				    int err)
  {
  	struct skcipher_request *req = skcipher_request_cast(base);
  	struct cryptd_skcipher_request_ctx *rctx = skcipher_request_ctx(req);
  	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
  	struct cryptd_skcipher_ctx *ctx = crypto_skcipher_ctx(tfm);
  	struct crypto_skcipher *child = ctx->child;
  	SKCIPHER_REQUEST_ON_STACK(subreq, child);
  
  	if (unlikely(err == -EINPROGRESS))
  		goto out;
  
  	skcipher_request_set_tfm(subreq, child);
  	skcipher_request_set_callback(subreq, CRYPTO_TFM_REQ_MAY_SLEEP,
  				      NULL, NULL);
  	skcipher_request_set_crypt(subreq, req->src, req->dst, req->cryptlen,
  				   req->iv);
  
  	err = crypto_skcipher_encrypt(subreq);
  	skcipher_request_zero(subreq);
  
  	req->base.complete = rctx->complete;
  
  out:
  	cryptd_skcipher_complete(req, err);
  }
  
  static void cryptd_skcipher_decrypt(struct crypto_async_request *base,
  				    int err)
  {
  	struct skcipher_request *req = skcipher_request_cast(base);
  	struct cryptd_skcipher_request_ctx *rctx = skcipher_request_ctx(req);
  	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
  	struct cryptd_skcipher_ctx *ctx = crypto_skcipher_ctx(tfm);
  	struct crypto_skcipher *child = ctx->child;
  	SKCIPHER_REQUEST_ON_STACK(subreq, child);
  
  	if (unlikely(err == -EINPROGRESS))
  		goto out;
  
  	skcipher_request_set_tfm(subreq, child);
  	skcipher_request_set_callback(subreq, CRYPTO_TFM_REQ_MAY_SLEEP,
  				      NULL, NULL);
  	skcipher_request_set_crypt(subreq, req->src, req->dst, req->cryptlen,
  				   req->iv);
  
  	err = crypto_skcipher_decrypt(subreq);
  	skcipher_request_zero(subreq);
  
  	req->base.complete = rctx->complete;
  
  out:
  	cryptd_skcipher_complete(req, err);
  }
  
  static int cryptd_skcipher_enqueue(struct skcipher_request *req,
  				   crypto_completion_t compl)
  {
  	struct cryptd_skcipher_request_ctx *rctx = skcipher_request_ctx(req);
  	struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(req);
  	struct cryptd_queue *queue;
  
  	queue = cryptd_get_queue(crypto_skcipher_tfm(tfm));
  	rctx->complete = req->base.complete;
  	req->base.complete = compl;
  
  	return cryptd_enqueue_request(queue, &req->base);
  }
  
  static int cryptd_skcipher_encrypt_enqueue(struct skcipher_request *req)
  {
  	return cryptd_skcipher_enqueue(req, cryptd_skcipher_encrypt);
  }
  
  static int cryptd_skcipher_decrypt_enqueue(struct skcipher_request *req)
  {
  	return cryptd_skcipher_enqueue(req, cryptd_skcipher_decrypt);
  }
  
  static int cryptd_skcipher_init_tfm(struct crypto_skcipher *tfm)
  {
  	struct skcipher_instance *inst = skcipher_alg_instance(tfm);
  	struct skcipherd_instance_ctx *ictx = skcipher_instance_ctx(inst);
  	struct crypto_skcipher_spawn *spawn = &ictx->spawn;
  	struct cryptd_skcipher_ctx *ctx = crypto_skcipher_ctx(tfm);
  	struct crypto_skcipher *cipher;
  
  	cipher = crypto_spawn_skcipher(spawn);
  	if (IS_ERR(cipher))
  		return PTR_ERR(cipher);
  
  	ctx->child = cipher;
  	crypto_skcipher_set_reqsize(
  		tfm, sizeof(struct cryptd_skcipher_request_ctx));
  	return 0;
  }
  
  static void cryptd_skcipher_exit_tfm(struct crypto_skcipher *tfm)
  {
  	struct cryptd_skcipher_ctx *ctx = crypto_skcipher_ctx(tfm);
  
  	crypto_free_skcipher(ctx->child);
  }
  
  static void cryptd_skcipher_free(struct skcipher_instance *inst)
  {
  	struct skcipherd_instance_ctx *ctx = skcipher_instance_ctx(inst);
  
  	crypto_drop_skcipher(&ctx->spawn);
  }
  
  static int cryptd_create_skcipher(struct crypto_template *tmpl,
  				  struct rtattr **tb,
  				  struct cryptd_queue *queue)
  {
  	struct skcipherd_instance_ctx *ctx;
  	struct skcipher_instance *inst;
  	struct skcipher_alg *alg;
  	const char *name;
  	u32 type;
  	u32 mask;
  	int err;
  
  	type = 0;
  	mask = CRYPTO_ALG_ASYNC;
  
  	cryptd_check_internal(tb, &type, &mask);
  
  	name = crypto_attr_alg_name(tb[1]);
  	if (IS_ERR(name))
  		return PTR_ERR(name);
  
  	inst = kzalloc(sizeof(*inst) + sizeof(*ctx), GFP_KERNEL);
  	if (!inst)
  		return -ENOMEM;
  
  	ctx = skcipher_instance_ctx(inst);
  	ctx->queue = queue;
  
  	crypto_set_skcipher_spawn(&ctx->spawn, skcipher_crypto_instance(inst));
  	err = crypto_grab_skcipher(&ctx->spawn, name, type, mask);
  	if (err)
  		goto out_free_inst;
  
  	alg = crypto_spawn_skcipher_alg(&ctx->spawn);
  	err = cryptd_init_instance(skcipher_crypto_instance(inst), &alg->base);
  	if (err)
  		goto out_drop_skcipher;
  
  	inst->alg.base.cra_flags = CRYPTO_ALG_ASYNC |
  				   (alg->base.cra_flags & CRYPTO_ALG_INTERNAL);
  
  	inst->alg.ivsize = crypto_skcipher_alg_ivsize(alg);
  	inst->alg.chunksize = crypto_skcipher_alg_chunksize(alg);
  	inst->alg.min_keysize = crypto_skcipher_alg_min_keysize(alg);
  	inst->alg.max_keysize = crypto_skcipher_alg_max_keysize(alg);
  
  	inst->alg.base.cra_ctxsize = sizeof(struct cryptd_skcipher_ctx);
  
  	inst->alg.init = cryptd_skcipher_init_tfm;
  	inst->alg.exit = cryptd_skcipher_exit_tfm;
  
  	inst->alg.setkey = cryptd_skcipher_setkey;
  	inst->alg.encrypt = cryptd_skcipher_encrypt_enqueue;
  	inst->alg.decrypt = cryptd_skcipher_decrypt_enqueue;
  
  	inst->free = cryptd_skcipher_free;
  
  	err = skcipher_register_instance(tmpl, inst);
  	if (err) {
  out_drop_skcipher:
  		crypto_drop_skcipher(&ctx->spawn);
  out_free_inst:
  		kfree(inst);
  	}
  	return err;
  }
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
629
630
631
  static int cryptd_hash_init_tfm(struct crypto_tfm *tfm)
  {
  	struct crypto_instance *inst = crypto_tfm_alg_instance(tfm);
46309d893   Herbert Xu   crypto: cryptd - ...
632
633
  	struct hashd_instance_ctx *ictx = crypto_instance_ctx(inst);
  	struct crypto_shash_spawn *spawn = &ictx->spawn;
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
634
  	struct cryptd_hash_ctx *ctx = crypto_tfm_ctx(tfm);
46309d893   Herbert Xu   crypto: cryptd - ...
635
  	struct crypto_shash *hash;
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
636

46309d893   Herbert Xu   crypto: cryptd - ...
637
638
639
  	hash = crypto_spawn_shash(spawn);
  	if (IS_ERR(hash))
  		return PTR_ERR(hash);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
640

46309d893   Herbert Xu   crypto: cryptd - ...
641
  	ctx->child = hash;
0d6669e2b   Herbert Xu   crypto: cryptd - ...
642
643
644
  	crypto_ahash_set_reqsize(__crypto_ahash_cast(tfm),
  				 sizeof(struct cryptd_hash_request_ctx) +
  				 crypto_shash_descsize(hash));
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
645
646
647
648
649
650
  	return 0;
  }
  
  static void cryptd_hash_exit_tfm(struct crypto_tfm *tfm)
  {
  	struct cryptd_hash_ctx *ctx = crypto_tfm_ctx(tfm);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
651

46309d893   Herbert Xu   crypto: cryptd - ...
652
  	crypto_free_shash(ctx->child);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
653
654
655
656
657
658
  }
  
  static int cryptd_hash_setkey(struct crypto_ahash *parent,
  				   const u8 *key, unsigned int keylen)
  {
  	struct cryptd_hash_ctx *ctx   = crypto_ahash_ctx(parent);
46309d893   Herbert Xu   crypto: cryptd - ...
659
  	struct crypto_shash *child = ctx->child;
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
660
  	int err;
46309d893   Herbert Xu   crypto: cryptd - ...
661
662
663
664
665
666
  	crypto_shash_clear_flags(child, CRYPTO_TFM_REQ_MASK);
  	crypto_shash_set_flags(child, crypto_ahash_get_flags(parent) &
  				      CRYPTO_TFM_REQ_MASK);
  	err = crypto_shash_setkey(child, key, keylen);
  	crypto_ahash_set_flags(parent, crypto_shash_get_flags(child) &
  				       CRYPTO_TFM_RES_MASK);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
667
668
669
670
  	return err;
  }
  
  static int cryptd_hash_enqueue(struct ahash_request *req,
3e3dc25fe   Mark Rustad   crypto: Resolve s...
671
  				crypto_completion_t compl)
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
672
673
674
  {
  	struct cryptd_hash_request_ctx *rctx = ahash_request_ctx(req);
  	struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
254eff771   Huang Ying   crypto: cryptd - ...
675
676
  	struct cryptd_queue *queue =
  		cryptd_get_queue(crypto_ahash_tfm(tfm));
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
677
678
  
  	rctx->complete = req->base.complete;
3e3dc25fe   Mark Rustad   crypto: Resolve s...
679
  	req->base.complete = compl;
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
680

254eff771   Huang Ying   crypto: cryptd - ...
681
  	return cryptd_enqueue_request(queue, &req->base);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
682
  }
81760ea6a   Herbert Xu   crypto: cryptd - ...
683
684
685
686
687
688
689
690
691
692
693
694
695
696
  static void cryptd_hash_complete(struct ahash_request *req, int err)
  {
  	struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
  	struct cryptd_hash_ctx *ctx = crypto_ahash_ctx(tfm);
  	struct cryptd_hash_request_ctx *rctx = ahash_request_ctx(req);
  	int refcnt = atomic_read(&ctx->refcnt);
  
  	local_bh_disable();
  	rctx->complete(&req->base, err);
  	local_bh_enable();
  
  	if (err != -EINPROGRESS && refcnt && atomic_dec_and_test(&ctx->refcnt))
  		crypto_free_ahash(tfm);
  }
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
697
698
  static void cryptd_hash_init(struct crypto_async_request *req_async, int err)
  {
46309d893   Herbert Xu   crypto: cryptd - ...
699
700
701
702
703
  	struct cryptd_hash_ctx *ctx = crypto_tfm_ctx(req_async->tfm);
  	struct crypto_shash *child = ctx->child;
  	struct ahash_request *req = ahash_request_cast(req_async);
  	struct cryptd_hash_request_ctx *rctx = ahash_request_ctx(req);
  	struct shash_desc *desc = &rctx->desc;
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
704
705
706
  
  	if (unlikely(err == -EINPROGRESS))
  		goto out;
46309d893   Herbert Xu   crypto: cryptd - ...
707
708
  	desc->tfm = child;
  	desc->flags = CRYPTO_TFM_REQ_MAY_SLEEP;
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
709

46309d893   Herbert Xu   crypto: cryptd - ...
710
  	err = crypto_shash_init(desc);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
711
712
713
714
  
  	req->base.complete = rctx->complete;
  
  out:
81760ea6a   Herbert Xu   crypto: cryptd - ...
715
  	cryptd_hash_complete(req, err);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
716
717
718
719
720
721
722
723
724
  }
  
  static int cryptd_hash_init_enqueue(struct ahash_request *req)
  {
  	return cryptd_hash_enqueue(req, cryptd_hash_init);
  }
  
  static void cryptd_hash_update(struct crypto_async_request *req_async, int err)
  {
46309d893   Herbert Xu   crypto: cryptd - ...
725
  	struct ahash_request *req = ahash_request_cast(req_async);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
726
  	struct cryptd_hash_request_ctx *rctx;
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
727
728
729
730
731
  
  	rctx = ahash_request_ctx(req);
  
  	if (unlikely(err == -EINPROGRESS))
  		goto out;
46309d893   Herbert Xu   crypto: cryptd - ...
732
  	err = shash_ahash_update(req, &rctx->desc);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
733
734
735
736
  
  	req->base.complete = rctx->complete;
  
  out:
81760ea6a   Herbert Xu   crypto: cryptd - ...
737
  	cryptd_hash_complete(req, err);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
738
739
740
741
742
743
744
745
746
  }
  
  static int cryptd_hash_update_enqueue(struct ahash_request *req)
  {
  	return cryptd_hash_enqueue(req, cryptd_hash_update);
  }
  
  static void cryptd_hash_final(struct crypto_async_request *req_async, int err)
  {
46309d893   Herbert Xu   crypto: cryptd - ...
747
748
  	struct ahash_request *req = ahash_request_cast(req_async);
  	struct cryptd_hash_request_ctx *rctx = ahash_request_ctx(req);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
749
750
751
  
  	if (unlikely(err == -EINPROGRESS))
  		goto out;
46309d893   Herbert Xu   crypto: cryptd - ...
752
  	err = crypto_shash_final(&rctx->desc, req->result);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
753
754
755
756
  
  	req->base.complete = rctx->complete;
  
  out:
81760ea6a   Herbert Xu   crypto: cryptd - ...
757
  	cryptd_hash_complete(req, err);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
758
759
760
761
762
763
  }
  
  static int cryptd_hash_final_enqueue(struct ahash_request *req)
  {
  	return cryptd_hash_enqueue(req, cryptd_hash_final);
  }
6fba00d17   Herbert Xu   crypto: cryptd - ...
764
765
766
767
768
769
770
771
772
773
774
775
776
  static void cryptd_hash_finup(struct crypto_async_request *req_async, int err)
  {
  	struct ahash_request *req = ahash_request_cast(req_async);
  	struct cryptd_hash_request_ctx *rctx = ahash_request_ctx(req);
  
  	if (unlikely(err == -EINPROGRESS))
  		goto out;
  
  	err = shash_ahash_finup(req, &rctx->desc);
  
  	req->base.complete = rctx->complete;
  
  out:
81760ea6a   Herbert Xu   crypto: cryptd - ...
777
  	cryptd_hash_complete(req, err);
6fba00d17   Herbert Xu   crypto: cryptd - ...
778
779
780
781
782
783
  }
  
  static int cryptd_hash_finup_enqueue(struct ahash_request *req)
  {
  	return cryptd_hash_enqueue(req, cryptd_hash_finup);
  }
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
784
785
  static void cryptd_hash_digest(struct crypto_async_request *req_async, int err)
  {
46309d893   Herbert Xu   crypto: cryptd - ...
786
787
788
789
790
  	struct cryptd_hash_ctx *ctx = crypto_tfm_ctx(req_async->tfm);
  	struct crypto_shash *child = ctx->child;
  	struct ahash_request *req = ahash_request_cast(req_async);
  	struct cryptd_hash_request_ctx *rctx = ahash_request_ctx(req);
  	struct shash_desc *desc = &rctx->desc;
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
791
792
793
  
  	if (unlikely(err == -EINPROGRESS))
  		goto out;
46309d893   Herbert Xu   crypto: cryptd - ...
794
795
  	desc->tfm = child;
  	desc->flags = CRYPTO_TFM_REQ_MAY_SLEEP;
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
796

46309d893   Herbert Xu   crypto: cryptd - ...
797
  	err = shash_ahash_digest(req, desc);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
798
799
800
801
  
  	req->base.complete = rctx->complete;
  
  out:
81760ea6a   Herbert Xu   crypto: cryptd - ...
802
  	cryptd_hash_complete(req, err);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
803
804
805
806
807
808
  }
  
  static int cryptd_hash_digest_enqueue(struct ahash_request *req)
  {
  	return cryptd_hash_enqueue(req, cryptd_hash_digest);
  }
6fba00d17   Herbert Xu   crypto: cryptd - ...
809
810
811
812
813
814
815
816
817
  static int cryptd_hash_export(struct ahash_request *req, void *out)
  {
  	struct cryptd_hash_request_ctx *rctx = ahash_request_ctx(req);
  
  	return crypto_shash_export(&rctx->desc, out);
  }
  
  static int cryptd_hash_import(struct ahash_request *req, const void *in)
  {
0bd222359   Ard Biesheuvel   crypto: cryptd - ...
818
819
820
821
822
823
  	struct crypto_ahash *tfm = crypto_ahash_reqtfm(req);
  	struct cryptd_hash_ctx *ctx = crypto_ahash_ctx(tfm);
  	struct shash_desc *desc = cryptd_shash_desc(req);
  
  	desc->tfm = ctx->child;
  	desc->flags = req->base.flags;
6fba00d17   Herbert Xu   crypto: cryptd - ...
824

0bd222359   Ard Biesheuvel   crypto: cryptd - ...
825
  	return crypto_shash_import(desc, in);
6fba00d17   Herbert Xu   crypto: cryptd - ...
826
  }
9cd899a32   Herbert Xu   crypto: cryptd - ...
827
828
  static int cryptd_create_hash(struct crypto_template *tmpl, struct rtattr **tb,
  			      struct cryptd_queue *queue)
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
829
  {
46309d893   Herbert Xu   crypto: cryptd - ...
830
  	struct hashd_instance_ctx *ctx;
0b535adfb   Herbert Xu   crypto: cryptd - ...
831
  	struct ahash_instance *inst;
46309d893   Herbert Xu   crypto: cryptd - ...
832
  	struct shash_alg *salg;
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
833
  	struct crypto_alg *alg;
466a7b9e3   Stephan Mueller   crypto: cryptd - ...
834
835
  	u32 type = 0;
  	u32 mask = 0;
46309d893   Herbert Xu   crypto: cryptd - ...
836
  	int err;
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
837

466a7b9e3   Stephan Mueller   crypto: cryptd - ...
838
839
840
  	cryptd_check_internal(tb, &type, &mask);
  
  	salg = shash_attr_alg(tb[1], type, mask);
46309d893   Herbert Xu   crypto: cryptd - ...
841
  	if (IS_ERR(salg))
9cd899a32   Herbert Xu   crypto: cryptd - ...
842
  		return PTR_ERR(salg);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
843

46309d893   Herbert Xu   crypto: cryptd - ...
844
  	alg = &salg->base;
0b535adfb   Herbert Xu   crypto: cryptd - ...
845
846
  	inst = cryptd_alloc_instance(alg, ahash_instance_headroom(),
  				     sizeof(*ctx));
05ed8758f   Steffen Klassert   crypto: cryptd - ...
847
  	err = PTR_ERR(inst);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
848
849
  	if (IS_ERR(inst))
  		goto out_put_alg;
0b535adfb   Herbert Xu   crypto: cryptd - ...
850
  	ctx = ahash_instance_ctx(inst);
46309d893   Herbert Xu   crypto: cryptd - ...
851
  	ctx->queue = queue;
0b535adfb   Herbert Xu   crypto: cryptd - ...
852
853
  	err = crypto_init_shash_spawn(&ctx->spawn, salg,
  				      ahash_crypto_instance(inst));
46309d893   Herbert Xu   crypto: cryptd - ...
854
855
  	if (err)
  		goto out_free_inst;
8d906d183   Eric Biggers   crypto: hash - an...
856
857
858
  	inst->alg.halg.base.cra_flags = CRYPTO_ALG_ASYNC |
  		(alg->cra_flags & (CRYPTO_ALG_INTERNAL |
  				   CRYPTO_ALG_OPTIONAL_KEY));
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
859

0b535adfb   Herbert Xu   crypto: cryptd - ...
860
  	inst->alg.halg.digestsize = salg->digestsize;
1a0783402   Wang, Rui Y   crypto: cryptd - ...
861
  	inst->alg.halg.statesize = salg->statesize;
0b535adfb   Herbert Xu   crypto: cryptd - ...
862
  	inst->alg.halg.base.cra_ctxsize = sizeof(struct cryptd_hash_ctx);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
863

0b535adfb   Herbert Xu   crypto: cryptd - ...
864
865
  	inst->alg.halg.base.cra_init = cryptd_hash_init_tfm;
  	inst->alg.halg.base.cra_exit = cryptd_hash_exit_tfm;
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
866

0b535adfb   Herbert Xu   crypto: cryptd - ...
867
868
869
  	inst->alg.init   = cryptd_hash_init_enqueue;
  	inst->alg.update = cryptd_hash_update_enqueue;
  	inst->alg.final  = cryptd_hash_final_enqueue;
6fba00d17   Herbert Xu   crypto: cryptd - ...
870
871
872
  	inst->alg.finup  = cryptd_hash_finup_enqueue;
  	inst->alg.export = cryptd_hash_export;
  	inst->alg.import = cryptd_hash_import;
ed7b0af0c   Eric Biggers   crypto: cryptd - ...
873
874
  	if (crypto_shash_alg_has_setkey(salg))
  		inst->alg.setkey = cryptd_hash_setkey;
0b535adfb   Herbert Xu   crypto: cryptd - ...
875
  	inst->alg.digest = cryptd_hash_digest_enqueue;
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
876

0b535adfb   Herbert Xu   crypto: cryptd - ...
877
  	err = ahash_register_instance(tmpl, inst);
9cd899a32   Herbert Xu   crypto: cryptd - ...
878
879
880
881
882
  	if (err) {
  		crypto_drop_shash(&ctx->spawn);
  out_free_inst:
  		kfree(inst);
  	}
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
883
884
  out_put_alg:
  	crypto_mod_put(alg);
9cd899a32   Herbert Xu   crypto: cryptd - ...
885
  	return err;
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
886
  }
92b9876bd   Herbert Xu   crypto: cryptd - ...
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
  static int cryptd_aead_setkey(struct crypto_aead *parent,
  			      const u8 *key, unsigned int keylen)
  {
  	struct cryptd_aead_ctx *ctx = crypto_aead_ctx(parent);
  	struct crypto_aead *child = ctx->child;
  
  	return crypto_aead_setkey(child, key, keylen);
  }
  
  static int cryptd_aead_setauthsize(struct crypto_aead *parent,
  				   unsigned int authsize)
  {
  	struct cryptd_aead_ctx *ctx = crypto_aead_ctx(parent);
  	struct crypto_aead *child = ctx->child;
  
  	return crypto_aead_setauthsize(child, authsize);
  }
298c926c6   Adrian Hoban   crypto: cryptd - ...
904
905
906
907
908
909
  static void cryptd_aead_crypt(struct aead_request *req,
  			struct crypto_aead *child,
  			int err,
  			int (*crypt)(struct aead_request *req))
  {
  	struct cryptd_aead_request_ctx *rctx;
81760ea6a   Herbert Xu   crypto: cryptd - ...
910
  	struct cryptd_aead_ctx *ctx;
ec9f2006f   Herbert Xu   crypto: cryptd - ...
911
  	crypto_completion_t compl;
81760ea6a   Herbert Xu   crypto: cryptd - ...
912
913
  	struct crypto_aead *tfm;
  	int refcnt;
ec9f2006f   Herbert Xu   crypto: cryptd - ...
914

298c926c6   Adrian Hoban   crypto: cryptd - ...
915
  	rctx = aead_request_ctx(req);
ec9f2006f   Herbert Xu   crypto: cryptd - ...
916
  	compl = rctx->complete;
298c926c6   Adrian Hoban   crypto: cryptd - ...
917

31bd44e76   Herbert Xu   crypto: cryptd - ...
918
  	tfm = crypto_aead_reqtfm(req);
298c926c6   Adrian Hoban   crypto: cryptd - ...
919
920
921
922
  	if (unlikely(err == -EINPROGRESS))
  		goto out;
  	aead_request_set_tfm(req, child);
  	err = crypt( req );
81760ea6a   Herbert Xu   crypto: cryptd - ...
923

298c926c6   Adrian Hoban   crypto: cryptd - ...
924
  out:
81760ea6a   Herbert Xu   crypto: cryptd - ...
925
926
  	ctx = crypto_aead_ctx(tfm);
  	refcnt = atomic_read(&ctx->refcnt);
298c926c6   Adrian Hoban   crypto: cryptd - ...
927
  	local_bh_disable();
ec9f2006f   Herbert Xu   crypto: cryptd - ...
928
  	compl(&req->base, err);
298c926c6   Adrian Hoban   crypto: cryptd - ...
929
  	local_bh_enable();
81760ea6a   Herbert Xu   crypto: cryptd - ...
930
931
932
  
  	if (err != -EINPROGRESS && refcnt && atomic_dec_and_test(&ctx->refcnt))
  		crypto_free_aead(tfm);
298c926c6   Adrian Hoban   crypto: cryptd - ...
933
934
935
936
937
938
939
940
941
  }
  
  static void cryptd_aead_encrypt(struct crypto_async_request *areq, int err)
  {
  	struct cryptd_aead_ctx *ctx = crypto_tfm_ctx(areq->tfm);
  	struct crypto_aead *child = ctx->child;
  	struct aead_request *req;
  
  	req = container_of(areq, struct aead_request, base);
ba3749a71   Herbert Xu   crypto: cryptd - ...
942
  	cryptd_aead_crypt(req, child, err, crypto_aead_alg(child)->encrypt);
298c926c6   Adrian Hoban   crypto: cryptd - ...
943
944
945
946
947
948
949
950
951
  }
  
  static void cryptd_aead_decrypt(struct crypto_async_request *areq, int err)
  {
  	struct cryptd_aead_ctx *ctx = crypto_tfm_ctx(areq->tfm);
  	struct crypto_aead *child = ctx->child;
  	struct aead_request *req;
  
  	req = container_of(areq, struct aead_request, base);
ba3749a71   Herbert Xu   crypto: cryptd - ...
952
  	cryptd_aead_crypt(req, child, err, crypto_aead_alg(child)->decrypt);
298c926c6   Adrian Hoban   crypto: cryptd - ...
953
954
955
  }
  
  static int cryptd_aead_enqueue(struct aead_request *req,
3e3dc25fe   Mark Rustad   crypto: Resolve s...
956
  				    crypto_completion_t compl)
298c926c6   Adrian Hoban   crypto: cryptd - ...
957
958
959
960
961
962
  {
  	struct cryptd_aead_request_ctx *rctx = aead_request_ctx(req);
  	struct crypto_aead *tfm = crypto_aead_reqtfm(req);
  	struct cryptd_queue *queue = cryptd_get_queue(crypto_aead_tfm(tfm));
  
  	rctx->complete = req->base.complete;
3e3dc25fe   Mark Rustad   crypto: Resolve s...
963
  	req->base.complete = compl;
298c926c6   Adrian Hoban   crypto: cryptd - ...
964
965
966
967
968
969
970
971
972
973
974
975
  	return cryptd_enqueue_request(queue, &req->base);
  }
  
  static int cryptd_aead_encrypt_enqueue(struct aead_request *req)
  {
  	return cryptd_aead_enqueue(req, cryptd_aead_encrypt );
  }
  
  static int cryptd_aead_decrypt_enqueue(struct aead_request *req)
  {
  	return cryptd_aead_enqueue(req, cryptd_aead_decrypt );
  }
f614e546f   Herbert Xu   crypto: cryptd - ...
976
  static int cryptd_aead_init_tfm(struct crypto_aead *tfm)
298c926c6   Adrian Hoban   crypto: cryptd - ...
977
  {
f614e546f   Herbert Xu   crypto: cryptd - ...
978
979
  	struct aead_instance *inst = aead_alg_instance(tfm);
  	struct aead_instance_ctx *ictx = aead_instance_ctx(inst);
298c926c6   Adrian Hoban   crypto: cryptd - ...
980
  	struct crypto_aead_spawn *spawn = &ictx->aead_spawn;
f614e546f   Herbert Xu   crypto: cryptd - ...
981
  	struct cryptd_aead_ctx *ctx = crypto_aead_ctx(tfm);
298c926c6   Adrian Hoban   crypto: cryptd - ...
982
983
984
985
986
  	struct crypto_aead *cipher;
  
  	cipher = crypto_spawn_aead(spawn);
  	if (IS_ERR(cipher))
  		return PTR_ERR(cipher);
298c926c6   Adrian Hoban   crypto: cryptd - ...
987
  	ctx->child = cipher;
ec9f2006f   Herbert Xu   crypto: cryptd - ...
988
989
990
  	crypto_aead_set_reqsize(
  		tfm, max((unsigned)sizeof(struct cryptd_aead_request_ctx),
  			 crypto_aead_reqsize(cipher)));
298c926c6   Adrian Hoban   crypto: cryptd - ...
991
992
  	return 0;
  }
f614e546f   Herbert Xu   crypto: cryptd - ...
993
  static void cryptd_aead_exit_tfm(struct crypto_aead *tfm)
298c926c6   Adrian Hoban   crypto: cryptd - ...
994
  {
f614e546f   Herbert Xu   crypto: cryptd - ...
995
  	struct cryptd_aead_ctx *ctx = crypto_aead_ctx(tfm);
298c926c6   Adrian Hoban   crypto: cryptd - ...
996
997
998
999
1000
1001
1002
1003
  	crypto_free_aead(ctx->child);
  }
  
  static int cryptd_create_aead(struct crypto_template *tmpl,
  		              struct rtattr **tb,
  			      struct cryptd_queue *queue)
  {
  	struct aead_instance_ctx *ctx;
f614e546f   Herbert Xu   crypto: cryptd - ...
1004
1005
  	struct aead_instance *inst;
  	struct aead_alg *alg;
9b8c456e0   Herbert Xu   crypto: cryptd - ...
1006
1007
  	const char *name;
  	u32 type = 0;
ec9f2006f   Herbert Xu   crypto: cryptd - ...
1008
  	u32 mask = CRYPTO_ALG_ASYNC;
298c926c6   Adrian Hoban   crypto: cryptd - ...
1009
  	int err;
466a7b9e3   Stephan Mueller   crypto: cryptd - ...
1010
  	cryptd_check_internal(tb, &type, &mask);
9b8c456e0   Herbert Xu   crypto: cryptd - ...
1011
1012
1013
  	name = crypto_attr_alg_name(tb[1]);
  	if (IS_ERR(name))
  		return PTR_ERR(name);
298c926c6   Adrian Hoban   crypto: cryptd - ...
1014

9b8c456e0   Herbert Xu   crypto: cryptd - ...
1015
1016
1017
  	inst = kzalloc(sizeof(*inst) + sizeof(*ctx), GFP_KERNEL);
  	if (!inst)
  		return -ENOMEM;
298c926c6   Adrian Hoban   crypto: cryptd - ...
1018

f614e546f   Herbert Xu   crypto: cryptd - ...
1019
  	ctx = aead_instance_ctx(inst);
298c926c6   Adrian Hoban   crypto: cryptd - ...
1020
  	ctx->queue = queue;
f614e546f   Herbert Xu   crypto: cryptd - ...
1021
  	crypto_set_aead_spawn(&ctx->aead_spawn, aead_crypto_instance(inst));
9b8c456e0   Herbert Xu   crypto: cryptd - ...
1022
  	err = crypto_grab_aead(&ctx->aead_spawn, name, type, mask);
298c926c6   Adrian Hoban   crypto: cryptd - ...
1023
1024
  	if (err)
  		goto out_free_inst;
f614e546f   Herbert Xu   crypto: cryptd - ...
1025
1026
  	alg = crypto_spawn_aead_alg(&ctx->aead_spawn);
  	err = cryptd_init_instance(aead_crypto_instance(inst), &alg->base);
9b8c456e0   Herbert Xu   crypto: cryptd - ...
1027
1028
  	if (err)
  		goto out_drop_aead;
f614e546f   Herbert Xu   crypto: cryptd - ...
1029
  	inst->alg.base.cra_flags = CRYPTO_ALG_ASYNC |
5e4b8c1fc   Herbert Xu   crypto: aead - Re...
1030
  				   (alg->base.cra_flags & CRYPTO_ALG_INTERNAL);
f614e546f   Herbert Xu   crypto: cryptd - ...
1031
  	inst->alg.base.cra_ctxsize = sizeof(struct cryptd_aead_ctx);
298c926c6   Adrian Hoban   crypto: cryptd - ...
1032

f614e546f   Herbert Xu   crypto: cryptd - ...
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
  	inst->alg.ivsize = crypto_aead_alg_ivsize(alg);
  	inst->alg.maxauthsize = crypto_aead_alg_maxauthsize(alg);
  
  	inst->alg.init = cryptd_aead_init_tfm;
  	inst->alg.exit = cryptd_aead_exit_tfm;
  	inst->alg.setkey = cryptd_aead_setkey;
  	inst->alg.setauthsize = cryptd_aead_setauthsize;
  	inst->alg.encrypt = cryptd_aead_encrypt_enqueue;
  	inst->alg.decrypt = cryptd_aead_decrypt_enqueue;
  
  	err = aead_register_instance(tmpl, inst);
298c926c6   Adrian Hoban   crypto: cryptd - ...
1044
  	if (err) {
9b8c456e0   Herbert Xu   crypto: cryptd - ...
1045
1046
  out_drop_aead:
  		crypto_drop_aead(&ctx->aead_spawn);
298c926c6   Adrian Hoban   crypto: cryptd - ...
1047
1048
1049
  out_free_inst:
  		kfree(inst);
  	}
298c926c6   Adrian Hoban   crypto: cryptd - ...
1050
1051
  	return err;
  }
254eff771   Huang Ying   crypto: cryptd - ...
1052
  static struct cryptd_queue queue;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
1053

9cd899a32   Herbert Xu   crypto: cryptd - ...
1054
  static int cryptd_create(struct crypto_template *tmpl, struct rtattr **tb)
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
1055
1056
1057
1058
1059
  {
  	struct crypto_attr_type *algt;
  
  	algt = crypto_get_attr_type(tb);
  	if (IS_ERR(algt))
9cd899a32   Herbert Xu   crypto: cryptd - ...
1060
  		return PTR_ERR(algt);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
1061
1062
1063
  
  	switch (algt->type & algt->mask & CRYPTO_ALG_TYPE_MASK) {
  	case CRYPTO_ALG_TYPE_BLKCIPHER:
4e0958d19   Herbert Xu   crypto: cryptd - ...
1064
1065
1066
1067
1068
  		if ((algt->type & CRYPTO_ALG_TYPE_MASK) ==
  		    CRYPTO_ALG_TYPE_BLKCIPHER)
  			return cryptd_create_blkcipher(tmpl, tb, &queue);
  
  		return cryptd_create_skcipher(tmpl, tb, &queue);
b8a28251c   Loc Ho   [CRYPTO] cryptd: ...
1069
  	case CRYPTO_ALG_TYPE_DIGEST:
9cd899a32   Herbert Xu   crypto: cryptd - ...
1070
  		return cryptd_create_hash(tmpl, tb, &queue);
298c926c6   Adrian Hoban   crypto: cryptd - ...
1071
1072
  	case CRYPTO_ALG_TYPE_AEAD:
  		return cryptd_create_aead(tmpl, tb, &queue);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
1073
  	}
9cd899a32   Herbert Xu   crypto: cryptd - ...
1074
  	return -EINVAL;
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
1075
1076
1077
1078
1079
  }
  
  static void cryptd_free(struct crypto_instance *inst)
  {
  	struct cryptd_instance_ctx *ctx = crypto_instance_ctx(inst);
0b535adfb   Herbert Xu   crypto: cryptd - ...
1080
  	struct hashd_instance_ctx *hctx = crypto_instance_ctx(inst);
298c926c6   Adrian Hoban   crypto: cryptd - ...
1081
  	struct aead_instance_ctx *aead_ctx = crypto_instance_ctx(inst);
0b535adfb   Herbert Xu   crypto: cryptd - ...
1082
1083
1084
1085
1086
1087
  
  	switch (inst->alg.cra_flags & CRYPTO_ALG_TYPE_MASK) {
  	case CRYPTO_ALG_TYPE_AHASH:
  		crypto_drop_shash(&hctx->spawn);
  		kfree(ahash_instance(inst));
  		return;
298c926c6   Adrian Hoban   crypto: cryptd - ...
1088
  	case CRYPTO_ALG_TYPE_AEAD:
f614e546f   Herbert Xu   crypto: cryptd - ...
1089
1090
  		crypto_drop_aead(&aead_ctx->aead_spawn);
  		kfree(aead_instance(inst));
298c926c6   Adrian Hoban   crypto: cryptd - ...
1091
1092
1093
1094
  		return;
  	default:
  		crypto_drop_spawn(&ctx->spawn);
  		kfree(inst);
0b535adfb   Herbert Xu   crypto: cryptd - ...
1095
  	}
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
1096
1097
1098
1099
  }
  
  static struct crypto_template cryptd_tmpl = {
  	.name = "cryptd",
9cd899a32   Herbert Xu   crypto: cryptd - ...
1100
  	.create = cryptd_create,
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
1101
1102
1103
  	.free = cryptd_free,
  	.module = THIS_MODULE,
  };
1cac2cbc7   Huang Ying   crypto: cryptd - ...
1104
1105
1106
1107
  struct cryptd_ablkcipher *cryptd_alloc_ablkcipher(const char *alg_name,
  						  u32 type, u32 mask)
  {
  	char cryptd_alg_name[CRYPTO_MAX_ALG_NAME];
81760ea6a   Herbert Xu   crypto: cryptd - ...
1108
  	struct cryptd_blkcipher_ctx *ctx;
505fd21d6   Huang Ying   crypto: cryptd - ...
1109
  	struct crypto_tfm *tfm;
1cac2cbc7   Huang Ying   crypto: cryptd - ...
1110
1111
1112
1113
  
  	if (snprintf(cryptd_alg_name, CRYPTO_MAX_ALG_NAME,
  		     "cryptd(%s)", alg_name) >= CRYPTO_MAX_ALG_NAME)
  		return ERR_PTR(-EINVAL);
c012a79d0   Alexander Kuleshov   crypto: cryptod -...
1114
  	type = crypto_skcipher_type(type);
505fd21d6   Huang Ying   crypto: cryptd - ...
1115
1116
1117
  	mask &= ~CRYPTO_ALG_TYPE_MASK;
  	mask |= (CRYPTO_ALG_GENIV | CRYPTO_ALG_TYPE_BLKCIPHER_MASK);
  	tfm = crypto_alloc_base(cryptd_alg_name, type, mask);
1cac2cbc7   Huang Ying   crypto: cryptd - ...
1118
1119
  	if (IS_ERR(tfm))
  		return ERR_CAST(tfm);
505fd21d6   Huang Ying   crypto: cryptd - ...
1120
1121
  	if (tfm->__crt_alg->cra_module != THIS_MODULE) {
  		crypto_free_tfm(tfm);
1cac2cbc7   Huang Ying   crypto: cryptd - ...
1122
1123
  		return ERR_PTR(-EINVAL);
  	}
81760ea6a   Herbert Xu   crypto: cryptd - ...
1124
1125
  	ctx = crypto_tfm_ctx(tfm);
  	atomic_set(&ctx->refcnt, 1);
505fd21d6   Huang Ying   crypto: cryptd - ...
1126
  	return __cryptd_ablkcipher_cast(__crypto_ablkcipher_cast(tfm));
1cac2cbc7   Huang Ying   crypto: cryptd - ...
1127
1128
1129
1130
1131
1132
1133
1134
1135
  }
  EXPORT_SYMBOL_GPL(cryptd_alloc_ablkcipher);
  
  struct crypto_blkcipher *cryptd_ablkcipher_child(struct cryptd_ablkcipher *tfm)
  {
  	struct cryptd_blkcipher_ctx *ctx = crypto_ablkcipher_ctx(&tfm->base);
  	return ctx->child;
  }
  EXPORT_SYMBOL_GPL(cryptd_ablkcipher_child);
81760ea6a   Herbert Xu   crypto: cryptd - ...
1136
1137
1138
1139
1140
1141
1142
  bool cryptd_ablkcipher_queued(struct cryptd_ablkcipher *tfm)
  {
  	struct cryptd_blkcipher_ctx *ctx = crypto_ablkcipher_ctx(&tfm->base);
  
  	return atomic_read(&ctx->refcnt) - 1;
  }
  EXPORT_SYMBOL_GPL(cryptd_ablkcipher_queued);
1cac2cbc7   Huang Ying   crypto: cryptd - ...
1143
1144
  void cryptd_free_ablkcipher(struct cryptd_ablkcipher *tfm)
  {
81760ea6a   Herbert Xu   crypto: cryptd - ...
1145
1146
1147
1148
  	struct cryptd_blkcipher_ctx *ctx = crypto_ablkcipher_ctx(&tfm->base);
  
  	if (atomic_dec_and_test(&ctx->refcnt))
  		crypto_free_ablkcipher(&tfm->base);
1cac2cbc7   Huang Ying   crypto: cryptd - ...
1149
1150
  }
  EXPORT_SYMBOL_GPL(cryptd_free_ablkcipher);
4e0958d19   Herbert Xu   crypto: cryptd - ...
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
  struct cryptd_skcipher *cryptd_alloc_skcipher(const char *alg_name,
  					      u32 type, u32 mask)
  {
  	char cryptd_alg_name[CRYPTO_MAX_ALG_NAME];
  	struct cryptd_skcipher_ctx *ctx;
  	struct crypto_skcipher *tfm;
  
  	if (snprintf(cryptd_alg_name, CRYPTO_MAX_ALG_NAME,
  		     "cryptd(%s)", alg_name) >= CRYPTO_MAX_ALG_NAME)
  		return ERR_PTR(-EINVAL);
  
  	tfm = crypto_alloc_skcipher(cryptd_alg_name, type, mask);
  	if (IS_ERR(tfm))
  		return ERR_CAST(tfm);
  
  	if (tfm->base.__crt_alg->cra_module != THIS_MODULE) {
  		crypto_free_skcipher(tfm);
  		return ERR_PTR(-EINVAL);
  	}
  
  	ctx = crypto_skcipher_ctx(tfm);
  	atomic_set(&ctx->refcnt, 1);
  
  	return container_of(tfm, struct cryptd_skcipher, base);
  }
  EXPORT_SYMBOL_GPL(cryptd_alloc_skcipher);
  
  struct crypto_skcipher *cryptd_skcipher_child(struct cryptd_skcipher *tfm)
  {
  	struct cryptd_skcipher_ctx *ctx = crypto_skcipher_ctx(&tfm->base);
  
  	return ctx->child;
  }
  EXPORT_SYMBOL_GPL(cryptd_skcipher_child);
  
  bool cryptd_skcipher_queued(struct cryptd_skcipher *tfm)
  {
  	struct cryptd_skcipher_ctx *ctx = crypto_skcipher_ctx(&tfm->base);
  
  	return atomic_read(&ctx->refcnt) - 1;
  }
  EXPORT_SYMBOL_GPL(cryptd_skcipher_queued);
  
  void cryptd_free_skcipher(struct cryptd_skcipher *tfm)
  {
  	struct cryptd_skcipher_ctx *ctx = crypto_skcipher_ctx(&tfm->base);
  
  	if (atomic_dec_and_test(&ctx->refcnt))
  		crypto_free_skcipher(&tfm->base);
  }
  EXPORT_SYMBOL_GPL(cryptd_free_skcipher);
ace136636   Huang Ying   crypto: cryptd - ...
1202
1203
1204
1205
  struct cryptd_ahash *cryptd_alloc_ahash(const char *alg_name,
  					u32 type, u32 mask)
  {
  	char cryptd_alg_name[CRYPTO_MAX_ALG_NAME];
81760ea6a   Herbert Xu   crypto: cryptd - ...
1206
  	struct cryptd_hash_ctx *ctx;
ace136636   Huang Ying   crypto: cryptd - ...
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
  	struct crypto_ahash *tfm;
  
  	if (snprintf(cryptd_alg_name, CRYPTO_MAX_ALG_NAME,
  		     "cryptd(%s)", alg_name) >= CRYPTO_MAX_ALG_NAME)
  		return ERR_PTR(-EINVAL);
  	tfm = crypto_alloc_ahash(cryptd_alg_name, type, mask);
  	if (IS_ERR(tfm))
  		return ERR_CAST(tfm);
  	if (tfm->base.__crt_alg->cra_module != THIS_MODULE) {
  		crypto_free_ahash(tfm);
  		return ERR_PTR(-EINVAL);
  	}
81760ea6a   Herbert Xu   crypto: cryptd - ...
1219
1220
  	ctx = crypto_ahash_ctx(tfm);
  	atomic_set(&ctx->refcnt, 1);
ace136636   Huang Ying   crypto: cryptd - ...
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
  	return __cryptd_ahash_cast(tfm);
  }
  EXPORT_SYMBOL_GPL(cryptd_alloc_ahash);
  
  struct crypto_shash *cryptd_ahash_child(struct cryptd_ahash *tfm)
  {
  	struct cryptd_hash_ctx *ctx = crypto_ahash_ctx(&tfm->base);
  
  	return ctx->child;
  }
  EXPORT_SYMBOL_GPL(cryptd_ahash_child);
0e1227d35   Huang Ying   crypto: ghash - A...
1232
1233
1234
1235
1236
1237
  struct shash_desc *cryptd_shash_desc(struct ahash_request *req)
  {
  	struct cryptd_hash_request_ctx *rctx = ahash_request_ctx(req);
  	return &rctx->desc;
  }
  EXPORT_SYMBOL_GPL(cryptd_shash_desc);
81760ea6a   Herbert Xu   crypto: cryptd - ...
1238
1239
1240
1241
1242
1243
1244
  bool cryptd_ahash_queued(struct cryptd_ahash *tfm)
  {
  	struct cryptd_hash_ctx *ctx = crypto_ahash_ctx(&tfm->base);
  
  	return atomic_read(&ctx->refcnt) - 1;
  }
  EXPORT_SYMBOL_GPL(cryptd_ahash_queued);
ace136636   Huang Ying   crypto: cryptd - ...
1245
1246
  void cryptd_free_ahash(struct cryptd_ahash *tfm)
  {
81760ea6a   Herbert Xu   crypto: cryptd - ...
1247
1248
1249
1250
  	struct cryptd_hash_ctx *ctx = crypto_ahash_ctx(&tfm->base);
  
  	if (atomic_dec_and_test(&ctx->refcnt))
  		crypto_free_ahash(&tfm->base);
ace136636   Huang Ying   crypto: cryptd - ...
1251
1252
  }
  EXPORT_SYMBOL_GPL(cryptd_free_ahash);
298c926c6   Adrian Hoban   crypto: cryptd - ...
1253
1254
1255
1256
  struct cryptd_aead *cryptd_alloc_aead(const char *alg_name,
  						  u32 type, u32 mask)
  {
  	char cryptd_alg_name[CRYPTO_MAX_ALG_NAME];
81760ea6a   Herbert Xu   crypto: cryptd - ...
1257
  	struct cryptd_aead_ctx *ctx;
298c926c6   Adrian Hoban   crypto: cryptd - ...
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
  	struct crypto_aead *tfm;
  
  	if (snprintf(cryptd_alg_name, CRYPTO_MAX_ALG_NAME,
  		     "cryptd(%s)", alg_name) >= CRYPTO_MAX_ALG_NAME)
  		return ERR_PTR(-EINVAL);
  	tfm = crypto_alloc_aead(cryptd_alg_name, type, mask);
  	if (IS_ERR(tfm))
  		return ERR_CAST(tfm);
  	if (tfm->base.__crt_alg->cra_module != THIS_MODULE) {
  		crypto_free_aead(tfm);
  		return ERR_PTR(-EINVAL);
  	}
81760ea6a   Herbert Xu   crypto: cryptd - ...
1270
1271
1272
  
  	ctx = crypto_aead_ctx(tfm);
  	atomic_set(&ctx->refcnt, 1);
298c926c6   Adrian Hoban   crypto: cryptd - ...
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
  	return __cryptd_aead_cast(tfm);
  }
  EXPORT_SYMBOL_GPL(cryptd_alloc_aead);
  
  struct crypto_aead *cryptd_aead_child(struct cryptd_aead *tfm)
  {
  	struct cryptd_aead_ctx *ctx;
  	ctx = crypto_aead_ctx(&tfm->base);
  	return ctx->child;
  }
  EXPORT_SYMBOL_GPL(cryptd_aead_child);
81760ea6a   Herbert Xu   crypto: cryptd - ...
1284
1285
1286
1287
1288
1289
1290
  bool cryptd_aead_queued(struct cryptd_aead *tfm)
  {
  	struct cryptd_aead_ctx *ctx = crypto_aead_ctx(&tfm->base);
  
  	return atomic_read(&ctx->refcnt) - 1;
  }
  EXPORT_SYMBOL_GPL(cryptd_aead_queued);
298c926c6   Adrian Hoban   crypto: cryptd - ...
1291
1292
  void cryptd_free_aead(struct cryptd_aead *tfm)
  {
81760ea6a   Herbert Xu   crypto: cryptd - ...
1293
1294
1295
1296
  	struct cryptd_aead_ctx *ctx = crypto_aead_ctx(&tfm->base);
  
  	if (atomic_dec_and_test(&ctx->refcnt))
  		crypto_free_aead(&tfm->base);
298c926c6   Adrian Hoban   crypto: cryptd - ...
1297
1298
  }
  EXPORT_SYMBOL_GPL(cryptd_free_aead);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
1299
1300
1301
  static int __init cryptd_init(void)
  {
  	int err;
254eff771   Huang Ying   crypto: cryptd - ...
1302
  	err = cryptd_init_queue(&queue, CRYPTD_MAX_CPU_QLEN);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
1303
1304
1305
1306
1307
  	if (err)
  		return err;
  
  	err = crypto_register_template(&cryptd_tmpl);
  	if (err)
254eff771   Huang Ying   crypto: cryptd - ...
1308
  		cryptd_fini_queue(&queue);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
1309
1310
1311
1312
1313
1314
  
  	return err;
  }
  
  static void __exit cryptd_exit(void)
  {
254eff771   Huang Ying   crypto: cryptd - ...
1315
  	cryptd_fini_queue(&queue);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
1316
1317
  	crypto_unregister_template(&cryptd_tmpl);
  }
b2bac6acf   Herbert Xu   crypto: cryptd - ...
1318
  subsys_initcall(cryptd_init);
124b53d02   Herbert Xu   [CRYPTO] cryptd: ...
1319
1320
1321
1322
  module_exit(cryptd_exit);
  
  MODULE_LICENSE("GPL");
  MODULE_DESCRIPTION("Software async crypto daemon");
4943ba16b   Kees Cook   crypto: include c...
1323
  MODULE_ALIAS_CRYPTO("cryptd");