Merge master.kernel.org:/pub/scm/linux/kernel/git/jejb/scsi-misc-2.6
[deliverable/linux.git] / include / linux / crypto.h
1 /*
2 * Scatterlist Cryptographic API.
3 *
4 * Copyright (c) 2002 James Morris <jmorris@intercode.com.au>
5 * Copyright (c) 2002 David S. Miller (davem@redhat.com)
6 * Copyright (c) 2005 Herbert Xu <herbert@gondor.apana.org.au>
7 *
8 * Portions derived from Cryptoapi, by Alexander Kjeldaas <astor@fast.no>
9 * and Nettle, by Niels Möller.
10 *
11 * This program is free software; you can redistribute it and/or modify it
12 * under the terms of the GNU General Public License as published by the Free
13 * Software Foundation; either version 2 of the License, or (at your option)
14 * any later version.
15 *
16 */
17 #ifndef _LINUX_CRYPTO_H
18 #define _LINUX_CRYPTO_H
19
20 #include <asm/atomic.h>
21 #include <linux/module.h>
22 #include <linux/kernel.h>
23 #include <linux/list.h>
24 #include <linux/slab.h>
25 #include <linux/string.h>
26 #include <linux/uaccess.h>
27
28 /*
29 * Algorithm masks and types.
30 */
31 #define CRYPTO_ALG_TYPE_MASK 0x0000000f
32 #define CRYPTO_ALG_TYPE_CIPHER 0x00000001
33 #define CRYPTO_ALG_TYPE_DIGEST 0x00000002
34 #define CRYPTO_ALG_TYPE_HASH 0x00000003
35 #define CRYPTO_ALG_TYPE_BLKCIPHER 0x00000004
36 #define CRYPTO_ALG_TYPE_COMPRESS 0x00000005
37 #define CRYPTO_ALG_TYPE_AEAD 0x00000006
38
39 #define CRYPTO_ALG_TYPE_HASH_MASK 0x0000000e
40
41 #define CRYPTO_ALG_LARVAL 0x00000010
42 #define CRYPTO_ALG_DEAD 0x00000020
43 #define CRYPTO_ALG_DYING 0x00000040
44 #define CRYPTO_ALG_ASYNC 0x00000080
45
46 /*
47 * Set this bit if and only if the algorithm requires another algorithm of
48 * the same type to handle corner cases.
49 */
50 #define CRYPTO_ALG_NEED_FALLBACK 0x00000100
51
52 /*
53 * Transform masks and values (for crt_flags).
54 */
55 #define CRYPTO_TFM_REQ_MASK 0x000fff00
56 #define CRYPTO_TFM_RES_MASK 0xfff00000
57
58 #define CRYPTO_TFM_REQ_WEAK_KEY 0x00000100
59 #define CRYPTO_TFM_REQ_MAY_SLEEP 0x00000200
60 #define CRYPTO_TFM_REQ_MAY_BACKLOG 0x00000400
61 #define CRYPTO_TFM_RES_WEAK_KEY 0x00100000
62 #define CRYPTO_TFM_RES_BAD_KEY_LEN 0x00200000
63 #define CRYPTO_TFM_RES_BAD_KEY_SCHED 0x00400000
64 #define CRYPTO_TFM_RES_BAD_BLOCK_LEN 0x00800000
65 #define CRYPTO_TFM_RES_BAD_FLAGS 0x01000000
66
67 /*
68 * Miscellaneous stuff.
69 */
70 #define CRYPTO_MAX_ALG_NAME 64
71
72 /*
73 * The macro CRYPTO_MINALIGN_ATTR (along with the void * type in the actual
74 * declaration) is used to ensure that the crypto_tfm context structure is
75 * aligned correctly for the given architecture so that there are no alignment
76 * faults for C data types. In particular, this is required on platforms such
77 * as arm where pointers are 32-bit aligned but there are data types such as
78 * u64 which require 64-bit alignment.
79 */
80 #if defined(ARCH_KMALLOC_MINALIGN)
81 #define CRYPTO_MINALIGN ARCH_KMALLOC_MINALIGN
82 #elif defined(ARCH_SLAB_MINALIGN)
83 #define CRYPTO_MINALIGN ARCH_SLAB_MINALIGN
84 #endif
85
86 #ifdef CRYPTO_MINALIGN
87 #define CRYPTO_MINALIGN_ATTR __attribute__ ((__aligned__(CRYPTO_MINALIGN)))
88 #else
89 #define CRYPTO_MINALIGN_ATTR
90 #endif
91
92 struct scatterlist;
93 struct crypto_ablkcipher;
94 struct crypto_async_request;
95 struct crypto_aead;
96 struct crypto_blkcipher;
97 struct crypto_hash;
98 struct crypto_tfm;
99 struct crypto_type;
100
101 typedef void (*crypto_completion_t)(struct crypto_async_request *req, int err);
102
103 struct crypto_async_request {
104 struct list_head list;
105 crypto_completion_t complete;
106 void *data;
107 struct crypto_tfm *tfm;
108
109 u32 flags;
110 };
111
112 struct ablkcipher_request {
113 struct crypto_async_request base;
114
115 unsigned int nbytes;
116
117 void *info;
118
119 struct scatterlist *src;
120 struct scatterlist *dst;
121
122 void *__ctx[] CRYPTO_MINALIGN_ATTR;
123 };
124
125 /**
126 * struct aead_request - AEAD request
127 * @base: Common attributes for async crypto requests
128 * @assoclen: Length in bytes of associated data for authentication
129 * @cryptlen: Length of data to be encrypted or decrypted
130 * @iv: Initialisation vector
131 * @assoc: Associated data
132 * @src: Source data
133 * @dst: Destination data
134 * @__ctx: Start of private context data
135 */
136 struct aead_request {
137 struct crypto_async_request base;
138
139 unsigned int assoclen;
140 unsigned int cryptlen;
141
142 u8 *iv;
143
144 struct scatterlist *assoc;
145 struct scatterlist *src;
146 struct scatterlist *dst;
147
148 void *__ctx[] CRYPTO_MINALIGN_ATTR;
149 };
150
151 struct blkcipher_desc {
152 struct crypto_blkcipher *tfm;
153 void *info;
154 u32 flags;
155 };
156
157 struct cipher_desc {
158 struct crypto_tfm *tfm;
159 void (*crfn)(struct crypto_tfm *tfm, u8 *dst, const u8 *src);
160 unsigned int (*prfn)(const struct cipher_desc *desc, u8 *dst,
161 const u8 *src, unsigned int nbytes);
162 void *info;
163 };
164
165 struct hash_desc {
166 struct crypto_hash *tfm;
167 u32 flags;
168 };
169
170 /*
171 * Algorithms: modular crypto algorithm implementations, managed
172 * via crypto_register_alg() and crypto_unregister_alg().
173 */
174 struct ablkcipher_alg {
175 int (*setkey)(struct crypto_ablkcipher *tfm, const u8 *key,
176 unsigned int keylen);
177 int (*encrypt)(struct ablkcipher_request *req);
178 int (*decrypt)(struct ablkcipher_request *req);
179
180 unsigned int min_keysize;
181 unsigned int max_keysize;
182 unsigned int ivsize;
183 };
184
185 struct aead_alg {
186 int (*setkey)(struct crypto_aead *tfm, const u8 *key,
187 unsigned int keylen);
188 int (*encrypt)(struct aead_request *req);
189 int (*decrypt)(struct aead_request *req);
190
191 unsigned int ivsize;
192 unsigned int authsize;
193 };
194
195 struct blkcipher_alg {
196 int (*setkey)(struct crypto_tfm *tfm, const u8 *key,
197 unsigned int keylen);
198 int (*encrypt)(struct blkcipher_desc *desc,
199 struct scatterlist *dst, struct scatterlist *src,
200 unsigned int nbytes);
201 int (*decrypt)(struct blkcipher_desc *desc,
202 struct scatterlist *dst, struct scatterlist *src,
203 unsigned int nbytes);
204
205 unsigned int min_keysize;
206 unsigned int max_keysize;
207 unsigned int ivsize;
208 };
209
210 struct cipher_alg {
211 unsigned int cia_min_keysize;
212 unsigned int cia_max_keysize;
213 int (*cia_setkey)(struct crypto_tfm *tfm, const u8 *key,
214 unsigned int keylen);
215 void (*cia_encrypt)(struct crypto_tfm *tfm, u8 *dst, const u8 *src);
216 void (*cia_decrypt)(struct crypto_tfm *tfm, u8 *dst, const u8 *src);
217 };
218
219 struct digest_alg {
220 unsigned int dia_digestsize;
221 void (*dia_init)(struct crypto_tfm *tfm);
222 void (*dia_update)(struct crypto_tfm *tfm, const u8 *data,
223 unsigned int len);
224 void (*dia_final)(struct crypto_tfm *tfm, u8 *out);
225 int (*dia_setkey)(struct crypto_tfm *tfm, const u8 *key,
226 unsigned int keylen);
227 };
228
229 struct hash_alg {
230 int (*init)(struct hash_desc *desc);
231 int (*update)(struct hash_desc *desc, struct scatterlist *sg,
232 unsigned int nbytes);
233 int (*final)(struct hash_desc *desc, u8 *out);
234 int (*digest)(struct hash_desc *desc, struct scatterlist *sg,
235 unsigned int nbytes, u8 *out);
236 int (*setkey)(struct crypto_hash *tfm, const u8 *key,
237 unsigned int keylen);
238
239 unsigned int digestsize;
240 };
241
242 struct compress_alg {
243 int (*coa_compress)(struct crypto_tfm *tfm, const u8 *src,
244 unsigned int slen, u8 *dst, unsigned int *dlen);
245 int (*coa_decompress)(struct crypto_tfm *tfm, const u8 *src,
246 unsigned int slen, u8 *dst, unsigned int *dlen);
247 };
248
249 #define cra_ablkcipher cra_u.ablkcipher
250 #define cra_aead cra_u.aead
251 #define cra_blkcipher cra_u.blkcipher
252 #define cra_cipher cra_u.cipher
253 #define cra_digest cra_u.digest
254 #define cra_hash cra_u.hash
255 #define cra_compress cra_u.compress
256
257 struct crypto_alg {
258 struct list_head cra_list;
259 struct list_head cra_users;
260
261 u32 cra_flags;
262 unsigned int cra_blocksize;
263 unsigned int cra_ctxsize;
264 unsigned int cra_alignmask;
265
266 int cra_priority;
267 atomic_t cra_refcnt;
268
269 char cra_name[CRYPTO_MAX_ALG_NAME];
270 char cra_driver_name[CRYPTO_MAX_ALG_NAME];
271
272 const struct crypto_type *cra_type;
273
274 union {
275 struct ablkcipher_alg ablkcipher;
276 struct aead_alg aead;
277 struct blkcipher_alg blkcipher;
278 struct cipher_alg cipher;
279 struct digest_alg digest;
280 struct hash_alg hash;
281 struct compress_alg compress;
282 } cra_u;
283
284 int (*cra_init)(struct crypto_tfm *tfm);
285 void (*cra_exit)(struct crypto_tfm *tfm);
286 void (*cra_destroy)(struct crypto_alg *alg);
287
288 struct module *cra_module;
289 };
290
291 /*
292 * Algorithm registration interface.
293 */
294 int crypto_register_alg(struct crypto_alg *alg);
295 int crypto_unregister_alg(struct crypto_alg *alg);
296
297 /*
298 * Algorithm query interface.
299 */
300 #ifdef CONFIG_CRYPTO
301 int crypto_has_alg(const char *name, u32 type, u32 mask);
302 #else
303 static inline int crypto_has_alg(const char *name, u32 type, u32 mask)
304 {
305 return 0;
306 }
307 #endif
308
309 /*
310 * Transforms: user-instantiated objects which encapsulate algorithms
311 * and core processing logic. Managed via crypto_alloc_*() and
312 * crypto_free_*(), as well as the various helpers below.
313 */
314
315 struct ablkcipher_tfm {
316 int (*setkey)(struct crypto_ablkcipher *tfm, const u8 *key,
317 unsigned int keylen);
318 int (*encrypt)(struct ablkcipher_request *req);
319 int (*decrypt)(struct ablkcipher_request *req);
320 unsigned int ivsize;
321 unsigned int reqsize;
322 };
323
324 struct aead_tfm {
325 int (*setkey)(struct crypto_aead *tfm, const u8 *key,
326 unsigned int keylen);
327 int (*encrypt)(struct aead_request *req);
328 int (*decrypt)(struct aead_request *req);
329 unsigned int ivsize;
330 unsigned int authsize;
331 unsigned int reqsize;
332 };
333
334 struct blkcipher_tfm {
335 void *iv;
336 int (*setkey)(struct crypto_tfm *tfm, const u8 *key,
337 unsigned int keylen);
338 int (*encrypt)(struct blkcipher_desc *desc, struct scatterlist *dst,
339 struct scatterlist *src, unsigned int nbytes);
340 int (*decrypt)(struct blkcipher_desc *desc, struct scatterlist *dst,
341 struct scatterlist *src, unsigned int nbytes);
342 };
343
344 struct cipher_tfm {
345 int (*cit_setkey)(struct crypto_tfm *tfm,
346 const u8 *key, unsigned int keylen);
347 void (*cit_encrypt_one)(struct crypto_tfm *tfm, u8 *dst, const u8 *src);
348 void (*cit_decrypt_one)(struct crypto_tfm *tfm, u8 *dst, const u8 *src);
349 };
350
351 struct hash_tfm {
352 int (*init)(struct hash_desc *desc);
353 int (*update)(struct hash_desc *desc,
354 struct scatterlist *sg, unsigned int nsg);
355 int (*final)(struct hash_desc *desc, u8 *out);
356 int (*digest)(struct hash_desc *desc, struct scatterlist *sg,
357 unsigned int nsg, u8 *out);
358 int (*setkey)(struct crypto_hash *tfm, const u8 *key,
359 unsigned int keylen);
360 unsigned int digestsize;
361 };
362
363 struct compress_tfm {
364 int (*cot_compress)(struct crypto_tfm *tfm,
365 const u8 *src, unsigned int slen,
366 u8 *dst, unsigned int *dlen);
367 int (*cot_decompress)(struct crypto_tfm *tfm,
368 const u8 *src, unsigned int slen,
369 u8 *dst, unsigned int *dlen);
370 };
371
372 #define crt_ablkcipher crt_u.ablkcipher
373 #define crt_aead crt_u.aead
374 #define crt_blkcipher crt_u.blkcipher
375 #define crt_cipher crt_u.cipher
376 #define crt_hash crt_u.hash
377 #define crt_compress crt_u.compress
378
379 struct crypto_tfm {
380
381 u32 crt_flags;
382
383 union {
384 struct ablkcipher_tfm ablkcipher;
385 struct aead_tfm aead;
386 struct blkcipher_tfm blkcipher;
387 struct cipher_tfm cipher;
388 struct hash_tfm hash;
389 struct compress_tfm compress;
390 } crt_u;
391
392 struct crypto_alg *__crt_alg;
393
394 void *__crt_ctx[] CRYPTO_MINALIGN_ATTR;
395 };
396
397 struct crypto_ablkcipher {
398 struct crypto_tfm base;
399 };
400
401 struct crypto_aead {
402 struct crypto_tfm base;
403 };
404
405 struct crypto_blkcipher {
406 struct crypto_tfm base;
407 };
408
409 struct crypto_cipher {
410 struct crypto_tfm base;
411 };
412
413 struct crypto_comp {
414 struct crypto_tfm base;
415 };
416
417 struct crypto_hash {
418 struct crypto_tfm base;
419 };
420
421 enum {
422 CRYPTOA_UNSPEC,
423 CRYPTOA_ALG,
424 CRYPTOA_TYPE,
425 CRYPTOA_U32,
426 __CRYPTOA_MAX,
427 };
428
429 #define CRYPTOA_MAX (__CRYPTOA_MAX - 1)
430
431 /* Maximum number of (rtattr) parameters for each template. */
432 #define CRYPTO_MAX_ATTRS 32
433
434 struct crypto_attr_alg {
435 char name[CRYPTO_MAX_ALG_NAME];
436 };
437
438 struct crypto_attr_type {
439 u32 type;
440 u32 mask;
441 };
442
443 struct crypto_attr_u32 {
444 u32 num;
445 };
446
447 /*
448 * Transform user interface.
449 */
450
451 struct crypto_tfm *crypto_alloc_tfm(const char *alg_name, u32 tfm_flags);
452 struct crypto_tfm *crypto_alloc_base(const char *alg_name, u32 type, u32 mask);
453 void crypto_free_tfm(struct crypto_tfm *tfm);
454
455 /*
456 * Transform helpers which query the underlying algorithm.
457 */
458 static inline const char *crypto_tfm_alg_name(struct crypto_tfm *tfm)
459 {
460 return tfm->__crt_alg->cra_name;
461 }
462
463 static inline const char *crypto_tfm_alg_driver_name(struct crypto_tfm *tfm)
464 {
465 return tfm->__crt_alg->cra_driver_name;
466 }
467
468 static inline int crypto_tfm_alg_priority(struct crypto_tfm *tfm)
469 {
470 return tfm->__crt_alg->cra_priority;
471 }
472
473 static inline const char *crypto_tfm_alg_modname(struct crypto_tfm *tfm)
474 {
475 return module_name(tfm->__crt_alg->cra_module);
476 }
477
478 static inline u32 crypto_tfm_alg_type(struct crypto_tfm *tfm)
479 {
480 return tfm->__crt_alg->cra_flags & CRYPTO_ALG_TYPE_MASK;
481 }
482
483 static inline unsigned int crypto_tfm_alg_blocksize(struct crypto_tfm *tfm)
484 {
485 return tfm->__crt_alg->cra_blocksize;
486 }
487
488 static inline unsigned int crypto_tfm_alg_alignmask(struct crypto_tfm *tfm)
489 {
490 return tfm->__crt_alg->cra_alignmask;
491 }
492
493 static inline u32 crypto_tfm_get_flags(struct crypto_tfm *tfm)
494 {
495 return tfm->crt_flags;
496 }
497
498 static inline void crypto_tfm_set_flags(struct crypto_tfm *tfm, u32 flags)
499 {
500 tfm->crt_flags |= flags;
501 }
502
503 static inline void crypto_tfm_clear_flags(struct crypto_tfm *tfm, u32 flags)
504 {
505 tfm->crt_flags &= ~flags;
506 }
507
508 static inline void *crypto_tfm_ctx(struct crypto_tfm *tfm)
509 {
510 return tfm->__crt_ctx;
511 }
512
513 static inline unsigned int crypto_tfm_ctx_alignment(void)
514 {
515 struct crypto_tfm *tfm;
516 return __alignof__(tfm->__crt_ctx);
517 }
518
519 /*
520 * API wrappers.
521 */
522 static inline struct crypto_ablkcipher *__crypto_ablkcipher_cast(
523 struct crypto_tfm *tfm)
524 {
525 return (struct crypto_ablkcipher *)tfm;
526 }
527
528 static inline struct crypto_ablkcipher *crypto_alloc_ablkcipher(
529 const char *alg_name, u32 type, u32 mask)
530 {
531 type &= ~CRYPTO_ALG_TYPE_MASK;
532 type |= CRYPTO_ALG_TYPE_BLKCIPHER;
533 mask |= CRYPTO_ALG_TYPE_MASK;
534
535 return __crypto_ablkcipher_cast(
536 crypto_alloc_base(alg_name, type, mask));
537 }
538
539 static inline struct crypto_tfm *crypto_ablkcipher_tfm(
540 struct crypto_ablkcipher *tfm)
541 {
542 return &tfm->base;
543 }
544
545 static inline void crypto_free_ablkcipher(struct crypto_ablkcipher *tfm)
546 {
547 crypto_free_tfm(crypto_ablkcipher_tfm(tfm));
548 }
549
550 static inline int crypto_has_ablkcipher(const char *alg_name, u32 type,
551 u32 mask)
552 {
553 type &= ~CRYPTO_ALG_TYPE_MASK;
554 type |= CRYPTO_ALG_TYPE_BLKCIPHER;
555 mask |= CRYPTO_ALG_TYPE_MASK;
556
557 return crypto_has_alg(alg_name, type, mask);
558 }
559
560 static inline struct ablkcipher_tfm *crypto_ablkcipher_crt(
561 struct crypto_ablkcipher *tfm)
562 {
563 return &crypto_ablkcipher_tfm(tfm)->crt_ablkcipher;
564 }
565
566 static inline unsigned int crypto_ablkcipher_ivsize(
567 struct crypto_ablkcipher *tfm)
568 {
569 return crypto_ablkcipher_crt(tfm)->ivsize;
570 }
571
572 static inline unsigned int crypto_ablkcipher_blocksize(
573 struct crypto_ablkcipher *tfm)
574 {
575 return crypto_tfm_alg_blocksize(crypto_ablkcipher_tfm(tfm));
576 }
577
578 static inline unsigned int crypto_ablkcipher_alignmask(
579 struct crypto_ablkcipher *tfm)
580 {
581 return crypto_tfm_alg_alignmask(crypto_ablkcipher_tfm(tfm));
582 }
583
584 static inline u32 crypto_ablkcipher_get_flags(struct crypto_ablkcipher *tfm)
585 {
586 return crypto_tfm_get_flags(crypto_ablkcipher_tfm(tfm));
587 }
588
589 static inline void crypto_ablkcipher_set_flags(struct crypto_ablkcipher *tfm,
590 u32 flags)
591 {
592 crypto_tfm_set_flags(crypto_ablkcipher_tfm(tfm), flags);
593 }
594
595 static inline void crypto_ablkcipher_clear_flags(struct crypto_ablkcipher *tfm,
596 u32 flags)
597 {
598 crypto_tfm_clear_flags(crypto_ablkcipher_tfm(tfm), flags);
599 }
600
601 static inline int crypto_ablkcipher_setkey(struct crypto_ablkcipher *tfm,
602 const u8 *key, unsigned int keylen)
603 {
604 return crypto_ablkcipher_crt(tfm)->setkey(tfm, key, keylen);
605 }
606
607 static inline struct crypto_ablkcipher *crypto_ablkcipher_reqtfm(
608 struct ablkcipher_request *req)
609 {
610 return __crypto_ablkcipher_cast(req->base.tfm);
611 }
612
613 static inline int crypto_ablkcipher_encrypt(struct ablkcipher_request *req)
614 {
615 struct ablkcipher_tfm *crt =
616 crypto_ablkcipher_crt(crypto_ablkcipher_reqtfm(req));
617 return crt->encrypt(req);
618 }
619
620 static inline int crypto_ablkcipher_decrypt(struct ablkcipher_request *req)
621 {
622 struct ablkcipher_tfm *crt =
623 crypto_ablkcipher_crt(crypto_ablkcipher_reqtfm(req));
624 return crt->decrypt(req);
625 }
626
627 static inline unsigned int crypto_ablkcipher_reqsize(
628 struct crypto_ablkcipher *tfm)
629 {
630 return crypto_ablkcipher_crt(tfm)->reqsize;
631 }
632
633 static inline void ablkcipher_request_set_tfm(
634 struct ablkcipher_request *req, struct crypto_ablkcipher *tfm)
635 {
636 req->base.tfm = crypto_ablkcipher_tfm(tfm);
637 }
638
639 static inline struct ablkcipher_request *ablkcipher_request_cast(
640 struct crypto_async_request *req)
641 {
642 return container_of(req, struct ablkcipher_request, base);
643 }
644
645 static inline struct ablkcipher_request *ablkcipher_request_alloc(
646 struct crypto_ablkcipher *tfm, gfp_t gfp)
647 {
648 struct ablkcipher_request *req;
649
650 req = kmalloc(sizeof(struct ablkcipher_request) +
651 crypto_ablkcipher_reqsize(tfm), gfp);
652
653 if (likely(req))
654 ablkcipher_request_set_tfm(req, tfm);
655
656 return req;
657 }
658
659 static inline void ablkcipher_request_free(struct ablkcipher_request *req)
660 {
661 kfree(req);
662 }
663
664 static inline void ablkcipher_request_set_callback(
665 struct ablkcipher_request *req,
666 u32 flags, crypto_completion_t complete, void *data)
667 {
668 req->base.complete = complete;
669 req->base.data = data;
670 req->base.flags = flags;
671 }
672
673 static inline void ablkcipher_request_set_crypt(
674 struct ablkcipher_request *req,
675 struct scatterlist *src, struct scatterlist *dst,
676 unsigned int nbytes, void *iv)
677 {
678 req->src = src;
679 req->dst = dst;
680 req->nbytes = nbytes;
681 req->info = iv;
682 }
683
684 static inline struct crypto_aead *__crypto_aead_cast(struct crypto_tfm *tfm)
685 {
686 return (struct crypto_aead *)tfm;
687 }
688
689 static inline struct crypto_aead *crypto_alloc_aead(const char *alg_name,
690 u32 type, u32 mask)
691 {
692 type &= ~CRYPTO_ALG_TYPE_MASK;
693 type |= CRYPTO_ALG_TYPE_AEAD;
694 mask |= CRYPTO_ALG_TYPE_MASK;
695
696 return __crypto_aead_cast(crypto_alloc_base(alg_name, type, mask));
697 }
698
699 static inline struct crypto_tfm *crypto_aead_tfm(struct crypto_aead *tfm)
700 {
701 return &tfm->base;
702 }
703
704 static inline void crypto_free_aead(struct crypto_aead *tfm)
705 {
706 crypto_free_tfm(crypto_aead_tfm(tfm));
707 }
708
709 static inline struct aead_tfm *crypto_aead_crt(struct crypto_aead *tfm)
710 {
711 return &crypto_aead_tfm(tfm)->crt_aead;
712 }
713
714 static inline unsigned int crypto_aead_ivsize(struct crypto_aead *tfm)
715 {
716 return crypto_aead_crt(tfm)->ivsize;
717 }
718
719 static inline unsigned int crypto_aead_authsize(struct crypto_aead *tfm)
720 {
721 return crypto_aead_crt(tfm)->authsize;
722 }
723
724 static inline unsigned int crypto_aead_blocksize(struct crypto_aead *tfm)
725 {
726 return crypto_tfm_alg_blocksize(crypto_aead_tfm(tfm));
727 }
728
729 static inline unsigned int crypto_aead_alignmask(struct crypto_aead *tfm)
730 {
731 return crypto_tfm_alg_alignmask(crypto_aead_tfm(tfm));
732 }
733
734 static inline u32 crypto_aead_get_flags(struct crypto_aead *tfm)
735 {
736 return crypto_tfm_get_flags(crypto_aead_tfm(tfm));
737 }
738
739 static inline void crypto_aead_set_flags(struct crypto_aead *tfm, u32 flags)
740 {
741 crypto_tfm_set_flags(crypto_aead_tfm(tfm), flags);
742 }
743
744 static inline void crypto_aead_clear_flags(struct crypto_aead *tfm, u32 flags)
745 {
746 crypto_tfm_clear_flags(crypto_aead_tfm(tfm), flags);
747 }
748
749 static inline int crypto_aead_setkey(struct crypto_aead *tfm, const u8 *key,
750 unsigned int keylen)
751 {
752 return crypto_aead_crt(tfm)->setkey(tfm, key, keylen);
753 }
754
755 static inline struct crypto_aead *crypto_aead_reqtfm(struct aead_request *req)
756 {
757 return __crypto_aead_cast(req->base.tfm);
758 }
759
760 static inline int crypto_aead_encrypt(struct aead_request *req)
761 {
762 return crypto_aead_crt(crypto_aead_reqtfm(req))->encrypt(req);
763 }
764
765 static inline int crypto_aead_decrypt(struct aead_request *req)
766 {
767 return crypto_aead_crt(crypto_aead_reqtfm(req))->decrypt(req);
768 }
769
770 static inline unsigned int crypto_aead_reqsize(struct crypto_aead *tfm)
771 {
772 return crypto_aead_crt(tfm)->reqsize;
773 }
774
775 static inline void aead_request_set_tfm(struct aead_request *req,
776 struct crypto_aead *tfm)
777 {
778 req->base.tfm = crypto_aead_tfm(tfm);
779 }
780
781 static inline struct aead_request *aead_request_alloc(struct crypto_aead *tfm,
782 gfp_t gfp)
783 {
784 struct aead_request *req;
785
786 req = kmalloc(sizeof(*req) + crypto_aead_reqsize(tfm), gfp);
787
788 if (likely(req))
789 aead_request_set_tfm(req, tfm);
790
791 return req;
792 }
793
794 static inline void aead_request_free(struct aead_request *req)
795 {
796 kfree(req);
797 }
798
799 static inline void aead_request_set_callback(struct aead_request *req,
800 u32 flags,
801 crypto_completion_t complete,
802 void *data)
803 {
804 req->base.complete = complete;
805 req->base.data = data;
806 req->base.flags = flags;
807 }
808
809 static inline void aead_request_set_crypt(struct aead_request *req,
810 struct scatterlist *src,
811 struct scatterlist *dst,
812 unsigned int cryptlen, u8 *iv)
813 {
814 req->src = src;
815 req->dst = dst;
816 req->cryptlen = cryptlen;
817 req->iv = iv;
818 }
819
820 static inline void aead_request_set_assoc(struct aead_request *req,
821 struct scatterlist *assoc,
822 unsigned int assoclen)
823 {
824 req->assoc = assoc;
825 req->assoclen = assoclen;
826 }
827
828 static inline struct crypto_blkcipher *__crypto_blkcipher_cast(
829 struct crypto_tfm *tfm)
830 {
831 return (struct crypto_blkcipher *)tfm;
832 }
833
834 static inline struct crypto_blkcipher *crypto_blkcipher_cast(
835 struct crypto_tfm *tfm)
836 {
837 BUG_ON(crypto_tfm_alg_type(tfm) != CRYPTO_ALG_TYPE_BLKCIPHER);
838 return __crypto_blkcipher_cast(tfm);
839 }
840
841 static inline struct crypto_blkcipher *crypto_alloc_blkcipher(
842 const char *alg_name, u32 type, u32 mask)
843 {
844 type &= ~(CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_ASYNC);
845 type |= CRYPTO_ALG_TYPE_BLKCIPHER;
846 mask |= CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_ASYNC;
847
848 return __crypto_blkcipher_cast(crypto_alloc_base(alg_name, type, mask));
849 }
850
851 static inline struct crypto_tfm *crypto_blkcipher_tfm(
852 struct crypto_blkcipher *tfm)
853 {
854 return &tfm->base;
855 }
856
857 static inline void crypto_free_blkcipher(struct crypto_blkcipher *tfm)
858 {
859 crypto_free_tfm(crypto_blkcipher_tfm(tfm));
860 }
861
862 static inline int crypto_has_blkcipher(const char *alg_name, u32 type, u32 mask)
863 {
864 type &= ~(CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_ASYNC);
865 type |= CRYPTO_ALG_TYPE_BLKCIPHER;
866 mask |= CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_ASYNC;
867
868 return crypto_has_alg(alg_name, type, mask);
869 }
870
871 static inline const char *crypto_blkcipher_name(struct crypto_blkcipher *tfm)
872 {
873 return crypto_tfm_alg_name(crypto_blkcipher_tfm(tfm));
874 }
875
876 static inline struct blkcipher_tfm *crypto_blkcipher_crt(
877 struct crypto_blkcipher *tfm)
878 {
879 return &crypto_blkcipher_tfm(tfm)->crt_blkcipher;
880 }
881
882 static inline struct blkcipher_alg *crypto_blkcipher_alg(
883 struct crypto_blkcipher *tfm)
884 {
885 return &crypto_blkcipher_tfm(tfm)->__crt_alg->cra_blkcipher;
886 }
887
888 static inline unsigned int crypto_blkcipher_ivsize(struct crypto_blkcipher *tfm)
889 {
890 return crypto_blkcipher_alg(tfm)->ivsize;
891 }
892
893 static inline unsigned int crypto_blkcipher_blocksize(
894 struct crypto_blkcipher *tfm)
895 {
896 return crypto_tfm_alg_blocksize(crypto_blkcipher_tfm(tfm));
897 }
898
899 static inline unsigned int crypto_blkcipher_alignmask(
900 struct crypto_blkcipher *tfm)
901 {
902 return crypto_tfm_alg_alignmask(crypto_blkcipher_tfm(tfm));
903 }
904
905 static inline u32 crypto_blkcipher_get_flags(struct crypto_blkcipher *tfm)
906 {
907 return crypto_tfm_get_flags(crypto_blkcipher_tfm(tfm));
908 }
909
910 static inline void crypto_blkcipher_set_flags(struct crypto_blkcipher *tfm,
911 u32 flags)
912 {
913 crypto_tfm_set_flags(crypto_blkcipher_tfm(tfm), flags);
914 }
915
916 static inline void crypto_blkcipher_clear_flags(struct crypto_blkcipher *tfm,
917 u32 flags)
918 {
919 crypto_tfm_clear_flags(crypto_blkcipher_tfm(tfm), flags);
920 }
921
922 static inline int crypto_blkcipher_setkey(struct crypto_blkcipher *tfm,
923 const u8 *key, unsigned int keylen)
924 {
925 return crypto_blkcipher_crt(tfm)->setkey(crypto_blkcipher_tfm(tfm),
926 key, keylen);
927 }
928
929 static inline int crypto_blkcipher_encrypt(struct blkcipher_desc *desc,
930 struct scatterlist *dst,
931 struct scatterlist *src,
932 unsigned int nbytes)
933 {
934 desc->info = crypto_blkcipher_crt(desc->tfm)->iv;
935 return crypto_blkcipher_crt(desc->tfm)->encrypt(desc, dst, src, nbytes);
936 }
937
938 static inline int crypto_blkcipher_encrypt_iv(struct blkcipher_desc *desc,
939 struct scatterlist *dst,
940 struct scatterlist *src,
941 unsigned int nbytes)
942 {
943 return crypto_blkcipher_crt(desc->tfm)->encrypt(desc, dst, src, nbytes);
944 }
945
946 static inline int crypto_blkcipher_decrypt(struct blkcipher_desc *desc,
947 struct scatterlist *dst,
948 struct scatterlist *src,
949 unsigned int nbytes)
950 {
951 desc->info = crypto_blkcipher_crt(desc->tfm)->iv;
952 return crypto_blkcipher_crt(desc->tfm)->decrypt(desc, dst, src, nbytes);
953 }
954
955 static inline int crypto_blkcipher_decrypt_iv(struct blkcipher_desc *desc,
956 struct scatterlist *dst,
957 struct scatterlist *src,
958 unsigned int nbytes)
959 {
960 return crypto_blkcipher_crt(desc->tfm)->decrypt(desc, dst, src, nbytes);
961 }
962
963 static inline void crypto_blkcipher_set_iv(struct crypto_blkcipher *tfm,
964 const u8 *src, unsigned int len)
965 {
966 memcpy(crypto_blkcipher_crt(tfm)->iv, src, len);
967 }
968
969 static inline void crypto_blkcipher_get_iv(struct crypto_blkcipher *tfm,
970 u8 *dst, unsigned int len)
971 {
972 memcpy(dst, crypto_blkcipher_crt(tfm)->iv, len);
973 }
974
975 static inline struct crypto_cipher *__crypto_cipher_cast(struct crypto_tfm *tfm)
976 {
977 return (struct crypto_cipher *)tfm;
978 }
979
980 static inline struct crypto_cipher *crypto_cipher_cast(struct crypto_tfm *tfm)
981 {
982 BUG_ON(crypto_tfm_alg_type(tfm) != CRYPTO_ALG_TYPE_CIPHER);
983 return __crypto_cipher_cast(tfm);
984 }
985
986 static inline struct crypto_cipher *crypto_alloc_cipher(const char *alg_name,
987 u32 type, u32 mask)
988 {
989 type &= ~CRYPTO_ALG_TYPE_MASK;
990 type |= CRYPTO_ALG_TYPE_CIPHER;
991 mask |= CRYPTO_ALG_TYPE_MASK;
992
993 return __crypto_cipher_cast(crypto_alloc_base(alg_name, type, mask));
994 }
995
996 static inline struct crypto_tfm *crypto_cipher_tfm(struct crypto_cipher *tfm)
997 {
998 return &tfm->base;
999 }
1000
1001 static inline void crypto_free_cipher(struct crypto_cipher *tfm)
1002 {
1003 crypto_free_tfm(crypto_cipher_tfm(tfm));
1004 }
1005
1006 static inline int crypto_has_cipher(const char *alg_name, u32 type, u32 mask)
1007 {
1008 type &= ~CRYPTO_ALG_TYPE_MASK;
1009 type |= CRYPTO_ALG_TYPE_CIPHER;
1010 mask |= CRYPTO_ALG_TYPE_MASK;
1011
1012 return crypto_has_alg(alg_name, type, mask);
1013 }
1014
1015 static inline struct cipher_tfm *crypto_cipher_crt(struct crypto_cipher *tfm)
1016 {
1017 return &crypto_cipher_tfm(tfm)->crt_cipher;
1018 }
1019
1020 static inline unsigned int crypto_cipher_blocksize(struct crypto_cipher *tfm)
1021 {
1022 return crypto_tfm_alg_blocksize(crypto_cipher_tfm(tfm));
1023 }
1024
1025 static inline unsigned int crypto_cipher_alignmask(struct crypto_cipher *tfm)
1026 {
1027 return crypto_tfm_alg_alignmask(crypto_cipher_tfm(tfm));
1028 }
1029
1030 static inline u32 crypto_cipher_get_flags(struct crypto_cipher *tfm)
1031 {
1032 return crypto_tfm_get_flags(crypto_cipher_tfm(tfm));
1033 }
1034
1035 static inline void crypto_cipher_set_flags(struct crypto_cipher *tfm,
1036 u32 flags)
1037 {
1038 crypto_tfm_set_flags(crypto_cipher_tfm(tfm), flags);
1039 }
1040
1041 static inline void crypto_cipher_clear_flags(struct crypto_cipher *tfm,
1042 u32 flags)
1043 {
1044 crypto_tfm_clear_flags(crypto_cipher_tfm(tfm), flags);
1045 }
1046
1047 static inline int crypto_cipher_setkey(struct crypto_cipher *tfm,
1048 const u8 *key, unsigned int keylen)
1049 {
1050 return crypto_cipher_crt(tfm)->cit_setkey(crypto_cipher_tfm(tfm),
1051 key, keylen);
1052 }
1053
1054 static inline void crypto_cipher_encrypt_one(struct crypto_cipher *tfm,
1055 u8 *dst, const u8 *src)
1056 {
1057 crypto_cipher_crt(tfm)->cit_encrypt_one(crypto_cipher_tfm(tfm),
1058 dst, src);
1059 }
1060
1061 static inline void crypto_cipher_decrypt_one(struct crypto_cipher *tfm,
1062 u8 *dst, const u8 *src)
1063 {
1064 crypto_cipher_crt(tfm)->cit_decrypt_one(crypto_cipher_tfm(tfm),
1065 dst, src);
1066 }
1067
1068 static inline struct crypto_hash *__crypto_hash_cast(struct crypto_tfm *tfm)
1069 {
1070 return (struct crypto_hash *)tfm;
1071 }
1072
1073 static inline struct crypto_hash *crypto_hash_cast(struct crypto_tfm *tfm)
1074 {
1075 BUG_ON((crypto_tfm_alg_type(tfm) ^ CRYPTO_ALG_TYPE_HASH) &
1076 CRYPTO_ALG_TYPE_HASH_MASK);
1077 return __crypto_hash_cast(tfm);
1078 }
1079
1080 static inline struct crypto_hash *crypto_alloc_hash(const char *alg_name,
1081 u32 type, u32 mask)
1082 {
1083 type &= ~CRYPTO_ALG_TYPE_MASK;
1084 type |= CRYPTO_ALG_TYPE_HASH;
1085 mask |= CRYPTO_ALG_TYPE_HASH_MASK;
1086
1087 return __crypto_hash_cast(crypto_alloc_base(alg_name, type, mask));
1088 }
1089
1090 static inline struct crypto_tfm *crypto_hash_tfm(struct crypto_hash *tfm)
1091 {
1092 return &tfm->base;
1093 }
1094
1095 static inline void crypto_free_hash(struct crypto_hash *tfm)
1096 {
1097 crypto_free_tfm(crypto_hash_tfm(tfm));
1098 }
1099
1100 static inline int crypto_has_hash(const char *alg_name, u32 type, u32 mask)
1101 {
1102 type &= ~CRYPTO_ALG_TYPE_MASK;
1103 type |= CRYPTO_ALG_TYPE_HASH;
1104 mask |= CRYPTO_ALG_TYPE_HASH_MASK;
1105
1106 return crypto_has_alg(alg_name, type, mask);
1107 }
1108
1109 static inline struct hash_tfm *crypto_hash_crt(struct crypto_hash *tfm)
1110 {
1111 return &crypto_hash_tfm(tfm)->crt_hash;
1112 }
1113
1114 static inline unsigned int crypto_hash_blocksize(struct crypto_hash *tfm)
1115 {
1116 return crypto_tfm_alg_blocksize(crypto_hash_tfm(tfm));
1117 }
1118
1119 static inline unsigned int crypto_hash_alignmask(struct crypto_hash *tfm)
1120 {
1121 return crypto_tfm_alg_alignmask(crypto_hash_tfm(tfm));
1122 }
1123
1124 static inline unsigned int crypto_hash_digestsize(struct crypto_hash *tfm)
1125 {
1126 return crypto_hash_crt(tfm)->digestsize;
1127 }
1128
1129 static inline u32 crypto_hash_get_flags(struct crypto_hash *tfm)
1130 {
1131 return crypto_tfm_get_flags(crypto_hash_tfm(tfm));
1132 }
1133
1134 static inline void crypto_hash_set_flags(struct crypto_hash *tfm, u32 flags)
1135 {
1136 crypto_tfm_set_flags(crypto_hash_tfm(tfm), flags);
1137 }
1138
1139 static inline void crypto_hash_clear_flags(struct crypto_hash *tfm, u32 flags)
1140 {
1141 crypto_tfm_clear_flags(crypto_hash_tfm(tfm), flags);
1142 }
1143
1144 static inline int crypto_hash_init(struct hash_desc *desc)
1145 {
1146 return crypto_hash_crt(desc->tfm)->init(desc);
1147 }
1148
1149 static inline int crypto_hash_update(struct hash_desc *desc,
1150 struct scatterlist *sg,
1151 unsigned int nbytes)
1152 {
1153 return crypto_hash_crt(desc->tfm)->update(desc, sg, nbytes);
1154 }
1155
1156 static inline int crypto_hash_final(struct hash_desc *desc, u8 *out)
1157 {
1158 return crypto_hash_crt(desc->tfm)->final(desc, out);
1159 }
1160
1161 static inline int crypto_hash_digest(struct hash_desc *desc,
1162 struct scatterlist *sg,
1163 unsigned int nbytes, u8 *out)
1164 {
1165 return crypto_hash_crt(desc->tfm)->digest(desc, sg, nbytes, out);
1166 }
1167
1168 static inline int crypto_hash_setkey(struct crypto_hash *hash,
1169 const u8 *key, unsigned int keylen)
1170 {
1171 return crypto_hash_crt(hash)->setkey(hash, key, keylen);
1172 }
1173
1174 static inline struct crypto_comp *__crypto_comp_cast(struct crypto_tfm *tfm)
1175 {
1176 return (struct crypto_comp *)tfm;
1177 }
1178
1179 static inline struct crypto_comp *crypto_comp_cast(struct crypto_tfm *tfm)
1180 {
1181 BUG_ON((crypto_tfm_alg_type(tfm) ^ CRYPTO_ALG_TYPE_COMPRESS) &
1182 CRYPTO_ALG_TYPE_MASK);
1183 return __crypto_comp_cast(tfm);
1184 }
1185
1186 static inline struct crypto_comp *crypto_alloc_comp(const char *alg_name,
1187 u32 type, u32 mask)
1188 {
1189 type &= ~CRYPTO_ALG_TYPE_MASK;
1190 type |= CRYPTO_ALG_TYPE_COMPRESS;
1191 mask |= CRYPTO_ALG_TYPE_MASK;
1192
1193 return __crypto_comp_cast(crypto_alloc_base(alg_name, type, mask));
1194 }
1195
1196 static inline struct crypto_tfm *crypto_comp_tfm(struct crypto_comp *tfm)
1197 {
1198 return &tfm->base;
1199 }
1200
1201 static inline void crypto_free_comp(struct crypto_comp *tfm)
1202 {
1203 crypto_free_tfm(crypto_comp_tfm(tfm));
1204 }
1205
1206 static inline int crypto_has_comp(const char *alg_name, u32 type, u32 mask)
1207 {
1208 type &= ~CRYPTO_ALG_TYPE_MASK;
1209 type |= CRYPTO_ALG_TYPE_COMPRESS;
1210 mask |= CRYPTO_ALG_TYPE_MASK;
1211
1212 return crypto_has_alg(alg_name, type, mask);
1213 }
1214
1215 static inline const char *crypto_comp_name(struct crypto_comp *tfm)
1216 {
1217 return crypto_tfm_alg_name(crypto_comp_tfm(tfm));
1218 }
1219
1220 static inline struct compress_tfm *crypto_comp_crt(struct crypto_comp *tfm)
1221 {
1222 return &crypto_comp_tfm(tfm)->crt_compress;
1223 }
1224
1225 static inline int crypto_comp_compress(struct crypto_comp *tfm,
1226 const u8 *src, unsigned int slen,
1227 u8 *dst, unsigned int *dlen)
1228 {
1229 return crypto_comp_crt(tfm)->cot_compress(crypto_comp_tfm(tfm),
1230 src, slen, dst, dlen);
1231 }
1232
1233 static inline int crypto_comp_decompress(struct crypto_comp *tfm,
1234 const u8 *src, unsigned int slen,
1235 u8 *dst, unsigned int *dlen)
1236 {
1237 return crypto_comp_crt(tfm)->cot_decompress(crypto_comp_tfm(tfm),
1238 src, slen, dst, dlen);
1239 }
1240
1241 #endif /* _LINUX_CRYPTO_H */
1242
This page took 0.081786 seconds and 6 git commands to generate.