]> git.proxmox.com Git - mirror_ubuntu-artful-kernel.git/blob - crypto/internal.h
Introduce "hcall" pointer to indicate pending hypercall.
[mirror_ubuntu-artful-kernel.git] / crypto / internal.h
1 /*
2 * Cryptographic API.
3 *
4 * Copyright (c) 2002 James Morris <jmorris@intercode.com.au>
5 * Copyright (c) 2005 Herbert Xu <herbert@gondor.apana.org.au>
6 *
7 * This program is free software; you can redistribute it and/or modify it
8 * under the terms of the GNU General Public License as published by the Free
9 * Software Foundation; either version 2 of the License, or (at your option)
10 * any later version.
11 *
12 */
13 #ifndef _CRYPTO_INTERNAL_H
14 #define _CRYPTO_INTERNAL_H
15
16 #include <crypto/algapi.h>
17 #include <linux/completion.h>
18 #include <linux/mm.h>
19 #include <linux/highmem.h>
20 #include <linux/interrupt.h>
21 #include <linux/init.h>
22 #include <linux/list.h>
23 #include <linux/module.h>
24 #include <linux/kernel.h>
25 #include <linux/notifier.h>
26 #include <linux/rwsem.h>
27 #include <linux/slab.h>
28 #include <asm/kmap_types.h>
29
30 /* Crypto notification events. */
31 enum {
32 CRYPTO_MSG_ALG_REQUEST,
33 CRYPTO_MSG_ALG_REGISTER,
34 CRYPTO_MSG_ALG_UNREGISTER,
35 CRYPTO_MSG_TMPL_REGISTER,
36 CRYPTO_MSG_TMPL_UNREGISTER,
37 };
38
39 struct crypto_instance;
40 struct crypto_template;
41
42 struct crypto_larval {
43 struct crypto_alg alg;
44 struct crypto_alg *adult;
45 struct completion completion;
46 u32 mask;
47 };
48
49 extern struct list_head crypto_alg_list;
50 extern struct rw_semaphore crypto_alg_sem;
51 extern struct blocking_notifier_head crypto_chain;
52
53 static inline enum km_type crypto_kmap_type(int out)
54 {
55 enum km_type type;
56
57 if (in_softirq())
58 type = out * (KM_SOFTIRQ1 - KM_SOFTIRQ0) + KM_SOFTIRQ0;
59 else
60 type = out * (KM_USER1 - KM_USER0) + KM_USER0;
61
62 return type;
63 }
64
65 static inline void *crypto_kmap(struct page *page, int out)
66 {
67 return kmap_atomic(page, crypto_kmap_type(out));
68 }
69
70 static inline void crypto_kunmap(void *vaddr, int out)
71 {
72 kunmap_atomic(vaddr, crypto_kmap_type(out));
73 }
74
75 static inline void crypto_yield(u32 flags)
76 {
77 if (flags & CRYPTO_TFM_REQ_MAY_SLEEP)
78 cond_resched();
79 }
80
81 #ifdef CONFIG_PROC_FS
82 void __init crypto_init_proc(void);
83 void __exit crypto_exit_proc(void);
84 #else
85 static inline void crypto_init_proc(void)
86 { }
87 static inline void crypto_exit_proc(void)
88 { }
89 #endif
90
91 static inline unsigned int crypto_digest_ctxsize(struct crypto_alg *alg)
92 {
93 unsigned int len = alg->cra_ctxsize;
94
95 if (alg->cra_alignmask) {
96 len = ALIGN(len, (unsigned long)alg->cra_alignmask + 1);
97 len += alg->cra_digest.dia_digestsize;
98 }
99
100 return len;
101 }
102
103 static inline unsigned int crypto_cipher_ctxsize(struct crypto_alg *alg)
104 {
105 return alg->cra_ctxsize;
106 }
107
108 static inline unsigned int crypto_compress_ctxsize(struct crypto_alg *alg)
109 {
110 return alg->cra_ctxsize;
111 }
112
113 struct crypto_alg *crypto_mod_get(struct crypto_alg *alg);
114 struct crypto_alg *__crypto_alg_lookup(const char *name, u32 type, u32 mask);
115 struct crypto_alg *crypto_alg_mod_lookup(const char *name, u32 type, u32 mask);
116
117 int crypto_init_digest_ops(struct crypto_tfm *tfm);
118 int crypto_init_cipher_ops(struct crypto_tfm *tfm);
119 int crypto_init_compress_ops(struct crypto_tfm *tfm);
120
121 void crypto_exit_digest_ops(struct crypto_tfm *tfm);
122 void crypto_exit_cipher_ops(struct crypto_tfm *tfm);
123 void crypto_exit_compress_ops(struct crypto_tfm *tfm);
124
125 void crypto_larval_error(const char *name, u32 type, u32 mask);
126
127 void crypto_shoot_alg(struct crypto_alg *alg);
128 struct crypto_tfm *__crypto_alloc_tfm(struct crypto_alg *alg, u32 type,
129 u32 mask);
130
131 int crypto_register_instance(struct crypto_template *tmpl,
132 struct crypto_instance *inst);
133
134 int crypto_register_notifier(struct notifier_block *nb);
135 int crypto_unregister_notifier(struct notifier_block *nb);
136
137 static inline void crypto_alg_put(struct crypto_alg *alg)
138 {
139 if (atomic_dec_and_test(&alg->cra_refcnt) && alg->cra_destroy)
140 alg->cra_destroy(alg);
141 }
142
143 static inline int crypto_tmpl_get(struct crypto_template *tmpl)
144 {
145 return try_module_get(tmpl->module);
146 }
147
148 static inline void crypto_tmpl_put(struct crypto_template *tmpl)
149 {
150 module_put(tmpl->module);
151 }
152
153 static inline int crypto_is_larval(struct crypto_alg *alg)
154 {
155 return alg->cra_flags & CRYPTO_ALG_LARVAL;
156 }
157
158 static inline int crypto_is_dead(struct crypto_alg *alg)
159 {
160 return alg->cra_flags & CRYPTO_ALG_DEAD;
161 }
162
163 static inline int crypto_is_moribund(struct crypto_alg *alg)
164 {
165 return alg->cra_flags & (CRYPTO_ALG_DEAD | CRYPTO_ALG_DYING);
166 }
167
168 static inline int crypto_notify(unsigned long val, void *v)
169 {
170 return blocking_notifier_call_chain(&crypto_chain, val, v);
171 }
172
173 #endif /* _CRYPTO_INTERNAL_H */
174