]> git.proxmox.com Git - mirror_ubuntu-eoan-kernel.git/blob - crypto/crypto_user_stat.c
7b668c659122193f3fd9f3c508f199f61878d017
[mirror_ubuntu-eoan-kernel.git] / crypto / crypto_user_stat.c
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * Crypto user configuration API.
4 *
5 * Copyright (C) 2017-2018 Corentin Labbe <clabbe@baylibre.com>
6 *
7 */
8
9 #include <linux/crypto.h>
10 #include <linux/cryptouser.h>
11 #include <linux/sched.h>
12 #include <net/netlink.h>
13 #include <crypto/internal/skcipher.h>
14 #include <crypto/internal/rng.h>
15 #include <crypto/akcipher.h>
16 #include <crypto/kpp.h>
17 #include <crypto/internal/cryptouser.h>
18
19 #include "internal.h"
20
21 #define null_terminated(x) (strnlen(x, sizeof(x)) < sizeof(x))
22
23 static DEFINE_MUTEX(crypto_cfg_mutex);
24
25 extern struct sock *crypto_nlsk;
26
27 struct crypto_dump_info {
28 struct sk_buff *in_skb;
29 struct sk_buff *out_skb;
30 u32 nlmsg_seq;
31 u16 nlmsg_flags;
32 };
33
34 static int crypto_report_aead(struct sk_buff *skb, struct crypto_alg *alg)
35 {
36 struct crypto_stat_aead raead;
37
38 memset(&raead, 0, sizeof(raead));
39
40 strscpy(raead.type, "aead", sizeof(raead.type));
41
42 raead.stat_encrypt_cnt = atomic64_read(&alg->encrypt_cnt);
43 raead.stat_encrypt_tlen = atomic64_read(&alg->encrypt_tlen);
44 raead.stat_decrypt_cnt = atomic64_read(&alg->decrypt_cnt);
45 raead.stat_decrypt_tlen = atomic64_read(&alg->decrypt_tlen);
46 raead.stat_aead_err_cnt = atomic64_read(&alg->aead_err_cnt);
47
48 return nla_put(skb, CRYPTOCFGA_STAT_AEAD, sizeof(raead), &raead);
49 }
50
51 static int crypto_report_cipher(struct sk_buff *skb, struct crypto_alg *alg)
52 {
53 struct crypto_stat_cipher rcipher;
54
55 memset(&rcipher, 0, sizeof(rcipher));
56
57 strscpy(rcipher.type, "cipher", sizeof(rcipher.type));
58
59 rcipher.stat_encrypt_cnt = atomic64_read(&alg->encrypt_cnt);
60 rcipher.stat_encrypt_tlen = atomic64_read(&alg->encrypt_tlen);
61 rcipher.stat_decrypt_cnt = atomic64_read(&alg->decrypt_cnt);
62 rcipher.stat_decrypt_tlen = atomic64_read(&alg->decrypt_tlen);
63 rcipher.stat_cipher_err_cnt = atomic64_read(&alg->cipher_err_cnt);
64
65 return nla_put(skb, CRYPTOCFGA_STAT_CIPHER, sizeof(rcipher), &rcipher);
66 }
67
68 static int crypto_report_comp(struct sk_buff *skb, struct crypto_alg *alg)
69 {
70 struct crypto_stat_compress rcomp;
71
72 memset(&rcomp, 0, sizeof(rcomp));
73
74 strscpy(rcomp.type, "compression", sizeof(rcomp.type));
75 rcomp.stat_compress_cnt = atomic64_read(&alg->compress_cnt);
76 rcomp.stat_compress_tlen = atomic64_read(&alg->compress_tlen);
77 rcomp.stat_decompress_cnt = atomic64_read(&alg->decompress_cnt);
78 rcomp.stat_decompress_tlen = atomic64_read(&alg->decompress_tlen);
79 rcomp.stat_compress_err_cnt = atomic64_read(&alg->compress_err_cnt);
80
81 return nla_put(skb, CRYPTOCFGA_STAT_COMPRESS, sizeof(rcomp), &rcomp);
82 }
83
84 static int crypto_report_acomp(struct sk_buff *skb, struct crypto_alg *alg)
85 {
86 struct crypto_stat_compress racomp;
87
88 memset(&racomp, 0, sizeof(racomp));
89
90 strscpy(racomp.type, "acomp", sizeof(racomp.type));
91 racomp.stat_compress_cnt = atomic64_read(&alg->compress_cnt);
92 racomp.stat_compress_tlen = atomic64_read(&alg->compress_tlen);
93 racomp.stat_decompress_cnt = atomic64_read(&alg->decompress_cnt);
94 racomp.stat_decompress_tlen = atomic64_read(&alg->decompress_tlen);
95 racomp.stat_compress_err_cnt = atomic64_read(&alg->compress_err_cnt);
96
97 return nla_put(skb, CRYPTOCFGA_STAT_ACOMP, sizeof(racomp), &racomp);
98 }
99
100 static int crypto_report_akcipher(struct sk_buff *skb, struct crypto_alg *alg)
101 {
102 struct crypto_stat_akcipher rakcipher;
103
104 memset(&rakcipher, 0, sizeof(rakcipher));
105
106 strscpy(rakcipher.type, "akcipher", sizeof(rakcipher.type));
107 rakcipher.stat_encrypt_cnt = atomic64_read(&alg->encrypt_cnt);
108 rakcipher.stat_encrypt_tlen = atomic64_read(&alg->encrypt_tlen);
109 rakcipher.stat_decrypt_cnt = atomic64_read(&alg->decrypt_cnt);
110 rakcipher.stat_decrypt_tlen = atomic64_read(&alg->decrypt_tlen);
111 rakcipher.stat_sign_cnt = atomic64_read(&alg->sign_cnt);
112 rakcipher.stat_verify_cnt = atomic64_read(&alg->verify_cnt);
113 rakcipher.stat_akcipher_err_cnt = atomic64_read(&alg->akcipher_err_cnt);
114
115 return nla_put(skb, CRYPTOCFGA_STAT_AKCIPHER,
116 sizeof(rakcipher), &rakcipher);
117 }
118
119 static int crypto_report_kpp(struct sk_buff *skb, struct crypto_alg *alg)
120 {
121 struct crypto_stat_kpp rkpp;
122
123 memset(&rkpp, 0, sizeof(rkpp));
124
125 strscpy(rkpp.type, "kpp", sizeof(rkpp.type));
126
127 rkpp.stat_setsecret_cnt = atomic64_read(&alg->setsecret_cnt);
128 rkpp.stat_generate_public_key_cnt = atomic64_read(&alg->generate_public_key_cnt);
129 rkpp.stat_compute_shared_secret_cnt = atomic64_read(&alg->compute_shared_secret_cnt);
130 rkpp.stat_kpp_err_cnt = atomic64_read(&alg->kpp_err_cnt);
131
132 return nla_put(skb, CRYPTOCFGA_STAT_KPP, sizeof(rkpp), &rkpp);
133 }
134
135 static int crypto_report_ahash(struct sk_buff *skb, struct crypto_alg *alg)
136 {
137 struct crypto_stat_hash rhash;
138
139 memset(&rhash, 0, sizeof(rhash));
140
141 strscpy(rhash.type, "ahash", sizeof(rhash.type));
142
143 rhash.stat_hash_cnt = atomic64_read(&alg->hash_cnt);
144 rhash.stat_hash_tlen = atomic64_read(&alg->hash_tlen);
145 rhash.stat_hash_err_cnt = atomic64_read(&alg->hash_err_cnt);
146
147 return nla_put(skb, CRYPTOCFGA_STAT_HASH, sizeof(rhash), &rhash);
148 }
149
150 static int crypto_report_shash(struct sk_buff *skb, struct crypto_alg *alg)
151 {
152 struct crypto_stat_hash rhash;
153
154 memset(&rhash, 0, sizeof(rhash));
155
156 strscpy(rhash.type, "shash", sizeof(rhash.type));
157
158 rhash.stat_hash_cnt = atomic64_read(&alg->hash_cnt);
159 rhash.stat_hash_tlen = atomic64_read(&alg->hash_tlen);
160 rhash.stat_hash_err_cnt = atomic64_read(&alg->hash_err_cnt);
161
162 return nla_put(skb, CRYPTOCFGA_STAT_HASH, sizeof(rhash), &rhash);
163 }
164
165 static int crypto_report_rng(struct sk_buff *skb, struct crypto_alg *alg)
166 {
167 struct crypto_stat_rng rrng;
168
169 memset(&rrng, 0, sizeof(rrng));
170
171 strscpy(rrng.type, "rng", sizeof(rrng.type));
172
173 rrng.stat_generate_cnt = atomic64_read(&alg->generate_cnt);
174 rrng.stat_generate_tlen = atomic64_read(&alg->generate_tlen);
175 rrng.stat_seed_cnt = atomic64_read(&alg->seed_cnt);
176 rrng.stat_rng_err_cnt = atomic64_read(&alg->rng_err_cnt);
177
178 return nla_put(skb, CRYPTOCFGA_STAT_RNG, sizeof(rrng), &rrng);
179 }
180
181 static int crypto_reportstat_one(struct crypto_alg *alg,
182 struct crypto_user_alg *ualg,
183 struct sk_buff *skb)
184 {
185 memset(ualg, 0, sizeof(*ualg));
186
187 strscpy(ualg->cru_name, alg->cra_name, sizeof(ualg->cru_name));
188 strscpy(ualg->cru_driver_name, alg->cra_driver_name,
189 sizeof(ualg->cru_driver_name));
190 strscpy(ualg->cru_module_name, module_name(alg->cra_module),
191 sizeof(ualg->cru_module_name));
192
193 ualg->cru_type = 0;
194 ualg->cru_mask = 0;
195 ualg->cru_flags = alg->cra_flags;
196 ualg->cru_refcnt = refcount_read(&alg->cra_refcnt);
197
198 if (nla_put_u32(skb, CRYPTOCFGA_PRIORITY_VAL, alg->cra_priority))
199 goto nla_put_failure;
200 if (alg->cra_flags & CRYPTO_ALG_LARVAL) {
201 struct crypto_stat_larval rl;
202
203 memset(&rl, 0, sizeof(rl));
204 strscpy(rl.type, "larval", sizeof(rl.type));
205 if (nla_put(skb, CRYPTOCFGA_STAT_LARVAL, sizeof(rl), &rl))
206 goto nla_put_failure;
207 goto out;
208 }
209
210 switch (alg->cra_flags & (CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_LARVAL)) {
211 case CRYPTO_ALG_TYPE_AEAD:
212 if (crypto_report_aead(skb, alg))
213 goto nla_put_failure;
214 break;
215 case CRYPTO_ALG_TYPE_SKCIPHER:
216 if (crypto_report_cipher(skb, alg))
217 goto nla_put_failure;
218 break;
219 case CRYPTO_ALG_TYPE_BLKCIPHER:
220 if (crypto_report_cipher(skb, alg))
221 goto nla_put_failure;
222 break;
223 case CRYPTO_ALG_TYPE_CIPHER:
224 if (crypto_report_cipher(skb, alg))
225 goto nla_put_failure;
226 break;
227 case CRYPTO_ALG_TYPE_COMPRESS:
228 if (crypto_report_comp(skb, alg))
229 goto nla_put_failure;
230 break;
231 case CRYPTO_ALG_TYPE_ACOMPRESS:
232 if (crypto_report_acomp(skb, alg))
233 goto nla_put_failure;
234 break;
235 case CRYPTO_ALG_TYPE_SCOMPRESS:
236 if (crypto_report_acomp(skb, alg))
237 goto nla_put_failure;
238 break;
239 case CRYPTO_ALG_TYPE_AKCIPHER:
240 if (crypto_report_akcipher(skb, alg))
241 goto nla_put_failure;
242 break;
243 case CRYPTO_ALG_TYPE_KPP:
244 if (crypto_report_kpp(skb, alg))
245 goto nla_put_failure;
246 break;
247 case CRYPTO_ALG_TYPE_AHASH:
248 if (crypto_report_ahash(skb, alg))
249 goto nla_put_failure;
250 break;
251 case CRYPTO_ALG_TYPE_HASH:
252 if (crypto_report_shash(skb, alg))
253 goto nla_put_failure;
254 break;
255 case CRYPTO_ALG_TYPE_RNG:
256 if (crypto_report_rng(skb, alg))
257 goto nla_put_failure;
258 break;
259 default:
260 pr_err("ERROR: Unhandled alg %d in %s\n",
261 alg->cra_flags & (CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_LARVAL),
262 __func__);
263 }
264
265 out:
266 return 0;
267
268 nla_put_failure:
269 return -EMSGSIZE;
270 }
271
272 static int crypto_reportstat_alg(struct crypto_alg *alg,
273 struct crypto_dump_info *info)
274 {
275 struct sk_buff *in_skb = info->in_skb;
276 struct sk_buff *skb = info->out_skb;
277 struct nlmsghdr *nlh;
278 struct crypto_user_alg *ualg;
279 int err = 0;
280
281 nlh = nlmsg_put(skb, NETLINK_CB(in_skb).portid, info->nlmsg_seq,
282 CRYPTO_MSG_GETSTAT, sizeof(*ualg), info->nlmsg_flags);
283 if (!nlh) {
284 err = -EMSGSIZE;
285 goto out;
286 }
287
288 ualg = nlmsg_data(nlh);
289
290 err = crypto_reportstat_one(alg, ualg, skb);
291 if (err) {
292 nlmsg_cancel(skb, nlh);
293 goto out;
294 }
295
296 nlmsg_end(skb, nlh);
297
298 out:
299 return err;
300 }
301
302 int crypto_reportstat(struct sk_buff *in_skb, struct nlmsghdr *in_nlh,
303 struct nlattr **attrs)
304 {
305 struct crypto_user_alg *p = nlmsg_data(in_nlh);
306 struct crypto_alg *alg;
307 struct sk_buff *skb;
308 struct crypto_dump_info info;
309 int err;
310
311 if (!null_terminated(p->cru_name) || !null_terminated(p->cru_driver_name))
312 return -EINVAL;
313
314 alg = crypto_alg_match(p, 0);
315 if (!alg)
316 return -ENOENT;
317
318 err = -ENOMEM;
319 skb = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_ATOMIC);
320 if (!skb)
321 goto drop_alg;
322
323 info.in_skb = in_skb;
324 info.out_skb = skb;
325 info.nlmsg_seq = in_nlh->nlmsg_seq;
326 info.nlmsg_flags = 0;
327
328 err = crypto_reportstat_alg(alg, &info);
329
330 drop_alg:
331 crypto_mod_put(alg);
332
333 if (err)
334 return err;
335
336 return nlmsg_unicast(crypto_nlsk, skb, NETLINK_CB(in_skb).portid);
337 }
338
339 int crypto_dump_reportstat(struct sk_buff *skb, struct netlink_callback *cb)
340 {
341 struct crypto_alg *alg;
342 struct crypto_dump_info info;
343 int err;
344
345 if (cb->args[0])
346 goto out;
347
348 cb->args[0] = 1;
349
350 info.in_skb = cb->skb;
351 info.out_skb = skb;
352 info.nlmsg_seq = cb->nlh->nlmsg_seq;
353 info.nlmsg_flags = NLM_F_MULTI;
354
355 list_for_each_entry(alg, &crypto_alg_list, cra_list) {
356 err = crypto_reportstat_alg(alg, &info);
357 if (err)
358 goto out_err;
359 }
360
361 out:
362 return skb->len;
363 out_err:
364 return err;
365 }
366
367 int crypto_dump_reportstat_done(struct netlink_callback *cb)
368 {
369 return 0;
370 }
371
372 MODULE_LICENSE("GPL");