Commit | Line | Data |
---|---|---|
8275d1aa TC |
1 | /* |
2 | * Cryptographic API. | |
3 | * | |
4 | * Glue code for the SHA256 Secure Hash Algorithm assembler | |
5 | * implementation using supplemental SSE3 / AVX / AVX2 instructions. | |
6 | * | |
7 | * This file is based on sha256_generic.c | |
8 | * | |
9 | * Copyright (C) 2013 Intel Corporation. | |
10 | * | |
11 | * Author: | |
12 | * Tim Chen <tim.c.chen@linux.intel.com> | |
13 | * | |
14 | * This program is free software; you can redistribute it and/or modify it | |
15 | * under the terms of the GNU General Public License as published by the Free | |
16 | * Software Foundation; either version 2 of the License, or (at your option) | |
17 | * any later version. | |
18 | * | |
19 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, | |
20 | * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF | |
21 | * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND | |
22 | * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS | |
23 | * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN | |
24 | * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN | |
25 | * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | |
26 | * SOFTWARE. | |
27 | */ | |
28 | ||
29 | ||
30 | #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt | |
31 | ||
32 | #include <crypto/internal/hash.h> | |
33 | #include <linux/init.h> | |
34 | #include <linux/module.h> | |
35 | #include <linux/mm.h> | |
36 | #include <linux/cryptohash.h> | |
37 | #include <linux/types.h> | |
38 | #include <crypto/sha.h> | |
1631030a | 39 | #include <crypto/sha256_base.h> |
df6b35f4 | 40 | #include <asm/fpu/api.h> |
8275d1aa TC |
41 | #include <linux/string.h> |
42 | ||
1631030a AB |
43 | asmlinkage void sha256_transform_ssse3(u32 *digest, const char *data, |
44 | u64 rounds); | |
5dda42fc | 45 | typedef void (sha256_transform_fn)(u32 *digest, const char *data, u64 rounds); |
8275d1aa | 46 | |
5dda42fc | 47 | static int sha256_update(struct shash_desc *desc, const u8 *data, |
48 | unsigned int len, sha256_transform_fn *sha256_xform) | |
8275d1aa TC |
49 | { |
50 | struct sha256_state *sctx = shash_desc_ctx(desc); | |
8275d1aa | 51 | |
1631030a AB |
52 | if (!irq_fpu_usable() || |
53 | (sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) | |
54 | return crypto_sha256_update(desc, data, len); | |
8275d1aa | 55 | |
1631030a AB |
56 | /* make sure casting to sha256_block_fn() is safe */ |
57 | BUILD_BUG_ON(offsetof(struct sha256_state, state) != 0); | |
8275d1aa | 58 | |
1631030a AB |
59 | kernel_fpu_begin(); |
60 | sha256_base_do_update(desc, data, len, | |
5dda42fc | 61 | (sha256_block_fn *)sha256_xform); |
1631030a | 62 | kernel_fpu_end(); |
8275d1aa TC |
63 | |
64 | return 0; | |
65 | } | |
66 | ||
5dda42fc | 67 | static int sha256_finup(struct shash_desc *desc, const u8 *data, |
68 | unsigned int len, u8 *out, sha256_transform_fn *sha256_xform) | |
8275d1aa | 69 | { |
1631030a AB |
70 | if (!irq_fpu_usable()) |
71 | return crypto_sha256_finup(desc, data, len, out); | |
8275d1aa | 72 | |
1631030a AB |
73 | kernel_fpu_begin(); |
74 | if (len) | |
75 | sha256_base_do_update(desc, data, len, | |
5dda42fc | 76 | (sha256_block_fn *)sha256_xform); |
77 | sha256_base_do_finalize(desc, (sha256_block_fn *)sha256_xform); | |
1631030a | 78 | kernel_fpu_end(); |
8275d1aa | 79 | |
1631030a | 80 | return sha256_base_finish(desc, out); |
8275d1aa TC |
81 | } |
82 | ||
5dda42fc | 83 | static int sha256_ssse3_update(struct shash_desc *desc, const u8 *data, |
84 | unsigned int len) | |
85 | { | |
86 | return sha256_update(desc, data, len, sha256_transform_ssse3); | |
87 | } | |
88 | ||
89 | static int sha256_ssse3_finup(struct shash_desc *desc, const u8 *data, | |
90 | unsigned int len, u8 *out) | |
91 | { | |
92 | return sha256_finup(desc, data, len, out, sha256_transform_ssse3); | |
93 | } | |
94 | ||
1631030a AB |
95 | /* Add padding and return the message digest. */ |
96 | static int sha256_ssse3_final(struct shash_desc *desc, u8 *out) | |
a710f761 | 97 | { |
1631030a | 98 | return sha256_ssse3_finup(desc, NULL, 0, out); |
a710f761 JK |
99 | } |
100 | ||
5dda42fc | 101 | static struct shash_alg sha256_ssse3_algs[] = { { |
8275d1aa | 102 | .digestsize = SHA256_DIGEST_SIZE, |
1631030a | 103 | .init = sha256_base_init, |
8275d1aa TC |
104 | .update = sha256_ssse3_update, |
105 | .final = sha256_ssse3_final, | |
1631030a | 106 | .finup = sha256_ssse3_finup, |
8275d1aa | 107 | .descsize = sizeof(struct sha256_state), |
8275d1aa TC |
108 | .base = { |
109 | .cra_name = "sha256", | |
110 | .cra_driver_name = "sha256-ssse3", | |
111 | .cra_priority = 150, | |
112 | .cra_flags = CRYPTO_ALG_TYPE_SHASH, | |
113 | .cra_blocksize = SHA256_BLOCK_SIZE, | |
114 | .cra_module = THIS_MODULE, | |
115 | } | |
a710f761 JK |
116 | }, { |
117 | .digestsize = SHA224_DIGEST_SIZE, | |
1631030a | 118 | .init = sha224_base_init, |
a710f761 | 119 | .update = sha256_ssse3_update, |
1631030a AB |
120 | .final = sha256_ssse3_final, |
121 | .finup = sha256_ssse3_finup, | |
a710f761 | 122 | .descsize = sizeof(struct sha256_state), |
a710f761 JK |
123 | .base = { |
124 | .cra_name = "sha224", | |
125 | .cra_driver_name = "sha224-ssse3", | |
126 | .cra_priority = 150, | |
127 | .cra_flags = CRYPTO_ALG_TYPE_SHASH, | |
128 | .cra_blocksize = SHA224_BLOCK_SIZE, | |
129 | .cra_module = THIS_MODULE, | |
130 | } | |
131 | } }; | |
8275d1aa | 132 | |
5dda42fc | 133 | static int register_sha256_ssse3(void) |
134 | { | |
135 | if (boot_cpu_has(X86_FEATURE_SSSE3)) | |
136 | return crypto_register_shashes(sha256_ssse3_algs, | |
137 | ARRAY_SIZE(sha256_ssse3_algs)); | |
138 | return 0; | |
139 | } | |
140 | ||
141 | static void unregister_sha256_ssse3(void) | |
142 | { | |
143 | if (boot_cpu_has(X86_FEATURE_SSSE3)) | |
144 | crypto_unregister_shashes(sha256_ssse3_algs, | |
145 | ARRAY_SIZE(sha256_ssse3_algs)); | |
146 | } | |
147 | ||
8275d1aa | 148 | #ifdef CONFIG_AS_AVX |
5dda42fc | 149 | asmlinkage void sha256_transform_avx(u32 *digest, const char *data, |
150 | u64 rounds); | |
151 | ||
152 | static int sha256_avx_update(struct shash_desc *desc, const u8 *data, | |
153 | unsigned int len) | |
154 | { | |
155 | return sha256_update(desc, data, len, sha256_transform_avx); | |
156 | } | |
157 | ||
158 | static int sha256_avx_finup(struct shash_desc *desc, const u8 *data, | |
159 | unsigned int len, u8 *out) | |
160 | { | |
161 | return sha256_finup(desc, data, len, out, sha256_transform_avx); | |
162 | } | |
163 | ||
164 | static int sha256_avx_final(struct shash_desc *desc, u8 *out) | |
165 | { | |
166 | return sha256_avx_finup(desc, NULL, 0, out); | |
167 | } | |
168 | ||
169 | static struct shash_alg sha256_avx_algs[] = { { | |
170 | .digestsize = SHA256_DIGEST_SIZE, | |
171 | .init = sha256_base_init, | |
172 | .update = sha256_avx_update, | |
173 | .final = sha256_avx_final, | |
174 | .finup = sha256_avx_finup, | |
175 | .descsize = sizeof(struct sha256_state), | |
176 | .base = { | |
177 | .cra_name = "sha256", | |
178 | .cra_driver_name = "sha256-avx", | |
179 | .cra_priority = 160, | |
180 | .cra_flags = CRYPTO_ALG_TYPE_SHASH, | |
181 | .cra_blocksize = SHA256_BLOCK_SIZE, | |
182 | .cra_module = THIS_MODULE, | |
183 | } | |
184 | }, { | |
185 | .digestsize = SHA224_DIGEST_SIZE, | |
186 | .init = sha224_base_init, | |
187 | .update = sha256_avx_update, | |
188 | .final = sha256_avx_final, | |
189 | .finup = sha256_avx_finup, | |
190 | .descsize = sizeof(struct sha256_state), | |
191 | .base = { | |
192 | .cra_name = "sha224", | |
193 | .cra_driver_name = "sha224-avx", | |
194 | .cra_priority = 160, | |
195 | .cra_flags = CRYPTO_ALG_TYPE_SHASH, | |
196 | .cra_blocksize = SHA224_BLOCK_SIZE, | |
197 | .cra_module = THIS_MODULE, | |
198 | } | |
199 | } }; | |
200 | ||
201 | static bool avx_usable(void) | |
8275d1aa | 202 | { |
d91cab78 | 203 | if (!cpu_has_xfeatures(XFEATURE_MASK_SSE | XFEATURE_MASK_YMM, NULL)) { |
da154e82 | 204 | if (boot_cpu_has(X86_FEATURE_AVX)) |
70d51eb6 | 205 | pr_info("AVX detected but unusable.\n"); |
8275d1aa TC |
206 | return false; |
207 | } | |
208 | ||
209 | return true; | |
210 | } | |
8275d1aa | 211 | |
5dda42fc | 212 | static int register_sha256_avx(void) |
8275d1aa | 213 | { |
5dda42fc | 214 | if (avx_usable()) |
215 | return crypto_register_shashes(sha256_avx_algs, | |
216 | ARRAY_SIZE(sha256_avx_algs)); | |
217 | return 0; | |
218 | } | |
8275d1aa | 219 | |
5dda42fc | 220 | static void unregister_sha256_avx(void) |
221 | { | |
222 | if (avx_usable()) | |
223 | crypto_unregister_shashes(sha256_avx_algs, | |
224 | ARRAY_SIZE(sha256_avx_algs)); | |
225 | } | |
8275d1aa | 226 | |
5dda42fc | 227 | #else |
228 | static inline int register_sha256_avx(void) { return 0; } | |
229 | static inline void unregister_sha256_avx(void) { } | |
8275d1aa | 230 | #endif |
5dda42fc | 231 | |
232 | #if defined(CONFIG_AS_AVX2) && defined(CONFIG_AS_AVX) | |
233 | asmlinkage void sha256_transform_rorx(u32 *digest, const char *data, | |
234 | u64 rounds); | |
235 | ||
236 | static int sha256_avx2_update(struct shash_desc *desc, const u8 *data, | |
237 | unsigned int len) | |
238 | { | |
239 | return sha256_update(desc, data, len, sha256_transform_rorx); | |
240 | } | |
241 | ||
242 | static int sha256_avx2_finup(struct shash_desc *desc, const u8 *data, | |
243 | unsigned int len, u8 *out) | |
244 | { | |
245 | return sha256_finup(desc, data, len, out, sha256_transform_rorx); | |
246 | } | |
247 | ||
248 | static int sha256_avx2_final(struct shash_desc *desc, u8 *out) | |
249 | { | |
250 | return sha256_avx2_finup(desc, NULL, 0, out); | |
251 | } | |
252 | ||
253 | static struct shash_alg sha256_avx2_algs[] = { { | |
254 | .digestsize = SHA256_DIGEST_SIZE, | |
255 | .init = sha256_base_init, | |
256 | .update = sha256_avx2_update, | |
257 | .final = sha256_avx2_final, | |
258 | .finup = sha256_avx2_finup, | |
259 | .descsize = sizeof(struct sha256_state), | |
260 | .base = { | |
261 | .cra_name = "sha256", | |
262 | .cra_driver_name = "sha256-avx2", | |
263 | .cra_priority = 170, | |
264 | .cra_flags = CRYPTO_ALG_TYPE_SHASH, | |
265 | .cra_blocksize = SHA256_BLOCK_SIZE, | |
266 | .cra_module = THIS_MODULE, | |
8275d1aa | 267 | } |
5dda42fc | 268 | }, { |
269 | .digestsize = SHA224_DIGEST_SIZE, | |
270 | .init = sha224_base_init, | |
271 | .update = sha256_avx2_update, | |
272 | .final = sha256_avx2_final, | |
273 | .finup = sha256_avx2_finup, | |
274 | .descsize = sizeof(struct sha256_state), | |
275 | .base = { | |
276 | .cra_name = "sha224", | |
277 | .cra_driver_name = "sha224-avx2", | |
278 | .cra_priority = 170, | |
279 | .cra_flags = CRYPTO_ALG_TYPE_SHASH, | |
280 | .cra_blocksize = SHA224_BLOCK_SIZE, | |
281 | .cra_module = THIS_MODULE, | |
282 | } | |
283 | } }; | |
8275d1aa | 284 | |
5dda42fc | 285 | static bool avx2_usable(void) |
286 | { | |
287 | if (avx_usable() && boot_cpu_has(X86_FEATURE_AVX2) && | |
288 | boot_cpu_has(X86_FEATURE_BMI2)) | |
289 | return true; | |
290 | ||
291 | return false; | |
292 | } | |
293 | ||
294 | static int register_sha256_avx2(void) | |
295 | { | |
296 | if (avx2_usable()) | |
297 | return crypto_register_shashes(sha256_avx2_algs, | |
298 | ARRAY_SIZE(sha256_avx2_algs)); | |
299 | return 0; | |
300 | } | |
301 | ||
302 | static void unregister_sha256_avx2(void) | |
303 | { | |
304 | if (avx2_usable()) | |
305 | crypto_unregister_shashes(sha256_avx2_algs, | |
306 | ARRAY_SIZE(sha256_avx2_algs)); | |
307 | } | |
308 | ||
309 | #else | |
310 | static inline int register_sha256_avx2(void) { return 0; } | |
311 | static inline void unregister_sha256_avx2(void) { } | |
8275d1aa | 312 | #endif |
5dda42fc | 313 | |
95fca7df | 314 | #ifdef CONFIG_AS_SHA256_NI |
5dda42fc | 315 | asmlinkage void sha256_ni_transform(u32 *digest, const char *data, |
316 | u64 rounds); /*unsigned int rounds);*/ | |
317 | ||
318 | static int sha256_ni_update(struct shash_desc *desc, const u8 *data, | |
319 | unsigned int len) | |
320 | { | |
321 | return sha256_update(desc, data, len, sha256_ni_transform); | |
322 | } | |
323 | ||
324 | static int sha256_ni_finup(struct shash_desc *desc, const u8 *data, | |
325 | unsigned int len, u8 *out) | |
326 | { | |
327 | return sha256_finup(desc, data, len, out, sha256_ni_transform); | |
328 | } | |
329 | ||
330 | static int sha256_ni_final(struct shash_desc *desc, u8 *out) | |
331 | { | |
332 | return sha256_ni_finup(desc, NULL, 0, out); | |
333 | } | |
334 | ||
335 | static struct shash_alg sha256_ni_algs[] = { { | |
336 | .digestsize = SHA256_DIGEST_SIZE, | |
337 | .init = sha256_base_init, | |
338 | .update = sha256_ni_update, | |
339 | .final = sha256_ni_final, | |
340 | .finup = sha256_ni_finup, | |
341 | .descsize = sizeof(struct sha256_state), | |
342 | .base = { | |
343 | .cra_name = "sha256", | |
344 | .cra_driver_name = "sha256-ni", | |
345 | .cra_priority = 250, | |
346 | .cra_flags = CRYPTO_ALG_TYPE_SHASH, | |
347 | .cra_blocksize = SHA256_BLOCK_SIZE, | |
348 | .cra_module = THIS_MODULE, | |
349 | } | |
350 | }, { | |
351 | .digestsize = SHA224_DIGEST_SIZE, | |
352 | .init = sha224_base_init, | |
353 | .update = sha256_ni_update, | |
354 | .final = sha256_ni_final, | |
355 | .finup = sha256_ni_finup, | |
356 | .descsize = sizeof(struct sha256_state), | |
357 | .base = { | |
358 | .cra_name = "sha224", | |
359 | .cra_driver_name = "sha224-ni", | |
360 | .cra_priority = 250, | |
361 | .cra_flags = CRYPTO_ALG_TYPE_SHASH, | |
362 | .cra_blocksize = SHA224_BLOCK_SIZE, | |
363 | .cra_module = THIS_MODULE, | |
8275d1aa | 364 | } |
5dda42fc | 365 | } }; |
366 | ||
367 | static int register_sha256_ni(void) | |
368 | { | |
369 | if (boot_cpu_has(X86_FEATURE_SHA_NI)) | |
370 | return crypto_register_shashes(sha256_ni_algs, | |
371 | ARRAY_SIZE(sha256_ni_algs)); | |
372 | return 0; | |
373 | } | |
374 | ||
375 | static void unregister_sha256_ni(void) | |
376 | { | |
377 | if (boot_cpu_has(X86_FEATURE_SHA_NI)) | |
378 | crypto_unregister_shashes(sha256_ni_algs, | |
379 | ARRAY_SIZE(sha256_ni_algs)); | |
380 | } | |
381 | ||
382 | #else | |
383 | static inline int register_sha256_ni(void) { return 0; } | |
384 | static inline void unregister_sha256_ni(void) { } | |
8275d1aa | 385 | #endif |
8275d1aa | 386 | |
5dda42fc | 387 | static int __init sha256_ssse3_mod_init(void) |
388 | { | |
389 | if (register_sha256_ssse3()) | |
390 | goto fail; | |
391 | ||
392 | if (register_sha256_avx()) { | |
393 | unregister_sha256_ssse3(); | |
394 | goto fail; | |
8275d1aa | 395 | } |
8275d1aa | 396 | |
5dda42fc | 397 | if (register_sha256_avx2()) { |
398 | unregister_sha256_avx(); | |
399 | unregister_sha256_ssse3(); | |
400 | goto fail; | |
401 | } | |
402 | ||
403 | if (register_sha256_ni()) { | |
404 | unregister_sha256_avx2(); | |
405 | unregister_sha256_avx(); | |
406 | unregister_sha256_ssse3(); | |
407 | goto fail; | |
408 | } | |
409 | ||
410 | return 0; | |
411 | fail: | |
8275d1aa TC |
412 | return -ENODEV; |
413 | } | |
414 | ||
415 | static void __exit sha256_ssse3_mod_fini(void) | |
416 | { | |
5dda42fc | 417 | unregister_sha256_ni(); |
418 | unregister_sha256_avx2(); | |
419 | unregister_sha256_avx(); | |
420 | unregister_sha256_ssse3(); | |
8275d1aa TC |
421 | } |
422 | ||
423 | module_init(sha256_ssse3_mod_init); | |
424 | module_exit(sha256_ssse3_mod_fini); | |
425 | ||
426 | MODULE_LICENSE("GPL"); | |
427 | MODULE_DESCRIPTION("SHA256 Secure Hash Algorithm, Supplemental SSE3 accelerated"); | |
428 | ||
5d26a105 KC |
429 | MODULE_ALIAS_CRYPTO("sha256"); |
430 | MODULE_ALIAS_CRYPTO("sha224"); |