Get rid of the shash boilerplate, and register the accelerated arm64 version of the CRC-T10DIF algorithm with the library interface instead.
Signed-off-by: Ard Biesheuvel <a...@kernel.org> --- arch/arm64/crypto/crct10dif-ce-glue.c | 85 ++++---------------- 1 file changed, 15 insertions(+), 70 deletions(-) diff --git a/arch/arm64/crypto/crct10dif-ce-glue.c b/arch/arm64/crypto/crct10dif-ce-glue.c index ccc3f6067742..70c730866d4a 100644 --- a/arch/arm64/crypto/crct10dif-ce-glue.c +++ b/arch/arm64/crypto/crct10dif-ce-glue.c @@ -12,7 +12,6 @@ #include <linux/module.h> #include <linux/string.h> -#include <crypto/internal/hash.h> #include <crypto/internal/simd.h> #include <asm/neon.h> @@ -23,97 +22,43 @@ asmlinkage u16 crc_t10dif_pmull_p8(u16 init_crc, const u8 *buf, size_t len); asmlinkage u16 crc_t10dif_pmull_p64(u16 init_crc, const u8 *buf, size_t len); -static int crct10dif_init(struct shash_desc *desc) +static u16 crct10dif_update_pmull_p8(u16 crc, const u8 *data, size_t length) { - u16 *crc = shash_desc_ctx(desc); - - *crc = 0; - return 0; -} - -static int crct10dif_update_pmull_p8(struct shash_desc *desc, const u8 *data, - unsigned int length) -{ - u16 *crc = shash_desc_ctx(desc); - if (length >= CRC_T10DIF_PMULL_CHUNK_SIZE && crypto_simd_usable()) { kernel_neon_begin(); - *crc = crc_t10dif_pmull_p8(*crc, data, length); + crc = crc_t10dif_pmull_p8(crc, data, length); kernel_neon_end(); } else { - *crc = crc_t10dif_generic(*crc, data, length); + crc = crc_t10dif_generic(crc, data, length); } - - return 0; + return crc; } -static int crct10dif_update_pmull_p64(struct shash_desc *desc, const u8 *data, - unsigned int length) +static u16 crct10dif_update_pmull_p64(u16 crc, const u8 *data, size_t length) { - u16 *crc = shash_desc_ctx(desc); - if (length >= CRC_T10DIF_PMULL_CHUNK_SIZE && crypto_simd_usable()) { kernel_neon_begin(); - *crc = crc_t10dif_pmull_p64(*crc, data, length); + crc = crc_t10dif_pmull_p64(crc, data, length); kernel_neon_end(); } else { - *crc = crc_t10dif_generic(*crc, data, length); + crc = crc_t10dif_generic(crc, data, length); } - - return 0; + return crc; } -static int crct10dif_final(struct shash_desc *desc, u8 *out) -{ - u16 *crc = shash_desc_ctx(desc); - - *(u16 *)out = *crc; - return 0; -} - -static struct shash_alg crc_t10dif_alg[] = {{ - .digestsize = CRC_T10DIF_DIGEST_SIZE, - .init = crct10dif_init, - .update = crct10dif_update_pmull_p8, - .final = crct10dif_final, - .descsize = CRC_T10DIF_DIGEST_SIZE, - - .base.cra_name = "crct10dif", - .base.cra_driver_name = "crct10dif-arm64-neon", - .base.cra_priority = 100, - .base.cra_blocksize = CRC_T10DIF_BLOCK_SIZE, - .base.cra_module = THIS_MODULE, -}, { - .digestsize = CRC_T10DIF_DIGEST_SIZE, - .init = crct10dif_init, - .update = crct10dif_update_pmull_p64, - .final = crct10dif_final, - .descsize = CRC_T10DIF_DIGEST_SIZE, - - .base.cra_name = "crct10dif", - .base.cra_driver_name = "crct10dif-arm64-ce", - .base.cra_priority = 200, - .base.cra_blocksize = CRC_T10DIF_BLOCK_SIZE, - .base.cra_module = THIS_MODULE, -}}; - static int __init crc_t10dif_mod_init(void) { if (cpu_have_named_feature(PMULL)) - return crypto_register_shashes(crc_t10dif_alg, - ARRAY_SIZE(crc_t10dif_alg)); + return crc_t10dif_register(crct10dif_update_pmull_p64, + "crct10dif-arm64-ce"); else - /* only register the first array element */ - return crypto_register_shash(crc_t10dif_alg); + return crc_t10dif_register(crct10dif_update_pmull_p8, + "crct10dif-arm64-neon"); } static void __exit crc_t10dif_mod_exit(void) { - if (cpu_have_named_feature(PMULL)) - crypto_unregister_shashes(crc_t10dif_alg, - ARRAY_SIZE(crc_t10dif_alg)); - else - crypto_unregister_shash(crc_t10dif_alg); + crc_t10dif_unregister(); } module_cpu_feature_match(ASIMD, crc_t10dif_mod_init); @@ -121,5 +66,5 @@ module_exit(crc_t10dif_mod_exit); MODULE_AUTHOR("Ard Biesheuvel <ard.biesheu...@linaro.org>"); MODULE_LICENSE("GPL v2"); -MODULE_ALIAS_CRYPTO("crct10dif"); -MODULE_ALIAS_CRYPTO("crct10dif-arm64-ce"); +MODULE_ALIAS("crct10dif-arch"); +MODULE_IMPORT_NS(CRYPTO_INTERNAL); -- 2.17.1