/* SPDX-License-Identifier: GPL-2.0-or-later */ /* * AES optimized using the CP Assist for Cryptographic Functions (CPACF) * * Copyright 2026 Google LLC */ #include #include static __ro_after_init DEFINE_STATIC_KEY_FALSE(have_cpacf_aes128); static __ro_after_init DEFINE_STATIC_KEY_FALSE(have_cpacf_aes192); static __ro_after_init DEFINE_STATIC_KEY_FALSE(have_cpacf_aes256); /* * When the CPU supports CPACF AES for the requested key length, we need only * save a copy of the raw AES key, as that's what the CPACF instructions need. * * When unsupported, fall back to the generic key expansion and en/decryption. */ static void aes_preparekey_arch(union aes_enckey_arch *k, union aes_invkey_arch *inv_k, const u8 *in_key, int key_len, int nrounds) { if (key_len == AES_KEYSIZE_128) { if (static_branch_likely(&have_cpacf_aes128)) { memcpy(k->raw_key, in_key, AES_KEYSIZE_128); return; } } else if (key_len == AES_KEYSIZE_192) { if (static_branch_likely(&have_cpacf_aes192)) { memcpy(k->raw_key, in_key, AES_KEYSIZE_192); return; } } else { if (static_branch_likely(&have_cpacf_aes256)) { memcpy(k->raw_key, in_key, AES_KEYSIZE_256); return; } } aes_expandkey_generic(k->rndkeys, inv_k ? inv_k->inv_rndkeys : NULL, in_key, key_len); } static inline bool aes_crypt_s390(const struct aes_enckey *key, u8 out[AES_BLOCK_SIZE], const u8 in[AES_BLOCK_SIZE], int decrypt) { if (key->len == AES_KEYSIZE_128) { if (static_branch_likely(&have_cpacf_aes128)) { cpacf_km(CPACF_KM_AES_128 | decrypt, (void *)key->k.raw_key, out, in, AES_BLOCK_SIZE); return true; } } else if (key->len == AES_KEYSIZE_192) { if (static_branch_likely(&have_cpacf_aes192)) { cpacf_km(CPACF_KM_AES_192 | decrypt, (void *)key->k.raw_key, out, in, AES_BLOCK_SIZE); return true; } } else { if (static_branch_likely(&have_cpacf_aes256)) { cpacf_km(CPACF_KM_AES_256 | decrypt, (void *)key->k.raw_key, out, in, AES_BLOCK_SIZE); return true; } } return false; } static void aes_encrypt_arch(const struct aes_enckey *key, u8 out[AES_BLOCK_SIZE], const u8 in[AES_BLOCK_SIZE]) { if (likely(aes_crypt_s390(key, out, in, 0))) return; aes_encrypt_generic(key->k.rndkeys, key->nrounds, out, in); } static void aes_decrypt_arch(const struct aes_key *key, u8 out[AES_BLOCK_SIZE], const u8 in[AES_BLOCK_SIZE]) { if (likely(aes_crypt_s390((const struct aes_enckey *)key, out, in, CPACF_DECRYPT))) return; aes_decrypt_generic(key->inv_k.inv_rndkeys, key->nrounds, out, in); } #define aes_mod_init_arch aes_mod_init_arch static void aes_mod_init_arch(void) { if (cpu_have_feature(S390_CPU_FEATURE_MSA)) { cpacf_mask_t km_functions; cpacf_query(CPACF_KM, &km_functions); if (cpacf_test_func(&km_functions, CPACF_KM_AES_128)) static_branch_enable(&have_cpacf_aes128); if (cpacf_test_func(&km_functions, CPACF_KM_AES_192)) static_branch_enable(&have_cpacf_aes192); if (cpacf_test_func(&km_functions, CPACF_KM_AES_256)) static_branch_enable(&have_cpacf_aes256); } }