/linux-6.1.9/lib/crypto/ |
D | aes.c | 199 ctx->key_enc[i] = get_unaligned_le32(in_key + i * sizeof(u32)); in aes_expandkey() 202 u32 *rki = ctx->key_enc + (i * kwords); in aes_expandkey() 231 ctx->key_dec[0] = ctx->key_enc[key_len + 24]; in aes_expandkey() 232 ctx->key_dec[1] = ctx->key_enc[key_len + 25]; in aes_expandkey() 233 ctx->key_dec[2] = ctx->key_enc[key_len + 26]; in aes_expandkey() 234 ctx->key_dec[3] = ctx->key_enc[key_len + 27]; in aes_expandkey() 237 ctx->key_dec[i] = inv_mix_columns(ctx->key_enc[j]); in aes_expandkey() 238 ctx->key_dec[i + 1] = inv_mix_columns(ctx->key_enc[j + 1]); in aes_expandkey() 239 ctx->key_dec[i + 2] = inv_mix_columns(ctx->key_enc[j + 2]); in aes_expandkey() 240 ctx->key_dec[i + 3] = inv_mix_columns(ctx->key_enc[j + 3]); in aes_expandkey() [all …]
|
/linux-6.1.9/arch/powerpc/crypto/ |
D | aes-spe-glue.c | 42 u32 key_enc[AES_MAX_KEYLENGTH_U32]; member 48 u32 key_enc[AES_MAX_KEYLENGTH_U32]; member 54 extern void ppc_encrypt_aes(u8 *out, const u8 *in, u32 *key_enc, u32 rounds); 56 extern void ppc_encrypt_ecb(u8 *out, const u8 *in, u32 *key_enc, u32 rounds, 60 extern void ppc_encrypt_cbc(u8 *out, const u8 *in, u32 *key_enc, u32 rounds, 64 extern void ppc_crypt_ctr (u8 *out, const u8 *in, u32 *key_enc, u32 rounds, 66 extern void ppc_encrypt_xts(u8 *out, const u8 *in, u32 *key_enc, u32 rounds, 71 extern void ppc_expand_key_128(u32 *key_enc, const u8 *key); 72 extern void ppc_expand_key_192(u32 *key_enc, const u8 *key); 73 extern void ppc_expand_key_256(u32 *key_enc, const u8 *key); [all …]
|
/linux-6.1.9/arch/arm64/crypto/ |
D | aes-ce-glue.c | 56 __aes_ce_encrypt(ctx->key_enc, dst, src, num_rounds(ctx)); in aes_cipher_encrypt() 85 struct aes_block *key_enc, *key_dec; in ce_aes_expandkey() local 95 ctx->key_enc[i] = get_unaligned_le32(in_key + i * sizeof(u32)); in ce_aes_expandkey() 99 u32 *rki = ctx->key_enc + (i * kwords); in ce_aes_expandkey() 128 key_enc = (struct aes_block *)ctx->key_enc; in ce_aes_expandkey() 132 key_dec[0] = key_enc[j]; in ce_aes_expandkey() 134 __aes_ce_invert(key_dec + i, key_enc + j); in ce_aes_expandkey() 135 key_dec[i] = key_enc[0]; in ce_aes_expandkey()
|
D | aes-ce-ccm-glue.c | 117 ctx->key_enc, num_rounds(ctx)); in ccm_calculate_auth_mac() 131 macp = ce_aes_ccm_auth_data(mac, p, n, macp, ctx->key_enc, in ccm_calculate_auth_mac() 179 walk.nbytes - tail, ctx->key_enc, in ccm_encrypt() 183 ce_aes_ccm_final(mac, buf, ctx->key_enc, num_rounds(ctx)); in ccm_encrypt() 237 walk.nbytes - tail, ctx->key_enc, in ccm_decrypt() 241 ce_aes_ccm_final(mac, buf, ctx->key_enc, num_rounds(ctx)); in ccm_decrypt()
|
D | aes-glue.c | 192 ctx->key_enc, rounds, blocks); in ecb_encrypt() 230 ctx->key_enc, rounds, blocks, walk->iv); in cbc_encrypt_walk() 328 ctx->key_enc, rounds, walk.nbytes, walk.iv); in cts_cbc_encrypt() 421 ctx->key1.key_enc, rounds, blocks, in essiv_cbc_encrypt() 422 req->iv, ctx->key2.key_enc); in essiv_cbc_encrypt() 444 req->iv, ctx->key2.key_enc); in essiv_cbc_decrypt() 482 aes_xctr_encrypt(dst, src, ctx->key_enc, rounds, nbytes, in xctr_encrypt() 527 aes_ctr_encrypt(dst, src, ctx->key_enc, rounds, nbytes, in ctr_encrypt() 584 ctx->key1.key_enc, rounds, nbytes, in xts_encrypt() 585 ctx->key2.key_enc, walk.iv, first); in xts_encrypt() [all …]
|
D | aes-neonbs-glue.c | 88 aesbs_convert_key(ctx->rk, rk.key_enc, ctx->rounds); in aesbs_setkey() 146 memcpy(ctx->enc, rk.key_enc, sizeof(ctx->enc)); in aesbs_cbc_ctr_setkey() 149 aesbs_convert_key(ctx->key.rk, rk.key_enc, ctx->key.rounds); in aesbs_cbc_ctr_setkey() 260 memcpy(ctx->twkey, rk.key_enc, sizeof(ctx->twkey)); in aesbs_xts_setkey() 327 neon_aes_xts_encrypt(out, in, ctx->cts.key_enc, in __xts_crypt() 361 neon_aes_xts_encrypt(out, in, ctx->cts.key_enc, ctx->key.rounds, in __xts_crypt()
|
D | aes-cipher-glue.c | 20 __aes_arm64_encrypt(ctx->key_enc, out, in, rounds); in aes_arm64_encrypt()
|
D | ghash-ce-glue.c | 382 dg, iv, ctx->aes_key.key_enc, nrounds, in gcm_encrypt() 454 dg, iv, ctx->aes_key.key_enc, in gcm_decrypt()
|
/linux-6.1.9/arch/arm/crypto/ |
D | aes-ce-glue.c | 80 struct aes_block *key_enc, *key_dec; in ce_aes_expandkey() local 90 ctx->key_enc[i] = get_unaligned_le32(in_key + i * sizeof(u32)); in ce_aes_expandkey() 94 u32 *rki = ctx->key_enc + (i * kwords); in ce_aes_expandkey() 124 key_enc = (struct aes_block *)ctx->key_enc; in ce_aes_expandkey() 128 key_dec[0] = key_enc[j]; in ce_aes_expandkey() 130 ce_aes_invert(key_dec + i, key_enc + j); in ce_aes_expandkey() 131 key_dec[i] = key_enc[0]; in ce_aes_expandkey() 180 ctx->key_enc, num_rounds(ctx), blocks); in ecb_encrypt() 218 ctx->key_enc, num_rounds(ctx), blocks, in cbc_encrypt_walk() 318 ctx->key_enc, num_rounds(ctx), walk.nbytes, in cts_cbc_encrypt() [all …]
|
D | aes-cipher-glue.c | 21 __aes_arm_encrypt(ctx->key_enc, rounds, in, out); in aes_arm_encrypt()
|
D | aes-neonbs-glue.c | 82 aesbs_convert_key(ctx->rk, rk.key_enc, ctx->rounds); in aesbs_setkey() 141 aesbs_convert_key(ctx->key.rk, rk.key_enc, ctx->key.rounds); in aesbs_cbc_setkey() 229 aesbs_convert_key(ctx->key.rk, ctx->fallback.key_enc, ctx->key.rounds); in aesbs_ctr_setkey_sync()
|
/linux-6.1.9/crypto/ |
D | aes_generic.c | 1180 const u32 *kp = ctx->key_enc + 4; in crypto_aes_encrypt() 1183 b0[0] = ctx->key_enc[0] ^ get_unaligned_le32(in); in crypto_aes_encrypt() 1184 b0[1] = ctx->key_enc[1] ^ get_unaligned_le32(in + 4); in crypto_aes_encrypt() 1185 b0[2] = ctx->key_enc[2] ^ get_unaligned_le32(in + 8); in crypto_aes_encrypt() 1186 b0[3] = ctx->key_enc[3] ^ get_unaligned_le32(in + 12); in crypto_aes_encrypt()
|
/linux-6.1.9/include/crypto/ |
D | aes.h | 26 u32 key_enc[AES_MAX_KEYLENGTH_U32]; member
|
/linux-6.1.9/drivers/crypto/inside-secure/ |
D | safexcel_cipher.c | 386 if (le32_to_cpu(ctx->key[i]) != aes.key_enc[i]) { in safexcel_skcipher_aes_setkey() 394 ctx->key[i] = cpu_to_le32(aes.key_enc[i]); in safexcel_skcipher_aes_setkey() 1449 if (le32_to_cpu(ctx->key[i]) != aes.key_enc[i]) { in safexcel_skcipher_aesctr_setkey() 1457 ctx->key[i] = cpu_to_le32(aes.key_enc[i]); in safexcel_skcipher_aesctr_setkey() 2487 if (le32_to_cpu(ctx->key[i]) != aes.key_enc[i]) { in safexcel_skcipher_aesxts_setkey() 2495 ctx->key[i] = cpu_to_le32(aes.key_enc[i]); in safexcel_skcipher_aesxts_setkey() 2505 aes.key_enc[i]) { in safexcel_skcipher_aesxts_setkey() 2514 cpu_to_le32(aes.key_enc[i]); in safexcel_skcipher_aesxts_setkey() 2596 if (le32_to_cpu(ctx->key[i]) != aes.key_enc[i]) { in safexcel_aead_gcm_setkey() 2604 ctx->key[i] = cpu_to_le32(aes.key_enc[i]); in safexcel_aead_gcm_setkey() [all …]
|
D | safexcel_hash.c | 2027 ctx->base.ipad.be[i + 8] = cpu_to_be32(aes.key_enc[i]); in safexcel_cbcmac_setkey() 2180 ctx->base.ipad.word[i + 8] = swab32(ctx->aes->key_enc[i]); in safexcel_cmac_setkey()
|
/linux-6.1.9/drivers/crypto/ |
D | sa2ul.c | 477 ctx.key_enc[52] = ctx.key_enc[51] ^ ctx.key_enc[46]; in sa_aes_inv_key() 478 ctx.key_enc[53] = ctx.key_enc[52] ^ ctx.key_enc[47]; in sa_aes_inv_key() 497 memcpy(inv_key, &ctx.key_enc[key_pos], key_sz); in sa_aes_inv_key()
|
D | padlock-aes.c | 150 memcpy(ctx->E, gen_aes.key_enc, AES_MAX_KEYLENGTH); in aes_set_key()
|
/linux-6.1.9/drivers/crypto/marvell/cesa/ |
D | cipher.c | 277 ctx->aes.key_dec[4 + i] = ctx->aes.key_enc[offset + i]; in mv_cesa_aes_setkey() 709 key = ctx->aes.key_enc; in mv_cesa_aes_op()
|
/linux-6.1.9/drivers/crypto/qat/qat_common/ |
D | qat_algs.c | 474 key = (u8 *)aes_expanded.key_enc + (AES_BLOCK_SIZE * nrounds); in qat_alg_xts_reverse_key() 479 key = (u8 *)aes_expanded.key_enc + (AES_BLOCK_SIZE * nrounds); in qat_alg_xts_reverse_key()
|
/linux-6.1.9/drivers/crypto/aspeed/ |
D | aspeed-hace-crypto.c | 626 memcpy(ctx->key, gen_aes_key.key_enc, AES_MAX_KEYLENGTH); in aspeed_aes_setkey()
|