/linux-5.19.10/arch/x86/crypto/ |
D | sm4_aesni_avx_glue.c | 151 u8 keystream[SM4_BLOCK_SIZE * 8]; in sm4_avx_cbc_decrypt() local 156 sm4_aesni_avx_crypt8(ctx->rkey_dec, keystream, in sm4_avx_cbc_decrypt() 165 &keystream[i * SM4_BLOCK_SIZE], in sm4_avx_cbc_decrypt() 170 crypto_xor_cpy(dst, walk.iv, keystream, SM4_BLOCK_SIZE); in sm4_avx_cbc_decrypt() 202 u8 keystream[SM4_BLOCK_SIZE]; in sm4_cfb_encrypt() local 208 sm4_crypt_block(ctx->rkey_enc, keystream, iv); in sm4_cfb_encrypt() 209 crypto_xor_cpy(dst, src, keystream, SM4_BLOCK_SIZE); in sm4_cfb_encrypt() 220 sm4_crypt_block(ctx->rkey_enc, keystream, walk.iv); in sm4_cfb_encrypt() 221 crypto_xor_cpy(dst, src, keystream, nbytes); in sm4_cfb_encrypt() 257 u8 keystream[SM4_BLOCK_SIZE * 8]; in sm4_avx_cfb_decrypt() local [all …]
|
D | aesni-intel_glue.c | 500 u8 keystream[AES_BLOCK_SIZE]; in ctr_crypt() local 517 aesni_enc(ctx, keystream, walk.iv); in ctr_crypt() 520 keystream, nbytes); in ctr_crypt()
|
/linux-5.19.10/arch/arm64/crypto/ |
D | sm4-neon-glue.c | 156 u8 keystream[SM4_BLOCK_SIZE * 8]; in sm4_cbc_decrypt() local 160 sm4_neon_crypt_blk1_8(ctx->rkey_dec, keystream, in sm4_cbc_decrypt() 169 &keystream[i * SM4_BLOCK_SIZE], in sm4_cbc_decrypt() 175 keystream, SM4_BLOCK_SIZE); in sm4_cbc_decrypt() 199 u8 keystream[SM4_BLOCK_SIZE]; in sm4_cfb_encrypt() local 205 sm4_crypt_block(ctx->rkey_enc, keystream, iv); in sm4_cfb_encrypt() 206 crypto_xor_cpy(dst, src, keystream, SM4_BLOCK_SIZE); in sm4_cfb_encrypt() 217 sm4_crypt_block(ctx->rkey_enc, keystream, walk.iv); in sm4_cfb_encrypt() 218 crypto_xor_cpy(dst, src, keystream, nbytes); in sm4_cfb_encrypt() 256 u8 keystream[SM4_BLOCK_SIZE * 8]; in sm4_cfb_decrypt() local [all …]
|
D | sm4-ce-glue.c | 186 u8 keystream[SM4_BLOCK_SIZE]; in sm4_cfb_encrypt() local 188 sm4_ce_crypt_block(ctx->rkey_enc, keystream, walk.iv); in sm4_cfb_encrypt() 189 crypto_xor_cpy(dst, src, keystream, nbytes); in sm4_cfb_encrypt() 228 u8 keystream[SM4_BLOCK_SIZE]; in sm4_cfb_decrypt() local 230 sm4_ce_crypt_block(ctx->rkey_enc, keystream, walk.iv); in sm4_cfb_decrypt() 231 crypto_xor_cpy(dst, src, keystream, nbytes); in sm4_cfb_decrypt() 270 u8 keystream[SM4_BLOCK_SIZE]; in sm4_ctr_crypt() local 272 sm4_ce_crypt_block(ctx->rkey_enc, keystream, walk.iv); in sm4_ctr_crypt() 274 crypto_xor_cpy(dst, src, keystream, nbytes); in sm4_ctr_crypt()
|
/linux-5.19.10/crypto/ |
D | ctr.c | 35 u8 *keystream = PTR_ALIGN(tmp + 0, alignmask + 1); in crypto_ctr_crypt_final() local 40 crypto_cipher_encrypt_one(tfm, keystream, ctrblk); in crypto_ctr_crypt_final() 41 crypto_xor_cpy(dst, keystream, src, nbytes); in crypto_ctr_crypt_final() 83 u8 *keystream = PTR_ALIGN(tmp + 0, alignmask + 1); in crypto_ctr_crypt_inplace() local 87 fn(crypto_cipher_tfm(tfm), keystream, ctrblk); in crypto_ctr_crypt_inplace() 88 crypto_xor(src, keystream, bsize); in crypto_ctr_crypt_inplace()
|
D | Kconfig | 450 stream cipher. It generates keystream blocks, which are then XORed
|
/linux-5.19.10/drivers/crypto/vmx/ |
D | aes_ctr.c | 73 u8 keystream[AES_BLOCK_SIZE]; in p8_aes_ctr_final() local 81 aes_p8_encrypt(ctrblk, keystream, &ctx->enc_key); in p8_aes_ctr_final() 86 crypto_xor_cpy(dst, keystream, src, nbytes); in p8_aes_ctr_final()
|
/linux-5.19.10/arch/sparc/crypto/ |
D | aes_glue.c | 323 u64 keystream[AES_BLOCK_SIZE / sizeof(u64)]; in ctr_crypt_final() local 329 keystream, AES_BLOCK_SIZE); in ctr_crypt_final() 330 crypto_xor_cpy(dst, (u8 *) keystream, src, nbytes); in ctr_crypt_final()
|