Lines Matching refs:AES_BLOCK_SIZE
192 nbytes = round_down(nbytes, AES_BLOCK_SIZE); in ppc_ecb_crypt()
231 nbytes = round_down(nbytes, AES_BLOCK_SIZE); in ppc_cbc_crypt()
273 nbytes = round_down(nbytes, AES_BLOCK_SIZE); in ppc_ctr_crypt()
300 nbytes = round_down(nbytes, AES_BLOCK_SIZE); in ppc_xts_crypt()
324 int tail = req->cryptlen % AES_BLOCK_SIZE; in ppc_xts_encrypt()
325 int offset = req->cryptlen - tail - AES_BLOCK_SIZE; in ppc_xts_encrypt()
327 u8 b[2][AES_BLOCK_SIZE]; in ppc_xts_encrypt()
330 if (req->cryptlen < AES_BLOCK_SIZE) in ppc_xts_encrypt()
344 scatterwalk_map_and_copy(b[0], req->dst, offset, AES_BLOCK_SIZE, 0); in ppc_xts_encrypt()
346 scatterwalk_map_and_copy(b[0], req->src, offset + AES_BLOCK_SIZE, tail, 0); in ppc_xts_encrypt()
349 ppc_encrypt_xts(b[0], b[0], ctx->key_enc, ctx->rounds, AES_BLOCK_SIZE, in ppc_xts_encrypt()
353 scatterwalk_map_and_copy(b[0], req->dst, offset, AES_BLOCK_SIZE + tail, 1); in ppc_xts_encrypt()
362 int tail = req->cryptlen % AES_BLOCK_SIZE; in ppc_xts_decrypt()
363 int offset = req->cryptlen - tail - AES_BLOCK_SIZE; in ppc_xts_decrypt()
365 u8 b[3][AES_BLOCK_SIZE]; in ppc_xts_decrypt()
369 if (req->cryptlen < AES_BLOCK_SIZE) in ppc_xts_decrypt()
383 scatterwalk_map_and_copy(b[1], req->src, offset, AES_BLOCK_SIZE + tail, 0); in ppc_xts_decrypt()
388 AES_BLOCK_SIZE); in ppc_xts_decrypt()
392 ppc_decrypt_xts(b[1], b[1], ctx->key_dec, ctx->rounds, AES_BLOCK_SIZE, in ppc_xts_decrypt()
395 memcpy(b[0] + tail, b[1] + tail, AES_BLOCK_SIZE - tail); in ppc_xts_decrypt()
396 ppc_decrypt_xts(b[0], b[0], ctx->key_dec, ctx->rounds, AES_BLOCK_SIZE, in ppc_xts_decrypt()
400 scatterwalk_map_and_copy(b[0], req->dst, offset, AES_BLOCK_SIZE + tail, 1); in ppc_xts_decrypt()
418 .cra_blocksize = AES_BLOCK_SIZE,
438 .base.cra_blocksize = AES_BLOCK_SIZE,
450 .base.cra_blocksize = AES_BLOCK_SIZE,
455 .ivsize = AES_BLOCK_SIZE,
468 .ivsize = AES_BLOCK_SIZE,
472 .chunksize = AES_BLOCK_SIZE,
477 .base.cra_blocksize = AES_BLOCK_SIZE,
482 .ivsize = AES_BLOCK_SIZE,