Lines Matching refs:AES_BLOCK_SIZE
54 u8 b[AES_BLOCK_SIZE];
177 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ecb_encrypt()
182 err = skcipher_walk_done(&walk, walk.nbytes % AES_BLOCK_SIZE); in ecb_encrypt()
197 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ecb_decrypt()
202 err = skcipher_walk_done(&walk, walk.nbytes % AES_BLOCK_SIZE); in ecb_decrypt()
215 while ((blocks = (walk->nbytes / AES_BLOCK_SIZE))) { in cbc_encrypt_walk()
221 err = skcipher_walk_done(walk, walk->nbytes % AES_BLOCK_SIZE); in cbc_encrypt_walk()
245 while ((blocks = (walk->nbytes / AES_BLOCK_SIZE))) { in cbc_decrypt_walk()
251 err = skcipher_walk_done(walk, walk->nbytes % AES_BLOCK_SIZE); in cbc_decrypt_walk()
271 int cbc_blocks = DIV_ROUND_UP(req->cryptlen, AES_BLOCK_SIZE) - 2; in cts_cbc_encrypt()
282 if (req->cryptlen <= AES_BLOCK_SIZE) { in cts_cbc_encrypt()
283 if (req->cryptlen < AES_BLOCK_SIZE) in cts_cbc_encrypt()
290 cbc_blocks * AES_BLOCK_SIZE, in cts_cbc_encrypt()
298 if (req->cryptlen == AES_BLOCK_SIZE) in cts_cbc_encrypt()
309 req->cryptlen - cbc_blocks * AES_BLOCK_SIZE, in cts_cbc_encrypt()
329 int cbc_blocks = DIV_ROUND_UP(req->cryptlen, AES_BLOCK_SIZE) - 2; in cts_cbc_decrypt()
340 if (req->cryptlen <= AES_BLOCK_SIZE) { in cts_cbc_decrypt()
341 if (req->cryptlen < AES_BLOCK_SIZE) in cts_cbc_decrypt()
348 cbc_blocks * AES_BLOCK_SIZE, in cts_cbc_decrypt()
356 if (req->cryptlen == AES_BLOCK_SIZE) in cts_cbc_decrypt()
367 req->cryptlen - cbc_blocks * AES_BLOCK_SIZE, in cts_cbc_decrypt()
392 while ((blocks = (walk.nbytes / AES_BLOCK_SIZE))) { in ctr_encrypt()
398 err = skcipher_walk_done(&walk, walk.nbytes % AES_BLOCK_SIZE); in ctr_encrypt()
401 u8 __aligned(8) tail[AES_BLOCK_SIZE]; in ctr_encrypt()
449 int tail = req->cryptlen % AES_BLOCK_SIZE; in xts_encrypt()
455 if (req->cryptlen < AES_BLOCK_SIZE) in xts_encrypt()
462 AES_BLOCK_SIZE) - 2; in xts_encrypt()
471 xts_blocks * AES_BLOCK_SIZE, in xts_encrypt()
479 for (first = 1; walk.nbytes >= AES_BLOCK_SIZE; first = 0) { in xts_encrypt()
483 nbytes &= ~(AES_BLOCK_SIZE - 1); in xts_encrypt()
500 skcipher_request_set_crypt(req, src, dst, AES_BLOCK_SIZE + tail, in xts_encrypt()
521 int tail = req->cryptlen % AES_BLOCK_SIZE; in xts_decrypt()
527 if (req->cryptlen < AES_BLOCK_SIZE) in xts_decrypt()
534 AES_BLOCK_SIZE) - 2; in xts_decrypt()
543 xts_blocks * AES_BLOCK_SIZE, in xts_decrypt()
551 for (first = 1; walk.nbytes >= AES_BLOCK_SIZE; first = 0) { in xts_decrypt()
555 nbytes &= ~(AES_BLOCK_SIZE - 1); in xts_decrypt()
572 skcipher_request_set_crypt(req, src, dst, AES_BLOCK_SIZE + tail, in xts_decrypt()
593 .base.cra_blocksize = AES_BLOCK_SIZE,
607 .base.cra_blocksize = AES_BLOCK_SIZE,
613 .ivsize = AES_BLOCK_SIZE,
622 .base.cra_blocksize = AES_BLOCK_SIZE,
628 .ivsize = AES_BLOCK_SIZE,
629 .walksize = 2 * AES_BLOCK_SIZE,
644 .ivsize = AES_BLOCK_SIZE,
645 .chunksize = AES_BLOCK_SIZE,
659 .ivsize = AES_BLOCK_SIZE,
660 .chunksize = AES_BLOCK_SIZE,
669 .base.cra_blocksize = AES_BLOCK_SIZE,
675 .ivsize = AES_BLOCK_SIZE,
676 .walksize = 2 * AES_BLOCK_SIZE,