/linux-6.1.9/drivers/crypto/allwinner/sun4i-ss/ |
D | sun4i-ss-cipher.c | 15 static int noinline_for_stack sun4i_ss_opti_poll(struct skcipher_request *areq) in sun4i_ss_opti_poll() argument 17 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); in sun4i_ss_opti_poll() 21 struct sun4i_cipher_req_ctx *ctx = skcipher_request_ctx(areq); in sun4i_ss_opti_poll() 30 unsigned int ileft = areq->cryptlen; in sun4i_ss_opti_poll() 31 unsigned int oleft = areq->cryptlen; in sun4i_ss_opti_poll() 41 if (!areq->cryptlen) in sun4i_ss_opti_poll() 44 if (!areq->src || !areq->dst) { in sun4i_ss_opti_poll() 49 if (areq->iv && ivsize > 0 && mode & SS_DECRYPTION) { in sun4i_ss_opti_poll() 50 scatterwalk_map_and_copy(ctx->backup_iv, areq->src, in sun4i_ss_opti_poll() 51 areq->cryptlen - ivsize, ivsize, 0); in sun4i_ss_opti_poll() [all …]
|
D | sun4i-ss-hash.c | 47 int sun4i_hash_init(struct ahash_request *areq) in sun4i_hash_init() argument 49 struct sun4i_req_ctx *op = ahash_request_ctx(areq); in sun4i_hash_init() 50 struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq); in sun4i_hash_init() 62 int sun4i_hash_export_md5(struct ahash_request *areq, void *out) in sun4i_hash_export_md5() argument 64 struct sun4i_req_ctx *op = ahash_request_ctx(areq); in sun4i_hash_export_md5() 85 int sun4i_hash_import_md5(struct ahash_request *areq, const void *in) in sun4i_hash_import_md5() argument 87 struct sun4i_req_ctx *op = ahash_request_ctx(areq); in sun4i_hash_import_md5() 91 sun4i_hash_init(areq); in sun4i_hash_import_md5() 104 int sun4i_hash_export_sha1(struct ahash_request *areq, void *out) in sun4i_hash_export_sha1() argument 106 struct sun4i_req_ctx *op = ahash_request_ctx(areq); in sun4i_hash_export_sha1() [all …]
|
D | sun4i-ss.h | 201 int sun4i_hash_init(struct ahash_request *areq); 202 int sun4i_hash_update(struct ahash_request *areq); 203 int sun4i_hash_final(struct ahash_request *areq); 204 int sun4i_hash_finup(struct ahash_request *areq); 205 int sun4i_hash_digest(struct ahash_request *areq); 206 int sun4i_hash_export_md5(struct ahash_request *areq, void *out); 207 int sun4i_hash_import_md5(struct ahash_request *areq, const void *in); 208 int sun4i_hash_export_sha1(struct ahash_request *areq, void *out); 209 int sun4i_hash_import_sha1(struct ahash_request *areq, const void *in); 211 int sun4i_ss_cbc_aes_encrypt(struct skcipher_request *areq); [all …]
|
/linux-6.1.9/drivers/crypto/allwinner/sun8i-ce/ |
D | sun8i-ce-cipher.c | 24 static int sun8i_ce_cipher_need_fallback(struct skcipher_request *areq) in sun8i_ce_cipher_need_fallback() argument 26 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); in sun8i_ce_cipher_need_fallback() 34 if (sg_nents_for_len(areq->src, areq->cryptlen) > MAX_SG || in sun8i_ce_cipher_need_fallback() 35 sg_nents_for_len(areq->dst, areq->cryptlen) > MAX_SG) { in sun8i_ce_cipher_need_fallback() 40 if (areq->cryptlen < crypto_skcipher_ivsize(tfm)) { in sun8i_ce_cipher_need_fallback() 45 if (areq->cryptlen == 0) { in sun8i_ce_cipher_need_fallback() 50 if (areq->cryptlen % 16) { in sun8i_ce_cipher_need_fallback() 55 len = areq->cryptlen; in sun8i_ce_cipher_need_fallback() 56 sg = areq->src; in sun8i_ce_cipher_need_fallback() 71 len = areq->cryptlen; in sun8i_ce_cipher_need_fallback() [all …]
|
D | sun8i-ce-hash.c | 74 int sun8i_ce_hash_init(struct ahash_request *areq) in sun8i_ce_hash_init() argument 76 struct sun8i_ce_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ce_hash_init() 77 struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq); in sun8i_ce_hash_init() 83 rctx->fallback_req.base.flags = areq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP; in sun8i_ce_hash_init() 88 int sun8i_ce_hash_export(struct ahash_request *areq, void *out) in sun8i_ce_hash_export() argument 90 struct sun8i_ce_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ce_hash_export() 91 struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq); in sun8i_ce_hash_export() 95 rctx->fallback_req.base.flags = areq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP; in sun8i_ce_hash_export() 100 int sun8i_ce_hash_import(struct ahash_request *areq, const void *in) in sun8i_ce_hash_import() argument 102 struct sun8i_ce_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ce_hash_import() [all …]
|
D | sun8i-ce.h | 350 int sun8i_ce_enqueue(struct crypto_async_request *areq, u32 type); 358 int sun8i_ce_skdecrypt(struct skcipher_request *areq); 359 int sun8i_ce_skencrypt(struct skcipher_request *areq); 367 int sun8i_ce_hash_init(struct ahash_request *areq); 368 int sun8i_ce_hash_export(struct ahash_request *areq, void *out); 369 int sun8i_ce_hash_import(struct ahash_request *areq, const void *in); 370 int sun8i_ce_hash(struct ahash_request *areq); 371 int sun8i_ce_hash_final(struct ahash_request *areq); 372 int sun8i_ce_hash_update(struct ahash_request *areq); 373 int sun8i_ce_hash_finup(struct ahash_request *areq); [all …]
|
/linux-6.1.9/drivers/crypto/gemini/ |
D | sl3516-ce-cipher.c | 21 static bool sl3516_ce_need_fallback(struct skcipher_request *areq) in sl3516_ce_need_fallback() argument 23 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); in sl3516_ce_need_fallback() 30 if (areq->cryptlen == 0 || areq->cryptlen % 16) { in sl3516_ce_need_fallback() 39 if (sg_nents(areq->src) > MAXDESC / 2) { in sl3516_ce_need_fallback() 44 if (sg_nents(areq->dst) > MAXDESC) { in sl3516_ce_need_fallback() 49 sg = areq->src; in sl3516_ce_need_fallback() 65 sg = areq->dst; in sl3516_ce_need_fallback() 83 in_sg = areq->src; in sl3516_ce_need_fallback() 84 out_sg = areq->dst; in sl3516_ce_need_fallback() 99 static int sl3516_ce_cipher_fallback(struct skcipher_request *areq) in sl3516_ce_cipher_fallback() argument [all …]
|
/linux-6.1.9/drivers/crypto/amlogic/ |
D | amlogic-gxl-cipher.c | 25 static bool meson_cipher_need_fallback(struct skcipher_request *areq) in meson_cipher_need_fallback() argument 27 struct scatterlist *src_sg = areq->src; in meson_cipher_need_fallback() 28 struct scatterlist *dst_sg = areq->dst; in meson_cipher_need_fallback() 30 if (areq->cryptlen == 0) in meson_cipher_need_fallback() 58 static int meson_cipher_do_fallback(struct skcipher_request *areq) in meson_cipher_do_fallback() argument 60 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); in meson_cipher_do_fallback() 62 struct meson_cipher_req_ctx *rctx = skcipher_request_ctx(areq); in meson_cipher_do_fallback() 72 skcipher_request_set_callback(&rctx->fallback_req, areq->base.flags, in meson_cipher_do_fallback() 73 areq->base.complete, areq->base.data); in meson_cipher_do_fallback() 74 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in meson_cipher_do_fallback() [all …]
|
/linux-6.1.9/drivers/crypto/allwinner/sun8i-ss/ |
D | sun8i-ss-cipher.c | 23 static bool sun8i_ss_need_fallback(struct skcipher_request *areq) in sun8i_ss_need_fallback() argument 25 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); in sun8i_ss_need_fallback() 28 struct scatterlist *in_sg = areq->src; in sun8i_ss_need_fallback() 29 struct scatterlist *out_sg = areq->dst; in sun8i_ss_need_fallback() 33 if (areq->cryptlen == 0 || areq->cryptlen % 16) { in sun8i_ss_need_fallback() 38 if (sg_nents_for_len(areq->src, areq->cryptlen) > 8 || in sun8i_ss_need_fallback() 39 sg_nents_for_len(areq->dst, areq->cryptlen) > 8) { in sun8i_ss_need_fallback() 44 len = areq->cryptlen; in sun8i_ss_need_fallback() 45 sg = areq->src; in sun8i_ss_need_fallback() 59 len = areq->cryptlen; in sun8i_ss_need_fallback() [all …]
|
D | sun8i-ss-hash.c | 164 int sun8i_ss_hash_init(struct ahash_request *areq) in sun8i_ss_hash_init() argument 166 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ss_hash_init() 167 struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq); in sun8i_ss_hash_init() 173 rctx->fallback_req.base.flags = areq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP; in sun8i_ss_hash_init() 178 int sun8i_ss_hash_export(struct ahash_request *areq, void *out) in sun8i_ss_hash_export() argument 180 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ss_hash_export() 181 struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq); in sun8i_ss_hash_export() 185 rctx->fallback_req.base.flags = areq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP; in sun8i_ss_hash_export() 190 int sun8i_ss_hash_import(struct ahash_request *areq, const void *in) in sun8i_ss_hash_import() argument 192 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ss_hash_import() [all …]
|
D | sun8i-ss.h | 296 int sun8i_ss_enqueue(struct crypto_async_request *areq, u32 type); 304 int sun8i_ss_skdecrypt(struct skcipher_request *areq); 305 int sun8i_ss_skencrypt(struct skcipher_request *areq); 318 int sun8i_ss_hash_init(struct ahash_request *areq); 319 int sun8i_ss_hash_export(struct ahash_request *areq, void *out); 320 int sun8i_ss_hash_import(struct ahash_request *areq, const void *in); 321 int sun8i_ss_hash_final(struct ahash_request *areq); 322 int sun8i_ss_hash_update(struct ahash_request *areq); 323 int sun8i_ss_hash_finup(struct ahash_request *areq); 324 int sun8i_ss_hash_digest(struct ahash_request *areq);
|
/linux-6.1.9/drivers/crypto/cavium/nitrox/ |
D | nitrox_aead.c | 192 struct aead_request *areq = arg; in nitrox_aead_callback() local 193 struct nitrox_aead_rctx *rctx = aead_request_ctx(areq); in nitrox_aead_callback() 202 areq->base.complete(&areq->base, err); in nitrox_aead_callback() 213 static int nitrox_aes_gcm_enc(struct aead_request *areq) in nitrox_aes_gcm_enc() argument 215 struct crypto_aead *aead = crypto_aead_reqtfm(areq); in nitrox_aes_gcm_enc() 217 struct nitrox_aead_rctx *rctx = aead_request_ctx(areq); in nitrox_aes_gcm_enc() 222 if (!nitrox_aes_gcm_assoclen_supported(areq->assoclen)) in nitrox_aes_gcm_enc() 225 memcpy(fctx->crypto.iv, areq->iv, GCM_AES_SALT_SIZE); in nitrox_aes_gcm_enc() 227 rctx->cryptlen = areq->cryptlen; in nitrox_aes_gcm_enc() 228 rctx->assoclen = areq->assoclen; in nitrox_aes_gcm_enc() [all …]
|
/linux-6.1.9/drivers/crypto/inside-secure/ |
D | safexcel_hash.c | 224 static int safexcel_ahash_enqueue(struct ahash_request *areq); 232 struct ahash_request *areq = ahash_request_cast(async); in safexcel_handle_req_result() local 233 struct crypto_ahash *ahash = crypto_ahash_reqtfm(areq); in safexcel_handle_req_result() 234 struct safexcel_ahash_req *sreq = ahash_request_ctx(areq); in safexcel_handle_req_result() 252 dma_unmap_sg(priv->dev, areq->src, sreq->nents, DMA_TO_DEVICE); in safexcel_handle_req_result() 285 areq->nbytes = 0; in safexcel_handle_req_result() 286 safexcel_ahash_enqueue(areq); in safexcel_handle_req_result() 295 *(__le32 *)areq->result = ~sreq->state[0]; in safexcel_handle_req_result() 297 memcpy(areq->result, sreq->state, in safexcel_handle_req_result() 314 struct ahash_request *areq = ahash_request_cast(async); in safexcel_ahash_send_req() local [all …]
|
/linux-6.1.9/crypto/ |
D | algif_skcipher.c | 60 struct af_alg_async_req *areq; in _skcipher_recvmsg() local 71 areq = af_alg_alloc_areq(sk, sizeof(struct af_alg_async_req) + in _skcipher_recvmsg() 73 if (IS_ERR(areq)) in _skcipher_recvmsg() 74 return PTR_ERR(areq); in _skcipher_recvmsg() 77 err = af_alg_get_rsgl(sk, msg, flags, areq, ctx->used, &len); in _skcipher_recvmsg() 92 areq->tsgl_entries = af_alg_count_tsgl(sk, len, 0); in _skcipher_recvmsg() 93 if (!areq->tsgl_entries) in _skcipher_recvmsg() 94 areq->tsgl_entries = 1; in _skcipher_recvmsg() 95 areq->tsgl = sock_kmalloc(sk, array_size(sizeof(*areq->tsgl), in _skcipher_recvmsg() 96 areq->tsgl_entries), in _skcipher_recvmsg() [all …]
|
D | algif_aead.c | 100 struct af_alg_async_req *areq; in _aead_recvmsg() local 153 areq = af_alg_alloc_areq(sk, sizeof(struct af_alg_async_req) + in _aead_recvmsg() 155 if (IS_ERR(areq)) in _aead_recvmsg() 156 return PTR_ERR(areq); in _aead_recvmsg() 159 err = af_alg_get_rsgl(sk, msg, flags, areq, outlen, &usedpages); in _aead_recvmsg() 213 rsgl_src = areq->first_rsgl.sgl.sg; in _aead_recvmsg() 227 areq->first_rsgl.sgl.sg, processed); in _aead_recvmsg() 245 areq->first_rsgl.sgl.sg, outlen); in _aead_recvmsg() 250 areq->tsgl_entries = af_alg_count_tsgl(sk, processed, in _aead_recvmsg() 252 if (!areq->tsgl_entries) in _aead_recvmsg() [all …]
|
D | af_alg.c | 659 static void af_alg_free_areq_sgls(struct af_alg_async_req *areq) in af_alg_free_areq_sgls() argument 661 struct sock *sk = areq->sk; in af_alg_free_areq_sgls() 669 list_for_each_entry_safe(rsgl, tmp, &areq->rsgl_list, list) { in af_alg_free_areq_sgls() 673 if (rsgl != &areq->first_rsgl) in af_alg_free_areq_sgls() 677 tsgl = areq->tsgl; in af_alg_free_areq_sgls() 679 for_each_sg(tsgl, sg, areq->tsgl_entries, i) { in af_alg_free_areq_sgls() 685 sock_kfree_s(sk, tsgl, areq->tsgl_entries * sizeof(*tsgl)); in af_alg_free_areq_sgls() 1045 void af_alg_free_resources(struct af_alg_async_req *areq) in af_alg_free_resources() argument 1047 struct sock *sk = areq->sk; in af_alg_free_resources() 1049 af_alg_free_areq_sgls(areq); in af_alg_free_resources() [all …]
|
D | chacha20poly1305.c | 118 static void chacha_decrypt_done(struct crypto_async_request *areq, int err) in chacha_decrypt_done() argument 120 async_done_continue(areq->data, err, poly_verify_tag); in chacha_decrypt_done() 164 static void poly_tail_done(struct crypto_async_request *areq, int err) in poly_tail_done() argument 166 async_done_continue(areq->data, err, poly_tail_continue); in poly_tail_done() 194 static void poly_cipherpad_done(struct crypto_async_request *areq, int err) in poly_cipherpad_done() argument 196 async_done_continue(areq->data, err, poly_tail); in poly_cipherpad_done() 223 static void poly_cipher_done(struct crypto_async_request *areq, int err) in poly_cipher_done() argument 225 async_done_continue(areq->data, err, poly_cipherpad); in poly_cipher_done() 253 static void poly_adpad_done(struct crypto_async_request *areq, int err) in poly_adpad_done() argument 255 async_done_continue(areq->data, err, poly_cipher); in poly_adpad_done() [all …]
|
D | ahash.c | 287 struct ahash_request *areq = req->data; in ahash_op_unaligned_done() local 290 ahash_notify_einprogress(areq); in ahash_op_unaligned_done() 304 ahash_restore_req(areq, err); in ahash_op_unaligned_done() 307 areq->base.complete(&areq->base, err); in ahash_op_unaligned_done() 387 struct ahash_request *areq = req->data; in ahash_def_finup_done2() local 392 ahash_restore_req(areq, err); in ahash_def_finup_done2() 394 areq->base.complete(&areq->base, err); in ahash_def_finup_done2() 415 struct ahash_request *areq = req->data; in ahash_def_finup_done1() local 418 ahash_notify_einprogress(areq); in ahash_def_finup_done1() 422 areq->base.flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP; in ahash_def_finup_done1() [all …]
|
/linux-6.1.9/drivers/crypto/ |
D | talitos.c | 989 struct aead_request *areq, bool encrypt) in ipsec_esp_unmap() argument 991 struct crypto_aead *aead = crypto_aead_reqtfm(areq); in ipsec_esp_unmap() 995 unsigned int cryptlen = areq->cryptlen - (encrypt ? 0 : authsize); in ipsec_esp_unmap() 1004 talitos_sg_unmap(dev, edesc, areq->src, areq->dst, in ipsec_esp_unmap() 1005 cryptlen + authsize, areq->assoclen); in ipsec_esp_unmap() 1014 sg_pcopy_to_buffer(areq->dst, dst_nents, ctx->iv, ivsize, in ipsec_esp_unmap() 1015 areq->assoclen + cryptlen - ivsize); in ipsec_esp_unmap() 1026 struct aead_request *areq = context; in ipsec_esp_encrypt_done() local 1027 struct crypto_aead *authenc = crypto_aead_reqtfm(areq); in ipsec_esp_encrypt_done() 1033 ipsec_esp_unmap(dev, edesc, areq, true); in ipsec_esp_encrypt_done() [all …]
|
D | atmel-i2c.h | 163 void (*cbk)(struct atmel_i2c_work_data *work_data, void *areq, 165 void *areq; member 174 void *areq, int status), 175 void *areq);
|
/linux-6.1.9/drivers/crypto/rockchip/ |
D | rk3288_crypto_ahash.c | 37 static int rk_ahash_digest_fb(struct ahash_request *areq) in rk_ahash_digest_fb() argument 39 struct rk_ahash_rctx *rctx = ahash_request_ctx(areq); in rk_ahash_digest_fb() 40 struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq); in rk_ahash_digest_fb() 44 rctx->fallback_req.base.flags = areq->base.flags & in rk_ahash_digest_fb() 47 rctx->fallback_req.nbytes = areq->nbytes; in rk_ahash_digest_fb() 48 rctx->fallback_req.src = areq->src; in rk_ahash_digest_fb() 49 rctx->fallback_req.result = areq->result; in rk_ahash_digest_fb() 220 struct ahash_request *areq = container_of(breq, struct ahash_request, base); in rk_hash_prepare() local 221 struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq); in rk_hash_prepare() 222 struct rk_ahash_rctx *rctx = ahash_request_ctx(areq); in rk_hash_prepare() [all …]
|
D | rk3288_crypto_skcipher.c | 55 static int rk_cipher_fallback(struct skcipher_request *areq) in rk_cipher_fallback() argument 57 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); in rk_cipher_fallback() 59 struct rk_cipher_rctx *rctx = skcipher_request_ctx(areq); in rk_cipher_fallback() 63 skcipher_request_set_callback(&rctx->fallback_req, areq->base.flags, in rk_cipher_fallback() 64 areq->base.complete, areq->base.data); in rk_cipher_fallback() 65 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in rk_cipher_fallback() 66 areq->cryptlen, areq->iv); in rk_cipher_fallback() 314 struct skcipher_request *areq = container_of(async_req, struct skcipher_request, base); in rk_cipher_run() local 315 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); in rk_cipher_run() 317 struct rk_cipher_rctx *rctx = skcipher_request_ctx(areq); in rk_cipher_run() [all …]
|
/linux-6.1.9/drivers/crypto/stm32/ |
D | stm32-cryp.c | 139 struct aead_request *areq; member 361 return is_encrypt(cryp) ? cryp->areq->cryptlen : in stm32_cryp_get_input_text_len() 362 cryp->areq->cryptlen - cryp->authsize; in stm32_cryp_get_input_text_len() 371 memcpy(iv, cryp->areq->iv, 12); in stm32_cryp_gcm_init() 386 if (cryp->areq->assoclen) { in stm32_cryp_gcm_init() 437 u32 alen = cryp->areq->assoclen; in stm32_cryp_write_ccm_first_header() 478 memcpy(iv, cryp->areq->iv, AES_BLOCK_SIZE); in stm32_cryp_ccm_init() 488 if (cryp->areq->assoclen) in stm32_cryp_ccm_init() 519 if (cryp->areq->assoclen) { in stm32_cryp_ccm_init() 643 crypto_finalize_aead_request(cryp->engine, cryp->areq, err); in stm32_cryp_finish_req() [all …]
|
/linux-6.1.9/drivers/net/ethernet/fungible/funcore/ |
D | fun_dev.c | 226 const struct fun_dev_params *areq) in fun_enable_admin_queue() argument 229 .cqe_size_log2 = areq->cqe_size_log2, in fun_enable_admin_queue() 230 .sqe_size_log2 = areq->sqe_size_log2, in fun_enable_admin_queue() 231 .cq_depth = areq->cq_depth, in fun_enable_admin_queue() 232 .sq_depth = areq->sq_depth, in fun_enable_admin_queue() 233 .rq_depth = areq->rq_depth, in fun_enable_admin_queue() 235 unsigned int ntags = areq->sq_depth - 1; in fun_enable_admin_queue() 242 if (areq->sq_depth < AQA_MIN_QUEUE_SIZE || in fun_enable_admin_queue() 243 areq->sq_depth > AQA_MAX_QUEUE_SIZE || in fun_enable_admin_queue() 244 areq->cq_depth < AQA_MIN_QUEUE_SIZE || in fun_enable_admin_queue() [all …]
|
/linux-6.1.9/drivers/crypto/xilinx/ |
D | zynqmp-aes-gcm.c | 207 struct aead_request *areq = in zynqmp_handle_aes_req() local 211 struct zynqmp_aead_req_ctx *rq_ctx = aead_request_ctx(areq); in zynqmp_handle_aes_req() 216 need_fallback = zynqmp_fallback_check(tfm_ctx, areq); in zynqmp_handle_aes_req() 221 aead_request_set_callback(subreq, areq->base.flags, in zynqmp_handle_aes_req() 223 aead_request_set_crypt(subreq, areq->src, areq->dst, in zynqmp_handle_aes_req() 224 areq->cryptlen, areq->iv); in zynqmp_handle_aes_req() 225 aead_request_set_ad(subreq, areq->assoclen); in zynqmp_handle_aes_req() 231 err = zynqmp_aes_aead_cipher(areq); in zynqmp_handle_aes_req() 234 crypto_finalize_aead_request(engine, areq, err); in zynqmp_handle_aes_req()
|