Home
last modified time | relevance | path

Searched refs:areq (Results 1 – 25 of 75) sorted by relevance

123

/linux-6.6.21/drivers/crypto/allwinner/sun4i-ss/
Dsun4i-ss-cipher.c15 static int noinline_for_stack sun4i_ss_opti_poll(struct skcipher_request *areq) in sun4i_ss_opti_poll() argument
17 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); in sun4i_ss_opti_poll()
21 struct sun4i_cipher_req_ctx *ctx = skcipher_request_ctx(areq); in sun4i_ss_opti_poll()
30 unsigned int ileft = areq->cryptlen; in sun4i_ss_opti_poll()
31 unsigned int oleft = areq->cryptlen; in sun4i_ss_opti_poll()
41 if (!areq->cryptlen) in sun4i_ss_opti_poll()
44 if (!areq->src || !areq->dst) { in sun4i_ss_opti_poll()
49 if (areq->iv && ivsize > 0 && mode & SS_DECRYPTION) { in sun4i_ss_opti_poll()
50 scatterwalk_map_and_copy(ctx->backup_iv, areq->src, in sun4i_ss_opti_poll()
51 areq->cryptlen - ivsize, ivsize, 0); in sun4i_ss_opti_poll()
[all …]
Dsun4i-ss-hash.c47 int sun4i_hash_init(struct ahash_request *areq) in sun4i_hash_init() argument
49 struct sun4i_req_ctx *op = ahash_request_ctx(areq); in sun4i_hash_init()
50 struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq); in sun4i_hash_init()
62 int sun4i_hash_export_md5(struct ahash_request *areq, void *out) in sun4i_hash_export_md5() argument
64 struct sun4i_req_ctx *op = ahash_request_ctx(areq); in sun4i_hash_export_md5()
85 int sun4i_hash_import_md5(struct ahash_request *areq, const void *in) in sun4i_hash_import_md5() argument
87 struct sun4i_req_ctx *op = ahash_request_ctx(areq); in sun4i_hash_import_md5()
91 sun4i_hash_init(areq); in sun4i_hash_import_md5()
104 int sun4i_hash_export_sha1(struct ahash_request *areq, void *out) in sun4i_hash_export_sha1() argument
106 struct sun4i_req_ctx *op = ahash_request_ctx(areq); in sun4i_hash_export_sha1()
[all …]
Dsun4i-ss.h201 int sun4i_hash_init(struct ahash_request *areq);
202 int sun4i_hash_update(struct ahash_request *areq);
203 int sun4i_hash_final(struct ahash_request *areq);
204 int sun4i_hash_finup(struct ahash_request *areq);
205 int sun4i_hash_digest(struct ahash_request *areq);
206 int sun4i_hash_export_md5(struct ahash_request *areq, void *out);
207 int sun4i_hash_import_md5(struct ahash_request *areq, const void *in);
208 int sun4i_hash_export_sha1(struct ahash_request *areq, void *out);
209 int sun4i_hash_import_sha1(struct ahash_request *areq, const void *in);
211 int sun4i_ss_cbc_aes_encrypt(struct skcipher_request *areq);
[all …]
/linux-6.6.21/drivers/crypto/allwinner/sun8i-ce/
Dsun8i-ce-cipher.c24 static int sun8i_ce_cipher_need_fallback(struct skcipher_request *areq) in sun8i_ce_cipher_need_fallback() argument
26 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); in sun8i_ce_cipher_need_fallback()
34 if (sg_nents_for_len(areq->src, areq->cryptlen) > MAX_SG || in sun8i_ce_cipher_need_fallback()
35 sg_nents_for_len(areq->dst, areq->cryptlen) > MAX_SG) { in sun8i_ce_cipher_need_fallback()
40 if (areq->cryptlen < crypto_skcipher_ivsize(tfm)) { in sun8i_ce_cipher_need_fallback()
45 if (areq->cryptlen == 0) { in sun8i_ce_cipher_need_fallback()
50 if (areq->cryptlen % 16) { in sun8i_ce_cipher_need_fallback()
55 len = areq->cryptlen; in sun8i_ce_cipher_need_fallback()
56 sg = areq->src; in sun8i_ce_cipher_need_fallback()
71 len = areq->cryptlen; in sun8i_ce_cipher_need_fallback()
[all …]
Dsun8i-ce-hash.c72 int sun8i_ce_hash_init(struct ahash_request *areq) in sun8i_ce_hash_init() argument
74 struct sun8i_ce_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ce_hash_init()
75 struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq); in sun8i_ce_hash_init()
81 rctx->fallback_req.base.flags = areq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP; in sun8i_ce_hash_init()
86 int sun8i_ce_hash_export(struct ahash_request *areq, void *out) in sun8i_ce_hash_export() argument
88 struct sun8i_ce_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ce_hash_export()
89 struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq); in sun8i_ce_hash_export()
93 rctx->fallback_req.base.flags = areq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP; in sun8i_ce_hash_export()
98 int sun8i_ce_hash_import(struct ahash_request *areq, const void *in) in sun8i_ce_hash_import() argument
100 struct sun8i_ce_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ce_hash_import()
[all …]
Dsun8i-ce.h352 int sun8i_ce_cipher_do_one(struct crypto_engine *engine, void *areq);
353 int sun8i_ce_skdecrypt(struct skcipher_request *areq);
354 int sun8i_ce_skencrypt(struct skcipher_request *areq);
362 int sun8i_ce_hash_init(struct ahash_request *areq);
363 int sun8i_ce_hash_export(struct ahash_request *areq, void *out);
364 int sun8i_ce_hash_import(struct ahash_request *areq, const void *in);
365 int sun8i_ce_hash_final(struct ahash_request *areq);
366 int sun8i_ce_hash_update(struct ahash_request *areq);
367 int sun8i_ce_hash_finup(struct ahash_request *areq);
368 int sun8i_ce_hash_digest(struct ahash_request *areq);
/linux-6.6.21/drivers/crypto/gemini/
Dsl3516-ce-cipher.c25 static bool sl3516_ce_need_fallback(struct skcipher_request *areq) in sl3516_ce_need_fallback() argument
27 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); in sl3516_ce_need_fallback()
34 if (areq->cryptlen == 0 || areq->cryptlen % 16) { in sl3516_ce_need_fallback()
43 if (sg_nents(areq->src) > MAXDESC / 2) { in sl3516_ce_need_fallback()
48 if (sg_nents(areq->dst) > MAXDESC) { in sl3516_ce_need_fallback()
53 sg = areq->src; in sl3516_ce_need_fallback()
69 sg = areq->dst; in sl3516_ce_need_fallback()
87 in_sg = areq->src; in sl3516_ce_need_fallback()
88 out_sg = areq->dst; in sl3516_ce_need_fallback()
103 static int sl3516_ce_cipher_fallback(struct skcipher_request *areq) in sl3516_ce_cipher_fallback() argument
[all …]
/linux-6.6.21/drivers/crypto/amlogic/
Damlogic-gxl-cipher.c25 static bool meson_cipher_need_fallback(struct skcipher_request *areq) in meson_cipher_need_fallback() argument
27 struct scatterlist *src_sg = areq->src; in meson_cipher_need_fallback()
28 struct scatterlist *dst_sg = areq->dst; in meson_cipher_need_fallback()
30 if (areq->cryptlen == 0) in meson_cipher_need_fallback()
58 static int meson_cipher_do_fallback(struct skcipher_request *areq) in meson_cipher_do_fallback() argument
60 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); in meson_cipher_do_fallback()
62 struct meson_cipher_req_ctx *rctx = skcipher_request_ctx(areq); in meson_cipher_do_fallback()
72 skcipher_request_set_callback(&rctx->fallback_req, areq->base.flags, in meson_cipher_do_fallback()
73 areq->base.complete, areq->base.data); in meson_cipher_do_fallback()
74 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in meson_cipher_do_fallback()
[all …]
Damlogic-gxl.h153 int meson_enqueue(struct crypto_async_request *areq, u32 type);
159 int meson_skdecrypt(struct skcipher_request *areq);
160 int meson_skencrypt(struct skcipher_request *areq);
161 int meson_handle_cipher_request(struct crypto_engine *engine, void *areq);
/linux-6.6.21/drivers/crypto/allwinner/sun8i-ss/
Dsun8i-ss-cipher.c23 static bool sun8i_ss_need_fallback(struct skcipher_request *areq) in sun8i_ss_need_fallback() argument
25 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); in sun8i_ss_need_fallback()
28 struct scatterlist *in_sg = areq->src; in sun8i_ss_need_fallback()
29 struct scatterlist *out_sg = areq->dst; in sun8i_ss_need_fallback()
33 if (areq->cryptlen == 0 || areq->cryptlen % 16) { in sun8i_ss_need_fallback()
38 if (sg_nents_for_len(areq->src, areq->cryptlen) > 8 || in sun8i_ss_need_fallback()
39 sg_nents_for_len(areq->dst, areq->cryptlen) > 8) { in sun8i_ss_need_fallback()
44 len = areq->cryptlen; in sun8i_ss_need_fallback()
45 sg = areq->src; in sun8i_ss_need_fallback()
59 len = areq->cryptlen; in sun8i_ss_need_fallback()
[all …]
Dsun8i-ss-hash.c161 int sun8i_ss_hash_init(struct ahash_request *areq) in sun8i_ss_hash_init() argument
163 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ss_hash_init()
164 struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq); in sun8i_ss_hash_init()
170 rctx->fallback_req.base.flags = areq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP; in sun8i_ss_hash_init()
175 int sun8i_ss_hash_export(struct ahash_request *areq, void *out) in sun8i_ss_hash_export() argument
177 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ss_hash_export()
178 struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq); in sun8i_ss_hash_export()
182 rctx->fallback_req.base.flags = areq->base.flags & CRYPTO_TFM_REQ_MAY_SLEEP; in sun8i_ss_hash_export()
187 int sun8i_ss_hash_import(struct ahash_request *areq, const void *in) in sun8i_ss_hash_import() argument
189 struct sun8i_ss_hash_reqctx *rctx = ahash_request_ctx(areq); in sun8i_ss_hash_import()
[all …]
Dsun8i-ss.h294 int sun8i_ss_handle_cipher_request(struct crypto_engine *engine, void *areq);
295 int sun8i_ss_skdecrypt(struct skcipher_request *areq);
296 int sun8i_ss_skencrypt(struct skcipher_request *areq);
309 int sun8i_ss_hash_init(struct ahash_request *areq);
310 int sun8i_ss_hash_export(struct ahash_request *areq, void *out);
311 int sun8i_ss_hash_import(struct ahash_request *areq, const void *in);
312 int sun8i_ss_hash_final(struct ahash_request *areq);
313 int sun8i_ss_hash_update(struct ahash_request *areq);
314 int sun8i_ss_hash_finup(struct ahash_request *areq);
315 int sun8i_ss_hash_digest(struct ahash_request *areq);
/linux-6.6.21/drivers/crypto/cavium/nitrox/
Dnitrox_aead.c192 struct aead_request *areq = arg; in nitrox_aead_callback() local
193 struct nitrox_aead_rctx *rctx = aead_request_ctx(areq); in nitrox_aead_callback()
202 aead_request_complete(areq, err); in nitrox_aead_callback()
213 static int nitrox_aes_gcm_enc(struct aead_request *areq) in nitrox_aes_gcm_enc() argument
215 struct crypto_aead *aead = crypto_aead_reqtfm(areq); in nitrox_aes_gcm_enc()
217 struct nitrox_aead_rctx *rctx = aead_request_ctx(areq); in nitrox_aes_gcm_enc()
222 if (!nitrox_aes_gcm_assoclen_supported(areq->assoclen)) in nitrox_aes_gcm_enc()
225 memcpy(fctx->crypto.iv, areq->iv, GCM_AES_SALT_SIZE); in nitrox_aes_gcm_enc()
227 rctx->cryptlen = areq->cryptlen; in nitrox_aes_gcm_enc()
228 rctx->assoclen = areq->assoclen; in nitrox_aes_gcm_enc()
[all …]
/linux-6.6.21/drivers/crypto/inside-secure/
Dsafexcel_hash.c224 static int safexcel_ahash_enqueue(struct ahash_request *areq);
232 struct ahash_request *areq = ahash_request_cast(async); in safexcel_handle_req_result() local
233 struct crypto_ahash *ahash = crypto_ahash_reqtfm(areq); in safexcel_handle_req_result()
234 struct safexcel_ahash_req *sreq = ahash_request_ctx_dma(areq); in safexcel_handle_req_result()
252 dma_unmap_sg(priv->dev, areq->src, sreq->nents, DMA_TO_DEVICE); in safexcel_handle_req_result()
285 areq->nbytes = 0; in safexcel_handle_req_result()
286 safexcel_ahash_enqueue(areq); in safexcel_handle_req_result()
295 *(__le32 *)areq->result = ~sreq->state[0]; in safexcel_handle_req_result()
297 memcpy(areq->result, sreq->state, in safexcel_handle_req_result()
314 struct ahash_request *areq = ahash_request_cast(async); in safexcel_ahash_send_req() local
[all …]
/linux-6.6.21/drivers/crypto/intel/qat/qat_common/
Dqat_comp_algs.c80 struct acomp_req *areq = qat_req->acompress_req; in qat_comp_resubmit() local
81 struct crypto_acomp *tfm = crypto_acomp_reqtfm(areq); in qat_comp_resubmit()
87 areq->dlen = dlen; in qat_comp_resubmit()
93 ret = qat_bl_realloc_map_new_dst(accel_dev, &areq->dst, dlen, qat_bufs, in qat_comp_resubmit()
94 qat_algs_alloc_flags(&areq->base)); in qat_comp_resubmit()
103 ret = qat_alg_send_dc_message(qat_req, inst, &areq->base); in qat_comp_resubmit()
109 acomp_request_complete(areq, ret); in qat_comp_resubmit()
140 struct acomp_req *areq = qat_req->acompress_req; in qat_comp_rfc1950_callback() local
150 scatterwalk_map_and_copy(&zlib_header, areq->dst, 0, QAT_RFC_1950_HDR_SIZE, 1); in qat_comp_rfc1950_callback()
151 areq->dlen += QAT_RFC_1950_HDR_SIZE; in qat_comp_rfc1950_callback()
[all …]
/linux-6.6.21/crypto/
Dalgif_skcipher.c60 struct af_alg_async_req *areq; in _skcipher_recvmsg() local
71 areq = af_alg_alloc_areq(sk, sizeof(struct af_alg_async_req) + in _skcipher_recvmsg()
73 if (IS_ERR(areq)) in _skcipher_recvmsg()
74 return PTR_ERR(areq); in _skcipher_recvmsg()
77 err = af_alg_get_rsgl(sk, msg, flags, areq, ctx->used, &len); in _skcipher_recvmsg()
92 areq->tsgl_entries = af_alg_count_tsgl(sk, len, 0); in _skcipher_recvmsg()
93 if (!areq->tsgl_entries) in _skcipher_recvmsg()
94 areq->tsgl_entries = 1; in _skcipher_recvmsg()
95 areq->tsgl = sock_kmalloc(sk, array_size(sizeof(*areq->tsgl), in _skcipher_recvmsg()
96 areq->tsgl_entries), in _skcipher_recvmsg()
[all …]
Dalgif_aead.c100 struct af_alg_async_req *areq; in _aead_recvmsg() local
153 areq = af_alg_alloc_areq(sk, sizeof(struct af_alg_async_req) + in _aead_recvmsg()
155 if (IS_ERR(areq)) in _aead_recvmsg()
156 return PTR_ERR(areq); in _aead_recvmsg()
159 err = af_alg_get_rsgl(sk, msg, flags, areq, outlen, &usedpages); in _aead_recvmsg()
213 rsgl_src = areq->first_rsgl.sgl.sgt.sgl; in _aead_recvmsg()
227 areq->first_rsgl.sgl.sgt.sgl, in _aead_recvmsg()
246 areq->first_rsgl.sgl.sgt.sgl, in _aead_recvmsg()
252 areq->tsgl_entries = af_alg_count_tsgl(sk, processed, in _aead_recvmsg()
254 if (!areq->tsgl_entries) in _aead_recvmsg()
[all …]
Daf_alg.c769 static void af_alg_free_areq_sgls(struct af_alg_async_req *areq) in af_alg_free_areq_sgls() argument
771 struct sock *sk = areq->sk; in af_alg_free_areq_sgls()
779 list_for_each_entry_safe(rsgl, tmp, &areq->rsgl_list, list) { in af_alg_free_areq_sgls()
783 if (rsgl != &areq->first_rsgl) in af_alg_free_areq_sgls()
787 tsgl = areq->tsgl; in af_alg_free_areq_sgls()
789 for_each_sg(tsgl, sg, areq->tsgl_entries, i) { in af_alg_free_areq_sgls()
795 sock_kfree_s(sk, tsgl, areq->tsgl_entries * sizeof(*tsgl)); in af_alg_free_areq_sgls()
1116 void af_alg_free_resources(struct af_alg_async_req *areq) in af_alg_free_resources() argument
1118 struct sock *sk = areq->sk; in af_alg_free_resources()
1121 af_alg_free_areq_sgls(areq); in af_alg_free_resources()
[all …]
/linux-6.6.21/drivers/crypto/
Dtalitos.c989 struct aead_request *areq, bool encrypt) in ipsec_esp_unmap() argument
991 struct crypto_aead *aead = crypto_aead_reqtfm(areq); in ipsec_esp_unmap()
995 unsigned int cryptlen = areq->cryptlen - (encrypt ? 0 : authsize); in ipsec_esp_unmap()
1004 talitos_sg_unmap(dev, edesc, areq->src, areq->dst, in ipsec_esp_unmap()
1005 cryptlen + authsize, areq->assoclen); in ipsec_esp_unmap()
1014 sg_pcopy_to_buffer(areq->dst, dst_nents, ctx->iv, ivsize, in ipsec_esp_unmap()
1015 areq->assoclen + cryptlen - ivsize); in ipsec_esp_unmap()
1026 struct aead_request *areq = context; in ipsec_esp_encrypt_done() local
1027 struct crypto_aead *authenc = crypto_aead_reqtfm(areq); in ipsec_esp_encrypt_done()
1033 ipsec_esp_unmap(dev, edesc, areq, true); in ipsec_esp_encrypt_done()
[all …]
Datmel-i2c.h163 void (*cbk)(struct atmel_i2c_work_data *work_data, void *areq,
165 void *areq; member
174 void *areq, int status),
175 void *areq);
/linux-6.6.21/drivers/crypto/rockchip/
Drk3288_crypto_ahash.c44 static int rk_ahash_digest_fb(struct ahash_request *areq) in rk_ahash_digest_fb() argument
46 struct rk_ahash_rctx *rctx = ahash_request_ctx(areq); in rk_ahash_digest_fb()
47 struct crypto_ahash *tfm = crypto_ahash_reqtfm(areq); in rk_ahash_digest_fb()
55 rctx->fallback_req.base.flags = areq->base.flags & in rk_ahash_digest_fb()
58 rctx->fallback_req.nbytes = areq->nbytes; in rk_ahash_digest_fb()
59 rctx->fallback_req.src = areq->src; in rk_ahash_digest_fb()
60 rctx->fallback_req.result = areq->result; in rk_ahash_digest_fb()
235 struct ahash_request *areq = container_of(breq, struct ahash_request, base); in rk_hash_prepare() local
236 struct rk_ahash_rctx *rctx = ahash_request_ctx(areq); in rk_hash_prepare()
240 ret = dma_map_sg(rkc->dev, areq->src, sg_nents(areq->src), DMA_TO_DEVICE); in rk_hash_prepare()
[all …]
Drk3288_crypto_skcipher.c68 static int rk_cipher_fallback(struct skcipher_request *areq) in rk_cipher_fallback() argument
70 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); in rk_cipher_fallback()
72 struct rk_cipher_rctx *rctx = skcipher_request_ctx(areq); in rk_cipher_fallback()
80 skcipher_request_set_callback(&rctx->fallback_req, areq->base.flags, in rk_cipher_fallback()
81 areq->base.complete, areq->base.data); in rk_cipher_fallback()
82 skcipher_request_set_crypt(&rctx->fallback_req, areq->src, areq->dst, in rk_cipher_fallback()
83 areq->cryptlen, areq->iv); in rk_cipher_fallback()
301 struct skcipher_request *areq = container_of(async_req, struct skcipher_request, base); in rk_cipher_run() local
302 struct crypto_skcipher *tfm = crypto_skcipher_reqtfm(areq); in rk_cipher_run()
303 struct rk_cipher_rctx *rctx = skcipher_request_ctx(areq); in rk_cipher_run()
[all …]
/linux-6.6.21/drivers/net/ethernet/fungible/funcore/
Dfun_dev.c225 const struct fun_dev_params *areq) in fun_enable_admin_queue() argument
228 .cqe_size_log2 = areq->cqe_size_log2, in fun_enable_admin_queue()
229 .sqe_size_log2 = areq->sqe_size_log2, in fun_enable_admin_queue()
230 .cq_depth = areq->cq_depth, in fun_enable_admin_queue()
231 .sq_depth = areq->sq_depth, in fun_enable_admin_queue()
232 .rq_depth = areq->rq_depth, in fun_enable_admin_queue()
234 unsigned int ntags = areq->sq_depth - 1; in fun_enable_admin_queue()
241 if (areq->sq_depth < AQA_MIN_QUEUE_SIZE || in fun_enable_admin_queue()
242 areq->sq_depth > AQA_MAX_QUEUE_SIZE || in fun_enable_admin_queue()
243 areq->cq_depth < AQA_MIN_QUEUE_SIZE || in fun_enable_admin_queue()
[all …]
/linux-6.6.21/drivers/crypto/xilinx/
Dzynqmp-aes-gcm.c207 struct aead_request *areq = in zynqmp_handle_aes_req() local
211 struct zynqmp_aead_req_ctx *rq_ctx = aead_request_ctx(areq); in zynqmp_handle_aes_req()
216 need_fallback = zynqmp_fallback_check(tfm_ctx, areq); in zynqmp_handle_aes_req()
221 aead_request_set_callback(subreq, areq->base.flags, in zynqmp_handle_aes_req()
223 aead_request_set_crypt(subreq, areq->src, areq->dst, in zynqmp_handle_aes_req()
224 areq->cryptlen, areq->iv); in zynqmp_handle_aes_req()
225 aead_request_set_ad(subreq, areq->assoclen); in zynqmp_handle_aes_req()
231 err = zynqmp_aes_aead_cipher(areq); in zynqmp_handle_aes_req()
234 crypto_finalize_aead_request(engine, areq, err); in zynqmp_handle_aes_req()
/linux-6.6.21/drivers/crypto/axis/
Dartpec6_crypto.c360 static int artpec6_crypto_prepare_aead(struct aead_request *areq);
361 static int artpec6_crypto_prepare_crypto(struct skcipher_request *areq);
362 static int artpec6_crypto_prepare_hash(struct ahash_request *areq);
1308 static int artpec6_crypto_prepare_hash(struct ahash_request *areq) in artpec6_crypto_prepare_hash() argument
1310 struct artpec6_hashalg_context *ctx = crypto_tfm_ctx(areq->base.tfm); in artpec6_crypto_prepare_hash()
1311 struct artpec6_hash_request_context *req_ctx = ahash_request_ctx(areq); in artpec6_crypto_prepare_hash()
1312 size_t digestsize = crypto_ahash_digestsize(crypto_ahash_reqtfm(areq)); in artpec6_crypto_prepare_hash()
1315 crypto_ahash_tfm(crypto_ahash_reqtfm(areq))); in artpec6_crypto_prepare_hash()
1401 size_t total_bytes = areq->nbytes + req_ctx->partial_bytes; in artpec6_crypto_prepare_hash()
1427 artpec6_crypto_walk_init(&walk, areq->src); in artpec6_crypto_prepare_hash()
[all …]

123