Home
last modified time | relevance | path

Searched refs:out_sg (Results 1 – 23 of 23) sorted by relevance

/linux-5.19.10/lib/
Dsg_split.c17 struct scatterlist *out_sg; member
80 struct scatterlist *in_sg, *out_sg; in sg_split_phys() local
85 out_sg = split->out_sg; in sg_split_phys()
86 for (j = 0; j < split->nents; j++, out_sg++) { in sg_split_phys()
87 *out_sg = *in_sg; in sg_split_phys()
89 out_sg->offset += split->skip_sg0; in sg_split_phys()
90 out_sg->length -= split->skip_sg0; in sg_split_phys()
92 out_sg->offset = 0; in sg_split_phys()
94 sg_dma_address(out_sg) = 0; in sg_split_phys()
95 sg_dma_len(out_sg) = 0; in sg_split_phys()
[all …]
/linux-5.19.10/drivers/crypto/nx/
Dnx-aes-xcbc.c63 struct nx_sg *in_sg, *out_sg; in nx_xcbc_empty() local
87 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *) keys, &len, in nx_xcbc_empty()
94 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_xcbc_empty()
115 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, in nx_xcbc_empty()
122 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_xcbc_empty()
173 struct nx_sg *out_sg; in nx_xcbc_update() local
202 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *)sctx->state, in nx_xcbc_update()
210 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_xcbc_update()
297 struct nx_sg *in_sg, *out_sg; in nx_xcbc_final() local
333 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, in nx_xcbc_final()
[all …]
Dnx-sha256.c67 struct nx_sg *out_sg; in nx_sha256_update() local
98 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *)sctx->state, in nx_sha256_update()
100 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_sha256_update()
186 struct nx_sg *in_sg, *out_sg; in nx_sha256_final() local
224 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, max_sg_len); in nx_sha256_final()
232 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_sha256_final()
Dnx-sha512.c67 struct nx_sg *out_sg; in nx_sha512_update() local
98 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *)sctx->state, in nx_sha512_update()
100 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_sha512_update()
189 struct nx_sg *in_sg, *out_sg; in nx_sha512_final() local
233 out_sg = nx_build_sg_list(nx_ctx->out_sg, out, &len, in nx_sha512_final()
237 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in nx_sha512_final()
Dnx.c270 struct nx_sg *nx_outsg = nx_ctx->out_sg; in nx_build_sg_lists()
295 nx_ctx->op.outlen = trim_sg_list(nx_ctx->out_sg, nx_outsg, delta, nbytes); in nx_build_sg_lists()
315 nx_ctx->op.out = __pa(nx_ctx->out_sg); in nx_ctx_init()
323 nx_ctx->op_aead.out = __pa(nx_ctx->out_sg); in nx_ctx_init()
678 nx_ctx->out_sg = (struct nx_sg *)((u8 *)nx_ctx->in_sg + NX_PAGE_SIZE); in nx_crypto_ctx_init()
682 (struct nx_csbcpb *)((u8 *)nx_ctx->out_sg + in nx_crypto_ctx_init()
754 nx_ctx->out_sg = NULL; in nx_crypto_ctx_exit()
Dnx-aes-gcm.c249 struct nx_sg *in_sg, *out_sg; in gcm_empty() local
276 out_sg = nx_build_sg_list(nx_ctx->out_sg, (u8 *) out, &len, in gcm_empty()
283 nx_ctx->op.outlen = (nx_ctx->out_sg - out_sg) * sizeof(struct nx_sg); in gcm_empty()
Dnx.h127 struct nx_sg *out_sg; /* aligned pointer into kmem to an sg list */ member
Dnx-aes-ccm.c165 struct nx_sg *nx_outsg = nx_ctx->out_sg; in generate_pat()
254 nx_ctx->op.outlen = (nx_ctx->out_sg - nx_outsg) * in generate_pat()
/linux-5.19.10/drivers/crypto/
Domap-des.c153 struct scatterlist *out_sg; member
377 struct scatterlist *in_sg, struct scatterlist *out_sg, in omap_des_crypt_dma() argument
388 scatterwalk_start(&dd->out_walk, dd->out_sg); in omap_des_crypt_dma()
434 tx_out = dmaengine_prep_slave_sg(dd->dma_lch_out, out_sg, out_sg_len, in omap_des_crypt_dma()
473 err = dma_map_sg(dd->dev, dd->out_sg, dd->out_sg_len, in omap_des_crypt_dma_start()
481 err = omap_des_crypt_dma(tfm, dd->in_sg, dd->out_sg, dd->in_sg_len, in omap_des_crypt_dma_start()
485 dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, in omap_des_crypt_dma_start()
544 dd->out_sg = req->dst; in omap_des_prepare_req()
557 ret = omap_crypto_align_sg(&dd->out_sg, dd->total, DES_BLOCK_SIZE, in omap_des_prepare_req()
567 dd->out_sg_len = sg_nents_for_len(dd->out_sg, dd->total); in omap_des_prepare_req()
[all …]
Domap-aes-gcm.c34 dd->out_sg = NULL; in omap_aes_gcm_finish_req()
54 dma_sync_sg_for_device(dd->dev, dd->out_sg, dd->out_sg_len, in omap_aes_gcm_done_task()
57 dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, DMA_FROM_DEVICE); in omap_aes_gcm_done_task()
60 omap_crypto_cleanup(dd->out_sg, dd->orig_out, in omap_aes_gcm_done_task()
150 dd->out_sg = req->dst; in omap_aes_gcm_copy_buffers()
153 dd->out_sg = scatterwalk_ffwd(sg_arr, req->dst, req->assoclen); in omap_aes_gcm_copy_buffers()
156 if (req->src == req->dst || dd->out_sg == sg_arr) in omap_aes_gcm_copy_buffers()
160 ret = omap_crypto_align_sg(&dd->out_sg, cryptlen, in omap_aes_gcm_copy_buffers()
169 dd->out_sg_len = sg_nents_for_len(dd->out_sg, clen); in omap_aes_gcm_copy_buffers()
Domap-aes.c269 struct scatterlist *out_sg, in omap_aes_crypt_dma() argument
279 scatterwalk_start(&dd->out_walk, dd->out_sg); in omap_aes_crypt_dma()
327 tx_out = dmaengine_prep_slave_sg(dd->dma_lch_out, out_sg, in omap_aes_crypt_dma()
377 err = dma_map_sg(dd->dev, dd->out_sg, dd->out_sg_len, in omap_aes_crypt_dma_start()
386 err = omap_aes_crypt_dma(dd, dd->in_sg, dd->out_sg, dd->in_sg_len, in omap_aes_crypt_dma_start()
391 dma_unmap_sg(dd->dev, dd->out_sg, dd->out_sg_len, in omap_aes_crypt_dma_start()
448 dd->out_sg = req->dst; in omap_aes_prepare_req()
461 ret = omap_crypto_align_sg(&dd->out_sg, dd->total, AES_BLOCK_SIZE, in omap_aes_prepare_req()
471 dd->out_sg_len = sg_nents_for_len(dd->out_sg, dd->total); in omap_aes_prepare_req()
512 dma_sync_sg_for_device(dd->dev, dd->out_sg, dd->out_sg_len, in omap_aes_done_task()
[all …]
Datmel-tdes.c118 struct scatterlist *out_sg; member
304 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_FROM_DEVICE); in atmel_tdes_crypt_pdc_stop()
311 count = atmel_tdes_sg_copy(&dd->out_sg, &dd->out_offset, in atmel_tdes_crypt_pdc_stop()
498 out = IS_ALIGNED((u32)dd->out_sg->offset, sizeof(u32)) && in atmel_tdes_crypt_start()
499 IS_ALIGNED(dd->out_sg->length, dd->ctx->block_size); in atmel_tdes_crypt_start()
502 if (sg_dma_len(dd->in_sg) != sg_dma_len(dd->out_sg)) in atmel_tdes_crypt_start()
509 count = min_t(size_t, count, sg_dma_len(dd->out_sg)); in atmel_tdes_crypt_start()
517 err = dma_map_sg(dd->dev, dd->out_sg, 1, in atmel_tdes_crypt_start()
527 addr_out = sg_dma_address(dd->out_sg); in atmel_tdes_crypt_start()
551 dma_unmap_sg(dd->dev, dd->out_sg, 1, DMA_TO_DEVICE); in atmel_tdes_crypt_start()
[all …]
Dsahara.c225 struct scatterlist *out_sg; member
477 dev->nb_out_sg = sg_nents_for_len(dev->out_sg, dev->total); in sahara_hw_descriptor_create()
494 ret = dma_map_sg(dev->device, dev->out_sg, dev->nb_out_sg, in sahara_hw_descriptor_create()
517 sg = dev->out_sg; in sahara_hw_descriptor_create()
543 dma_unmap_sg(dev->device, dev->out_sg, dev->nb_out_sg, in sahara_hw_descriptor_create()
568 dev->out_sg = req->dst; in sahara_aes_process()
594 dma_unmap_sg(dev->device, dev->out_sg, dev->nb_out_sg, in sahara_aes_process()
Domap-aes.h184 struct scatterlist *out_sg; member
/linux-5.19.10/drivers/crypto/gemini/
Dsl3516-ce-cipher.c27 struct scatterlist *out_sg; in sl3516_ce_need_fallback() local
84 out_sg = areq->dst; in sl3516_ce_need_fallback()
85 while (in_sg && out_sg) { in sl3516_ce_need_fallback()
86 if (in_sg->length != out_sg->length) { in sl3516_ce_need_fallback()
91 out_sg = sg_next(out_sg); in sl3516_ce_need_fallback()
93 if (in_sg || out_sg) in sl3516_ce_need_fallback()
/linux-5.19.10/drivers/crypto/allwinner/sun8i-ss/
Dsun8i-ss-cipher.c29 struct scatterlist *out_sg = areq->dst; in sun8i_ss_need_fallback() local
77 out_sg = areq->dst; in sun8i_ss_need_fallback()
78 while (in_sg && out_sg) { in sun8i_ss_need_fallback()
79 if (in_sg->length != out_sg->length) in sun8i_ss_need_fallback()
82 out_sg = sg_next(out_sg); in sun8i_ss_need_fallback()
84 if (in_sg || out_sg) in sun8i_ss_need_fallback()
/linux-5.19.10/drivers/crypto/allwinner/sun4i-ss/
Dsun4i-ss-cipher.c183 struct scatterlist *out_sg = areq->dst; in sun4i_ss_cipher_poll() local
230 while (out_sg && no_chunk == 1) { in sun4i_ss_cipher_poll()
231 if ((out_sg->length | out_sg->offset) & 3u) in sun4i_ss_cipher_poll()
233 out_sg = sg_next(out_sg); in sun4i_ss_cipher_poll()
/linux-5.19.10/crypto/
Drsa-pkcs1pad.c102 struct scatterlist in_sg[2], out_sg[1]; member
360 pkcs1pad_sg_set_buf(req_ctx->out_sg, req_ctx->out_buf, in pkcs1pad_decrypt()
369 req_ctx->out_sg, req->src_len, in pkcs1pad_decrypt()
552 pkcs1pad_sg_set_buf(req_ctx->out_sg, req_ctx->out_buf, in pkcs1pad_verify()
561 req_ctx->out_sg, sig_size, ctx->key_size); in pkcs1pad_verify()
/linux-5.19.10/crypto/asymmetric_keys/
Dpublic_key.c218 struct scatterlist in_sg, out_sg; in software_key_eds_op() local
258 sg_init_one(&out_sg, out, params->out_len); in software_key_eds_op()
259 akcipher_request_set_crypt(req, &in_sg, &out_sg, params->in_len, in software_key_eds_op()
/linux-5.19.10/net/vmw_vsock/
Dvirtio_transport.c106 int ret, in_sg = 0, out_sg = 0; in virtio_transport_send_pkt_work() local
125 sgs[out_sg++] = &hdr; in virtio_transport_send_pkt_work()
128 sgs[out_sg++] = &buf; in virtio_transport_send_pkt_work()
131 ret = virtqueue_add_sgs(vq, sgs, out_sg, in_sg, pkt, GFP_KERNEL); in virtio_transport_send_pkt_work()
/linux-5.19.10/arch/um/drivers/
Dvirt-pci.c72 struct scatterlist out_sg, extra_sg, in_sg; in um_pci_send_cmd() local
74 [0] = &out_sg, in um_pci_send_cmd()
122 sg_init_one(&out_sg, cmd, cmd_size); in um_pci_send_cmd()
/linux-5.19.10/net/tls/
Dtls_sw.c1409 struct scatterlist *out_sg, in decrypt_internal() argument
1427 if (darg->zc && (out_iov || out_sg)) { in decrypt_internal()
1432 n_sgout = sg_nents(out_sg); in decrypt_internal()
1524 } else if (out_sg) { in decrypt_internal()
1525 memcpy(sgout, out_sg, n_sgout * sizeof(*sgout)); in decrypt_internal()
/linux-5.19.10/drivers/crypto/stm32/
Dstm32-cryp.c148 struct scatterlist *out_sg; member
1066 cryp->out_sg = req ? req->dst : areq->dst; in stm32_cryp_prepare_req()
1067 scatterwalk_start(&cryp->out_walk, cryp->out_sg); in stm32_cryp_prepare_req()