Home
last modified time | relevance | path

Searched refs:sg_src (Results 1 – 16 of 16) sorted by relevance

/linux-5.19.10/drivers/crypto/rockchip/
Drk3288_crypto.c68 static int check_alignment(struct scatterlist *sg_src, in check_alignment() argument
74 in = IS_ALIGNED((uint32_t)sg_src->offset, 4) && in check_alignment()
75 IS_ALIGNED((uint32_t)sg_src->length, align_mask); in check_alignment()
82 return (align && (sg_src->length == sg_dst->length)); in check_alignment()
86 struct scatterlist *sg_src, in rk_load_data() argument
92 check_alignment(sg_src, sg_dst, dev->align_size) : in rk_load_data()
95 count = min(dev->left_bytes, sg_src->length); in rk_load_data()
98 if (!dma_map_sg(dev->dev, sg_src, 1, DMA_TO_DEVICE)) { in rk_load_data()
103 dev->addr_in = sg_dma_address(sg_src); in rk_load_data()
110 dma_unmap_sg(dev->dev, sg_src, 1, in rk_load_data()
[all …]
Drk3288_crypto_skcipher.c251 u8 *src_last_blk = page_address(sg_page(dev->sg_src)) + in rk_set_data_start()
252 dev->sg_src->offset + dev->sg_src->length - ivsize; in rk_set_data_start()
263 err = dev->load_data(dev, dev->sg_src, dev->sg_dst); in rk_set_data_start()
278 dev->sg_src = req->src; in rk_ablk_start()
357 if (sg_is_last(dev->sg_src)) { in rk_ablk_rx()
363 dev->sg_src = sg_next(dev->sg_src); in rk_ablk_rx()
Drk3288_crypto_ahash.c188 err = dev->load_data(dev, dev->sg_src, NULL); in rk_ahash_set_data_start()
205 dev->sg_src = req->src; in rk_ahash_start()
239 if (sg_is_last(dev->sg_src)) { in rk_ahash_crypto_rx()
245 dev->sg_src = sg_next(dev->sg_src); in rk_ahash_crypto_rx()
Drk3288_crypto.h205 struct scatterlist *sg_src; member
226 struct scatterlist *sg_src,
/linux-5.19.10/arch/arm/crypto/
Daes-ce-glue.c273 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_encrypt() local
301 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_encrypt()
331 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_decrypt() local
359 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_decrypt()
450 struct scatterlist sg_src[2], sg_dst[2]; in xts_encrypt() local
496 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_encrypt()
522 struct scatterlist sg_src[2], sg_dst[2]; in xts_decrypt() local
568 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_decrypt()
/linux-5.19.10/drivers/crypto/nx/
Dnx.c150 struct scatterlist *sg_src, in nx_walk_and_build() argument
161 scatterwalk_start(&walk, sg_src); in nx_walk_and_build()
163 if (start < offset + sg_src->length) in nx_walk_and_build()
166 offset += sg_src->length; in nx_walk_and_build()
167 sg_src = sg_next(sg_src); in nx_walk_and_build()
/linux-5.19.10/arch/arm64/crypto/
Daes-glue.c278 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_encrypt() local
305 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_encrypt()
335 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_decrypt() local
362 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_decrypt()
487 struct scatterlist sg_src[2], sg_dst[2]; in xts_encrypt() local
533 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_encrypt()
559 struct scatterlist sg_src[2], sg_dst[2]; in xts_decrypt() local
605 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_decrypt()
Daes-neonbs-glue.c272 struct scatterlist sg_src[2], sg_dst[2]; in __xts_crypt() local
344 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in __xts_crypt()
/linux-5.19.10/lib/crypto/
Dchacha20poly1305-selftest.c8891 struct scatterlist sg_src[3]; in chacha20poly1305_selftest() local
8926 sg_init_one(sg_src, computed_output, in chacha20poly1305_selftest()
8928 ret = chacha20poly1305_encrypt_sg_inplace(sg_src, in chacha20poly1305_selftest()
8968 sg_init_one(sg_src, computed_output, in chacha20poly1305_selftest()
8970 ret = chacha20poly1305_decrypt_sg_inplace(sg_src, in chacha20poly1305_selftest()
9032 sg_init_table(sg_src, 3); in chacha20poly1305_selftest()
9034 sg_set_buf(&sg_src[k++], input, i); in chacha20poly1305_selftest()
9036 sg_set_buf(&sg_src[k++], input + i, j - i); in chacha20poly1305_selftest()
9038 sg_set_buf(&sg_src[k++], input + j, total_len - j); in chacha20poly1305_selftest()
9039 sg_init_marker(sg_src, k); in chacha20poly1305_selftest()
[all …]
/linux-5.19.10/arch/x86/crypto/
Daegis128-aesni-glue.c70 struct aegis_state *state, struct scatterlist *sg_src, in crypto_aegis128_aesni_process_ad() argument
77 scatterwalk_start(&walk, sg_src); in crypto_aegis128_aesni_process_ad()
Daesni-intel_glue.c372 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_encrypt() local
399 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_encrypt()
428 struct scatterlist sg_src[2], sg_dst[2]; in cts_cbc_decrypt() local
455 dst = src = scatterwalk_ffwd(sg_src, req->src, subreq.cryptlen); in cts_cbc_decrypt()
903 struct scatterlist sg_src[2], sg_dst[2]; in xts_crypt() local
906 dst = src = scatterwalk_ffwd(sg_src, req->src, req->cryptlen); in xts_crypt()
/linux-5.19.10/drivers/crypto/ux500/cryp/
Dcryp.h215 struct scatterlist *sg_src; member
Dcryp_core.c534 ctx->device->dma.sg_src = sg; in cryp_set_dma_transfer()
536 ctx->device->dma.sg_src, in cryp_set_dma_transfer()
551 ctx->device->dma.sg_src, in cryp_set_dma_transfer()
611 dma_unmap_sg(chan->device->dev, ctx->device->dma.sg_src, in cryp_dma_done()
/linux-5.19.10/drivers/crypto/
Ds5p-sss.c311 struct scatterlist *sg_src; member
512 dma_unmap_sg(dev->dev, dev->sg_src, 1, DMA_TO_DEVICE); in s5p_unset_indata()
562 dev->sg_src = sg; in s5p_set_indata()
602 if (!sg_is_last(dev->sg_src)) { in s5p_aes_rx()
603 ret = s5p_set_indata(dev, sg_next(dev->sg_src)); in s5p_aes_rx()
747 s5p_set_dma_indata(dev, dev->sg_src); in s5p_aes_interrupt()
1955 s5p_set_dma_indata(dev, dev->sg_src); in s5p_aes_crypt_start()
/linux-5.19.10/drivers/dma/
Dste_dma40.c2110 struct scatterlist *sg_src, struct scatterlist *sg_dst, in d40_prep_sg_log() argument
2119 ret = d40_log_sg_to_lli(sg_src, sg_len, in d40_prep_sg_log()
2138 struct scatterlist *sg_src, struct scatterlist *sg_dst, in d40_prep_sg_phy() argument
2151 ret = d40_phy_sg_to_lli(sg_src, sg_len, src_dev_addr, in d40_prep_sg_phy()
2208 d40_prep_sg(struct dma_chan *dchan, struct scatterlist *sg_src, in d40_prep_sg() argument
2228 desc = d40_prep_desc(chan, sg_src, sg_len, dma_flags); in d40_prep_sg()
2232 if (sg_next(&sg_src[sg_len - 1]) == sg_src) in d40_prep_sg()
2243 ret = d40_prep_sg_log(chan, desc, sg_src, sg_dst, in d40_prep_sg()
2246 ret = d40_prep_sg_phy(chan, desc, sg_src, sg_dst, in d40_prep_sg()
/linux-5.19.10/crypto/
Daegis128-core.c277 struct scatterlist *sg_src, in crypto_aegis128_process_ad() argument
285 scatterwalk_start(&walk, sg_src); in crypto_aegis128_process_ad()