/linux-6.1.9/crypto/async_tx/ |
D | async_xor.c | 32 int src_cnt = unmap->to_cnt; in do_async_xor() local 37 while (src_cnt) { in do_async_xor() 41 xor_src_cnt = min(src_cnt, (int)dma->max_xor); in do_async_xor() 45 if (src_cnt > xor_src_cnt) { in do_async_xor() 86 if (src_cnt > xor_src_cnt) { in do_async_xor() 88 src_cnt -= xor_src_cnt; in do_async_xor() 90 src_cnt++; in do_async_xor() 102 int src_cnt, size_t len, struct async_submit_ctl *submit) in do_sync_xor_offs() argument 116 for (i = 0; i < src_cnt; i++) in do_sync_xor_offs() 120 src_cnt = xor_src_cnt; in do_sync_xor_offs() [all …]
|
D | async_pq.c | 46 int src_cnt = disks - 2; in do_async_gen_syndrome() local 51 while (src_cnt > 0) { in do_async_gen_syndrome() 53 pq_src_cnt = min(src_cnt, dma_maxpq(dma, dma_flags)); in do_async_gen_syndrome() 58 if (src_cnt > pq_src_cnt) { in do_async_gen_syndrome() 94 src_cnt -= pq_src_cnt; in do_async_gen_syndrome() 144 int src_cnt, size_t len) in is_dma_pq_aligned_offs() argument 148 for (i = 0; i < src_cnt; i++) { in is_dma_pq_aligned_offs() 180 int src_cnt = disks - 2; in async_gen_syndrome() local 183 blocks, src_cnt, len); in async_gen_syndrome() 194 (src_cnt <= dma_maxpq(device, 0) || in async_gen_syndrome() [all …]
|
/linux-6.1.9/drivers/dma/ioat/ |
D | prep.c | 153 dma_addr_t dest, dma_addr_t *src, unsigned int src_cnt, in __ioat_prep_xor_lock() argument 168 BUG_ON(src_cnt < 2); in __ioat_prep_xor_lock() 174 if (src_cnt > 5) { in __ioat_prep_xor_lock() 209 for (s = 0; s < src_cnt; s++) in __ioat_prep_xor_lock() 215 xor->ctl_f.src_cnt = src_cnt_to_hw(src_cnt); in __ioat_prep_xor_lock() 246 unsigned int src_cnt, size_t len, unsigned long flags) in ioat_prep_xor() argument 253 return __ioat_prep_xor_lock(chan, NULL, dest, src, src_cnt, len, flags); in ioat_prep_xor() 258 unsigned int src_cnt, size_t len, in ioat_prep_xor_val() argument 272 src_cnt - 1, len, flags); in ioat_prep_xor_val() 283 int src_cnt = src_cnt_to_sw(pq->ctl_f.src_cnt); in dump_pq_desc_dbg() local [all …]
|
D | dma.h | 362 unsigned int src_cnt, size_t len, unsigned long flags); 365 unsigned int src_cnt, size_t len, 369 unsigned int src_cnt, const unsigned char *scf, size_t len, 373 unsigned int src_cnt, const unsigned char *scf, size_t len, 377 unsigned int src_cnt, size_t len, unsigned long flags); 380 unsigned int src_cnt, size_t len,
|
D | hw.h | 115 unsigned int src_cnt:3; member 162 unsigned int src_cnt:3; member 211 unsigned int src_cnt:3; member
|
D | dma.c | 506 if (src_cnt_to_sw(xor->ctl_f.src_cnt) > 5) in desc_has_ext() 512 if (src_cnt_to_sw(pq->ctl_f.src_cnt) > 3) in desc_has_ext()
|
/linux-6.1.9/drivers/dma/ppc4xx/ |
D | adma.c | 115 int src_cnt, u32 addr); 181 unsigned int src_cnt) in prep_dma_xor_dbg() argument 186 for (i = 0; i < src_cnt; i++) in prep_dma_xor_dbg() 192 unsigned int src_cnt) in prep_dma_pq_dbg() argument 197 for (i = 0; i < src_cnt; i++) in prep_dma_pq_dbg() 205 unsigned int src_cnt, in prep_dma_pqzero_sum_dbg() argument 212 for (i = 0; i < src_cnt; i++) in prep_dma_pqzero_sum_dbg() 215 for (i = 0; i < src_cnt; i++) in prep_dma_pqzero_sum_dbg() 221 pr_debug("\t0x%016llx ", src[src_cnt + i]); in prep_dma_pqzero_sum_dbg() 264 desc->src_cnt = 0; in ppc440spe_desc_init_null_xor() [all …]
|
D | adma.h | 156 u8 src_cnt; member
|
/linux-6.1.9/drivers/dma/ |
D | iop-adma.h | 174 iop_chan_pq_slot_count(size_t len, int src_cnt, int *slots_per_op) in iop_chan_pq_slot_count() argument 181 iop_desc_init_pq(struct iop_adma_desc_slot *desc, int src_cnt, in iop_desc_init_pq() argument 201 iop_chan_pq_zero_sum_slot_count(size_t len, int src_cnt, int *slots_per_op) in iop_chan_pq_zero_sum_slot_count() argument 208 iop_desc_init_pq_zero_sum(struct iop_adma_desc_slot *desc, int src_cnt, in iop_desc_init_pq_zero_sum() argument 305 static inline int iop3xx_aau_xor_slot_count(size_t len, int src_cnt, in iop3xx_aau_xor_slot_count() argument 318 *slots_per_op = slot_count_table[src_cnt - 1]; in iop3xx_aau_xor_slot_count() 337 static inline int iop_chan_xor_slot_count(size_t len, int src_cnt, in iop_chan_xor_slot_count() argument 340 int slot_cnt = iop3xx_aau_xor_slot_count(len, src_cnt, slots_per_op); in iop_chan_xor_slot_count() 359 static inline int iop_chan_zero_sum_slot_count(size_t len, int src_cnt, in iop_chan_zero_sum_slot_count() argument 362 int slot_cnt = iop3xx_aau_xor_slot_count(len, src_cnt, slots_per_op); in iop_chan_zero_sum_slot_count() [all …]
|
D | iop-adma.c | 538 dma_addr_t *dma_src, unsigned int src_cnt, size_t len, in iop_adma_prep_dma_xor() argument 551 __func__, src_cnt, len, flags); in iop_adma_prep_dma_xor() 554 slot_cnt = iop_chan_xor_slot_count(len, src_cnt, &slots_per_op); in iop_adma_prep_dma_xor() 558 iop_desc_init_xor(grp_start, src_cnt, flags); in iop_adma_prep_dma_xor() 562 while (src_cnt--) in iop_adma_prep_dma_xor() 563 iop_desc_set_xor_src_addr(grp_start, src_cnt, in iop_adma_prep_dma_xor() 564 dma_src[src_cnt]); in iop_adma_prep_dma_xor() 573 unsigned int src_cnt, size_t len, u32 *result, in iop_adma_prep_dma_xor_val() argument 584 __func__, src_cnt, len); in iop_adma_prep_dma_xor_val() 587 slot_cnt = iop_chan_zero_sum_slot_count(len, src_cnt, &slots_per_op); in iop_adma_prep_dma_xor_val() [all …]
|
D | fsl_raid.c | 316 unsigned int src_cnt, const unsigned char *scf, size_t len, in fsl_re_prep_dma_genq() argument 325 unsigned int save_src_cnt = src_cnt; in fsl_re_prep_dma_genq() 341 src_cnt += 1; in fsl_re_prep_dma_genq() 346 cdb |= (src_cnt - 1) << FSL_RE_CDB_NRCS_SHIFT; in fsl_re_prep_dma_genq() 361 for (i = 0; i < src_cnt; i++) in fsl_re_prep_dma_genq() 391 unsigned int src_cnt, size_t len, unsigned long flags) in fsl_re_prep_dma_xor() argument 394 return fsl_re_prep_dma_genq(chan, dest, src, src_cnt, NULL, len, flags); in fsl_re_prep_dma_xor() 403 unsigned int src_cnt, const unsigned char *scf, size_t len, in fsl_re_prep_dma_pq() argument 413 unsigned int save_src_cnt = src_cnt; in fsl_re_prep_dma_pq() 427 if (src_cnt == 1) { in fsl_re_prep_dma_pq() [all …]
|
D | bcm-sba-raid.c | 762 dma_addr_t dst, dma_addr_t *src, u32 src_cnt) in sba_fillup_xor_msg() argument 788 for (i = 1; i < src_cnt; i++) { in sba_fillup_xor_msg() 842 u32 src_cnt, size_t len, unsigned long flags) in sba_prep_dma_xor_req() argument 855 off, len, dst, src, src_cnt); in sba_prep_dma_xor_req() 866 u32 src_cnt, size_t len, unsigned long flags) in sba_prep_dma_xor() argument 874 if (unlikely(src_cnt > sba->max_xor_srcs)) in sba_prep_dma_xor() 881 req = sba_prep_dma_xor_req(sba, off, dst, src, src_cnt, in sba_prep_dma_xor() 907 const u8 *scf, dma_addr_t *src, u32 src_cnt) in sba_fillup_pq_msg() argument 968 for (i = 0; i < src_cnt; i++) { in sba_fillup_pq_msg() 1053 u32 src_cnt, const u8 *scf, size_t len, unsigned long flags) in sba_prep_dma_pq_req() argument [all …]
|
D | mv_xor_v2.c | 432 unsigned int src_cnt, size_t len, unsigned long flags) in mv_xor_v2_prep_dma_xor() argument 440 if (src_cnt > MV_XOR_V2_CMD_LINE_NUM_MAX_D_BUF || src_cnt < 1) in mv_xor_v2_prep_dma_xor() 445 __func__, src_cnt, len, &dest, flags); in mv_xor_v2_prep_dma_xor() 468 for (i = 0; i < src_cnt; i++) in mv_xor_v2_prep_dma_xor() 472 src_cnt << DESC_NUM_ACTIVE_D_BUF_SHIFT; in mv_xor_v2_prep_dma_xor()
|
D | xgene-dma.c | 358 static u8 xgene_dma_encode_xor_flyby(u32 src_cnt) in xgene_dma_encode_xor_flyby() argument 369 return flyby_type[src_cnt]; in xgene_dma_encode_xor_flyby() 416 u32 src_cnt, size_t *nbytes, in xgene_dma_prep_xor_desc() argument 437 desc1->m2 |= cpu_to_le64(xgene_dma_encode_xor_flyby(src_cnt)); in xgene_dma_prep_xor_desc() 440 for (i = 0; i < src_cnt; i++) { in xgene_dma_prep_xor_desc() 839 u32 src_cnt, size_t len, unsigned long flags) in xgene_dma_prep_xor() argument 859 src_cnt, &len, multi); in xgene_dma_prep_xor() 887 u32 src_cnt, const u8 *scf, size_t len, unsigned long flags) in xgene_dma_prep_pq() argument 905 memcpy(_src, src, sizeof(*src) * src_cnt); in xgene_dma_prep_pq() 934 src_cnt, &len, multi); in xgene_dma_prep_pq() [all …]
|
D | mv_xor.c | 556 unsigned int src_cnt, size_t len, unsigned long flags) in mv_xor_prep_dma_xor() argument 569 __func__, src_cnt, len, &dest, flags); in mv_xor_prep_dma_xor() 583 while (src_cnt--) { in mv_xor_prep_dma_xor() 585 ret = mv_xor_add_io_win(mv_chan, src[src_cnt]); in mv_xor_prep_dma_xor() 588 mv_desc_set_src_addr(sw_desc, src_cnt, src[src_cnt]); in mv_xor_prep_dma_xor()
|
/linux-6.1.9/include/linux/ |
D | async_tx.h | 163 int src_cnt, size_t len, struct async_submit_ctl *submit); 168 int src_cnt, size_t len, struct async_submit_ctl *submit); 172 int src_cnt, size_t len, enum sum_check_flags *result, 178 int src_cnt, size_t len, enum sum_check_flags *result, 189 async_gen_syndrome(struct page **blocks, unsigned int *offsets, int src_cnt, 193 async_syndrome_val(struct page **blocks, unsigned int *offsets, int src_cnt,
|
D | dmaengine.h | 891 unsigned int src_cnt, size_t len, unsigned long flags); 893 struct dma_chan *chan, dma_addr_t *src, unsigned int src_cnt, 897 unsigned int src_cnt, const unsigned char *scf, 901 unsigned int src_cnt, const unsigned char *scf, size_t len,
|
/linux-6.1.9/arch/powerpc/include/asm/ |
D | async_tx.h | 14 int src_cnt, size_t src_sz); 17 src_cnt, src_sz) \ argument 19 src_cnt, src_sz)
|
/linux-6.1.9/fs/btrfs/ |
D | raid56.c | 450 static void run_xor(void **pages, int src_cnt, ssize_t len) in run_xor() argument 454 void *dest = pages[src_cnt]; in run_xor() 456 while(src_cnt > 0) { in run_xor() 457 xor_src_cnt = min(src_cnt, MAX_XOR_BLOCKS); in run_xor() 460 src_cnt -= xor_src_cnt; in run_xor()
|
/linux-6.1.9/drivers/crypto/ |
D | hifn_795x.c | 550 volatile __le16 src_cnt; /* 15:0 of source count */ member
|