/linux-6.1.9/arch/riscv/lib/ |
D | memcpy.S | 12 move t6, a0 /* Preserve return value */ 18 andi a3, t6, SZREG-1 34 sb a5, 0(t6) 35 addi t6, t6, 1 54 REG_S a4, 0(t6) 55 REG_S a5, SZREG(t6) 56 REG_S a6, 2*SZREG(t6) 57 REG_S a7, 3*SZREG(t6) 58 REG_S t0, 4*SZREG(t6) 59 REG_S t1, 5*SZREG(t6) [all …]
|
D | uaccess.S | 17 li t6, SR_SUM 18 csrs CSR_STATUS, t6 171 csrc CSR_STATUS, t6 178 csrc CSR_STATUS, t6 190 li t6, SR_SUM 191 csrs CSR_STATUS, t6 213 csrc CSR_STATUS, t6 230 csrc CSR_STATUS, t6
|
D | memmove.S | 69 andi t6, t4, -SZREG 103 addi a2, t6, SZREG /* The other breakpoint for the unrolled loop*/ 151 bne t3, t6, 1b 153 mv t3, t6 /* Fix the dest pointer in case the loop was broken */ 236 bne t3, t6, 1b 273 beq t4, t6, 2f 279 bne t4, t6, 1b
|
/linux-6.1.9/arch/alpha/lib/ |
D | stxcpy.S | 73 negq t8, t6 # e0 : find low bit set 74 and t8, t6, t12 # e1 (stall) 78 and t12, 0x80, t6 # e0 : 79 bne t6, 1f # .. e1 (zdb) 84 subq t12, 1, t6 # .. e1 : 85 zapnot t1, t6, t1 # e0 : clear src bytes >= null 86 or t12, t6, t8 # .. e1 : 142 or t1, t6, t6 # e0 : 143 cmpbge zero, t6, t8 # .. e1 : 144 lda t6, -1 # e0 : for masking just below [all …]
|
D | stxncpy.S | 91 and t12, 0x80, t6 # e0 : 92 bne t6, 1f # .. e1 (zdb) 97 subq t12, 1, t6 # .. e1 : 98 or t12, t6, t8 # e0 : 165 or t0, t6, t6 # e1 : mask original data for zero test 166 cmpbge zero, t6, t8 # e0 : 168 lda t6, -1 # e0 : 171 mskql t6, a1, t6 # e0 : mask out bits already seen 174 or t6, t2, t2 # .. e1 : 243 or t8, t10, t6 # e1 : [all …]
|
D | ev6-stxcpy.S | 89 negq t8, t6 # E : find low bit set 90 and t8, t6, t12 # E : (stall) 93 and t12, 0x80, t6 # E : (stall) 94 bne t6, 1f # U : (stall) 99 subq t12, 1, t6 # E : 100 zapnot t1, t6, t1 # U : clear src bytes >= null (stall) 101 or t12, t6, t8 # E : (stall) 165 or t1, t6, t6 # E : 166 cmpbge zero, t6, t8 # E : (stall) 167 lda t6, -1 # E : for masking just below [all …]
|
D | ev6-stxncpy.S | 117 and t12, 0x80, t6 # E : (stall) 118 bne t6, 1f # U : (stall) 123 subq t12, 1, t6 # E : 124 or t12, t6, t8 # E : (stall) 205 or t0, t6, t6 # E : mask original data for zero test (stall) 207 cmpbge zero, t6, t8 # E : 209 lda t6, -1 # E : 213 mskql t6, a1, t6 # U : mask out bits already seen 215 or t6, t2, t2 # E : (stall) 288 or t8, t10, t6 # E : (stall) [all …]
|
D | strrchr.S | 23 mov zero, t6 # .. e1 : t6 is last match aligned addr 46 cmovne t3, v0, t6 # .. e1 : save previous comparisons match 63 cmovne t3, v0, t6 # e0 : 80 addq t6, t0, v0 # .. e1 : add our aligned base ptr to the mix
|
D | ev67-strrchr.S | 40 mov zero, t6 # E : t6 is last match aligned addr 68 cmovne t3, v0, t6 # E : save previous comparisons match 94 cmovne t3, v0, t6 # E : 105 addq t6, t5, v0 # E : and add to quadword address
|
/linux-6.1.9/arch/arm64/crypto/ |
D | crct10dif-ce-core.S | 86 t6 .req v20 138 ext t6.8b, ad.8b, ad.8b, #3 // A3 144 pmull t6.8h, t6.8b, fold_consts.8b // J = A3*B 152 tbl t6.16b, {ad.16b}, perm3.16b // A3 158 pmull2 t6.8h, t6.16b, fold_consts.16b // J = A3*B 164 eor t6.16b, t6.16b, t9.16b // N = I + J 168 uzp1 t7.2d, t6.2d, t3.2d 169 uzp2 t6.2d, t6.2d, t3.2d 178 eor t7.16b, t7.16b, t6.16b 179 and t6.16b, t6.16b, k00_16.16b [all …]
|
D | aes-neonbs-core.S | 250 t0, t1, t2, t3, t4, t5, t6, t7, inv 262 ext \t6\().16b, \x6\().16b, \x6\().16b, #12 265 eor \x6\().16b, \x6\().16b, \t6\().16b 275 eor \t6\().16b, \t6\().16b, \x5\().16b 293 eor \x3\().16b, \x3\().16b, \t6\().16b 298 eor \x2\().16b, \x3\().16b, \t6\().16b 306 t0, t1, t2, t3, t4, t5, t6, t7 308 ext \t6\().16b, \x6\().16b, \x6\().16b, #8 312 eor \t6\().16b, \t6\().16b, \x6\().16b 323 eor \x0\().16b, \x0\().16b, \t6\().16b [all …]
|
/linux-6.1.9/arch/x86/crypto/ |
D | camellia-aesni-avx-asm_64.S | 50 #define roundsm16(x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2, t3, t4, t5, t6, \ argument 73 filter_8bit(x0, t0, t1, t7, t6); \ 74 filter_8bit(x7, t0, t1, t7, t6); \ 75 filter_8bit(x1, t0, t1, t7, t6); \ 76 filter_8bit(x4, t0, t1, t7, t6); \ 77 filter_8bit(x2, t0, t1, t7, t6); \ 78 filter_8bit(x5, t0, t1, t7, t6); \ 82 filter_8bit(x3, t2, t3, t7, t6); \ 83 filter_8bit(x6, t2, t3, t7, t6); \ 100 filter_8bit(x0, t0, t1, t7, t6); \ [all …]
|
D | camellia-aesni-avx2-asm_64.S | 62 #define roundsm32(x0, x1, x2, x3, x4, x5, x6, x7, t0, t1, t2, t3, t4, t5, t6, \ argument 70 vbroadcasti128 .Lpre_tf_hi_s1, t6; \ 86 filter_8bit(x0, t5, t6, t7, t4); \ 87 filter_8bit(x7, t5, t6, t7, t4); \ 94 filter_8bit(x2, t5, t6, t7, t4); \ 95 filter_8bit(x5, t5, t6, t7, t4); \ 96 filter_8bit(x1, t5, t6, t7, t4); \ 97 filter_8bit(x4, t5, t6, t7, t4); \ 102 vextracti128 $1, x2, t6##_x; \ 121 vaesenclast t4##_x, t6##_x, t6##_x; \ [all …]
|
/linux-6.1.9/arch/riscv/kernel/ |
D | mcount.S | 67 mv t6, s0 71 mv a0, t6 90 la t6, ftrace_graph_entry_stub 91 bne t2, t6, do_ftrace_graph_caller
|
/linux-6.1.9/arch/riscv/include/asm/ |
D | compat.h | 52 compat_ulong_t t6; member 89 cregs->t6 = (compat_ulong_t) regs->t6; in regs_to_cregs() 126 regs->t6 = (unsigned long) cregs->t6; in cregs_to_regs()
|
/linux-6.1.9/arch/loongarch/mm/ |
D | page.S | 50 ld.d t6, a1, 48 65 st.d t6, a0, 48 67 ld.d t6, a1, 112 78 st.d t6, a0, -16
|
/linux-6.1.9/scripts/ |
D | makelst | 29 t6=`printf "%lu" $((0x$t4 - 0x$t5))` 32 $3 -r --source --adjust-vma=${t6:-0} $1
|
/linux-6.1.9/arch/mips/kernel/ |
D | scall32-o32.S | 62 load_a5: user_lw(t6, 20(t0)) # argument #6 from usp 68 sw t6, 20(sp) # argument #6 to ksp 158 li t6, 0 194 lw t6, 28(sp) 197 sw t6, 24(sp)
|
/linux-6.1.9/arch/ia64/lib/ |
D | copy_page_mck.S | 82 #define t6 t2 // alias! macro 87 #define t12 t6 // alias! 158 (p[D]) ld8 t6 = [src0], 3*8 165 (p[D]) st8 [dst0] = t6, 3*8
|
D | copy_page.S | 45 t5[PIPE_DEPTH], t6[PIPE_DEPTH], t7[PIPE_DEPTH], t8[PIPE_DEPTH] 84 (p[0]) ld8 t6[0]=[src2],16 85 (EPI) st8 [tgt2]=t6[PIPE_DEPTH-1],16
|
/linux-6.1.9/arch/sparc/lib/ |
D | blockops.S | 28 #define MIRROR_BLOCK(dst, src, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ argument 32 ldd [src + offset + 0x00], t6; \ 36 std t6, [dst + offset + 0x00];
|
D | copy_user.S | 68 #define MOVE_BIGCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ argument 72 LD(ldd, src, offset + 0x18, t6, bigchunk_fault) \ 79 ST(st, dst, offset + 0x18, t6, bigchunk_fault) \ 83 #define MOVE_BIGALIGNCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ argument 87 LD(ldd, src, offset + 0x18, t6, bigchunk_fault) \ 91 ST(std, dst, offset + 0x18, t6, bigchunk_fault)
|
D | memcpy.S | 19 #define MOVE_BIGCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ argument 23 ldd [%src + (offset) + 0x18], %t6; \ 30 st %t6, [%dst + (offset) + 0x18]; \ 33 #define MOVE_BIGALIGNCHUNK(src, dst, offset, t0, t1, t2, t3, t4, t5, t6, t7) \ argument 37 ldd [%src + (offset) + 0x18], %t6; \ 41 std %t6, [%dst + (offset) + 0x18];
|
/linux-6.1.9/drivers/media/pci/cx88/ |
D | cx88-dsp.c | 72 u32 t2, t4, t6, t8; in int_cos() local 87 t6 = t4 * x / 32768 * x / 32768 / 5 / 6; in int_cos() 88 t8 = t6 * x / 32768 * x / 32768 / 7 / 8; in int_cos() 89 ret = 32768 - t2 + t4 - t6 + t8; in int_cos()
|
/linux-6.1.9/arch/arm/crypto/ |
D | aes-neonbs-core.S | 297 t0, t1, t2, t3, t4, t5, t6, t7, inv 309 vext.8 \t6, \x6, \x6, #12 312 veor \x6, \x6, \t6 322 veor \t6, \t6, \x5 340 veor \x3, \x3, \t6 345 veor \x2, \x3, \t6 353 t0, t1, t2, t3, t4, t5, t6, t7 360 vld1.8 {\t6-\t7}, [bskey, :256] 365 veor \x6, \x6, \t6 368 vext.8 \t6, \x6, \x6, #8 [all …]
|