/linux-6.6.21/arch/xtensa/lib/ |
D | memcopy.S | 76 s8i a6, a5, 0 77 addi a5, a5, 1 96 s8i a6, a5, 0 97 addi a5, a5, 1 98 _bbci.l a5, 1, .Ldstaligned # if dst is now aligned, then 107 s8i a6, a5, 0 108 s8i a7, a5, 1 109 addi a5, a5, 2 117 mov a5, a2 # copy dst so that a2 is return value 140 s32i a6, a5, 0 [all …]
|
D | memset.S | 44 mov a5, a2 # copy dst so that a2 is return value 65 add a6, a6, a5 # a6 = end of last 16B chunk 68 EX(10f) s32i a3, a5, 0 69 EX(10f) s32i a3, a5, 4 70 EX(10f) s32i a3, a5, 8 71 EX(10f) s32i a3, a5, 12 72 addi a5, a5, 16 74 blt a5, a6, .Loop1 79 EX(10f) s32i a3, a5, 0 80 EX(10f) s32i a3, a5, 4 [all …]
|
D | checksum.S | 47 extui a5, a2, 0, 2 48 bnez a5, 8f /* branch if 2-byte aligned */ 51 srli a5, a3, 5 /* 32-byte chunks */ 53 loopgtz a5, 2f 55 beqz a5, 2f 56 slli a5, a5, 5 57 add a5, a5, a2 /* a5 = end of last 32-byte chunk */ 78 blt a2, a5, .Loop1 81 extui a5, a3, 2, 3 /* remaining 4-byte chunks */ 83 loopgtz a5, 3f [all …]
|
D | usercopy.S | 70 mov a5, a2 # copy dst so that a2 is return value 95 EX(10f) s8i a6, a5, 0 96 addi a5, a5, 1 98 bbci.l a5, 1, .Ldstaligned # if dst is now aligned, then 106 EX(10f) s8i a6, a5, 0 107 EX(10f) s8i a7, a5, 1 108 addi a5, a5, 2 128 EX(10f) s8i a6, a5, 0 129 addi a5, a5, 1 154 EX(10f) s32i a6, a5, 0 [all …]
|
D | mulsi3.S | 48 srai a5, a3, 16 50 mul16u a6, a5, a2 59 rsr a5, ACCLO 62 slli a5, a5, 16 63 add a2, a4, a5 72 xor a5, a2, a3 /* Top bit is 1 if one input is negative. */ 100 movltz a2, a3, a5 127 movltz a2, a3, a5
|
D | divsi3.S | 16 do_nsau a5, a6, a2, a8 /* udividend_shift = nsau (udividend) */ 18 bgeu a5, a4, .Lspecial 20 sub a4, a4, a5 /* count = udivisor_shift - udividend_shift */ 45 neg a5, a2 46 movltz a2, a5, a7 /* return (sign < 0) ? -quotient : quotient */
|
D | umodsi3.S | 14 do_nsau a5, a2, a6, a7 /* dividend_shift = nsau (dividend) */ 16 bgeu a5, a4, .Lspecial 18 sub a4, a4, a5 /* count = divisor_shift - dividend_shift */
|
D | udivsi3.S | 15 do_nsau a5, a6, a2, a7 /* dividend_shift = nsau (dividend) */ 17 bgeu a5, a4, .Lspecial 19 sub a4, a4, a5 /* count = divisor_shift - dividend_shift */
|
/linux-6.6.21/arch/xtensa/include/asm/ |
D | initialize_mmu.h | 105 addi a5, a2, -XCHAL_SPANNING_WAY 106 add a4, a4, a5 113 add a5, a2, a4 114 3: idtlb a5 115 iitlb a5 116 add a5, a5, a4 117 bne a5, a2, 3b 126 movi a5, XCHAL_KSEG_CACHED_VADDR + XCHAL_KSEG_TLB_WAY 128 wdtlb a4, a5 129 witlb a4, a5 [all …]
|
/linux-6.6.21/arch/loongarch/lib/ |
D | memmove.S | 82 sub.d a5, a2, t1 98 st.d t0, a5, -8 99 st.d t1, a5, -16 100 st.d t2, a5, -24 101 st.d t3, a5, -32 102 st.d t4, a5, -40 103 st.d t5, a5, -48 104 st.d t6, a5, -56 105 st.d t7, a5, -64 106 addi.d a5, a5, -64 [all …]
|
D | memcpy.S | 134 add.d a5, a0, t0 150 st.d t0, a5, 0 151 st.d t1, a5, 8 152 st.d t2, a5, 16 153 st.d t3, a5, 24 154 st.d t4, a5, 32 155 st.d t5, a5, 40 156 st.d t6, a5, 48 157 st.d t7, a5, 56 158 addi.d a5, a5, 64 [all …]
|
/linux-6.6.21/arch/xtensa/kernel/ |
D | align.S | 191 l32i a5, a3, 0 193 __src_b a3, a5, a6 # a3 has the data word 203 extui a5, a4, INSN_OP1, 4 204 _beqi a5, OP1_L32I, .Lload_w 205 bnei a5, OP1_L8UI, .Lload16 233 extui a5, a4, INSN_OP0, 4 # get insn.op0 nibble 236 _beqi a5, OP0_L32I_N, .Lload # L32I.N, jump 237 addi a6, a5, -OP0_S32I_N 247 .Lstore:movi a5, .Lstore_table # table 249 addx8 a5, a6, a5 [all …]
|
D | vectors.S | 707 s32e a0, a5, -16 708 s32e a1, a5, -12 709 s32e a2, a5, -8 710 s32e a3, a5, -4 719 l32e a0, a5, -16 720 l32e a1, a5, -12 721 l32e a2, a5, -8 722 l32e a3, a5, -4 737 s32e a5, a0, -28 754 l32e a5, a7, -28 [all …]
|
/linux-6.6.21/arch/riscv/lib/ |
D | tishift.S | 11 li a5,64 12 sub a5,a5,a2 13 sext.w a4,a5 14 blez a5, .L2 34 li a5,64 35 sub a5,a5,a2 36 sext.w a4,a5 37 blez a5, .L4 57 li a5,64 58 sub a5,a5,a2 [all …]
|
D | memcpy.S | 32 lb a5, 0(a1) 34 sb a5, 0(t6) 45 REG_L a5, SZREG(a1) 55 REG_S a5, SZREG(t6) 65 REG_L a5, 11*SZREG(a1) 72 REG_S a5, 11*SZREG(t6) 87 or a5, a1, t6 88 or a5, a5, a3 89 andi a5, a5, 3 90 bnez a5, 5f
|
D | uaccess.S | 53 fixup lb a5, 0(a1), 10f 55 fixup sb a5, 0(a0), 10f 81 fixup REG_L a5, SZREG(a1), 10f 89 fixup REG_S a5, SZREG(a0), 10f 129 li a5, SZREG*8 130 sub t4, a5, t3 133 fixup REG_L a5, 0(a1), 10f 144 srl a4, a5, t3 145 fixup REG_L a5, SZREG(a1), 10f 147 sll a2, a5, t4 [all …]
|
D | memmove.S | 99 andi a5, a1, (SZREG - 1) /* Find the alignment offset of src (a1) */ 100 slli a6, a5, 3 /* Multiply by 8 to convert that to bits to shift */ 101 sub a5, a1, t3 /* Find the difference between src and dest */ 155 add a1, t3, a5 /* Restore the src pointer */ 161 andi a5, a4, (SZREG - 1) /* Find the alignment offset of src (a4) */ 162 slli a6, a5, 3 /* Multiply by 8 to convert that to bits to shift */ 163 sub a5, a4, t4 /* Find the difference between src and dest */ 217 add a4, t4, a5 /* Restore the src pointer */
|
/linux-6.6.21/arch/xtensa/boot/boot-redboot/ |
D | bootstrap.S | 60 rsr a5, windowbase 61 ssl a5 87 movi a5, __start_a0 89 sub a0, a4, a5 92 movi a5, __reloc_end 96 # a5: compiled end address 112 blt a8, a5, 1b 119 ___flush_dcache_all a5 a6 123 ___invalidate_icache_all a5 a6 156 movi.n a5, 0 [all …]
|
/linux-6.6.21/arch/mips/kernel/ |
D | linux32.c | 77 unsigned long, unused, unsigned long, a4, unsigned long, a5) 79 return ksys_pread64(fd, buf, count, merge_64(a4, a5)); 83 size_t, count, u32, unused, u64, a4, u64, a5) 85 return ksys_pwrite64(fd, buf, count, merge_64(a4, a5)); 110 unsigned long a4, unsigned long a5, in sys32_sync_file_range() argument 114 merge_64(a2, a3), merge_64(a4, a5), in sys32_sync_file_range() 120 unsigned long a4, unsigned long a5, in sys32_fadvise64_64() argument 124 merge_64(a2, a3), merge_64(a4, a5), in sys32_fadvise64_64()
|
/linux-6.6.21/include/linux/ |
D | arm-smccc.h | 272 unsigned long a5; member 353 unsigned long a5, unsigned long a6, unsigned long a7, 358 unsigned long a5, unsigned long a6, unsigned long a7, in __arm_smccc_smc() argument 379 unsigned long a5, unsigned long a6, unsigned long a7, 461 #define __declare_arg_7(a0, a1, a2, a3, a4, a5, res) \ argument 462 typeof(a5) __a5 = a5; \ 464 register typeof(a5) arg5 asm("r5") = __a5 466 #define __declare_arg_8(a0, a1, a2, a3, a4, a5, a6, res) \ argument 468 __declare_arg_7(a0, a1, a2, a3, a4, a5, res); \ 471 #define __declare_arg_9(a0, a1, a2, a3, a4, a5, a6, a7, res) \ argument [all …]
|
/linux-6.6.21/arch/powerpc/platforms/powernv/ |
D | opal-call.c | 17 s64 a4, s64 a5, s64 a6, s64 a7, in __trace_opal_entry() argument 33 args[5] = a5; in __trace_opal_entry() 70 s64 a4, s64 a5, s64 a6, s64 a7, in __opal_call_trace() argument 75 __trace_opal_entry(a0, a1, a2, a3, a4, a5, a6, a7, opcode); in __opal_call_trace() 76 ret = __opal_call(a0, a1, a2, a3, a4, a5, a6, a7, opcode, msr); in __opal_call_trace() 87 s64 a4, s64 a5, s64 a6, s64 a7, in __opal_call_trace() argument 97 int64_t a4, int64_t a5, int64_t a6, int64_t a7, int64_t opcode) in opal_call() argument 110 return __opal_call(a0, a1, a2, a3, a4, a5, a6, a7, opcode, msr); in opal_call() 116 ret = __opal_call_trace(a0, a1, a2, a3, a4, a5, a6, a7, opcode, msr); in opal_call() 118 ret = __opal_call(a0, a1, a2, a3, a4, a5, a6, a7, opcode, msr); in opal_call() [all …]
|
/linux-6.6.21/arch/arm/boot/dts/arm/ |
D | vexpress-v2p-ca5s.dts | 40 compatible = "arm,cortex-a5"; 47 compatible = "arm,cortex-a5"; 97 compatible = "arm,cortex-a5-scu"; 102 compatible = "arm,cortex-a5-twd-timer"; 108 compatible = "arm,cortex-a5-global-timer", 116 compatible = "arm,cortex-a5-twd-wdt"; 122 compatible = "arm,cortex-a5-gic", "arm,cortex-a9-gic"; 139 compatible = "arm,cortex-a5-pmu";
|
/linux-6.6.21/arch/riscv/kernel/ |
D | copy-unaligned.S | 18 REG_L a5, SZREG(a1) 26 REG_S a5, SZREG(a0) 50 lb a5, 1(a1) 58 sb a5, 1(a0)
|
/linux-6.6.21/arch/arm64/crypto/ |
D | chacha-neon-core.S | 173 a5 .req w17 225 mov a5, v5.s[0] 248 add a1, a1, a5 288 eor a5, a5, a9 302 ror a5, a5, #20 315 add a1, a1, a5 355 eor a5, a5, a9 369 ror a5, a5, #25 380 add a0, a0, a5 420 eor a5, a5, a10 [all …]
|
/linux-6.6.21/arch/x86/platform/uv/ |
D | bios_uv.c | 24 u64 a4, u64 a5) in __uv_bios_call() argument 35 ret = efi_call_virt_pointer(tab, function, (u64)which, a1, a2, a3, a4, a5); in __uv_bios_call() 41 u64 a5) in uv_bios_call() argument 48 ret = __uv_bios_call(which, a1, a2, a3, a4, a5); in uv_bios_call() 55 u64 a4, u64 a5) in uv_bios_call_irqsave() argument 64 ret = __uv_bios_call(which, a1, a2, a3, a4, a5); in uv_bios_call_irqsave()
|