/linux-6.1.9/arch/x86/lib/ |
D | memmove_64.S | 35 add %rdx, %r8 48 cmp $680, %rdx 57 sub $0x20, %rdx 62 sub $0x20, %rdx 75 addq $0x20, %rdx 82 movq %rdx, %rcx 83 movq -8(%rsi, %rdx), %r11 84 lea -8(%rdi, %rdx), %r10 96 movq %rdx, %rcx 99 leaq -8(%rsi, %rdx), %rsi [all …]
|
D | memcpy_64.S | 36 movq %rdx, %rcx 55 movq %rdx, %rcx 63 cmpq $0x20, %rdx 72 subq $0x20, %rdx 74 subq $0x20, %rdx 98 addq %rdx, %rsi 99 addq %rdx, %rdi 100 subq $0x20, %rdx 107 subq $0x20, %rdx 124 subq %rdx, %rsi [all …]
|
D | hweight.S | 42 pushq %rdx 44 movq %rdi, %rdx # w -> t 46 shrq %rdx # t >>= 1 47 andq %rdx, %rax # t &= 0x5555555555555555 48 movabsq $0x3333333333333333, %rdx 53 andq %rdx, %rax # t &= 0x3333333333333333 54 andq %rdi, %rdx # w_tmp &= 0x3333333333333333 55 addq %rdx, %rax # w = w_tmp + t 57 movq %rax, %rdx # w -> t 58 shrq $4, %rdx # t >>= 4 [all …]
|
D | memset_64.S | 31 movq %rdx,%rcx 63 movq %rdx,%rcx 83 movq %rdx,%rcx 131 cmpq $7,%rdx 137 subq %r8,%rdx
|
D | copy_page_64.S | 35 movq 0x8*2(%rsi), %rdx 46 movq %rdx, 0x8*2(%rdi) 65 movq 0x8*2(%rsi), %rdx 74 movq %rdx, 0x8*2(%rdi)
|
D | getuser.S | 42 ALTERNATIVE __stringify(mov $((1 << 47) - 4096 - (n)),%rdx), \ 43 __stringify(mov $((1 << 56) - 4096 - (n)),%rdx), X86_FEATURE_LA57 100 4: movq (%_ASM_AX),%rdx 155 9: movq (%_ASM_AX),%rdx
|
/linux-6.1.9/tools/arch/x86/lib/ |
D | memcpy_64.S | 35 movq %rdx, %rcx 54 movq %rdx, %rcx 62 cmpq $0x20, %rdx 71 subq $0x20, %rdx 73 subq $0x20, %rdx 97 addq %rdx, %rsi 98 addq %rdx, %rdi 99 subq $0x20, %rdx 106 subq $0x20, %rdx 123 subq %rdx, %rsi [all …]
|
D | memset_64.S | 31 movq %rdx,%rcx 63 movq %rdx,%rcx 83 movq %rdx,%rcx 131 cmpq $7,%rdx 137 subq %r8,%rdx
|
/linux-6.1.9/arch/x86/crypto/ |
D | sm4-aesni-avx-asm_64.S | 159 vmovdqu 0*16(%rdx), RA0; 165 vmovdqu 1*16(%rdx), RA1; 167 vmovdqu 2*16(%rdx), RA2; 170 vmovdqu 3*16(%rdx), RA3; 380 vmovdqu (0 * 16)(%rdx), RA0; 381 vmovdqu (1 * 16)(%rdx), RA1; 382 vmovdqu (2 * 16)(%rdx), RA2; 383 vmovdqu (3 * 16)(%rdx), RA3; 384 vmovdqu (4 * 16)(%rdx), RB0; 389 vmovdqu (5 * 16)(%rdx), RB1; [all …]
|
D | sm4-aesni-avx2-asm_64.S | 371 vpxor (0 * 32)(%rdx), RA0, RA0; 372 vpxor (1 * 32)(%rdx), RA1, RA1; 373 vpxor (2 * 32)(%rdx), RA2, RA2; 374 vpxor (3 * 32)(%rdx), RA3, RA3; 375 vpxor (4 * 32)(%rdx), RB0, RB0; 376 vpxor (5 * 32)(%rdx), RB1, RB1; 377 vpxor (6 * 32)(%rdx), RB2, RB2; 378 vpxor (7 * 32)(%rdx), RB3, RB3; 410 vmovdqu (0 * 32)(%rdx), RA0; 411 vmovdqu (1 * 32)(%rdx), RA1; [all …]
|
D | cast5-avx-x86_64-asm_64.S | 63 #define RGI1 %rdx 374 vmovdqu (0*4*4)(%rdx), RL1; 375 vmovdqu (1*4*4)(%rdx), RR1; 376 vmovdqu (2*4*4)(%rdx), RL2; 377 vmovdqu (3*4*4)(%rdx), RR2; 378 vmovdqu (4*4*4)(%rdx), RL3; 379 vmovdqu (5*4*4)(%rdx), RR3; 380 vmovdqu (6*4*4)(%rdx), RL4; 381 vmovdqu (7*4*4)(%rdx), RR4; 412 vmovdqu (0*4*4)(%rdx), RL1; [all …]
|
D | chacha-avx512vl-x86_64.S | 30 # %rdx: up to 2 data blocks input, i 117 vpxord 0x00(%rdx),%xmm7,%xmm6 124 vpxord 0x10(%rdx),%xmm7,%xmm6 131 vpxord 0x20(%rdx),%xmm7,%xmm6 138 vpxord 0x30(%rdx),%xmm7,%xmm6 146 vpxord 0x40(%rdx),%xmm7,%xmm6 152 vpxord 0x50(%rdx),%xmm7,%xmm6 158 vpxord 0x60(%rdx),%xmm7,%xmm6 164 vpxord 0x70(%rdx),%xmm7,%xmm6 184 vmovdqu8 (%rdx,%r9),%xmm1{%k1}{z} [all …]
|
D | chacha-avx2-x86_64.S | 40 # %rdx: up to 2 data blocks input, i 144 vpxor 0x00(%rdx),%xmm7,%xmm6 151 vpxor 0x10(%rdx),%xmm7,%xmm6 158 vpxor 0x20(%rdx),%xmm7,%xmm6 165 vpxor 0x30(%rdx),%xmm7,%xmm6 173 vpxor 0x40(%rdx),%xmm7,%xmm6 179 vpxor 0x50(%rdx),%xmm7,%xmm6 185 vpxor 0x60(%rdx),%xmm7,%xmm6 191 vpxor 0x70(%rdx),%xmm7,%xmm6 211 lea (%rdx,%rax),%rsi [all …]
|
D | chacha-ssse3-x86_64.S | 117 # %rdx: up to 1 data block input, i 139 movdqu 0x00(%rdx),%xmm4 147 movdqu 0x10(%rdx),%xmm0 155 movdqu 0x20(%rdx),%xmm0 163 movdqu 0x30(%rdx),%xmm0 184 lea (%rdx,%rax),%rsi 226 # %rdx: up to 4 data blocks input, i 654 movdqu 0x00(%rdx),%xmm1 661 movdqu 0x10(%rdx),%xmm1 668 movdqu 0x20(%rdx),%xmm1 [all …]
|
/linux-6.1.9/arch/x86/mm/ |
D | mem_encrypt_boot.S | 44 movq %rdx, %r12 /* Area length */ 55 movq %r8, %rdx /* Pagetables used for encryption */ 99 mov %rdx, %cr3 102 mov %cr4, %rdx 103 andq $~X86_CR4_PGE, %rdx 104 mov %rdx, %cr4 105 orq $X86_CR4_PGE, %rdx 106 mov %rdx, %cr4 118 mov %rdx, %r15 /* Save original PAT value */ 151 mov %r15, %rdx /* Restore original PAT value */
|
/linux-6.1.9/arch/x86/power/ |
D | hibernate_asm_64.S | 34 movq %rax, %rdx 35 andq $~(X86_CR4_PGE), %rdx 36 movq %rdx, %cr4; # turn off PGE 49 movq pt_regs_dx(%rax), %rdx 80 movq %rdx, pt_regs_dx(%rax) 112 movq restore_pblist(%rip), %rdx 132 testq %rdx, %rdx 136 movq pbe_address(%rdx), %rsi 137 movq pbe_orig_address(%rdx), %rdi 143 movq pbe_next(%rdx), %rdx
|
/linux-6.1.9/arch/x86/entry/ |
D | calling.h | 68 .macro PUSH_REGS rdx=%rdx rcx=%rcx rax=%rax save_ret=0 77 pushq \rdx /* pt_regs->dx */ 120 .macro PUSH_AND_CLEAR_REGS rdx=%rdx rcx=%rcx rax=%rax save_ret=0 121 PUSH_REGS rdx=\rdx, rcx=\rcx, rax=\rax, save_ret=\save_ret 138 popq %rdx 306 shl $32, %rdx 307 or %rdx, %rax 316 movq PER_CPU_VAR(x86_spec_ctrl_current), %rdx 318 shr $32, %rdx 334 mov \save_reg, %rdx [all …]
|
D | entry_64.S | 1189 pushq %rdx 1207 SWITCH_TO_KERNEL_CR3 scratch_reg=%rdx 1208 movq %rsp, %rdx 1210 UNWIND_HINT_IRET_REGS base=%rdx offset=8 1211 pushq 5*8(%rdx) /* pt_regs->ss */ 1212 pushq 4*8(%rdx) /* pt_regs->rsp */ 1213 pushq 3*8(%rdx) /* pt_regs->flags */ 1214 pushq 2*8(%rdx) /* pt_regs->cs */ 1215 pushq 1*8(%rdx) /* pt_regs->rip */ 1218 PUSH_AND_CLEAR_REGS rdx=(%rdx) [all …]
|
/linux-6.1.9/arch/x86/kernel/ |
D | sev_verify_cbit.S | 38 movq %rsi, %rdx 39 andq $(~X86_CR4_PGE), %rdx 40 movq %rdx, %cr4 49 1: rdrand %rdx 53 movq %rdx, sev_check_data(%rip) 65 cmpq %rdx, sev_check_data(%rip)
|
D | ftrace_64.S | 84 movq %rdx, RDX(%rsp) 95 movq MCOUNT_REG_SIZE-8(%rsp), %rdx 97 movq %rbp, %rdx 99 movq %rdx, RBP(%rsp) 125 movq RDX(%rsp), %rdx 159 movq function_trace_op(%rip), %rdx 199 movq function_trace_op(%rip), %rdx 320 movq %rdx, 8(%rsp) 326 movq 8(%rsp), %rdx
|
D | relocate_kernel_64.S | 119 pushq %rdx 211 popq %rdx 214 call *%rdx 290 movq %rdi, %rdx 298 movq %rdx, %rsi 302 movq %rdx, %rdi
|
/linux-6.1.9/tools/testing/selftests/kvm/x86_64/ |
D | xen_vmcall_test.c | 33 unsigned long rdx = ARGVALUE(3); in guest_code() local 42 "a"(rax), "D"(rdi), "S"(rsi), "d"(rdx), in guest_code() 64 "a"(rax), "D"(rdi), "S"(rsi), "d"(rdx), in guest_code() 71 rdx = 0x5a5a5a5a; /* ingpa (badly aligned) */ in guest_code() 74 "a"(rax), "c"(rcx), "d"(rdx), in guest_code()
|
/linux-6.1.9/tools/testing/selftests/sgx/ |
D | test_encl_bootstrap.S | 66 xor %rdx, %rdx 75 add %rdx, %rdx # OF = SF = AF = CF = 0; ZF = PF = 1
|
/linux-6.1.9/arch/x86/kernel/acpi/ |
D | wakeup_64.S | 21 movq $0x123456789abcdef0, %rdx 22 cmpq %rdx, %rax 61 movq %rdx, pt_regs_dx(%rax) 108 movq pt_regs_dx(%rax), %rdx
|
/linux-6.1.9/arch/x86/purgatory/ |
D | entry64.S | 41 movq rdx(%rip), %rdx 64 rdx: .quad 0x0 label
|