/glibc-2.36/sysdeps/x86_64/multiarch/ |
D | strcat-sse2-unaligned.S | 44 xor %rax, %rax 55 mov %rdi, %rax 56 and $-16, %rax 59 mov %rdi, %rax 60 and $-16, %rax 61 pcmpeqb (%rax), %xmm0 63 sub %rax, %rcx 74 pcmpeqb 16(%rax), %xmm0 79 pcmpeqb 32(%rax), %xmm1 84 pcmpeqb 48(%rax), %xmm2 [all …]
|
D | strcat-avx2.S | 56 mov %rdi, %rax 57 and $-VEC_SIZE, %rax 60 mov %rdi, %rax 61 and $-VEC_SIZE, %rax 62 vpcmpeqb (%rax), %ymm6, %ymm0 64 sub %rax, %rcx 71 vpcmpeqb VEC_SIZE(%rax), %ymm6, %ymm0 76 vpcmpeqb (VEC_SIZE * 2)(%rax), %ymm6, %ymm1 81 vpcmpeqb (VEC_SIZE * 3)(%rax), %ymm6, %ymm2 86 vpcmpeqb (VEC_SIZE * 4)(%rax), %ymm6, %ymm3 [all …]
|
D | strcat-evex.S | 61 mov %rdi, %rax 62 and $-VEC_SIZE, %rax 65 mov %rdi, %rax 66 and $-VEC_SIZE, %rax 67 vpcmpb $0, (%rax), %YMMZERO, %k0 69 sub %rax, %rcx 76 vpcmpb $0, VEC_SIZE(%rax), %YMMZERO, %k0 81 vpcmpb $0, (VEC_SIZE * 2)(%rax), %YMMZERO, %k1 86 vpcmpb $0, (VEC_SIZE * 3)(%rax), %YMMZERO, %k2 91 vpcmpb $0, (VEC_SIZE * 4)(%rax), %YMMZERO, %k3 [all …]
|
D | strlen-sse2.S | 36 # define SHIFT_RETURN shrq $2, %rax 61 PCMPEQ (%rax), %xmm0; \ 62 PCMPEQ 16(%rax), %xmm1; \ 63 PCMPEQ 32(%rax), %xmm2; \ 64 PCMPEQ 48(%rax), %xmm3; \ 80 xor %rax, %rax 105 movq %rdi, %rax 117 subq %rax, %rsi; \ 118 andq $-64, %rax; \ 122 # define STRNLEN_PROLOG andq $-64, %rax; [all …]
|
D | strlen-evex-base.S | 46 # define RAX rax 102 cmpq %rsi, %rax 103 cmovnb %rsi, %rax 111 movq %rsi, %rax 116 leaq VEC_SIZE(%rdi), %rax 118 andq $-VEC_SIZE, %rax 120 movq %rax, %rdx 134 VPCMP $0, (%rax), %VMM0, %k0 144 VPCMP $0, VEC_SIZE(%rax), %VMM0, %k0 154 VPCMP $0, (VEC_SIZE * 2)(%rax), %VMM0, %k0 [all …]
|
D | strcat-sse2.S | 36 movq %rdi, %rax /* Duplicate destination pointer. */ 45 0: cmpb $0x0,(%rax) /* is byte NUL? */ 47 incq %rax /* increment pointer */ 57 movq (%rax), %rcx /* get double word (= 8 bytes) in question */ 58 addq $8,%rax /* adjust pointer for next word */ 71 movq (%rax), %rcx /* get double word (= 8 bytes) in question */ 72 addq $8,%rax /* adjust pointer for next word */ 85 movq (%rax), %rcx /* get double word (= 8 bytes) in question */ 86 addq $8,%rax /* adjust pointer for next word */ 99 movq (%rax), %rcx /* get double word (= 8 bytes) in question */ [all …]
|
D | memrchr-evex.S | 49 leaq -1(%rdi, %rdx), %rax 71 subq %rcx, %rax 81 decq %rax 84 subq %rcx, %rax 93 andq $-VEC_SIZE, %rax 96 movq %rax, %rdx 99 vpcmpb $0, -(VEC_SIZE)(%rax), %VECMATCH, %k0 109 decq %rax 124 subq %rcx, %rax 132 movq %rax, %rsi [all …]
|
D | wcslen-sse2.S | 51 lea 32(%rdi), %rax 53 and $-16, %rax 55 pcmpeqd (%rax), %xmm0 58 addq $16, %rax 62 pcmpeqd (%rax), %xmm1 65 addq $16, %rax 69 pcmpeqd (%rax), %xmm2 72 addq $16, %rax 76 pcmpeqd (%rax), %xmm3 78 addq $16, %rax [all …]
|
D | memrchr-avx2.S | 53 leaq -1(%rdx, %rdi), %rax 61 vpcmpeqb -(VEC_SIZE - 1)(%rax), %ymm0, %ymm1 76 subq %rcx, %rax 87 subq %rcx, %rax 97 andq $-VEC_SIZE, %rax 100 movq %rax, %rdx 102 vpcmpeqb -(VEC_SIZE)(%rax), %ymm0, %ymm1 104 decq %rax 116 vpcmpeqb -(VEC_SIZE * 2 - 1)(%rax), %ymm0, %ymm1 123 subq %rcx, %rax [all …]
|
D | wcschr-sse2.S | 49 pmovmskb %xmm0, %rax 50 or %rax, %rdx 60 pmovmskb %xmm0, %rax 61 or %rax, %rdx 73 pmovmskb %xmm0, %rax 76 sar %cl, %rax 77 test %rax, %rax 80 bsf %rax, %rax 84 cmp %rdx, %rax 88 add %rdi, %rax [all …]
|
D | strstr-sse2-unaligned.S | 31 movq %rdi, %rax 34 cmpq $4031, %rax 60 salq $16, %rax 61 orq %rax, %r8 64 bsf %r8, %rax 65 addq %rdi, %rax 66 cmpb $0, (%rax) 71 cmpb 2(%rax), %dl 84 cmpb 2(%rax,%rdx), %cl 98 leaq -1(%r8), %rax [all …]
|
D | strrchr-avx2.S | 82 addq %rdi, %rax 87 andq $-CHAR_SIZE, %rax 110 addq %r8, %rax 112 andq $-CHAR_SIZE, %rax 128 leaq 1(%rdi, %rax), %rax 130 andq $-CHAR_SIZE, %rax 146 leaq (VEC_SIZE + 1)(%rdi, %rax), %rax 148 andq $-CHAR_SIZE, %rax 215 addq %rax, %rdx 221 salq $32, %rax [all …]
|
D | strchr-sse2.S | 52 leaq (%rdi,%rax), %rax 55 leaq (%rdi,%rax), %rax 56 cmpb %sil, (%rax) 57 cmovne %rdx, %rax 78 salq $32, %rax 80 orq %rcx, %rax 84 orq %rcx, %rax 85 testq %rax, %rax 128 salq $16, %rax 132 orq %r8, %rax [all …]
|
D | strcpy-sse2.S | 59 movq (%rsi), %rax /* Read double word (8 bytes). */ 61 movq %rax, %r9 /* Save a copy for NUL finding. */ 66 xorq %rax, %r9 /* (word+magic)^word */ 73 movq %rax, (%rdx) /* Write value to destination. */ 77 movq (%rsi), %rax /* Read double word (8 bytes). */ 79 movq %rax, %r9 /* Save a copy for NUL finding. */ 84 xorq %rax, %r9 /* (word+magic)^word */ 91 movq %rax, (%rdx) /* Write value to destination. */ 95 movq (%rsi), %rax /* Read double word (8 bytes). */ 97 movq %rax, %r9 /* Save a copy for NUL finding. */ [all …]
|
D | memcmp-sse2.S | 114 shlq $32, %rax 118 orq %rdi, %rax 120 cmpq %rcx, %rax 129 movq (%rsi), %rax 130 subq (%rdi), %rax 134 orq %rcx, %rax 141 movq (%rdi), %rax 143 cmpq %rcx, %rax 147 movq -8(%rdi, %rdx, CHAR_SIZE), %rax 149 cmpq %rcx, %rax [all …]
|
D | strrchr-evex.S | 50 # define maskm_2x rax 108 leaq (%rdi, %rax, CHAR_SIZE), %rax 110 addq %rdi, %rax 136 leaq (%rsi, %rax, CHAR_SIZE), %rax 138 addq %rsi, %rax 155 kmovq %k3, %rax 156 bsrq %rax, %rax 157 leaq (VEC_SIZE)(%r8, %rax, CHAR_SIZE), %rax 169 leaq (VEC_SIZE * 3)(%rdi, %rax, CHAR_SIZE), %rax 182 leaq (VEC_SIZE)(%rdi, %rax, CHAR_SIZE), %rax [all …]
|
D | memchr-evex.S | 124 leaq (%rdi, %rax, CHAR_SIZE), %rax 126 addq %rdi, %rax 148 cmovg %rdi, %rax 187 leaq (%RAW_PTR_REG, %rax, CHAR_SIZE), %rax 189 addq %RAW_PTR_REG, %rax 196 leaq VEC_SIZE(%rdi, %rax, CHAR_SIZE), %rax 202 leaq (VEC_SIZE * 2)(%rdi, %rax, CHAR_SIZE), %rax 208 leaq (VEC_SIZE * 3)(%rdi, %rax, CHAR_SIZE), %rax 214 leaq (VEC_SIZE * 4)(%rdi, %rax, CHAR_SIZE), %rax 362 addq %rax, %rcx [all …]
|
D | strchr-evex.S | 93 cmp (%rdi, %rax, CHAR_SIZE), %CHAR_REG 106 leaq (%rdi, %rax, CHAR_SIZE), %rax 108 addq %rdi, %rax 132 leaq (VEC_SIZE * 4)(%rdi, %rax, CHAR_SIZE), %rax 149 cmp (VEC_SIZE)(%rdi, %rax, CHAR_SIZE), %CHAR_REG 155 leaq (VEC_SIZE)(%rdi, %rax, CHAR_SIZE), %rax 176 leaq (VEC_SIZE * 2)(%rdi, %rax, CHAR_SIZE), %rax 186 cmp (VEC_SIZE * 3)(%rdi, %rax, CHAR_SIZE), %CHAR_REG 191 leaq (VEC_SIZE * 3)(%rdi, %rax, CHAR_SIZE), %rax 305 orq %rcx, %rax [all …]
|
D | strlen-evex.S | 99 btsq %rsi, %rax 112 leal -(CHAR_PER_VEC * 4 + 1)(%rcx, %rax), %eax 119 leal CHAR_PER_VEC(%rdi, %rax), %eax 131 leal -(CHAR_PER_VEC * 3 + 1)(%rcx, %rax), %eax 138 leal (CHAR_PER_VEC * 2)(%rdi, %rax), %eax 150 leal -(CHAR_PER_VEC * 2 + 1)(%rcx, %rax), %eax 157 leal (CHAR_PER_VEC * 3)(%rdi, %rax), %eax 169 leal -(CHAR_PER_VEC + 1)(%rcx, %rax), %eax 176 leal (CHAR_PER_VEC * 4)(%rdi, %rax), %eax 282 salq $CHAR_PER_VEC, %rax [all …]
|
/glibc-2.36/sysdeps/x86_64/fpu/multiarch/ |
D | svml_d_pow8_core_avx512.S | 94 movq __svml_dpow_data@GOTPCREL(%rip), %rax 103 vmovups _dbOne(%rax), %zmm6 109 vpsubd _i3fe7fe00(%rax), %zmm10, %zmm14{%k1} 110 vpandd _iIndexMask(%rax), %zmm10, %zmm5{%k1} 113 vpaddd _HIDELTA(%rax), %zmm10, %zmm3{%k1} 114 vpaddd _iIndexAdd(%rax), %zmm5, %zmm5{%k1} 116 vpaddd _i2p20_2p19(%rax), %zmm14, %zmm14{%k1} 117 vpcmpd $1, _LORANGE(%rax), %zmm3, %k2{%k1} 119 vpandd _ABSMASK(%rax), %zmm15, %zmm2{%k1} 125 vgatherdpd 11712(%rax,%ymm8), %zmm9{%k2} [all …]
|
D | svml_d_pow4_core_avx2.S | 92 movq __svml_dpow_data@GOTPCREL(%rip), %rax 104 vandps _iIndexMask(%rax), %xmm5, %xmm3 105 vpaddd _iIndexAdd(%rax), %xmm3, %xmm6 115 vandpd _iMantissaMask(%rax), %ymm0, %ymm4 116 vorpd _dbOne(%rax), %ymm4, %ymm13 121 vpsubd _i3fe7fe00(%rax), %xmm5, %xmm3 122 vpaddd _HIDELTA(%rax), %xmm5, %xmm5 125 vpand _ABSMASK(%rax), %xmm2, %xmm10 127 vgatherdpd %ymm2, 11712(%rax,%xmm9), %ymm1 128 vmovups _LORANGE(%rax), %xmm7 [all …]
|
D | svml_d_sincos8_core_avx512.S | 47 movq __svml_d_trig_data@GOTPCREL(%rip), %rax 50 vmovups __dSignMask(%rax), %zmm12 51 vmovups __dInvPI(%rax), %zmm5 56 vmovups __dPI1_FMA(%rax), %zmm7 57 vmovups __dPI3_FMA(%rax), %zmm9 66 vfmadd213pd __dRShifter(%rax), %zmm3, %zmm5 67 vmovups __dC6(%rax), %zmm13 70 vsubpd __dRShifter(%rax), %zmm5, %zmm1 75 vcmppd $22, __dRangeVal(%rax), %zmm3, %k1 82 vfnmadd231pd __dPI2_FMA(%rax), %zmm1, %zmm8 [all …]
|
D | svml_s_sincosf8_core_avx2.S | 59 movq __svml_s_trig_data@GOTPCREL(%rip), %rax 62 vmovups __sAbsMask(%rax), %ymm2 63 vmovups __sInvPI(%rax), %ymm1 64 vmovups __sPI1_FMA(%rax), %ymm13 72 vfmadd213ps __sRShifter(%rax), %ymm4, %ymm1 78 vsubps __sRShifter(%rax), %ymm1, %ymm1 84 vfnmadd231ps __sPI2_FMA(%rax), %ymm1, %ymm7 85 vandps __sSignMask(%rax), %ymm7, %ymm15 86 vxorps __sOneHalf(%rax), %ymm15, %ymm6 92 vmovups __sPI3_FMA(%rax), %ymm13 [all …]
|
/glibc-2.36/sysdeps/x86_64/ |
D | mul_1.S | 36 mov (up), %rax /* read first u limb early */ 41 add %r10, %rax 52 mov %rax, (rp) 59 mov %rax, %r9 60 mov (up,n,8), %rax 68 mov %rax, %r8 75 mov %rax, %rbx 84 mov %rax, %r10 85 mov 24(up,n,8), %rax 91 add %rax, %r9 [all …]
|
D | dl-tlsdesc.S | 41 movq 8(%rax), %rax 63 movq 8(%rax), %rax 64 subq %fs:0, %rax 107 movq TLSDESC_ARG(%rax), %rdi 108 movq (%rsi), %rax 109 cmpq %rax, TLSDESC_GEN_COUNT(%rdi) 111 movq TLSDESC_MODID(%rdi), %rax 112 salq $4, %rax 113 movq (%rax,%rsi), %rax 114 cmpq $-1, %rax [all …]
|