Home
last modified time | relevance | path

Searched refs:rax (Results 1 – 25 of 164) sorted by relevance

1234567

/glibc-2.36/sysdeps/x86_64/multiarch/
Dstrcat-sse2-unaligned.S44 xor %rax, %rax
55 mov %rdi, %rax
56 and $-16, %rax
59 mov %rdi, %rax
60 and $-16, %rax
61 pcmpeqb (%rax), %xmm0
63 sub %rax, %rcx
74 pcmpeqb 16(%rax), %xmm0
79 pcmpeqb 32(%rax), %xmm1
84 pcmpeqb 48(%rax), %xmm2
[all …]
Dstrcat-avx2.S56 mov %rdi, %rax
57 and $-VEC_SIZE, %rax
60 mov %rdi, %rax
61 and $-VEC_SIZE, %rax
62 vpcmpeqb (%rax), %ymm6, %ymm0
64 sub %rax, %rcx
71 vpcmpeqb VEC_SIZE(%rax), %ymm6, %ymm0
76 vpcmpeqb (VEC_SIZE * 2)(%rax), %ymm6, %ymm1
81 vpcmpeqb (VEC_SIZE * 3)(%rax), %ymm6, %ymm2
86 vpcmpeqb (VEC_SIZE * 4)(%rax), %ymm6, %ymm3
[all …]
Dstrcat-evex.S61 mov %rdi, %rax
62 and $-VEC_SIZE, %rax
65 mov %rdi, %rax
66 and $-VEC_SIZE, %rax
67 vpcmpb $0, (%rax), %YMMZERO, %k0
69 sub %rax, %rcx
76 vpcmpb $0, VEC_SIZE(%rax), %YMMZERO, %k0
81 vpcmpb $0, (VEC_SIZE * 2)(%rax), %YMMZERO, %k1
86 vpcmpb $0, (VEC_SIZE * 3)(%rax), %YMMZERO, %k2
91 vpcmpb $0, (VEC_SIZE * 4)(%rax), %YMMZERO, %k3
[all …]
Dstrlen-sse2.S36 # define SHIFT_RETURN shrq $2, %rax
61 PCMPEQ (%rax), %xmm0; \
62 PCMPEQ 16(%rax), %xmm1; \
63 PCMPEQ 32(%rax), %xmm2; \
64 PCMPEQ 48(%rax), %xmm3; \
80 xor %rax, %rax
105 movq %rdi, %rax
117 subq %rax, %rsi; \
118 andq $-64, %rax; \
122 # define STRNLEN_PROLOG andq $-64, %rax;
[all …]
Dstrlen-evex-base.S46 # define RAX rax
102 cmpq %rsi, %rax
103 cmovnb %rsi, %rax
111 movq %rsi, %rax
116 leaq VEC_SIZE(%rdi), %rax
118 andq $-VEC_SIZE, %rax
120 movq %rax, %rdx
134 VPCMP $0, (%rax), %VMM0, %k0
144 VPCMP $0, VEC_SIZE(%rax), %VMM0, %k0
154 VPCMP $0, (VEC_SIZE * 2)(%rax), %VMM0, %k0
[all …]
Dstrcat-sse2.S36 movq %rdi, %rax /* Duplicate destination pointer. */
45 0: cmpb $0x0,(%rax) /* is byte NUL? */
47 incq %rax /* increment pointer */
57 movq (%rax), %rcx /* get double word (= 8 bytes) in question */
58 addq $8,%rax /* adjust pointer for next word */
71 movq (%rax), %rcx /* get double word (= 8 bytes) in question */
72 addq $8,%rax /* adjust pointer for next word */
85 movq (%rax), %rcx /* get double word (= 8 bytes) in question */
86 addq $8,%rax /* adjust pointer for next word */
99 movq (%rax), %rcx /* get double word (= 8 bytes) in question */
[all …]
Dmemrchr-evex.S49 leaq -1(%rdi, %rdx), %rax
71 subq %rcx, %rax
81 decq %rax
84 subq %rcx, %rax
93 andq $-VEC_SIZE, %rax
96 movq %rax, %rdx
99 vpcmpb $0, -(VEC_SIZE)(%rax), %VECMATCH, %k0
109 decq %rax
124 subq %rcx, %rax
132 movq %rax, %rsi
[all …]
Dwcslen-sse2.S51 lea 32(%rdi), %rax
53 and $-16, %rax
55 pcmpeqd (%rax), %xmm0
58 addq $16, %rax
62 pcmpeqd (%rax), %xmm1
65 addq $16, %rax
69 pcmpeqd (%rax), %xmm2
72 addq $16, %rax
76 pcmpeqd (%rax), %xmm3
78 addq $16, %rax
[all …]
Dmemrchr-avx2.S53 leaq -1(%rdx, %rdi), %rax
61 vpcmpeqb -(VEC_SIZE - 1)(%rax), %ymm0, %ymm1
76 subq %rcx, %rax
87 subq %rcx, %rax
97 andq $-VEC_SIZE, %rax
100 movq %rax, %rdx
102 vpcmpeqb -(VEC_SIZE)(%rax), %ymm0, %ymm1
104 decq %rax
116 vpcmpeqb -(VEC_SIZE * 2 - 1)(%rax), %ymm0, %ymm1
123 subq %rcx, %rax
[all …]
Dwcschr-sse2.S49 pmovmskb %xmm0, %rax
50 or %rax, %rdx
60 pmovmskb %xmm0, %rax
61 or %rax, %rdx
73 pmovmskb %xmm0, %rax
76 sar %cl, %rax
77 test %rax, %rax
80 bsf %rax, %rax
84 cmp %rdx, %rax
88 add %rdi, %rax
[all …]
Dstrstr-sse2-unaligned.S31 movq %rdi, %rax
34 cmpq $4031, %rax
60 salq $16, %rax
61 orq %rax, %r8
64 bsf %r8, %rax
65 addq %rdi, %rax
66 cmpb $0, (%rax)
71 cmpb 2(%rax), %dl
84 cmpb 2(%rax,%rdx), %cl
98 leaq -1(%r8), %rax
[all …]
Dstrrchr-avx2.S82 addq %rdi, %rax
87 andq $-CHAR_SIZE, %rax
110 addq %r8, %rax
112 andq $-CHAR_SIZE, %rax
128 leaq 1(%rdi, %rax), %rax
130 andq $-CHAR_SIZE, %rax
146 leaq (VEC_SIZE + 1)(%rdi, %rax), %rax
148 andq $-CHAR_SIZE, %rax
215 addq %rax, %rdx
221 salq $32, %rax
[all …]
Dstrchr-sse2.S52 leaq (%rdi,%rax), %rax
55 leaq (%rdi,%rax), %rax
56 cmpb %sil, (%rax)
57 cmovne %rdx, %rax
78 salq $32, %rax
80 orq %rcx, %rax
84 orq %rcx, %rax
85 testq %rax, %rax
128 salq $16, %rax
132 orq %r8, %rax
[all …]
Dstrcpy-sse2.S59 movq (%rsi), %rax /* Read double word (8 bytes). */
61 movq %rax, %r9 /* Save a copy for NUL finding. */
66 xorq %rax, %r9 /* (word+magic)^word */
73 movq %rax, (%rdx) /* Write value to destination. */
77 movq (%rsi), %rax /* Read double word (8 bytes). */
79 movq %rax, %r9 /* Save a copy for NUL finding. */
84 xorq %rax, %r9 /* (word+magic)^word */
91 movq %rax, (%rdx) /* Write value to destination. */
95 movq (%rsi), %rax /* Read double word (8 bytes). */
97 movq %rax, %r9 /* Save a copy for NUL finding. */
[all …]
Dmemcmp-sse2.S114 shlq $32, %rax
118 orq %rdi, %rax
120 cmpq %rcx, %rax
129 movq (%rsi), %rax
130 subq (%rdi), %rax
134 orq %rcx, %rax
141 movq (%rdi), %rax
143 cmpq %rcx, %rax
147 movq -8(%rdi, %rdx, CHAR_SIZE), %rax
149 cmpq %rcx, %rax
[all …]
Dstrrchr-evex.S50 # define maskm_2x rax
108 leaq (%rdi, %rax, CHAR_SIZE), %rax
110 addq %rdi, %rax
136 leaq (%rsi, %rax, CHAR_SIZE), %rax
138 addq %rsi, %rax
155 kmovq %k3, %rax
156 bsrq %rax, %rax
157 leaq (VEC_SIZE)(%r8, %rax, CHAR_SIZE), %rax
169 leaq (VEC_SIZE * 3)(%rdi, %rax, CHAR_SIZE), %rax
182 leaq (VEC_SIZE)(%rdi, %rax, CHAR_SIZE), %rax
[all …]
Dmemchr-evex.S124 leaq (%rdi, %rax, CHAR_SIZE), %rax
126 addq %rdi, %rax
148 cmovg %rdi, %rax
187 leaq (%RAW_PTR_REG, %rax, CHAR_SIZE), %rax
189 addq %RAW_PTR_REG, %rax
196 leaq VEC_SIZE(%rdi, %rax, CHAR_SIZE), %rax
202 leaq (VEC_SIZE * 2)(%rdi, %rax, CHAR_SIZE), %rax
208 leaq (VEC_SIZE * 3)(%rdi, %rax, CHAR_SIZE), %rax
214 leaq (VEC_SIZE * 4)(%rdi, %rax, CHAR_SIZE), %rax
362 addq %rax, %rcx
[all …]
Dstrchr-evex.S93 cmp (%rdi, %rax, CHAR_SIZE), %CHAR_REG
106 leaq (%rdi, %rax, CHAR_SIZE), %rax
108 addq %rdi, %rax
132 leaq (VEC_SIZE * 4)(%rdi, %rax, CHAR_SIZE), %rax
149 cmp (VEC_SIZE)(%rdi, %rax, CHAR_SIZE), %CHAR_REG
155 leaq (VEC_SIZE)(%rdi, %rax, CHAR_SIZE), %rax
176 leaq (VEC_SIZE * 2)(%rdi, %rax, CHAR_SIZE), %rax
186 cmp (VEC_SIZE * 3)(%rdi, %rax, CHAR_SIZE), %CHAR_REG
191 leaq (VEC_SIZE * 3)(%rdi, %rax, CHAR_SIZE), %rax
305 orq %rcx, %rax
[all …]
Dstrlen-evex.S99 btsq %rsi, %rax
112 leal -(CHAR_PER_VEC * 4 + 1)(%rcx, %rax), %eax
119 leal CHAR_PER_VEC(%rdi, %rax), %eax
131 leal -(CHAR_PER_VEC * 3 + 1)(%rcx, %rax), %eax
138 leal (CHAR_PER_VEC * 2)(%rdi, %rax), %eax
150 leal -(CHAR_PER_VEC * 2 + 1)(%rcx, %rax), %eax
157 leal (CHAR_PER_VEC * 3)(%rdi, %rax), %eax
169 leal -(CHAR_PER_VEC + 1)(%rcx, %rax), %eax
176 leal (CHAR_PER_VEC * 4)(%rdi, %rax), %eax
282 salq $CHAR_PER_VEC, %rax
[all …]
/glibc-2.36/sysdeps/x86_64/fpu/multiarch/
Dsvml_d_pow8_core_avx512.S94 movq __svml_dpow_data@GOTPCREL(%rip), %rax
103 vmovups _dbOne(%rax), %zmm6
109 vpsubd _i3fe7fe00(%rax), %zmm10, %zmm14{%k1}
110 vpandd _iIndexMask(%rax), %zmm10, %zmm5{%k1}
113 vpaddd _HIDELTA(%rax), %zmm10, %zmm3{%k1}
114 vpaddd _iIndexAdd(%rax), %zmm5, %zmm5{%k1}
116 vpaddd _i2p20_2p19(%rax), %zmm14, %zmm14{%k1}
117 vpcmpd $1, _LORANGE(%rax), %zmm3, %k2{%k1}
119 vpandd _ABSMASK(%rax), %zmm15, %zmm2{%k1}
125 vgatherdpd 11712(%rax,%ymm8), %zmm9{%k2}
[all …]
Dsvml_d_pow4_core_avx2.S92 movq __svml_dpow_data@GOTPCREL(%rip), %rax
104 vandps _iIndexMask(%rax), %xmm5, %xmm3
105 vpaddd _iIndexAdd(%rax), %xmm3, %xmm6
115 vandpd _iMantissaMask(%rax), %ymm0, %ymm4
116 vorpd _dbOne(%rax), %ymm4, %ymm13
121 vpsubd _i3fe7fe00(%rax), %xmm5, %xmm3
122 vpaddd _HIDELTA(%rax), %xmm5, %xmm5
125 vpand _ABSMASK(%rax), %xmm2, %xmm10
127 vgatherdpd %ymm2, 11712(%rax,%xmm9), %ymm1
128 vmovups _LORANGE(%rax), %xmm7
[all …]
Dsvml_d_sincos8_core_avx512.S47 movq __svml_d_trig_data@GOTPCREL(%rip), %rax
50 vmovups __dSignMask(%rax), %zmm12
51 vmovups __dInvPI(%rax), %zmm5
56 vmovups __dPI1_FMA(%rax), %zmm7
57 vmovups __dPI3_FMA(%rax), %zmm9
66 vfmadd213pd __dRShifter(%rax), %zmm3, %zmm5
67 vmovups __dC6(%rax), %zmm13
70 vsubpd __dRShifter(%rax), %zmm5, %zmm1
75 vcmppd $22, __dRangeVal(%rax), %zmm3, %k1
82 vfnmadd231pd __dPI2_FMA(%rax), %zmm1, %zmm8
[all …]
Dsvml_s_sincosf8_core_avx2.S59 movq __svml_s_trig_data@GOTPCREL(%rip), %rax
62 vmovups __sAbsMask(%rax), %ymm2
63 vmovups __sInvPI(%rax), %ymm1
64 vmovups __sPI1_FMA(%rax), %ymm13
72 vfmadd213ps __sRShifter(%rax), %ymm4, %ymm1
78 vsubps __sRShifter(%rax), %ymm1, %ymm1
84 vfnmadd231ps __sPI2_FMA(%rax), %ymm1, %ymm7
85 vandps __sSignMask(%rax), %ymm7, %ymm15
86 vxorps __sOneHalf(%rax), %ymm15, %ymm6
92 vmovups __sPI3_FMA(%rax), %ymm13
[all …]
/glibc-2.36/sysdeps/x86_64/
Dmul_1.S36 mov (up), %rax /* read first u limb early */
41 add %r10, %rax
52 mov %rax, (rp)
59 mov %rax, %r9
60 mov (up,n,8), %rax
68 mov %rax, %r8
75 mov %rax, %rbx
84 mov %rax, %r10
85 mov 24(up,n,8), %rax
91 add %rax, %r9
[all …]
Ddl-tlsdesc.S41 movq 8(%rax), %rax
63 movq 8(%rax), %rax
64 subq %fs:0, %rax
107 movq TLSDESC_ARG(%rax), %rdi
108 movq (%rsi), %rax
109 cmpq %rax, TLSDESC_GEN_COUNT(%rdi)
111 movq TLSDESC_MODID(%rdi), %rax
112 salq $4, %rax
113 movq (%rax,%rsi), %rax
114 cmpq $-1, %rax
[all …]

1234567