/glibc-2.36/sysdeps/x86_64/multiarch/ |
D | memcmpeq-evex.S | 81 VPCMP $4,(%rdi), %YMM2, %k1{%k2} 82 kmovd %k1, %eax 88 VPCMP $4, -(VEC_SIZE * 1)(%rdi, %rdx), %YMM1, %k1 89 kmovd %k1, %eax 100 VPCMP $4,(%rdi), %YMM1, %k1 101 kmovd %k1, %eax 110 VPCMP $4, VEC_SIZE(%rdi), %YMM2, %k1 111 kmovd %k1, %eax 121 VPCMP $4,(VEC_SIZE * 2)(%rdi), %YMM3, %k1 122 kmovd %k1, %eax [all …]
|
D | strrchr-evex.S | 96 VPCMP $0, %YMMMATCH, %YMM1, %k1 97 kmovd %k1, %eax 119 VPCMP $0, %YMMMATCH, %YMM2, %k1 120 kmovd %k1, %eax 130 VPCMP $0, %YMMMATCH, %YMM1, %k1 131 kmovd %k1, %eax 162 VPCMP $0, %YMMMATCH, %YMM4, %k1 163 kmovd %k1, %eax 174 VPCMP $0, %YMMMATCH, %YMM2, %k1 175 kmovd %k1, %eax [all …]
|
D | memcmp-evex-movbe.S | 122 VPCMP $4,(%rdi), %YMM2, %k1{%k2} 123 kmovd %k1, %eax 152 VPCMP $4,(%rdi), %YMM1, %k1 153 kmovd %k1, %eax 165 VPCMP $4, VEC_SIZE(%rdi), %YMM2, %k1 166 kmovd %k1, %eax 176 VPCMP $4,(VEC_SIZE * 2)(%rdi), %YMM3, %k1 177 kmovd %k1, %eax 182 VPCMP $4,(VEC_SIZE * 3)(%rdi), %YMM4, %k1 183 kmovd %k1, %ecx [all …]
|
D | strcmp-evex.S | 295 CMP_R1_S2_YMM (%YMM0, (%rsi), %YMM1, %k1){%k2} 296 kmovd %k1, %ecx 470 CMP_R1_S2_YMM (%YMM0, VEC_SIZE(%rsi), %YMM1, %k1){%k2} 471 kmovd %k1, %ecx 482 CMP_R1_S2_YMM (%YMM0, (VEC_SIZE * 2)(%rsi), %YMM1, %k1){%k2} 483 kmovd %k1, %ecx 489 CMP_R1_S2_YMM (%YMM0, (VEC_SIZE * 3)(%rsi), %YMM1, %k1){%k2} 490 kmovd %k1, %ecx 580 VPTESTM %YMM9, %YMM9, %k1 607 VPTESTNM %YMM6, %YMM6, %k0{%k1} [all …]
|
D | strcat-evex.S | 81 vpcmpb $0, (VEC_SIZE * 2)(%rax), %YMMZERO, %k1 82 kmovd %k1, %edx 102 vpcmpb $0, (VEC_SIZE * 2)(%rax), %YMMZERO, %k1 103 kmovd %k1, %edx 123 vpcmpb $0, (VEC_SIZE * 2)(%rax), %YMMZERO, %k1 124 kmovd %k1, %edx 144 vpcmpb $0, (VEC_SIZE * 2)(%rax), %YMMZERO, %k1 145 kmovd %k1, %edx 189 vpcmpb $0, VEC_SIZE(%rax), %YMMZERO, %k1 191 kmovd %k1, %edx [all …]
|
D | strchr-evex.S | 120 kmovd %k1, %ecx 126 kord %k0, %k1, %k0 164 kmovd %k1, %ecx 170 kord %k0, %k1, %k0 221 VPTESTN %YMM1, %YMM1, %k1 222 kortestd %k0, %k1 239 VPTESTN %YMM1, %YMM1, %k1 240 kortestd %k0, %k1 280 VPTESTN %YMM4, %YMM4, %k1 281 kmovd %k1, %ecx
|
D | strstr-avx512.c | 145 __mmask64 k1 = _mm512_cmpeq_epi8_mask (hay0, ned1); in __strstr_avx512() local 146 k1 = kshiftri_mask64 (k1, 1); in __strstr_avx512() 148 uint64_t k2 = cvtmask64_u64 (kand_mask64 (k0, k1)) & cmpmask; in __strstr_avx512() 189 k1 = _mm512_cmpeq_epi8_mask (hay1, ned1); in __strstr_avx512() 191 k2 = cvtmask64_u64 (kand_mask64 (k0, k1)) & cmpmask; in __strstr_avx512()
|
D | memchr-evex.S | 311 VPCMP $4, (VEC_SIZE * 4)(%rdi), %YMMMATCH, %k1 317 VPMINU %YMM2, %YMM3, %YMM3{%k1}{z} 325 VPCMP $4, (%rdi), %YMMMATCH, %k1 327 VPCMP $0, (%rdi), %YMMMATCH, %k1 337 vpternlogd $1, %ymm2, %ymm3, %ymm4{%k1}{z} 343 kmovd %k1, %eax 431 kmovd %k1, %eax
|
D | strlen-evex.S | 260 VPCMP $0, %YMM4, %YMMZERO, %k1 262 kortestd %k0, %k1 342 kunpckbw %k0, %k1, %k0 346 kunpckdq %k0, %k1, %k0
|
D | strlen-evex-base.S | 208 VPTESTN %VMM4, %VMM4, %k1 211 KORTEST %k0, %k1 231 KMOV %k1, %RCX
|
D | memrchr-evex.S | 246 vpcmpb $4, (VEC_SIZE * 3)(%rax), %VECMATCH, %k1 255 vpminub %VEC(2), %VEC(3), %VEC(3){%k1}{z} 307 kmovd %k1, %ecx
|
/glibc-2.36/sysdeps/x86_64/fpu/multiarch/ |
D | svml_d_pow8_core_avx512.S | 98 kmovw %edx, %k1 109 vpsubd _i3fe7fe00(%rax), %zmm10, %zmm14{%k1} 110 vpandd _iIndexMask(%rax), %zmm10, %zmm5{%k1} 111 vpsrad $20, %zmm14, %zmm14{%k1} 113 vpaddd _HIDELTA(%rax), %zmm10, %zmm3{%k1} 114 vpaddd _iIndexAdd(%rax), %zmm5, %zmm5{%k1} 116 vpaddd _i2p20_2p19(%rax), %zmm14, %zmm14{%k1} 117 vpcmpd $1, _LORANGE(%rax), %zmm3, %k2{%k1} 118 vpsrld $10, %zmm5, %zmm5{%k1} 119 vpandd _ABSMASK(%rax), %zmm15, %zmm2{%k1} [all …]
|
D | svml_s_powf16_core_avx512.S | 114 vpcmpd $5, _NMAXVAL(%rdx), %zmm3, %k1 125 vpbroadcastd %eax, %zmm7{%k1}{z} 126 kxnorw %k1, %k1, %k1 157 vgatherqpd _Log2Rcp_lookup(%rdx,%zmm12), %zmm6{%k1} 203 vpcmpgtd __iDomainRange(%rdx), %zmm11, %k1 205 vpbroadcastd %eax, %zmm10{%k1}{z} 371 vpcmpd $1, _NMAXVAL(%rax), %zmm9, %k1 385 vpandnd %zmm9, %zmm9, %zmm26{%k1} 387 kxnorw %k1, %k1, %k1 426 vgatherdpd _Log2Rcp_lookup(%rax,%ymm5), %zmm11{%k1} [all …]
|
D | svml_s_atanf16_core_avx512.S | 60 vcmpps $29, {sae}, %zmm3, %zmm7, %k1 71 vblendmps MOne+__svml_satan_data_internal_avx512(%rip), %zmm5, %zmm9{%k1} 75 vminps {sae}, %zmm7, %zmm6, %zmm8{%k1} 86 vblendmps Pi2+__svml_satan_data_internal_avx512(%rip), %zmm3, %zmm9{%k1}
|
D | svml_d_log8_core_avx512.S | 62 vcmppd $17, _MinNorm(%rdx), %zmm0, %k1 66 vpbroadcastq %rax, %zmm6{%k1}{z} 81 vcmppd $30, _Threshold(%rdx), %zmm11, %k1 96 vpbroadcastq %rax, %zmm12{%k1}{z} 262 vcmppd $21, _MinNorm(%rax), %zmm3, %k1 297 vpandnq %zmm3, %zmm3, %zmm8{%k1} 298 vcmppd $21, %zmm0, %zmm14, %k1 299 vpandnq %zmm14, %zmm14, %zmm1{%k1}
|
D | svml_s_atan2f16_core_avx512.S | 94 vcmpps $17, {sae}, %zmm2, %zmm1, %k1 98 vblendmps %zmm1, %zmm5, %zmm11{%k1} 99 vblendmps %zmm2, %zmm1, %zmm5{%k1} 100 vxorps %zmm4, %zmm4, %zmm4{%k1} 283 vcmpps $3, {sae}, %zmm8, %zmm8, %k1 313 vpandnd %zmm8, %zmm8, %zmm12{%k1}
|
D | svml_d_exp8_core_avx512.S | 87 vpcmpgtd __iDomainRange(%rax), %zmm2, %k1{%k2} 91 vpbroadcastd %ecx, %zmm3{%k1}{z} 267 kxnorw %k1, %k1, %k1 310 vgatherdpd (%rax,%ymm14,8), %zmm15{%k1}
|
D | svml_d_atan28_core_avx512.S | 101 vcmppd $17, {sae}, %zmm2, %zmm11, %k1 107 vblendmpd %zmm11, %zmm12, %zmm13{%k1} 108 vblendmpd %zmm2, %zmm11, %zmm15{%k1} 115 vxorpd %zmm3, %zmm3, %zmm3{%k1} 321 vcmppd $3, {sae}, %zmm7, %zmm7, %k1 334 vpandnq %zmm7, %zmm7, %zmm12{%k1}
|
D | svml_s_asinhf16_core_avx512.S | 142 vcmpps $21, {sae}, %zmm7, %zmm12, %k1 165 vmulps {rn-sae}, %zmm6, %zmm12, %zmm3{%k1} 186 vxorps %zmm7, %zmm7, %zmm7{%k1} 202 vsubps {rn-sae}, %zmm3, %zmm1, %zmm1{%k1}
|
D | svml_s_logf16_core_avx512.S | 63 vpcmpd $1, _iLoRange(%rax), %zmm1, %k1 68 vpbroadcastd %ecx, %zmm5{%k1}{z} 243 vpcmpd $5, _iLoRange(%rax), %zmm1, %k1 266 vpandnd %zmm1, %zmm1, %zmm6{%k1}
|
D | svml_s_sinf16_core_avx512.S | 84 vcmpps $22, __sRangeReductionVal(%rax), %zmm12, %k1 85 vpbroadcastd %edx, %zmm13{%k1}{z} 297 vcmpps $18, __sRangeReductionVal(%rax), %zmm13, %k1 337 vpandnd %zmm13, %zmm13, %zmm14{%k1}
|
D | svml_s_cosf16_core_avx512.S | 87 vcmpps $22, __sRangeReductionVal(%rdx), %zmm1, %k1 88 vpbroadcastd %eax, %zmm12{%k1}{z} 291 vcmpps $18, __sRangeReductionVal(%rax), %zmm1, %k1 325 vpandnd %zmm1, %zmm1, %zmm12{%k1}
|
D | svml_s_expf16_core_avx512.S | 73 vpcmpgtd __iDomainRange(%rax), %zmm1, %k1 78 vpbroadcastd %ecx, %zmm2{%k1}{z} 300 vpcmpd $2, __iDomainRange(%rax), %zmm2, %k1 301 vpandnd %zmm2, %zmm2, %zmm3{%k1}
|
D | svml_d_asinh8_core_avx512.S | 135 vcmppd $21, {sae}, %zmm15, %zmm1, %k1 183 vmulpd {rn-sae}, %zmm8, %zmm1, %zmm10{%k1} 201 vxorpd %zmm14, %zmm14, %zmm14{%k1} 217 vsubpd {rn-sae}, %zmm10, %zmm8, %zmm8{%k1}
|
/glibc-2.36/crypt/ |
D | crypt_util.c | 668 ufc_long v1, v2, *k1; in _ufc_mk_keytab_r() local 679 v1 = v2 = 0; k1 = &do_pc1[0][0][0]; in _ufc_mk_keytab_r() 681 v1 |= k1[*key & 0x7f]; k1 += 128; in _ufc_mk_keytab_r() 682 v2 |= k1[*key++ & 0x7f]; k1 += 128; in _ufc_mk_keytab_r() 686 k1 = &do_pc2[0][0]; in _ufc_mk_keytab_r() 689 v = k1[(v1 >> 21) & 0x7f]; k1 += 128; in _ufc_mk_keytab_r() 690 v |= k1[(v1 >> 14) & 0x7f]; k1 += 128; in _ufc_mk_keytab_r() 691 v |= k1[(v1 >> 7) & 0x7f]; k1 += 128; in _ufc_mk_keytab_r() 692 v |= k1[(v1 ) & 0x7f]; k1 += 128; in _ufc_mk_keytab_r() 703 v |= k1[(v2 >> 21) & 0x7f]; k1 += 128; in _ufc_mk_keytab_r() [all …]
|