Lines Matching refs:zSig0

573      int8 roundingPrecision, flag zSign, int32 zExp, bits64 zSig0, bits64 zSig1  in roundAndPackFloatx80()  argument
594 zSig0 |= ( zSig1 != 0 ); in roundAndPackFloatx80()
609 roundBits = zSig0 & roundMask; in roundAndPackFloatx80()
612 || ( ( zExp == 0x7FFE ) && ( zSig0 + roundIncrement < zSig0 ) ) in roundAndPackFloatx80()
620 || ( zSig0 <= zSig0 + roundIncrement ); in roundAndPackFloatx80()
621 shift64RightJamming( zSig0, 1 - zExp, &zSig0 ); in roundAndPackFloatx80()
623 roundBits = zSig0 & roundMask; in roundAndPackFloatx80()
626 zSig0 += roundIncrement; in roundAndPackFloatx80()
627 if ( (sbits64) zSig0 < 0 ) zExp = 1; in roundAndPackFloatx80()
632 zSig0 &= ~ roundMask; in roundAndPackFloatx80()
633 return packFloatx80( zSign, zExp, zSig0 ); in roundAndPackFloatx80()
637 zSig0 += roundIncrement; in roundAndPackFloatx80()
638 if ( zSig0 < roundIncrement ) { in roundAndPackFloatx80()
640 zSig0 = LIT64( 0x8000000000000000 ); in roundAndPackFloatx80()
646 zSig0 &= ~ roundMask; in roundAndPackFloatx80()
647 if ( zSig0 == 0 ) zExp = 0; in roundAndPackFloatx80()
648 return packFloatx80( zSign, zExp, zSig0 ); in roundAndPackFloatx80()
667 && ( zSig0 == LIT64( 0xFFFFFFFFFFFFFFFF ) ) in roundAndPackFloatx80()
687 || ( zSig0 < LIT64( 0xFFFFFFFFFFFFFFFF ) ); in roundAndPackFloatx80()
688 shift64ExtraRightJamming( zSig0, zSig1, 1 - zExp, &zSig0, &zSig1 ); in roundAndPackFloatx80()
704 ++zSig0; in roundAndPackFloatx80()
705 zSig0 &= ~ ( ( zSig1 + zSig1 == 0 ) & roundNearestEven ); in roundAndPackFloatx80()
706 if ( (sbits64) zSig0 < 0 ) zExp = 1; in roundAndPackFloatx80()
708 return packFloatx80( zSign, zExp, zSig0 ); in roundAndPackFloatx80()
713 ++zSig0; in roundAndPackFloatx80()
714 if ( zSig0 == 0 ) { in roundAndPackFloatx80()
716 zSig0 = LIT64( 0x8000000000000000 ); in roundAndPackFloatx80()
719 zSig0 &= ~ ( ( zSig1 + zSig1 == 0 ) & roundNearestEven ); in roundAndPackFloatx80()
723 if ( zSig0 == 0 ) zExp = 0; in roundAndPackFloatx80()
726 return packFloatx80( zSign, zExp, zSig0 ); in roundAndPackFloatx80()
741 int8 roundingPrecision, flag zSign, int32 zExp, bits64 zSig0, bits64 zSig1 in normalizeRoundAndPackFloatx80() argument
746 if ( zSig0 == 0 ) { in normalizeRoundAndPackFloatx80()
747 zSig0 = zSig1; in normalizeRoundAndPackFloatx80()
751 shiftCount = countLeadingZeros64( zSig0 ); in normalizeRoundAndPackFloatx80()
752 shortShift128Left( zSig0, zSig1, shiftCount, &zSig0, &zSig1 ); in normalizeRoundAndPackFloatx80()
755 roundAndPackFloatx80( roundingPrecision, zSign, zExp, zSig0, zSig1 ); in normalizeRoundAndPackFloatx80()
2123 bits64 aSig, bSig, zSig0, zSig1; in float64_mul() local
2161 mul64To128( aSig, bSig, &zSig0, &zSig1 ); in float64_mul()
2162 zSig0 |= ( zSig1 != 0 ); in float64_mul()
2163 if ( 0 <= (sbits64) ( zSig0<<1 ) ) { in float64_mul()
2164 zSig0 <<= 1; in float64_mul()
2167 return roundAndPackFloat64( zSign, zExp, zSig0 ); in float64_mul()
2756 bits64 aSig, bSig, zSig0, zSig1; in addFloatx80Sigs() local
2790 zSig0 = aSig + bSig; in addFloatx80Sigs()
2792 normalizeFloatx80Subnormal( zSig0, &zExp, &zSig0 ); in addFloatx80Sigs()
2799 zSig0 = aSig + bSig; in addFloatx80Sigs()
2801 if ( (sbits64) zSig0 < 0 ) goto roundAndPack; in addFloatx80Sigs()
2803 shift64ExtraRightJamming( zSig0, zSig1, 1, &zSig0, &zSig1 ); in addFloatx80Sigs()
2804 zSig0 |= LIT64( 0x8000000000000000 ); in addFloatx80Sigs()
2809 floatx80_rounding_precision, zSign, zExp, zSig0, zSig1 ); in addFloatx80Sigs()
2825 bits64 aSig, bSig, zSig0, zSig1; in subFloatx80Sigs() local
2861 sub128( bSig, 0, aSig, zSig1, &zSig0, &zSig1 ); in subFloatx80Sigs()
2873 sub128( aSig, 0, bSig, zSig1, &zSig0, &zSig1 ); in subFloatx80Sigs()
2878 floatx80_rounding_precision, zSign, zExp, zSig0, zSig1 ); in subFloatx80Sigs()
2937 bits64 aSig, bSig, zSig0, zSig1; in floatx80_mul() local
2975 mul64To128( aSig, bSig, &zSig0, &zSig1 ); in floatx80_mul()
2976 if ( 0 < (sbits64) zSig0 ) { in floatx80_mul()
2977 shortShift128Left( zSig0, zSig1, 1, &zSig0, &zSig1 ); in floatx80_mul()
2982 floatx80_rounding_precision, zSign, zExp, zSig0, zSig1 ); in floatx80_mul()
2997 bits64 aSig, bSig, zSig0, zSig1; in floatx80_div() local
3044 zSig0 = estimateDiv128To64( aSig, rem1, bSig ); in floatx80_div()
3045 mul64To128( bSig, zSig0, &term0, &term1 ); in floatx80_div()
3048 --zSig0; in floatx80_div()
3063 floatx80_rounding_precision, zSign, zExp, zSig0, zSig1 ); in floatx80_div()
3176 bits64 aSig0, aSig1, zSig0, zSig1; in floatx80_sqrt() local
3202 zSig0 = estimateSqrt32( aExp, aSig0>>32 ); in floatx80_sqrt()
3203 zSig0 <<= 31; in floatx80_sqrt()
3206 zSig0 = estimateDiv128To64( aSig0, aSig1, zSig0 ) + zSig0 + 4; in floatx80_sqrt()
3207 if ( 0 <= (sbits64) zSig0 ) zSig0 = LIT64( 0xFFFFFFFFFFFFFFFF ); in floatx80_sqrt()
3209 mul64To128( zSig0, zSig0, &term0, &term1 ); in floatx80_sqrt()
3212 --zSig0; in floatx80_sqrt()
3213 shortShift128Left( 0, zSig0, 1, &term0, &term1 ); in floatx80_sqrt()
3218 zSig1 = estimateDiv128To64( shiftedRem0, shiftedRem1, zSig0 ); in floatx80_sqrt()
3221 mul64To128( zSig0, zSig1, &term1, &term2 ); in floatx80_sqrt()
3228 shortShift192Left( 0, zSig0, zSig1, 1, &term1, &term2, &term3 ); in floatx80_sqrt()
3237 floatx80_rounding_precision, 0, zExp, zSig0, zSig1 ); in floatx80_sqrt()