Lines Matching refs:aSig
155 normalizeFloat32Subnormal( bits32 aSig, int16 *zExpPtr, bits32 *zSigPtr ) in normalizeFloat32Subnormal() argument
159 shiftCount = countLeadingZeros32( aSig ) - 8; in normalizeFloat32Subnormal()
160 *zSigPtr = aSig<<shiftCount; in normalizeFloat32Subnormal()
336 normalizeFloat64Subnormal( bits64 aSig, int16 *zExpPtr, bits64 *zSigPtr ) in normalizeFloat64Subnormal() argument
340 shiftCount = countLeadingZeros64( aSig ) - 11; in normalizeFloat64Subnormal()
341 *zSigPtr = aSig<<shiftCount; in normalizeFloat64Subnormal()
512 normalizeFloatx80Subnormal( bits64 aSig, int32 *zExpPtr, bits64 *zSigPtr ) in normalizeFloatx80Subnormal() argument
516 shiftCount = countLeadingZeros64( aSig ); in normalizeFloatx80Subnormal()
517 *zSigPtr = aSig<<shiftCount; in normalizeFloatx80Subnormal()
839 bits32 aSig; in float32_to_int32() local
842 aSig = extractFloat32Frac( a ); in float32_to_int32()
845 if ( ( aExp == 0x7FF ) && aSig ) aSign = 0; in float32_to_int32()
846 if ( aExp ) aSig |= 0x00800000; in float32_to_int32()
848 zSig = aSig; in float32_to_int32()
870 bits32 aSig; in float32_to_int32_round_to_zero() local
873 aSig = extractFloat32Frac( a ); in float32_to_int32_round_to_zero()
880 if ( ! aSign || ( ( aExp == 0xFF ) && aSig ) ) return 0x7FFFFFFF; in float32_to_int32_round_to_zero()
884 if ( aExp | aSig ) float_raise( float_flag_inexact ); in float32_to_int32_round_to_zero()
887 aSig = ( aSig | 0x00800000 )<<8; in float32_to_int32_round_to_zero()
888 z = aSig>>( - shiftCount ); in float32_to_int32_round_to_zero()
889 if ( (bits32) ( aSig<<( shiftCount & 31 ) ) ) { in float32_to_int32_round_to_zero()
908 bits32 aSig; in float32_to_float64() local
910 aSig = extractFloat32Frac( a ); in float32_to_float64()
914 if ( aSig ) return commonNaNToFloat64( float32ToCommonNaN( a ) ); in float32_to_float64()
918 if ( aSig == 0 ) return packFloat64( aSign, 0, 0 ); in float32_to_float64()
919 normalizeFloat32Subnormal( aSig, &aExp, &aSig ); in float32_to_float64()
922 return packFloat64( aSign, aExp + 0x380, ( (bits64) aSig )<<29 ); in float32_to_float64()
940 bits32 aSig; in float32_to_floatx80() local
942 aSig = extractFloat32Frac( a ); in float32_to_floatx80()
946 if ( aSig ) return commonNaNToFloatx80( float32ToCommonNaN( a ) ); in float32_to_floatx80()
950 if ( aSig == 0 ) return packFloatx80( aSign, 0, 0 ); in float32_to_floatx80()
951 normalizeFloat32Subnormal( aSig, &aExp, &aSig ); in float32_to_floatx80()
953 aSig |= 0x00800000; in float32_to_floatx80()
954 return packFloatx80( aSign, aExp + 0x3F80, ( (bits64) aSig )<<40 ); in float32_to_floatx80()
1032 bits32 aSig, bSig, zSig; in addFloat32Sigs() local
1035 aSig = extractFloat32Frac( a ); in addFloat32Sigs()
1040 aSig <<= 6; in addFloat32Sigs()
1044 if ( aSig ) return propagateFloat32NaN( a, b ); in addFloat32Sigs()
1065 aSig |= 0x20000000; in addFloat32Sigs()
1067 shift32RightJamming( aSig, - expDiff, &aSig ); in addFloat32Sigs()
1072 if ( aSig | bSig ) return propagateFloat32NaN( a, b ); in addFloat32Sigs()
1075 if ( aExp == 0 ) return packFloat32( zSign, 0, ( aSig + bSig )>>6 ); in addFloat32Sigs()
1076 zSig = 0x40000000 + aSig + bSig; in addFloat32Sigs()
1080 aSig |= 0x20000000; in addFloat32Sigs()
1081 zSig = ( aSig + bSig )<<1; in addFloat32Sigs()
1084 zSig = aSig + bSig; in addFloat32Sigs()
1104 bits32 aSig, bSig, zSig; in subFloat32Sigs() local
1107 aSig = extractFloat32Frac( a ); in subFloat32Sigs()
1112 aSig <<= 7; in subFloat32Sigs()
1117 if ( aSig | bSig ) return propagateFloat32NaN( a, b ); in subFloat32Sigs()
1125 if ( bSig < aSig ) goto aBigger; in subFloat32Sigs()
1126 if ( aSig < bSig ) goto bBigger; in subFloat32Sigs()
1137 aSig |= 0x40000000; in subFloat32Sigs()
1139 shift32RightJamming( aSig, - expDiff, &aSig ); in subFloat32Sigs()
1142 zSig = bSig - aSig; in subFloat32Sigs()
1148 if ( aSig ) return propagateFloat32NaN( a, b ); in subFloat32Sigs()
1158 aSig |= 0x40000000; in subFloat32Sigs()
1160 zSig = aSig - bSig; in subFloat32Sigs()
1223 bits32 aSig, bSig; in float32_mul() local
1227 aSig = extractFloat32Frac( a ); in float32_mul()
1235 if ( aSig || ( ( bExp == 0xFF ) && bSig ) ) { in float32_mul()
1246 if ( ( aExp | aSig ) == 0 ) { in float32_mul()
1253 if ( aSig == 0 ) return packFloat32( zSign, 0, 0 ); in float32_mul()
1254 normalizeFloat32Subnormal( aSig, &aExp, &aSig ); in float32_mul()
1261 aSig = ( aSig | 0x00800000 )<<7; in float32_mul()
1263 shift64RightJamming( ( (bits64) aSig ) * bSig, 32, &zSig64 ); in float32_mul()
1284 bits32 aSig, bSig, zSig; in float32_div() local
1286 aSig = extractFloat32Frac( a ); in float32_div()
1294 if ( aSig ) return propagateFloat32NaN( a, b ); in float32_div()
1308 if ( ( aExp | aSig ) == 0 ) { in float32_div()
1318 if ( aSig == 0 ) return packFloat32( zSign, 0, 0 ); in float32_div()
1319 normalizeFloat32Subnormal( aSig, &aExp, &aSig ); in float32_div()
1322 aSig = ( aSig | 0x00800000 )<<7; in float32_div()
1324 if ( bSig <= ( aSig + aSig ) ) { in float32_div()
1325 aSig >>= 1; in float32_div()
1329 bits64 tmp = ( (bits64) aSig )<<32; in float32_div()
1334 zSig |= ( ( (bits64) bSig ) * zSig != ( (bits64) aSig )<<32 ); in float32_div()
1351 bits32 aSig, bSig; in float32_rem() local
1357 aSig = extractFloat32Frac( a ); in float32_rem()
1364 if ( aSig || ( ( bExp == 0xFF ) && bSig ) ) { in float32_rem()
1382 if ( aSig == 0 ) return a; in float32_rem()
1383 normalizeFloat32Subnormal( aSig, &aExp, &aSig ); in float32_rem()
1386 aSig |= 0x00800000; in float32_rem()
1389 aSig <<= 8; in float32_rem()
1393 aSig >>= 1; in float32_rem()
1395 q = ( bSig <= aSig ); in float32_rem()
1396 if ( q ) aSig -= bSig; in float32_rem()
1398 bits64 tmp = ( (bits64) aSig )<<32; in float32_rem()
1403 aSig = ( ( aSig>>1 )<<( expDiff - 1 ) ) - bSig * q; in float32_rem()
1406 aSig >>= 2; in float32_rem()
1411 if ( bSig <= aSig ) aSig -= bSig; in float32_rem()
1412 aSig64 = ( (bits64) aSig )<<40; in float32_rem()
1426 aSig = ( ( aSig64>>33 )<<( expDiff - 1 ) ) - bSig * q; in float32_rem()
1429 alternateASig = aSig; in float32_rem()
1431 aSig -= bSig; in float32_rem()
1432 } while ( 0 <= (sbits32) aSig ); in float32_rem()
1433 sigMean = aSig + alternateASig; in float32_rem()
1435 aSig = alternateASig; in float32_rem()
1437 zSign = ( (sbits32) aSig < 0 ); in float32_rem()
1438 if ( zSign ) aSig = - aSig; in float32_rem()
1439 return normalizeRoundAndPackFloat32( roundData, aSign ^ zSign, bExp, aSig ); in float32_rem()
1454 bits32 aSig, zSig; in float32_sqrt() local
1457 aSig = extractFloat32Frac( a ); in float32_sqrt()
1461 if ( aSig ) return propagateFloat32NaN( a, 0 ); in float32_sqrt()
1467 if ( ( aExp | aSig ) == 0 ) return a; in float32_sqrt()
1472 if ( aSig == 0 ) return 0; in float32_sqrt()
1473 normalizeFloat32Subnormal( aSig, &aExp, &aSig ); in float32_sqrt()
1476 aSig = ( aSig | 0x00800000 )<<8; in float32_sqrt()
1477 zSig = estimateSqrt32( aExp, aSig ) + 2; in float32_sqrt()
1483 aSig >>= aExp & 1; in float32_sqrt()
1485 rem = ( ( (bits64) aSig )<<32 ) - term; in float32_sqrt()
1656 bits64 aSig; in float64_to_int32() local
1658 aSig = extractFloat64Frac( a ); in float64_to_int32()
1661 if ( ( aExp == 0x7FF ) && aSig ) aSign = 0; in float64_to_int32()
1662 if ( aExp ) aSig |= LIT64( 0x0010000000000000 ); in float64_to_int32()
1664 if ( 0 < shiftCount ) shift64RightJamming( aSig, shiftCount, &aSig ); in float64_to_int32()
1665 return roundAndPackInt32( roundData, aSign, aSig ); in float64_to_int32()
1684 bits64 aSig, savedASig; in float64_to_int32_round_to_zero() local
1687 aSig = extractFloat64Frac( a ); in float64_to_int32_round_to_zero()
1692 if ( ( aExp == 0x7FF ) && aSig ) aSign = 0; in float64_to_int32_round_to_zero()
1696 if ( aExp || aSig ) float_raise( float_flag_inexact ); in float64_to_int32_round_to_zero()
1699 aSig |= LIT64( 0x0010000000000000 ); in float64_to_int32_round_to_zero()
1700 savedASig = aSig; in float64_to_int32_round_to_zero()
1701 aSig >>= shiftCount; in float64_to_int32_round_to_zero()
1702 z = aSig; in float64_to_int32_round_to_zero()
1709 if ( ( aSig<<shiftCount ) != savedASig ) { in float64_to_int32_round_to_zero()
1731 bits64 aSig; in float64_to_uint32() local
1733 aSig = extractFloat64Frac( a ); in float64_to_uint32()
1737 if ( aExp ) aSig |= LIT64( 0x0010000000000000 ); in float64_to_uint32()
1739 if ( 0 < shiftCount ) shift64RightJamming( aSig, shiftCount, &aSig ); in float64_to_uint32()
1740 return roundAndPackInt32( roundData, aSign, aSig ); in float64_to_uint32()
1757 bits64 aSig, savedASig; in float64_to_uint32_round_to_zero() local
1760 aSig = extractFloat64Frac( a ); in float64_to_uint32_round_to_zero()
1765 if ( ( aExp == 0x7FF ) && aSig ) aSign = 0; in float64_to_uint32_round_to_zero()
1769 if ( aExp || aSig ) float_raise( float_flag_inexact ); in float64_to_uint32_round_to_zero()
1772 aSig |= LIT64( 0x0010000000000000 ); in float64_to_uint32_round_to_zero()
1773 savedASig = aSig; in float64_to_uint32_round_to_zero()
1774 aSig >>= shiftCount; in float64_to_uint32_round_to_zero()
1775 z = aSig; in float64_to_uint32_round_to_zero()
1782 if ( ( aSig<<shiftCount ) != savedASig ) { in float64_to_uint32_round_to_zero()
1800 bits64 aSig; in float64_to_float32() local
1803 aSig = extractFloat64Frac( a ); in float64_to_float32()
1807 if ( aSig ) return commonNaNToFloat32( float64ToCommonNaN( a ) ); in float64_to_float32()
1810 shift64RightJamming( aSig, 22, &aSig ); in float64_to_float32()
1811 zSig = aSig; in float64_to_float32()
1834 bits64 aSig; in float64_to_floatx80() local
1836 aSig = extractFloat64Frac( a ); in float64_to_floatx80()
1840 if ( aSig ) return commonNaNToFloatx80( float64ToCommonNaN( a ) ); in float64_to_floatx80()
1844 if ( aSig == 0 ) return packFloatx80( aSign, 0, 0 ); in float64_to_floatx80()
1845 normalizeFloat64Subnormal( aSig, &aExp, &aSig ); in float64_to_floatx80()
1849 aSign, aExp + 0x3C00, ( aSig | LIT64( 0x0010000000000000 ) )<<11 ); in float64_to_floatx80()
1928 bits64 aSig, bSig, zSig; in addFloat64Sigs() local
1931 aSig = extractFloat64Frac( a ); in addFloat64Sigs()
1936 aSig <<= 9; in addFloat64Sigs()
1940 if ( aSig ) return propagateFloat64NaN( a, b ); in addFloat64Sigs()
1961 aSig |= LIT64( 0x2000000000000000 ); in addFloat64Sigs()
1963 shift64RightJamming( aSig, - expDiff, &aSig ); in addFloat64Sigs()
1968 if ( aSig | bSig ) return propagateFloat64NaN( a, b ); in addFloat64Sigs()
1971 if ( aExp == 0 ) return packFloat64( zSign, 0, ( aSig + bSig )>>9 ); in addFloat64Sigs()
1972 zSig = LIT64( 0x4000000000000000 ) + aSig + bSig; in addFloat64Sigs()
1976 aSig |= LIT64( 0x2000000000000000 ); in addFloat64Sigs()
1977 zSig = ( aSig + bSig )<<1; in addFloat64Sigs()
1980 zSig = aSig + bSig; in addFloat64Sigs()
2000 bits64 aSig, bSig, zSig; in subFloat64Sigs() local
2003 aSig = extractFloat64Frac( a ); in subFloat64Sigs()
2008 aSig <<= 10; in subFloat64Sigs()
2013 if ( aSig | bSig ) return propagateFloat64NaN( a, b ); in subFloat64Sigs()
2021 if ( bSig < aSig ) goto aBigger; in subFloat64Sigs()
2022 if ( aSig < bSig ) goto bBigger; in subFloat64Sigs()
2033 aSig |= LIT64( 0x4000000000000000 ); in subFloat64Sigs()
2035 shift64RightJamming( aSig, - expDiff, &aSig ); in subFloat64Sigs()
2038 zSig = bSig - aSig; in subFloat64Sigs()
2044 if ( aSig ) return propagateFloat64NaN( a, b ); in subFloat64Sigs()
2054 aSig |= LIT64( 0x4000000000000000 ); in subFloat64Sigs()
2056 zSig = aSig - bSig; in subFloat64Sigs()
2119 bits64 aSig, bSig, zSig0, zSig1; in float64_mul() local
2121 aSig = extractFloat64Frac( a ); in float64_mul()
2129 if ( aSig || ( ( bExp == 0x7FF ) && bSig ) ) { in float64_mul()
2140 if ( ( aExp | aSig ) == 0 ) { in float64_mul()
2147 if ( aSig == 0 ) return packFloat64( zSign, 0, 0 ); in float64_mul()
2148 normalizeFloat64Subnormal( aSig, &aExp, &aSig ); in float64_mul()
2155 aSig = ( aSig | LIT64( 0x0010000000000000 ) )<<10; in float64_mul()
2157 mul64To128( aSig, bSig, &zSig0, &zSig1 ); in float64_mul()
2178 bits64 aSig, bSig, zSig; in float64_div() local
2182 aSig = extractFloat64Frac( a ); in float64_div()
2190 if ( aSig ) return propagateFloat64NaN( a, b ); in float64_div()
2204 if ( ( aExp | aSig ) == 0 ) { in float64_div()
2214 if ( aSig == 0 ) return packFloat64( zSign, 0, 0 ); in float64_div()
2215 normalizeFloat64Subnormal( aSig, &aExp, &aSig ); in float64_div()
2218 aSig = ( aSig | LIT64( 0x0010000000000000 ) )<<10; in float64_div()
2220 if ( bSig <= ( aSig + aSig ) ) { in float64_div()
2221 aSig >>= 1; in float64_div()
2224 zSig = estimateDiv128To64( aSig, 0, bSig ); in float64_div()
2227 sub128( aSig, 0, term0, term1, &rem0, &rem1 ); in float64_div()
2249 bits64 aSig, bSig; in float64_rem() local
2253 aSig = extractFloat64Frac( a ); in float64_rem()
2260 if ( aSig || ( ( bExp == 0x7FF ) && bSig ) ) { in float64_rem()
2278 if ( aSig == 0 ) return a; in float64_rem()
2279 normalizeFloat64Subnormal( aSig, &aExp, &aSig ); in float64_rem()
2282 aSig = ( aSig | LIT64( 0x0010000000000000 ) )<<11; in float64_rem()
2286 aSig >>= 1; in float64_rem()
2288 q = ( bSig <= aSig ); in float64_rem()
2289 if ( q ) aSig -= bSig; in float64_rem()
2292 q = estimateDiv128To64( aSig, 0, bSig ); in float64_rem()
2294 aSig = - ( ( bSig>>2 ) * q ); in float64_rem()
2299 q = estimateDiv128To64( aSig, 0, bSig ); in float64_rem()
2303 aSig = ( ( aSig>>1 )<<( expDiff - 1 ) ) - bSig * q; in float64_rem()
2306 aSig >>= 2; in float64_rem()
2310 alternateASig = aSig; in float64_rem()
2312 aSig -= bSig; in float64_rem()
2313 } while ( 0 <= (sbits64) aSig ); in float64_rem()
2314 sigMean = aSig + alternateASig; in float64_rem()
2316 aSig = alternateASig; in float64_rem()
2318 zSign = ( (sbits64) aSig < 0 ); in float64_rem()
2319 if ( zSign ) aSig = - aSig; in float64_rem()
2320 return normalizeRoundAndPackFloat64( roundData, aSign ^ zSign, bExp, aSig ); in float64_rem()
2335 bits64 aSig, zSig; in float64_sqrt() local
2339 aSig = extractFloat64Frac( a ); in float64_sqrt()
2343 if ( aSig ) return propagateFloat64NaN( a, a ); in float64_sqrt()
2349 if ( ( aExp | aSig ) == 0 ) return a; in float64_sqrt()
2354 if ( aSig == 0 ) return 0; in float64_sqrt()
2355 normalizeFloat64Subnormal( aSig, &aExp, &aSig ); in float64_sqrt()
2358 aSig |= LIT64( 0x0010000000000000 ); in float64_sqrt()
2359 zSig = estimateSqrt32( aExp, aSig>>21 ); in float64_sqrt()
2361 aSig <<= 9 - ( aExp & 1 ); in float64_sqrt()
2362 zSig = estimateDiv128To64( aSig, 0, zSig ) + zSig + 2; in float64_sqrt()
2368 aSig <<= 2; in float64_sqrt()
2370 sub128( aSig, 0, term0, term1, &rem0, &rem1 ); in float64_sqrt()
2545 bits64 aSig; in floatx80_to_int32() local
2547 aSig = extractFloatx80Frac( a ); in floatx80_to_int32()
2550 if ( ( aExp == 0x7FFF ) && (bits64) ( aSig<<1 ) ) aSign = 0; in floatx80_to_int32()
2553 shift64RightJamming( aSig, shiftCount, &aSig ); in floatx80_to_int32()
2554 return roundAndPackInt32( roundData, aSign, aSig ); in floatx80_to_int32()
2573 bits64 aSig, savedASig; in floatx80_to_int32_round_to_zero() local
2576 aSig = extractFloatx80Frac( a ); in floatx80_to_int32_round_to_zero()
2581 if ( ( aExp == 0x7FFF ) && (bits64) ( aSig<<1 ) ) aSign = 0; in floatx80_to_int32_round_to_zero()
2585 if ( aExp || aSig ) float_raise( float_flag_inexact ); in floatx80_to_int32_round_to_zero()
2588 savedASig = aSig; in floatx80_to_int32_round_to_zero()
2589 aSig >>= shiftCount; in floatx80_to_int32_round_to_zero()
2590 z = aSig; in floatx80_to_int32_round_to_zero()
2597 if ( ( aSig<<shiftCount ) != savedASig ) { in floatx80_to_int32_round_to_zero()
2616 bits64 aSig; in floatx80_to_float32() local
2618 aSig = extractFloatx80Frac( a ); in floatx80_to_float32()
2622 if ( (bits64) ( aSig<<1 ) ) { in floatx80_to_float32()
2627 shift64RightJamming( aSig, 33, &aSig ); in floatx80_to_float32()
2628 if ( aExp || aSig ) aExp -= 0x3F81; in floatx80_to_float32()
2629 return roundAndPackFloat32( roundData, aSign, aExp, aSig ); in floatx80_to_float32()
2645 bits64 aSig, zSig; in floatx80_to_float64() local
2647 aSig = extractFloatx80Frac( a ); in floatx80_to_float64()
2651 if ( (bits64) ( aSig<<1 ) ) { in floatx80_to_float64()
2656 shift64RightJamming( aSig, 1, &zSig ); in floatx80_to_float64()
2657 if ( aExp || aSig ) aExp -= 0x3C01; in floatx80_to_float64()
2748 bits64 aSig, bSig, zSig0, zSig1; in addFloatx80Sigs() local
2751 aSig = extractFloatx80Frac( a ); in addFloatx80Sigs()
2758 if ( (bits64) ( aSig<<1 ) ) return propagateFloatx80NaN( a, b ); in addFloatx80Sigs()
2771 shift64ExtraRightJamming( aSig, 0, - expDiff, &aSig, &zSig1 ); in addFloatx80Sigs()
2776 if ( (bits64) ( ( aSig | bSig )<<1 ) ) { in addFloatx80Sigs()
2782 zSig0 = aSig + bSig; in addFloatx80Sigs()
2791 zSig0 = aSig + bSig; in addFloatx80Sigs()
2817 bits64 aSig, bSig, zSig0, zSig1; in subFloatx80Sigs() local
2821 aSig = extractFloatx80Frac( a ); in subFloatx80Sigs()
2829 if ( (bits64) ( ( aSig | bSig )<<1 ) ) { in subFloatx80Sigs()
2843 if ( bSig < aSig ) goto aBigger; in subFloatx80Sigs()
2844 if ( aSig < bSig ) goto bBigger; in subFloatx80Sigs()
2852 shift128RightJamming( aSig, 0, - expDiff, &aSig, &zSig1 ); in subFloatx80Sigs()
2854 sub128( bSig, 0, aSig, zSig1, &zSig0, &zSig1 ); in subFloatx80Sigs()
2860 if ( (bits64) ( aSig<<1 ) ) return propagateFloatx80NaN( a, b ); in subFloatx80Sigs()
2866 sub128( aSig, 0, bSig, zSig1, &zSig0, &zSig1 ); in subFloatx80Sigs()
2930 bits64 aSig, bSig, zSig0, zSig1; in floatx80_mul() local
2933 aSig = extractFloatx80Frac( a ); in floatx80_mul()
2941 if ( (bits64) ( aSig<<1 ) in floatx80_mul()
2950 if ( ( aExp | aSig ) == 0 ) { in floatx80_mul()
2961 if ( aSig == 0 ) return packFloatx80( zSign, 0, 0 ); in floatx80_mul()
2962 normalizeFloatx80Subnormal( aSig, &aExp, &aSig ); in floatx80_mul()
2969 mul64To128( aSig, bSig, &zSig0, &zSig1 ); in floatx80_mul()
2991 bits64 aSig, bSig, zSig0, zSig1; in floatx80_div() local
2995 aSig = extractFloatx80Frac( a ); in floatx80_div()
3003 if ( (bits64) ( aSig<<1 ) ) return propagateFloatx80NaN( a, b ); in floatx80_div()
3016 if ( ( aExp | aSig ) == 0 ) { in floatx80_div()
3030 if ( aSig == 0 ) return packFloatx80( zSign, 0, 0 ); in floatx80_div()
3031 normalizeFloatx80Subnormal( aSig, &aExp, &aSig ); in floatx80_div()
3035 if ( bSig <= aSig ) { in floatx80_div()
3036 shift128Right( aSig, 0, 1, &aSig, &rem1 ); in floatx80_div()
3039 zSig0 = estimateDiv128To64( aSig, rem1, bSig ); in floatx80_div()
3041 sub128( aSig, rem1, term0, term1, &rem0, &rem1 ); in floatx80_div()