aSig 155 arch/arm/nwfpe/softfloat.c normalizeFloat32Subnormal( bits32 aSig, int16 *zExpPtr, bits32 *zSigPtr ) aSig 159 arch/arm/nwfpe/softfloat.c shiftCount = countLeadingZeros32( aSig ) - 8; aSig 160 arch/arm/nwfpe/softfloat.c *zSigPtr = aSig<<shiftCount; aSig 336 arch/arm/nwfpe/softfloat.c normalizeFloat64Subnormal( bits64 aSig, int16 *zExpPtr, bits64 *zSigPtr ) aSig 340 arch/arm/nwfpe/softfloat.c shiftCount = countLeadingZeros64( aSig ) - 11; aSig 341 arch/arm/nwfpe/softfloat.c *zSigPtr = aSig<<shiftCount; aSig 512 arch/arm/nwfpe/softfloat.c normalizeFloatx80Subnormal( bits64 aSig, int32 *zExpPtr, bits64 *zSigPtr ) aSig 516 arch/arm/nwfpe/softfloat.c shiftCount = countLeadingZeros64( aSig ); aSig 517 arch/arm/nwfpe/softfloat.c *zSigPtr = aSig<<shiftCount; aSig 839 arch/arm/nwfpe/softfloat.c bits32 aSig; aSig 842 arch/arm/nwfpe/softfloat.c aSig = extractFloat32Frac( a ); aSig 845 arch/arm/nwfpe/softfloat.c if ( ( aExp == 0x7FF ) && aSig ) aSign = 0; aSig 846 arch/arm/nwfpe/softfloat.c if ( aExp ) aSig |= 0x00800000; aSig 848 arch/arm/nwfpe/softfloat.c zSig = aSig; aSig 870 arch/arm/nwfpe/softfloat.c bits32 aSig; aSig 873 arch/arm/nwfpe/softfloat.c aSig = extractFloat32Frac( a ); aSig 880 arch/arm/nwfpe/softfloat.c if ( ! aSign || ( ( aExp == 0xFF ) && aSig ) ) return 0x7FFFFFFF; aSig 884 arch/arm/nwfpe/softfloat.c if ( aExp | aSig ) float_raise( float_flag_inexact ); aSig 887 arch/arm/nwfpe/softfloat.c aSig = ( aSig | 0x00800000 )<<8; aSig 888 arch/arm/nwfpe/softfloat.c z = aSig>>( - shiftCount ); aSig 889 arch/arm/nwfpe/softfloat.c if ( (bits32) ( aSig<<( shiftCount & 31 ) ) ) { aSig 908 arch/arm/nwfpe/softfloat.c bits32 aSig; aSig 910 arch/arm/nwfpe/softfloat.c aSig = extractFloat32Frac( a ); aSig 914 arch/arm/nwfpe/softfloat.c if ( aSig ) return commonNaNToFloat64( float32ToCommonNaN( a ) ); aSig 918 arch/arm/nwfpe/softfloat.c if ( aSig == 0 ) return packFloat64( aSign, 0, 0 ); aSig 919 arch/arm/nwfpe/softfloat.c normalizeFloat32Subnormal( aSig, &aExp, &aSig ); aSig 922 arch/arm/nwfpe/softfloat.c return packFloat64( aSign, aExp + 0x380, ( (bits64) aSig )<<29 ); aSig 940 arch/arm/nwfpe/softfloat.c bits32 aSig; aSig 942 arch/arm/nwfpe/softfloat.c aSig = extractFloat32Frac( a ); aSig 946 arch/arm/nwfpe/softfloat.c if ( aSig ) return commonNaNToFloatx80( float32ToCommonNaN( a ) ); aSig 950 arch/arm/nwfpe/softfloat.c if ( aSig == 0 ) return packFloatx80( aSign, 0, 0 ); aSig 951 arch/arm/nwfpe/softfloat.c normalizeFloat32Subnormal( aSig, &aExp, &aSig ); aSig 953 arch/arm/nwfpe/softfloat.c aSig |= 0x00800000; aSig 954 arch/arm/nwfpe/softfloat.c return packFloatx80( aSign, aExp + 0x3F80, ( (bits64) aSig )<<40 ); aSig 1032 arch/arm/nwfpe/softfloat.c bits32 aSig, bSig, zSig; aSig 1035 arch/arm/nwfpe/softfloat.c aSig = extractFloat32Frac( a ); aSig 1040 arch/arm/nwfpe/softfloat.c aSig <<= 6; aSig 1044 arch/arm/nwfpe/softfloat.c if ( aSig ) return propagateFloat32NaN( a, b ); aSig 1065 arch/arm/nwfpe/softfloat.c aSig |= 0x20000000; aSig 1067 arch/arm/nwfpe/softfloat.c shift32RightJamming( aSig, - expDiff, &aSig ); aSig 1072 arch/arm/nwfpe/softfloat.c if ( aSig | bSig ) return propagateFloat32NaN( a, b ); aSig 1075 arch/arm/nwfpe/softfloat.c if ( aExp == 0 ) return packFloat32( zSign, 0, ( aSig + bSig )>>6 ); aSig 1076 arch/arm/nwfpe/softfloat.c zSig = 0x40000000 + aSig + bSig; aSig 1080 arch/arm/nwfpe/softfloat.c aSig |= 0x20000000; aSig 1081 arch/arm/nwfpe/softfloat.c zSig = ( aSig + bSig )<<1; aSig 1084 arch/arm/nwfpe/softfloat.c zSig = aSig + bSig; aSig 1104 arch/arm/nwfpe/softfloat.c bits32 aSig, bSig, zSig; aSig 1107 arch/arm/nwfpe/softfloat.c aSig = extractFloat32Frac( a ); aSig 1112 arch/arm/nwfpe/softfloat.c aSig <<= 7; aSig 1117 arch/arm/nwfpe/softfloat.c if ( aSig | bSig ) return propagateFloat32NaN( a, b ); aSig 1125 arch/arm/nwfpe/softfloat.c if ( bSig < aSig ) goto aBigger; aSig 1126 arch/arm/nwfpe/softfloat.c if ( aSig < bSig ) goto bBigger; aSig 1137 arch/arm/nwfpe/softfloat.c aSig |= 0x40000000; aSig 1139 arch/arm/nwfpe/softfloat.c shift32RightJamming( aSig, - expDiff, &aSig ); aSig 1142 arch/arm/nwfpe/softfloat.c zSig = bSig - aSig; aSig 1148 arch/arm/nwfpe/softfloat.c if ( aSig ) return propagateFloat32NaN( a, b ); aSig 1158 arch/arm/nwfpe/softfloat.c aSig |= 0x40000000; aSig 1160 arch/arm/nwfpe/softfloat.c zSig = aSig - bSig; aSig 1223 arch/arm/nwfpe/softfloat.c bits32 aSig, bSig; aSig 1227 arch/arm/nwfpe/softfloat.c aSig = extractFloat32Frac( a ); aSig 1235 arch/arm/nwfpe/softfloat.c if ( aSig || ( ( bExp == 0xFF ) && bSig ) ) { aSig 1246 arch/arm/nwfpe/softfloat.c if ( ( aExp | aSig ) == 0 ) { aSig 1253 arch/arm/nwfpe/softfloat.c if ( aSig == 0 ) return packFloat32( zSign, 0, 0 ); aSig 1254 arch/arm/nwfpe/softfloat.c normalizeFloat32Subnormal( aSig, &aExp, &aSig ); aSig 1261 arch/arm/nwfpe/softfloat.c aSig = ( aSig | 0x00800000 )<<7; aSig 1263 arch/arm/nwfpe/softfloat.c shift64RightJamming( ( (bits64) aSig ) * bSig, 32, &zSig64 ); aSig 1284 arch/arm/nwfpe/softfloat.c bits32 aSig, bSig, zSig; aSig 1286 arch/arm/nwfpe/softfloat.c aSig = extractFloat32Frac( a ); aSig 1294 arch/arm/nwfpe/softfloat.c if ( aSig ) return propagateFloat32NaN( a, b ); aSig 1308 arch/arm/nwfpe/softfloat.c if ( ( aExp | aSig ) == 0 ) { aSig 1318 arch/arm/nwfpe/softfloat.c if ( aSig == 0 ) return packFloat32( zSign, 0, 0 ); aSig 1319 arch/arm/nwfpe/softfloat.c normalizeFloat32Subnormal( aSig, &aExp, &aSig ); aSig 1322 arch/arm/nwfpe/softfloat.c aSig = ( aSig | 0x00800000 )<<7; aSig 1324 arch/arm/nwfpe/softfloat.c if ( bSig <= ( aSig + aSig ) ) { aSig 1325 arch/arm/nwfpe/softfloat.c aSig >>= 1; aSig 1329 arch/arm/nwfpe/softfloat.c bits64 tmp = ( (bits64) aSig )<<32; aSig 1334 arch/arm/nwfpe/softfloat.c zSig |= ( ( (bits64) bSig ) * zSig != ( (bits64) aSig )<<32 ); aSig 1351 arch/arm/nwfpe/softfloat.c bits32 aSig, bSig; aSig 1357 arch/arm/nwfpe/softfloat.c aSig = extractFloat32Frac( a ); aSig 1364 arch/arm/nwfpe/softfloat.c if ( aSig || ( ( bExp == 0xFF ) && bSig ) ) { aSig 1382 arch/arm/nwfpe/softfloat.c if ( aSig == 0 ) return a; aSig 1383 arch/arm/nwfpe/softfloat.c normalizeFloat32Subnormal( aSig, &aExp, &aSig ); aSig 1386 arch/arm/nwfpe/softfloat.c aSig |= 0x00800000; aSig 1389 arch/arm/nwfpe/softfloat.c aSig <<= 8; aSig 1393 arch/arm/nwfpe/softfloat.c aSig >>= 1; aSig 1395 arch/arm/nwfpe/softfloat.c q = ( bSig <= aSig ); aSig 1396 arch/arm/nwfpe/softfloat.c if ( q ) aSig -= bSig; aSig 1398 arch/arm/nwfpe/softfloat.c bits64 tmp = ( (bits64) aSig )<<32; aSig 1403 arch/arm/nwfpe/softfloat.c aSig = ( ( aSig>>1 )<<( expDiff - 1 ) ) - bSig * q; aSig 1406 arch/arm/nwfpe/softfloat.c aSig >>= 2; aSig 1411 arch/arm/nwfpe/softfloat.c if ( bSig <= aSig ) aSig -= bSig; aSig 1412 arch/arm/nwfpe/softfloat.c aSig64 = ( (bits64) aSig )<<40; aSig 1426 arch/arm/nwfpe/softfloat.c aSig = ( ( aSig64>>33 )<<( expDiff - 1 ) ) - bSig * q; aSig 1429 arch/arm/nwfpe/softfloat.c alternateASig = aSig; aSig 1431 arch/arm/nwfpe/softfloat.c aSig -= bSig; aSig 1432 arch/arm/nwfpe/softfloat.c } while ( 0 <= (sbits32) aSig ); aSig 1433 arch/arm/nwfpe/softfloat.c sigMean = aSig + alternateASig; aSig 1435 arch/arm/nwfpe/softfloat.c aSig = alternateASig; aSig 1437 arch/arm/nwfpe/softfloat.c zSign = ( (sbits32) aSig < 0 ); aSig 1438 arch/arm/nwfpe/softfloat.c if ( zSign ) aSig = - aSig; aSig 1439 arch/arm/nwfpe/softfloat.c return normalizeRoundAndPackFloat32( roundData, aSign ^ zSign, bExp, aSig ); aSig 1454 arch/arm/nwfpe/softfloat.c bits32 aSig, zSig; aSig 1457 arch/arm/nwfpe/softfloat.c aSig = extractFloat32Frac( a ); aSig 1461 arch/arm/nwfpe/softfloat.c if ( aSig ) return propagateFloat32NaN( a, 0 ); aSig 1467 arch/arm/nwfpe/softfloat.c if ( ( aExp | aSig ) == 0 ) return a; aSig 1472 arch/arm/nwfpe/softfloat.c if ( aSig == 0 ) return 0; aSig 1473 arch/arm/nwfpe/softfloat.c normalizeFloat32Subnormal( aSig, &aExp, &aSig ); aSig 1476 arch/arm/nwfpe/softfloat.c aSig = ( aSig | 0x00800000 )<<8; aSig 1477 arch/arm/nwfpe/softfloat.c zSig = estimateSqrt32( aExp, aSig ) + 2; aSig 1483 arch/arm/nwfpe/softfloat.c aSig >>= aExp & 1; aSig 1485 arch/arm/nwfpe/softfloat.c rem = ( ( (bits64) aSig )<<32 ) - term; aSig 1656 arch/arm/nwfpe/softfloat.c bits64 aSig; aSig 1658 arch/arm/nwfpe/softfloat.c aSig = extractFloat64Frac( a ); aSig 1661 arch/arm/nwfpe/softfloat.c if ( ( aExp == 0x7FF ) && aSig ) aSign = 0; aSig 1662 arch/arm/nwfpe/softfloat.c if ( aExp ) aSig |= LIT64( 0x0010000000000000 ); aSig 1664 arch/arm/nwfpe/softfloat.c if ( 0 < shiftCount ) shift64RightJamming( aSig, shiftCount, &aSig ); aSig 1665 arch/arm/nwfpe/softfloat.c return roundAndPackInt32( roundData, aSign, aSig ); aSig 1684 arch/arm/nwfpe/softfloat.c bits64 aSig, savedASig; aSig 1687 arch/arm/nwfpe/softfloat.c aSig = extractFloat64Frac( a ); aSig 1692 arch/arm/nwfpe/softfloat.c if ( ( aExp == 0x7FF ) && aSig ) aSign = 0; aSig 1696 arch/arm/nwfpe/softfloat.c if ( aExp || aSig ) float_raise( float_flag_inexact ); aSig 1699 arch/arm/nwfpe/softfloat.c aSig |= LIT64( 0x0010000000000000 ); aSig 1700 arch/arm/nwfpe/softfloat.c savedASig = aSig; aSig 1701 arch/arm/nwfpe/softfloat.c aSig >>= shiftCount; aSig 1702 arch/arm/nwfpe/softfloat.c z = aSig; aSig 1709 arch/arm/nwfpe/softfloat.c if ( ( aSig<<shiftCount ) != savedASig ) { aSig 1731 arch/arm/nwfpe/softfloat.c bits64 aSig; aSig 1733 arch/arm/nwfpe/softfloat.c aSig = extractFloat64Frac( a ); aSig 1737 arch/arm/nwfpe/softfloat.c if ( aExp ) aSig |= LIT64( 0x0010000000000000 ); aSig 1739 arch/arm/nwfpe/softfloat.c if ( 0 < shiftCount ) shift64RightJamming( aSig, shiftCount, &aSig ); aSig 1740 arch/arm/nwfpe/softfloat.c return roundAndPackInt32( roundData, aSign, aSig ); aSig 1757 arch/arm/nwfpe/softfloat.c bits64 aSig, savedASig; aSig 1760 arch/arm/nwfpe/softfloat.c aSig = extractFloat64Frac( a ); aSig 1765 arch/arm/nwfpe/softfloat.c if ( ( aExp == 0x7FF ) && aSig ) aSign = 0; aSig 1769 arch/arm/nwfpe/softfloat.c if ( aExp || aSig ) float_raise( float_flag_inexact ); aSig 1772 arch/arm/nwfpe/softfloat.c aSig |= LIT64( 0x0010000000000000 ); aSig 1773 arch/arm/nwfpe/softfloat.c savedASig = aSig; aSig 1774 arch/arm/nwfpe/softfloat.c aSig >>= shiftCount; aSig 1775 arch/arm/nwfpe/softfloat.c z = aSig; aSig 1782 arch/arm/nwfpe/softfloat.c if ( ( aSig<<shiftCount ) != savedASig ) { aSig 1800 arch/arm/nwfpe/softfloat.c bits64 aSig; aSig 1803 arch/arm/nwfpe/softfloat.c aSig = extractFloat64Frac( a ); aSig 1807 arch/arm/nwfpe/softfloat.c if ( aSig ) return commonNaNToFloat32( float64ToCommonNaN( a ) ); aSig 1810 arch/arm/nwfpe/softfloat.c shift64RightJamming( aSig, 22, &aSig ); aSig 1811 arch/arm/nwfpe/softfloat.c zSig = aSig; aSig 1834 arch/arm/nwfpe/softfloat.c bits64 aSig; aSig 1836 arch/arm/nwfpe/softfloat.c aSig = extractFloat64Frac( a ); aSig 1840 arch/arm/nwfpe/softfloat.c if ( aSig ) return commonNaNToFloatx80( float64ToCommonNaN( a ) ); aSig 1844 arch/arm/nwfpe/softfloat.c if ( aSig == 0 ) return packFloatx80( aSign, 0, 0 ); aSig 1845 arch/arm/nwfpe/softfloat.c normalizeFloat64Subnormal( aSig, &aExp, &aSig ); aSig 1849 arch/arm/nwfpe/softfloat.c aSign, aExp + 0x3C00, ( aSig | LIT64( 0x0010000000000000 ) )<<11 ); aSig 1928 arch/arm/nwfpe/softfloat.c bits64 aSig, bSig, zSig; aSig 1931 arch/arm/nwfpe/softfloat.c aSig = extractFloat64Frac( a ); aSig 1936 arch/arm/nwfpe/softfloat.c aSig <<= 9; aSig 1940 arch/arm/nwfpe/softfloat.c if ( aSig ) return propagateFloat64NaN( a, b ); aSig 1961 arch/arm/nwfpe/softfloat.c aSig |= LIT64( 0x2000000000000000 ); aSig 1963 arch/arm/nwfpe/softfloat.c shift64RightJamming( aSig, - expDiff, &aSig ); aSig 1968 arch/arm/nwfpe/softfloat.c if ( aSig | bSig ) return propagateFloat64NaN( a, b ); aSig 1971 arch/arm/nwfpe/softfloat.c if ( aExp == 0 ) return packFloat64( zSign, 0, ( aSig + bSig )>>9 ); aSig 1972 arch/arm/nwfpe/softfloat.c zSig = LIT64( 0x4000000000000000 ) + aSig + bSig; aSig 1976 arch/arm/nwfpe/softfloat.c aSig |= LIT64( 0x2000000000000000 ); aSig 1977 arch/arm/nwfpe/softfloat.c zSig = ( aSig + bSig )<<1; aSig 1980 arch/arm/nwfpe/softfloat.c zSig = aSig + bSig; aSig 2000 arch/arm/nwfpe/softfloat.c bits64 aSig, bSig, zSig; aSig 2003 arch/arm/nwfpe/softfloat.c aSig = extractFloat64Frac( a ); aSig 2008 arch/arm/nwfpe/softfloat.c aSig <<= 10; aSig 2013 arch/arm/nwfpe/softfloat.c if ( aSig | bSig ) return propagateFloat64NaN( a, b ); aSig 2021 arch/arm/nwfpe/softfloat.c if ( bSig < aSig ) goto aBigger; aSig 2022 arch/arm/nwfpe/softfloat.c if ( aSig < bSig ) goto bBigger; aSig 2033 arch/arm/nwfpe/softfloat.c aSig |= LIT64( 0x4000000000000000 ); aSig 2035 arch/arm/nwfpe/softfloat.c shift64RightJamming( aSig, - expDiff, &aSig ); aSig 2038 arch/arm/nwfpe/softfloat.c zSig = bSig - aSig; aSig 2044 arch/arm/nwfpe/softfloat.c if ( aSig ) return propagateFloat64NaN( a, b ); aSig 2054 arch/arm/nwfpe/softfloat.c aSig |= LIT64( 0x4000000000000000 ); aSig 2056 arch/arm/nwfpe/softfloat.c zSig = aSig - bSig; aSig 2119 arch/arm/nwfpe/softfloat.c bits64 aSig, bSig, zSig0, zSig1; aSig 2121 arch/arm/nwfpe/softfloat.c aSig = extractFloat64Frac( a ); aSig 2129 arch/arm/nwfpe/softfloat.c if ( aSig || ( ( bExp == 0x7FF ) && bSig ) ) { aSig 2140 arch/arm/nwfpe/softfloat.c if ( ( aExp | aSig ) == 0 ) { aSig 2147 arch/arm/nwfpe/softfloat.c if ( aSig == 0 ) return packFloat64( zSign, 0, 0 ); aSig 2148 arch/arm/nwfpe/softfloat.c normalizeFloat64Subnormal( aSig, &aExp, &aSig ); aSig 2155 arch/arm/nwfpe/softfloat.c aSig = ( aSig | LIT64( 0x0010000000000000 ) )<<10; aSig 2157 arch/arm/nwfpe/softfloat.c mul64To128( aSig, bSig, &zSig0, &zSig1 ); aSig 2178 arch/arm/nwfpe/softfloat.c bits64 aSig, bSig, zSig; aSig 2182 arch/arm/nwfpe/softfloat.c aSig = extractFloat64Frac( a ); aSig 2190 arch/arm/nwfpe/softfloat.c if ( aSig ) return propagateFloat64NaN( a, b ); aSig 2204 arch/arm/nwfpe/softfloat.c if ( ( aExp | aSig ) == 0 ) { aSig 2214 arch/arm/nwfpe/softfloat.c if ( aSig == 0 ) return packFloat64( zSign, 0, 0 ); aSig 2215 arch/arm/nwfpe/softfloat.c normalizeFloat64Subnormal( aSig, &aExp, &aSig ); aSig 2218 arch/arm/nwfpe/softfloat.c aSig = ( aSig | LIT64( 0x0010000000000000 ) )<<10; aSig 2220 arch/arm/nwfpe/softfloat.c if ( bSig <= ( aSig + aSig ) ) { aSig 2221 arch/arm/nwfpe/softfloat.c aSig >>= 1; aSig 2224 arch/arm/nwfpe/softfloat.c zSig = estimateDiv128To64( aSig, 0, bSig ); aSig 2227 arch/arm/nwfpe/softfloat.c sub128( aSig, 0, term0, term1, &rem0, &rem1 ); aSig 2249 arch/arm/nwfpe/softfloat.c bits64 aSig, bSig; aSig 2253 arch/arm/nwfpe/softfloat.c aSig = extractFloat64Frac( a ); aSig 2260 arch/arm/nwfpe/softfloat.c if ( aSig || ( ( bExp == 0x7FF ) && bSig ) ) { aSig 2278 arch/arm/nwfpe/softfloat.c if ( aSig == 0 ) return a; aSig 2279 arch/arm/nwfpe/softfloat.c normalizeFloat64Subnormal( aSig, &aExp, &aSig ); aSig 2282 arch/arm/nwfpe/softfloat.c aSig = ( aSig | LIT64( 0x0010000000000000 ) )<<11; aSig 2286 arch/arm/nwfpe/softfloat.c aSig >>= 1; aSig 2288 arch/arm/nwfpe/softfloat.c q = ( bSig <= aSig ); aSig 2289 arch/arm/nwfpe/softfloat.c if ( q ) aSig -= bSig; aSig 2292 arch/arm/nwfpe/softfloat.c q = estimateDiv128To64( aSig, 0, bSig ); aSig 2294 arch/arm/nwfpe/softfloat.c aSig = - ( ( bSig>>2 ) * q ); aSig 2299 arch/arm/nwfpe/softfloat.c q = estimateDiv128To64( aSig, 0, bSig ); aSig 2303 arch/arm/nwfpe/softfloat.c aSig = ( ( aSig>>1 )<<( expDiff - 1 ) ) - bSig * q; aSig 2306 arch/arm/nwfpe/softfloat.c aSig >>= 2; aSig 2310 arch/arm/nwfpe/softfloat.c alternateASig = aSig; aSig 2312 arch/arm/nwfpe/softfloat.c aSig -= bSig; aSig 2313 arch/arm/nwfpe/softfloat.c } while ( 0 <= (sbits64) aSig ); aSig 2314 arch/arm/nwfpe/softfloat.c sigMean = aSig + alternateASig; aSig 2316 arch/arm/nwfpe/softfloat.c aSig = alternateASig; aSig 2318 arch/arm/nwfpe/softfloat.c zSign = ( (sbits64) aSig < 0 ); aSig 2319 arch/arm/nwfpe/softfloat.c if ( zSign ) aSig = - aSig; aSig 2320 arch/arm/nwfpe/softfloat.c return normalizeRoundAndPackFloat64( roundData, aSign ^ zSign, bExp, aSig ); aSig 2335 arch/arm/nwfpe/softfloat.c bits64 aSig, zSig; aSig 2339 arch/arm/nwfpe/softfloat.c aSig = extractFloat64Frac( a ); aSig 2343 arch/arm/nwfpe/softfloat.c if ( aSig ) return propagateFloat64NaN( a, a ); aSig 2349 arch/arm/nwfpe/softfloat.c if ( ( aExp | aSig ) == 0 ) return a; aSig 2354 arch/arm/nwfpe/softfloat.c if ( aSig == 0 ) return 0; aSig 2355 arch/arm/nwfpe/softfloat.c normalizeFloat64Subnormal( aSig, &aExp, &aSig ); aSig 2358 arch/arm/nwfpe/softfloat.c aSig |= LIT64( 0x0010000000000000 ); aSig 2359 arch/arm/nwfpe/softfloat.c zSig = estimateSqrt32( aExp, aSig>>21 ); aSig 2361 arch/arm/nwfpe/softfloat.c aSig <<= 9 - ( aExp & 1 ); aSig 2362 arch/arm/nwfpe/softfloat.c zSig = estimateDiv128To64( aSig, 0, zSig ) + zSig + 2; aSig 2368 arch/arm/nwfpe/softfloat.c aSig <<= 2; aSig 2370 arch/arm/nwfpe/softfloat.c sub128( aSig, 0, term0, term1, &rem0, &rem1 ); aSig 2545 arch/arm/nwfpe/softfloat.c bits64 aSig; aSig 2547 arch/arm/nwfpe/softfloat.c aSig = extractFloatx80Frac( a ); aSig 2550 arch/arm/nwfpe/softfloat.c if ( ( aExp == 0x7FFF ) && (bits64) ( aSig<<1 ) ) aSign = 0; aSig 2553 arch/arm/nwfpe/softfloat.c shift64RightJamming( aSig, shiftCount, &aSig ); aSig 2554 arch/arm/nwfpe/softfloat.c return roundAndPackInt32( roundData, aSign, aSig ); aSig 2573 arch/arm/nwfpe/softfloat.c bits64 aSig, savedASig; aSig 2576 arch/arm/nwfpe/softfloat.c aSig = extractFloatx80Frac( a ); aSig 2581 arch/arm/nwfpe/softfloat.c if ( ( aExp == 0x7FFF ) && (bits64) ( aSig<<1 ) ) aSign = 0; aSig 2585 arch/arm/nwfpe/softfloat.c if ( aExp || aSig ) float_raise( float_flag_inexact ); aSig 2588 arch/arm/nwfpe/softfloat.c savedASig = aSig; aSig 2589 arch/arm/nwfpe/softfloat.c aSig >>= shiftCount; aSig 2590 arch/arm/nwfpe/softfloat.c z = aSig; aSig 2597 arch/arm/nwfpe/softfloat.c if ( ( aSig<<shiftCount ) != savedASig ) { aSig 2616 arch/arm/nwfpe/softfloat.c bits64 aSig; aSig 2618 arch/arm/nwfpe/softfloat.c aSig = extractFloatx80Frac( a ); aSig 2622 arch/arm/nwfpe/softfloat.c if ( (bits64) ( aSig<<1 ) ) { aSig 2627 arch/arm/nwfpe/softfloat.c shift64RightJamming( aSig, 33, &aSig ); aSig 2628 arch/arm/nwfpe/softfloat.c if ( aExp || aSig ) aExp -= 0x3F81; aSig 2629 arch/arm/nwfpe/softfloat.c return roundAndPackFloat32( roundData, aSign, aExp, aSig ); aSig 2645 arch/arm/nwfpe/softfloat.c bits64 aSig, zSig; aSig 2647 arch/arm/nwfpe/softfloat.c aSig = extractFloatx80Frac( a ); aSig 2651 arch/arm/nwfpe/softfloat.c if ( (bits64) ( aSig<<1 ) ) { aSig 2656 arch/arm/nwfpe/softfloat.c shift64RightJamming( aSig, 1, &zSig ); aSig 2657 arch/arm/nwfpe/softfloat.c if ( aExp || aSig ) aExp -= 0x3C01; aSig 2748 arch/arm/nwfpe/softfloat.c bits64 aSig, bSig, zSig0, zSig1; aSig 2751 arch/arm/nwfpe/softfloat.c aSig = extractFloatx80Frac( a ); aSig 2758 arch/arm/nwfpe/softfloat.c if ( (bits64) ( aSig<<1 ) ) return propagateFloatx80NaN( a, b ); aSig 2771 arch/arm/nwfpe/softfloat.c shift64ExtraRightJamming( aSig, 0, - expDiff, &aSig, &zSig1 ); aSig 2776 arch/arm/nwfpe/softfloat.c if ( (bits64) ( ( aSig | bSig )<<1 ) ) { aSig 2782 arch/arm/nwfpe/softfloat.c zSig0 = aSig + bSig; aSig 2791 arch/arm/nwfpe/softfloat.c zSig0 = aSig + bSig; aSig 2817 arch/arm/nwfpe/softfloat.c bits64 aSig, bSig, zSig0, zSig1; aSig 2821 arch/arm/nwfpe/softfloat.c aSig = extractFloatx80Frac( a ); aSig 2829 arch/arm/nwfpe/softfloat.c if ( (bits64) ( ( aSig | bSig )<<1 ) ) { aSig 2843 arch/arm/nwfpe/softfloat.c if ( bSig < aSig ) goto aBigger; aSig 2844 arch/arm/nwfpe/softfloat.c if ( aSig < bSig ) goto bBigger; aSig 2852 arch/arm/nwfpe/softfloat.c shift128RightJamming( aSig, 0, - expDiff, &aSig, &zSig1 ); aSig 2854 arch/arm/nwfpe/softfloat.c sub128( bSig, 0, aSig, zSig1, &zSig0, &zSig1 ); aSig 2860 arch/arm/nwfpe/softfloat.c if ( (bits64) ( aSig<<1 ) ) return propagateFloatx80NaN( a, b ); aSig 2866 arch/arm/nwfpe/softfloat.c sub128( aSig, 0, bSig, zSig1, &zSig0, &zSig1 ); aSig 2930 arch/arm/nwfpe/softfloat.c bits64 aSig, bSig, zSig0, zSig1; aSig 2933 arch/arm/nwfpe/softfloat.c aSig = extractFloatx80Frac( a ); aSig 2941 arch/arm/nwfpe/softfloat.c if ( (bits64) ( aSig<<1 ) aSig 2950 arch/arm/nwfpe/softfloat.c if ( ( aExp | aSig ) == 0 ) { aSig 2961 arch/arm/nwfpe/softfloat.c if ( aSig == 0 ) return packFloatx80( zSign, 0, 0 ); aSig 2962 arch/arm/nwfpe/softfloat.c normalizeFloatx80Subnormal( aSig, &aExp, &aSig ); aSig 2969 arch/arm/nwfpe/softfloat.c mul64To128( aSig, bSig, &zSig0, &zSig1 ); aSig 2991 arch/arm/nwfpe/softfloat.c bits64 aSig, bSig, zSig0, zSig1; aSig 2995 arch/arm/nwfpe/softfloat.c aSig = extractFloatx80Frac( a ); aSig 3003 arch/arm/nwfpe/softfloat.c if ( (bits64) ( aSig<<1 ) ) return propagateFloatx80NaN( a, b ); aSig 3016 arch/arm/nwfpe/softfloat.c if ( ( aExp | aSig ) == 0 ) { aSig 3030 arch/arm/nwfpe/softfloat.c if ( aSig == 0 ) return packFloatx80( zSign, 0, 0 ); aSig 3031 arch/arm/nwfpe/softfloat.c normalizeFloatx80Subnormal( aSig, &aExp, &aSig ); aSig 3035 arch/arm/nwfpe/softfloat.c if ( bSig <= aSig ) { aSig 3036 arch/arm/nwfpe/softfloat.c shift128Right( aSig, 0, 1, &aSig, &rem1 ); aSig 3039 arch/arm/nwfpe/softfloat.c zSig0 = estimateDiv128To64( aSig, rem1, bSig ); aSig 3041 arch/arm/nwfpe/softfloat.c sub128( aSig, rem1, term0, term1, &rem0, &rem1 ); aSig 108 arch/sh/kernel/cpu/sh4/softfloat.c static void normalizeFloat64Subnormal(bits64 aSig, int16 * zExpPtr, aSig 111 arch/sh/kernel/cpu/sh4/softfloat.c static void normalizeFloat32Subnormal(bits32 aSig, int16 * zExpPtr, aSig 227 arch/sh/kernel/cpu/sh4/softfloat.c bits64 aSig, bSig, zSig; aSig 230 arch/sh/kernel/cpu/sh4/softfloat.c aSig = extractFloat64Frac(a); aSig 235 arch/sh/kernel/cpu/sh4/softfloat.c aSig <<= 10; aSig 245 arch/sh/kernel/cpu/sh4/softfloat.c if (bSig < aSig) aSig 247 arch/sh/kernel/cpu/sh4/softfloat.c if (aSig < bSig) aSig 257 arch/sh/kernel/cpu/sh4/softfloat.c aSig |= LIT64(0x4000000000000000); aSig 259 arch/sh/kernel/cpu/sh4/softfloat.c shift64RightJamming(aSig, -expDiff, &aSig); aSig 262 arch/sh/kernel/cpu/sh4/softfloat.c zSig = bSig - aSig; aSig 276 arch/sh/kernel/cpu/sh4/softfloat.c aSig |= LIT64(0x4000000000000000); aSig 278 arch/sh/kernel/cpu/sh4/softfloat.c zSig = aSig - bSig; aSig 288 arch/sh/kernel/cpu/sh4/softfloat.c bits64 aSig, bSig, zSig; aSig 291 arch/sh/kernel/cpu/sh4/softfloat.c aSig = extractFloat64Frac(a); aSig 296 arch/sh/kernel/cpu/sh4/softfloat.c aSig <<= 9; aSig 316 arch/sh/kernel/cpu/sh4/softfloat.c aSig |= LIT64(0x2000000000000000); aSig 318 arch/sh/kernel/cpu/sh4/softfloat.c shift64RightJamming(aSig, -expDiff, &aSig); aSig 325 arch/sh/kernel/cpu/sh4/softfloat.c return packFloat64(zSign, 0, (aSig + bSig) >> 9); aSig 326 arch/sh/kernel/cpu/sh4/softfloat.c zSig = LIT64(0x4000000000000000) + aSig + bSig; aSig 330 arch/sh/kernel/cpu/sh4/softfloat.c aSig |= LIT64(0x2000000000000000); aSig 331 arch/sh/kernel/cpu/sh4/softfloat.c zSig = (aSig + bSig) << 1; aSig 334 arch/sh/kernel/cpu/sh4/softfloat.c zSig = aSig + bSig; aSig 457 arch/sh/kernel/cpu/sh4/softfloat.c bits32 aSig, bSig, zSig; aSig 460 arch/sh/kernel/cpu/sh4/softfloat.c aSig = extractFloat32Frac(a); aSig 465 arch/sh/kernel/cpu/sh4/softfloat.c aSig <<= 7; aSig 475 arch/sh/kernel/cpu/sh4/softfloat.c if (bSig < aSig) aSig 477 arch/sh/kernel/cpu/sh4/softfloat.c if (aSig < bSig) aSig 487 arch/sh/kernel/cpu/sh4/softfloat.c aSig |= 0x40000000; aSig 489 arch/sh/kernel/cpu/sh4/softfloat.c shift32RightJamming(aSig, -expDiff, &aSig); aSig 492 arch/sh/kernel/cpu/sh4/softfloat.c zSig = bSig - aSig; aSig 506 arch/sh/kernel/cpu/sh4/softfloat.c aSig |= 0x40000000; aSig 508 arch/sh/kernel/cpu/sh4/softfloat.c zSig = aSig - bSig; aSig 519 arch/sh/kernel/cpu/sh4/softfloat.c bits32 aSig, bSig, zSig; aSig 522 arch/sh/kernel/cpu/sh4/softfloat.c aSig = extractFloat32Frac(a); aSig 527 arch/sh/kernel/cpu/sh4/softfloat.c aSig <<= 6; aSig 547 arch/sh/kernel/cpu/sh4/softfloat.c aSig |= 0x20000000; aSig 549 arch/sh/kernel/cpu/sh4/softfloat.c shift32RightJamming(aSig, -expDiff, &aSig); aSig 556 arch/sh/kernel/cpu/sh4/softfloat.c return packFloat32(zSign, 0, (aSig + bSig) >> 6); aSig 557 arch/sh/kernel/cpu/sh4/softfloat.c zSig = 0x40000000 + aSig + bSig; aSig 561 arch/sh/kernel/cpu/sh4/softfloat.c aSig |= 0x20000000; aSig 562 arch/sh/kernel/cpu/sh4/softfloat.c zSig = (aSig + bSig) << 1; aSig 565 arch/sh/kernel/cpu/sh4/softfloat.c zSig = aSig + bSig; aSig 629 arch/sh/kernel/cpu/sh4/softfloat.c normalizeFloat64Subnormal(bits64 aSig, int16 * zExpPtr, bits64 * zSigPtr) aSig 633 arch/sh/kernel/cpu/sh4/softfloat.c shiftCount = countLeadingZeros64(aSig) - 11; aSig 634 arch/sh/kernel/cpu/sh4/softfloat.c *zSigPtr = aSig << shiftCount; aSig 705 arch/sh/kernel/cpu/sh4/softfloat.c static void normalizeFloat32Subnormal(bits32 aSig, int16 * zExpPtr, aSig 710 arch/sh/kernel/cpu/sh4/softfloat.c shiftCount = countLeadingZeros32(aSig) - 8; aSig 711 arch/sh/kernel/cpu/sh4/softfloat.c *zSigPtr = aSig << shiftCount; aSig 720 arch/sh/kernel/cpu/sh4/softfloat.c bits64 aSig, bSig, zSig; aSig 724 arch/sh/kernel/cpu/sh4/softfloat.c aSig = extractFloat64Frac(a); aSig 741 arch/sh/kernel/cpu/sh4/softfloat.c if ((aExp | aSig) == 0) { aSig 749 arch/sh/kernel/cpu/sh4/softfloat.c if (aSig == 0) aSig 751 arch/sh/kernel/cpu/sh4/softfloat.c normalizeFloat64Subnormal(aSig, &aExp, &aSig); aSig 754 arch/sh/kernel/cpu/sh4/softfloat.c aSig = (aSig | LIT64(0x0010000000000000)) << 10; aSig 756 arch/sh/kernel/cpu/sh4/softfloat.c if (bSig <= (aSig + aSig)) { aSig 757 arch/sh/kernel/cpu/sh4/softfloat.c aSig >>= 1; aSig 760 arch/sh/kernel/cpu/sh4/softfloat.c zSig = estimateDiv128To64(aSig, 0, bSig); aSig 763 arch/sh/kernel/cpu/sh4/softfloat.c sub128(aSig, 0, term0, term1, &rem0, &rem1); aSig 778 arch/sh/kernel/cpu/sh4/softfloat.c bits32 aSig, bSig; aSig 781 arch/sh/kernel/cpu/sh4/softfloat.c aSig = extractFloat32Frac(a); aSig 803 arch/sh/kernel/cpu/sh4/softfloat.c if (aSig == 0) aSig 805 arch/sh/kernel/cpu/sh4/softfloat.c normalizeFloat32Subnormal(aSig, &aExp, &aSig); aSig 808 arch/sh/kernel/cpu/sh4/softfloat.c aSig = (aSig | 0x00800000) << 7; aSig 810 arch/sh/kernel/cpu/sh4/softfloat.c if (bSig <= (aSig + aSig)) { aSig 811 arch/sh/kernel/cpu/sh4/softfloat.c aSig >>= 1; aSig 814 arch/sh/kernel/cpu/sh4/softfloat.c zSig = (((bits64) aSig) << 32); aSig 818 arch/sh/kernel/cpu/sh4/softfloat.c zSig |= (((bits64) bSig) * zSig != ((bits64) aSig) << 32); aSig 828 arch/sh/kernel/cpu/sh4/softfloat.c unsigned int aSig, bSig; aSig 832 arch/sh/kernel/cpu/sh4/softfloat.c aSig = extractFloat32Frac(a); aSig 840 arch/sh/kernel/cpu/sh4/softfloat.c if (aSig == 0) aSig 842 arch/sh/kernel/cpu/sh4/softfloat.c normalizeFloat32Subnormal(aSig, &aExp, &aSig); aSig 849 arch/sh/kernel/cpu/sh4/softfloat.c if ((bExp == 0xff && bSig == 0) || (aExp == 0xff && aSig == 0)) aSig 853 arch/sh/kernel/cpu/sh4/softfloat.c aSig = (aSig | 0x00800000) << 7; aSig 855 arch/sh/kernel/cpu/sh4/softfloat.c shift64RightJamming(((unsigned long long)aSig) * bSig, 32, &zSig64); aSig 869 arch/sh/kernel/cpu/sh4/softfloat.c unsigned long long int aSig, bSig, zSig0, zSig1; aSig 871 arch/sh/kernel/cpu/sh4/softfloat.c aSig = extractFloat64Frac(a); aSig 880 arch/sh/kernel/cpu/sh4/softfloat.c if (aSig == 0) aSig 882 arch/sh/kernel/cpu/sh4/softfloat.c normalizeFloat64Subnormal(aSig, &aExp, &aSig); aSig 889 arch/sh/kernel/cpu/sh4/softfloat.c if ((aExp == 0x7ff && aSig == 0) || (bExp == 0x7ff && bSig == 0)) aSig 893 arch/sh/kernel/cpu/sh4/softfloat.c aSig = (aSig | 0x0010000000000000LL) << 10; aSig 895 arch/sh/kernel/cpu/sh4/softfloat.c mul64To128(aSig, bSig, &zSig0, &zSig1); aSig 916 arch/sh/kernel/cpu/sh4/softfloat.c bits64 aSig; aSig 919 arch/sh/kernel/cpu/sh4/softfloat.c aSig = extractFloat64Frac( a ); aSig 923 arch/sh/kernel/cpu/sh4/softfloat.c shift64RightJamming( aSig, 22, &aSig ); aSig 924 arch/sh/kernel/cpu/sh4/softfloat.c zSig = aSig;