Lines Matching refs:m32

27 		reg->lowmant = reg->mant.m32[1] << (8 - cnt);  in fp_denormalize()
28 reg->mant.m32[1] = (reg->mant.m32[1] >> cnt) | in fp_denormalize()
29 (reg->mant.m32[0] << (32 - cnt)); in fp_denormalize()
30 reg->mant.m32[0] = reg->mant.m32[0] >> cnt; in fp_denormalize()
33 reg->lowmant = reg->mant.m32[1] >> (cnt - 8); in fp_denormalize()
34 if (reg->mant.m32[1] << (40 - cnt)) in fp_denormalize()
36 reg->mant.m32[1] = (reg->mant.m32[1] >> cnt) | in fp_denormalize()
37 (reg->mant.m32[0] << (32 - cnt)); in fp_denormalize()
38 reg->mant.m32[0] = reg->mant.m32[0] >> cnt; in fp_denormalize()
42 : "m" (reg->mant.m32[0]), "d" (64 - cnt)); in fp_denormalize()
43 if (reg->mant.m32[1] << (40 - cnt)) in fp_denormalize()
45 reg->mant.m32[1] = reg->mant.m32[0] >> (cnt - 32); in fp_denormalize()
46 reg->mant.m32[0] = 0; in fp_denormalize()
49 reg->lowmant = reg->mant.m32[0] >> (cnt - 40); in fp_denormalize()
50 if ((reg->mant.m32[0] << (72 - cnt)) || reg->mant.m32[1]) in fp_denormalize()
52 reg->mant.m32[1] = reg->mant.m32[0] >> (cnt - 32); in fp_denormalize()
53 reg->mant.m32[0] = 0; in fp_denormalize()
56 reg->lowmant = reg->mant.m32[0] || reg->mant.m32[1]; in fp_denormalize()
57 reg->mant.m32[0] = 0; in fp_denormalize()
58 reg->mant.m32[1] = 0; in fp_denormalize()
67 if (reg->mant.m32[0]) { in fp_overnormalize()
68 asm ("bfffo %1{#0,#32},%0" : "=d" (shift) : "dm" (reg->mant.m32[0])); in fp_overnormalize()
69 reg->mant.m32[0] = (reg->mant.m32[0] << shift) | (reg->mant.m32[1] >> (32 - shift)); in fp_overnormalize()
70 reg->mant.m32[1] = (reg->mant.m32[1] << shift); in fp_overnormalize()
72 asm ("bfffo %1{#0,#32},%0" : "=d" (shift) : "dm" (reg->mant.m32[1])); in fp_overnormalize()
73 reg->mant.m32[0] = (reg->mant.m32[1] << shift); in fp_overnormalize()
74 reg->mant.m32[1] = 0; in fp_overnormalize()
88 asm volatile ("addx.l %1,%0" : "=d" (dest->mant.m32[1]) in fp_addmant()
89 : "d" (src->mant.m32[1]), "0" (dest->mant.m32[1])); in fp_addmant()
90 asm volatile ("addx.l %1,%0" : "=d" (dest->mant.m32[0]) in fp_addmant()
91 : "d" (src->mant.m32[0]), "0" (dest->mant.m32[0])); in fp_addmant()
106 reg->lowmant = (reg->mant.m32[1] << 7) | (reg->lowmant ? 1 : 0); in fp_addcarry()
107 reg->mant.m32[1] = (reg->mant.m32[1] >> 1) | in fp_addcarry()
108 (reg->mant.m32[0] << 31); in fp_addcarry()
109 reg->mant.m32[0] = (reg->mant.m32[0] >> 1) | 0x80000000; in fp_addcarry()
120 asm volatile ("subx.l %1,%0" : "=d" (dest->mant.m32[1]) in fp_submant()
121 : "d" (src2->mant.m32[1]), "0" (src1->mant.m32[1])); in fp_submant()
122 asm volatile ("subx.l %1,%0" : "=d" (dest->mant.m32[0]) in fp_submant()
123 : "d" (src2->mant.m32[0]), "0" (src1->mant.m32[0])); in fp_submant()
141 asm volatile ("add.l %1,%0" : "=d,g" (dest->m32[2]) \
142 : "g,d" (temp.m32[1]), "0,0" (dest->m32[2])); \
143 asm volatile ("addx.l %1,%0" : "=d" (dest->m32[1]) \
144 : "d" (temp.m32[0]), "0" (dest->m32[1])); \
145 asm volatile ("addx.l %1,%0" : "=d" (dest->m32[0]) \
146 : "d" (0), "0" (dest->m32[0])); \
149 asm ("sub.l %1,%0" : "=d,dm" (dest.m32[1]) \
150 : "dm,d" (src.m32[1]), "0,0" (dest.m32[1])); \
151 asm ("subx.l %1,%0" : "=d" (dest.m32[0]) \
152 : "d" (src.m32[0]), "0" (dest.m32[0])); \
156 asm ("sub.l %1,%0" : "=d,dm" (dest.m32[2]) \
157 : "dm,d" (srcl), "0,0" (dest.m32[2])); \
158 asm ("subx.l %1,%0" : "=d" (dest.m32[1]) \
159 : "d" (srcm), "0" (dest.m32[1])); \
160 asm ("subx.l %2,%1; scs %0" : "=d" (carry), "=d" (dest.m32[0]) \
161 : "d" (srch), "1" (dest.m32[0])); \
170 fp_mul64(dest->m32[0], dest->m32[1], src1->mant.m32[0], src2->mant.m32[0]); in fp_multiplymant()
171 fp_mul64(dest->m32[2], dest->m32[3], src1->mant.m32[1], src2->mant.m32[1]); in fp_multiplymant()
173 fp_mul64(temp.m32[0], temp.m32[1], src1->mant.m32[0], src2->mant.m32[1]); in fp_multiplymant()
176 fp_mul64(temp.m32[0], temp.m32[1], src1->mant.m32[1], src2->mant.m32[0]); in fp_multiplymant()
185 unsigned long *mantp = dest->m32; in fp_dividemant()
208 dummy = div->mant.m32[1] / div->mant.m32[0] + 1; in fp_dividemant()
214 if (src->mant.m32[0] == div->mant.m32[0]) { in fp_dividemant()
215 fp_div64(first, rem, 0, src->mant.m32[1], div->mant.m32[0]); in fp_dividemant()
220 fp_div64(first, rem, src->mant.m32[0], src->mant.m32[1], div->mant.m32[0]); in fp_dividemant()
225 fp_mul64(tmp.m32[0], tmp.m32[1], div->mant.m32[0], first - *mantp); in fp_dividemant()
226 fp_add64(tmp.m32[0], tmp.m32[1], 0, rem); in fp_dividemant()
227 tmp.m32[2] = 0; in fp_dividemant()
229 fp_mul64(tmp64.m32[0], tmp64.m32[1], *mantp, div->mant.m32[1]); in fp_dividemant()
230 fp_sub96c(tmp, 0, tmp64.m32[0], tmp64.m32[1]); in fp_dividemant()
232 src->mant.m32[0] = tmp.m32[1]; in fp_dividemant()
233 src->mant.m32[1] = tmp.m32[2]; in fp_dividemant()
235 while (!fp_sub96c(tmp, 0, div->mant.m32[0], div->mant.m32[1])) { in fp_dividemant()
236 src->mant.m32[0] = tmp.m32[1]; in fp_dividemant()
237 src->mant.m32[1] = tmp.m32[2]; in fp_dividemant()
251 dest->lowmant = src->m32[2] >> 24; in fp_putmant128()
252 if (src->m32[3] || (src->m32[2] << 8)) in fp_putmant128()
257 : "=d" (tmp) : "0" (src->m32[2])); in fp_putmant128()
259 : "=d" (dest->mant.m32[1]) : "0" (src->m32[1])); in fp_putmant128()
261 : "=d" (dest->mant.m32[0]) : "0" (src->m32[0])); in fp_putmant128()
263 if (src->m32[3] || (tmp << 8)) in fp_putmant128()
268 : "=d" (dest->mant.m32[0]) in fp_putmant128()
269 : "d" (src->m32[0]), "0" (src->m32[1])); in fp_putmant128()
271 : "=d" (dest->mant.m32[1]) : "0" (src->m32[2])); in fp_putmant128()
273 : "=d" (tmp) : "0" (src->m32[3])); in fp_putmant128()
275 if (src->m32[3] << 7) in fp_putmant128()
279 dest->mant.m32[0] = src->m32[1]; in fp_putmant128()
280 dest->mant.m32[1] = src->m32[2]; in fp_putmant128()
281 dest->lowmant = src->m32[3] >> 24; in fp_putmant128()
282 if (src->m32[3] << 8) in fp_putmant128()