int32 eax stack32 eax_stack int32 ebx stack32 ebx_stack int32 esi stack32 esi_stack int32 edi stack32 edi_stack int32 ebp stack32 ebp_stack stack32 q2_stack int32 q2 stack32 q3_stack int32 q3 stack32 x1overy1z1t1_stack int32 x1overy1z1t1 stack32 surface_stack int32 surface stack32 e_stack int32 e int32 selectbit int32 thisword int32 bytes int32 bits stack32 selectbit_stack stack32 thisword_stack stack32 bits_stack stack32 bytes_stack stack64 xovery0 stack64 xovery1 stack64 xovery2 stack64 xovery3 stack64 xovery4 stack64 xoverz0 stack64 xoverz1 stack64 xoverz2 stack64 xoverz3 stack64 xoverz4 stack64 xovert0 stack64 xovert1 stack64 xovert2 stack64 xovert3 stack64 xovert4 stack64 aoverb0 stack64 aoverb1 stack64 aoverb2 stack64 aoverb3 stack64 aoverb4 stack64 aoverc0 stack64 aoverc1 stack64 aoverc2 stack64 aoverc3 stack64 aoverc4 stack64 aoverd0 stack64 aoverd1 stack64 aoverd2 stack64 aoverd3 stack64 aoverd4 stack64 AAoverBB0 stack64 AAoverBB1 stack64 AAoverBB2 stack64 AAoverBB3 stack64 AAoverBB4 stack64 AAoverCC0 stack64 AAoverCC1 stack64 AAoverCC2 stack64 AAoverCC3 stack64 AAoverCC4 stack64 AAoverDD0 stack64 AAoverDD1 stack64 AAoverDD2 stack64 AAoverDD3 stack64 AAoverDD4 stack64 q2x0 stack64 q2x1 stack64 q2x2 stack64 q2x3 stack64 q2x4 stack64 q2y0 stack64 q2y1 stack64 q2y2 stack64 q2y3 stack64 q2y4 stack64 q2z0 stack64 q2z1 stack64 q2z2 stack64 q2z3 stack64 q2z4 stack64 q2t0 stack64 q2t1 stack64 q2t2 stack64 q2t3 stack64 q2t4 stack64 q3x0 stack64 q3x1 stack64 q3x2 stack64 q3x3 stack64 q3x4 stack64 q3y0 stack64 q3y1 stack64 q3y2 stack64 q3y3 stack64 q3y4 stack64 q3z0 stack64 q3z1 stack64 q3z2 stack64 q3z3 stack64 q3z4 stack64 q3t0 stack64 q3t1 stack64 q3t2 stack64 q3t3 stack64 q3t4 stack64 exponentbitshifted float80 use1 float80 x0 float80 x1 float80 x2 float80 x3 float80 x4 float80 x5 float80 x6 float80 x7 float80 x8 float80 x9 float80 y0 float80 y1 float80 y2 float80 y3 float80 y4 float80 y5 float80 y6 float80 y7 float80 y8 float80 y9 stack64 leftpmx0 stack64 leftpmx1 stack64 leftpmx2 stack64 leftpmx3 stack64 leftpmx4 stack64 leftpmy0 stack64 leftpmy1 stack64 leftpmy2 stack64 leftpmy3 stack64 leftpmy4 stack64 leftpmz0 stack64 leftpmz1 stack64 leftpmz2 stack64 leftpmz3 stack64 leftpmz4 stack64 leftpmt0 stack64 leftpmt1 stack64 leftpmt2 stack64 leftpmt3 stack64 leftpmt4 stack64 rightpmx0 stack64 rightpmx1 stack64 rightpmx2 stack64 rightpmx3 stack64 rightpmx4 stack64 rightpmy0 stack64 rightpmy1 stack64 rightpmy2 stack64 rightpmy3 stack64 rightpmy4 stack64 rightpmz0 stack64 rightpmz1 stack64 rightpmz2 stack64 rightpmz3 stack64 rightpmz4 stack64 rightpmt0 stack64 rightpmt1 stack64 rightpmt2 stack64 rightpmt3 stack64 rightpmt4 stack64 left0x0 stack64 left0x1 stack64 left0x2 stack64 left0x3 stack64 left0x4 stack64 left0y0 stack64 left0y1 stack64 left0y2 stack64 left0y3 stack64 left0y4 stack64 left0z0 stack64 left0z1 stack64 left0z2 stack64 left0z3 stack64 left0z4 stack64 left0t0 stack64 left0t1 stack64 left0t2 stack64 left0t3 stack64 left0t4 stack64 right0x0 stack64 right0x1 stack64 right0x2 stack64 right0x3 stack64 right0x4 stack64 right0y0 stack64 right0y1 stack64 right0y2 stack64 right0y3 stack64 right0y4 stack64 right0z0 stack64 right0z1 stack64 right0z2 stack64 right0z3 stack64 right0z4 stack64 right0t0 stack64 right0t1 stack64 right0t2 stack64 right0t3 stack64 right0t4 stack64 left1x0 stack64 left1x1 stack64 left1x2 stack64 left1x3 stack64 left1x4 stack64 left1y0 stack64 left1y1 stack64 left1y2 stack64 left1y3 stack64 left1y4 stack64 left1z0 stack64 left1z1 stack64 left1z2 stack64 left1z3 stack64 left1z4 stack64 left1t0 stack64 left1t1 stack64 left1t2 stack64 left1t3 stack64 left1t4 stack64 right1x0 stack64 right1x1 stack64 right1x2 stack64 right1x3 stack64 right1x4 stack64 right1y0 stack64 right1y1 stack64 right1y2 stack64 right1y3 stack64 right1y4 stack64 right1z0 stack64 right1z1 stack64 right1z2 stack64 right1z3 stack64 right1z4 stack64 right1t0 stack64 right1t1 stack64 right1t2 stack64 right1t3 stack64 right1t4 stack64 left2x0 stack64 left2x1 stack64 left2x2 stack64 left2x3 stack64 left2x4 stack64 left2y0 stack64 left2y1 stack64 left2y2 stack64 left2y3 stack64 left2y4 stack64 left2z0 stack64 left2z1 stack64 left2z2 stack64 left2z3 stack64 left2z4 stack64 left2t0 stack64 left2t1 stack64 left2t2 stack64 left2t3 stack64 left2t4 stack64 right2x0 stack64 right2x1 stack64 right2x2 stack64 right2x3 stack64 right2x4 stack64 right2y0 stack64 right2y1 stack64 right2y2 stack64 right2y3 stack64 right2y4 stack64 right2z0 stack64 right2z1 stack64 right2z2 stack64 right2z3 stack64 right2z4 stack64 right2t0 stack64 right2t1 stack64 right2t2 stack64 right2t3 stack64 right2t4 stack64 left3x0 stack64 left3x1 stack64 left3x2 stack64 left3x3 stack64 left3x4 stack64 left3y0 stack64 left3y1 stack64 left3y2 stack64 left3y3 stack64 left3y4 stack64 left3z0 stack64 left3z1 stack64 left3z2 stack64 left3z3 stack64 left3z4 stack64 left3t0 stack64 left3t1 stack64 left3t2 stack64 left3t3 stack64 left3t4 stack64 left4x0 stack64 left4x1 stack64 left4x2 stack64 left4x3 stack64 left4x4 stack64 left4y0 stack64 left4y1 stack64 left4y2 stack64 left4y3 stack64 left4y4 stack64 left4z0 stack64 left4z1 stack64 left4z2 stack64 left4z3 stack64 left4z4 stack64 left4t0 stack64 left4t1 stack64 left4t2 stack64 left4t3 stack64 left4t4 stack64 right4x0 stack64 right4x1 stack64 right4x2 stack64 right4x3 stack64 right4x4 stack64 right4y0 stack64 right4y1 stack64 right4y2 stack64 right4y3 stack64 right4y4 stack64 right4z0 stack64 right4z1 stack64 right4z2 stack64 right4z3 stack64 right4z4 stack64 right4t0 stack64 right4t1 stack64 right4t2 stack64 right4t3 stack64 right4t4 stack64 left5x0 stack64 left5x1 stack64 left5x2 stack64 left5x3 stack64 left5x4 stack64 left5y0 stack64 left5y1 stack64 left5y2 stack64 left5y3 stack64 left5y4 stack64 left5z0 stack64 left5z1 stack64 left5z2 stack64 left5z3 stack64 left5z4 stack64 left5t0 stack64 left5t1 stack64 left5t2 stack64 left5t3 stack64 left5t4 stack64 right5x0 stack64 right5x1 stack64 right5x2 stack64 right5x3 stack64 right5x4 stack64 right5y0 stack64 right5y1 stack64 right5y2 stack64 right5y3 stack64 right5y4 stack64 right5z0 stack64 right5z1 stack64 right5z2 stack64 right5z3 stack64 right5z4 stack64 right5t0 stack64 right5t1 stack64 right5t2 stack64 right5t3 stack64 right5t4 float80 out0 float80 out1 float80 out2 float80 out3 float80 out4 float80 carry0 float80 carry1 float80 carry2 float80 carry3 float80 carry4 float80 u0v0 float80 u0v1 float80 u0v2 float80 u0v3 float80 u0v4 float80 u1v0 float80 u1v1 float80 u1v2 float80 u1v3 float80 u1v4 float80 u2v0 float80 u2v1 float80 u2v2 float80 u2v3 float80 u2v4 float80 u3v0 float80 u3v1 float80 u3v2 float80 u3v3 float80 u3v4 float80 u4v0 float80 u4v1 float80 u4v2 float80 u4v3 float80 u4v4 float80 u22 float80 u32 float80 u42 stack64 u22_stack stack64 u32_stack float80 u0u0 float80 u0u1 float80 u0u2 float80 u0u3 float80 u1u1 float80 u1u2 float80 u1u3 float80 u2u2 float80 u2u3 float80 u3u3 float80 d0 float80 d1 float80 d2 float80 d3 float80 d4 float80 e0 float80 e1 float80 e2 float80 e3 float80 e4 float80 f0 float80 f1 float80 f2 float80 f3 float80 f4 float80 g0 float80 g1 float80 g2 float80 g3 float80 g4 float80 h0 float80 h1 float80 h2 float80 h3 float80 h4 float80 i0 float80 i1 float80 i2 float80 i3 float80 i4 float80 j0 float80 j1 float80 j2 float80 j3 float80 j4 float80 k0 float80 k1 float80 k2 float80 k3 float80 k4 enter gaudry_pm_surface_doubleadd input q2_stack input q3_stack input x1overy1z1t1_stack input surface_stack input e_stack caller eax caller ebx caller esi caller edi caller ebp eax_stack = eax ebx_stack = ebx esi_stack = esi edi_stack = edi ebp_stack = ebp x1overy1z1t1 = x1overy1z1t1_stack x0 = *(float64 *) (x1overy1z1t1 + 0) x1 = *(float64 *) (x1overy1z1t1 + 8) x2 = *(float64 *) (x1overy1z1t1 + 16) x3 = *(float64 *) (x1overy1z1t1 + 24) x4 = *(float64 *) (x1overy1z1t1 + 32) *(float64 *) &xovery0 = x0 *(float64 *) &xovery1 = x1 *(float64 *) &xovery2 = x2 *(float64 *) &xovery3 = x3 *(float64 *) &xovery4 = x4 x0 = *(float64 *) (x1overy1z1t1 + 40) x1 = *(float64 *) (x1overy1z1t1 + 48) x2 = *(float64 *) (x1overy1z1t1 + 56) x3 = *(float64 *) (x1overy1z1t1 + 64) x4 = *(float64 *) (x1overy1z1t1 + 72) *(float64 *) &xoverz0 = x0 *(float64 *) &xoverz1 = x1 *(float64 *) &xoverz2 = x2 *(float64 *) &xoverz3 = x3 *(float64 *) &xoverz4 = x4 x0 = *(float64 *) (x1overy1z1t1 + 80) x1 = *(float64 *) (x1overy1z1t1 + 88) x2 = *(float64 *) (x1overy1z1t1 + 96) x3 = *(float64 *) (x1overy1z1t1 + 104) x4 = *(float64 *) (x1overy1z1t1 + 112) *(float64 *) &xovert0 = x0 *(float64 *) &xovert1 = x1 *(float64 *) &xovert2 = x2 *(float64 *) &xovert3 = x3 *(float64 *) &xovert4 = x4 surface = surface_stack x0 = *(float64 *) (surface + 0) x1 = *(float64 *) (surface + 8) x2 = *(float64 *) (surface + 16) x3 = *(float64 *) (surface + 24) x4 = *(float64 *) (surface + 32) *(float64 *) &AAoverBB0 = x0 *(float64 *) &AAoverBB1 = x1 *(float64 *) &AAoverBB2 = x2 *(float64 *) &AAoverBB3 = x3 *(float64 *) &AAoverBB4 = x4 x0 = *(float64 *) (surface + 40) x1 = *(float64 *) (surface + 48) x2 = *(float64 *) (surface + 56) x3 = *(float64 *) (surface + 64) x4 = *(float64 *) (surface + 72) *(float64 *) &AAoverCC0 = x0 *(float64 *) &AAoverCC1 = x1 *(float64 *) &AAoverCC2 = x2 *(float64 *) &AAoverCC3 = x3 *(float64 *) &AAoverCC4 = x4 x0 = *(float64 *) (surface + 80) x1 = *(float64 *) (surface + 88) x2 = *(float64 *) (surface + 96) x3 = *(float64 *) (surface + 104) x4 = *(float64 *) (surface + 112) *(float64 *) &AAoverDD0 = x0 *(float64 *) &AAoverDD1 = x1 *(float64 *) &AAoverDD2 = x2 *(float64 *) &AAoverDD3 = x3 *(float64 *) &AAoverDD4 = x4 x0 = *(float64 *) (surface + 120) x1 = *(float64 *) (surface + 128) x2 = *(float64 *) (surface + 136) x3 = *(float64 *) (surface + 144) x4 = *(float64 *) (surface + 152) *(float64 *) &aoverb0 = x0 *(float64 *) &aoverb1 = x1 *(float64 *) &aoverb2 = x2 *(float64 *) &aoverb3 = x3 *(float64 *) &aoverb4 = x4 x0 = *(float64 *) (surface + 160) x1 = *(float64 *) (surface + 168) x2 = *(float64 *) (surface + 176) x3 = *(float64 *) (surface + 184) x4 = *(float64 *) (surface + 192) *(float64 *) &aoverc0 = x0 *(float64 *) &aoverc1 = x1 *(float64 *) &aoverc2 = x2 *(float64 *) &aoverc3 = x3 *(float64 *) &aoverc4 = x4 x0 = *(float64 *) (surface + 200) x1 = *(float64 *) (surface + 208) x2 = *(float64 *) (surface + 216) x3 = *(float64 *) (surface + 224) x4 = *(float64 *) (surface + 232) *(float64 *) &aoverd0 = x0 *(float64 *) &aoverd1 = x1 *(float64 *) &aoverd2 = x2 *(float64 *) &aoverd3 = x3 *(float64 *) &aoverd4 = x4 q2 = q2_stack x0 = *(float64 *) (q2 + 0) x1 = *(float64 *) (q2 + 8) x2 = *(float64 *) (q2 + 16) x3 = *(float64 *) (q2 + 24) x4 = *(float64 *) (q2 + 32) *(float64 *) &q2x0 = x0 *(float64 *) &q2x1 = x1 *(float64 *) &q2x2 = x2 *(float64 *) &q2x3 = x3 *(float64 *) &q2x4 = x4 x0 = *(float64 *) (q2 + 40) x1 = *(float64 *) (q2 + 48) x2 = *(float64 *) (q2 + 56) x3 = *(float64 *) (q2 + 64) x4 = *(float64 *) (q2 + 72) *(float64 *) &q2y0 = x0 *(float64 *) &q2y1 = x1 *(float64 *) &q2y2 = x2 *(float64 *) &q2y3 = x3 *(float64 *) &q2y4 = x4 x0 = *(float64 *) (q2 + 80) x1 = *(float64 *) (q2 + 88) x2 = *(float64 *) (q2 + 96) x3 = *(float64 *) (q2 + 104) x4 = *(float64 *) (q2 + 112) *(float64 *) &q2z0 = x0 *(float64 *) &q2z1 = x1 *(float64 *) &q2z2 = x2 *(float64 *) &q2z3 = x3 *(float64 *) &q2z4 = x4 x0 = *(float64 *) (q2 + 120) x1 = *(float64 *) (q2 + 128) x2 = *(float64 *) (q2 + 136) x3 = *(float64 *) (q2 + 144) x4 = *(float64 *) (q2 + 152) *(float64 *) &q2t0 = x0 *(float64 *) &q2t1 = x1 *(float64 *) &q2t2 = x2 *(float64 *) &q2t3 = x3 *(float64 *) &q2t4 = x4 q3 = q3_stack x0 = *(float64 *) (q3 + 0) x1 = *(float64 *) (q3 + 8) x2 = *(float64 *) (q3 + 16) x3 = *(float64 *) (q3 + 24) x4 = *(float64 *) (q3 + 32) *(float64 *) &q3x0 = x0 *(float64 *) &q3x1 = x1 *(float64 *) &q3x2 = x2 *(float64 *) &q3x3 = x3 *(float64 *) &q3x4 = x4 x0 = *(float64 *) (q3 + 40) x1 = *(float64 *) (q3 + 48) x2 = *(float64 *) (q3 + 56) x3 = *(float64 *) (q3 + 64) x4 = *(float64 *) (q3 + 72) *(float64 *) &q3y0 = x0 *(float64 *) &q3y1 = x1 *(float64 *) &q3y2 = x2 *(float64 *) &q3y3 = x3 *(float64 *) &q3y4 = x4 x0 = *(float64 *) (q3 + 80) x1 = *(float64 *) (q3 + 88) x2 = *(float64 *) (q3 + 96) x3 = *(float64 *) (q3 + 104) x4 = *(float64 *) (q3 + 112) *(float64 *) &q3z0 = x0 *(float64 *) &q3z1 = x1 *(float64 *) &q3z2 = x2 *(float64 *) &q3z3 = x3 *(float64 *) &q3z4 = x4 x0 = *(float64 *) (q3 + 120) x1 = *(float64 *) (q3 + 128) x2 = *(float64 *) (q3 + 136) x3 = *(float64 *) (q3 + 144) x4 = *(float64 *) (q3 + 152) *(float64 *) &q3t0 = x0 *(float64 *) &q3t1 = x1 *(float64 *) &q3t2 = x2 *(float64 *) &q3t3 = x3 *(float64 *) &q3t4 = x4 # selectbit = 0; # for (pos = 254;pos >= 0;--pos) { # ebit = 1 & (e[pos / 8] >> (pos % 8)); # selectbit = selectbit ^ ebit; # ... # selectbit = ebit; # } bytes = 28 bits = 31 e = e_stack thisword = *(uint32 *) (e + 28) thisword <<<= 1 selectbit = 0 morebytes: bytes_stack = bytes morebits: thisword <<<= 1 thisword_stack = thisword bits_stack = bits thisword &= 1 selectbit ^= thisword exponentbitshifted top = 0x43300000 inplace exponentbitshifted bottom = selectbit selectbit_stack = thisword # left0x = square(q2x) out3 = *(float64 *) &q2x4 out3 *= out3 u22 = *(float64 *) &q2x2 u22 += u22 u32 = *(float64 *) &q2x3 u32 += u32 u42 = *(float64 *) &q2x4 u42 += u42 out3 *= *(float64 *) &gaudry_pm_scale u1u2 = *(float64 *) &q2x1 u1u2 *= u22 out3 += u1u2 u0u3 = *(float64 *) &q2x0 u0u3 *= u32 out3 += u0u3 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 *(float64 *) &u22_stack = u22 *(float64 *) &u32_stack = u32 out4 = *(float64 *) &q2x0 out4 *= u42 u1u3 = *(float64 *) &q2x1 u1u3 *= u32 out4 += u1u3 u2u2 = *(float64 *) &q2x2 u2u2 *= u2u2 out4 += u2u2 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &q2x1 out0 *= u42 u2u3 = *(float64 *) &q2x2 u2u3 *= *(float64 *) &u32_stack out0 += u2u3 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0u0 = *(float64 *) &q2x0 u0u0 *= u0u0 out0 += u0u0 out1 = *(float64 *) &q2x2 out1 *= u42 u3u3 = *(float64 *) &q2x3 u3u3 *= u3u3 out1 += u3u3 out1 *= *(float64 *) &gaudry_pm_scale u0u1 = *(float64 *) &q2x0 u0u1 += u0u1 u0u1 *= *(float64 *) &q2x1 out1 += u0u1 out2 = *(float64 *) &q2x3 out2 *= u42 out2 *= *(float64 *) &gaudry_pm_scale u0u2 = *(float64 *) &q2x0 u0u2 *= *(float64 *) &u22_stack out2 += u0u2 u1u1 = *(float64 *) &q2x1 u1u1 *= u1u1 out2 += u1u1 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 *(float64 *) &left0x0 = out0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out1 -= carry1 *(float64 *) &left0x1 = out1 out2 -= carry2 *(float64 *) &left0x2 = out2 out3 -= carry3 *(float64 *) &left0x3 = out3 out4 += carry3 *(float64 *) &left0x4 = out4 # left0y = square(q2y) out3 = *(float64 *) &q2y4 out3 *= out3 u22 = *(float64 *) &q2y2 u22 += u22 u32 = *(float64 *) &q2y3 u32 += u32 u42 = *(float64 *) &q2y4 u42 += u42 out3 *= *(float64 *) &gaudry_pm_scale u1u2 = *(float64 *) &q2y1 u1u2 *= u22 out3 += u1u2 u0u3 = *(float64 *) &q2y0 u0u3 *= u32 out3 += u0u3 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 *(float64 *) &u22_stack = u22 *(float64 *) &u32_stack = u32 out4 = *(float64 *) &q2y0 out4 *= u42 u1u3 = *(float64 *) &q2y1 u1u3 *= u32 out4 += u1u3 u2u2 = *(float64 *) &q2y2 u2u2 *= u2u2 out4 += u2u2 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &q2y1 out0 *= u42 u2u3 = *(float64 *) &q2y2 u2u3 *= *(float64 *) &u32_stack out0 += u2u3 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0u0 = *(float64 *) &q2y0 u0u0 *= u0u0 out0 += u0u0 out1 = *(float64 *) &q2y2 out1 *= u42 u3u3 = *(float64 *) &q2y3 u3u3 *= u3u3 out1 += u3u3 out1 *= *(float64 *) &gaudry_pm_scale u0u1 = *(float64 *) &q2y0 u0u1 += u0u1 u0u1 *= *(float64 *) &q2y1 out1 += u0u1 out2 = *(float64 *) &q2y3 out2 *= u42 out2 *= *(float64 *) &gaudry_pm_scale u0u2 = *(float64 *) &q2y0 u0u2 *= *(float64 *) &u22_stack out2 += u0u2 u1u1 = *(float64 *) &q2y1 u1u1 *= u1u1 out2 += u1u1 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 *(float64 *) &left0y0 = out0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out1 -= carry1 *(float64 *) &left0y1 = out1 out2 -= carry2 *(float64 *) &left0y2 = out2 out3 -= carry3 *(float64 *) &left0y3 = out3 out4 += carry3 *(float64 *) &left0y4 = out4 # left0z = square(q2z) out3 = *(float64 *) &q2z4 out3 *= out3 u22 = *(float64 *) &q2z2 u22 += u22 u32 = *(float64 *) &q2z3 u32 += u32 u42 = *(float64 *) &q2z4 u42 += u42 out3 *= *(float64 *) &gaudry_pm_scale u1u2 = *(float64 *) &q2z1 u1u2 *= u22 out3 += u1u2 u0u3 = *(float64 *) &q2z0 u0u3 *= u32 out3 += u0u3 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 *(float64 *) &u22_stack = u22 *(float64 *) &u32_stack = u32 out4 = *(float64 *) &q2z0 out4 *= u42 u1u3 = *(float64 *) &q2z1 u1u3 *= u32 out4 += u1u3 u2u2 = *(float64 *) &q2z2 u2u2 *= u2u2 out4 += u2u2 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &q2z1 out0 *= u42 u2u3 = *(float64 *) &q2z2 u2u3 *= *(float64 *) &u32_stack out0 += u2u3 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0u0 = *(float64 *) &q2z0 u0u0 *= u0u0 out0 += u0u0 out1 = *(float64 *) &q2z2 out1 *= u42 u3u3 = *(float64 *) &q2z3 u3u3 *= u3u3 out1 += u3u3 out1 *= *(float64 *) &gaudry_pm_scale u0u1 = *(float64 *) &q2z0 u0u1 += u0u1 u0u1 *= *(float64 *) &q2z1 out1 += u0u1 out2 = *(float64 *) &q2z3 out2 *= u42 out2 *= *(float64 *) &gaudry_pm_scale u0u2 = *(float64 *) &q2z0 u0u2 *= *(float64 *) &u22_stack out2 += u0u2 u1u1 = *(float64 *) &q2z1 u1u1 *= u1u1 out2 += u1u1 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 *(float64 *) &left0z0 = out0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out1 -= carry1 *(float64 *) &left0z1 = out1 out2 -= carry2 *(float64 *) &left0z2 = out2 out3 -= carry3 *(float64 *) &left0z3 = out3 out4 += carry3 *(float64 *) &left0z4 = out4 # left0t = square(q2t) out3 = *(float64 *) &q2t4 out3 *= out3 u22 = *(float64 *) &q2t2 u22 += u22 u32 = *(float64 *) &q2t3 u32 += u32 u42 = *(float64 *) &q2t4 u42 += u42 out3 *= *(float64 *) &gaudry_pm_scale u1u2 = *(float64 *) &q2t1 u1u2 *= u22 out3 += u1u2 u0u3 = *(float64 *) &q2t0 u0u3 *= u32 out3 += u0u3 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 *(float64 *) &u22_stack = u22 *(float64 *) &u32_stack = u32 out4 = *(float64 *) &q2t0 out4 *= u42 u1u3 = *(float64 *) &q2t1 u1u3 *= u32 out4 += u1u3 u2u2 = *(float64 *) &q2t2 u2u2 *= u2u2 out4 += u2u2 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &q2t1 out0 *= u42 u2u3 = *(float64 *) &q2t2 u2u3 *= *(float64 *) &u32_stack out0 += u2u3 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0u0 = *(float64 *) &q2t0 u0u0 *= u0u0 out0 += u0u0 out1 = *(float64 *) &q2t2 out1 *= u42 u3u3 = *(float64 *) &q2t3 u3u3 *= u3u3 out1 += u3u3 out1 *= *(float64 *) &gaudry_pm_scale u0u1 = *(float64 *) &q2t0 u0u1 += u0u1 u0u1 *= *(float64 *) &q2t1 out1 += u0u1 out2 = *(float64 *) &q2t3 out2 *= u42 out2 *= *(float64 *) &gaudry_pm_scale u0u2 = *(float64 *) &q2t0 u0u2 *= *(float64 *) &u22_stack out2 += u0u2 u1u1 = *(float64 *) &q2t1 u1u1 *= u1u1 out2 += u1u1 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 *(float64 *) &left0t0 = out0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out1 -= carry1 *(float64 *) &left0t1 = out1 out2 -= carry2 *(float64 *) &left0t2 = out2 out3 -= carry3 *(float64 *) &left0t3 = out3 out4 += carry3 *(float64 *) &left0t4 = out4 # right0x = square(q3x) out3 = *(float64 *) &q3x4 out3 *= out3 u22 = *(float64 *) &q3x2 u22 += u22 u32 = *(float64 *) &q3x3 u32 += u32 u42 = *(float64 *) &q3x4 u42 += u42 out3 *= *(float64 *) &gaudry_pm_scale u1u2 = *(float64 *) &q3x1 u1u2 *= u22 out3 += u1u2 u0u3 = *(float64 *) &q3x0 u0u3 *= u32 out3 += u0u3 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 *(float64 *) &u22_stack = u22 *(float64 *) &u32_stack = u32 out4 = *(float64 *) &q3x0 out4 *= u42 u1u3 = *(float64 *) &q3x1 u1u3 *= u32 out4 += u1u3 u2u2 = *(float64 *) &q3x2 u2u2 *= u2u2 out4 += u2u2 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &q3x1 out0 *= u42 u2u3 = *(float64 *) &q3x2 u2u3 *= *(float64 *) &u32_stack out0 += u2u3 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0u0 = *(float64 *) &q3x0 u0u0 *= u0u0 out0 += u0u0 out1 = *(float64 *) &q3x2 out1 *= u42 u3u3 = *(float64 *) &q3x3 u3u3 *= u3u3 out1 += u3u3 out1 *= *(float64 *) &gaudry_pm_scale u0u1 = *(float64 *) &q3x0 u0u1 += u0u1 u0u1 *= *(float64 *) &q3x1 out1 += u0u1 out2 = *(float64 *) &q3x3 out2 *= u42 out2 *= *(float64 *) &gaudry_pm_scale u0u2 = *(float64 *) &q3x0 u0u2 *= *(float64 *) &u22_stack out2 += u0u2 u1u1 = *(float64 *) &q3x1 u1u1 *= u1u1 out2 += u1u1 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 *(float64 *) &right0x0 = out0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out1 -= carry1 *(float64 *) &right0x1 = out1 out2 -= carry2 *(float64 *) &right0x2 = out2 out3 -= carry3 *(float64 *) &right0x3 = out3 out4 += carry3 *(float64 *) &right0x4 = out4 # right0y = square(q3y) out3 = *(float64 *) &q3y4 out3 *= out3 u22 = *(float64 *) &q3y2 u22 += u22 u32 = *(float64 *) &q3y3 u32 += u32 u42 = *(float64 *) &q3y4 u42 += u42 out3 *= *(float64 *) &gaudry_pm_scale u1u2 = *(float64 *) &q3y1 u1u2 *= u22 out3 += u1u2 u0u3 = *(float64 *) &q3y0 u0u3 *= u32 out3 += u0u3 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 *(float64 *) &u22_stack = u22 *(float64 *) &u32_stack = u32 out4 = *(float64 *) &q3y0 out4 *= u42 u1u3 = *(float64 *) &q3y1 u1u3 *= u32 out4 += u1u3 u2u2 = *(float64 *) &q3y2 u2u2 *= u2u2 out4 += u2u2 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &q3y1 out0 *= u42 u2u3 = *(float64 *) &q3y2 u2u3 *= *(float64 *) &u32_stack out0 += u2u3 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0u0 = *(float64 *) &q3y0 u0u0 *= u0u0 out0 += u0u0 out1 = *(float64 *) &q3y2 out1 *= u42 u3u3 = *(float64 *) &q3y3 u3u3 *= u3u3 out1 += u3u3 out1 *= *(float64 *) &gaudry_pm_scale u0u1 = *(float64 *) &q3y0 u0u1 += u0u1 u0u1 *= *(float64 *) &q3y1 out1 += u0u1 out2 = *(float64 *) &q3y3 out2 *= u42 out2 *= *(float64 *) &gaudry_pm_scale u0u2 = *(float64 *) &q3y0 u0u2 *= *(float64 *) &u22_stack out2 += u0u2 u1u1 = *(float64 *) &q3y1 u1u1 *= u1u1 out2 += u1u1 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 *(float64 *) &right0y0 = out0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out1 -= carry1 *(float64 *) &right0y1 = out1 out2 -= carry2 *(float64 *) &right0y2 = out2 out3 -= carry3 *(float64 *) &right0y3 = out3 out4 += carry3 *(float64 *) &right0y4 = out4 # right0z = square(q3z) out3 = *(float64 *) &q3z4 out3 *= out3 u22 = *(float64 *) &q3z2 u22 += u22 u32 = *(float64 *) &q3z3 u32 += u32 u42 = *(float64 *) &q3z4 u42 += u42 out3 *= *(float64 *) &gaudry_pm_scale u1u2 = *(float64 *) &q3z1 u1u2 *= u22 out3 += u1u2 u0u3 = *(float64 *) &q3z0 u0u3 *= u32 out3 += u0u3 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 *(float64 *) &u22_stack = u22 *(float64 *) &u32_stack = u32 out4 = *(float64 *) &q3z0 out4 *= u42 u1u3 = *(float64 *) &q3z1 u1u3 *= u32 out4 += u1u3 u2u2 = *(float64 *) &q3z2 u2u2 *= u2u2 out4 += u2u2 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &q3z1 out0 *= u42 u2u3 = *(float64 *) &q3z2 u2u3 *= *(float64 *) &u32_stack out0 += u2u3 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0u0 = *(float64 *) &q3z0 u0u0 *= u0u0 out0 += u0u0 out1 = *(float64 *) &q3z2 out1 *= u42 u3u3 = *(float64 *) &q3z3 u3u3 *= u3u3 out1 += u3u3 out1 *= *(float64 *) &gaudry_pm_scale u0u1 = *(float64 *) &q3z0 u0u1 += u0u1 u0u1 *= *(float64 *) &q3z1 out1 += u0u1 out2 = *(float64 *) &q3z3 out2 *= u42 out2 *= *(float64 *) &gaudry_pm_scale u0u2 = *(float64 *) &q3z0 u0u2 *= *(float64 *) &u22_stack out2 += u0u2 u1u1 = *(float64 *) &q3z1 u1u1 *= u1u1 out2 += u1u1 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 *(float64 *) &right0z0 = out0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out1 -= carry1 *(float64 *) &right0z1 = out1 out2 -= carry2 *(float64 *) &right0z2 = out2 out3 -= carry3 *(float64 *) &right0z3 = out3 out4 += carry3 *(float64 *) &right0z4 = out4 # right0t = square(q3t) out3 = *(float64 *) &q3t4 out3 *= out3 u22 = *(float64 *) &q3t2 u22 += u22 u32 = *(float64 *) &q3t3 u32 += u32 u42 = *(float64 *) &q3t4 u42 += u42 out3 *= *(float64 *) &gaudry_pm_scale u1u2 = *(float64 *) &q3t1 u1u2 *= u22 out3 += u1u2 u0u3 = *(float64 *) &q3t0 u0u3 *= u32 out3 += u0u3 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 *(float64 *) &u22_stack = u22 *(float64 *) &u32_stack = u32 out4 = *(float64 *) &q3t0 out4 *= u42 u1u3 = *(float64 *) &q3t1 u1u3 *= u32 out4 += u1u3 u2u2 = *(float64 *) &q3t2 u2u2 *= u2u2 out4 += u2u2 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &q3t1 out0 *= u42 u2u3 = *(float64 *) &q3t2 u2u3 *= *(float64 *) &u32_stack out0 += u2u3 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0u0 = *(float64 *) &q3t0 u0u0 *= u0u0 out0 += u0u0 out1 = *(float64 *) &q3t2 out1 *= u42 u3u3 = *(float64 *) &q3t3 u3u3 *= u3u3 out1 += u3u3 out1 *= *(float64 *) &gaudry_pm_scale u0u1 = *(float64 *) &q3t0 u0u1 += u0u1 u0u1 *= *(float64 *) &q3t1 out1 += u0u1 out2 = *(float64 *) &q3t3 out2 *= u42 out2 *= *(float64 *) &gaudry_pm_scale u0u2 = *(float64 *) &q3t0 u0u2 *= *(float64 *) &u22_stack out2 += u0u2 u1u1 = *(float64 *) &q3t1 u1u1 *= u1u1 out2 += u1u1 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 *(float64 *) &right0t0 = out0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out1 -= carry1 *(float64 *) &right0t1 = out1 out2 -= carry2 *(float64 *) &right0t2 = out2 out3 -= carry3 *(float64 *) &right0t3 = out3 out4 += carry3 *(float64 *) &right0t4 = out4 # left1, right1 = select(left0, right0, bit) use1 = *(float64 *) &exponentbitshifted use1 -= *(float64 *) &gaudry_pm_in0offset x0 = *(float64 *) &left0x0 y0 = *(float64 *) &right0x0 y0 -= x0; negate y0 *= use1 x0 -= y0 y0 += *(float64 *) &right0x0 *(float64 *) &left1x0 = x0 *(float64 *) &right1x0 = y0 x1 = *(float64 *) &left0x1 y1 = *(float64 *) &right0x1 y1 -= x1; negate y1 *= use1 x1 -= y1 y1 += *(float64 *) &right0x1 *(float64 *) &left1x1 = x1 *(float64 *) &right1x1 = y1 x2 = *(float64 *) &left0x2 y2 = *(float64 *) &right0x2 y2 -= x2; negate y2 *= use1 x2 -= y2 y2 += *(float64 *) &right0x2 *(float64 *) &left1x2 = x2 *(float64 *) &right1x2 = y2 x3 = *(float64 *) &left0x3 y3 = *(float64 *) &right0x3 y3 -= x3; negate y3 *= use1 x3 -= y3 y3 += *(float64 *) &right0x3 *(float64 *) &left1x3 = x3 *(float64 *) &right1x3 = y3 x4 = *(float64 *) &left0x4 y4 = *(float64 *) &right0x4 y4 -= x4; negate y4 *= use1 x4 -= y4 y4 += *(float64 *) &right0x4 *(float64 *) &left1x4 = x4 *(float64 *) &right1x4 = y4 x0 = *(float64 *) &left0y0 y0 = *(float64 *) &right0y0 y0 -= x0; negate y0 *= use1 x0 -= y0 y0 += *(float64 *) &right0y0 *(float64 *) &left1y0 = x0 *(float64 *) &right1y0 = y0 x1 = *(float64 *) &left0y1 y1 = *(float64 *) &right0y1 y1 -= x1; negate y1 *= use1 x1 -= y1 y1 += *(float64 *) &right0y1 *(float64 *) &left1y1 = x1 *(float64 *) &right1y1 = y1 x2 = *(float64 *) &left0y2 y2 = *(float64 *) &right0y2 y2 -= x2; negate y2 *= use1 x2 -= y2 y2 += *(float64 *) &right0y2 *(float64 *) &left1y2 = x2 *(float64 *) &right1y2 = y2 x3 = *(float64 *) &left0y3 y3 = *(float64 *) &right0y3 y3 -= x3; negate y3 *= use1 x3 -= y3 y3 += *(float64 *) &right0y3 *(float64 *) &left1y3 = x3 *(float64 *) &right1y3 = y3 x4 = *(float64 *) &left0y4 y4 = *(float64 *) &right0y4 y4 -= x4; negate y4 *= use1 x4 -= y4 y4 += *(float64 *) &right0y4 *(float64 *) &left1y4 = x4 *(float64 *) &right1y4 = y4 x0 = *(float64 *) &left0z0 y0 = *(float64 *) &right0z0 y0 -= x0; negate y0 *= use1 x0 -= y0 y0 += *(float64 *) &right0z0 *(float64 *) &left1z0 = x0 *(float64 *) &right1z0 = y0 x1 = *(float64 *) &left0z1 y1 = *(float64 *) &right0z1 y1 -= x1; negate y1 *= use1 x1 -= y1 y1 += *(float64 *) &right0z1 *(float64 *) &left1z1 = x1 *(float64 *) &right1z1 = y1 x2 = *(float64 *) &left0z2 y2 = *(float64 *) &right0z2 y2 -= x2; negate y2 *= use1 x2 -= y2 y2 += *(float64 *) &right0z2 *(float64 *) &left1z2 = x2 *(float64 *) &right1z2 = y2 x3 = *(float64 *) &left0z3 y3 = *(float64 *) &right0z3 y3 -= x3; negate y3 *= use1 x3 -= y3 y3 += *(float64 *) &right0z3 *(float64 *) &left1z3 = x3 *(float64 *) &right1z3 = y3 x4 = *(float64 *) &left0z4 y4 = *(float64 *) &right0z4 y4 -= x4; negate y4 *= use1 x4 -= y4 y4 += *(float64 *) &right0z4 *(float64 *) &left1z4 = x4 *(float64 *) &right1z4 = y4 x0 = *(float64 *) &left0t0 y0 = *(float64 *) &right0t0 y0 -= x0; negate y0 *= use1 x0 -= y0 y0 += *(float64 *) &right0t0 *(float64 *) &left1t0 = x0 *(float64 *) &right1t0 = y0 x1 = *(float64 *) &left0t1 y1 = *(float64 *) &right0t1 y1 -= x1; negate y1 *= use1 x1 -= y1 y1 += *(float64 *) &right0t1 *(float64 *) &left1t1 = x1 *(float64 *) &right1t1 = y1 x2 = *(float64 *) &left0t2 y2 = *(float64 *) &right0t2 y2 -= x2; negate y2 *= use1 x2 -= y2 y2 += *(float64 *) &right0t2 *(float64 *) &left1t2 = x2 *(float64 *) &right1t2 = y2 x3 = *(float64 *) &left0t3 y3 = *(float64 *) &right0t3 y3 -= x3; negate y3 *= use1 x3 -= y3 y3 += *(float64 *) &right0t3 *(float64 *) &left1t3 = x3 *(float64 *) &right1t3 = y3 x4 = *(float64 *) &left0t4 y4 = *(float64 *) &right0t4 y4 -= x4; negate y4 *= use1 x4 -= y4 y4 += *(float64 *) &right0t4 *(float64 *) &left1t4 = x4 *(float64 *) &right1t4 = y4 # left2 = hadamard(left1) f0 = *(float64 *) &left1x0 d0 = *(float64 *) &left1y0 g0 = *(float64 *) &left1z0 e0 = *(float64 *) &left1t0 d0 += f0 e0 += g0 f0 -= *(float64 *) &left1y0 g0 -= *(float64 *) &left1t0 *(float64 *) &leftpmx0 = d0 *(float64 *) &leftpmy0 = e0 *(float64 *) &leftpmz0 = f0 *(float64 *) &leftpmt0 = g0 f1 = *(float64 *) &left1x1 d1 = *(float64 *) &left1y1 g1 = *(float64 *) &left1z1 e1 = *(float64 *) &left1t1 d1 += f1 e1 += g1 f1 -= *(float64 *) &left1y1 g1 -= *(float64 *) &left1t1 *(float64 *) &leftpmx1 = d1 *(float64 *) &leftpmy1 = e1 *(float64 *) &leftpmz1 = f1 *(float64 *) &leftpmt1 = g1 f2 = *(float64 *) &left1x2 d2 = *(float64 *) &left1y2 g2 = *(float64 *) &left1z2 e2 = *(float64 *) &left1t2 d2 += f2 e2 += g2 f2 -= *(float64 *) &left1y2 g2 -= *(float64 *) &left1t2 *(float64 *) &leftpmx2 = d2 *(float64 *) &leftpmy2 = e2 *(float64 *) &leftpmz2 = f2 *(float64 *) &leftpmt2 = g2 f3 = *(float64 *) &left1x3 d3 = *(float64 *) &left1y3 g3 = *(float64 *) &left1z3 e3 = *(float64 *) &left1t3 d3 += f3 e3 += g3 f3 -= *(float64 *) &left1y3 g3 -= *(float64 *) &left1t3 *(float64 *) &leftpmx3 = d3 *(float64 *) &leftpmy3 = e3 *(float64 *) &leftpmz3 = f3 *(float64 *) &leftpmt3 = g3 f4 = *(float64 *) &left1x4 d4 = *(float64 *) &left1y4 g4 = *(float64 *) &left1z4 e4 = *(float64 *) &left1t4 d4 += f4 e4 += g4 f4 -= *(float64 *) &left1y4 g4 -= *(float64 *) &left1t4 *(float64 *) &leftpmx4 = d4 *(float64 *) &leftpmy4 = e4 *(float64 *) &leftpmz4 = f4 *(float64 *) &leftpmt4 = g4 j0 = *(float64 *) &leftpmx0 h0 = *(float64 *) &leftpmy0 k0 = *(float64 *) &leftpmz0 i0 = *(float64 *) &leftpmt0 h0 += j0 i0 += k0 j0 -= *(float64 *) &leftpmy0 k0 -= *(float64 *) &leftpmt0 *(float64 *) &left2x0 = h0 *(float64 *) &left2y0 = j0 *(float64 *) &left2z0 = i0 *(float64 *) &left2t0 = k0 j1 = *(float64 *) &leftpmx1 h1 = *(float64 *) &leftpmy1 k1 = *(float64 *) &leftpmz1 i1 = *(float64 *) &leftpmt1 h1 += j1 i1 += k1 j1 -= *(float64 *) &leftpmy1 k1 -= *(float64 *) &leftpmt1 *(float64 *) &left2x1 = h1 *(float64 *) &left2y1 = j1 *(float64 *) &left2z1 = i1 *(float64 *) &left2t1 = k1 j2 = *(float64 *) &leftpmx2 h2 = *(float64 *) &leftpmy2 k2 = *(float64 *) &leftpmz2 i2 = *(float64 *) &leftpmt2 h2 += j2 i2 += k2 j2 -= *(float64 *) &leftpmy2 k2 -= *(float64 *) &leftpmt2 *(float64 *) &left2x2 = h2 *(float64 *) &left2y2 = j2 *(float64 *) &left2z2 = i2 *(float64 *) &left2t2 = k2 j3 = *(float64 *) &leftpmx3 h3 = *(float64 *) &leftpmy3 k3 = *(float64 *) &leftpmz3 i3 = *(float64 *) &leftpmt3 h3 += j3 i3 += k3 j3 -= *(float64 *) &leftpmy3 k3 -= *(float64 *) &leftpmt3 *(float64 *) &left2x3 = h3 *(float64 *) &left2y3 = j3 *(float64 *) &left2z3 = i3 *(float64 *) &left2t3 = k3 j4 = *(float64 *) &leftpmx4 h4 = *(float64 *) &leftpmy4 k4 = *(float64 *) &leftpmz4 i4 = *(float64 *) &leftpmt4 h4 += j4 i4 += k4 j4 -= *(float64 *) &leftpmy4 k4 -= *(float64 *) &leftpmt4 *(float64 *) &left2x4 = h4 *(float64 *) &left2y4 = j4 *(float64 *) &left2z4 = i4 *(float64 *) &left2t4 = k4 # right2 = hadamard(right1) f0 = *(float64 *) &right1x0 d0 = *(float64 *) &right1y0 g0 = *(float64 *) &right1z0 e0 = *(float64 *) &right1t0 d0 += f0 e0 += g0 f0 -= *(float64 *) &right1y0 g0 -= *(float64 *) &right1t0 *(float64 *) &rightpmx0 = d0 *(float64 *) &rightpmy0 = e0 *(float64 *) &rightpmz0 = f0 *(float64 *) &rightpmt0 = g0 f1 = *(float64 *) &right1x1 d1 = *(float64 *) &right1y1 g1 = *(float64 *) &right1z1 e1 = *(float64 *) &right1t1 d1 += f1 e1 += g1 f1 -= *(float64 *) &right1y1 g1 -= *(float64 *) &right1t1 *(float64 *) &rightpmx1 = d1 *(float64 *) &rightpmy1 = e1 *(float64 *) &rightpmz1 = f1 *(float64 *) &rightpmt1 = g1 f2 = *(float64 *) &right1x2 d2 = *(float64 *) &right1y2 g2 = *(float64 *) &right1z2 e2 = *(float64 *) &right1t2 d2 += f2 e2 += g2 f2 -= *(float64 *) &right1y2 g2 -= *(float64 *) &right1t2 *(float64 *) &rightpmx2 = d2 *(float64 *) &rightpmy2 = e2 *(float64 *) &rightpmz2 = f2 *(float64 *) &rightpmt2 = g2 f3 = *(float64 *) &right1x3 d3 = *(float64 *) &right1y3 g3 = *(float64 *) &right1z3 e3 = *(float64 *) &right1t3 d3 += f3 e3 += g3 f3 -= *(float64 *) &right1y3 g3 -= *(float64 *) &right1t3 *(float64 *) &rightpmx3 = d3 *(float64 *) &rightpmy3 = e3 *(float64 *) &rightpmz3 = f3 *(float64 *) &rightpmt3 = g3 f4 = *(float64 *) &right1x4 d4 = *(float64 *) &right1y4 g4 = *(float64 *) &right1z4 e4 = *(float64 *) &right1t4 d4 += f4 e4 += g4 f4 -= *(float64 *) &right1y4 g4 -= *(float64 *) &right1t4 *(float64 *) &rightpmx4 = d4 *(float64 *) &rightpmy4 = e4 *(float64 *) &rightpmz4 = f4 *(float64 *) &rightpmt4 = g4 j0 = *(float64 *) &rightpmx0 h0 = *(float64 *) &rightpmy0 k0 = *(float64 *) &rightpmz0 i0 = *(float64 *) &rightpmt0 h0 += j0 i0 += k0 j0 -= *(float64 *) &rightpmy0 k0 -= *(float64 *) &rightpmt0 *(float64 *) &right2x0 = h0 *(float64 *) &right2y0 = j0 *(float64 *) &right2z0 = i0 *(float64 *) &right2t0 = k0 j1 = *(float64 *) &rightpmx1 h1 = *(float64 *) &rightpmy1 k1 = *(float64 *) &rightpmz1 i1 = *(float64 *) &rightpmt1 h1 += j1 i1 += k1 j1 -= *(float64 *) &rightpmy1 k1 -= *(float64 *) &rightpmt1 *(float64 *) &right2x1 = h1 *(float64 *) &right2y1 = j1 *(float64 *) &right2z1 = i1 *(float64 *) &right2t1 = k1 j2 = *(float64 *) &rightpmx2 h2 = *(float64 *) &rightpmy2 k2 = *(float64 *) &rightpmz2 i2 = *(float64 *) &rightpmt2 h2 += j2 i2 += k2 j2 -= *(float64 *) &rightpmy2 k2 -= *(float64 *) &rightpmt2 *(float64 *) &right2x2 = h2 *(float64 *) &right2y2 = j2 *(float64 *) &right2z2 = i2 *(float64 *) &right2t2 = k2 j3 = *(float64 *) &rightpmx3 h3 = *(float64 *) &rightpmy3 k3 = *(float64 *) &rightpmz3 i3 = *(float64 *) &rightpmt3 h3 += j3 i3 += k3 j3 -= *(float64 *) &rightpmy3 k3 -= *(float64 *) &rightpmt3 *(float64 *) &right2x3 = h3 *(float64 *) &right2y3 = j3 *(float64 *) &right2z3 = i3 *(float64 *) &right2t3 = k3 j4 = *(float64 *) &rightpmx4 h4 = *(float64 *) &rightpmy4 k4 = *(float64 *) &rightpmz4 i4 = *(float64 *) &rightpmt4 h4 += j4 i4 += k4 j4 -= *(float64 *) &rightpmy4 k4 -= *(float64 *) &rightpmt4 *(float64 *) &right2x4 = h4 *(float64 *) &right2y4 = j4 *(float64 *) &right2z4 = i4 *(float64 *) &right2t4 = k4 # left3y = mult(left2y,AAoverBB) out3 = *(float64 *) &left2y4 out3 *= *(float64 *) &AAoverBB4 out3 *= *(float64 *) &gaudry_pm_scale u0v3 = *(float64 *) &left2y0 u0v3 *= *(float64 *) &AAoverBB3 out3 += u0v3 u1v2 = *(float64 *) &left2y1 u1v2 *= *(float64 *) &AAoverBB2 out3 += u1v2 u2v1 = *(float64 *) &left2y2 u2v1 *= *(float64 *) &AAoverBB1 out3 += u2v1 u3v0 = *(float64 *) &left2y3 u3v0 *= *(float64 *) &AAoverBB0 out3 += u3v0 out4 = *(float64 *) &left2y0 out4 *= *(float64 *) &AAoverBB4 u1v3 = *(float64 *) &left2y1 u1v3 *= *(float64 *) &AAoverBB3 out4 += u1v3 u2v2 = *(float64 *) &left2y2 u2v2 *= *(float64 *) &AAoverBB2 out4 += u2v2 u3v1 = *(float64 *) &left2y3 u3v1 *= *(float64 *) &AAoverBB1 out4 += u3v1 u4v0 = *(float64 *) &left2y4 u4v0 *= *(float64 *) &AAoverBB0 out4 += u4v0 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &left2y1 out0 *= *(float64 *) &AAoverBB4 u2v3 = *(float64 *) &left2y2 u2v3 *= *(float64 *) &AAoverBB3 out0 += u2v3 u3v2 = *(float64 *) &left2y3 u3v2 *= *(float64 *) &AAoverBB2 out0 += u3v2 u4v1 = *(float64 *) &left2y4 u4v1 *= *(float64 *) &AAoverBB1 out0 += u4v1 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0v0 = *(float64 *) &left2y0 u0v0 *= *(float64 *) &AAoverBB0 out0 += u0v0 out1 = *(float64 *) &left2y2 out1 *= *(float64 *) &AAoverBB4 u3v3 = *(float64 *) &left2y3 u3v3 *= *(float64 *) &AAoverBB3 out1 += u3v3 u4v2 = *(float64 *) &left2y4 u4v2 *= *(float64 *) &AAoverBB2 out1 += u4v2 out1 *= *(float64 *) &gaudry_pm_scale u0v1 = *(float64 *) &left2y0 u0v1 *= *(float64 *) &AAoverBB1 out1 += u0v1 u1v0 = *(float64 *) &left2y1 u1v0 *= *(float64 *) &AAoverBB0 out1 += u1v0 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 out2 = *(float64 *) &left2y3 out2 *= *(float64 *) &AAoverBB4 u4v3 = *(float64 *) &left2y4 u4v3 *= *(float64 *) &AAoverBB3 out2 += u4v3 out2 *= *(float64 *) &gaudry_pm_scale u0v2 = *(float64 *) &left2y0 u0v2 *= *(float64 *) &AAoverBB2 out2 += u0v2 u1v1 = *(float64 *) &left2y1 u1v1 *= *(float64 *) &AAoverBB1 out2 += u1v1 u2v0 = *(float64 *) &left2y2 u2v0 *= *(float64 *) &AAoverBB0 out2 += u2v0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 out1 -= carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 out2 -= carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 *(float64 *) &left3y0 = out0 *(float64 *) &left3y1 = out1 *(float64 *) &left3y2 = out2 *(float64 *) &left3y3 = out3 *(float64 *) &left3y4 = out4 # left3z = mult(left2z,AAoverCC) out3 = *(float64 *) &left2z4 out3 *= *(float64 *) &AAoverCC4 out3 *= *(float64 *) &gaudry_pm_scale u0v3 = *(float64 *) &left2z0 u0v3 *= *(float64 *) &AAoverCC3 out3 += u0v3 u1v2 = *(float64 *) &left2z1 u1v2 *= *(float64 *) &AAoverCC2 out3 += u1v2 u2v1 = *(float64 *) &left2z2 u2v1 *= *(float64 *) &AAoverCC1 out3 += u2v1 u3v0 = *(float64 *) &left2z3 u3v0 *= *(float64 *) &AAoverCC0 out3 += u3v0 out4 = *(float64 *) &left2z0 out4 *= *(float64 *) &AAoverCC4 u1v3 = *(float64 *) &left2z1 u1v3 *= *(float64 *) &AAoverCC3 out4 += u1v3 u2v2 = *(float64 *) &left2z2 u2v2 *= *(float64 *) &AAoverCC2 out4 += u2v2 u3v1 = *(float64 *) &left2z3 u3v1 *= *(float64 *) &AAoverCC1 out4 += u3v1 u4v0 = *(float64 *) &left2z4 u4v0 *= *(float64 *) &AAoverCC0 out4 += u4v0 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &left2z1 out0 *= *(float64 *) &AAoverCC4 u2v3 = *(float64 *) &left2z2 u2v3 *= *(float64 *) &AAoverCC3 out0 += u2v3 u3v2 = *(float64 *) &left2z3 u3v2 *= *(float64 *) &AAoverCC2 out0 += u3v2 u4v1 = *(float64 *) &left2z4 u4v1 *= *(float64 *) &AAoverCC1 out0 += u4v1 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0v0 = *(float64 *) &left2z0 u0v0 *= *(float64 *) &AAoverCC0 out0 += u0v0 out1 = *(float64 *) &left2z2 out1 *= *(float64 *) &AAoverCC4 u3v3 = *(float64 *) &left2z3 u3v3 *= *(float64 *) &AAoverCC3 out1 += u3v3 u4v2 = *(float64 *) &left2z4 u4v2 *= *(float64 *) &AAoverCC2 out1 += u4v2 out1 *= *(float64 *) &gaudry_pm_scale u0v1 = *(float64 *) &left2z0 u0v1 *= *(float64 *) &AAoverCC1 out1 += u0v1 u1v0 = *(float64 *) &left2z1 u1v0 *= *(float64 *) &AAoverCC0 out1 += u1v0 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 out2 = *(float64 *) &left2z3 out2 *= *(float64 *) &AAoverCC4 u4v3 = *(float64 *) &left2z4 u4v3 *= *(float64 *) &AAoverCC3 out2 += u4v3 out2 *= *(float64 *) &gaudry_pm_scale u0v2 = *(float64 *) &left2z0 u0v2 *= *(float64 *) &AAoverCC2 out2 += u0v2 u1v1 = *(float64 *) &left2z1 u1v1 *= *(float64 *) &AAoverCC1 out2 += u1v1 u2v0 = *(float64 *) &left2z2 u2v0 *= *(float64 *) &AAoverCC0 out2 += u2v0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 out1 -= carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 out2 -= carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 *(float64 *) &left3z0 = out0 *(float64 *) &left3z1 = out1 *(float64 *) &left3z2 = out2 *(float64 *) &left3z3 = out3 *(float64 *) &left3z4 = out4 # left3t = mult(left2t,AAoverDD) out3 = *(float64 *) &left2t4 out3 *= *(float64 *) &AAoverDD4 out3 *= *(float64 *) &gaudry_pm_scale u0v3 = *(float64 *) &left2t0 u0v3 *= *(float64 *) &AAoverDD3 out3 += u0v3 u1v2 = *(float64 *) &left2t1 u1v2 *= *(float64 *) &AAoverDD2 out3 += u1v2 u2v1 = *(float64 *) &left2t2 u2v1 *= *(float64 *) &AAoverDD1 out3 += u2v1 u3v0 = *(float64 *) &left2t3 u3v0 *= *(float64 *) &AAoverDD0 out3 += u3v0 out4 = *(float64 *) &left2t0 out4 *= *(float64 *) &AAoverDD4 u1v3 = *(float64 *) &left2t1 u1v3 *= *(float64 *) &AAoverDD3 out4 += u1v3 u2v2 = *(float64 *) &left2t2 u2v2 *= *(float64 *) &AAoverDD2 out4 += u2v2 u3v1 = *(float64 *) &left2t3 u3v1 *= *(float64 *) &AAoverDD1 out4 += u3v1 u4v0 = *(float64 *) &left2t4 u4v0 *= *(float64 *) &AAoverDD0 out4 += u4v0 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &left2t1 out0 *= *(float64 *) &AAoverDD4 u2v3 = *(float64 *) &left2t2 u2v3 *= *(float64 *) &AAoverDD3 out0 += u2v3 u3v2 = *(float64 *) &left2t3 u3v2 *= *(float64 *) &AAoverDD2 out0 += u3v2 u4v1 = *(float64 *) &left2t4 u4v1 *= *(float64 *) &AAoverDD1 out0 += u4v1 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0v0 = *(float64 *) &left2t0 u0v0 *= *(float64 *) &AAoverDD0 out0 += u0v0 out1 = *(float64 *) &left2t2 out1 *= *(float64 *) &AAoverDD4 u3v3 = *(float64 *) &left2t3 u3v3 *= *(float64 *) &AAoverDD3 out1 += u3v3 u4v2 = *(float64 *) &left2t4 u4v2 *= *(float64 *) &AAoverDD2 out1 += u4v2 out1 *= *(float64 *) &gaudry_pm_scale u0v1 = *(float64 *) &left2t0 u0v1 *= *(float64 *) &AAoverDD1 out1 += u0v1 u1v0 = *(float64 *) &left2t1 u1v0 *= *(float64 *) &AAoverDD0 out1 += u1v0 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 out2 = *(float64 *) &left2t3 out2 *= *(float64 *) &AAoverDD4 u4v3 = *(float64 *) &left2t4 u4v3 *= *(float64 *) &AAoverDD3 out2 += u4v3 out2 *= *(float64 *) &gaudry_pm_scale u0v2 = *(float64 *) &left2t0 u0v2 *= *(float64 *) &AAoverDD2 out2 += u0v2 u1v1 = *(float64 *) &left2t1 u1v1 *= *(float64 *) &AAoverDD1 out2 += u1v1 u2v0 = *(float64 *) &left2t2 u2v0 *= *(float64 *) &AAoverDD0 out2 += u2v0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 out1 -= carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 out2 -= carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 *(float64 *) &left3t0 = out0 *(float64 *) &left3t1 = out1 *(float64 *) &left3t2 = out2 *(float64 *) &left3t3 = out3 *(float64 *) &left3t4 = out4 # left4x = square(left2x) out3 = *(float64 *) &left2x4 out3 *= out3 u22 = *(float64 *) &left2x2 u22 += u22 u32 = *(float64 *) &left2x3 u32 += u32 u42 = *(float64 *) &left2x4 u42 += u42 out3 *= *(float64 *) &gaudry_pm_scale u1u2 = *(float64 *) &left2x1 u1u2 *= u22 out3 += u1u2 u0u3 = *(float64 *) &left2x0 u0u3 *= u32 out3 += u0u3 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 *(float64 *) &u22_stack = u22 *(float64 *) &u32_stack = u32 out4 = *(float64 *) &left2x0 out4 *= u42 u1u3 = *(float64 *) &left2x1 u1u3 *= u32 out4 += u1u3 u2u2 = *(float64 *) &left2x2 u2u2 *= u2u2 out4 += u2u2 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &left2x1 out0 *= u42 u2u3 = *(float64 *) &left2x2 u2u3 *= *(float64 *) &u32_stack out0 += u2u3 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0u0 = *(float64 *) &left2x0 u0u0 *= u0u0 out0 += u0u0 out1 = *(float64 *) &left2x2 out1 *= u42 u3u3 = *(float64 *) &left2x3 u3u3 *= u3u3 out1 += u3u3 out1 *= *(float64 *) &gaudry_pm_scale u0u1 = *(float64 *) &left2x0 u0u1 += u0u1 u0u1 *= *(float64 *) &left2x1 out1 += u0u1 out2 = *(float64 *) &left2x3 out2 *= u42 out2 *= *(float64 *) &gaudry_pm_scale u0u2 = *(float64 *) &left2x0 u0u2 *= *(float64 *) &u22_stack out2 += u0u2 u1u1 = *(float64 *) &left2x1 u1u1 *= u1u1 out2 += u1u1 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 *(float64 *) &left4x0 = out0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out1 -= carry1 *(float64 *) &left4x1 = out1 out2 -= carry2 *(float64 *) &left4x2 = out2 out3 -= carry3 *(float64 *) &left4x3 = out3 out4 += carry3 *(float64 *) &left4x4 = out4 # left4y = mult(left2y,left3y) out3 = *(float64 *) &left2y4 out3 *= *(float64 *) &left3y4 out3 *= *(float64 *) &gaudry_pm_scale u0v3 = *(float64 *) &left2y0 u0v3 *= *(float64 *) &left3y3 out3 += u0v3 u1v2 = *(float64 *) &left2y1 u1v2 *= *(float64 *) &left3y2 out3 += u1v2 u2v1 = *(float64 *) &left2y2 u2v1 *= *(float64 *) &left3y1 out3 += u2v1 u3v0 = *(float64 *) &left2y3 u3v0 *= *(float64 *) &left3y0 out3 += u3v0 out4 = *(float64 *) &left2y0 out4 *= *(float64 *) &left3y4 u1v3 = *(float64 *) &left2y1 u1v3 *= *(float64 *) &left3y3 out4 += u1v3 u2v2 = *(float64 *) &left2y2 u2v2 *= *(float64 *) &left3y2 out4 += u2v2 u3v1 = *(float64 *) &left2y3 u3v1 *= *(float64 *) &left3y1 out4 += u3v1 u4v0 = *(float64 *) &left2y4 u4v0 *= *(float64 *) &left3y0 out4 += u4v0 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &left2y1 out0 *= *(float64 *) &left3y4 u2v3 = *(float64 *) &left2y2 u2v3 *= *(float64 *) &left3y3 out0 += u2v3 u3v2 = *(float64 *) &left2y3 u3v2 *= *(float64 *) &left3y2 out0 += u3v2 u4v1 = *(float64 *) &left2y4 u4v1 *= *(float64 *) &left3y1 out0 += u4v1 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0v0 = *(float64 *) &left2y0 u0v0 *= *(float64 *) &left3y0 out0 += u0v0 out1 = *(float64 *) &left2y2 out1 *= *(float64 *) &left3y4 u3v3 = *(float64 *) &left2y3 u3v3 *= *(float64 *) &left3y3 out1 += u3v3 u4v2 = *(float64 *) &left2y4 u4v2 *= *(float64 *) &left3y2 out1 += u4v2 out1 *= *(float64 *) &gaudry_pm_scale u0v1 = *(float64 *) &left2y0 u0v1 *= *(float64 *) &left3y1 out1 += u0v1 u1v0 = *(float64 *) &left2y1 u1v0 *= *(float64 *) &left3y0 out1 += u1v0 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 out2 = *(float64 *) &left2y3 out2 *= *(float64 *) &left3y4 u4v3 = *(float64 *) &left2y4 u4v3 *= *(float64 *) &left3y3 out2 += u4v3 out2 *= *(float64 *) &gaudry_pm_scale u0v2 = *(float64 *) &left2y0 u0v2 *= *(float64 *) &left3y2 out2 += u0v2 u1v1 = *(float64 *) &left2y1 u1v1 *= *(float64 *) &left3y1 out2 += u1v1 u2v0 = *(float64 *) &left2y2 u2v0 *= *(float64 *) &left3y0 out2 += u2v0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 out1 -= carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 out2 -= carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 *(float64 *) &left4y0 = out0 *(float64 *) &left4y1 = out1 *(float64 *) &left4y2 = out2 *(float64 *) &left4y3 = out3 *(float64 *) &left4y4 = out4 # left4z = mult(left2z,left3z) out3 = *(float64 *) &left2z4 out3 *= *(float64 *) &left3z4 out3 *= *(float64 *) &gaudry_pm_scale u0v3 = *(float64 *) &left2z0 u0v3 *= *(float64 *) &left3z3 out3 += u0v3 u1v2 = *(float64 *) &left2z1 u1v2 *= *(float64 *) &left3z2 out3 += u1v2 u2v1 = *(float64 *) &left2z2 u2v1 *= *(float64 *) &left3z1 out3 += u2v1 u3v0 = *(float64 *) &left2z3 u3v0 *= *(float64 *) &left3z0 out3 += u3v0 out4 = *(float64 *) &left2z0 out4 *= *(float64 *) &left3z4 u1v3 = *(float64 *) &left2z1 u1v3 *= *(float64 *) &left3z3 out4 += u1v3 u2v2 = *(float64 *) &left2z2 u2v2 *= *(float64 *) &left3z2 out4 += u2v2 u3v1 = *(float64 *) &left2z3 u3v1 *= *(float64 *) &left3z1 out4 += u3v1 u4v0 = *(float64 *) &left2z4 u4v0 *= *(float64 *) &left3z0 out4 += u4v0 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &left2z1 out0 *= *(float64 *) &left3z4 u2v3 = *(float64 *) &left2z2 u2v3 *= *(float64 *) &left3z3 out0 += u2v3 u3v2 = *(float64 *) &left2z3 u3v2 *= *(float64 *) &left3z2 out0 += u3v2 u4v1 = *(float64 *) &left2z4 u4v1 *= *(float64 *) &left3z1 out0 += u4v1 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0v0 = *(float64 *) &left2z0 u0v0 *= *(float64 *) &left3z0 out0 += u0v0 out1 = *(float64 *) &left2z2 out1 *= *(float64 *) &left3z4 u3v3 = *(float64 *) &left2z3 u3v3 *= *(float64 *) &left3z3 out1 += u3v3 u4v2 = *(float64 *) &left2z4 u4v2 *= *(float64 *) &left3z2 out1 += u4v2 out1 *= *(float64 *) &gaudry_pm_scale u0v1 = *(float64 *) &left2z0 u0v1 *= *(float64 *) &left3z1 out1 += u0v1 u1v0 = *(float64 *) &left2z1 u1v0 *= *(float64 *) &left3z0 out1 += u1v0 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 out2 = *(float64 *) &left2z3 out2 *= *(float64 *) &left3z4 u4v3 = *(float64 *) &left2z4 u4v3 *= *(float64 *) &left3z3 out2 += u4v3 out2 *= *(float64 *) &gaudry_pm_scale u0v2 = *(float64 *) &left2z0 u0v2 *= *(float64 *) &left3z2 out2 += u0v2 u1v1 = *(float64 *) &left2z1 u1v1 *= *(float64 *) &left3z1 out2 += u1v1 u2v0 = *(float64 *) &left2z2 u2v0 *= *(float64 *) &left3z0 out2 += u2v0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 out1 -= carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 out2 -= carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 *(float64 *) &left4z0 = out0 *(float64 *) &left4z1 = out1 *(float64 *) &left4z2 = out2 *(float64 *) &left4z3 = out3 *(float64 *) &left4z4 = out4 # left4t = mult(left2t,left3t) out3 = *(float64 *) &left2t4 out3 *= *(float64 *) &left3t4 out3 *= *(float64 *) &gaudry_pm_scale u0v3 = *(float64 *) &left2t0 u0v3 *= *(float64 *) &left3t3 out3 += u0v3 u1v2 = *(float64 *) &left2t1 u1v2 *= *(float64 *) &left3t2 out3 += u1v2 u2v1 = *(float64 *) &left2t2 u2v1 *= *(float64 *) &left3t1 out3 += u2v1 u3v0 = *(float64 *) &left2t3 u3v0 *= *(float64 *) &left3t0 out3 += u3v0 out4 = *(float64 *) &left2t0 out4 *= *(float64 *) &left3t4 u1v3 = *(float64 *) &left2t1 u1v3 *= *(float64 *) &left3t3 out4 += u1v3 u2v2 = *(float64 *) &left2t2 u2v2 *= *(float64 *) &left3t2 out4 += u2v2 u3v1 = *(float64 *) &left2t3 u3v1 *= *(float64 *) &left3t1 out4 += u3v1 u4v0 = *(float64 *) &left2t4 u4v0 *= *(float64 *) &left3t0 out4 += u4v0 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &left2t1 out0 *= *(float64 *) &left3t4 u2v3 = *(float64 *) &left2t2 u2v3 *= *(float64 *) &left3t3 out0 += u2v3 u3v2 = *(float64 *) &left2t3 u3v2 *= *(float64 *) &left3t2 out0 += u3v2 u4v1 = *(float64 *) &left2t4 u4v1 *= *(float64 *) &left3t1 out0 += u4v1 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0v0 = *(float64 *) &left2t0 u0v0 *= *(float64 *) &left3t0 out0 += u0v0 out1 = *(float64 *) &left2t2 out1 *= *(float64 *) &left3t4 u3v3 = *(float64 *) &left2t3 u3v3 *= *(float64 *) &left3t3 out1 += u3v3 u4v2 = *(float64 *) &left2t4 u4v2 *= *(float64 *) &left3t2 out1 += u4v2 out1 *= *(float64 *) &gaudry_pm_scale u0v1 = *(float64 *) &left2t0 u0v1 *= *(float64 *) &left3t1 out1 += u0v1 u1v0 = *(float64 *) &left2t1 u1v0 *= *(float64 *) &left3t0 out1 += u1v0 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 out2 = *(float64 *) &left2t3 out2 *= *(float64 *) &left3t4 u4v3 = *(float64 *) &left2t4 u4v3 *= *(float64 *) &left3t3 out2 += u4v3 out2 *= *(float64 *) &gaudry_pm_scale u0v2 = *(float64 *) &left2t0 u0v2 *= *(float64 *) &left3t2 out2 += u0v2 u1v1 = *(float64 *) &left2t1 u1v1 *= *(float64 *) &left3t1 out2 += u1v1 u2v0 = *(float64 *) &left2t2 u2v0 *= *(float64 *) &left3t0 out2 += u2v0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 out1 -= carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 out2 -= carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 *(float64 *) &left4t0 = out0 *(float64 *) &left4t1 = out1 *(float64 *) &left4t2 = out2 *(float64 *) &left4t3 = out3 *(float64 *) &left4t4 = out4 # right4x = mult(right2x,left2x) out3 = *(float64 *) &right2x4 out3 *= *(float64 *) &left2x4 out3 *= *(float64 *) &gaudry_pm_scale u0v3 = *(float64 *) &right2x0 u0v3 *= *(float64 *) &left2x3 out3 += u0v3 u1v2 = *(float64 *) &right2x1 u1v2 *= *(float64 *) &left2x2 out3 += u1v2 u2v1 = *(float64 *) &right2x2 u2v1 *= *(float64 *) &left2x1 out3 += u2v1 u3v0 = *(float64 *) &right2x3 u3v0 *= *(float64 *) &left2x0 out3 += u3v0 out4 = *(float64 *) &right2x0 out4 *= *(float64 *) &left2x4 u1v3 = *(float64 *) &right2x1 u1v3 *= *(float64 *) &left2x3 out4 += u1v3 u2v2 = *(float64 *) &right2x2 u2v2 *= *(float64 *) &left2x2 out4 += u2v2 u3v1 = *(float64 *) &right2x3 u3v1 *= *(float64 *) &left2x1 out4 += u3v1 u4v0 = *(float64 *) &right2x4 u4v0 *= *(float64 *) &left2x0 out4 += u4v0 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &right2x1 out0 *= *(float64 *) &left2x4 u2v3 = *(float64 *) &right2x2 u2v3 *= *(float64 *) &left2x3 out0 += u2v3 u3v2 = *(float64 *) &right2x3 u3v2 *= *(float64 *) &left2x2 out0 += u3v2 u4v1 = *(float64 *) &right2x4 u4v1 *= *(float64 *) &left2x1 out0 += u4v1 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0v0 = *(float64 *) &right2x0 u0v0 *= *(float64 *) &left2x0 out0 += u0v0 out1 = *(float64 *) &right2x2 out1 *= *(float64 *) &left2x4 u3v3 = *(float64 *) &right2x3 u3v3 *= *(float64 *) &left2x3 out1 += u3v3 u4v2 = *(float64 *) &right2x4 u4v2 *= *(float64 *) &left2x2 out1 += u4v2 out1 *= *(float64 *) &gaudry_pm_scale u0v1 = *(float64 *) &right2x0 u0v1 *= *(float64 *) &left2x1 out1 += u0v1 u1v0 = *(float64 *) &right2x1 u1v0 *= *(float64 *) &left2x0 out1 += u1v0 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 out2 = *(float64 *) &right2x3 out2 *= *(float64 *) &left2x4 u4v3 = *(float64 *) &right2x4 u4v3 *= *(float64 *) &left2x3 out2 += u4v3 out2 *= *(float64 *) &gaudry_pm_scale u0v2 = *(float64 *) &right2x0 u0v2 *= *(float64 *) &left2x2 out2 += u0v2 u1v1 = *(float64 *) &right2x1 u1v1 *= *(float64 *) &left2x1 out2 += u1v1 u2v0 = *(float64 *) &right2x2 u2v0 *= *(float64 *) &left2x0 out2 += u2v0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 out1 -= carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 out2 -= carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 *(float64 *) &right4x0 = out0 *(float64 *) &right4x1 = out1 *(float64 *) &right4x2 = out2 *(float64 *) &right4x3 = out3 *(float64 *) &right4x4 = out4 # right4y = mult(right2y,left3y) out3 = *(float64 *) &right2y4 out3 *= *(float64 *) &left3y4 out3 *= *(float64 *) &gaudry_pm_scale u0v3 = *(float64 *) &right2y0 u0v3 *= *(float64 *) &left3y3 out3 += u0v3 u1v2 = *(float64 *) &right2y1 u1v2 *= *(float64 *) &left3y2 out3 += u1v2 u2v1 = *(float64 *) &right2y2 u2v1 *= *(float64 *) &left3y1 out3 += u2v1 u3v0 = *(float64 *) &right2y3 u3v0 *= *(float64 *) &left3y0 out3 += u3v0 out4 = *(float64 *) &right2y0 out4 *= *(float64 *) &left3y4 u1v3 = *(float64 *) &right2y1 u1v3 *= *(float64 *) &left3y3 out4 += u1v3 u2v2 = *(float64 *) &right2y2 u2v2 *= *(float64 *) &left3y2 out4 += u2v2 u3v1 = *(float64 *) &right2y3 u3v1 *= *(float64 *) &left3y1 out4 += u3v1 u4v0 = *(float64 *) &right2y4 u4v0 *= *(float64 *) &left3y0 out4 += u4v0 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &right2y1 out0 *= *(float64 *) &left3y4 u2v3 = *(float64 *) &right2y2 u2v3 *= *(float64 *) &left3y3 out0 += u2v3 u3v2 = *(float64 *) &right2y3 u3v2 *= *(float64 *) &left3y2 out0 += u3v2 u4v1 = *(float64 *) &right2y4 u4v1 *= *(float64 *) &left3y1 out0 += u4v1 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0v0 = *(float64 *) &right2y0 u0v0 *= *(float64 *) &left3y0 out0 += u0v0 out1 = *(float64 *) &right2y2 out1 *= *(float64 *) &left3y4 u3v3 = *(float64 *) &right2y3 u3v3 *= *(float64 *) &left3y3 out1 += u3v3 u4v2 = *(float64 *) &right2y4 u4v2 *= *(float64 *) &left3y2 out1 += u4v2 out1 *= *(float64 *) &gaudry_pm_scale u0v1 = *(float64 *) &right2y0 u0v1 *= *(float64 *) &left3y1 out1 += u0v1 u1v0 = *(float64 *) &right2y1 u1v0 *= *(float64 *) &left3y0 out1 += u1v0 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 out2 = *(float64 *) &right2y3 out2 *= *(float64 *) &left3y4 u4v3 = *(float64 *) &right2y4 u4v3 *= *(float64 *) &left3y3 out2 += u4v3 out2 *= *(float64 *) &gaudry_pm_scale u0v2 = *(float64 *) &right2y0 u0v2 *= *(float64 *) &left3y2 out2 += u0v2 u1v1 = *(float64 *) &right2y1 u1v1 *= *(float64 *) &left3y1 out2 += u1v1 u2v0 = *(float64 *) &right2y2 u2v0 *= *(float64 *) &left3y0 out2 += u2v0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 out1 -= carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 out2 -= carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 *(float64 *) &right4y0 = out0 *(float64 *) &right4y1 = out1 *(float64 *) &right4y2 = out2 *(float64 *) &right4y3 = out3 *(float64 *) &right4y4 = out4 # right4z = mult(right2z,left3z) out3 = *(float64 *) &right2z4 out3 *= *(float64 *) &left3z4 out3 *= *(float64 *) &gaudry_pm_scale u0v3 = *(float64 *) &right2z0 u0v3 *= *(float64 *) &left3z3 out3 += u0v3 u1v2 = *(float64 *) &right2z1 u1v2 *= *(float64 *) &left3z2 out3 += u1v2 u2v1 = *(float64 *) &right2z2 u2v1 *= *(float64 *) &left3z1 out3 += u2v1 u3v0 = *(float64 *) &right2z3 u3v0 *= *(float64 *) &left3z0 out3 += u3v0 out4 = *(float64 *) &right2z0 out4 *= *(float64 *) &left3z4 u1v3 = *(float64 *) &right2z1 u1v3 *= *(float64 *) &left3z3 out4 += u1v3 u2v2 = *(float64 *) &right2z2 u2v2 *= *(float64 *) &left3z2 out4 += u2v2 u3v1 = *(float64 *) &right2z3 u3v1 *= *(float64 *) &left3z1 out4 += u3v1 u4v0 = *(float64 *) &right2z4 u4v0 *= *(float64 *) &left3z0 out4 += u4v0 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &right2z1 out0 *= *(float64 *) &left3z4 u2v3 = *(float64 *) &right2z2 u2v3 *= *(float64 *) &left3z3 out0 += u2v3 u3v2 = *(float64 *) &right2z3 u3v2 *= *(float64 *) &left3z2 out0 += u3v2 u4v1 = *(float64 *) &right2z4 u4v1 *= *(float64 *) &left3z1 out0 += u4v1 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0v0 = *(float64 *) &right2z0 u0v0 *= *(float64 *) &left3z0 out0 += u0v0 out1 = *(float64 *) &right2z2 out1 *= *(float64 *) &left3z4 u3v3 = *(float64 *) &right2z3 u3v3 *= *(float64 *) &left3z3 out1 += u3v3 u4v2 = *(float64 *) &right2z4 u4v2 *= *(float64 *) &left3z2 out1 += u4v2 out1 *= *(float64 *) &gaudry_pm_scale u0v1 = *(float64 *) &right2z0 u0v1 *= *(float64 *) &left3z1 out1 += u0v1 u1v0 = *(float64 *) &right2z1 u1v0 *= *(float64 *) &left3z0 out1 += u1v0 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 out2 = *(float64 *) &right2z3 out2 *= *(float64 *) &left3z4 u4v3 = *(float64 *) &right2z4 u4v3 *= *(float64 *) &left3z3 out2 += u4v3 out2 *= *(float64 *) &gaudry_pm_scale u0v2 = *(float64 *) &right2z0 u0v2 *= *(float64 *) &left3z2 out2 += u0v2 u1v1 = *(float64 *) &right2z1 u1v1 *= *(float64 *) &left3z1 out2 += u1v1 u2v0 = *(float64 *) &right2z2 u2v0 *= *(float64 *) &left3z0 out2 += u2v0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 out1 -= carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 out2 -= carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 *(float64 *) &right4z0 = out0 *(float64 *) &right4z1 = out1 *(float64 *) &right4z2 = out2 *(float64 *) &right4z3 = out3 *(float64 *) &right4z4 = out4 # right4t = mult(right2t,left3t) out3 = *(float64 *) &right2t4 out3 *= *(float64 *) &left3t4 out3 *= *(float64 *) &gaudry_pm_scale u0v3 = *(float64 *) &right2t0 u0v3 *= *(float64 *) &left3t3 out3 += u0v3 u1v2 = *(float64 *) &right2t1 u1v2 *= *(float64 *) &left3t2 out3 += u1v2 u2v1 = *(float64 *) &right2t2 u2v1 *= *(float64 *) &left3t1 out3 += u2v1 u3v0 = *(float64 *) &right2t3 u3v0 *= *(float64 *) &left3t0 out3 += u3v0 out4 = *(float64 *) &right2t0 out4 *= *(float64 *) &left3t4 u1v3 = *(float64 *) &right2t1 u1v3 *= *(float64 *) &left3t3 out4 += u1v3 u2v2 = *(float64 *) &right2t2 u2v2 *= *(float64 *) &left3t2 out4 += u2v2 u3v1 = *(float64 *) &right2t3 u3v1 *= *(float64 *) &left3t1 out4 += u3v1 u4v0 = *(float64 *) &right2t4 u4v0 *= *(float64 *) &left3t0 out4 += u4v0 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &right2t1 out0 *= *(float64 *) &left3t4 u2v3 = *(float64 *) &right2t2 u2v3 *= *(float64 *) &left3t3 out0 += u2v3 u3v2 = *(float64 *) &right2t3 u3v2 *= *(float64 *) &left3t2 out0 += u3v2 u4v1 = *(float64 *) &right2t4 u4v1 *= *(float64 *) &left3t1 out0 += u4v1 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0v0 = *(float64 *) &right2t0 u0v0 *= *(float64 *) &left3t0 out0 += u0v0 out1 = *(float64 *) &right2t2 out1 *= *(float64 *) &left3t4 u3v3 = *(float64 *) &right2t3 u3v3 *= *(float64 *) &left3t3 out1 += u3v3 u4v2 = *(float64 *) &right2t4 u4v2 *= *(float64 *) &left3t2 out1 += u4v2 out1 *= *(float64 *) &gaudry_pm_scale u0v1 = *(float64 *) &right2t0 u0v1 *= *(float64 *) &left3t1 out1 += u0v1 u1v0 = *(float64 *) &right2t1 u1v0 *= *(float64 *) &left3t0 out1 += u1v0 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 out2 = *(float64 *) &right2t3 out2 *= *(float64 *) &left3t4 u4v3 = *(float64 *) &right2t4 u4v3 *= *(float64 *) &left3t3 out2 += u4v3 out2 *= *(float64 *) &gaudry_pm_scale u0v2 = *(float64 *) &right2t0 u0v2 *= *(float64 *) &left3t2 out2 += u0v2 u1v1 = *(float64 *) &right2t1 u1v1 *= *(float64 *) &left3t1 out2 += u1v1 u2v0 = *(float64 *) &right2t2 u2v0 *= *(float64 *) &left3t0 out2 += u2v0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 out1 -= carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 out2 -= carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 *(float64 *) &right4t0 = out0 *(float64 *) &right4t1 = out1 *(float64 *) &right4t2 = out2 *(float64 *) &right4t3 = out3 *(float64 *) &right4t4 = out4 # left5 = hadamard(left4); q2x = left5x f0 = *(float64 *) &left4x0 d0 = *(float64 *) &left4y0 g0 = *(float64 *) &left4z0 e0 = *(float64 *) &left4t0 d0 += f0 e0 += g0 f0 -= *(float64 *) &left4y0 g0 -= *(float64 *) &left4t0 *(float64 *) &leftpmx0 = d0 *(float64 *) &leftpmy0 = e0 *(float64 *) &leftpmz0 = f0 *(float64 *) &leftpmt0 = g0 f1 = *(float64 *) &left4x1 d1 = *(float64 *) &left4y1 g1 = *(float64 *) &left4z1 e1 = *(float64 *) &left4t1 d1 += f1 e1 += g1 f1 -= *(float64 *) &left4y1 g1 -= *(float64 *) &left4t1 *(float64 *) &leftpmx1 = d1 *(float64 *) &leftpmy1 = e1 *(float64 *) &leftpmz1 = f1 *(float64 *) &leftpmt1 = g1 f2 = *(float64 *) &left4x2 d2 = *(float64 *) &left4y2 g2 = *(float64 *) &left4z2 e2 = *(float64 *) &left4t2 d2 += f2 e2 += g2 f2 -= *(float64 *) &left4y2 g2 -= *(float64 *) &left4t2 *(float64 *) &leftpmx2 = d2 *(float64 *) &leftpmy2 = e2 *(float64 *) &leftpmz2 = f2 *(float64 *) &leftpmt2 = g2 f3 = *(float64 *) &left4x3 d3 = *(float64 *) &left4y3 g3 = *(float64 *) &left4z3 e3 = *(float64 *) &left4t3 d3 += f3 e3 += g3 f3 -= *(float64 *) &left4y3 g3 -= *(float64 *) &left4t3 *(float64 *) &leftpmx3 = d3 *(float64 *) &leftpmy3 = e3 *(float64 *) &leftpmz3 = f3 *(float64 *) &leftpmt3 = g3 f4 = *(float64 *) &left4x4 d4 = *(float64 *) &left4y4 g4 = *(float64 *) &left4z4 e4 = *(float64 *) &left4t4 d4 += f4 e4 += g4 f4 -= *(float64 *) &left4y4 g4 -= *(float64 *) &left4t4 *(float64 *) &leftpmx4 = d4 *(float64 *) &leftpmy4 = e4 *(float64 *) &leftpmz4 = f4 *(float64 *) &leftpmt4 = g4 j0 = *(float64 *) &leftpmx0 h0 = *(float64 *) &leftpmy0 k0 = *(float64 *) &leftpmz0 i0 = *(float64 *) &leftpmt0 h0 += j0 i0 += k0 j0 -= *(float64 *) &leftpmy0 k0 -= *(float64 *) &leftpmt0 *(float64 *) &q2x0 = h0 *(float64 *) &left5y0 = j0 *(float64 *) &left5z0 = i0 *(float64 *) &left5t0 = k0 j1 = *(float64 *) &leftpmx1 h1 = *(float64 *) &leftpmy1 k1 = *(float64 *) &leftpmz1 i1 = *(float64 *) &leftpmt1 h1 += j1 i1 += k1 j1 -= *(float64 *) &leftpmy1 k1 -= *(float64 *) &leftpmt1 *(float64 *) &q2x1 = h1 *(float64 *) &left5y1 = j1 *(float64 *) &left5z1 = i1 *(float64 *) &left5t1 = k1 j2 = *(float64 *) &leftpmx2 h2 = *(float64 *) &leftpmy2 k2 = *(float64 *) &leftpmz2 i2 = *(float64 *) &leftpmt2 h2 += j2 i2 += k2 j2 -= *(float64 *) &leftpmy2 k2 -= *(float64 *) &leftpmt2 *(float64 *) &q2x2 = h2 *(float64 *) &left5y2 = j2 *(float64 *) &left5z2 = i2 *(float64 *) &left5t2 = k2 j3 = *(float64 *) &leftpmx3 h3 = *(float64 *) &leftpmy3 k3 = *(float64 *) &leftpmz3 i3 = *(float64 *) &leftpmt3 h3 += j3 i3 += k3 j3 -= *(float64 *) &leftpmy3 k3 -= *(float64 *) &leftpmt3 *(float64 *) &q2x3 = h3 *(float64 *) &left5y3 = j3 *(float64 *) &left5z3 = i3 *(float64 *) &left5t3 = k3 j4 = *(float64 *) &leftpmx4 h4 = *(float64 *) &leftpmy4 k4 = *(float64 *) &leftpmz4 i4 = *(float64 *) &leftpmt4 h4 += j4 i4 += k4 j4 -= *(float64 *) &leftpmy4 k4 -= *(float64 *) &leftpmt4 *(float64 *) &q2x4 = h4 *(float64 *) &left5y4 = j4 *(float64 *) &left5z4 = i4 *(float64 *) &left5t4 = k4 # right5 = hadamard(right4); q3x = right5x f0 = *(float64 *) &right4x0 d0 = *(float64 *) &right4y0 g0 = *(float64 *) &right4z0 e0 = *(float64 *) &right4t0 d0 += f0 e0 += g0 f0 -= *(float64 *) &right4y0 g0 -= *(float64 *) &right4t0 *(float64 *) &rightpmx0 = d0 *(float64 *) &rightpmy0 = e0 *(float64 *) &rightpmz0 = f0 *(float64 *) &rightpmt0 = g0 f1 = *(float64 *) &right4x1 d1 = *(float64 *) &right4y1 g1 = *(float64 *) &right4z1 e1 = *(float64 *) &right4t1 d1 += f1 e1 += g1 f1 -= *(float64 *) &right4y1 g1 -= *(float64 *) &right4t1 *(float64 *) &rightpmx1 = d1 *(float64 *) &rightpmy1 = e1 *(float64 *) &rightpmz1 = f1 *(float64 *) &rightpmt1 = g1 f2 = *(float64 *) &right4x2 d2 = *(float64 *) &right4y2 g2 = *(float64 *) &right4z2 e2 = *(float64 *) &right4t2 d2 += f2 e2 += g2 f2 -= *(float64 *) &right4y2 g2 -= *(float64 *) &right4t2 *(float64 *) &rightpmx2 = d2 *(float64 *) &rightpmy2 = e2 *(float64 *) &rightpmz2 = f2 *(float64 *) &rightpmt2 = g2 f3 = *(float64 *) &right4x3 d3 = *(float64 *) &right4y3 g3 = *(float64 *) &right4z3 e3 = *(float64 *) &right4t3 d3 += f3 e3 += g3 f3 -= *(float64 *) &right4y3 g3 -= *(float64 *) &right4t3 *(float64 *) &rightpmx3 = d3 *(float64 *) &rightpmy3 = e3 *(float64 *) &rightpmz3 = f3 *(float64 *) &rightpmt3 = g3 f4 = *(float64 *) &right4x4 d4 = *(float64 *) &right4y4 g4 = *(float64 *) &right4z4 e4 = *(float64 *) &right4t4 d4 += f4 e4 += g4 f4 -= *(float64 *) &right4y4 g4 -= *(float64 *) &right4t4 *(float64 *) &rightpmx4 = d4 *(float64 *) &rightpmy4 = e4 *(float64 *) &rightpmz4 = f4 *(float64 *) &rightpmt4 = g4 j0 = *(float64 *) &rightpmx0 h0 = *(float64 *) &rightpmy0 k0 = *(float64 *) &rightpmz0 i0 = *(float64 *) &rightpmt0 h0 += j0 i0 += k0 j0 -= *(float64 *) &rightpmy0 k0 -= *(float64 *) &rightpmt0 *(float64 *) &q3x0 = h0 *(float64 *) &right5y0 = j0 *(float64 *) &right5z0 = i0 *(float64 *) &right5t0 = k0 j1 = *(float64 *) &rightpmx1 h1 = *(float64 *) &rightpmy1 k1 = *(float64 *) &rightpmz1 i1 = *(float64 *) &rightpmt1 h1 += j1 i1 += k1 j1 -= *(float64 *) &rightpmy1 k1 -= *(float64 *) &rightpmt1 *(float64 *) &q3x1 = h1 *(float64 *) &right5y1 = j1 *(float64 *) &right5z1 = i1 *(float64 *) &right5t1 = k1 j2 = *(float64 *) &rightpmx2 h2 = *(float64 *) &rightpmy2 k2 = *(float64 *) &rightpmz2 i2 = *(float64 *) &rightpmt2 h2 += j2 i2 += k2 j2 -= *(float64 *) &rightpmy2 k2 -= *(float64 *) &rightpmt2 *(float64 *) &q3x2 = h2 *(float64 *) &right5y2 = j2 *(float64 *) &right5z2 = i2 *(float64 *) &right5t2 = k2 j3 = *(float64 *) &rightpmx3 h3 = *(float64 *) &rightpmy3 k3 = *(float64 *) &rightpmz3 i3 = *(float64 *) &rightpmt3 h3 += j3 i3 += k3 j3 -= *(float64 *) &rightpmy3 k3 -= *(float64 *) &rightpmt3 *(float64 *) &q3x3 = h3 *(float64 *) &right5y3 = j3 *(float64 *) &right5z3 = i3 *(float64 *) &right5t3 = k3 j4 = *(float64 *) &rightpmx4 h4 = *(float64 *) &rightpmy4 k4 = *(float64 *) &rightpmz4 i4 = *(float64 *) &rightpmt4 h4 += j4 i4 += k4 j4 -= *(float64 *) &rightpmy4 k4 -= *(float64 *) &rightpmt4 *(float64 *) &q3x4 = h4 *(float64 *) &right5y4 = j4 *(float64 *) &right5z4 = i4 *(float64 *) &right5t4 = k4 # q2y = mult(left5y,aoverb) out3 = *(float64 *) &left5y4 out3 *= *(float64 *) &aoverb4 out3 *= *(float64 *) &gaudry_pm_scale u0v3 = *(float64 *) &left5y0 u0v3 *= *(float64 *) &aoverb3 out3 += u0v3 u1v2 = *(float64 *) &left5y1 u1v2 *= *(float64 *) &aoverb2 out3 += u1v2 u2v1 = *(float64 *) &left5y2 u2v1 *= *(float64 *) &aoverb1 out3 += u2v1 u3v0 = *(float64 *) &left5y3 u3v0 *= *(float64 *) &aoverb0 out3 += u3v0 out4 = *(float64 *) &left5y0 out4 *= *(float64 *) &aoverb4 u1v3 = *(float64 *) &left5y1 u1v3 *= *(float64 *) &aoverb3 out4 += u1v3 u2v2 = *(float64 *) &left5y2 u2v2 *= *(float64 *) &aoverb2 out4 += u2v2 u3v1 = *(float64 *) &left5y3 u3v1 *= *(float64 *) &aoverb1 out4 += u3v1 u4v0 = *(float64 *) &left5y4 u4v0 *= *(float64 *) &aoverb0 out4 += u4v0 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &left5y1 out0 *= *(float64 *) &aoverb4 u2v3 = *(float64 *) &left5y2 u2v3 *= *(float64 *) &aoverb3 out0 += u2v3 u3v2 = *(float64 *) &left5y3 u3v2 *= *(float64 *) &aoverb2 out0 += u3v2 u4v1 = *(float64 *) &left5y4 u4v1 *= *(float64 *) &aoverb1 out0 += u4v1 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0v0 = *(float64 *) &left5y0 u0v0 *= *(float64 *) &aoverb0 out0 += u0v0 out1 = *(float64 *) &left5y2 out1 *= *(float64 *) &aoverb4 u3v3 = *(float64 *) &left5y3 u3v3 *= *(float64 *) &aoverb3 out1 += u3v3 u4v2 = *(float64 *) &left5y4 u4v2 *= *(float64 *) &aoverb2 out1 += u4v2 out1 *= *(float64 *) &gaudry_pm_scale u0v1 = *(float64 *) &left5y0 u0v1 *= *(float64 *) &aoverb1 out1 += u0v1 u1v0 = *(float64 *) &left5y1 u1v0 *= *(float64 *) &aoverb0 out1 += u1v0 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 out2 = *(float64 *) &left5y3 out2 *= *(float64 *) &aoverb4 u4v3 = *(float64 *) &left5y4 u4v3 *= *(float64 *) &aoverb3 out2 += u4v3 out2 *= *(float64 *) &gaudry_pm_scale u0v2 = *(float64 *) &left5y0 u0v2 *= *(float64 *) &aoverb2 out2 += u0v2 u1v1 = *(float64 *) &left5y1 u1v1 *= *(float64 *) &aoverb1 out2 += u1v1 u2v0 = *(float64 *) &left5y2 u2v0 *= *(float64 *) &aoverb0 out2 += u2v0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 out1 -= carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 out2 -= carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 *(float64 *) &q2y0 = out0 *(float64 *) &q2y1 = out1 *(float64 *) &q2y2 = out2 *(float64 *) &q2y3 = out3 *(float64 *) &q2y4 = out4 # q2z = mult(left5z,aoverc) out3 = *(float64 *) &left5z4 out3 *= *(float64 *) &aoverc4 out3 *= *(float64 *) &gaudry_pm_scale u0v3 = *(float64 *) &left5z0 u0v3 *= *(float64 *) &aoverc3 out3 += u0v3 u1v2 = *(float64 *) &left5z1 u1v2 *= *(float64 *) &aoverc2 out3 += u1v2 u2v1 = *(float64 *) &left5z2 u2v1 *= *(float64 *) &aoverc1 out3 += u2v1 u3v0 = *(float64 *) &left5z3 u3v0 *= *(float64 *) &aoverc0 out3 += u3v0 out4 = *(float64 *) &left5z0 out4 *= *(float64 *) &aoverc4 u1v3 = *(float64 *) &left5z1 u1v3 *= *(float64 *) &aoverc3 out4 += u1v3 u2v2 = *(float64 *) &left5z2 u2v2 *= *(float64 *) &aoverc2 out4 += u2v2 u3v1 = *(float64 *) &left5z3 u3v1 *= *(float64 *) &aoverc1 out4 += u3v1 u4v0 = *(float64 *) &left5z4 u4v0 *= *(float64 *) &aoverc0 out4 += u4v0 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &left5z1 out0 *= *(float64 *) &aoverc4 u2v3 = *(float64 *) &left5z2 u2v3 *= *(float64 *) &aoverc3 out0 += u2v3 u3v2 = *(float64 *) &left5z3 u3v2 *= *(float64 *) &aoverc2 out0 += u3v2 u4v1 = *(float64 *) &left5z4 u4v1 *= *(float64 *) &aoverc1 out0 += u4v1 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0v0 = *(float64 *) &left5z0 u0v0 *= *(float64 *) &aoverc0 out0 += u0v0 out1 = *(float64 *) &left5z2 out1 *= *(float64 *) &aoverc4 u3v3 = *(float64 *) &left5z3 u3v3 *= *(float64 *) &aoverc3 out1 += u3v3 u4v2 = *(float64 *) &left5z4 u4v2 *= *(float64 *) &aoverc2 out1 += u4v2 out1 *= *(float64 *) &gaudry_pm_scale u0v1 = *(float64 *) &left5z0 u0v1 *= *(float64 *) &aoverc1 out1 += u0v1 u1v0 = *(float64 *) &left5z1 u1v0 *= *(float64 *) &aoverc0 out1 += u1v0 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 out2 = *(float64 *) &left5z3 out2 *= *(float64 *) &aoverc4 u4v3 = *(float64 *) &left5z4 u4v3 *= *(float64 *) &aoverc3 out2 += u4v3 out2 *= *(float64 *) &gaudry_pm_scale u0v2 = *(float64 *) &left5z0 u0v2 *= *(float64 *) &aoverc2 out2 += u0v2 u1v1 = *(float64 *) &left5z1 u1v1 *= *(float64 *) &aoverc1 out2 += u1v1 u2v0 = *(float64 *) &left5z2 u2v0 *= *(float64 *) &aoverc0 out2 += u2v0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 out1 -= carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 out2 -= carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 *(float64 *) &q2z0 = out0 *(float64 *) &q2z1 = out1 *(float64 *) &q2z2 = out2 *(float64 *) &q2z3 = out3 *(float64 *) &q2z4 = out4 # q2t = mult(left5t,aoverd) out3 = *(float64 *) &left5t4 out3 *= *(float64 *) &aoverd4 out3 *= *(float64 *) &gaudry_pm_scale u0v3 = *(float64 *) &left5t0 u0v3 *= *(float64 *) &aoverd3 out3 += u0v3 u1v2 = *(float64 *) &left5t1 u1v2 *= *(float64 *) &aoverd2 out3 += u1v2 u2v1 = *(float64 *) &left5t2 u2v1 *= *(float64 *) &aoverd1 out3 += u2v1 u3v0 = *(float64 *) &left5t3 u3v0 *= *(float64 *) &aoverd0 out3 += u3v0 out4 = *(float64 *) &left5t0 out4 *= *(float64 *) &aoverd4 u1v3 = *(float64 *) &left5t1 u1v3 *= *(float64 *) &aoverd3 out4 += u1v3 u2v2 = *(float64 *) &left5t2 u2v2 *= *(float64 *) &aoverd2 out4 += u2v2 u3v1 = *(float64 *) &left5t3 u3v1 *= *(float64 *) &aoverd1 out4 += u3v1 u4v0 = *(float64 *) &left5t4 u4v0 *= *(float64 *) &aoverd0 out4 += u4v0 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &left5t1 out0 *= *(float64 *) &aoverd4 u2v3 = *(float64 *) &left5t2 u2v3 *= *(float64 *) &aoverd3 out0 += u2v3 u3v2 = *(float64 *) &left5t3 u3v2 *= *(float64 *) &aoverd2 out0 += u3v2 u4v1 = *(float64 *) &left5t4 u4v1 *= *(float64 *) &aoverd1 out0 += u4v1 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0v0 = *(float64 *) &left5t0 u0v0 *= *(float64 *) &aoverd0 out0 += u0v0 out1 = *(float64 *) &left5t2 out1 *= *(float64 *) &aoverd4 u3v3 = *(float64 *) &left5t3 u3v3 *= *(float64 *) &aoverd3 out1 += u3v3 u4v2 = *(float64 *) &left5t4 u4v2 *= *(float64 *) &aoverd2 out1 += u4v2 out1 *= *(float64 *) &gaudry_pm_scale u0v1 = *(float64 *) &left5t0 u0v1 *= *(float64 *) &aoverd1 out1 += u0v1 u1v0 = *(float64 *) &left5t1 u1v0 *= *(float64 *) &aoverd0 out1 += u1v0 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 out2 = *(float64 *) &left5t3 out2 *= *(float64 *) &aoverd4 u4v3 = *(float64 *) &left5t4 u4v3 *= *(float64 *) &aoverd3 out2 += u4v3 out2 *= *(float64 *) &gaudry_pm_scale u0v2 = *(float64 *) &left5t0 u0v2 *= *(float64 *) &aoverd2 out2 += u0v2 u1v1 = *(float64 *) &left5t1 u1v1 *= *(float64 *) &aoverd1 out2 += u1v1 u2v0 = *(float64 *) &left5t2 u2v0 *= *(float64 *) &aoverd0 out2 += u2v0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 out1 -= carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 out2 -= carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 *(float64 *) &q2t0 = out0 *(float64 *) &q2t1 = out1 *(float64 *) &q2t2 = out2 *(float64 *) &q2t3 = out3 *(float64 *) &q2t4 = out4 # q3y = mult(right5y,x1overy1) out3 = *(float64 *) &right5y4 out3 *= *(float64 *) &xovery4 out3 *= *(float64 *) &gaudry_pm_scale u0v3 = *(float64 *) &right5y0 u0v3 *= *(float64 *) &xovery3 out3 += u0v3 u1v2 = *(float64 *) &right5y1 u1v2 *= *(float64 *) &xovery2 out3 += u1v2 u2v1 = *(float64 *) &right5y2 u2v1 *= *(float64 *) &xovery1 out3 += u2v1 u3v0 = *(float64 *) &right5y3 u3v0 *= *(float64 *) &xovery0 out3 += u3v0 out4 = *(float64 *) &right5y0 out4 *= *(float64 *) &xovery4 u1v3 = *(float64 *) &right5y1 u1v3 *= *(float64 *) &xovery3 out4 += u1v3 u2v2 = *(float64 *) &right5y2 u2v2 *= *(float64 *) &xovery2 out4 += u2v2 u3v1 = *(float64 *) &right5y3 u3v1 *= *(float64 *) &xovery1 out4 += u3v1 u4v0 = *(float64 *) &right5y4 u4v0 *= *(float64 *) &xovery0 out4 += u4v0 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &right5y1 out0 *= *(float64 *) &xovery4 u2v3 = *(float64 *) &right5y2 u2v3 *= *(float64 *) &xovery3 out0 += u2v3 u3v2 = *(float64 *) &right5y3 u3v2 *= *(float64 *) &xovery2 out0 += u3v2 u4v1 = *(float64 *) &right5y4 u4v1 *= *(float64 *) &xovery1 out0 += u4v1 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0v0 = *(float64 *) &right5y0 u0v0 *= *(float64 *) &xovery0 out0 += u0v0 out1 = *(float64 *) &right5y2 out1 *= *(float64 *) &xovery4 u3v3 = *(float64 *) &right5y3 u3v3 *= *(float64 *) &xovery3 out1 += u3v3 u4v2 = *(float64 *) &right5y4 u4v2 *= *(float64 *) &xovery2 out1 += u4v2 out1 *= *(float64 *) &gaudry_pm_scale u0v1 = *(float64 *) &right5y0 u0v1 *= *(float64 *) &xovery1 out1 += u0v1 u1v0 = *(float64 *) &right5y1 u1v0 *= *(float64 *) &xovery0 out1 += u1v0 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 out2 = *(float64 *) &right5y3 out2 *= *(float64 *) &xovery4 u4v3 = *(float64 *) &right5y4 u4v3 *= *(float64 *) &xovery3 out2 += u4v3 out2 *= *(float64 *) &gaudry_pm_scale u0v2 = *(float64 *) &right5y0 u0v2 *= *(float64 *) &xovery2 out2 += u0v2 u1v1 = *(float64 *) &right5y1 u1v1 *= *(float64 *) &xovery1 out2 += u1v1 u2v0 = *(float64 *) &right5y2 u2v0 *= *(float64 *) &xovery0 out2 += u2v0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 out1 -= carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 out2 -= carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 *(float64 *) &q3y0 = out0 *(float64 *) &q3y1 = out1 *(float64 *) &q3y2 = out2 *(float64 *) &q3y3 = out3 *(float64 *) &q3y4 = out4 # q3z = mult(right5z,x1overz1) out3 = *(float64 *) &right5z4 out3 *= *(float64 *) &xoverz4 out3 *= *(float64 *) &gaudry_pm_scale u0v3 = *(float64 *) &right5z0 u0v3 *= *(float64 *) &xoverz3 out3 += u0v3 u1v2 = *(float64 *) &right5z1 u1v2 *= *(float64 *) &xoverz2 out3 += u1v2 u2v1 = *(float64 *) &right5z2 u2v1 *= *(float64 *) &xoverz1 out3 += u2v1 u3v0 = *(float64 *) &right5z3 u3v0 *= *(float64 *) &xoverz0 out3 += u3v0 out4 = *(float64 *) &right5z0 out4 *= *(float64 *) &xoverz4 u1v3 = *(float64 *) &right5z1 u1v3 *= *(float64 *) &xoverz3 out4 += u1v3 u2v2 = *(float64 *) &right5z2 u2v2 *= *(float64 *) &xoverz2 out4 += u2v2 u3v1 = *(float64 *) &right5z3 u3v1 *= *(float64 *) &xoverz1 out4 += u3v1 u4v0 = *(float64 *) &right5z4 u4v0 *= *(float64 *) &xoverz0 out4 += u4v0 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &right5z1 out0 *= *(float64 *) &xoverz4 u2v3 = *(float64 *) &right5z2 u2v3 *= *(float64 *) &xoverz3 out0 += u2v3 u3v2 = *(float64 *) &right5z3 u3v2 *= *(float64 *) &xoverz2 out0 += u3v2 u4v1 = *(float64 *) &right5z4 u4v1 *= *(float64 *) &xoverz1 out0 += u4v1 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0v0 = *(float64 *) &right5z0 u0v0 *= *(float64 *) &xoverz0 out0 += u0v0 out1 = *(float64 *) &right5z2 out1 *= *(float64 *) &xoverz4 u3v3 = *(float64 *) &right5z3 u3v3 *= *(float64 *) &xoverz3 out1 += u3v3 u4v2 = *(float64 *) &right5z4 u4v2 *= *(float64 *) &xoverz2 out1 += u4v2 out1 *= *(float64 *) &gaudry_pm_scale u0v1 = *(float64 *) &right5z0 u0v1 *= *(float64 *) &xoverz1 out1 += u0v1 u1v0 = *(float64 *) &right5z1 u1v0 *= *(float64 *) &xoverz0 out1 += u1v0 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 out2 = *(float64 *) &right5z3 out2 *= *(float64 *) &xoverz4 u4v3 = *(float64 *) &right5z4 u4v3 *= *(float64 *) &xoverz3 out2 += u4v3 out2 *= *(float64 *) &gaudry_pm_scale u0v2 = *(float64 *) &right5z0 u0v2 *= *(float64 *) &xoverz2 out2 += u0v2 u1v1 = *(float64 *) &right5z1 u1v1 *= *(float64 *) &xoverz1 out2 += u1v1 u2v0 = *(float64 *) &right5z2 u2v0 *= *(float64 *) &xoverz0 out2 += u2v0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 out1 -= carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 out2 -= carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 *(float64 *) &q3z0 = out0 *(float64 *) &q3z1 = out1 *(float64 *) &q3z2 = out2 *(float64 *) &q3z3 = out3 *(float64 *) &q3z4 = out4 # q3t = mult(right5t,x1overt1) out3 = *(float64 *) &right5t4 out3 *= *(float64 *) &xovert4 out3 *= *(float64 *) &gaudry_pm_scale u0v3 = *(float64 *) &right5t0 u0v3 *= *(float64 *) &xovert3 out3 += u0v3 u1v2 = *(float64 *) &right5t1 u1v2 *= *(float64 *) &xovert2 out3 += u1v2 u2v1 = *(float64 *) &right5t2 u2v1 *= *(float64 *) &xovert1 out3 += u2v1 u3v0 = *(float64 *) &right5t3 u3v0 *= *(float64 *) &xovert0 out3 += u3v0 out4 = *(float64 *) &right5t0 out4 *= *(float64 *) &xovert4 u1v3 = *(float64 *) &right5t1 u1v3 *= *(float64 *) &xovert3 out4 += u1v3 u2v2 = *(float64 *) &right5t2 u2v2 *= *(float64 *) &xovert2 out4 += u2v2 u3v1 = *(float64 *) &right5t3 u3v1 *= *(float64 *) &xovert1 out4 += u3v1 u4v0 = *(float64 *) &right5t4 u4v0 *= *(float64 *) &xovert0 out4 += u4v0 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 out0 = *(float64 *) &right5t1 out0 *= *(float64 *) &xovert4 u2v3 = *(float64 *) &right5t2 u2v3 *= *(float64 *) &xovert3 out0 += u2v3 u3v2 = *(float64 *) &right5t3 u3v2 *= *(float64 *) &xovert2 out0 += u3v2 u4v1 = *(float64 *) &right5t4 u4v1 *= *(float64 *) &xovert1 out0 += u4v1 carry4 = *(float64 *) &gaudry_pm_alpha127 carry4 += out4 carry4 -= *(float64 *) &gaudry_pm_alpha127 out0 += carry4 out4 -= carry4 out0 *= *(float64 *) &gaudry_pm_scale u0v0 = *(float64 *) &right5t0 u0v0 *= *(float64 *) &xovert0 out0 += u0v0 out1 = *(float64 *) &right5t2 out1 *= *(float64 *) &xovert4 u3v3 = *(float64 *) &right5t3 u3v3 *= *(float64 *) &xovert3 out1 += u3v3 u4v2 = *(float64 *) &right5t4 u4v2 *= *(float64 *) &xovert2 out1 += u4v2 out1 *= *(float64 *) &gaudry_pm_scale u0v1 = *(float64 *) &right5t0 u0v1 *= *(float64 *) &xovert1 out1 += u0v1 u1v0 = *(float64 *) &right5t1 u1v0 *= *(float64 *) &xovert0 out1 += u1v0 carry0 = *(float64 *) &gaudry_pm_alpha26 carry0 += out0 carry0 -= *(float64 *) &gaudry_pm_alpha26 out1 += carry0 out0 -= carry0 out2 = *(float64 *) &right5t3 out2 *= *(float64 *) &xovert4 u4v3 = *(float64 *) &right5t4 u4v3 *= *(float64 *) &xovert3 out2 += u4v3 out2 *= *(float64 *) &gaudry_pm_scale u0v2 = *(float64 *) &right5t0 u0v2 *= *(float64 *) &xovert2 out2 += u0v2 u1v1 = *(float64 *) &right5t1 u1v1 *= *(float64 *) &xovert1 out2 += u1v1 u2v0 = *(float64 *) &right5t2 u2v0 *= *(float64 *) &xovert0 out2 += u2v0 carry1 = *(float64 *) &gaudry_pm_alpha51 carry1 += out1 carry1 -= *(float64 *) &gaudry_pm_alpha51 out2 += carry1 out1 -= carry1 carry2 = *(float64 *) &gaudry_pm_alpha77 carry2 += out2 carry2 -= *(float64 *) &gaudry_pm_alpha77 out3 += carry2 out2 -= carry2 carry3 = *(float64 *) &gaudry_pm_alpha102 carry3 += out3 carry3 -= *(float64 *) &gaudry_pm_alpha102 out4 += carry3 out3 -= carry3 *(float64 *) &q3t0 = out0 *(float64 *) &q3t1 = out1 *(float64 *) &q3t2 = out2 *(float64 *) &q3t3 = out3 *(float64 *) &q3t4 = out4 # loop bottom selectbit = selectbit_stack thisword = thisword_stack bits = bits_stack bits -= 1 goto morebits if unsigned> bytes = bytes_stack bytes -= 4 goto done if unsigned< e = e_stack thisword = *(uint32 *) (e + bytes) bits = 32 goto morebytes done: q2 = q2_stack x0 = *(float64 *) &q2x0 x1 = *(float64 *) &q2x1 x2 = *(float64 *) &q2x2 x3 = *(float64 *) &q2x3 x4 = *(float64 *) &q2x4 *(float64 *) (q2 + 0) = x0 *(float64 *) (q2 + 8) = x1 *(float64 *) (q2 + 16) = x2 *(float64 *) (q2 + 24) = x3 *(float64 *) (q2 + 32) = x4 x0 = *(float64 *) &q2y0 x1 = *(float64 *) &q2y1 x2 = *(float64 *) &q2y2 x3 = *(float64 *) &q2y3 x4 = *(float64 *) &q2y4 *(float64 *) (q2 + 40) = x0 *(float64 *) (q2 + 48) = x1 *(float64 *) (q2 + 56) = x2 *(float64 *) (q2 + 64) = x3 *(float64 *) (q2 + 72) = x4 x0 = *(float64 *) &q2z0 x1 = *(float64 *) &q2z1 x2 = *(float64 *) &q2z2 x3 = *(float64 *) &q2z3 x4 = *(float64 *) &q2z4 *(float64 *) (q2 + 80) = x0 *(float64 *) (q2 + 88) = x1 *(float64 *) (q2 + 96) = x2 *(float64 *) (q2 + 104) = x3 *(float64 *) (q2 + 112) = x4 x0 = *(float64 *) &q2t0 x1 = *(float64 *) &q2t1 x2 = *(float64 *) &q2t2 x3 = *(float64 *) &q2t3 x4 = *(float64 *) &q2t4 *(float64 *) (q2 + 120) = x0 *(float64 *) (q2 + 128) = x1 *(float64 *) (q2 + 136) = x2 *(float64 *) (q2 + 144) = x3 *(float64 *) (q2 + 152) = x4 q3 = q3_stack x0 = *(float64 *) &q3x0 x1 = *(float64 *) &q3x1 x2 = *(float64 *) &q3x2 x3 = *(float64 *) &q3x3 x4 = *(float64 *) &q3x4 *(float64 *) (q3 + 0) = x0 *(float64 *) (q3 + 8) = x1 *(float64 *) (q3 + 16) = x2 *(float64 *) (q3 + 24) = x3 *(float64 *) (q3 + 32) = x4 x0 = *(float64 *) &q3y0 x1 = *(float64 *) &q3y1 x2 = *(float64 *) &q3y2 x3 = *(float64 *) &q3y3 x4 = *(float64 *) &q3y4 *(float64 *) (q3 + 40) = x0 *(float64 *) (q3 + 48) = x1 *(float64 *) (q3 + 56) = x2 *(float64 *) (q3 + 64) = x3 *(float64 *) (q3 + 72) = x4 x0 = *(float64 *) &q3z0 x1 = *(float64 *) &q3z1 x2 = *(float64 *) &q3z2 x3 = *(float64 *) &q3z3 x4 = *(float64 *) &q3z4 *(float64 *) (q3 + 80) = x0 *(float64 *) (q3 + 88) = x1 *(float64 *) (q3 + 96) = x2 *(float64 *) (q3 + 104) = x3 *(float64 *) (q3 + 112) = x4 x0 = *(float64 *) &q3t0 x1 = *(float64 *) &q3t1 x2 = *(float64 *) &q3t2 x3 = *(float64 *) &q3t3 x4 = *(float64 *) &q3t4 *(float64 *) (q3 + 120) = x0 *(float64 *) (q3 + 128) = x1 *(float64 *) (q3 + 136) = x2 *(float64 *) (q3 + 144) = x3 *(float64 *) (q3 + 152) = x4 eax = eax_stack ebx = ebx_stack esi = esi_stack edi = edi_stack ebp = ebp_stack leave