/* mpc_mul -- Multiply two complex numbers Copyright (C) 2002, 2004, 2005, 2008, 2009, 2010, 2011, 2012, 2016, 2020, 2022 INRIA This file is part of GNU MPC. GNU MPC is free software; you can redistribute it and/or modify it under he terms of the GNU Lesser General Public License as published by the Free Software Foundation; either version 3 of the License, or (at your option) any later version. GNU MPC is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public License for more details. You should have received a copy of the GNU Lesser General Public License along with this program. If not, see http://www.gnu.org/licenses/ . */ #include /* for MPC_ASSERT */ #include "mpc-impl.h" #define mpz_add_si(z,x,y) do { \ if (y >= 0) \ mpz_add_ui (z, x, (long int) y); \ else \ mpz_sub_ui (z, x, (long int) (-y)); \ } while (0); /* compute z=x*y when x has an infinite part */ static int mul_infinite (mpc_ptr z, mpc_srcptr x, mpc_srcptr y) { /* Let x=xr+i*xi and y=yr+i*yi; extract the signs of the operands */ int xrs = mpfr_signbit (mpc_realref (x)) ? -1 : 1; int xis = mpfr_signbit (mpc_imagref (x)) ? -1 : 1; int yrs = mpfr_signbit (mpc_realref (y)) ? -1 : 1; int yis = mpfr_signbit (mpc_imagref (y)) ? -1 : 1; int u, v; /* compute the sign of u = xrs * yrs * xr * yr - xis * yis * xi * yi v = xrs * yis * xr * yi + xis * yrs * xi * yr +1 if positive, -1 if negative, 0 if NaN */ if ( mpfr_nan_p (mpc_realref (x)) || mpfr_nan_p (mpc_imagref (x)) || mpfr_nan_p (mpc_realref (y)) || mpfr_nan_p (mpc_imagref (y))) { u = 0; v = 0; } else if (mpfr_inf_p (mpc_realref (x))) { /* x = (+/-inf) xr + i*xi */ u = ( mpfr_zero_p (mpc_realref (y)) || (mpfr_inf_p (mpc_imagref (x)) && mpfr_zero_p (mpc_imagref (y))) || (mpfr_zero_p (mpc_imagref (x)) && mpfr_inf_p (mpc_imagref (y))) || ( (mpfr_inf_p (mpc_imagref (x)) || mpfr_inf_p (mpc_imagref (y))) && xrs*yrs == xis*yis) ? 0 : xrs * yrs); v = ( mpfr_zero_p (mpc_imagref (y)) || (mpfr_inf_p (mpc_imagref (x)) && mpfr_zero_p (mpc_realref (y))) || (mpfr_zero_p (mpc_imagref (x)) && mpfr_inf_p (mpc_realref (y))) || ( (mpfr_inf_p (mpc_imagref (x)) || mpfr_inf_p (mpc_imagref (x))) && xrs*yis != xis*yrs) ? 0 : xrs * yis); } else { /* x = xr + i*(+/-inf) with |xr| != inf */ u = ( mpfr_zero_p (mpc_imagref (y)) || (mpfr_zero_p (mpc_realref (x)) && mpfr_inf_p (mpc_realref (y))) || (mpfr_inf_p (mpc_realref (y)) && xrs*yrs == xis*yis) ? 0 : -xis * yis); v = ( mpfr_zero_p (mpc_realref (y)) || (mpfr_zero_p (mpc_realref (x)) && mpfr_inf_p (mpc_imagref (y))) || (mpfr_inf_p (mpc_imagref (y)) && xrs*yis != xis*yrs) ? 0 : xis * yrs); } if (u == 0 && v == 0) { /* Naive result is NaN+i*NaN. Obtain an infinity using the algorithm given in Annex G.5.1 of the ISO C99 standard */ int xr = (mpfr_zero_p (mpc_realref (x)) || mpfr_nan_p (mpc_realref (x)) ? 0 : (mpfr_inf_p (mpc_realref (x)) ? 1 : 0)); int xi = (mpfr_zero_p (mpc_imagref (x)) || mpfr_nan_p (mpc_imagref (x)) ? 0 : (mpfr_inf_p (mpc_imagref (x)) ? 1 : 0)); int yr = (mpfr_zero_p (mpc_realref (y)) || mpfr_nan_p (mpc_realref (y)) ? 0 : 1); int yi = (mpfr_zero_p (mpc_imagref (y)) || mpfr_nan_p (mpc_imagref (y)) ? 0 : 1); if (mpc_inf_p (y)) { yr = mpfr_inf_p (mpc_realref (y)) ? 1 : 0; yi = mpfr_inf_p (mpc_imagref (y)) ? 1 : 0; } u = xrs * xr * yrs * yr - xis * xi * yis * yi; v = xrs * xr * yis * yi + xis * xi * yrs * yr; } if (u == 0) mpfr_set_nan (mpc_realref (z)); else mpfr_set_inf (mpc_realref (z), u); if (v == 0) mpfr_set_nan (mpc_imagref (z)); else mpfr_set_inf (mpc_imagref (z), v); return MPC_INEX (0, 0); /* exact */ } /* compute z = x*y for Im(y) == 0 */ static int mul_real (mpc_ptr z, mpc_srcptr x, mpc_srcptr y, mpc_rnd_t rnd) { int xrs, xis, yrs, yis; int inex; /* save signs of operands */ xrs = MPFR_SIGNBIT (mpc_realref (x)); xis = MPFR_SIGNBIT (mpc_imagref (x)); yrs = MPFR_SIGNBIT (mpc_realref (y)); yis = MPFR_SIGNBIT (mpc_imagref (y)); inex = mpc_mul_fr (z, x, mpc_realref (y), rnd); /* Signs of zeroes may be wrong. Their correction does not change the inexact flag. */ if (mpfr_zero_p (mpc_realref (z))) mpfr_setsign (mpc_realref (z), mpc_realref (z), MPC_RND_RE(rnd) == MPFR_RNDD || (xrs != yrs && xis == yis), MPFR_RNDN); if (mpfr_zero_p (mpc_imagref (z))) mpfr_setsign (mpc_imagref (z), mpc_imagref (z), MPC_RND_IM (rnd) == MPFR_RNDD || (xrs != yis && xis != yrs), MPFR_RNDN); return inex; } /* compute z = x*y for Re(y) == 0, and Im(x) != 0 and Im(y) != 0 */ static int mul_imag (mpc_ptr z, mpc_srcptr x, mpc_srcptr y, mpc_rnd_t rnd) { int sign; int inex_re, inex_im; int overlap = z == x || z == y; mpc_t rop; if (overlap) mpc_init3 (rop, MPC_PREC_RE (z), MPC_PREC_IM (z)); else rop [0] = z[0]; sign = (MPFR_SIGNBIT (mpc_realref (y)) != MPFR_SIGNBIT (mpc_imagref (x))) && (MPFR_SIGNBIT (mpc_imagref (y)) != MPFR_SIGNBIT (mpc_realref (x))); inex_re = -mpfr_mul (mpc_realref (rop), mpc_imagref (x), mpc_imagref (y), INV_RND (MPC_RND_RE (rnd))); mpfr_neg (mpc_realref (rop), mpc_realref (rop), MPFR_RNDN); /* exact */ inex_im = mpfr_mul (mpc_imagref (rop), mpc_realref (x), mpc_imagref (y), MPC_RND_IM (rnd)); mpc_set (z, rop, MPC_RNDNN); /* Sign of zeroes may be wrong (note that Re(z) cannot be zero) */ if (mpfr_zero_p (mpc_imagref (z))) mpfr_setsign (mpc_imagref (z), mpc_imagref (z), MPC_RND_IM (rnd) == MPFR_RNDD || sign, MPFR_RNDN); if (overlap) mpc_clear (rop); return MPC_INEX (inex_re, inex_im); } int mpc_mul_naive (mpc_ptr z, mpc_srcptr x, mpc_srcptr y, mpc_rnd_t rnd) { /* computes z=x*y by the schoolbook method, where x and y are assumed to be finite and without zero parts */ int overlap, inex_re, inex_im; mpc_t rop; MPC_ASSERT ( mpfr_regular_p (mpc_realref (x)) && mpfr_regular_p (mpc_imagref (x)) && mpfr_regular_p (mpc_realref (y)) && mpfr_regular_p (mpc_imagref (y))); overlap = (z == x) || (z == y); if (overlap) mpc_init3 (rop, MPC_PREC_RE (z), MPC_PREC_IM (z)); else rop [0] = z [0]; inex_re = mpfr_fmms (mpc_realref (rop), mpc_realref (x), mpc_realref (y), mpc_imagref (x), mpc_imagref (y), MPC_RND_RE (rnd)); inex_im = mpfr_fmma (mpc_imagref (rop), mpc_realref (x), mpc_imagref (y), mpc_imagref (x), mpc_realref (y), MPC_RND_IM (rnd)); mpc_set (z, rop, MPC_RNDNN); if (overlap) mpc_clear (rop); return MPC_INEX (inex_re, inex_im); } int mpc_mul_karatsuba (mpc_ptr rop, mpc_srcptr op1, mpc_srcptr op2, mpc_rnd_t rnd) { /* computes rop=op1*op2 by a Karatsuba algorithm, where op1 and op2 are assumed to be finite and without zero parts */ mpfr_srcptr a, b, c, d; int mul_i, ok, inexact, mul_a, mul_c, inex_re = 0, inex_im = 0, sign_x, sign_u; mpfr_t u, v, w, x; mpfr_prec_t prec, prec_re, prec_u, prec_v, prec_w; mpfr_rnd_t rnd_re, rnd_u; int overlap; /* true if rop == op1 or rop == op2 */ mpc_t result; /* overlap is quite difficult to handle, because we have to tentatively round the variable u in the end to either the real or the imaginary part of rop (it is not possible to tell now whether the real or imaginary part is used). If this fails, we have to start again and need the correct values of op1 and op2. So we just create a new variable for the result in this case. */ mpfr_ptr ref; int loop; const int MAX_MUL_LOOP = 1; overlap = (rop == op1) || (rop == op2); if (overlap) mpc_init3 (result, MPC_PREC_RE (rop), MPC_PREC_IM (rop)); else result [0] = rop [0]; a = mpc_realref(op1); b = mpc_imagref(op1); c = mpc_realref(op2); d = mpc_imagref(op2); /* (a + i*b) * (c + i*d) = [ac - bd] + i*[ad + bc] */ mul_i = 0; /* number of multiplications by i */ mul_a = 1; /* implicit factor for a */ mul_c = 1; /* implicit factor for c */ if (mpfr_cmp_abs (a, b) < 0) { MPFR_SWAP (a, b); mul_i ++; mul_a = -1; /* consider i * (a+i*b) = -b + i*a */ } if (mpfr_cmp_abs (c, d) < 0) { MPFR_SWAP (c, d); mul_i ++; mul_c = -1; /* consider -d + i*c instead of c + i*d */ } /* find the precision and rounding mode for the new real part */ if (mul_i % 2) { prec_re = MPC_PREC_IM(rop); rnd_re = MPC_RND_IM(rnd); } else /* mul_i = 0 or 2 */ { prec_re = MPC_PREC_RE(rop); rnd_re = MPC_RND_RE(rnd); } if (mul_i) rnd_re = INV_RND(rnd_re); /* now |a| >= |b| and |c| >= |d| */ prec = MPC_MAX_PREC(rop); mpfr_init2 (v, prec_v = mpfr_get_prec (a) + mpfr_get_prec (d)); mpfr_init2 (w, prec_w = mpfr_get_prec (b) + mpfr_get_prec (c)); mpfr_init2 (u, 2); mpfr_init2 (x, 2); inexact = mpfr_mul (v, a, d, MPFR_RNDN); if (inexact) { /* over- or underflow */ ok = 0; goto clear; } if (mul_a == -1) mpfr_neg (v, v, MPFR_RNDN); inexact = mpfr_mul (w, b, c, MPFR_RNDN); if (inexact) { /* over- or underflow */ ok = 0; goto clear; } if (mul_c == -1) mpfr_neg (w, w, MPFR_RNDN); /* compute sign(v-w) */ sign_x = mpfr_cmp_abs (v, w); if (sign_x > 0) sign_x = 2 * mpfr_sgn (v) - mpfr_sgn (w); else if (sign_x == 0) sign_x = mpfr_sgn (v) - mpfr_sgn (w); else sign_x = mpfr_sgn (v) - 2 * mpfr_sgn (w); sign_u = mul_a * mpfr_sgn (a) * mul_c * mpfr_sgn (c); if (sign_x * sign_u < 0) { /* swap inputs */ MPFR_SWAP (a, c); MPFR_SWAP (b, d); mpfr_swap (v, w); { int tmp; tmp = mul_a; mul_a = mul_c; mul_c = tmp; } sign_x = - sign_x; } /* now sign_x * sign_u >= 0 */ loop = 0; do { loop++; /* the following should give failures with prob. <= 1/prec */ prec += mpc_ceil_log2 (prec) + 3; mpfr_set_prec (u, prec_u = prec); mpfr_set_prec (x, prec); /* first compute away(b +/- a) and store it in u */ inexact = (mul_a == -1 ? mpfr_sub (u, b, a, MPFR_RNDA) : mpfr_add (u, b, a, MPFR_RNDA)); /* then compute away(+/-c - d) and store it in x */ inexact |= (mul_c == -1 ? mpfr_add (x, c, d, MPFR_RNDA) : mpfr_sub (x, c, d, MPFR_RNDA)); if (mul_c == -1) mpfr_neg (x, x, MPFR_RNDN); if (inexact == 0) mpfr_prec_round (u, prec_u = 2 * prec, MPFR_RNDN); /* compute away(u*x) and store it in u */ inexact |= mpfr_mul (u, u, x, MPFR_RNDA); /* (a+b)*(c-d) */ /* if all computations are exact up to here, it may be that the real part is exact, thus we need if possible to compute v - w exactly */ if (inexact == 0) { mpfr_prec_t prec_x; /* v and w are different from 0, so mpfr_get_exp is safe to use */ prec_x = SAFE_ABS (mpfr_exp_t, mpfr_get_exp (v) - mpfr_get_exp (w)) + MPC_MAX (prec_v, prec_w) + 1; /* +1 is necessary for a potential carry */ /* ensure we do not use a too large precision */ if (prec_x > prec_u) prec_x = prec_u; if (prec_x > prec) mpfr_prec_round (x, prec_x, MPFR_RNDN); } rnd_u = (sign_u > 0) ? MPFR_RNDU : MPFR_RNDD; inexact |= mpfr_sub (x, v, w, rnd_u); /* ad - bc */ /* in case u=0, ensure that rnd_u rounds x away from zero */ if (mpfr_sgn (u) == 0) rnd_u = (mpfr_sgn (x) > 0) ? MPFR_RNDU : MPFR_RNDD; inexact |= mpfr_add (u, u, x, rnd_u); /* ac - bd */ ok = inexact == 0 || mpfr_can_round (u, prec_u - 3, rnd_u, MPFR_RNDZ, prec_re + (rnd_re == MPFR_RNDN)); /* this ensures both we can round correctly and determine the correct inexact flag (for rounding to nearest) */ } while (!ok && loop <= MAX_MUL_LOOP); /* after MAX_MUL_LOOP rounds, use mpc_naive instead */ if (ok) { /* if inexact is zero, then u is exactly ac-bd, otherwise fix the sign of the inexact flag for u, which was rounded away from ac-bd */ if (inexact != 0) inexact = mpfr_sgn (u); if (mul_i == 0) { inex_re = mpfr_set (mpc_realref(result), u, MPC_RND_RE(rnd)); if (inex_re == 0) { inex_re = inexact; /* u is rounded away from 0 */ inex_im = mpfr_add (mpc_imagref(result), v, w, MPC_RND_IM(rnd)); } else inex_im = mpfr_add (mpc_imagref(result), v, w, MPC_RND_IM(rnd)); } else if (mul_i == 1) /* (x+i*y)/i = y - i*x */ { inex_im = mpfr_neg (mpc_imagref(result), u, MPC_RND_IM(rnd)); if (inex_im == 0) { inex_im = -inexact; /* u is rounded away from 0 */ inex_re = mpfr_add (mpc_realref(result), v, w, MPC_RND_RE(rnd)); } else inex_re = mpfr_add (mpc_realref(result), v, w, MPC_RND_RE(rnd)); } else /* mul_i = 2, z/i^2 = -z */ { inex_re = mpfr_neg (mpc_realref(result), u, MPC_RND_RE(rnd)); if (inex_re == 0) { inex_re = -inexact; /* u is rounded away from 0 */ inex_im = -mpfr_add (mpc_imagref(result), v, w, INV_RND(MPC_RND_IM(rnd))); mpfr_neg (mpc_imagref(result), mpc_imagref(result), MPC_RND_IM(rnd)); } else { inex_im = -mpfr_add (mpc_imagref(result), v, w, INV_RND(MPC_RND_IM(rnd))); mpfr_neg (mpc_imagref(result), mpc_imagref(result), MPC_RND_IM(rnd)); } } /* Clear potential signs of 0. */ if (!inex_re) { ref = mpc_realref (result); if (mpfr_zero_p (ref) && mpfr_signbit (ref)) MPFR_CHANGE_SIGN (ref); } if (!inex_im) { ref = mpc_imagref (result); if (mpfr_zero_p (ref) && mpfr_signbit (ref)) MPFR_CHANGE_SIGN (ref); } mpc_set (rop, result, MPC_RNDNN); } clear: mpfr_clear (u); mpfr_clear (v); mpfr_clear (w); mpfr_clear (x); if (overlap) mpc_clear (result); if (ok) return MPC_INEX(inex_re, inex_im); else return mpc_mul_naive (rop, op1, op2, rnd); } int mpc_mul (mpc_ptr a, mpc_srcptr b, mpc_srcptr c, mpc_rnd_t rnd) { /* Conforming to ISO C99 standard (G.5.1 multiplicative operators), infinities are treated specially if both parts are NaN when computed naively. */ if (mpc_inf_p (b)) return mul_infinite (a, b, c); if (mpc_inf_p (c)) return mul_infinite (a, c, b); /* NaN contamination of both parts in result */ if (mpfr_nan_p (mpc_realref (b)) || mpfr_nan_p (mpc_imagref (b)) || mpfr_nan_p (mpc_realref (c)) || mpfr_nan_p (mpc_imagref (c))) { mpfr_set_nan (mpc_realref (a)); mpfr_set_nan (mpc_imagref (a)); return MPC_INEX (0, 0); } /* check for real multiplication */ if (mpfr_zero_p (mpc_imagref (b))) return mul_real (a, c, b, rnd); if (mpfr_zero_p (mpc_imagref (c))) return mul_real (a, b, c, rnd); /* check for purely imaginary multiplication */ if (mpfr_zero_p (mpc_realref (b))) return mul_imag (a, c, b, rnd); if (mpfr_zero_p (mpc_realref (c))) return mul_imag (a, b, c, rnd); /* If the real and imaginary part of one argument have a very different */ /* exponent, it is not reasonable to use Karatsuba multiplication. */ if ( SAFE_ABS (mpfr_exp_t, mpfr_get_exp (mpc_realref (b)) - mpfr_get_exp (mpc_imagref (b))) > (mpfr_exp_t) MPC_MAX_PREC (b) / 2 || SAFE_ABS (mpfr_exp_t, mpfr_get_exp (mpc_realref (c)) - mpfr_get_exp (mpc_imagref (c))) > (mpfr_exp_t) MPC_MAX_PREC (c) / 2) return mpc_mul_naive (a, b, c, rnd); else return ((MPC_MAX_PREC(a) <= (mpfr_prec_t) MUL_KARATSUBA_THRESHOLD * BITS_PER_MP_LIMB) ? mpc_mul_naive : mpc_mul_karatsuba) (a, b, c, rnd); }