aboutsummaryrefslogtreecommitdiff
path: root/contrib/arm-optimized-routines/math/aarch64/experimental/sve/powi.c
blob: 62dd1b11497073c8c8c39448a1726cb6c310871e (plain) (blame)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
/*
 * Double-precision SVE powi(x, n) function.
 *
 * Copyright (c) 2020-2024, Arm Limited.
 * SPDX-License-Identifier: MIT OR Apache-2.0 WITH LLVM-exception
 */

#include "sv_math.h"

/* Optimized double-precision vector powi (double base, long integer power).
   powi is developed for environments in which accuracy is of much less
   importance than performance, hence we provide no estimate for worst-case
   error.  */
svfloat64_t
_ZGVsMxvv_powk (svfloat64_t as, svint64_t ns, svbool_t p)
{
  /* Compute powi by successive squaring, right to left.  */
  svfloat64_t acc = sv_f64 (1.0);
  svbool_t want_recip = svcmplt (p, ns, 0);
  svuint64_t ns_abs = svreinterpret_u64 (svabs_x (p, ns));

  /* We use a max to avoid needing to check whether any lane != 0 on each
     iteration.  */
  uint64_t max_n = svmaxv (p, ns_abs);

  svfloat64_t c = as;
  /* Successively square c, and use merging predication (_m) to determine
     whether or not to perform the multiplication or keep the previous
     iteration.  */
  while (true)
    {
      svbool_t px = svcmpeq (p, svand_x (p, ns_abs, 1ull), 1ull);
      acc = svmul_m (px, acc, c);
      max_n >>= 1;
      if (max_n == 0)
	break;

      ns_abs = svlsr_x (p, ns_abs, 1);
      c = svmul_x (p, c, c);
    }

  /* Negative powers are handled by computing the abs(n) version and then
     taking the reciprocal.  */
  if (svptest_any (want_recip, want_recip))
    acc = svdivr_m (want_recip, acc, 1.0);

  return acc;
}
CLOSE_SVE_ATTR