9ed2e15ff4
This patch continues cleaning up math_private.h by moving the math_opt_barrier and math_force_eval macros to a separate header math-barriers.h. At present, those macros are inside a "#ifndef math_opt_barrier" in math_private.h to allow architectures to override them and then use a separate math-barriers.h header, no such #ifndef or #include_next is needed; architectures just have their own alternative version of math-barriers.h when providing their own optimized versions that avoid going through memory unnecessarily. The generic math-barriers.h has a comment added to document these two macros. In this patch, math_private.h is made to #include <math-barriers.h>, so files using these macros do not need updating yet. That is because of uses of math_force_eval in math_check_force_underflow and math_check_force_underflow_nonneg, which are still defined in math_private.h. Once those are moved out to a separate header, that separate header can be made to include <math-barriers.h>, as can the other files directly using these barrier macros, and then the include of <math-barriers.h> from math_private.h can be removed. Tested for x86_64 and x86. Also tested with build-many-glibcs.py that installed stripped shared libraries are unchanged by this patch. * sysdeps/generic/math-barriers.h: New file. * sysdeps/generic/math_private.h [!math_opt_barrier] (math_opt_barrier): Move to math-barriers.h. [!math_opt_barrier] (math_force_eval): Likewise. * sysdeps/aarch64/fpu/math-barriers.h: New file. * sysdeps/aarch64/fpu/math_private.h (math_opt_barrier): Move to math-barriers.h. (math_force_eval): Likewise. * sysdeps/alpha/fpu/math-barriers.h: New file. * sysdeps/alpha/fpu/math_private.h (math_opt_barrier): Move to math-barriers.h. (math_force_eval): Likewise. * sysdeps/x86/fpu/math-barriers.h: New file. * sysdeps/i386/fpu/fenv_private.h (math_opt_barrier): Move to math-barriers.h. (math_force_eval): Likewise. * sysdeps/m68k/m680x0/fpu/math_private.h: Move to.... * sysdeps/m68k/m680x0/fpu/math-barriers.h: ... here. Adjust multiple-include guard for rename. * sysdeps/powerpc/fpu/math-barriers.h: New file. * sysdeps/powerpc/fpu/math_private.h (math_opt_barrier): Move to math-barriers.h. (math_force_eval): Likewise.
324 lines
8.7 KiB
C
324 lines
8.7 KiB
C
/* Private floating point rounding and exceptions handling. AArch64 version.
|
|
Copyright (C) 2014-2018 Free Software Foundation, Inc.
|
|
This file is part of the GNU C Library.
|
|
|
|
The GNU C Library is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU Lesser General Public
|
|
License as published by the Free Software Foundation; either
|
|
version 2.1 of the License, or (at your option) any later version.
|
|
|
|
The GNU C Library is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
Lesser General Public License for more details.
|
|
|
|
You should have received a copy of the GNU Lesser General Public
|
|
License along with the GNU C Library; if not, see
|
|
<http://www.gnu.org/licenses/>. */
|
|
|
|
#ifndef AARCH64_MATH_PRIVATE_H
|
|
#define AARCH64_MATH_PRIVATE_H 1
|
|
|
|
#include <fenv.h>
|
|
#include <fpu_control.h>
|
|
|
|
static __always_inline void
|
|
libc_feholdexcept_aarch64 (fenv_t *envp)
|
|
{
|
|
fpu_control_t fpcr;
|
|
fpu_control_t new_fpcr;
|
|
fpu_fpsr_t fpsr;
|
|
fpu_fpsr_t new_fpsr;
|
|
|
|
_FPU_GETCW (fpcr);
|
|
_FPU_GETFPSR (fpsr);
|
|
envp->__fpcr = fpcr;
|
|
envp->__fpsr = fpsr;
|
|
|
|
/* Clear exception flags and set all exceptions to non-stop. */
|
|
new_fpcr = fpcr & ~(FE_ALL_EXCEPT << FE_EXCEPT_SHIFT);
|
|
new_fpsr = fpsr & ~FE_ALL_EXCEPT;
|
|
|
|
if (__glibc_unlikely (new_fpcr != fpcr))
|
|
_FPU_SETCW (new_fpcr);
|
|
|
|
if (new_fpsr != fpsr)
|
|
_FPU_SETFPSR (new_fpsr);
|
|
}
|
|
|
|
#define libc_feholdexcept libc_feholdexcept_aarch64
|
|
#define libc_feholdexceptf libc_feholdexcept_aarch64
|
|
#define libc_feholdexceptl libc_feholdexcept_aarch64
|
|
|
|
static __always_inline void
|
|
libc_fesetround_aarch64 (int round)
|
|
{
|
|
fpu_control_t fpcr;
|
|
|
|
_FPU_GETCW (fpcr);
|
|
|
|
/* Check whether rounding modes are different. */
|
|
round = (fpcr ^ round) & _FPU_FPCR_RM_MASK;
|
|
|
|
/* Set new rounding mode if different. */
|
|
if (__glibc_unlikely (round != 0))
|
|
_FPU_SETCW (fpcr ^ round);
|
|
}
|
|
|
|
#define libc_fesetround libc_fesetround_aarch64
|
|
#define libc_fesetroundf libc_fesetround_aarch64
|
|
#define libc_fesetroundl libc_fesetround_aarch64
|
|
|
|
static __always_inline void
|
|
libc_feholdexcept_setround_aarch64 (fenv_t *envp, int round)
|
|
{
|
|
fpu_control_t fpcr;
|
|
fpu_control_t new_fpcr;
|
|
fpu_fpsr_t fpsr;
|
|
fpu_fpsr_t new_fpsr;
|
|
|
|
_FPU_GETCW (fpcr);
|
|
_FPU_GETFPSR (fpsr);
|
|
envp->__fpcr = fpcr;
|
|
envp->__fpsr = fpsr;
|
|
|
|
/* Clear exception flags, set all exceptions to non-stop,
|
|
and set new rounding mode. */
|
|
new_fpcr = fpcr & ~((FE_ALL_EXCEPT << FE_EXCEPT_SHIFT) | _FPU_FPCR_RM_MASK);
|
|
new_fpcr |= round;
|
|
new_fpsr = fpsr & ~FE_ALL_EXCEPT;
|
|
|
|
if (__glibc_unlikely (new_fpcr != fpcr))
|
|
_FPU_SETCW (new_fpcr);
|
|
|
|
if (new_fpsr != fpsr)
|
|
_FPU_SETFPSR (new_fpsr);
|
|
}
|
|
|
|
#define libc_feholdexcept_setround libc_feholdexcept_setround_aarch64
|
|
#define libc_feholdexcept_setroundf libc_feholdexcept_setround_aarch64
|
|
#define libc_feholdexcept_setroundl libc_feholdexcept_setround_aarch64
|
|
|
|
static __always_inline int
|
|
libc_fetestexcept_aarch64 (int ex)
|
|
{
|
|
fpu_fpsr_t fpsr;
|
|
|
|
_FPU_GETFPSR (fpsr);
|
|
return fpsr & ex & FE_ALL_EXCEPT;
|
|
}
|
|
|
|
#define libc_fetestexcept libc_fetestexcept_aarch64
|
|
#define libc_fetestexceptf libc_fetestexcept_aarch64
|
|
#define libc_fetestexceptl libc_fetestexcept_aarch64
|
|
|
|
static __always_inline void
|
|
libc_fesetenv_aarch64 (const fenv_t *envp)
|
|
{
|
|
fpu_control_t fpcr;
|
|
fpu_control_t new_fpcr;
|
|
|
|
_FPU_GETCW (fpcr);
|
|
new_fpcr = envp->__fpcr;
|
|
|
|
if (__glibc_unlikely (fpcr != new_fpcr))
|
|
_FPU_SETCW (new_fpcr);
|
|
|
|
_FPU_SETFPSR (envp->__fpsr);
|
|
}
|
|
|
|
#define libc_fesetenv libc_fesetenv_aarch64
|
|
#define libc_fesetenvf libc_fesetenv_aarch64
|
|
#define libc_fesetenvl libc_fesetenv_aarch64
|
|
#define libc_feresetround_noex libc_fesetenv_aarch64
|
|
#define libc_feresetround_noexf libc_fesetenv_aarch64
|
|
#define libc_feresetround_noexl libc_fesetenv_aarch64
|
|
|
|
static __always_inline int
|
|
libc_feupdateenv_test_aarch64 (const fenv_t *envp, int ex)
|
|
{
|
|
fpu_control_t fpcr;
|
|
fpu_control_t new_fpcr;
|
|
fpu_fpsr_t fpsr;
|
|
fpu_fpsr_t new_fpsr;
|
|
int excepts;
|
|
|
|
_FPU_GETCW (fpcr);
|
|
_FPU_GETFPSR (fpsr);
|
|
|
|
/* Merge current exception flags with the saved fenv. */
|
|
excepts = fpsr & FE_ALL_EXCEPT;
|
|
new_fpcr = envp->__fpcr;
|
|
new_fpsr = envp->__fpsr | excepts;
|
|
|
|
if (__glibc_unlikely (fpcr != new_fpcr))
|
|
_FPU_SETCW (new_fpcr);
|
|
|
|
if (fpsr != new_fpsr)
|
|
_FPU_SETFPSR (new_fpsr);
|
|
|
|
/* Raise the exceptions if enabled in the new FP state. */
|
|
if (__glibc_unlikely (excepts & (new_fpcr >> FE_EXCEPT_SHIFT)))
|
|
__feraiseexcept (excepts);
|
|
|
|
return excepts & ex;
|
|
}
|
|
|
|
#define libc_feupdateenv_test libc_feupdateenv_test_aarch64
|
|
#define libc_feupdateenv_testf libc_feupdateenv_test_aarch64
|
|
#define libc_feupdateenv_testl libc_feupdateenv_test_aarch64
|
|
|
|
static __always_inline void
|
|
libc_feupdateenv_aarch64 (const fenv_t *envp)
|
|
{
|
|
libc_feupdateenv_test_aarch64 (envp, 0);
|
|
}
|
|
|
|
#define libc_feupdateenv libc_feupdateenv_aarch64
|
|
#define libc_feupdateenvf libc_feupdateenv_aarch64
|
|
#define libc_feupdateenvl libc_feupdateenv_aarch64
|
|
|
|
static __always_inline void
|
|
libc_feholdsetround_aarch64 (fenv_t *envp, int round)
|
|
{
|
|
fpu_control_t fpcr;
|
|
fpu_fpsr_t fpsr;
|
|
|
|
_FPU_GETCW (fpcr);
|
|
_FPU_GETFPSR (fpsr);
|
|
envp->__fpcr = fpcr;
|
|
envp->__fpsr = fpsr;
|
|
|
|
/* Check whether rounding modes are different. */
|
|
round = (fpcr ^ round) & _FPU_FPCR_RM_MASK;
|
|
|
|
/* Set new rounding mode if different. */
|
|
if (__glibc_unlikely (round != 0))
|
|
_FPU_SETCW (fpcr ^ round);
|
|
}
|
|
|
|
#define libc_feholdsetround libc_feholdsetround_aarch64
|
|
#define libc_feholdsetroundf libc_feholdsetround_aarch64
|
|
#define libc_feholdsetroundl libc_feholdsetround_aarch64
|
|
|
|
static __always_inline void
|
|
libc_feresetround_aarch64 (fenv_t *envp)
|
|
{
|
|
fpu_control_t fpcr;
|
|
int round;
|
|
|
|
_FPU_GETCW (fpcr);
|
|
|
|
/* Check whether rounding modes are different. */
|
|
round = (envp->__fpcr ^ fpcr) & _FPU_FPCR_RM_MASK;
|
|
|
|
/* Restore the rounding mode if it was changed. */
|
|
if (__glibc_unlikely (round != 0))
|
|
_FPU_SETCW (fpcr ^ round);
|
|
}
|
|
|
|
#define libc_feresetround libc_feresetround_aarch64
|
|
#define libc_feresetroundf libc_feresetround_aarch64
|
|
#define libc_feresetroundl libc_feresetround_aarch64
|
|
|
|
/* We have support for rounding mode context. */
|
|
#define HAVE_RM_CTX 1
|
|
|
|
static __always_inline void
|
|
libc_feholdsetround_aarch64_ctx (struct rm_ctx *ctx, int r)
|
|
{
|
|
fpu_control_t fpcr;
|
|
int round;
|
|
|
|
_FPU_GETCW (fpcr);
|
|
ctx->env.__fpcr = fpcr;
|
|
|
|
/* Check whether rounding modes are different. */
|
|
round = (fpcr ^ r) & _FPU_FPCR_RM_MASK;
|
|
ctx->updated_status = round != 0;
|
|
|
|
/* Set the rounding mode if changed. */
|
|
if (__glibc_unlikely (round != 0))
|
|
_FPU_SETCW (fpcr ^ round);
|
|
}
|
|
|
|
#define libc_feholdsetround_ctx libc_feholdsetround_aarch64_ctx
|
|
#define libc_feholdsetroundf_ctx libc_feholdsetround_aarch64_ctx
|
|
#define libc_feholdsetroundl_ctx libc_feholdsetround_aarch64_ctx
|
|
|
|
static __always_inline void
|
|
libc_feresetround_aarch64_ctx (struct rm_ctx *ctx)
|
|
{
|
|
/* Restore the rounding mode if updated. */
|
|
if (__glibc_unlikely (ctx->updated_status))
|
|
_FPU_SETCW (ctx->env.__fpcr);
|
|
}
|
|
|
|
#define libc_feresetround_ctx libc_feresetround_aarch64_ctx
|
|
#define libc_feresetroundf_ctx libc_feresetround_aarch64_ctx
|
|
#define libc_feresetroundl_ctx libc_feresetround_aarch64_ctx
|
|
|
|
static __always_inline void
|
|
libc_feholdsetround_noex_aarch64_ctx (struct rm_ctx *ctx, int r)
|
|
{
|
|
fpu_control_t fpcr;
|
|
fpu_fpsr_t fpsr;
|
|
int round;
|
|
|
|
_FPU_GETCW (fpcr);
|
|
_FPU_GETFPSR (fpsr);
|
|
ctx->env.__fpcr = fpcr;
|
|
ctx->env.__fpsr = fpsr;
|
|
|
|
/* Check whether rounding modes are different. */
|
|
round = (fpcr ^ r) & _FPU_FPCR_RM_MASK;
|
|
ctx->updated_status = round != 0;
|
|
|
|
/* Set the rounding mode if changed. */
|
|
if (__glibc_unlikely (round != 0))
|
|
_FPU_SETCW (fpcr ^ round);
|
|
}
|
|
|
|
#define libc_feholdsetround_noex_ctx libc_feholdsetround_noex_aarch64_ctx
|
|
#define libc_feholdsetround_noexf_ctx libc_feholdsetround_noex_aarch64_ctx
|
|
#define libc_feholdsetround_noexl_ctx libc_feholdsetround_noex_aarch64_ctx
|
|
|
|
static __always_inline void
|
|
libc_feresetround_noex_aarch64_ctx (struct rm_ctx *ctx)
|
|
{
|
|
/* Restore the rounding mode if updated. */
|
|
if (__glibc_unlikely (ctx->updated_status))
|
|
_FPU_SETCW (ctx->env.__fpcr);
|
|
|
|
/* Write new FPSR to restore exception flags. */
|
|
_FPU_SETFPSR (ctx->env.__fpsr);
|
|
}
|
|
|
|
#define libc_feresetround_noex_ctx libc_feresetround_noex_aarch64_ctx
|
|
#define libc_feresetround_noexf_ctx libc_feresetround_noex_aarch64_ctx
|
|
#define libc_feresetround_noexl_ctx libc_feresetround_noex_aarch64_ctx
|
|
|
|
/* Hack: only include the large arm_neon.h when needed. */
|
|
#ifdef _MATH_CONFIG_H
|
|
# include <arm_neon.h>
|
|
|
|
/* ACLE intrinsics for frintn and fcvtns instructions. */
|
|
# define TOINT_INTRINSICS 1
|
|
|
|
static inline double_t
|
|
roundtoint (double_t x)
|
|
{
|
|
return vget_lane_f64 (vrndn_f64 (vld1_f64 (&x)), 0);
|
|
}
|
|
|
|
static inline uint64_t
|
|
converttoint (double_t x)
|
|
{
|
|
return vcvtnd_s64_f64 (x);
|
|
}
|
|
#endif
|
|
|
|
#include_next <math_private.h>
|
|
|
|
#endif
|