[v5,18/18] x86-64: Add vector asinh/asinhf implementation to libmvec

Message ID 20211229064000.1465621-19-skpgkp2@gmail.com
State Committed
Commit e682d0157854955e4b5fc91731d24a854e810eee
Headers
Series x86-64: Add vector math functions to libmvec |

Checks

Context Check Description
dj/TryBot-apply_patch success Patch applied to master at the time it was sent
dj/TryBot-32bit success Build for i686

Commit Message

Sunil Pandey Dec. 29, 2021, 6:40 a.m. UTC
  Implement vectorized asinh/asinhf containing SSE, AVX, AVX2 and
AVX512 versions for libmvec as per vector ABI.  It also contains
accuracy and ABI tests for vector asinh/asinhf with regenerated ulps.
---
 bits/libm-simd-decl-stubs.h                   |   11 +
 math/bits/mathcalls.h                         |    2 +-
 .../unix/sysv/linux/x86_64/libmvec.abilist    |    8 +
 sysdeps/x86/fpu/bits/math-vector.h            |    4 +
 .../x86/fpu/finclude/math-vector-fortran.h    |    4 +
 sysdeps/x86_64/fpu/Makeconfig                 |    1 +
 sysdeps/x86_64/fpu/Versions                   |    2 +
 sysdeps/x86_64/fpu/libm-test-ulps             |   17 +
 .../fpu/multiarch/svml_d_asinh2_core-sse2.S   |   20 +
 .../x86_64/fpu/multiarch/svml_d_asinh2_core.c |   27 +
 .../fpu/multiarch/svml_d_asinh2_core_sse4.S   | 1662 +++++++++++++++++
 .../fpu/multiarch/svml_d_asinh4_core-sse.S    |   20 +
 .../x86_64/fpu/multiarch/svml_d_asinh4_core.c |   27 +
 .../fpu/multiarch/svml_d_asinh4_core_avx2.S   | 1601 ++++++++++++++++
 .../fpu/multiarch/svml_d_asinh8_core-avx2.S   |   20 +
 .../x86_64/fpu/multiarch/svml_d_asinh8_core.c |   27 +
 .../fpu/multiarch/svml_d_asinh8_core_avx512.S |  510 +++++
 .../fpu/multiarch/svml_s_asinhf16_core-avx2.S |   20 +
 .../fpu/multiarch/svml_s_asinhf16_core.c      |   28 +
 .../multiarch/svml_s_asinhf16_core_avx512.S   |  476 +++++
 .../fpu/multiarch/svml_s_asinhf4_core-sse2.S  |   20 +
 .../fpu/multiarch/svml_s_asinhf4_core.c       |   28 +
 .../fpu/multiarch/svml_s_asinhf4_core_sse4.S  |  509 +++++
 .../fpu/multiarch/svml_s_asinhf8_core-sse.S   |   20 +
 .../fpu/multiarch/svml_s_asinhf8_core.c       |   28 +
 .../fpu/multiarch/svml_s_asinhf8_core_avx2.S  |  457 +++++
 sysdeps/x86_64/fpu/svml_d_asinh2_core.S       |   29 +
 sysdeps/x86_64/fpu/svml_d_asinh4_core.S       |   29 +
 sysdeps/x86_64/fpu/svml_d_asinh4_core_avx.S   |   25 +
 sysdeps/x86_64/fpu/svml_d_asinh8_core.S       |   25 +
 sysdeps/x86_64/fpu/svml_s_asinhf16_core.S     |   25 +
 sysdeps/x86_64/fpu/svml_s_asinhf4_core.S      |   29 +
 sysdeps/x86_64/fpu/svml_s_asinhf8_core.S      |   29 +
 sysdeps/x86_64/fpu/svml_s_asinhf8_core_avx.S  |   25 +
 .../fpu/test-double-libmvec-asinh-avx.c       |    1 +
 .../fpu/test-double-libmvec-asinh-avx2.c      |    1 +
 .../fpu/test-double-libmvec-asinh-avx512f.c   |    1 +
 .../x86_64/fpu/test-double-libmvec-asinh.c    |    3 +
 .../x86_64/fpu/test-double-vlen2-wrappers.c   |    1 +
 .../fpu/test-double-vlen4-avx2-wrappers.c     |    1 +
 .../x86_64/fpu/test-double-vlen4-wrappers.c   |    1 +
 .../x86_64/fpu/test-double-vlen8-wrappers.c   |    1 +
 .../fpu/test-float-libmvec-asinhf-avx.c       |    1 +
 .../fpu/test-float-libmvec-asinhf-avx2.c      |    1 +
 .../fpu/test-float-libmvec-asinhf-avx512f.c   |    1 +
 .../x86_64/fpu/test-float-libmvec-asinhf.c    |    3 +
 .../x86_64/fpu/test-float-vlen16-wrappers.c   |    1 +
 .../x86_64/fpu/test-float-vlen4-wrappers.c    |    1 +
 .../fpu/test-float-vlen8-avx2-wrappers.c      |    1 +
 .../x86_64/fpu/test-float-vlen8-wrappers.c    |    1 +
 50 files changed, 5784 insertions(+), 1 deletion(-)
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core-sse2.S
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core.c
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core_sse4.S
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core-sse.S
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core.c
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core_avx2.S
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core-avx2.S
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core.c
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core_avx512.S
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core-avx2.S
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core.c
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core_avx512.S
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core-sse2.S
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core.c
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core_sse4.S
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core-sse.S
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core.c
 create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core_avx2.S
 create mode 100644 sysdeps/x86_64/fpu/svml_d_asinh2_core.S
 create mode 100644 sysdeps/x86_64/fpu/svml_d_asinh4_core.S
 create mode 100644 sysdeps/x86_64/fpu/svml_d_asinh4_core_avx.S
 create mode 100644 sysdeps/x86_64/fpu/svml_d_asinh8_core.S
 create mode 100644 sysdeps/x86_64/fpu/svml_s_asinhf16_core.S
 create mode 100644 sysdeps/x86_64/fpu/svml_s_asinhf4_core.S
 create mode 100644 sysdeps/x86_64/fpu/svml_s_asinhf8_core.S
 create mode 100644 sysdeps/x86_64/fpu/svml_s_asinhf8_core_avx.S
 create mode 100644 sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx.c
 create mode 100644 sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx2.c
 create mode 100644 sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx512f.c
 create mode 100644 sysdeps/x86_64/fpu/test-double-libmvec-asinh.c
 create mode 100644 sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx.c
 create mode 100644 sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx2.c
 create mode 100644 sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx512f.c
 create mode 100644 sysdeps/x86_64/fpu/test-float-libmvec-asinhf.c
  

Comments

H.J. Lu Dec. 29, 2021, 9:27 p.m. UTC | #1
On Tue, Dec 28, 2021 at 10:40:00PM -0800, Sunil K Pandey wrote:
> Implement vectorized asinh/asinhf containing SSE, AVX, AVX2 and
> AVX512 versions for libmvec as per vector ABI.  It also contains
> accuracy and ABI tests for vector asinh/asinhf with regenerated ulps.
> ---
>  bits/libm-simd-decl-stubs.h                   |   11 +
>  math/bits/mathcalls.h                         |    2 +-
>  .../unix/sysv/linux/x86_64/libmvec.abilist    |    8 +
>  sysdeps/x86/fpu/bits/math-vector.h            |    4 +
>  .../x86/fpu/finclude/math-vector-fortran.h    |    4 +
>  sysdeps/x86_64/fpu/Makeconfig                 |    1 +
>  sysdeps/x86_64/fpu/Versions                   |    2 +
>  sysdeps/x86_64/fpu/libm-test-ulps             |   17 +
>  .../fpu/multiarch/svml_d_asinh2_core-sse2.S   |   20 +
>  .../x86_64/fpu/multiarch/svml_d_asinh2_core.c |   27 +
>  .../fpu/multiarch/svml_d_asinh2_core_sse4.S   | 1662 +++++++++++++++++
>  .../fpu/multiarch/svml_d_asinh4_core-sse.S    |   20 +
>  .../x86_64/fpu/multiarch/svml_d_asinh4_core.c |   27 +
>  .../fpu/multiarch/svml_d_asinh4_core_avx2.S   | 1601 ++++++++++++++++
>  .../fpu/multiarch/svml_d_asinh8_core-avx2.S   |   20 +
>  .../x86_64/fpu/multiarch/svml_d_asinh8_core.c |   27 +
>  .../fpu/multiarch/svml_d_asinh8_core_avx512.S |  510 +++++
>  .../fpu/multiarch/svml_s_asinhf16_core-avx2.S |   20 +
>  .../fpu/multiarch/svml_s_asinhf16_core.c      |   28 +
>  .../multiarch/svml_s_asinhf16_core_avx512.S   |  476 +++++
>  .../fpu/multiarch/svml_s_asinhf4_core-sse2.S  |   20 +
>  .../fpu/multiarch/svml_s_asinhf4_core.c       |   28 +
>  .../fpu/multiarch/svml_s_asinhf4_core_sse4.S  |  509 +++++
>  .../fpu/multiarch/svml_s_asinhf8_core-sse.S   |   20 +
>  .../fpu/multiarch/svml_s_asinhf8_core.c       |   28 +
>  .../fpu/multiarch/svml_s_asinhf8_core_avx2.S  |  457 +++++
>  sysdeps/x86_64/fpu/svml_d_asinh2_core.S       |   29 +
>  sysdeps/x86_64/fpu/svml_d_asinh4_core.S       |   29 +
>  sysdeps/x86_64/fpu/svml_d_asinh4_core_avx.S   |   25 +
>  sysdeps/x86_64/fpu/svml_d_asinh8_core.S       |   25 +
>  sysdeps/x86_64/fpu/svml_s_asinhf16_core.S     |   25 +
>  sysdeps/x86_64/fpu/svml_s_asinhf4_core.S      |   29 +
>  sysdeps/x86_64/fpu/svml_s_asinhf8_core.S      |   29 +
>  sysdeps/x86_64/fpu/svml_s_asinhf8_core_avx.S  |   25 +
>  .../fpu/test-double-libmvec-asinh-avx.c       |    1 +
>  .../fpu/test-double-libmvec-asinh-avx2.c      |    1 +
>  .../fpu/test-double-libmvec-asinh-avx512f.c   |    1 +
>  .../x86_64/fpu/test-double-libmvec-asinh.c    |    3 +
>  .../x86_64/fpu/test-double-vlen2-wrappers.c   |    1 +
>  .../fpu/test-double-vlen4-avx2-wrappers.c     |    1 +
>  .../x86_64/fpu/test-double-vlen4-wrappers.c   |    1 +
>  .../x86_64/fpu/test-double-vlen8-wrappers.c   |    1 +
>  .../fpu/test-float-libmvec-asinhf-avx.c       |    1 +
>  .../fpu/test-float-libmvec-asinhf-avx2.c      |    1 +
>  .../fpu/test-float-libmvec-asinhf-avx512f.c   |    1 +
>  .../x86_64/fpu/test-float-libmvec-asinhf.c    |    3 +
>  .../x86_64/fpu/test-float-vlen16-wrappers.c   |    1 +
>  .../x86_64/fpu/test-float-vlen4-wrappers.c    |    1 +
>  .../fpu/test-float-vlen8-avx2-wrappers.c      |    1 +
>  .../x86_64/fpu/test-float-vlen8-wrappers.c    |    1 +
>  50 files changed, 5784 insertions(+), 1 deletion(-)
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core-sse2.S
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core.c
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core_sse4.S
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core-sse.S
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core.c
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core_avx2.S
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core-avx2.S
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core.c
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core_avx512.S
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core-avx2.S
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core.c
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core_avx512.S
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core-sse2.S
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core.c
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core_sse4.S
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core-sse.S
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core.c
>  create mode 100644 sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core_avx2.S
>  create mode 100644 sysdeps/x86_64/fpu/svml_d_asinh2_core.S
>  create mode 100644 sysdeps/x86_64/fpu/svml_d_asinh4_core.S
>  create mode 100644 sysdeps/x86_64/fpu/svml_d_asinh4_core_avx.S
>  create mode 100644 sysdeps/x86_64/fpu/svml_d_asinh8_core.S
>  create mode 100644 sysdeps/x86_64/fpu/svml_s_asinhf16_core.S
>  create mode 100644 sysdeps/x86_64/fpu/svml_s_asinhf4_core.S
>  create mode 100644 sysdeps/x86_64/fpu/svml_s_asinhf8_core.S
>  create mode 100644 sysdeps/x86_64/fpu/svml_s_asinhf8_core_avx.S
>  create mode 100644 sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx.c
>  create mode 100644 sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx2.c
>  create mode 100644 sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx512f.c
>  create mode 100644 sysdeps/x86_64/fpu/test-double-libmvec-asinh.c
>  create mode 100644 sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx.c
>  create mode 100644 sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx2.c
>  create mode 100644 sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx512f.c
>  create mode 100644 sysdeps/x86_64/fpu/test-float-libmvec-asinhf.c
> 
> diff --git a/bits/libm-simd-decl-stubs.h b/bits/libm-simd-decl-stubs.h
> index 21f1a43232..bcaddb7a0e 100644
> --- a/bits/libm-simd-decl-stubs.h
> +++ b/bits/libm-simd-decl-stubs.h
> @@ -296,4 +296,15 @@
>  #define __DECL_SIMD_tanhf32x
>  #define __DECL_SIMD_tanhf64x
>  #define __DECL_SIMD_tanhf128x
> +
> +#define __DECL_SIMD_asinh
> +#define __DECL_SIMD_asinhf
> +#define __DECL_SIMD_asinhl
> +#define __DECL_SIMD_asinhf16
> +#define __DECL_SIMD_asinhf32
> +#define __DECL_SIMD_asinhf64
> +#define __DECL_SIMD_asinhf128
> +#define __DECL_SIMD_asinhf32x
> +#define __DECL_SIMD_asinhf64x
> +#define __DECL_SIMD_asinhf128x
>  #endif
> diff --git a/math/bits/mathcalls.h b/math/bits/mathcalls.h
> index 3d1c2056d5..40e055e579 100644
> --- a/math/bits/mathcalls.h
> +++ b/math/bits/mathcalls.h
> @@ -84,7 +84,7 @@ __MATHDECL_VEC (void,sincos,,
>  /* Hyperbolic arc cosine of X.  */
>  __MATHCALL_VEC (acosh,, (_Mdouble_ __x));
>  /* Hyperbolic arc sine of X.  */
> -__MATHCALL (asinh,, (_Mdouble_ __x));
> +__MATHCALL_VEC (asinh,, (_Mdouble_ __x));
>  /* Hyperbolic arc tangent of X.  */
>  __MATHCALL_VEC (atanh,, (_Mdouble_ __x));
>  #endif
> diff --git a/sysdeps/unix/sysv/linux/x86_64/libmvec.abilist b/sysdeps/unix/sysv/linux/x86_64/libmvec.abilist
> index e178cef683..df265d6a12 100644
> --- a/sysdeps/unix/sysv/linux/x86_64/libmvec.abilist
> +++ b/sysdeps/unix/sysv/linux/x86_64/libmvec.abilist
> @@ -49,6 +49,7 @@ GLIBC_2.22 _ZGVeN8vvv_sincos F
>  GLIBC_2.35 _ZGVbN2v_acos F
>  GLIBC_2.35 _ZGVbN2v_acosh F
>  GLIBC_2.35 _ZGVbN2v_asin F
> +GLIBC_2.35 _ZGVbN2v_asinh F
>  GLIBC_2.35 _ZGVbN2v_atan F
>  GLIBC_2.35 _ZGVbN2v_atanh F
>  GLIBC_2.35 _ZGVbN2v_cbrt F
> @@ -67,6 +68,7 @@ GLIBC_2.35 _ZGVbN2vv_hypot F
>  GLIBC_2.35 _ZGVbN4v_acosf F
>  GLIBC_2.35 _ZGVbN4v_acoshf F
>  GLIBC_2.35 _ZGVbN4v_asinf F
> +GLIBC_2.35 _ZGVbN4v_asinhf F
>  GLIBC_2.35 _ZGVbN4v_atanf F
>  GLIBC_2.35 _ZGVbN4v_atanhf F
>  GLIBC_2.35 _ZGVbN4v_cbrtf F
> @@ -85,6 +87,7 @@ GLIBC_2.35 _ZGVbN4vv_hypotf F
>  GLIBC_2.35 _ZGVcN4v_acos F
>  GLIBC_2.35 _ZGVcN4v_acosh F
>  GLIBC_2.35 _ZGVcN4v_asin F
> +GLIBC_2.35 _ZGVcN4v_asinh F
>  GLIBC_2.35 _ZGVcN4v_atan F
>  GLIBC_2.35 _ZGVcN4v_atanh F
>  GLIBC_2.35 _ZGVcN4v_cbrt F
> @@ -103,6 +106,7 @@ GLIBC_2.35 _ZGVcN4vv_hypot F
>  GLIBC_2.35 _ZGVcN8v_acosf F
>  GLIBC_2.35 _ZGVcN8v_acoshf F
>  GLIBC_2.35 _ZGVcN8v_asinf F
> +GLIBC_2.35 _ZGVcN8v_asinhf F
>  GLIBC_2.35 _ZGVcN8v_atanf F
>  GLIBC_2.35 _ZGVcN8v_atanhf F
>  GLIBC_2.35 _ZGVcN8v_cbrtf F
> @@ -121,6 +125,7 @@ GLIBC_2.35 _ZGVcN8vv_hypotf F
>  GLIBC_2.35 _ZGVdN4v_acos F
>  GLIBC_2.35 _ZGVdN4v_acosh F
>  GLIBC_2.35 _ZGVdN4v_asin F
> +GLIBC_2.35 _ZGVdN4v_asinh F
>  GLIBC_2.35 _ZGVdN4v_atan F
>  GLIBC_2.35 _ZGVdN4v_atanh F
>  GLIBC_2.35 _ZGVdN4v_cbrt F
> @@ -139,6 +144,7 @@ GLIBC_2.35 _ZGVdN4vv_hypot F
>  GLIBC_2.35 _ZGVdN8v_acosf F
>  GLIBC_2.35 _ZGVdN8v_acoshf F
>  GLIBC_2.35 _ZGVdN8v_asinf F
> +GLIBC_2.35 _ZGVdN8v_asinhf F
>  GLIBC_2.35 _ZGVdN8v_atanf F
>  GLIBC_2.35 _ZGVdN8v_atanhf F
>  GLIBC_2.35 _ZGVdN8v_cbrtf F
> @@ -157,6 +163,7 @@ GLIBC_2.35 _ZGVdN8vv_hypotf F
>  GLIBC_2.35 _ZGVeN16v_acosf F
>  GLIBC_2.35 _ZGVeN16v_acoshf F
>  GLIBC_2.35 _ZGVeN16v_asinf F
> +GLIBC_2.35 _ZGVeN16v_asinhf F
>  GLIBC_2.35 _ZGVeN16v_atanf F
>  GLIBC_2.35 _ZGVeN16v_atanhf F
>  GLIBC_2.35 _ZGVeN16v_cbrtf F
> @@ -175,6 +182,7 @@ GLIBC_2.35 _ZGVeN16vv_hypotf F
>  GLIBC_2.35 _ZGVeN8v_acos F
>  GLIBC_2.35 _ZGVeN8v_acosh F
>  GLIBC_2.35 _ZGVeN8v_asin F
> +GLIBC_2.35 _ZGVeN8v_asinh F
>  GLIBC_2.35 _ZGVeN8v_atan F
>  GLIBC_2.35 _ZGVeN8v_atanh F
>  GLIBC_2.35 _ZGVeN8v_cbrt F
> diff --git a/sysdeps/x86/fpu/bits/math-vector.h b/sysdeps/x86/fpu/bits/math-vector.h
> index 3c657f6108..71b7d660db 100644
> --- a/sysdeps/x86/fpu/bits/math-vector.h
> +++ b/sysdeps/x86/fpu/bits/math-vector.h
> @@ -130,6 +130,10 @@
>  #  define __DECL_SIMD_tanh __DECL_SIMD_x86_64
>  #  undef __DECL_SIMD_tanhf
>  #  define __DECL_SIMD_tanhf __DECL_SIMD_x86_64
> +#  undef __DECL_SIMD_asinh
> +#  define __DECL_SIMD_asinh __DECL_SIMD_x86_64
> +#  undef __DECL_SIMD_asinhf
> +#  define __DECL_SIMD_asinhf __DECL_SIMD_x86_64
>  
>  # endif
>  #endif
> diff --git a/sysdeps/x86/fpu/finclude/math-vector-fortran.h b/sysdeps/x86/fpu/finclude/math-vector-fortran.h
> index c7f81945fe..4d3afdf753 100644
> --- a/sysdeps/x86/fpu/finclude/math-vector-fortran.h
> +++ b/sysdeps/x86/fpu/finclude/math-vector-fortran.h
> @@ -64,6 +64,8 @@
>  !GCC$ builtin (erff) attributes simd (notinbranch) if('x86_64')
>  !GCC$ builtin (tanh) attributes simd (notinbranch) if('x86_64')
>  !GCC$ builtin (tanhf) attributes simd (notinbranch) if('x86_64')
> +!GCC$ builtin (asinh) attributes simd (notinbranch) if('x86_64')
> +!GCC$ builtin (asinhf) attributes simd (notinbranch) if('x86_64')
>  
>  !GCC$ builtin (cos) attributes simd (notinbranch) if('x32')
>  !GCC$ builtin (cosf) attributes simd (notinbranch) if('x32')
> @@ -113,3 +115,5 @@
>  !GCC$ builtin (erff) attributes simd (notinbranch) if('x32')
>  !GCC$ builtin (tanh) attributes simd (notinbranch) if('x32')
>  !GCC$ builtin (tanhf) attributes simd (notinbranch) if('x32')
> +!GCC$ builtin (asinh) attributes simd (notinbranch) if('x32')
> +!GCC$ builtin (asinhf) attributes simd (notinbranch) if('x32')
> diff --git a/sysdeps/x86_64/fpu/Makeconfig b/sysdeps/x86_64/fpu/Makeconfig
> index 26df8d47bf..2ff33c7dd8 100644
> --- a/sysdeps/x86_64/fpu/Makeconfig
> +++ b/sysdeps/x86_64/fpu/Makeconfig
> @@ -25,6 +25,7 @@ libmvec-funcs = \
>    acos \
>    acosh \
>    asin \
> +  asinh \
>    atan \
>    atan2 \
>    atanh \
> diff --git a/sysdeps/x86_64/fpu/Versions b/sysdeps/x86_64/fpu/Versions
> index adcbe0fefb..e6ead13085 100644
> --- a/sysdeps/x86_64/fpu/Versions
> +++ b/sysdeps/x86_64/fpu/Versions
> @@ -17,6 +17,7 @@ libmvec {
>      _ZGVbN2v_acos; _ZGVcN4v_acos; _ZGVdN4v_acos; _ZGVeN8v_acos;
>      _ZGVbN2v_acosh; _ZGVcN4v_acosh; _ZGVdN4v_acosh; _ZGVeN8v_acosh;
>      _ZGVbN2v_asin; _ZGVcN4v_asin; _ZGVdN4v_asin; _ZGVeN8v_asin;
> +    _ZGVbN2v_asinh; _ZGVcN4v_asinh; _ZGVdN4v_asinh; _ZGVeN8v_asinh;
>      _ZGVbN2v_atan; _ZGVcN4v_atan; _ZGVdN4v_atan; _ZGVeN8v_atan;
>      _ZGVbN2v_atanh; _ZGVcN4v_atanh; _ZGVdN4v_atanh; _ZGVeN8v_atanh;
>      _ZGVbN2v_cbrt; _ZGVcN4v_cbrt; _ZGVdN4v_cbrt; _ZGVeN8v_cbrt;
> @@ -35,6 +36,7 @@ libmvec {
>      _ZGVbN4v_acosf; _ZGVcN8v_acosf; _ZGVdN8v_acosf; _ZGVeN16v_acosf;
>      _ZGVbN4v_acoshf; _ZGVcN8v_acoshf; _ZGVdN8v_acoshf; _ZGVeN16v_acoshf;
>      _ZGVbN4v_asinf; _ZGVcN8v_asinf; _ZGVdN8v_asinf; _ZGVeN16v_asinf;
> +    _ZGVbN4v_asinhf; _ZGVcN8v_asinhf; _ZGVdN8v_asinhf; _ZGVeN16v_asinhf;
>      _ZGVbN4v_atanf; _ZGVcN8v_atanf; _ZGVdN8v_atanf; _ZGVeN16v_atanf;
>      _ZGVbN4v_atanhf; _ZGVcN8v_atanhf; _ZGVdN8v_atanhf; _ZGVeN16v_atanhf;
>      _ZGVbN4v_cbrtf; _ZGVcN8v_cbrtf; _ZGVdN8v_cbrtf; _ZGVeN16v_cbrtf;
> diff --git a/sysdeps/x86_64/fpu/libm-test-ulps b/sysdeps/x86_64/fpu/libm-test-ulps
> index bfaad7acef..71e9fced02 100644
> --- a/sysdeps/x86_64/fpu/libm-test-ulps
> +++ b/sysdeps/x86_64/fpu/libm-test-ulps
> @@ -157,6 +157,23 @@ float: 3
>  float128: 4
>  ldouble: 5
>  
> +Function: "asinh_vlen2":
> +double: 1
> +
> +Function: "asinh_vlen4":
> +double: 1
> +float: 1
> +
> +Function: "asinh_vlen4_avx2":
> +double: 1
> +
> +Function: "asinh_vlen8":
> +double: 1
> +float: 1
> +
> +Function: "asinh_vlen8_avx2":
> +float: 1
> +
>  Function: "atan":
>  double: 1
>  float: 1
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core-sse2.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core-sse2.S
> new file mode 100644
> index 0000000000..ddd1c3ca24
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core-sse2.S
> @@ -0,0 +1,20 @@
> +/* SSE2 version of vectorized asinh, vector length is 2.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#define _ZGVbN2v_asinh _ZGVbN2v_asinh_sse2
> +#include "../svml_d_asinh2_core.S"
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core.c b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core.c
> new file mode 100644
> index 0000000000..37452d0f92
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core.c
> @@ -0,0 +1,27 @@
> +/* Multiple versions of vectorized asinh, vector length is 2.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#define SYMBOL_NAME _ZGVbN2v_asinh
> +#include "ifunc-mathvec-sse4_1.h"
> +
> +libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
> +
> +#ifdef SHARED
> +__hidden_ver1 (_ZGVbN2v_asinh, __GI__ZGVbN2v_asinh, __redirect__ZGVbN2v_asinh)
> +  __attribute__ ((visibility ("hidden")));
> +#endif
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core_sse4.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core_sse4.S
> new file mode 100644
> index 0000000000..0fe130f20a
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core_sse4.S
> @@ -0,0 +1,1662 @@
> +/* Function asinh vectorized with SSE4.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   https://www.gnu.org/licenses/.  */
> +
> +/*
> + * ALGORITHM DESCRIPTION:
> + *
> + *   Compute asinh(x) as log(x + sqrt(x*x + 1))
> + *
> + *   Special cases:
> + *
> + *   asinh(NaN) = quiet NaN, and raise invalid exception
> + *   asinh(INF) = that INF
> + *   asinh(0)   = that 0
> + *
> + */
> +
> +/* Offsets for data table __svml_dasinh_data_internal
> + */
> +#define Log_HA_table                  	0
> +#define Log_LA_table                  	8208
> +#define poly_coeff                    	12320
> +#define ExpMask                       	12384
> +#define Two10                         	12400
> +#define MinLog1p                      	12416
> +#define MaxLog1p                      	12432
> +#define One                           	12448
> +#define SgnMask                       	12464
> +#define XThreshold                    	12480
> +#define XhMask                        	12496
> +#define Threshold                     	12512
> +#define Bias                          	12528
> +#define Bias1                         	12544
> +#define ExpMask0                      	12560
> +#define ExpMask2                      	12576
> +#define L2                            	12592
> +#define dBigThreshold                 	12608
> +#define dC2                           	12624
> +#define dC3                           	12640
> +#define dC4                           	12656
> +#define dC5                           	12672
> +#define dHalf                         	12688
> +#define dLargestFinite                	12704
> +#define dLittleThreshold              	12720
> +#define dSign                         	12736
> +#define dThirtyOne                    	12752
> +#define dTopMask12                    	12768
> +#define dTopMask26                    	12784
> +#define dTopMask29                    	12800
> +#define XScale                        	12816
> +
> +/* Lookup bias for data table __svml_dasinh_data_internal.  */
> +#define Table_Lookup_Bias               -0x405ff0
> +
> +#include <sysdep.h>
> +
> +        .text
> +	.section .text.sse4,"ax",@progbits
> +ENTRY(_ZGVbN2v_asinh_sse4)
> +        pushq     %rbp
> +        cfi_def_cfa_offset(16)
> +        movq      %rsp, %rbp
> +        cfi_def_cfa(6, 16)
> +        cfi_offset(6, -16)
> +        andq      $-32, %rsp
> +        subq      $64, %rsp
> +        movaps    %xmm0, %xmm13
> +
> +/*
> + * Split X into high and low parts, XHi (<= 26 bits) and XLo (<= 27 bits)
> + * We could use either X or |X| here, but it doesn't seem to matter
> + */
> +        movups    dTopMask26+__svml_dasinh_data_internal(%rip), %xmm15
> +        movaps    %xmm13, %xmm7
> +        andps     %xmm13, %xmm15
> +        lea       Table_Lookup_Bias+__svml_dasinh_data_internal(%rip), %rsi
> +
> +/*
> + * Compute X^2 = (XHi + XLo)^2 = XHi^2 + XLo * (X + XHi)
> + * The two parts are shifted off by around 26 bits. So even though
> + * the low bit will not in general be exact, it's near enough
> + */
> +        movaps    %xmm15, %xmm8
> +        mulpd     %xmm15, %xmm8
> +        subpd     %xmm15, %xmm7
> +        addpd     %xmm13, %xmm15
> +
> +/* Load the constant 1 and a sign mask */
> +        movups    One+__svml_dasinh_data_internal(%rip), %xmm12
> +
> +/*
> + * Finally, express Y + W = X^2 + 1 accurately where Y has <= 29 bits.
> + * If |X| <= 1 then |XHi| <= 1 and so |X2Hi| <= 1, so we can treat 1
> + * as the dominant component in the compensated summation. Otherwise,
> + * if |X| >= 1, then since X2Hi only has 52 significant bits, the basic
> + * addition will be exact anyway until we get to |X| >= 2^53. But by
> + * that time the log function is well-conditioned enough that the
> + * rounding error doesn't matter. Hence we can treat 1 as dominant even
> + * if it literally isn't.
> + */
> +        movaps    %xmm12, %xmm3
> +        movaps    %xmm12, %xmm5
> +        addpd     %xmm8, %xmm3
> +        mulpd     %xmm15, %xmm7
> +        subpd     %xmm3, %xmm5
> +        movups    dTopMask29+__svml_dasinh_data_internal(%rip), %xmm6
> +        andps     %xmm3, %xmm6
> +
> +/*
> + * Compute R = 1/sqrt(Y + W) * (1 + d)
> + * Force R to <= 12 significant bits in case it isn't already
> + * This means that R * Y and R^2 * Y are exactly representable.
> + */
> +        cvtpd2ps  %xmm6, %xmm1
> +        addpd     %xmm8, %xmm5
> +        subpd     %xmm6, %xmm3
> +
> +/*
> + * Unfortunately, we can still be in trouble if |X| <= 2^-10, since
> + * the absolute error 2^-(12+53)-ish in sqrt(1 + X^2) gets scaled up
> + * by 1/X and comes close to our threshold. Hence if |X| <= 2^-9,
> + * perform an alternative computation
> + * sqrt(1 + X^2) - 1 = X^2/2 - X^4/8 + X^6/16
> + * X2 = X^2
> + */
> +        addpd     %xmm7, %xmm8
> +        addpd     %xmm7, %xmm5
> +        movlhps   %xmm1, %xmm1
> +        rsqrtps   %xmm1, %xmm4
> +        addpd     %xmm3, %xmm5
> +        cvtps2pd  %xmm4, %xmm2
> +        andps     dTopMask12+__svml_dasinh_data_internal(%rip), %xmm2
> +
> +/*
> + * Compute e = -(2 * d + d^2)
> + * The first FMR is exact, and the rounding error in the other is acceptable
> + * since d and e are ~ 2^-12
> + */
> +        movaps    %xmm12, %xmm1
> +
> +/*
> + * Compute S = (Y/sqrt(Y + W)) * (1 + d)
> + * and T = (W/sqrt(Y + W)) * (1 + d)
> + * so that S + T = sqrt(Y + W) * (1 + d)
> + * S is exact, and the rounding error in T is OK.
> + */
> +        mulpd     %xmm2, %xmm6
> +        mulpd     %xmm2, %xmm5
> +        movaps    %xmm2, %xmm0
> +
> +/*
> + * Obtain sqrt(1 + X^2) - 1 in two pieces
> + * sqrt(1 + X^2) - 1
> + * = sqrt(Y + W) - 1
> + * = (S + T) * (1 + Corr) - 1
> + * = [S - 1] + [T + (S + T) * Corr]
> + * We need a compensated summation for the last part. We treat S - 1
> + * as the larger part; it certainly is until about X < 2^-4, and in that
> + * case, the error is affordable since X dominates over sqrt(1 + X^2) - 1
> + * Final sum is dTmp5 (hi) + dTmp7 (lo)
> + */
> +        movaps    %xmm6, %xmm3
> +        mulpd     %xmm6, %xmm0
> +        mulpd     %xmm5, %xmm2
> +        subpd     %xmm0, %xmm1
> +        addpd     %xmm5, %xmm3
> +        subpd     %xmm12, %xmm6
> +        subpd     %xmm2, %xmm1
> +        movups    SgnMask+__svml_dasinh_data_internal(%rip), %xmm9
> +        movaps    %xmm12, %xmm4
> +
> +/*
> + * Get the absolute value of the input, since we will exploit antisymmetry
> + * and mostly assume X >= 0 in the core computation
> + */
> +        movaps    %xmm9, %xmm10
> +        andps     %xmm13, %xmm10
> +
> +/*
> + * Check whether the input is finite, by checking |X| <= MaxFloat
> + * Otherwise set the rangemask so that the callout will get used.
> + * Note that this will also use the callout for NaNs since not(NaN <= MaxFloat)
> + */
> +        movaps    %xmm10, %xmm14
> +
> +/*
> + * The following computation can go wrong for very large X, basically
> + * because X^2 overflows. But for large X we have
> + * asinh(X) / log(2 X) - 1 =~= 1/(4 * X^2), so for X >= 2^30
> + * we can just later stick X back into the log and tweak up the exponent.
> + * Actually we scale X by 2^-30 and tweak the exponent up by 31,
> + * to stay in the safe range for the later log computation.
> + * Compute a flag now telling us when do do this.
> + */
> +        movaps    %xmm10, %xmm11
> +        cmpnlepd  dLargestFinite+__svml_dasinh_data_internal(%rip), %xmm14
> +        cmpltpd   dBigThreshold+__svml_dasinh_data_internal(%rip), %xmm11
> +        movmskpd  %xmm14, %edx
> +
> +/*
> + * Now       1 / (1 + d)
> + * = 1 / (1 + (sqrt(1 - e) - 1))
> + * = 1 / sqrt(1 - e)
> + * = 1 + 1/2 * e + 3/8 * e^2 + 5/16 * e^3 + 35/128 * e^4 +
> + * 63/256 * e^5 + 231/1024 * e^6 + ....
> + * So compute the first five nonconstant terms of that, so that
> + * we have a relative correction (1 + Corr) to apply to S etc.
> + * C1 = 1/2
> + * C2 = 3/8
> + * C3 = 5/16
> + * C4 = 35/128
> + * C5 = 63/256
> + */
> +        movups    dC5+__svml_dasinh_data_internal(%rip), %xmm14
> +        movups    dHalf+__svml_dasinh_data_internal(%rip), %xmm15
> +        mulpd     %xmm1, %xmm14
> +
> +/* dX2over2 = X^2/2 */
> +        mulpd     %xmm15, %xmm8
> +        addpd     dC4+__svml_dasinh_data_internal(%rip), %xmm14
> +        mulpd     %xmm1, %xmm14
> +        addpd     dC3+__svml_dasinh_data_internal(%rip), %xmm14
> +        mulpd     %xmm1, %xmm14
> +        addpd     dC2+__svml_dasinh_data_internal(%rip), %xmm14
> +        mulpd     %xmm1, %xmm14
> +        addpd     %xmm15, %xmm14
> +        mulpd     %xmm14, %xmm1
> +        mulpd     %xmm3, %xmm1
> +        addpd     %xmm1, %xmm5
> +        addpd     %xmm6, %xmm5
> +
> +/* dX4over4 = X^4/4 */
> +        movaps    %xmm8, %xmm6
> +
> +/* dX46 = -X^4/4 + X^6/8 */
> +        movaps    %xmm8, %xmm7
> +        mulpd     %xmm8, %xmm6
> +        mulpd     %xmm6, %xmm7
> +        subpd     %xmm6, %xmm7
> +
> +/* dX46over2 = -X^4/8 + x^6/16 */
> +        mulpd     %xmm7, %xmm15
> +
> +/* Now multiplex the two possible computations */
> +        movaps    %xmm10, %xmm3
> +        cmplepd   dLittleThreshold+__svml_dasinh_data_internal(%rip), %xmm3
> +        addpd     %xmm15, %xmm8
> +        movaps    %xmm3, %xmm1
> +        andps     %xmm3, %xmm8
> +        andnps    %xmm5, %xmm1
> +
> +/*
> + * Now we feed into the log1p code, using H in place of _VARG1 and
> + * also adding L into Xl.
> + * compute 1+x as high, low parts
> + */
> +        movaps    %xmm12, %xmm5
> +        orps      %xmm8, %xmm1
> +        movaps    %xmm11, %xmm3
> +
> +/*
> + * Now do another compensated sum to add |X| + [sqrt(1 + X^2) - 1].
> + * It's always safe to assume |X| is larger.
> + * This is the final 2-part argument to the log1p function
> + */
> +        addpd     %xmm10, %xmm1
> +        maxpd     %xmm1, %xmm5
> +        minpd     %xmm1, %xmm4
> +
> +/* Now multiplex to the case X = 2^-30 * |input|, Xl = dL = 0 in the "big" case. */
> +        movups    XScale+__svml_dasinh_data_internal(%rip), %xmm8
> +        andps     %xmm9, %xmm1
> +        mulpd     %xmm8, %xmm10
> +        cmpltpd   XThreshold+__svml_dasinh_data_internal(%rip), %xmm1
> +        movaps    %xmm5, %xmm9
> +        andnps    %xmm10, %xmm3
> +        addpd     %xmm4, %xmm9
> +        orps      XhMask+__svml_dasinh_data_internal(%rip), %xmm1
> +        andps     %xmm1, %xmm9
> +        subpd     %xmm9, %xmm5
> +        andps     %xmm11, %xmm9
> +
> +/* Now resume the main code. */
> +        movups    ExpMask+__svml_dasinh_data_internal(%rip), %xmm10
> +        orps      %xmm9, %xmm3
> +
> +/* preserve mantissa, set input exponent to 2^(-10) */
> +        andps     %xmm3, %xmm10
> +
> +/* exponent bits */
> +        movaps    %xmm3, %xmm7
> +        orps      Two10+__svml_dasinh_data_internal(%rip), %xmm10
> +        psrlq     $20, %xmm7
> +
> +/* reciprocal approximation good to at least 11 bits */
> +        cvtpd2ps  %xmm10, %xmm1
> +        addpd     %xmm5, %xmm4
> +        movlhps   %xmm1, %xmm1
> +        andps     %xmm11, %xmm4
> +        rcpps     %xmm1, %xmm0
> +        cvtps2pd  %xmm0, %xmm0
> +
> +/* round reciprocal to nearest integer, will have 1+9 mantissa bits */
> +        movups    .FLT_30(%rip), %xmm6
> +        movaps    %xmm11, %xmm1
> +        addpd     %xmm6, %xmm0
> +        subpd     %xmm6, %xmm0
> +
> +/* exponent of X needed to scale Xl */
> +        movdqu    ExpMask0+__svml_dasinh_data_internal(%rip), %xmm5
> +
> +/* 2^ (-10-exp(X) ) */
> +        movdqu    ExpMask2+__svml_dasinh_data_internal(%rip), %xmm2
> +        pand      %xmm3, %xmm5
> +        psubq     %xmm5, %xmm2
> +
> +/* scale DblRcp */
> +        mulpd     %xmm0, %xmm2
> +
> +/* argument reduction */
> +        mulpd     %xmm2, %xmm3
> +        mulpd     %xmm2, %xmm4
> +        subpd     %xmm12, %xmm3
> +        addpd     %xmm4, %xmm3
> +
> +/* polynomial */
> +        movups    poly_coeff+__svml_dasinh_data_internal(%rip), %xmm12
> +        movaps    %xmm3, %xmm2
> +        pshufd    $221, %xmm7, %xmm8
> +        mulpd     %xmm3, %xmm12
> +
> +/* biased exponent in DP format */
> +        cvtdq2pd  %xmm8, %xmm14
> +        addpd     poly_coeff+16+__svml_dasinh_data_internal(%rip), %xmm12
> +        mulpd     %xmm3, %xmm2
> +
> +/* Add 31 to the exponent in the "large" case to get log(2 * input) */
> +        movups    dThirtyOne+__svml_dasinh_data_internal(%rip), %xmm9
> +
> +/* exponent*log(2.0) */
> +        movups    Threshold+__svml_dasinh_data_internal(%rip), %xmm5
> +        addpd     %xmm14, %xmm9
> +        cmpltpd   %xmm0, %xmm5
> +        mulpd     %xmm2, %xmm12
> +        andps     %xmm11, %xmm14
> +
> +/*
> + * prepare table index
> + * table lookup
> + */
> +        movaps    %xmm0, %xmm11
> +        movups    poly_coeff+32+__svml_dasinh_data_internal(%rip), %xmm0
> +        andnps    %xmm9, %xmm1
> +        mulpd     %xmm3, %xmm0
> +        addpd     poly_coeff+48+__svml_dasinh_data_internal(%rip), %xmm0
> +        addpd     %xmm12, %xmm0
> +
> +/* reconstruction */
> +        mulpd     %xmm0, %xmm2
> +        andps     Bias+__svml_dasinh_data_internal(%rip), %xmm5
> +        psrlq     $40, %xmm11
> +        orps      Bias1+__svml_dasinh_data_internal(%rip), %xmm5
> +        orps      %xmm14, %xmm1
> +        movd      %xmm11, %eax
> +        pshufd    $2, %xmm11, %xmm11
> +
> +/* Finally, reincorporate the original sign. */
> +        movups    dSign+__svml_dasinh_data_internal(%rip), %xmm0
> +        subpd     %xmm5, %xmm1
> +        addpd     %xmm2, %xmm3
> +        movd      %xmm11, %ecx
> +        mulpd     L2+__svml_dasinh_data_internal(%rip), %xmm1
> +        movslq    %eax, %rax
> +        andps     %xmm13, %xmm0
> +        movslq    %ecx, %rcx
> +        movsd     (%rsi,%rax), %xmm6
> +        movhpd    (%rsi,%rcx), %xmm6
> +        addpd     %xmm3, %xmm6
> +        addpd     %xmm6, %xmm1
> +        pxor      %xmm1, %xmm0
> +        testl     %edx, %edx
> +
> +/* Go to special inputs processing branch */
> +        jne       L(SPECIAL_VALUES_BRANCH)
> +                                # LOE rbx r12 r13 r14 r15 edx xmm0 xmm13
> +
> +/* Restore registers
> + * and exit the function
> + */
> +
> +L(EXIT):
> +        movq      %rbp, %rsp
> +        popq      %rbp
> +        cfi_def_cfa(7, 8)
> +        cfi_restore(6)
> +        ret
> +        cfi_def_cfa(6, 16)
> +        cfi_offset(6, -16)
> +
> +/* Branch to process
> + * special inputs
> + */
> +
> +L(SPECIAL_VALUES_BRANCH):
> +        movups    %xmm13, 32(%rsp)
> +        movups    %xmm0, 48(%rsp)
> +                                # LOE rbx r12 r13 r14 r15 edx
> +
> +        xorl      %eax, %eax
> +        movq      %r12, 16(%rsp)
> +        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -48; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xd0, 0xff, 0xff, 0xff, 0x22
> +        movl      %eax, %r12d
> +        movq      %r13, 8(%rsp)
> +        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -56; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xc8, 0xff, 0xff, 0xff, 0x22
> +        movl      %edx, %r13d
> +        movq      %r14, (%rsp)
> +        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -64; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x22
> +                                # LOE rbx r15 r12d r13d
> +
> +/* Range mask
> + * bits check
> + */
> +
> +L(RANGEMASK_CHECK):
> +        btl       %r12d, %r13d
> +
> +/* Call scalar math function */
> +        jc        L(SCALAR_MATH_CALL)
> +                                # LOE rbx r15 r12d r13d
> +
> +/* Special inputs
> + * processing loop
> + */
> +
> +L(SPECIAL_VALUES_LOOP):
> +        incl      %r12d
> +        cmpl      $2, %r12d
> +
> +/* Check bits in range mask */
> +        jl        L(RANGEMASK_CHECK)
> +                                # LOE rbx r15 r12d r13d
> +
> +        movq      16(%rsp), %r12
> +        cfi_restore(12)
> +        movq      8(%rsp), %r13
> +        cfi_restore(13)
> +        movq      (%rsp), %r14
> +        cfi_restore(14)
> +        movups    48(%rsp), %xmm0
> +
> +/* Go to exit */
> +        jmp       L(EXIT)
> +        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -48; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xd0, 0xff, 0xff, 0xff, 0x22
> +        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -56; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xc8, 0xff, 0xff, 0xff, 0x22
> +        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -64; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x22
> +                                # LOE rbx r12 r13 r14 r15 xmm0
> +
> +/* Scalar math fucntion call
> + * to process special input
> + */
> +
> +L(SCALAR_MATH_CALL):
> +        movl      %r12d, %r14d
> +        movsd     32(%rsp,%r14,8), %xmm0
> +        call      asinh@PLT
> +                                # LOE rbx r14 r15 r12d r13d xmm0
> +
> +        movsd     %xmm0, 48(%rsp,%r14,8)
> +
> +/* Process special inputs in loop */
> +        jmp       L(SPECIAL_VALUES_LOOP)
> +                                # LOE rbx r15 r12d r13d
> +END(_ZGVbN2v_asinh_sse4)
> +
> +        .section .rodata, "a"
> +        .align 16
> +
> +#ifdef __svml_dasinh_data_internal_typedef
> +typedef unsigned int VUINT32;
> +typedef struct {
> +        __declspec(align(16)) VUINT32 Log_HA_table[(1<<10)+2][2];
> +        __declspec(align(16)) VUINT32 Log_LA_table[(1<<9)+1][2];
> +        __declspec(align(16)) VUINT32 poly_coeff[4][2][2];
> +        __declspec(align(16)) VUINT32 ExpMask[2][2];
> +        __declspec(align(16)) VUINT32 Two10[2][2];
> +        __declspec(align(16)) VUINT32 MinLog1p[2][2];
> +        __declspec(align(16)) VUINT32 MaxLog1p[2][2];
> +        __declspec(align(16)) VUINT32 One[2][2];
> +        __declspec(align(16)) VUINT32 SgnMask[2][2];
> +        __declspec(align(16)) VUINT32 XThreshold[2][2];
> +        __declspec(align(16)) VUINT32 XhMask[2][2];
> +        __declspec(align(16)) VUINT32 Threshold[2][2];
> +        __declspec(align(16)) VUINT32 Bias[2][2];
> +        __declspec(align(16)) VUINT32 Bias1[2][2];
> +        __declspec(align(16)) VUINT32 ExpMask0[2][2];
> +        __declspec(align(16)) VUINT32 ExpMask2[2][2];
> +        __declspec(align(16)) VUINT32 L2[2][2];
> +        __declspec(align(16)) VUINT32 dBigThreshold[2][2];
> +        __declspec(align(16)) VUINT32 dC2[2][2];
> +        __declspec(align(16)) VUINT32 dC3[2][2];
> +        __declspec(align(16)) VUINT32 dC4[2][2];
> +        __declspec(align(16)) VUINT32 dC5[2][2];
> +        __declspec(align(16)) VUINT32 dHalf[2][2];
> +        __declspec(align(16)) VUINT32 dLargestFinite[2][2];
> +        __declspec(align(16)) VUINT32 dLittleThreshold[2][2];
> +        __declspec(align(16)) VUINT32 dSign[2][2];
> +        __declspec(align(16)) VUINT32 dThirtyOne[2][2];
> +        __declspec(align(16)) VUINT32 dTopMask12[2][2];
> +        __declspec(align(16)) VUINT32 dTopMask26[2][2];
> +        __declspec(align(16)) VUINT32 dTopMask29[2][2];
> +        __declspec(align(16)) VUINT32 XScale[2][2];
> +} __svml_dasinh_data_internal;
> +#endif
> +__svml_dasinh_data_internal:
> +        /* Log_HA_table */
> +        .quad 0xc086232bdd7a8300, 0xbe1ce91eef3fb100
> +        .quad 0xc086232fdc7ad828, 0xbe1cefcffda73b6a
> +        .quad 0xc0862333d97d2ba0, 0xbe1cef406748f1ff
> +        .quad 0xc0862337d48378e0, 0xbe1cef2a9429925a
> +        .quad 0xc086233bcd8fb878, 0xbe1cf138d17ebecb
> +        .quad 0xc086233fc4a3e018, 0xbe1ceff2dbbbb29e
> +        .quad 0xc0862343b9c1e270, 0xbe1cf1a42aae437b
> +        .quad 0xc0862347acebaf68, 0xbe1cef3b152048af
> +        .quad 0xc086234b9e2333f0, 0xbe1cef20e127805e
> +        .quad 0xc086234f8d6a5a30, 0xbe1cf00ad6052cf4
> +        .quad 0xc08623537ac30980, 0xbe1cefc4642ee597
> +        .quad 0xc0862357662f2660, 0xbe1cf1f277d36e16
> +        .quad 0xc086235b4fb092a0, 0xbe1ceed009e8d8e6
> +        .quad 0xc086235f37492d28, 0xbe1cf1e4038cb362
> +        .quad 0xc08623631cfad250, 0xbe1cf0b0873b8557
> +        .quad 0xc086236700c75b98, 0xbe1cf15bb3227c0b
> +        .quad 0xc086236ae2b09fe0, 0xbe1cf151ef8ca9ed
> +        .quad 0xc086236ec2b87358, 0xbe1cefe1dc2cd2ed
> +        .quad 0xc0862372a0e0a780, 0xbe1cf0d1eec5454f
> +        .quad 0xc08623767d2b0b48, 0xbe1ceeefd570bbce
> +        .quad 0xc086237a57996af0, 0xbe1cee99ae91b3a7
> +        .quad 0xc086237e302d9028, 0xbe1cf0412830fbd1
> +        .quad 0xc086238206e94218, 0xbe1ceee898588610
> +        .quad 0xc0862385dbce4548, 0xbe1cee9a1fbcaaea
> +        .quad 0xc0862389aede5bc0, 0xbe1ceed8e7cc1ad6
> +        .quad 0xc086238d801b4500, 0xbe1cf10c8d059da6
> +        .quad 0xc08623914f86be18, 0xbe1ceee6c63a8165
> +        .quad 0xc08623951d228180, 0xbe1cf0c3592d2ff1
> +        .quad 0xc0862398e8f04758, 0xbe1cf0026cc4cb1b
> +        .quad 0xc086239cb2f1c538, 0xbe1cf15d48d8e670
> +        .quad 0xc08623a07b28ae60, 0xbe1cef359363787c
> +        .quad 0xc08623a44196b390, 0xbe1cefdf1ab2e82c
> +        .quad 0xc08623a8063d8338, 0xbe1cefe43c02aa84
> +        .quad 0xc08623abc91ec960, 0xbe1cf044f5ae35b7
> +        .quad 0xc08623af8a3c2fb8, 0xbe1cf0b0b4001e1b
> +        .quad 0xc08623b349975d98, 0xbe1cf1bae76dfbcf
> +        .quad 0xc08623b70731f810, 0xbe1cef0a72e13a62
> +        .quad 0xc08623bac30da1c8, 0xbe1cf184007d2b6b
> +        .quad 0xc08623be7d2bfb40, 0xbe1cf16f4b239e98
> +        .quad 0xc08623c2358ea2a0, 0xbe1cf0976acada87
> +        .quad 0xc08623c5ec3733d0, 0xbe1cf066318a16ff
> +        .quad 0xc08623c9a1274880, 0xbe1ceffaa7148798
> +        .quad 0xc08623cd54607820, 0xbe1cf23ab02e9b6e
> +        .quad 0xc08623d105e45800, 0xbe1cefdfef7d4fde
> +        .quad 0xc08623d4b5b47b20, 0xbe1cf17fece44f2b
> +        .quad 0xc08623d863d27270, 0xbe1cf18f907d0d7c
> +        .quad 0xc08623dc103fccb0, 0xbe1cee61fe072c98
> +        .quad 0xc08623dfbafe1668, 0xbe1cf022dd891e2f
> +        .quad 0xc08623e3640eda20, 0xbe1ceecc1daf4358
> +        .quad 0xc08623e70b73a028, 0xbe1cf0173c4fa380
> +        .quad 0xc08623eab12deec8, 0xbe1cf16a2150c2f4
> +        .quad 0xc08623ee553f4a30, 0xbe1cf1bf980b1f4b
> +        .quad 0xc08623f1f7a93480, 0xbe1cef8b731663c2
> +        .quad 0xc08623f5986d2dc0, 0xbe1cee9a664d7ef4
> +        .quad 0xc08623f9378cb3f0, 0xbe1cf1eda2af6400
> +        .quad 0xc08623fcd5094320, 0xbe1cf1923f9d68d7
> +        .quad 0xc086240070e45548, 0xbe1cf0747cd3e03a
> +        .quad 0xc08624040b1f6260, 0xbe1cf22ee855bd6d
> +        .quad 0xc0862407a3bbe078, 0xbe1cf0d57360c00b
> +        .quad 0xc086240b3abb4398, 0xbe1ceebc815cd575
> +        .quad 0xc086240ed01efdd0, 0xbe1cf03bfb970951
> +        .quad 0xc086241263e87f50, 0xbe1cf16e74768529
> +        .quad 0xc0862415f6193658, 0xbe1cefec64b8becb
> +        .quad 0xc086241986b28f30, 0xbe1cf0838d210baa
> +        .quad 0xc086241d15b5f448, 0xbe1cf0ea86e75b11
> +        .quad 0xc0862420a324ce28, 0xbe1cf1708d11d805
> +        .quad 0xc08624242f008380, 0xbe1ceea988c5a417
> +        .quad 0xc0862427b94a7910, 0xbe1cef166a7bbca5
> +        .quad 0xc086242b420411d0, 0xbe1cf0c9d9e86a38
> +        .quad 0xc086242ec92eaee8, 0xbe1cef0946455411
> +        .quad 0xc08624324ecbaf98, 0xbe1cefea60907739
> +        .quad 0xc0862435d2dc7160, 0xbe1cf1ed0934ce42
> +        .quad 0xc086243955624ff8, 0xbe1cf191ba746c7d
> +        .quad 0xc086243cd65ea548, 0xbe1ceeec78cf2a7e
> +        .quad 0xc086244055d2c968, 0xbe1cef345284c119
> +        .quad 0xc0862443d3c012b8, 0xbe1cf24f77355219
> +        .quad 0xc08624475027d5e8, 0xbe1cf05bf087e114
> +        .quad 0xc086244acb0b65d0, 0xbe1cef3504a32189
> +        .quad 0xc086244e446c1398, 0xbe1ceff54b2a406f
> +        .quad 0xc0862451bc4b2eb8, 0xbe1cf0757d54ed4f
> +        .quad 0xc086245532aa04f0, 0xbe1cf0c8099fdfd5
> +        .quad 0xc0862458a789e250, 0xbe1cf0b173796a31
> +        .quad 0xc086245c1aec1138, 0xbe1cf11d8734540d
> +        .quad 0xc086245f8cd1da60, 0xbe1cf1916a723ceb
> +        .quad 0xc0862462fd3c84d8, 0xbe1cf19a911e1da7
> +        .quad 0xc08624666c2d5608, 0xbe1cf23a9ef72e4f
> +        .quad 0xc0862469d9a591c0, 0xbe1cef503d947663
> +        .quad 0xc086246d45a67a18, 0xbe1cf0fceeb1a0b2
> +        .quad 0xc0862470b0314fa8, 0xbe1cf107e27e4fbc
> +        .quad 0xc086247419475160, 0xbe1cf03dd9922331
> +        .quad 0xc086247780e9bc98, 0xbe1cefce1a10e129
> +        .quad 0xc086247ae719cd18, 0xbe1ceea47f73c4f6
> +        .quad 0xc086247e4bd8bd10, 0xbe1ceec0ac56d100
> +        .quad 0xc0862481af27c528, 0xbe1cee8a6593278a
> +        .quad 0xc086248511081c70, 0xbe1cf2231dd9dec7
> +        .quad 0xc0862488717af888, 0xbe1cf0b4b8ed7da8
> +        .quad 0xc086248bd0818d68, 0xbe1cf1bd8d835002
> +        .quad 0xc086248f2e1d0d98, 0xbe1cf259acc107f4
> +        .quad 0xc08624928a4eaa20, 0xbe1cee897636b00c
> +        .quad 0xc0862495e5179270, 0xbe1cee757f20c326
> +        .quad 0xc08624993e78f490, 0xbe1cefafd3aa54a4
> +        .quad 0xc086249c9673fd10, 0xbe1cee7298d38b97
> +        .quad 0xc086249fed09d6f8, 0xbe1ceedc158d4ceb
> +        .quad 0xc08624a3423babe0, 0xbe1cf2282987cb2e
> +        .quad 0xc08624a6960aa400, 0xbe1cefe7381ecc4b
> +        .quad 0xc08624a9e877e600, 0xbe1cef328dbbce80
> +        .quad 0xc08624ad39849728, 0xbe1cefde45f3cc71
> +        .quad 0xc08624b08931db58, 0xbe1cefa8b89433b9
> +        .quad 0xc08624b3d780d500, 0xbe1cef6773c0b139
> +        .quad 0xc08624b72472a528, 0xbe1cf031c931c11f
> +        .quad 0xc08624ba70086b78, 0xbe1cf088f49275e7
> +        .quad 0xc08624bdba434630, 0xbe1cf17de0eaa86d
> +        .quad 0xc08624c103245238, 0xbe1cefd492f1ba75
> +        .quad 0xc08624c44aacab08, 0xbe1cf1253e154466
> +        .quad 0xc08624c790dd6ad0, 0xbe1cf0fb09ee6d55
> +        .quad 0xc08624cad5b7aa58, 0xbe1cf1f08dd048fe
> +        .quad 0xc08624ce193c8120, 0xbe1ceeca0809697f
> +        .quad 0xc08624d15b6d0538, 0xbe1cef8d5662d968
> +        .quad 0xc08624d49c4a4b78, 0xbe1cee97b556ed78
> +        .quad 0xc08624d7dbd56750, 0xbe1cf1b14b6acb75
> +        .quad 0xc08624db1a0f6b00, 0xbe1cef1e860623f2
> +        .quad 0xc08624de56f96758, 0xbe1ceeaf4d156f3d
> +        .quad 0xc08624e192946bf0, 0xbe1ceecc12b400ed
> +        .quad 0xc08624e4cce18710, 0xbe1cf180c40c794f
> +        .quad 0xc08624e805e1c5c8, 0xbe1cf185a08f7f65
> +        .quad 0xc08624eb3d9633d8, 0xbe1cef45fc924078
> +        .quad 0xc08624ee73ffdbb0, 0xbe1cf1e4f457f32a
> +        .quad 0xc08624f1a91fc6a0, 0xbe1cf040147b8a5a
> +        .quad 0xc08624f4dcf6fc98, 0xbe1cf1effca0dfb2
> +        .quad 0xc08624f80f868468, 0xbe1cf0470146e5bc
> +        .quad 0xc08624fb40cf6390, 0xbe1cef4dd186e501
> +        .quad 0xc08624fe70d29e60, 0xbe1ceebe257f66c7
> +        .quad 0xc08625019f9137f0, 0xbe1ceefb7a1c395c
> +        .quad 0xc0862504cd0c3220, 0xbe1cf209dedfed8c
> +        .quad 0xc0862507f9448db0, 0xbe1cf082da464994
> +        .quad 0xc086250b243b4a18, 0xbe1cee88694a73cf
> +        .quad 0xc086250e4df165a0, 0xbe1cf0b61e8f0531
> +        .quad 0xc08625117667dd78, 0xbe1cf1106599c962
> +        .quad 0xc08625149d9fad98, 0xbe1ceff1ee88af1f
> +        .quad 0xc0862517c399d0c8, 0xbe1cf0f746994ef6
> +        .quad 0xc086251ae85740b8, 0xbe1cefe8a1d077e4
> +        .quad 0xc086251e0bd8f5e0, 0xbe1cf1a1da036092
> +        .quad 0xc08625212e1fe7a8, 0xbe1cf0f8a7786fcd
> +        .quad 0xc08625244f2d0c48, 0xbe1cefa1174a07a7
> +        .quad 0xc08625276f0158d8, 0xbe1cef1043aa5b25
> +        .quad 0xc086252a8d9dc150, 0xbe1cf15d521c169d
> +        .quad 0xc086252dab033898, 0xbe1cf220bba8861f
> +        .quad 0xc0862530c732b078, 0xbe1cef51e310eae2
> +        .quad 0xc0862533e22d1988, 0xbe1cf222fcedd8ae
> +        .quad 0xc0862536fbf36370, 0xbe1cefdb4da4bda8
> +        .quad 0xc086253a14867ca0, 0xbe1ceeafc1112171
> +        .quad 0xc086253d2be75280, 0xbe1cee99dfb4b408
> +        .quad 0xc08625404216d160, 0xbe1cf22d2536f06b
> +        .quad 0xc08625435715e498, 0xbe1cef6abbf2e268
> +        .quad 0xc08625466ae57648, 0xbe1cf093a14789f5
> +        .quad 0xc08625497d866fa0, 0xbe1cf0f93655603c
> +        .quad 0xc086254c8ef9b8b8, 0xbe1cf1cc40c9aafc
> +        .quad 0xc086254f9f4038a8, 0xbe1ceeea5f4e9157
> +        .quad 0xc0862552ae5ad568, 0xbe1cefa9f52d4997
> +        .quad 0xc0862555bc4a7400, 0xbe1cefa490a638ff
> +        .quad 0xc0862558c90ff868, 0xbe1cef7fcf797d6f
> +        .quad 0xc086255bd4ac4590, 0xbe1cf1b4c51113c9
> +        .quad 0xc086255edf203d78, 0xbe1cef55e5b4a55d
> +        .quad 0xc0862561e86cc100, 0xbe1cf0d37a25f9dc
> +        .quad 0xc0862564f092b028, 0xbe1ceebe9efc19d9
> +        .quad 0xc0862567f792e9d8, 0xbe1cee8ad30a57b5
> +        .quad 0xc086256afd6e4c08, 0xbe1cef4e1817b90b
> +        .quad 0xc086256e0225b3b8, 0xbe1cee7fa9229996
> +        .quad 0xc086257105b9fce0, 0xbe1cf0b54963d945
> +        .quad 0xc0862574082c0298, 0xbe1cee5f2f3c7995
> +        .quad 0xc0862577097c9ee0, 0xbe1cf0828e303a2c
> +        .quad 0xc086257a09acaae0, 0xbe1cf172c3078947
> +        .quad 0xc086257d08bcfec0, 0xbe1cf189252afa22
> +        .quad 0xc086258006ae71b8, 0xbe1cefdb80426923
> +        .quad 0xc08625830381da08, 0xbe1ceef1391a0372
> +        .quad 0xc0862585ff380d00, 0xbe1cf17720c78d13
> +        .quad 0xc0862588f9d1df18, 0xbe1ceef1f9027d83
> +        .quad 0xc086258bf35023b8, 0xbe1cf06fac99dec9
> +        .quad 0xc086258eebb3ad78, 0xbe1cf1373eeb45c0
> +        .quad 0xc0862591e2fd4e00, 0xbe1cef777536bb81
> +        .quad 0xc0862594d92dd600, 0xbe1cf0f43ca40766
> +        .quad 0xc0862597ce461558, 0xbe1cefb2cfc6766b
> +        .quad 0xc086259ac246daf0, 0xbe1ceea49e64ffa2
> +        .quad 0xc086259db530f4c8, 0xbe1cf250fa457dec
> +        .quad 0xc08625a0a7053018, 0xbe1cf17d8bb2a44e
> +        .quad 0xc08625a397c45918, 0xbe1cf1d5906d54b7
> +        .quad 0xc08625a6876f3b30, 0xbe1cf08fe7b31780
> +        .quad 0xc08625a97606a0e0, 0xbe1cef13edfc9d11
> +        .quad 0xc08625ac638b53c8, 0xbe1cef9d2b107219
> +        .quad 0xc08625af4ffe1cb0, 0xbe1cf1ddd4ff6160
> +        .quad 0xc08625b23b5fc390, 0xbe1cefa02a996495
> +        .quad 0xc08625b525b10f68, 0xbe1cf166a7e37ee5
> +        .quad 0xc08625b80ef2c680, 0xbe1cef0b171068a5
> +        .quad 0xc08625baf725ae28, 0xbe1cf05c80779283
> +        .quad 0xc08625bdde4a8af0, 0xbe1cf1bbfbffb889
> +        .quad 0xc08625c0c4622090, 0xbe1cf0b8666c0124
> +        .quad 0xc08625c3a96d31e0, 0xbe1cf0a8fcf47a86
> +        .quad 0xc08625c68d6c80f0, 0xbe1cef46e18cb092
> +        .quad 0xc08625c97060cef0, 0xbe1cf1458a350efb
> +        .quad 0xc08625cc524adc58, 0xbe1ceeea1dadce12
> +        .quad 0xc08625cf332b68b0, 0xbe1cf0a1bfdc44c7
> +        .quad 0xc08625d2130332d0, 0xbe1cef96d02da73e
> +        .quad 0xc08625d4f1d2f8a8, 0xbe1cf2451c3c7701
> +        .quad 0xc08625d7cf9b7778, 0xbe1cf10d08f83812
> +        .quad 0xc08625daac5d6ba0, 0xbe1ceec5b4895c5e
> +        .quad 0xc08625dd881990b0, 0xbe1cf14e1325c5e4
> +        .quad 0xc08625e062d0a188, 0xbe1cf21d0904be12
> +        .quad 0xc08625e33c835838, 0xbe1ceed0839bcf21
> +        .quad 0xc08625e615326df0, 0xbe1cf1bb944889d2
> +        .quad 0xc08625e8ecde9b48, 0xbe1cee738e85eece
> +        .quad 0xc08625ebc38897e0, 0xbe1cf25c2bc6ef12
> +        .quad 0xc08625ee99311ac8, 0xbe1cf132b70a41ad
> +        .quad 0xc08625f16dd8da28, 0xbe1cf1984236a6e3
> +        .quad 0xc08625f441808b78, 0xbe1cf19ae74998f9
> +        .quad 0xc08625f71428e370, 0xbe1cef3e175d61a1
> +        .quad 0xc08625f9e5d295f8, 0xbe1cf101f9868fd9
> +        .quad 0xc08625fcb67e5658, 0xbe1cee69db83dcd2
> +        .quad 0xc08625ff862cd6f8, 0xbe1cf081b636af51
> +        .quad 0xc086260254dec9a8, 0xbe1cee62c7d59b3e
> +        .quad 0xc08626052294df58, 0xbe1cf1b745c57716
> +        .quad 0xc0862607ef4fc868, 0xbe1cef3d2800ea23
> +        .quad 0xc086260abb103458, 0xbe1cef480ff1acd2
> +        .quad 0xc086260d85d6d200, 0xbe1cf2424c9a17ef
> +        .quad 0xc08626104fa44f90, 0xbe1cf12cfde90fd5
> +        .quad 0xc086261318795a68, 0xbe1cf21f590dd5b6
> +        .quad 0xc0862615e0569f48, 0xbe1cf0c50f9cd28a
> +        .quad 0xc0862618a73cca30, 0xbe1ceedbdb520545
> +        .quad 0xc086261b6d2c8668, 0xbe1cf0b030396011
> +        .quad 0xc086261e32267e98, 0xbe1cf19917010e96
> +        .quad 0xc0862620f62b5cb0, 0xbe1cf07331355985
> +        .quad 0xc0862623b93bc9e8, 0xbe1cf01ae921a1c3
> +        .quad 0xc08626267b586ed0, 0xbe1cefe5cf0dbf0c
> +        .quad 0xc08626293c81f348, 0xbe1cf01b258aeb50
> +        .quad 0xc086262bfcb8fe88, 0xbe1cee6b9e7f4c68
> +        .quad 0xc086262ebbfe3710, 0xbe1cee684a9b21c9
> +        .quad 0xc08626317a5242b8, 0xbe1cf1f8bcde9a8b
> +        .quad 0xc086263437b5c6c0, 0xbe1cf1d063d36238
> +        .quad 0xc0862636f42967a8, 0xbe1cf1e31a19075e
> +        .quad 0xc0862639afadc950, 0xbe1cf1d8efdf7e7d
> +        .quad 0xc086263c6a438ef0, 0xbe1cf1812ee72dba
> +        .quad 0xc086263f23eb5b18, 0xbe1cf1449a9a2279
> +        .quad 0xc0862641dca5cfb8, 0xbe1cee96edce5085
> +        .quad 0xc086264494738e08, 0xbe1cf06797bd03b2
> +        .quad 0xc08626474b5536b8, 0xbe1cef91b9b7ffc1
> +        .quad 0xc086264a014b69c0, 0xbe1cef4b6721278f
> +        .quad 0xc086264cb656c678, 0xbe1cf1942925eb4a
> +        .quad 0xc086264f6a77eba8, 0xbe1cefa2c7bc2e39
> +        .quad 0xc08626521daf7758, 0xbe1cf252595aceb3
> +        .quad 0xc0862654cffe0718, 0xbe1cee8e9ae47ec2
> +        .quad 0xc0862657816437a8, 0xbe1cf1bf913828fa
> +        .quad 0xc086265a31e2a558, 0xbe1cf23475d6b366
> +        .quad 0xc086265ce179ebc8, 0xbe1cef8df00a922b
> +        .quad 0xc086265f902aa5f0, 0xbe1cef279bfa43e0
> +        .quad 0xc08626623df56e38, 0xbe1cf080e10b8365
> +        .quad 0xc0862664eadade70, 0xbe1cf1a518f9b544
> +        .quad 0xc086266796db8fd0, 0xbe1cef9308fed9e9
> +        .quad 0xc086266a41f81ae8, 0xbe1ceea3ae6b19c9
> +        .quad 0xc086266cec3117b8, 0xbe1ceef06003d4c2
> +        .quad 0xc086266f95871da8, 0xbe1cf0b8457ffb0c
> +        .quad 0xc08626723dfac390, 0xbe1cf0c526745ad6
> +        .quad 0xc0862674e58c9fa8, 0xbe1cf0cf91ff7b5d
> +        .quad 0xc08626778c3d4798, 0xbe1cefe260819380
> +        .quad 0xc086267a320d5070, 0xbe1ceebd90aa27a3
> +        .quad 0xc086267cd6fd4ea8, 0xbe1cf0388121dffa
> +        .quad 0xc086267f7b0dd630, 0xbe1cf1a3881435f1
> +        .quad 0xc08626821e3f7a68, 0xbe1cef28e9d9ac52
> +        .quad 0xc0862684c092ce08, 0xbe1cf02d300062dd
> +        .quad 0xc086268762086350, 0xbe1cefaee1edfa35
> +        .quad 0xc086268a02a0cbe0, 0xbe1cf0a5a052e936
> +        .quad 0xc086268ca25c98d8, 0xbe1cee60a4a497ed
> +        .quad 0xc086268f413c5ab0, 0xbe1cf0e4a5d0cf49
> +        .quad 0xc0862691df40a170, 0xbe1cf149235a4e6e
> +        .quad 0xc08626947c69fc80, 0xbe1cf215180b9fcc
> +        .quad 0xc086269718b8fac8, 0xbe1cef9b156a9840
> +        .quad 0xc0862699b42e2a90, 0xbe1cf054c91441be
> +        .quad 0xc086269c4eca19a8, 0xbe1cf13ded26512c
> +        .quad 0xc086269ee88d5550, 0xbe1cf22ea4d8ac06
> +        .quad 0xc08626a181786a40, 0xbe1cf2354666ee2e
> +        .quad 0xc08626a4198be4a8, 0xbe1cefef936752b3
> +        .quad 0xc08626a6b0c85020, 0xbe1cf1e360a9db68
> +        .quad 0xc08626a9472e37d8, 0xbe1ceed6aeb812c5
> +        .quad 0xc08626abdcbe2650, 0xbe1cf227340b4986
> +        .quad 0xc08626ae7178a5b0, 0xbe1cf0215a0cbe0d
> +        .quad 0xc08626b1055e3f70, 0xbe1cf256adf0ae26
> +        .quad 0xc08626b3986f7ca8, 0xbe1ceff3c67aed06
> +        .quad 0xc08626b62aace5c8, 0xbe1cf2159fb93652
> +        .quad 0xc08626b8bc1702e0, 0xbe1cf01e6dbd1c7f
> +        .quad 0xc08626bb4cae5b60, 0xbe1cf009e75d1c0c
> +        .quad 0xc08626bddc737648, 0xbe1ceec10a020e73
> +        .quad 0xc08626c06b66da08, 0xbe1cf06d5783eee7
> +        .quad 0xc08626c2f9890ca0, 0xbe1cf0cb8f169ffe
> +        .quad 0xc08626c586da9388, 0xbe1cef7de2452430
> +        .quad 0xc08626c8135bf3b0, 0xbe1cf05da6f783ae
> +        .quad 0xc08626ca9f0db198, 0xbe1cefcc877d681d
> +        .quad 0xc08626cd29f05138, 0xbe1cef0531954ab3
> +        .quad 0xc08626cfb4045608, 0xbe1cf06b8565ea3d
> +        .quad 0xc08626d23d4a4310, 0xbe1cefdc455d9d7e
> +        .quad 0xc08626d4c5c29ad0, 0xbe1ceefc47e8fa64
> +        .quad 0xc08626d74d6ddf48, 0xbe1cf1872bf033f2
> +        .quad 0xc08626d9d44c9210, 0xbe1cf19d91087f9d
> +        .quad 0xc08626dc5a5f3438, 0xbe1cf012d444c6ab
> +        .quad 0xc08626dedfa64650, 0xbe1cf0ba528ee153
> +        .quad 0xc08626e164224880, 0xbe1ceeb431709788
> +        .quad 0xc08626e3e7d3ba60, 0xbe1cf0b9af31a6a5
> +        .quad 0xc08626e66abb1b28, 0xbe1cf168fb2e135b
> +        .quad 0xc08626e8ecd8e990, 0xbe1cef9097461c93
> +        .quad 0xc08626eb6e2da3d0, 0xbe1cee7a434735d8
> +        .quad 0xc08626edeeb9c7a8, 0xbe1cf235732b86f2
> +        .quad 0xc08626f06e7dd280, 0xbe1cefe1510b89e6
> +        .quad 0xc08626f2ed7a4120, 0xbe1cf1f64b9b80ef
> +        .quad 0xc08626f56baf9000, 0xbe1cf08f320ca339
> +        .quad 0xc08626f7e91e3b08, 0xbe1cf1b1de2808a1
> +        .quad 0xc08626fa65c6bdc0, 0xbe1cf1976d778b28
> +        .quad 0xc08626fce1a99338, 0xbe1ceef40a4f076f
> +        .quad 0xc08626ff5cc73600, 0xbe1cef3e45869ce3
> +        .quad 0xc0862701d7202048, 0xbe1ceef601b4c9d6
> +        .quad 0xc086270450b4cbc0, 0xbe1cf1eaf0b57fd6
> +        .quad 0xc0862706c985b1c0, 0xbe1cef82a44990f3
> +        .quad 0xc086270941934b10, 0xbe1ceefe32981f2c
> +        .quad 0xc086270bb8de1018, 0xbe1cefbf6f5a0445
> +        .quad 0xc086270e2f6678d0, 0xbe1cf18dba75792c
> +        .quad 0xc0862710a52cfcc8, 0xbe1cf0da64ce995f
> +        .quad 0xc08627131a321318, 0xbe1cef04ac0fb802
> +        .quad 0xc08627158e763268, 0xbe1cee9d4e2ad9bd
> +        .quad 0xc086271801f9d0f8, 0xbe1cefa9b55407b5
> +        .quad 0xc086271a74bd64a0, 0xbe1cefe6bd329570
> +        .quad 0xc086271ce6c162c8, 0xbe1cef0b1205dc85
> +        .quad 0xc086271f58064068, 0xbe1cef092a785e3f
> +        .quad 0xc0862721c88c7210, 0xbe1cf050dcdaac30
> +        .quad 0xc086272438546be8, 0xbe1cf210907ded8b
> +        .quad 0xc0862726a75ea1b8, 0xbe1cee760be44f99
> +        .quad 0xc086272915ab86c0, 0xbe1ceeeee07c2bcc
> +        .quad 0xc086272b833b8df0, 0xbe1cf06874992df5
> +        .quad 0xc086272df00f29d0, 0xbe1cef8fac5d4899
> +        .quad 0xc08627305c26cc70, 0xbe1cf1103241cc99
> +        .quad 0xc0862732c782e788, 0xbe1cf1d35fef83fe
> +        .quad 0xc08627353223ec68, 0xbe1cef3ec8133e1d
> +        .quad 0xc08627379c0a4be8, 0xbe1cef7261daccd8
> +        .quad 0xc086273a05367688, 0xbe1cf18656c50806
> +        .quad 0xc086273c6da8dc68, 0xbe1cf1c8736e049a
> +        .quad 0xc086273ed561ed38, 0xbe1cf1f93bff4911
> +        .quad 0xc08627413c621848, 0xbe1cf188a4ea680c
> +        .quad 0xc0862743a2a9cc80, 0xbe1cf1d270930c80
> +        .quad 0xc086274608397868, 0xbe1cf25a328c28e2
> +        .quad 0xc08627486d118a28, 0xbe1cf106f90aa3b8
> +        .quad 0xc086274ad1326f80, 0xbe1cee5e9d2e885a
> +        .quad 0xc086274d349c95c0, 0xbe1cf1c0bac27228
> +        .quad 0xc086274f975069f8, 0xbe1cf1a1500f9b1c
> +        .quad 0xc0862751f94e58c0, 0xbe1cefc30663ac44
> +        .quad 0xc08627545a96ce48, 0xbe1cf17123e427a2
> +        .quad 0xc0862756bb2a3678, 0xbe1cefb92749fea4
> +        .quad 0xc08627591b08fcc0, 0xbe1cefa40e1ea74a
> +        .quad 0xc086275b7a338c40, 0xbe1cee6f4612c3e9
> +        .quad 0xc086275dd8aa4fa8, 0xbe1cf1c54a053627
> +        .quad 0xc0862760366db168, 0xbe1ceff5eb503d9e
> +        .quad 0xc0862762937e1b70, 0xbe1cf02e47f10cee
> +        .quad 0xc0862764efdbf768, 0xbe1ceeb06e1d0dad
> +        .quad 0xc08627674b87ae88, 0xbe1cf10aadd6dba5
> +        .quad 0xc0862769a681a9c0, 0xbe1cf24e9913d30f
> +        .quad 0xc086276c00ca51a0, 0xbe1cef47b301e312
> +        .quad 0xc086276e5a620e48, 0xbe1ceeb1cefc2e85
> +        .quad 0xc0862770b3494788, 0xbe1cf16f1fbbe011
> +        .quad 0xc08627730b8064e8, 0xbe1ceebdf75174c7
> +        .quad 0xc08627756307cd70, 0xbe1cf06e3871a0da
> +        .quad 0xc0862777b9dfe7f0, 0xbe1cef16799fd554
> +        .quad 0xc086277a10091ac0, 0xbe1cf248dabf5377
> +        .quad 0xc086277c6583cc00, 0xbe1cf0c78d92a2cd
> +        .quad 0xc086277eba506158, 0xbe1cf0b911b029f0
> +        .quad 0xc08627810e6f4028, 0xbe1cefdc24719766
> +        .quad 0xc086278361e0cd70, 0xbe1cefbb6562b7e7
> +        .quad 0xc0862785b4a56dd8, 0xbe1cf1e0afb349ec
> +        .quad 0xc086278806bd85c0, 0xbe1cf008292e52fc
> +        .quad 0xc086278a58297918, 0xbe1cf053073872bf
> +        .quad 0xc086278ca8e9ab88, 0xbe1cf17a0a55a947
> +        .quad 0xc086278ef8fe8068, 0xbe1ceeffb0b60234
> +        .quad 0xc086279148685aa0, 0xbe1cf162204794a8
> +        .quad 0xc086279397279ce0, 0xbe1cf24cc8cb48ac
> +        .quad 0xc0862795e53ca978, 0xbe1cf0c9be68d5c3
> +        .quad 0xc086279832a7e258, 0xbe1cf172cd3d7388
> +        .quad 0xc086279a7f69a930, 0xbe1ceea2465fbce5
> +        .quad 0xc086279ccb825f40, 0xbe1cf0a386d2500f
> +        .quad 0xc086279f16f26590, 0xbe1cf1e338ddc18a
> +        .quad 0xc08627a161ba1cd0, 0xbe1cef1f5049867f
> +        .quad 0xc08627a3abd9e548, 0xbe1cef96c1ea8b1f
> +        .quad 0xc08627a5f5521f00, 0xbe1cf138f6fd3c26
> +        .quad 0xc08627a83e2329b0, 0xbe1cf0d4fcbfdf3a
> +        .quad 0xc08627aa864d64b0, 0xbe1cf24870c12c81
> +        .quad 0xc08627accdd12f18, 0xbe1cf0ae2a56348d
> +        .quad 0xc08627af14aee7a0, 0xbe1cee8ca1a9b893
> +        .quad 0xc08627b15ae6eca8, 0xbe1cf20414d637b0
> +        .quad 0xc08627b3a0799c60, 0xbe1cf0fc6b7b12d8
> +        .quad 0xc08627b5e5675488, 0xbe1cf152d93c4a00
> +        .quad 0xc08627b829b072a0, 0xbe1cf1073f9b77c2
> +        .quad 0xc08627ba6d5553d8, 0xbe1cee694f97d5a4
> +        .quad 0xc08627bcb0565500, 0xbe1cf0456b8239d7
> +        .quad 0xc08627bef2b3d2b0, 0xbe1cf211497127e3
> +        .quad 0xc08627c1346e2930, 0xbe1cf01856c0384d
> +        .quad 0xc08627c37585b468, 0xbe1cefa7dd05479e
> +        .quad 0xc08627c5b5fad000, 0xbe1cef3ae8e50b93
> +        .quad 0xc08627c7f5cdd750, 0xbe1ceea5f32fdd3a
> +        .quad 0xc08627ca34ff2560, 0xbe1cef424caeb8d9
> +        .quad 0xc08627cc738f14f0, 0xbe1cf0194d07a81f
> +        .quad 0xc08627ceb17e0070, 0xbe1cf20f452000c1
> +        .quad 0xc08627d0eecc4210, 0xbe1cf00e356218e4
> +        .quad 0xc08627d32b7a33a0, 0xbe1cef30484b4bcb
> +        .quad 0xc08627d567882eb0, 0xbe1ceeea11a6641b
> +        .quad 0xc08627d7a2f68c80, 0xbe1cf13492d5bd7b
> +        .quad 0xc08627d9ddc5a618, 0xbe1ceeb7048fad96
> +        .quad 0xc08627dc17f5d418, 0xbe1ceef0666f0477
> +        .quad 0xc08627de51876ee8, 0xbe1cf060d4b8b5c2
> +        .quad 0xc08627e08a7acea8, 0xbe1cf0b2a4b6ff8c
> +        .quad 0xc08627e2c2d04b28, 0xbe1cf0e34809a875
> +        .quad 0xc08627e4fa883bf0, 0xbe1cf16bf74a3522
> +        .quad 0xc08627e731a2f848, 0xbe1cee6a24623d57
> +        .quad 0xc08627e96820d718, 0xbe1cefc7b4f1528e
> +        .quad 0xc08627eb9e022f18, 0xbe1cf163051f3548
> +        .quad 0xc08627edd34756b8, 0xbe1cef36b3366305
> +        .quad 0xc08627f007f0a408, 0xbe1cf18134625550
> +        .quad 0xc08627f23bfe6cf0, 0xbe1cf0ec32ec1a11
> +        .quad 0xc08627f46f710700, 0xbe1ceeb3b64f3edc
> +        .quad 0xc08627f6a248c778, 0xbe1cf0cd15805bc8
> +        .quad 0xc08627f8d4860368, 0xbe1cf20db3bddebe
> +        .quad 0xc08627fb06290f90, 0xbe1cf25188430e25
> +        .quad 0xc08627fd37324070, 0xbe1ceea1713490f9
> +        .quad 0xc08627ff67a1ea28, 0xbe1cf159521d234c
> +        .quad 0xc0862801977860b8, 0xbe1cf24dfe50783b
> +        .quad 0xc0862803c6b5f7d0, 0xbe1ceef2ef89a60b
> +        .quad 0xc0862805f55b02c8, 0xbe1cee7fc919d62c
> +        .quad 0xc08628082367d4c0, 0xbe1cf215a7fb513a
> +        .quad 0xc086280a50dcc0a8, 0xbe1cf0e4401c5ed4
> +        .quad 0xc086280c7dba1910, 0xbe1cf04ec734d256
> +        .quad 0xc086280eaa003050, 0xbe1cf010ad787fea
> +        .quad 0xc0862810d5af5880, 0xbe1cee622478393d
> +        .quad 0xc086281300c7e368, 0xbe1cf01c7482564f
> +        .quad 0xc08628152b4a22a0, 0xbe1cf0de20d33536
> +        .quad 0xc086281755366778, 0xbe1cef2edae5837d
> +        .quad 0xc08628197e8d02f0, 0xbe1cf0a345318cc9
> +        .quad 0xc086281ba74e45d8, 0xbe1cf20085aa34b8
> +        .quad 0xc086281dcf7a80c0, 0xbe1cef5fa845ad83
> +        .quad 0xc086281ff71203e0, 0xbe1cf050d1df69c4
> +        .quad 0xc08628221e151f48, 0xbe1ceffe43c035b9
> +        .quad 0xc0862824448422b8, 0xbe1cf14f3018d3c2
> +        .quad 0xc08628266a5f5dc0, 0xbe1cef0a5fbae83d
> +        .quad 0xc08628288fa71f98, 0xbe1ceff8a95b72a1
> +        .quad 0xc086282ab45bb750, 0xbe1cef073aa9849b
> +        .quad 0xc086282cd87d73a8, 0xbe1cef69b3835c02
> +        .quad 0xc086282efc0ca328, 0xbe1cf0bc139379a9
> +        .quad 0xc08628311f099420, 0xbe1cef247a9ec596
> +        .quad 0xc086283341749490, 0xbe1cef74bbcc488a
> +        .quad 0xc0862835634df248, 0xbe1cef4bc42e7b8e
> +        .quad 0xc08628378495fad0, 0xbe1cf136d4d5a810
> +        .quad 0xc0862839a54cfb80, 0xbe1cf0d290b24dd8
> +        .quad 0xc086283bc5734168, 0xbe1ceeebde8e0065
> +        .quad 0xc086283de5091950, 0xbe1cf1a09f60aa1e
> +        .quad 0xc0862840040ecfe0, 0xbe1cf0803947a234
> +        .quad 0xc08628422284b168, 0xbe1cf0abf7638127
> +        .quad 0xc0862844406b0a08, 0xbe1cf0f73ee12058
> +        .quad 0xc08628465dc225a0, 0xbe1cf2079971b26c
> +        .quad 0xc08628487a8a4fe0, 0xbe1cee74957564b1
> +        .quad 0xc086284a96c3d420, 0xbe1ceee77c1b7d43
> +        .quad 0xc086284cb26efd90, 0xbe1cf23addba6e09
> +        .quad 0xc086284ecd8c1730, 0xbe1cf199f4a1da60
> +        .quad 0xc0862850e81b6bb0, 0xbe1cf09fdea81393
> +        .quad 0xc0862853021d4588, 0xbe1cf176adb417f7
> +        .quad 0xc08628551b91ef00, 0xbe1cf0f64f84a8da
> +        .quad 0xc08628573479b220, 0xbe1ceec34cf49523
> +        .quad 0xc08628594cd4d8a8, 0xbe1cf16d60fbe0bb
> +        .quad 0xc086285b64a3ac40, 0xbe1cee8de7acfc7b
> +        .quad 0xc086285d7be67630, 0xbe1ceee6256cce8d
> +        .quad 0xc086285f929d7fa0, 0xbe1cee7d66a3d8a5
> +        .quad 0xc0862861a8c91170, 0xbe1cf0bef8265792
> +        .quad 0xc0862863be697458, 0xbe1cf097f890c6f8
> +        .quad 0xc0862865d37ef0c8, 0xbe1cf09502d5c3fc
> +        .quad 0xc0862867e809cf00, 0xbe1ceeffb239dac7
> +        .quad 0xc0862869fc0a56f8, 0xbe1cf1fbfff95c98
> +        .quad 0xc086286c0f80d090, 0xbe1cefa57ad3eef7
> +        .quad 0xc086286e226d8348, 0xbe1cf22c58b9183d
> +        .quad 0xc086287034d0b690, 0xbe1ceff262d0a248
> +        .quad 0xc086287246aab180, 0xbe1cefa7bc194186
> +        .quad 0xc086287457fbbb08, 0xbe1cf06782d784d9
> +        .quad 0xc086287668c419e0, 0xbe1cf1d44d0eaa07
> +        .quad 0xc086287879041490, 0xbe1cf034803c8a48
> +        .quad 0xc086287a88bbf158, 0xbe1cf08e84916b6f
> +        .quad 0xc086287c97ebf650, 0xbe1cf0c4d3dc1bc7
> +        .quad 0xc086287ea6946958, 0xbe1cefb1e4625943
> +        .quad 0xc0862880b4b59010, 0xbe1cf143efdd1fd0
> +        .quad 0xc0862882c24faff8, 0xbe1cee9896d016da
> +        .quad 0xc0862884cf630e38, 0xbe1cf2186072f2cc
> +        .quad 0xc0862886dbefeff0, 0xbe1cef9217633d34
> +        .quad 0xc0862888e7f699e0, 0xbe1cf05603549486
> +        .quad 0xc086288af37750b0, 0xbe1cef50fff513d3
> +        .quad 0xc086288cfe7258c0, 0xbe1cf127713b32d0
> +        .quad 0xc086288f08e7f650, 0xbe1cf05015520f3d
> +        .quad 0xc086289112d86d58, 0xbe1cf12eb458b26f
> +        .quad 0xc08628931c4401a8, 0xbe1cf22eae2887ed
> +        .quad 0xc0862895252af6e0, 0xbe1cefdd6656dd2d
> +        .quad 0xc08628972d8d9058, 0xbe1cf1048ea4e646
> +        .quad 0xc0862899356c1150, 0xbe1ceec4501167e9
> +        .quad 0xc086289b3cc6bcb8, 0xbe1cf0ad52becc3f
> +        .quad 0xc086289d439dd568, 0xbe1cf0daa4e00e35
> +        .quad 0xc086289f49f19df8, 0xbe1cf00b80de8d6a
> +        .quad 0xc08628a14fc258c8, 0xbe1cf1bcf2ea8464
> +        .quad 0xc08628a355104818, 0xbe1cf0435e2782b0
> +        .quad 0xc08628a559dbade0, 0xbe1cf0e3e1a5f56c
> +        .quad 0xc08628a75e24cbf8, 0xbe1cefed9d5a721d
> +        .quad 0xc08628a961ebe3f8, 0xbe1cf0d2d74321e2
> +        .quad 0xc08628ab65313750, 0xbe1cf24200eb55e9
> +        .quad 0xc08628ad67f50740, 0xbe1cf23e9d7cf979
> +        .quad 0xc08628af6a3794d0, 0xbe1cf23a088f421c
> +        .quad 0xc08628b16bf920e0, 0xbe1cef2c1de1ab32
> +        .quad 0xc08628b36d39ec08, 0xbe1cf1abc231f7b2
> +        .quad 0xc08628b56dfa36d0, 0xbe1cf2074d5ba303
> +        .quad 0xc08628b76e3a4180, 0xbe1cf05cd5eed880
> +        /*== Log_LA_table ==*/
> +        .align 16
> +        .quad 0x8000000000000000
> +        .quad 0xbf5ff802a9ab10e6
> +        .quad 0xbf6ff00aa2b10bc0
> +        .quad 0xbf77ee11ebd82e94
> +        .quad 0xbf7fe02a6b106789
> +        .quad 0xbf83e7295d25a7d9
> +        .quad 0xbf87dc475f810a77
> +        .quad 0xbf8bcf712c74384c
> +        .quad 0xbf8fc0a8b0fc03e4
> +        .quad 0xbf91d7f7eb9eebe7
> +        .quad 0xbf93cea44346a575
> +        .quad 0xbf95c45a51b8d389
> +        .quad 0xbf97b91b07d5b11b
> +        .quad 0xbf99ace7551cc514
> +        .quad 0xbf9b9fc027af9198
> +        .quad 0xbf9d91a66c543cc4
> +        .quad 0xbf9f829b0e783300
> +        .quad 0xbfa0b94f7c196176
> +        .quad 0xbfa1b0d98923d980
> +        .quad 0xbfa2a7ec2214e873
> +        .quad 0xbfa39e87b9febd60
> +        .quad 0xbfa494acc34d911c
> +        .quad 0xbfa58a5bafc8e4d5
> +        .quad 0xbfa67f94f094bd98
> +        .quad 0xbfa77458f632dcfc
> +        .quad 0xbfa868a83083f6cf
> +        .quad 0xbfa95c830ec8e3eb
> +        .quad 0xbfaa4fe9ffa3d235
> +        .quad 0xbfab42dd711971bf
> +        .quad 0xbfac355dd0921f2d
> +        .quad 0xbfad276b8adb0b52
> +        .quad 0xbfae19070c276016
> +        .quad 0xbfaf0a30c01162a6
> +        .quad 0xbfaffae9119b9303
> +        .quad 0xbfb075983598e471
> +        .quad 0xbfb0ed839b5526fe
> +        .quad 0xbfb16536eea37ae1
> +        .quad 0xbfb1dcb263db1944
> +        .quad 0xbfb253f62f0a1417
> +        .quad 0xbfb2cb0283f5de1f
> +        .quad 0xbfb341d7961bd1d1
> +        .quad 0xbfb3b87598b1b6ee
> +        .quad 0xbfb42edcbea646f0
> +        .quad 0xbfb4a50d3aa1b040
> +        .quad 0xbfb51b073f06183f
> +        .quad 0xbfb590cafdf01c28
> +        .quad 0xbfb60658a93750c4
> +        .quad 0xbfb67bb0726ec0fc
> +        .quad 0xbfb6f0d28ae56b4c
> +        .quad 0xbfb765bf23a6be13
> +        .quad 0xbfb7da766d7b12cd
> +        .quad 0xbfb84ef898e8282a
> +        .quad 0xbfb8c345d6319b21
> +        .quad 0xbfb9375e55595ede
> +        .quad 0xbfb9ab42462033ad
> +        .quad 0xbfba1ef1d8061cd4
> +        .quad 0xbfba926d3a4ad563
> +        .quad 0xbfbb05b49bee43fe
> +        .quad 0xbfbb78c82bb0eda1
> +        .quad 0xbfbbeba818146765
> +        .quad 0xbfbc5e548f5bc743
> +        .quad 0xbfbcd0cdbf8c13e1
> +        .quad 0xbfbd4313d66cb35d
> +        .quad 0xbfbdb5270187d927
> +        .quad 0xbfbe27076e2af2e6
> +        .quad 0xbfbe98b549671467
> +        .quad 0xbfbf0a30c01162a6
> +        .quad 0xbfbf7b79fec37ddf
> +        .quad 0xbfbfec9131dbeabb
> +        .quad 0xbfc02ebb42bf3d4b
> +        .quad 0xbfc0671512ca596e
> +        .quad 0xbfc09f561ee719c3
> +        .quad 0xbfc0d77e7cd08e59
> +        .quad 0xbfc10f8e422539b1
> +        .quad 0xbfc14785846742ac
> +        .quad 0xbfc17f6458fca611
> +        .quad 0xbfc1b72ad52f67a0
> +        .quad 0xbfc1eed90e2dc2c3
> +        .quad 0xbfc2266f190a5acb
> +        .quad 0xbfc25ded0abc6ad2
> +        .quad 0xbfc29552f81ff523
> +        .quad 0xbfc2cca0f5f5f251
> +        .quad 0xbfc303d718e47fd3
> +        .quad 0xbfc33af575770e4f
> +        .quad 0xbfc371fc201e8f74
> +        .quad 0xbfc3a8eb2d31a376
> +        .quad 0xbfc3dfc2b0ecc62a
> +        .quad 0xbfc41682bf727bc0
> +        .quad 0xbfc44d2b6ccb7d1e
> +        .quad 0xbfc483bccce6e3dd
> +        .quad 0xbfc4ba36f39a55e5
> +        .quad 0xbfc4f099f4a230b2
> +        .quad 0xbfc526e5e3a1b438
> +        .quad 0xbfc55d1ad4232d6f
> +        .quad 0xbfc59338d9982086
> +        .quad 0xbfc5c940075972b9
> +        .quad 0xbfc5ff3070a793d4
> +        .quad 0xbfc6350a28aaa758
> +        .quad 0xbfc66acd4272ad51
> +        .quad 0xbfc6a079d0f7aad2
> +        .quad 0xbfc6d60fe719d21d
> +        .quad 0xbfc70b8f97a1aa75
> +        .quad 0xbfc740f8f54037a5
> +        .quad 0xbfc7764c128f2127
> +        .quad 0xbfc7ab890210d909
> +        .quad 0xbfc7e0afd630c274
> +        .quad 0xbfc815c0a14357eb
> +        .quad 0xbfc84abb75865139
> +        .quad 0xbfc87fa06520c911
> +        .quad 0xbfc8b46f8223625b
> +        .quad 0xbfc8e928de886d41
> +        .quad 0xbfc91dcc8c340bde
> +        .quad 0xbfc9525a9cf456b4
> +        .quad 0xbfc986d3228180ca
> +        .quad 0xbfc9bb362e7dfb83
> +        .quad 0xbfc9ef83d2769a34
> +        .quad 0xbfca23bc1fe2b563
> +        .quad 0xbfca57df28244dcd
> +        .quad 0xbfca8becfc882f19
> +        .quad 0xbfcabfe5ae46124c
> +        .quad 0xbfcaf3c94e80bff3
> +        .quad 0xbfcb2797ee46320c
> +        .quad 0xbfcb5b519e8fb5a4
> +        .quad 0xbfcb8ef670420c3b
> +        .quad 0xbfcbc286742d8cd6
> +        .quad 0xbfcbf601bb0e44e2
> +        .quad 0xbfcc2968558c18c1
> +        .quad 0xbfcc5cba543ae425
> +        .quad 0xbfcc8ff7c79a9a22
> +        .quad 0xbfccc320c0176502
> +        .quad 0xbfccf6354e09c5dc
> +        .quad 0xbfcd293581b6b3e7
> +        .quad 0xbfcd5c216b4fbb91
> +        .quad 0xbfcd8ef91af31d5e
> +        .quad 0xbfcdc1bca0abec7d
> +        .quad 0xbfcdf46c0c722d2f
> +        .quad 0xbfce27076e2af2e6
> +        .quad 0xbfce598ed5a87e2f
> +        .quad 0xbfce8c0252aa5a60
> +        .quad 0xbfcebe61f4dd7b0b
> +        .quad 0xbfcef0adcbdc5936
> +        .quad 0xbfcf22e5e72f105d
> +        .quad 0xbfcf550a564b7b37
> +        .quad 0xbfcf871b28955045
> +        .quad 0xbfcfb9186d5e3e2b
> +        .quad 0xbfcfeb0233e607cc
> +        .quad 0xbfd00e6c45ad501d
> +        .quad 0xbfd0274dc16c232f
> +        .quad 0xbfd0402594b4d041
> +        .quad 0xbfd058f3c703ebc6
> +        .quad 0xbfd071b85fcd590d
> +        .quad 0xbfd08a73667c57af
> +        .quad 0xbfd0a324e27390e3
> +        .quad 0xbfd0bbccdb0d24bd
> +        .quad 0xbfd0d46b579ab74b
> +        .quad 0xbfd0ed005f657da4
> +        .quad 0xbfd1058bf9ae4ad5
> +        .quad 0xbfd11e0e2dad9cb7
> +        .quad 0xbfd136870293a8b0
> +        .quad 0xbfd14ef67f88685a
> +        .quad 0xbfd1675cababa60e
> +        .quad 0xbfd17fb98e15095d
> +        .quad 0xbfd1980d2dd4236f
> +        .quad 0xbfd1b05791f07b49
> +        .quad 0xbfd1c898c16999fb
> +        .quad 0xbfd1e0d0c33716be
> +        .quad 0xbfd1f8ff9e48a2f3
> +        .quad 0xbfd211255986160c
> +        .quad 0xbfd22941fbcf7966
> +        .quad 0xbfd241558bfd1404
> +        .quad 0xbfd2596010df763a
> +        .quad 0xbfd27161913f853d
> +        .quad 0xbfd2895a13de86a3
> +        .quad 0xbfd2a1499f762bc9
> +        .quad 0xbfd2b9303ab89d25
> +        .quad 0xbfd2d10dec508583
> +        .quad 0xbfd2e8e2bae11d31
> +        .quad 0xbfd300aead06350c
> +        .quad 0xbfd31871c9544185
> +        .quad 0xbfd3302c16586588
> +        .quad 0xbfd347dd9a987d55
> +        .quad 0xbfd35f865c93293e
> +        .quad 0xbfd3772662bfd85b
> +        .quad 0xbfd38ebdb38ed321
> +        .quad 0xbfd3a64c556945ea
> +        .quad 0xbfd3bdd24eb14b6a
> +        .quad 0xbfd3d54fa5c1f710
> +        .quad 0xbfd3ecc460ef5f50
> +        .quad 0xbfd404308686a7e4
> +        .quad 0xbfd41b941cce0bee
> +        .quad 0xbfd432ef2a04e814
> +        .quad 0xbfd44a41b463c47c
> +        .quad 0xbfd4618bc21c5ec2
> +        .quad 0xbfd478cd5959b3d9
> +        .quad 0xbfd49006804009d1
> +        .quad 0xbfd4a7373cecf997
> +        .quad 0xbfd4be5f957778a1
> +        .quad 0xbfd4d57f8fefe27f
> +        .quad 0xbfd4ec973260026a
> +        .quad 0xbfd503a682cb1cb3
> +        .quad 0xbfd51aad872df82d
> +        .quad 0xbfd531ac457ee77e
> +        .quad 0xbfd548a2c3add263
> +        .quad 0xbfd55f9107a43ee2
> +        .quad 0xbfd5767717455a6c
> +        .quad 0xbfd58d54f86e02f2
> +        .quad 0xbfd5a42ab0f4cfe2
> +        .quad 0xbfd5baf846aa1b19
> +        .quad 0xbfd5d1bdbf5809ca
> +        .quad 0xbfd5e87b20c2954a
> +        .quad 0xbfd5ff3070a793d4
> +        .quad 0xbfd615ddb4bec13c
> +        .quad 0xbfd62c82f2b9c795
> +        .quad 0x3fd61965cdb02c1f
> +        .quad 0x3fd602d08af091ec
> +        .quad 0x3fd5ec433d5c35ae
> +        .quad 0x3fd5d5bddf595f30
> +        .quad 0x3fd5bf406b543db2
> +        .quad 0x3fd5a8cadbbedfa1
> +        .quad 0x3fd5925d2b112a59
> +        .quad 0x3fd57bf753c8d1fb
> +        .quad 0x3fd565995069514c
> +        .quad 0x3fd54f431b7be1a9
> +        .quad 0x3fd538f4af8f72fe
> +        .quad 0x3fd522ae0738a3d8
> +        .quad 0x3fd50c6f1d11b97c
> +        .quad 0x3fd4f637ebba9810
> +        .quad 0x3fd4e0086dd8baca
> +        .quad 0x3fd4c9e09e172c3c
> +        .quad 0x3fd4b3c077267e9a
> +        .quad 0x3fd49da7f3bcc41f
> +        .quad 0x3fd487970e958770
> +        .quad 0x3fd4718dc271c41b
> +        .quad 0x3fd45b8c0a17df13
> +        .quad 0x3fd44591e0539f49
> +        .quad 0x3fd42f9f3ff62642
> +        .quad 0x3fd419b423d5e8c7
> +        .quad 0x3fd403d086cea79c
> +        .quad 0x3fd3edf463c1683e
> +        .quad 0x3fd3d81fb5946dba
> +        .quad 0x3fd3c25277333184
> +        .quad 0x3fd3ac8ca38e5c5f
> +        .quad 0x3fd396ce359bbf54
> +        .quad 0x3fd3811728564cb2
> +        .quad 0x3fd36b6776be1117
> +        .quad 0x3fd355bf1bd82c8b
> +        .quad 0x3fd3401e12aecba1
> +        .quad 0x3fd32a84565120a8
> +        .quad 0x3fd314f1e1d35ce4
> +        .quad 0x3fd2ff66b04ea9d4
> +        .quad 0x3fd2e9e2bce12286
> +        .quad 0x3fd2d46602adccee
> +        .quad 0x3fd2bef07cdc9354
> +        .quad 0x3fd2a982269a3dbf
> +        .quad 0x3fd2941afb186b7c
> +        .quad 0x3fd27ebaf58d8c9d
> +        .quad 0x3fd269621134db92
> +        .quad 0x3fd25410494e56c7
> +        .quad 0x3fd23ec5991eba49
> +        .quad 0x3fd22981fbef797b
> +        .quad 0x3fd214456d0eb8d4
> +        .quad 0x3fd1ff0fe7cf47a7
> +        .quad 0x3fd1e9e1678899f4
> +        .quad 0x3fd1d4b9e796c245
> +        .quad 0x3fd1bf99635a6b95
> +        .quad 0x3fd1aa7fd638d33f
> +        .quad 0x3fd1956d3b9bc2fa
> +        .quad 0x3fd180618ef18adf
> +        .quad 0x3fd16b5ccbacfb73
> +        .quad 0x3fd1565eed455fc3
> +        .quad 0x3fd14167ef367783
> +        .quad 0x3fd12c77cd00713b
> +        .quad 0x3fd1178e8227e47c
> +        .quad 0x3fd102ac0a35cc1c
> +        .quad 0x3fd0edd060b78081
> +        .quad 0x3fd0d8fb813eb1ef
> +        .quad 0x3fd0c42d676162e3
> +        .quad 0x3fd0af660eb9e279
> +        .quad 0x3fd09aa572e6c6d4
> +        .quad 0x3fd085eb8f8ae797
> +        .quad 0x3fd07138604d5862
> +        .quad 0x3fd05c8be0d9635a
> +        .quad 0x3fd047e60cde83b8
> +        .quad 0x3fd03346e0106062
> +        .quad 0x3fd01eae5626c691
> +        .quad 0x3fd00a1c6adda473
> +        .quad 0x3fcfeb2233ea07cd
> +        .quad 0x3fcfc218be620a5e
> +        .quad 0x3fcf991c6cb3b379
> +        .quad 0x3fcf702d36777df0
> +        .quad 0x3fcf474b134df229
> +        .quad 0x3fcf1e75fadf9bde
> +        .quad 0x3fcef5ade4dcffe6
> +        .quad 0x3fceccf2c8fe920a
> +        .quad 0x3fcea4449f04aaf5
> +        .quad 0x3fce7ba35eb77e2a
> +        .quad 0x3fce530effe71012
> +        .quad 0x3fce2a877a6b2c12
> +        .quad 0x3fce020cc6235ab5
> +        .quad 0x3fcdd99edaf6d7e9
> +        .quad 0x3fcdb13db0d48940
> +        .quad 0x3fcd88e93fb2f450
> +        .quad 0x3fcd60a17f903515
> +        .quad 0x3fcd38666871f465
> +        .quad 0x3fcd1037f2655e7b
> +        .quad 0x3fcce816157f1988
> +        .quad 0x3fccc000c9db3c52
> +        .quad 0x3fcc97f8079d44ec
> +        .quad 0x3fcc6ffbc6f00f71
> +        .quad 0x3fcc480c0005ccd1
> +        .quad 0x3fcc2028ab17f9b4
> +        .quad 0x3fcbf851c067555f
> +        .quad 0x3fcbd087383bd8ad
> +        .quad 0x3fcba8c90ae4ad19
> +        .quad 0x3fcb811730b823d2
> +        .quad 0x3fcb5971a213acdb
> +        .quad 0x3fcb31d8575bce3d
> +        .quad 0x3fcb0a4b48fc1b46
> +        .quad 0x3fcae2ca6f672bd4
> +        .quad 0x3fcabb55c31693ad
> +        .quad 0x3fca93ed3c8ad9e3
> +        .quad 0x3fca6c90d44b704e
> +        .quad 0x3fca454082e6ab05
> +        .quad 0x3fca1dfc40f1b7f1
> +        .quad 0x3fc9f6c407089664
> +        .quad 0x3fc9cf97cdce0ec3
> +        .quad 0x3fc9a8778debaa38
> +        .quad 0x3fc981634011aa75
> +        .quad 0x3fc95a5adcf7017f
> +        .quad 0x3fc9335e5d594989
> +        .quad 0x3fc90c6db9fcbcd9
> +        .quad 0x3fc8e588ebac2dbf
> +        .quad 0x3fc8beafeb38fe8c
> +        .quad 0x3fc897e2b17b19a5
> +        .quad 0x3fc871213750e994
> +        .quad 0x3fc84a6b759f512f
> +        .quad 0x3fc823c16551a3c2
> +        .quad 0x3fc7fd22ff599d4f
> +        .quad 0x3fc7d6903caf5ad0
> +        .quad 0x3fc7b0091651528c
> +        .quad 0x3fc7898d85444c73
> +        .quad 0x3fc7631d82935a86
> +        .quad 0x3fc73cb9074fd14d
> +        .quad 0x3fc716600c914054
> +        .quad 0x3fc6f0128b756abc
> +        .quad 0x3fc6c9d07d203fc7
> +        .quad 0x3fc6a399dabbd383
> +        .quad 0x3fc67d6e9d785771
> +        .quad 0x3fc6574ebe8c133a
> +        .quad 0x3fc6313a37335d76
> +        .quad 0x3fc60b3100b09476
> +        .quad 0x3fc5e533144c1719
> +        .quad 0x3fc5bf406b543db2
> +        .quad 0x3fc59958ff1d52f1
> +        .quad 0x3fc5737cc9018cdd
> +        .quad 0x3fc54dabc26105d2
> +        .quad 0x3fc527e5e4a1b58d
> +        .quad 0x3fc5022b292f6a45
> +        .quad 0x3fc4dc7b897bc1c8
> +        .quad 0x3fc4b6d6fefe22a4
> +        .quad 0x3fc4913d8333b561
> +        .quad 0x3fc46baf0f9f5db7
> +        .quad 0x3fc4462b9dc9b3dc
> +        .quad 0x3fc420b32740fdd4
> +        .quad 0x3fc3fb45a59928cc
> +        .quad 0x3fc3d5e3126bc27f
> +        .quad 0x3fc3b08b6757f2a9
> +        .quad 0x3fc38b3e9e027479
> +        .quad 0x3fc365fcb0159016
> +        .quad 0x3fc340c59741142e
> +        .quad 0x3fc31b994d3a4f85
> +        .quad 0x3fc2f677cbbc0a96
> +        .quad 0x3fc2d1610c86813a
> +        .quad 0x3fc2ac55095f5c59
> +        .quad 0x3fc28753bc11aba5
> +        .quad 0x3fc2625d1e6ddf57
> +        .quad 0x3fc23d712a49c202
> +        .quad 0x3fc2188fd9807263
> +        .quad 0x3fc1f3b925f25d41
> +        .quad 0x3fc1ceed09853752
> +        .quad 0x3fc1aa2b7e23f72a
> +        .quad 0x3fc185747dbecf34
> +        .quad 0x3fc160c8024b27b1
> +        .quad 0x3fc13c2605c398c3
> +        .quad 0x3fc1178e8227e47c
> +        .quad 0x3fc0f301717cf0fb
> +        .quad 0x3fc0ce7ecdccc28d
> +        .quad 0x3fc0aa06912675d5
> +        .quad 0x3fc08598b59e3a07
> +        .quad 0x3fc06135354d4b18
> +        .quad 0x3fc03cdc0a51ec0d
> +        .quad 0x3fc0188d2ecf6140
> +        .quad 0x3fbfe89139dbd566
> +        .quad 0x3fbfa01c9db57ce2
> +        .quad 0x3fbf57bc7d9005db
> +        .quad 0x3fbf0f70cdd992e3
> +        .quad 0x3fbec739830a1120
> +        .quad 0x3fbe7f1691a32d3e
> +        .quad 0x3fbe3707ee30487b
> +        .quad 0x3fbdef0d8d466db9
> +        .quad 0x3fbda727638446a2
> +        .quad 0x3fbd5f55659210e2
> +        .quad 0x3fbd179788219364
> +        .quad 0x3fbccfedbfee13a8
> +        .quad 0x3fbc885801bc4b23
> +        .quad 0x3fbc40d6425a5cb1
> +        .quad 0x3fbbf968769fca11
> +        .quad 0x3fbbb20e936d6974
> +        .quad 0x3fbb6ac88dad5b1c
> +        .quad 0x3fbb23965a52ff00
> +        .quad 0x3fbadc77ee5aea8c
> +        .quad 0x3fba956d3ecade63
> +        .quad 0x3fba4e7640b1bc38
> +        .quad 0x3fba0792e9277cac
> +        .quad 0x3fb9c0c32d4d2548
> +        .quad 0x3fb97a07024cbe74
> +        .quad 0x3fb9335e5d594989
> +        .quad 0x3fb8ecc933aeb6e8
> +        .quad 0x3fb8a6477a91dc29
> +        .quad 0x3fb85fd927506a48
> +        .quad 0x3fb8197e2f40e3f0
> +        .quad 0x3fb7d33687c293c9
> +        .quad 0x3fb78d02263d82d3
> +        .quad 0x3fb746e100226ed9
> +        .quad 0x3fb700d30aeac0e1
> +        .quad 0x3fb6bad83c1883b6
> +        .quad 0x3fb674f089365a7a
> +        .quad 0x3fb62f1be7d77743
> +        .quad 0x3fb5e95a4d9791cb
> +        .quad 0x3fb5a3abb01ade25
> +        .quad 0x3fb55e10050e0384
> +        .quad 0x3fb518874226130a
> +        .quad 0x3fb4d3115d207eac
> +        .quad 0x3fb48dae4bc31018
> +        .quad 0x3fb4485e03dbdfad
> +        .quad 0x3fb403207b414b7f
> +        .quad 0x3fb3bdf5a7d1ee64
> +        .quad 0x3fb378dd7f749714
> +        .quad 0x3fb333d7f8183f4b
> +        .quad 0x3fb2eee507b40301
> +        .quad 0x3fb2aa04a44717a5
> +        .quad 0x3fb26536c3d8c369
> +        .quad 0x3fb2207b5c78549e
> +        .quad 0x3fb1dbd2643d190b
> +        .quad 0x3fb1973bd1465567
> +        .quad 0x3fb152b799bb3cc9
> +        .quad 0x3fb10e45b3cae831
> +        .quad 0x3fb0c9e615ac4e17
> +        .quad 0x3fb08598b59e3a07
> +        .quad 0x3fb0415d89e74444
> +        .quad 0x3faffa6911ab9301
> +        .quad 0x3faf723b517fc523
> +        .quad 0x3faeea31c006b87c
> +        .quad 0x3fae624c4a0b5e1b
> +        .quad 0x3fadda8adc67ee4e
> +        .quad 0x3fad52ed6405d86f
> +        .quad 0x3faccb73cdddb2cc
> +        .quad 0x3fac441e06f72a9e
> +        .quad 0x3fabbcebfc68f420
> +        .quad 0x3fab35dd9b58baad
> +        .quad 0x3faaaef2d0fb10fc
> +        .quad 0x3faa282b8a936171
> +        .quad 0x3fa9a187b573de7c
> +        .quad 0x3fa91b073efd7314
> +        .quad 0x3fa894aa149fb343
> +        .quad 0x3fa80e7023d8ccc4
> +        .quad 0x3fa788595a3577ba
> +        .quad 0x3fa70265a550e777
> +        .quad 0x3fa67c94f2d4bb58
> +        .quad 0x3fa5f6e73078efb8
> +        .quad 0x3fa5715c4c03ceef
> +        .quad 0x3fa4ebf43349e26f
> +        .quad 0x3fa466aed42de3ea
> +        .quad 0x3fa3e18c1ca0ae92
> +        .quad 0x3fa35c8bfaa1306b
> +        .quad 0x3fa2d7ae5c3c5bae
> +        .quad 0x3fa252f32f8d183f
> +        .quad 0x3fa1ce5a62bc353a
> +        .quad 0x3fa149e3e4005a8d
> +        .quad 0x3fa0c58fa19dfaaa
> +        .quad 0x3fa0415d89e74444
> +        .quad 0x3f9f7a9b16782856
> +        .quad 0x3f9e72bf2813ce51
> +        .quad 0x3f9d6b2725979802
> +        .quad 0x3f9c63d2ec14aaf2
> +        .quad 0x3f9b5cc258b718e6
> +        .quad 0x3f9a55f548c5c43f
> +        .quad 0x3f994f6b99a24475
> +        .quad 0x3f98492528c8cabf
> +        .quad 0x3f974321d3d006d3
> +        .quad 0x3f963d6178690bd6
> +        .quad 0x3f9537e3f45f3565
> +        .quad 0x3f9432a925980cc1
> +        .quad 0x3f932db0ea132e22
> +        .quad 0x3f9228fb1fea2e28
> +        .quad 0x3f912487a5507f70
> +        .quad 0x3f90205658935847
> +        .quad 0x3f8e38ce3033310c
> +        .quad 0x3f8c317384c75f06
> +        .quad 0x3f8a2a9c6c170462
> +        .quad 0x3f882448a388a2aa
> +        .quad 0x3f861e77e8b53fc6
> +        .quad 0x3f841929f96832f0
> +        .quad 0x3f82145e939ef1e9
> +        .quad 0x3f8010157588de71
> +        .quad 0x3f7c189cbb0e27fb
> +        .quad 0x3f78121214586b54
> +        .quad 0x3f740c8a747878e2
> +        .quad 0x3f70080559588b35
> +        .quad 0x3f680904828985c0
> +        .quad 0x3f60040155d5889e
> +        .quad 0x3f50020055655889
> +        .quad 0x0000000000000000
> +        /*== poly_coeff[4] ==*/
> +        .align 16
> +        .quad 0x3fc9999CACDB4D0A, 0x3fc9999CACDB4D0A /* coeff4 */
> +        .quad 0xbfd0000148058EE1, 0xbfd0000148058EE1 /* coeff3 */
> +        .quad 0x3fd55555555543C5, 0x3fd55555555543C5 /* coeff2 */
> +        .quad 0xbfdFFFFFFFFFF81F, 0xbfdFFFFFFFFFF81F /* coeff1 */
> +        /*== ExpMask ==*/
> +        .align 16
> +        .quad 0x000fffffffffffff, 0x000fffffffffffff
> +        /*== Two10 ==*/
> +        .align 16
> +        .quad 0x3f50000000000000, 0x3f50000000000000
> +        /*== MinLog1p = -1+2^(-53) ==*/
> +        .align 16
> +        .quad 0xbfefffffffffffff, 0xbfefffffffffffff
> +        /*== MaxLog1p ==*/
> +        .align 16
> +        .quad 0x7f3ffffffffff000, 0x7f3ffffffffff000
> +        /*== One ==*/
> +        .align 16
> +        .quad 0x3ff0000000000000, 0x3ff0000000000000
> +        /*== SgnMask ==*/
> +        .align 16
> +        .quad 0x7fffffffffffffff, 0x7fffffffffffffff
> +        /*== XThreshold ==*/
> +        .align 16
> +        .quad 0x3e00000000000000, 0x3e00000000000000
> +        /*== XhMask ==*/
> +        .align 16
> +        .quad 0xfffffffffffffc00, 0xfffffffffffffc00
> +        /*== Threshold ==*/
> +        .align 16
> +        .quad 0x4086a00000000000, 0x4086a00000000000
> +        /*== Bias ==*/
> +        .align 16
> +        .quad 0x408ff80000000000, 0x408ff80000000000
> +        /*== Bias1 ==*/
> +        .align 16
> +        .quad 0x408ff00000000000, 0x408ff00000000000
> +        /*== ExpMask ==*/
> +        .align 16
> +        .quad 0x7ff0000000000000, 0x7ff0000000000000
> +        /*== ExpMask2 ==*/
> +        .align 16
> +        .quad 0x7f40000000000000, 0x7f40000000000000
> +        /*== L2L ==*/
> +        .align 16
> +        .quad 0x3fe62E42FEFA39EF, 0x3fe62E42FEFA39EF
> +        /*== dBigThreshold ==*/
> +        .align 16
> +        .quad 0x41D0000000000000, 0x41D0000000000000
> +        /*== dC2 ==*/
> +        .align 16
> +        .quad 0x3FD8000000000000, 0x3FD8000000000000
> +        /*== dC3 ==*/
> +        .align 16
> +        .quad 0x3FD4000000000000, 0x3FD4000000000000
> +        /*== dC4 ==*/
> +        .align 16
> +        .quad 0x3FD1800000000000, 0x3FD1800000000000
> +        /*== dC5 ==*/
> +        .align 16
> +        .quad 0x3FCF800000000000, 0x3FCF800000000000
> +        /*== dHalf ==*/
> +        .align 16
> +        .quad 0x3FE0000000000000, 0x3FE0000000000000
> +        /*== dLargestFinite ==*/
> +        .align 16
> +        .quad 0x7FEFFFFFFFFFFFFF, 0x7FEFFFFFFFFFFFFF
> +        /*== dLittleThreshold ==*/
> +        .align 16
> +        .quad 0x3F60000000000000, 0x3F60000000000000
> +        /*== dSign ==*/
> +        .align 16
> +        .quad 0x8000000000000000, 0x8000000000000000
> +        /*== dThirtyOne ==*/
> +        .align 16
> +        .quad 0x403F000000000000, 0x403F000000000000
> +        /*== dTopMask12 ==*/
> +        .align 16
> +        .quad 0xFFFFFE0000000000, 0xFFFFFE0000000000
> +        /*== dTopMask26 ==*/
> +        .align 16
> +        .quad 0xFFFFFFFFF8000000, 0xFFFFFFFFF8000000
> +        /*== dTopMask29 ==*/
> +        .align 16
> +        .quad 0xFFFFFFFFFF000000, 0xFFFFFFFFFF000000
> +        /*== XScale ==*/
> +        .align 16
> +        .quad 0x3E10000000000000, 0x3E10000000000000
> +        .align 16
> +        .type	__svml_dasinh_data_internal,@object
> +        .size	__svml_dasinh_data_internal,.-__svml_dasinh_data_internal
> +        .align 16
> +
> +.FLT_30:
> +        .long	0x00000000,0x43380000,0x00000000,0x43380000
> +        .type	.FLT_30,@object
> +        .size	.FLT_30,16
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core-sse.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core-sse.S
> new file mode 100644
> index 0000000000..903b5f0fb5
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core-sse.S
> @@ -0,0 +1,20 @@
> +/* SSE version of vectorized asinh, vector length is 4.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#define _ZGVdN4v_asinh _ZGVdN4v_asinh_sse_wrapper
> +#include "../svml_d_asinh4_core.S"
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core.c b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core.c
> new file mode 100644
> index 0000000000..e7acd032b5
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core.c
> @@ -0,0 +1,27 @@
> +/* Multiple versions of vectorized asinh, vector length is 4.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#define SYMBOL_NAME _ZGVdN4v_asinh
> +#include "ifunc-mathvec-avx2.h"
> +
> +libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
> +
> +#ifdef SHARED
> +__hidden_ver1 (_ZGVdN4v_asinh, __GI__ZGVdN4v_asinh, __redirect__ZGVdN4v_asinh)
> +  __attribute__ ((visibility ("hidden")));
> +#endif
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core_avx2.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core_avx2.S
> new file mode 100644
> index 0000000000..d691d1ec6f
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core_avx2.S
> @@ -0,0 +1,1601 @@
> +/* Function asinh vectorized with AVX2.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   https://www.gnu.org/licenses/.  */
> +
> +/*
> + * ALGORITHM DESCRIPTION:
> + *
> + *   Compute asinh(x) as log(x + sqrt(x*x + 1))
> + *
> + *   Special cases:
> + *
> + *   asinh(NaN) = quiet NaN, and raise invalid exception
> + *   asinh(INF) = that INF
> + *   asinh(0)   = that 0
> + *
> + */
> +
> +/* Offsets for data table __svml_dasinh_data_internal
> + */
> +#define Log_HA_table                  	0
> +#define Log_LA_table                  	8224
> +#define poly_coeff                    	12352
> +#define ExpMask                       	12480
> +#define Two10                         	12512
> +#define MinLog1p                      	12544
> +#define MaxLog1p                      	12576
> +#define One                           	12608
> +#define SgnMask                       	12640
> +#define XThreshold                    	12672
> +#define XhMask                        	12704
> +#define Threshold                     	12736
> +#define Bias                          	12768
> +#define Bias1                         	12800
> +#define ExpMask0                      	12832
> +#define ExpMask2                      	12864
> +#define L2                            	12896
> +#define dBigThreshold                 	12928
> +#define dC2                           	12960
> +#define dC3                           	12992
> +#define dC4                           	13024
> +#define dC5                           	13056
> +#define dHalf                         	13088
> +#define dLargestFinite                	13120
> +#define dLittleThreshold              	13152
> +#define dSign                         	13184
> +#define dThirtyOne                    	13216
> +#define dTopMask12                    	13248
> +#define dTopMask29                    	13280
> +#define XScale                        	13312
> +
> +/* Lookup bias for data table __svml_dasinh_data_internal.  */
> +#define Table_Lookup_Bias               -0x405fe0
> +
> +#include <sysdep.h>
> +
> +        .text
> +	.section .text.avx2,"ax",@progbits
> +ENTRY(_ZGVdN4v_asinh_avx2)
> +        pushq     %rbp
> +        cfi_def_cfa_offset(16)
> +        movq      %rsp, %rbp
> +        cfi_def_cfa(6, 16)
> +        cfi_offset(6, -16)
> +        andq      $-32, %rsp
> +        subq      $96, %rsp
> +        lea       Table_Lookup_Bias+__svml_dasinh_data_internal(%rip), %r8
> +        vmovapd   %ymm0, %ymm13
> +        vmovupd   SgnMask+__svml_dasinh_data_internal(%rip), %ymm9
> +
> +/* Load the constant 1 and a sign mask */
> +        vmovupd   One+__svml_dasinh_data_internal(%rip), %ymm12
> +
> +/* No need to split X when FMA is available in hardware. */
> +        vmulpd    %ymm13, %ymm13, %ymm8
> +
> +/*
> + * Get the absolute value of the input, since we will exploit antisymmetry
> + * and mostly assume X >= 0 in the core computation
> + */
> +        vandpd    %ymm9, %ymm13, %ymm10
> +
> +/*
> + * Check whether the input is finite, by checking |X| <= MaxFloat
> + * Otherwise set the rangemask so that the callout will get used.
> + * Note that this will also use the callout for NaNs since not(NaN <= MaxFloat)
> + */
> +        vcmpnle_uqpd dLargestFinite+__svml_dasinh_data_internal(%rip), %ymm10, %ymm14
> +
> +/*
> + * Finally, express Y + W = X^2 + 1 accurately where Y has <= 29 bits.
> + * If |X| <= 1 then |XHi| <= 1 and so |X2Hi| <= 1, so we can treat 1
> + * as the dominant component in the compensated summation. Otherwise,
> + * if |X| >= 1, then since X2Hi only has 52 significant bits, the basic
> + * addition will be exact anyway until we get to |X| >= 2^53. But by
> + * that time the log function is well-conditioned enough that the
> + * rounding error doesn't matter. Hence we can treat 1 as dominant even
> + * if it literally isn't.
> + */
> +        vaddpd    %ymm8, %ymm12, %ymm5
> +
> +/*
> + * The following computation can go wrong for very large X, basically
> + * because X^2 overflows. But for large X we have
> + * asinh(X) / log(2 X) - 1 =~= 1/(4 * X^2), so for X >= 2^30
> + * we can just later stick X back into the log and tweak up the exponent.
> + * Actually we scale X by 2^-30 and tweak the exponent up by 31,
> + * to stay in the safe range for the later log computation.
> + * Compute a flag now telling us when do do this.
> + */
> +        vcmplt_oqpd dBigThreshold+__svml_dasinh_data_internal(%rip), %ymm10, %ymm11
> +        vsubpd    %ymm5, %ymm12, %ymm15
> +        vmovmskpd %ymm14, %eax
> +        vandpd    dTopMask29+__svml_dasinh_data_internal(%rip), %ymm5, %ymm14
> +
> +/*
> + * Compute R = 1/sqrt(Y + W) * (1 + d)
> + * Force R to <= 12 significant bits in case it isn't already
> + * This means that R * Y and R^2 * Y are exactly representable.
> + */
> +        vcvtpd2ps %ymm14, %xmm1
> +        vaddpd    %ymm15, %ymm8, %ymm0
> +        vsubpd    %ymm14, %ymm5, %ymm2
> +        vrsqrtps  %xmm1, %xmm3
> +        vmovapd   %ymm13, %ymm7
> +        vfmsub213pd %ymm8, %ymm13, %ymm7
> +        vcvtps2pd %xmm3, %ymm6
> +        vaddpd    %ymm0, %ymm7, %ymm4
> +
> +/*
> + * Unfortunately, we can still be in trouble if |X| <= 2^-10, since
> + * the absolute error 2^-(12+53)-ish in sqrt(1 + X^2) gets scaled up
> + * by 1/X and comes close to our threshold. Hence if |X| <= 2^-9,
> + * perform an alternative computation
> + * sqrt(1 + X^2) - 1 = X^2/2 - X^4/8 + X^6/16
> + * X2 = X^2
> + */
> +        vaddpd    %ymm7, %ymm8, %ymm7
> +        vaddpd    %ymm2, %ymm4, %ymm15
> +
> +/*
> + * Now       1 / (1 + d)
> + * = 1 / (1 + (sqrt(1 - e) - 1))
> + * = 1 / sqrt(1 - e)
> + * = 1 + 1/2 * e + 3/8 * e^2 + 5/16 * e^3 + 35/128 * e^4 +
> + * 63/256 * e^5 + 231/1024 * e^6 + ....
> + * So compute the first five nonconstant terms of that, so that
> + * we have a relative correction (1 + Corr) to apply to S etc.
> + * C1 = 1/2
> + * C2 = 3/8
> + * C3 = 5/16
> + * C4 = 35/128
> + * C5 = 63/256
> + */
> +        vmovupd   dC5+__svml_dasinh_data_internal(%rip), %ymm4
> +        vandpd    dTopMask12+__svml_dasinh_data_internal(%rip), %ymm6, %ymm0
> +
> +/*
> + * Compute S = (Y/sqrt(Y + W)) * (1 + d)
> + * and T = (W/sqrt(Y + W)) * (1 + d)
> + * so that S + T = sqrt(Y + W) * (1 + d)
> + * S is exact, and the rounding error in T is OK.
> + */
> +        vmulpd    %ymm0, %ymm14, %ymm3
> +        vmulpd    %ymm15, %ymm0, %ymm1
> +        vmovupd   dHalf+__svml_dasinh_data_internal(%rip), %ymm6
> +        vsubpd    %ymm12, %ymm3, %ymm14
> +
> +/*
> + * Obtain sqrt(1 + X^2) - 1 in two pieces
> + * sqrt(1 + X^2) - 1
> + * = sqrt(Y + W) - 1
> + * = (S + T) * (1 + Corr) - 1
> + * = [S - 1] + [T + (S + T) * Corr]
> + * We need a compensated summation for the last part. We treat S - 1
> + * as the larger part; it certainly is until about X < 2^-4, and in that
> + * case, the error is affordable since X dominates over sqrt(1 + X^2) - 1
> + * Final sum is dTmp5 (hi) + dTmp7 (lo)
> + */
> +        vaddpd    %ymm1, %ymm3, %ymm2
> +
> +/*
> + * Compute e = -(2 * d + d^2)
> + * The first FMR is exact, and the rounding error in the other is acceptable
> + * since d and e are ~ 2^-12
> + */
> +        vmovapd   %ymm12, %ymm5
> +        vfnmadd231pd %ymm3, %ymm0, %ymm5
> +        vfnmadd231pd %ymm1, %ymm0, %ymm5
> +        vfmadd213pd dC4+__svml_dasinh_data_internal(%rip), %ymm5, %ymm4
> +        vfmadd213pd dC3+__svml_dasinh_data_internal(%rip), %ymm5, %ymm4
> +        vfmadd213pd dC2+__svml_dasinh_data_internal(%rip), %ymm5, %ymm4
> +        vfmadd213pd %ymm6, %ymm5, %ymm4
> +        vmulpd    %ymm4, %ymm5, %ymm0
> +        vfmadd213pd %ymm1, %ymm2, %ymm0
> +
> +/* Now multiplex the two possible computations */
> +        vcmple_oqpd dLittleThreshold+__svml_dasinh_data_internal(%rip), %ymm10, %ymm2
> +        vaddpd    %ymm14, %ymm0, %ymm15
> +
> +/* dX2over2 = X^2/2 */
> +        vmulpd    %ymm7, %ymm6, %ymm0
> +
> +/* dX4over4 = X^4/4 */
> +        vmulpd    %ymm0, %ymm0, %ymm8
> +
> +/* dX46 = -X^4/4 + X^6/8 */
> +        vfmsub231pd %ymm0, %ymm8, %ymm8
> +
> +/* dX46over2 = -X^4/8 + x^6/16 */
> +        vmulpd    %ymm8, %ymm6, %ymm5
> +
> +/* 2^ (-10-exp(X) ) */
> +        vmovupd   ExpMask2+__svml_dasinh_data_internal(%rip), %ymm8
> +        vaddpd    %ymm5, %ymm0, %ymm4
> +        vblendvpd %ymm2, %ymm4, %ymm15, %ymm1
> +
> +/*
> + * Now do another compensated sum to add |X| + [sqrt(1 + X^2) - 1].
> + * It's always safe to assume |X| is larger.
> + * This is the final 2-part argument to the log1p function
> + */
> +        vaddpd    %ymm1, %ymm10, %ymm3
> +
> +/* Now multiplex to the case X = 2^-30 * |input|, Xl = dL = 0 in the "big" case. */
> +        vmulpd    XScale+__svml_dasinh_data_internal(%rip), %ymm10, %ymm10
> +
> +/*
> + * Now we feed into the log1p code, using H in place of _VARG1 and
> + * also adding L into Xl.
> + * compute 1+x as high, low parts
> + */
> +        vmaxpd    %ymm3, %ymm12, %ymm6
> +        vminpd    %ymm3, %ymm12, %ymm7
> +        vandpd    %ymm9, %ymm3, %ymm9
> +        vcmplt_oqpd XThreshold+__svml_dasinh_data_internal(%rip), %ymm9, %ymm0
> +        vaddpd    %ymm7, %ymm6, %ymm5
> +        vorpd     XhMask+__svml_dasinh_data_internal(%rip), %ymm0, %ymm4
> +        vandpd    %ymm4, %ymm5, %ymm1
> +        vblendvpd %ymm11, %ymm1, %ymm10, %ymm5
> +        vsubpd    %ymm1, %ymm6, %ymm2
> +
> +/* exponent bits */
> +        vpsrlq    $20, %ymm5, %ymm10
> +        vaddpd    %ymm2, %ymm7, %ymm3
> +
> +/*
> + * Now resume the main code.
> + * preserve mantissa, set input exponent to 2^(-10)
> + */
> +        vandpd    ExpMask+__svml_dasinh_data_internal(%rip), %ymm5, %ymm0
> +        vorpd     Two10+__svml_dasinh_data_internal(%rip), %ymm0, %ymm2
> +
> +/* reciprocal approximation good to at least 11 bits */
> +        vcvtpd2ps %ymm2, %xmm6
> +        vrcpps    %xmm6, %xmm7
> +        vcvtps2pd %xmm7, %ymm15
> +
> +/* exponent of X needed to scale Xl */
> +        vandps    ExpMask0+__svml_dasinh_data_internal(%rip), %ymm5, %ymm9
> +        vpsubq    %ymm9, %ymm8, %ymm0
> +        vandpd    %ymm11, %ymm3, %ymm4
> +
> +/* round reciprocal to nearest integer, will have 1+9 mantissa bits */
> +        vroundpd  $0, %ymm15, %ymm3
> +
> +/* scale DblRcp */
> +        vmulpd    %ymm0, %ymm3, %ymm2
> +
> +/* argument reduction */
> +        vfmsub213pd %ymm12, %ymm2, %ymm5
> +        vmulpd    %ymm2, %ymm4, %ymm12
> +        vmovupd   poly_coeff+64+__svml_dasinh_data_internal(%rip), %ymm2
> +        vaddpd    %ymm12, %ymm5, %ymm5
> +        vfmadd213pd poly_coeff+96+__svml_dasinh_data_internal(%rip), %ymm5, %ymm2
> +        vmulpd    %ymm5, %ymm5, %ymm4
> +        vextractf128 $1, %ymm10, %xmm14
> +        vshufps   $221, %xmm14, %xmm10, %xmm1
> +
> +/* biased exponent in DP format */
> +        vcvtdq2pd %xmm1, %ymm7
> +
> +/* exponent*log(2.0) */
> +        vmovupd   Threshold+__svml_dasinh_data_internal(%rip), %ymm10
> +
> +/* Add 31 to the exponent in the "large" case to get log(2 * input) */
> +        vaddpd    dThirtyOne+__svml_dasinh_data_internal(%rip), %ymm7, %ymm6
> +        vblendvpd %ymm11, %ymm7, %ymm6, %ymm1
> +
> +/*
> + * prepare table index
> + * table lookup
> + */
> +        vpsrlq    $40, %ymm3, %ymm11
> +        vcmplt_oqpd %ymm3, %ymm10, %ymm3
> +        vandpd    Bias+__svml_dasinh_data_internal(%rip), %ymm3, %ymm14
> +        vorpd     Bias1+__svml_dasinh_data_internal(%rip), %ymm14, %ymm15
> +        vsubpd    %ymm15, %ymm1, %ymm1
> +        vmulpd    L2+__svml_dasinh_data_internal(%rip), %ymm1, %ymm3
> +
> +/* polynomial */
> +        vmovupd   poly_coeff+__svml_dasinh_data_internal(%rip), %ymm1
> +        vfmadd213pd poly_coeff+32+__svml_dasinh_data_internal(%rip), %ymm5, %ymm1
> +        vfmadd213pd %ymm2, %ymm4, %ymm1
> +
> +/* reconstruction */
> +        vfmadd213pd %ymm5, %ymm4, %ymm1
> +        vextractf128 $1, %ymm11, %xmm7
> +        vmovd     %xmm11, %edx
> +        vmovd     %xmm7, %esi
> +        movslq    %edx, %rdx
> +        vpextrd   $2, %xmm11, %ecx
> +        movslq    %esi, %rsi
> +        vpextrd   $2, %xmm7, %edi
> +        movslq    %ecx, %rcx
> +        movslq    %edi, %rdi
> +        vmovsd    (%r8,%rdx), %xmm0
> +        vmovsd    (%r8,%rsi), %xmm8
> +        vmovhpd   (%r8,%rcx), %xmm0, %xmm6
> +        vmovhpd   (%r8,%rdi), %xmm8, %xmm9
> +        vinsertf128 $1, %xmm9, %ymm6, %ymm0
> +        vaddpd    %ymm1, %ymm0, %ymm0
> +        vaddpd    %ymm0, %ymm3, %ymm7
> +
> +/* Finally, reincorporate the original sign. */
> +        vandpd    dSign+__svml_dasinh_data_internal(%rip), %ymm13, %ymm6
> +        vxorpd    %ymm7, %ymm6, %ymm0
> +        testl     %eax, %eax
> +
> +/* Go to special inputs processing branch */
> +        jne       L(SPECIAL_VALUES_BRANCH)
> +                                # LOE rbx r12 r13 r14 r15 eax ymm0 ymm13
> +
> +/* Restore registers
> + * and exit the function
> + */
> +
> +L(EXIT):
> +        movq      %rbp, %rsp
> +        popq      %rbp
> +        cfi_def_cfa(7, 8)
> +        cfi_restore(6)
> +        ret
> +        cfi_def_cfa(6, 16)
> +        cfi_offset(6, -16)
> +
> +/* Branch to process
> + * special inputs
> + */
> +
> +L(SPECIAL_VALUES_BRANCH):
> +        vmovupd   %ymm13, 32(%rsp)
> +        vmovupd   %ymm0, 64(%rsp)
> +                                # LOE rbx r12 r13 r14 r15 eax ymm0
> +
> +        xorl      %edx, %edx
> +                                # LOE rbx r12 r13 r14 r15 eax edx
> +
> +        vzeroupper
> +        movq      %r12, 16(%rsp)
> +        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -80; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xb0, 0xff, 0xff, 0xff, 0x22
> +        movl      %edx, %r12d
> +        movq      %r13, 8(%rsp)
> +        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -88; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa8, 0xff, 0xff, 0xff, 0x22
> +        movl      %eax, %r13d
> +        movq      %r14, (%rsp)
> +        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -96; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa0, 0xff, 0xff, 0xff, 0x22
> +                                # LOE rbx r15 r12d r13d
> +
> +/* Range mask
> + * bits check
> + */
> +
> +L(RANGEMASK_CHECK):
> +        btl       %r12d, %r13d
> +
> +/* Call scalar math function */
> +        jc        L(SCALAR_MATH_CALL)
> +                                # LOE rbx r15 r12d r13d
> +
> +/* Special inputs
> + * processing loop
> + */
> +
> +L(SPECIAL_VALUES_LOOP):
> +        incl      %r12d
> +        cmpl      $4, %r12d
> +
> +/* Check bits in range mask */
> +        jl        L(RANGEMASK_CHECK)
> +                                # LOE rbx r15 r12d r13d
> +
> +        movq      16(%rsp), %r12
> +        cfi_restore(12)
> +        movq      8(%rsp), %r13
> +        cfi_restore(13)
> +        movq      (%rsp), %r14
> +        cfi_restore(14)
> +        vmovupd   64(%rsp), %ymm0
> +
> +/* Go to exit */
> +        jmp       L(EXIT)
> +        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -80; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xb0, 0xff, 0xff, 0xff, 0x22
> +        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -88; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa8, 0xff, 0xff, 0xff, 0x22
> +        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -96; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa0, 0xff, 0xff, 0xff, 0x22
> +                                # LOE rbx r12 r13 r14 r15 ymm0
> +
> +/* Scalar math fucntion call
> + * to process special input
> + */
> +
> +L(SCALAR_MATH_CALL):
> +        movl      %r12d, %r14d
> +        movsd     32(%rsp,%r14,8), %xmm0
> +        call      asinh@PLT
> +                                # LOE rbx r14 r15 r12d r13d xmm0
> +
> +        movsd     %xmm0, 64(%rsp,%r14,8)
> +
> +/* Process special inputs in loop */
> +        jmp       L(SPECIAL_VALUES_LOOP)
> +                                # LOE rbx r15 r12d r13d
> +END(_ZGVdN4v_asinh_avx2)
> +
> +        .section .rodata, "a"
> +        .align 32
> +
> +#ifdef __svml_dasinh_data_internal_typedef
> +typedef unsigned int VUINT32;
> +typedef struct {
> +        __declspec(align(32)) VUINT32 Log_HA_table[(1<<10)+2][2];
> +        __declspec(align(32)) VUINT32 Log_LA_table[(1<<9)+1][2];
> +        __declspec(align(32)) VUINT32 poly_coeff[4][4][2];
> +        __declspec(align(32)) VUINT32 ExpMask[4][2];
> +        __declspec(align(32)) VUINT32 Two10[4][2];
> +        __declspec(align(32)) VUINT32 MinLog1p[4][2];
> +        __declspec(align(32)) VUINT32 MaxLog1p[4][2];
> +        __declspec(align(32)) VUINT32 One[4][2];
> +        __declspec(align(32)) VUINT32 SgnMask[4][2];
> +        __declspec(align(32)) VUINT32 XThreshold[4][2];
> +        __declspec(align(32)) VUINT32 XhMask[4][2];
> +        __declspec(align(32)) VUINT32 Threshold[4][2];
> +        __declspec(align(32)) VUINT32 Bias[4][2];
> +        __declspec(align(32)) VUINT32 Bias1[4][2];
> +        __declspec(align(32)) VUINT32 ExpMask0[4][2];
> +        __declspec(align(32)) VUINT32 ExpMask2[4][2];
> +        __declspec(align(32)) VUINT32 L2[4][2];
> +        __declspec(align(32)) VUINT32 dBigThreshold[4][2];
> +        __declspec(align(32)) VUINT32 dC2[4][2];
> +        __declspec(align(32)) VUINT32 dC3[4][2];
> +        __declspec(align(32)) VUINT32 dC4[4][2];
> +        __declspec(align(32)) VUINT32 dC5[4][2];
> +        __declspec(align(32)) VUINT32 dHalf[4][2];
> +        __declspec(align(32)) VUINT32 dLargestFinite[4][2];
> +        __declspec(align(32)) VUINT32 dLittleThreshold[4][2];
> +        __declspec(align(32)) VUINT32 dSign[4][2];
> +        __declspec(align(32)) VUINT32 dThirtyOne[4][2];
> +        __declspec(align(32)) VUINT32 dTopMask12[4][2];
> +        __declspec(align(32)) VUINT32 dTopMask29[4][2];
> +        __declspec(align(32)) VUINT32 XScale[4][2];
> +} __svml_dasinh_data_internal;
> +#endif
> +__svml_dasinh_data_internal:
> +        /* Log_HA_table */
> +        .quad 0xc086232bdd7a8300, 0xbe1ce91eef3fb100
> +        .quad 0xc086232fdc7ad828, 0xbe1cefcffda73b6a
> +        .quad 0xc0862333d97d2ba0, 0xbe1cef406748f1ff
> +        .quad 0xc0862337d48378e0, 0xbe1cef2a9429925a
> +        .quad 0xc086233bcd8fb878, 0xbe1cf138d17ebecb
> +        .quad 0xc086233fc4a3e018, 0xbe1ceff2dbbbb29e
> +        .quad 0xc0862343b9c1e270, 0xbe1cf1a42aae437b
> +        .quad 0xc0862347acebaf68, 0xbe1cef3b152048af
> +        .quad 0xc086234b9e2333f0, 0xbe1cef20e127805e
> +        .quad 0xc086234f8d6a5a30, 0xbe1cf00ad6052cf4
> +        .quad 0xc08623537ac30980, 0xbe1cefc4642ee597
> +        .quad 0xc0862357662f2660, 0xbe1cf1f277d36e16
> +        .quad 0xc086235b4fb092a0, 0xbe1ceed009e8d8e6
> +        .quad 0xc086235f37492d28, 0xbe1cf1e4038cb362
> +        .quad 0xc08623631cfad250, 0xbe1cf0b0873b8557
> +        .quad 0xc086236700c75b98, 0xbe1cf15bb3227c0b
> +        .quad 0xc086236ae2b09fe0, 0xbe1cf151ef8ca9ed
> +        .quad 0xc086236ec2b87358, 0xbe1cefe1dc2cd2ed
> +        .quad 0xc0862372a0e0a780, 0xbe1cf0d1eec5454f
> +        .quad 0xc08623767d2b0b48, 0xbe1ceeefd570bbce
> +        .quad 0xc086237a57996af0, 0xbe1cee99ae91b3a7
> +        .quad 0xc086237e302d9028, 0xbe1cf0412830fbd1
> +        .quad 0xc086238206e94218, 0xbe1ceee898588610
> +        .quad 0xc0862385dbce4548, 0xbe1cee9a1fbcaaea
> +        .quad 0xc0862389aede5bc0, 0xbe1ceed8e7cc1ad6
> +        .quad 0xc086238d801b4500, 0xbe1cf10c8d059da6
> +        .quad 0xc08623914f86be18, 0xbe1ceee6c63a8165
> +        .quad 0xc08623951d228180, 0xbe1cf0c3592d2ff1
> +        .quad 0xc0862398e8f04758, 0xbe1cf0026cc4cb1b
> +        .quad 0xc086239cb2f1c538, 0xbe1cf15d48d8e670
> +        .quad 0xc08623a07b28ae60, 0xbe1cef359363787c
> +        .quad 0xc08623a44196b390, 0xbe1cefdf1ab2e82c
> +        .quad 0xc08623a8063d8338, 0xbe1cefe43c02aa84
> +        .quad 0xc08623abc91ec960, 0xbe1cf044f5ae35b7
> +        .quad 0xc08623af8a3c2fb8, 0xbe1cf0b0b4001e1b
> +        .quad 0xc08623b349975d98, 0xbe1cf1bae76dfbcf
> +        .quad 0xc08623b70731f810, 0xbe1cef0a72e13a62
> +        .quad 0xc08623bac30da1c8, 0xbe1cf184007d2b6b
> +        .quad 0xc08623be7d2bfb40, 0xbe1cf16f4b239e98
> +        .quad 0xc08623c2358ea2a0, 0xbe1cf0976acada87
> +        .quad 0xc08623c5ec3733d0, 0xbe1cf066318a16ff
> +        .quad 0xc08623c9a1274880, 0xbe1ceffaa7148798
> +        .quad 0xc08623cd54607820, 0xbe1cf23ab02e9b6e
> +        .quad 0xc08623d105e45800, 0xbe1cefdfef7d4fde
> +        .quad 0xc08623d4b5b47b20, 0xbe1cf17fece44f2b
> +        .quad 0xc08623d863d27270, 0xbe1cf18f907d0d7c
> +        .quad 0xc08623dc103fccb0, 0xbe1cee61fe072c98
> +        .quad 0xc08623dfbafe1668, 0xbe1cf022dd891e2f
> +        .quad 0xc08623e3640eda20, 0xbe1ceecc1daf4358
> +        .quad 0xc08623e70b73a028, 0xbe1cf0173c4fa380
> +        .quad 0xc08623eab12deec8, 0xbe1cf16a2150c2f4
> +        .quad 0xc08623ee553f4a30, 0xbe1cf1bf980b1f4b
> +        .quad 0xc08623f1f7a93480, 0xbe1cef8b731663c2
> +        .quad 0xc08623f5986d2dc0, 0xbe1cee9a664d7ef4
> +        .quad 0xc08623f9378cb3f0, 0xbe1cf1eda2af6400
> +        .quad 0xc08623fcd5094320, 0xbe1cf1923f9d68d7
> +        .quad 0xc086240070e45548, 0xbe1cf0747cd3e03a
> +        .quad 0xc08624040b1f6260, 0xbe1cf22ee855bd6d
> +        .quad 0xc0862407a3bbe078, 0xbe1cf0d57360c00b
> +        .quad 0xc086240b3abb4398, 0xbe1ceebc815cd575
> +        .quad 0xc086240ed01efdd0, 0xbe1cf03bfb970951
> +        .quad 0xc086241263e87f50, 0xbe1cf16e74768529
> +        .quad 0xc0862415f6193658, 0xbe1cefec64b8becb
> +        .quad 0xc086241986b28f30, 0xbe1cf0838d210baa
> +        .quad 0xc086241d15b5f448, 0xbe1cf0ea86e75b11
> +        .quad 0xc0862420a324ce28, 0xbe1cf1708d11d805
> +        .quad 0xc08624242f008380, 0xbe1ceea988c5a417
> +        .quad 0xc0862427b94a7910, 0xbe1cef166a7bbca5
> +        .quad 0xc086242b420411d0, 0xbe1cf0c9d9e86a38
> +        .quad 0xc086242ec92eaee8, 0xbe1cef0946455411
> +        .quad 0xc08624324ecbaf98, 0xbe1cefea60907739
> +        .quad 0xc0862435d2dc7160, 0xbe1cf1ed0934ce42
> +        .quad 0xc086243955624ff8, 0xbe1cf191ba746c7d
> +        .quad 0xc086243cd65ea548, 0xbe1ceeec78cf2a7e
> +        .quad 0xc086244055d2c968, 0xbe1cef345284c119
> +        .quad 0xc0862443d3c012b8, 0xbe1cf24f77355219
> +        .quad 0xc08624475027d5e8, 0xbe1cf05bf087e114
> +        .quad 0xc086244acb0b65d0, 0xbe1cef3504a32189
> +        .quad 0xc086244e446c1398, 0xbe1ceff54b2a406f
> +        .quad 0xc0862451bc4b2eb8, 0xbe1cf0757d54ed4f
> +        .quad 0xc086245532aa04f0, 0xbe1cf0c8099fdfd5
> +        .quad 0xc0862458a789e250, 0xbe1cf0b173796a31
> +        .quad 0xc086245c1aec1138, 0xbe1cf11d8734540d
> +        .quad 0xc086245f8cd1da60, 0xbe1cf1916a723ceb
> +        .quad 0xc0862462fd3c84d8, 0xbe1cf19a911e1da7
> +        .quad 0xc08624666c2d5608, 0xbe1cf23a9ef72e4f
> +        .quad 0xc0862469d9a591c0, 0xbe1cef503d947663
> +        .quad 0xc086246d45a67a18, 0xbe1cf0fceeb1a0b2
> +        .quad 0xc0862470b0314fa8, 0xbe1cf107e27e4fbc
> +        .quad 0xc086247419475160, 0xbe1cf03dd9922331
> +        .quad 0xc086247780e9bc98, 0xbe1cefce1a10e129
> +        .quad 0xc086247ae719cd18, 0xbe1ceea47f73c4f6
> +        .quad 0xc086247e4bd8bd10, 0xbe1ceec0ac56d100
> +        .quad 0xc0862481af27c528, 0xbe1cee8a6593278a
> +        .quad 0xc086248511081c70, 0xbe1cf2231dd9dec7
> +        .quad 0xc0862488717af888, 0xbe1cf0b4b8ed7da8
> +        .quad 0xc086248bd0818d68, 0xbe1cf1bd8d835002
> +        .quad 0xc086248f2e1d0d98, 0xbe1cf259acc107f4
> +        .quad 0xc08624928a4eaa20, 0xbe1cee897636b00c
> +        .quad 0xc0862495e5179270, 0xbe1cee757f20c326
> +        .quad 0xc08624993e78f490, 0xbe1cefafd3aa54a4
> +        .quad 0xc086249c9673fd10, 0xbe1cee7298d38b97
> +        .quad 0xc086249fed09d6f8, 0xbe1ceedc158d4ceb
> +        .quad 0xc08624a3423babe0, 0xbe1cf2282987cb2e
> +        .quad 0xc08624a6960aa400, 0xbe1cefe7381ecc4b
> +        .quad 0xc08624a9e877e600, 0xbe1cef328dbbce80
> +        .quad 0xc08624ad39849728, 0xbe1cefde45f3cc71
> +        .quad 0xc08624b08931db58, 0xbe1cefa8b89433b9
> +        .quad 0xc08624b3d780d500, 0xbe1cef6773c0b139
> +        .quad 0xc08624b72472a528, 0xbe1cf031c931c11f
> +        .quad 0xc08624ba70086b78, 0xbe1cf088f49275e7
> +        .quad 0xc08624bdba434630, 0xbe1cf17de0eaa86d
> +        .quad 0xc08624c103245238, 0xbe1cefd492f1ba75
> +        .quad 0xc08624c44aacab08, 0xbe1cf1253e154466
> +        .quad 0xc08624c790dd6ad0, 0xbe1cf0fb09ee6d55
> +        .quad 0xc08624cad5b7aa58, 0xbe1cf1f08dd048fe
> +        .quad 0xc08624ce193c8120, 0xbe1ceeca0809697f
> +        .quad 0xc08624d15b6d0538, 0xbe1cef8d5662d968
> +        .quad 0xc08624d49c4a4b78, 0xbe1cee97b556ed78
> +        .quad 0xc08624d7dbd56750, 0xbe1cf1b14b6acb75
> +        .quad 0xc08624db1a0f6b00, 0xbe1cef1e860623f2
> +        .quad 0xc08624de56f96758, 0xbe1ceeaf4d156f3d
> +        .quad 0xc08624e192946bf0, 0xbe1ceecc12b400ed
> +        .quad 0xc08624e4cce18710, 0xbe1cf180c40c794f
> +        .quad 0xc08624e805e1c5c8, 0xbe1cf185a08f7f65
> +        .quad 0xc08624eb3d9633d8, 0xbe1cef45fc924078
> +        .quad 0xc08624ee73ffdbb0, 0xbe1cf1e4f457f32a
> +        .quad 0xc08624f1a91fc6a0, 0xbe1cf040147b8a5a
> +        .quad 0xc08624f4dcf6fc98, 0xbe1cf1effca0dfb2
> +        .quad 0xc08624f80f868468, 0xbe1cf0470146e5bc
> +        .quad 0xc08624fb40cf6390, 0xbe1cef4dd186e501
> +        .quad 0xc08624fe70d29e60, 0xbe1ceebe257f66c7
> +        .quad 0xc08625019f9137f0, 0xbe1ceefb7a1c395c
> +        .quad 0xc0862504cd0c3220, 0xbe1cf209dedfed8c
> +        .quad 0xc0862507f9448db0, 0xbe1cf082da464994
> +        .quad 0xc086250b243b4a18, 0xbe1cee88694a73cf
> +        .quad 0xc086250e4df165a0, 0xbe1cf0b61e8f0531
> +        .quad 0xc08625117667dd78, 0xbe1cf1106599c962
> +        .quad 0xc08625149d9fad98, 0xbe1ceff1ee88af1f
> +        .quad 0xc0862517c399d0c8, 0xbe1cf0f746994ef6
> +        .quad 0xc086251ae85740b8, 0xbe1cefe8a1d077e4
> +        .quad 0xc086251e0bd8f5e0, 0xbe1cf1a1da036092
> +        .quad 0xc08625212e1fe7a8, 0xbe1cf0f8a7786fcd
> +        .quad 0xc08625244f2d0c48, 0xbe1cefa1174a07a7
> +        .quad 0xc08625276f0158d8, 0xbe1cef1043aa5b25
> +        .quad 0xc086252a8d9dc150, 0xbe1cf15d521c169d
> +        .quad 0xc086252dab033898, 0xbe1cf220bba8861f
> +        .quad 0xc0862530c732b078, 0xbe1cef51e310eae2
> +        .quad 0xc0862533e22d1988, 0xbe1cf222fcedd8ae
> +        .quad 0xc0862536fbf36370, 0xbe1cefdb4da4bda8
> +        .quad 0xc086253a14867ca0, 0xbe1ceeafc1112171
> +        .quad 0xc086253d2be75280, 0xbe1cee99dfb4b408
> +        .quad 0xc08625404216d160, 0xbe1cf22d2536f06b
> +        .quad 0xc08625435715e498, 0xbe1cef6abbf2e268
> +        .quad 0xc08625466ae57648, 0xbe1cf093a14789f5
> +        .quad 0xc08625497d866fa0, 0xbe1cf0f93655603c
> +        .quad 0xc086254c8ef9b8b8, 0xbe1cf1cc40c9aafc
> +        .quad 0xc086254f9f4038a8, 0xbe1ceeea5f4e9157
> +        .quad 0xc0862552ae5ad568, 0xbe1cefa9f52d4997
> +        .quad 0xc0862555bc4a7400, 0xbe1cefa490a638ff
> +        .quad 0xc0862558c90ff868, 0xbe1cef7fcf797d6f
> +        .quad 0xc086255bd4ac4590, 0xbe1cf1b4c51113c9
> +        .quad 0xc086255edf203d78, 0xbe1cef55e5b4a55d
> +        .quad 0xc0862561e86cc100, 0xbe1cf0d37a25f9dc
> +        .quad 0xc0862564f092b028, 0xbe1ceebe9efc19d9
> +        .quad 0xc0862567f792e9d8, 0xbe1cee8ad30a57b5
> +        .quad 0xc086256afd6e4c08, 0xbe1cef4e1817b90b
> +        .quad 0xc086256e0225b3b8, 0xbe1cee7fa9229996
> +        .quad 0xc086257105b9fce0, 0xbe1cf0b54963d945
> +        .quad 0xc0862574082c0298, 0xbe1cee5f2f3c7995
> +        .quad 0xc0862577097c9ee0, 0xbe1cf0828e303a2c
> +        .quad 0xc086257a09acaae0, 0xbe1cf172c3078947
> +        .quad 0xc086257d08bcfec0, 0xbe1cf189252afa22
> +        .quad 0xc086258006ae71b8, 0xbe1cefdb80426923
> +        .quad 0xc08625830381da08, 0xbe1ceef1391a0372
> +        .quad 0xc0862585ff380d00, 0xbe1cf17720c78d13
> +        .quad 0xc0862588f9d1df18, 0xbe1ceef1f9027d83
> +        .quad 0xc086258bf35023b8, 0xbe1cf06fac99dec9
> +        .quad 0xc086258eebb3ad78, 0xbe1cf1373eeb45c0
> +        .quad 0xc0862591e2fd4e00, 0xbe1cef777536bb81
> +        .quad 0xc0862594d92dd600, 0xbe1cf0f43ca40766
> +        .quad 0xc0862597ce461558, 0xbe1cefb2cfc6766b
> +        .quad 0xc086259ac246daf0, 0xbe1ceea49e64ffa2
> +        .quad 0xc086259db530f4c8, 0xbe1cf250fa457dec
> +        .quad 0xc08625a0a7053018, 0xbe1cf17d8bb2a44e
> +        .quad 0xc08625a397c45918, 0xbe1cf1d5906d54b7
> +        .quad 0xc08625a6876f3b30, 0xbe1cf08fe7b31780
> +        .quad 0xc08625a97606a0e0, 0xbe1cef13edfc9d11
> +        .quad 0xc08625ac638b53c8, 0xbe1cef9d2b107219
> +        .quad 0xc08625af4ffe1cb0, 0xbe1cf1ddd4ff6160
> +        .quad 0xc08625b23b5fc390, 0xbe1cefa02a996495
> +        .quad 0xc08625b525b10f68, 0xbe1cf166a7e37ee5
> +        .quad 0xc08625b80ef2c680, 0xbe1cef0b171068a5
> +        .quad 0xc08625baf725ae28, 0xbe1cf05c80779283
> +        .quad 0xc08625bdde4a8af0, 0xbe1cf1bbfbffb889
> +        .quad 0xc08625c0c4622090, 0xbe1cf0b8666c0124
> +        .quad 0xc08625c3a96d31e0, 0xbe1cf0a8fcf47a86
> +        .quad 0xc08625c68d6c80f0, 0xbe1cef46e18cb092
> +        .quad 0xc08625c97060cef0, 0xbe1cf1458a350efb
> +        .quad 0xc08625cc524adc58, 0xbe1ceeea1dadce12
> +        .quad 0xc08625cf332b68b0, 0xbe1cf0a1bfdc44c7
> +        .quad 0xc08625d2130332d0, 0xbe1cef96d02da73e
> +        .quad 0xc08625d4f1d2f8a8, 0xbe1cf2451c3c7701
> +        .quad 0xc08625d7cf9b7778, 0xbe1cf10d08f83812
> +        .quad 0xc08625daac5d6ba0, 0xbe1ceec5b4895c5e
> +        .quad 0xc08625dd881990b0, 0xbe1cf14e1325c5e4
> +        .quad 0xc08625e062d0a188, 0xbe1cf21d0904be12
> +        .quad 0xc08625e33c835838, 0xbe1ceed0839bcf21
> +        .quad 0xc08625e615326df0, 0xbe1cf1bb944889d2
> +        .quad 0xc08625e8ecde9b48, 0xbe1cee738e85eece
> +        .quad 0xc08625ebc38897e0, 0xbe1cf25c2bc6ef12
> +        .quad 0xc08625ee99311ac8, 0xbe1cf132b70a41ad
> +        .quad 0xc08625f16dd8da28, 0xbe1cf1984236a6e3
> +        .quad 0xc08625f441808b78, 0xbe1cf19ae74998f9
> +        .quad 0xc08625f71428e370, 0xbe1cef3e175d61a1
> +        .quad 0xc08625f9e5d295f8, 0xbe1cf101f9868fd9
> +        .quad 0xc08625fcb67e5658, 0xbe1cee69db83dcd2
> +        .quad 0xc08625ff862cd6f8, 0xbe1cf081b636af51
> +        .quad 0xc086260254dec9a8, 0xbe1cee62c7d59b3e
> +        .quad 0xc08626052294df58, 0xbe1cf1b745c57716
> +        .quad 0xc0862607ef4fc868, 0xbe1cef3d2800ea23
> +        .quad 0xc086260abb103458, 0xbe1cef480ff1acd2
> +        .quad 0xc086260d85d6d200, 0xbe1cf2424c9a17ef
> +        .quad 0xc08626104fa44f90, 0xbe1cf12cfde90fd5
> +        .quad 0xc086261318795a68, 0xbe1cf21f590dd5b6
> +        .quad 0xc0862615e0569f48, 0xbe1cf0c50f9cd28a
> +        .quad 0xc0862618a73cca30, 0xbe1ceedbdb520545
> +        .quad 0xc086261b6d2c8668, 0xbe1cf0b030396011
> +        .quad 0xc086261e32267e98, 0xbe1cf19917010e96
> +        .quad 0xc0862620f62b5cb0, 0xbe1cf07331355985
> +        .quad 0xc0862623b93bc9e8, 0xbe1cf01ae921a1c3
> +        .quad 0xc08626267b586ed0, 0xbe1cefe5cf0dbf0c
> +        .quad 0xc08626293c81f348, 0xbe1cf01b258aeb50
> +        .quad 0xc086262bfcb8fe88, 0xbe1cee6b9e7f4c68
> +        .quad 0xc086262ebbfe3710, 0xbe1cee684a9b21c9
> +        .quad 0xc08626317a5242b8, 0xbe1cf1f8bcde9a8b
> +        .quad 0xc086263437b5c6c0, 0xbe1cf1d063d36238
> +        .quad 0xc0862636f42967a8, 0xbe1cf1e31a19075e
> +        .quad 0xc0862639afadc950, 0xbe1cf1d8efdf7e7d
> +        .quad 0xc086263c6a438ef0, 0xbe1cf1812ee72dba
> +        .quad 0xc086263f23eb5b18, 0xbe1cf1449a9a2279
> +        .quad 0xc0862641dca5cfb8, 0xbe1cee96edce5085
> +        .quad 0xc086264494738e08, 0xbe1cf06797bd03b2
> +        .quad 0xc08626474b5536b8, 0xbe1cef91b9b7ffc1
> +        .quad 0xc086264a014b69c0, 0xbe1cef4b6721278f
> +        .quad 0xc086264cb656c678, 0xbe1cf1942925eb4a
> +        .quad 0xc086264f6a77eba8, 0xbe1cefa2c7bc2e39
> +        .quad 0xc08626521daf7758, 0xbe1cf252595aceb3
> +        .quad 0xc0862654cffe0718, 0xbe1cee8e9ae47ec2
> +        .quad 0xc0862657816437a8, 0xbe1cf1bf913828fa
> +        .quad 0xc086265a31e2a558, 0xbe1cf23475d6b366
> +        .quad 0xc086265ce179ebc8, 0xbe1cef8df00a922b
> +        .quad 0xc086265f902aa5f0, 0xbe1cef279bfa43e0
> +        .quad 0xc08626623df56e38, 0xbe1cf080e10b8365
> +        .quad 0xc0862664eadade70, 0xbe1cf1a518f9b544
> +        .quad 0xc086266796db8fd0, 0xbe1cef9308fed9e9
> +        .quad 0xc086266a41f81ae8, 0xbe1ceea3ae6b19c9
> +        .quad 0xc086266cec3117b8, 0xbe1ceef06003d4c2
> +        .quad 0xc086266f95871da8, 0xbe1cf0b8457ffb0c
> +        .quad 0xc08626723dfac390, 0xbe1cf0c526745ad6
> +        .quad 0xc0862674e58c9fa8, 0xbe1cf0cf91ff7b5d
> +        .quad 0xc08626778c3d4798, 0xbe1cefe260819380
> +        .quad 0xc086267a320d5070, 0xbe1ceebd90aa27a3
> +        .quad 0xc086267cd6fd4ea8, 0xbe1cf0388121dffa
> +        .quad 0xc086267f7b0dd630, 0xbe1cf1a3881435f1
> +        .quad 0xc08626821e3f7a68, 0xbe1cef28e9d9ac52
> +        .quad 0xc0862684c092ce08, 0xbe1cf02d300062dd
> +        .quad 0xc086268762086350, 0xbe1cefaee1edfa35
> +        .quad 0xc086268a02a0cbe0, 0xbe1cf0a5a052e936
> +        .quad 0xc086268ca25c98d8, 0xbe1cee60a4a497ed
> +        .quad 0xc086268f413c5ab0, 0xbe1cf0e4a5d0cf49
> +        .quad 0xc0862691df40a170, 0xbe1cf149235a4e6e
> +        .quad 0xc08626947c69fc80, 0xbe1cf215180b9fcc
> +        .quad 0xc086269718b8fac8, 0xbe1cef9b156a9840
> +        .quad 0xc0862699b42e2a90, 0xbe1cf054c91441be
> +        .quad 0xc086269c4eca19a8, 0xbe1cf13ded26512c
> +        .quad 0xc086269ee88d5550, 0xbe1cf22ea4d8ac06
> +        .quad 0xc08626a181786a40, 0xbe1cf2354666ee2e
> +        .quad 0xc08626a4198be4a8, 0xbe1cefef936752b3
> +        .quad 0xc08626a6b0c85020, 0xbe1cf1e360a9db68
> +        .quad 0xc08626a9472e37d8, 0xbe1ceed6aeb812c5
> +        .quad 0xc08626abdcbe2650, 0xbe1cf227340b4986
> +        .quad 0xc08626ae7178a5b0, 0xbe1cf0215a0cbe0d
> +        .quad 0xc08626b1055e3f70, 0xbe1cf256adf0ae26
> +        .quad 0xc08626b3986f7ca8, 0xbe1ceff3c67aed06
> +        .quad 0xc08626b62aace5c8, 0xbe1cf2159fb93652
> +        .quad 0xc08626b8bc1702e0, 0xbe1cf01e6dbd1c7f
> +        .quad 0xc08626bb4cae5b60, 0xbe1cf009e75d1c0c
> +        .quad 0xc08626bddc737648, 0xbe1ceec10a020e73
> +        .quad 0xc08626c06b66da08, 0xbe1cf06d5783eee7
> +        .quad 0xc08626c2f9890ca0, 0xbe1cf0cb8f169ffe
> +        .quad 0xc08626c586da9388, 0xbe1cef7de2452430
> +        .quad 0xc08626c8135bf3b0, 0xbe1cf05da6f783ae
> +        .quad 0xc08626ca9f0db198, 0xbe1cefcc877d681d
> +        .quad 0xc08626cd29f05138, 0xbe1cef0531954ab3
> +        .quad 0xc08626cfb4045608, 0xbe1cf06b8565ea3d
> +        .quad 0xc08626d23d4a4310, 0xbe1cefdc455d9d7e
> +        .quad 0xc08626d4c5c29ad0, 0xbe1ceefc47e8fa64
> +        .quad 0xc08626d74d6ddf48, 0xbe1cf1872bf033f2
> +        .quad 0xc08626d9d44c9210, 0xbe1cf19d91087f9d
> +        .quad 0xc08626dc5a5f3438, 0xbe1cf012d444c6ab
> +        .quad 0xc08626dedfa64650, 0xbe1cf0ba528ee153
> +        .quad 0xc08626e164224880, 0xbe1ceeb431709788
> +        .quad 0xc08626e3e7d3ba60, 0xbe1cf0b9af31a6a5
> +        .quad 0xc08626e66abb1b28, 0xbe1cf168fb2e135b
> +        .quad 0xc08626e8ecd8e990, 0xbe1cef9097461c93
> +        .quad 0xc08626eb6e2da3d0, 0xbe1cee7a434735d8
> +        .quad 0xc08626edeeb9c7a8, 0xbe1cf235732b86f2
> +        .quad 0xc08626f06e7dd280, 0xbe1cefe1510b89e6
> +        .quad 0xc08626f2ed7a4120, 0xbe1cf1f64b9b80ef
> +        .quad 0xc08626f56baf9000, 0xbe1cf08f320ca339
> +        .quad 0xc08626f7e91e3b08, 0xbe1cf1b1de2808a1
> +        .quad 0xc08626fa65c6bdc0, 0xbe1cf1976d778b28
> +        .quad 0xc08626fce1a99338, 0xbe1ceef40a4f076f
> +        .quad 0xc08626ff5cc73600, 0xbe1cef3e45869ce3
> +        .quad 0xc0862701d7202048, 0xbe1ceef601b4c9d6
> +        .quad 0xc086270450b4cbc0, 0xbe1cf1eaf0b57fd6
> +        .quad 0xc0862706c985b1c0, 0xbe1cef82a44990f3
> +        .quad 0xc086270941934b10, 0xbe1ceefe32981f2c
> +        .quad 0xc086270bb8de1018, 0xbe1cefbf6f5a0445
> +        .quad 0xc086270e2f6678d0, 0xbe1cf18dba75792c
> +        .quad 0xc0862710a52cfcc8, 0xbe1cf0da64ce995f
> +        .quad 0xc08627131a321318, 0xbe1cef04ac0fb802
> +        .quad 0xc08627158e763268, 0xbe1cee9d4e2ad9bd
> +        .quad 0xc086271801f9d0f8, 0xbe1cefa9b55407b5
> +        .quad 0xc086271a74bd64a0, 0xbe1cefe6bd329570
> +        .quad 0xc086271ce6c162c8, 0xbe1cef0b1205dc85
> +        .quad 0xc086271f58064068, 0xbe1cef092a785e3f
> +        .quad 0xc0862721c88c7210, 0xbe1cf050dcdaac30
> +        .quad 0xc086272438546be8, 0xbe1cf210907ded8b
> +        .quad 0xc0862726a75ea1b8, 0xbe1cee760be44f99
> +        .quad 0xc086272915ab86c0, 0xbe1ceeeee07c2bcc
> +        .quad 0xc086272b833b8df0, 0xbe1cf06874992df5
> +        .quad 0xc086272df00f29d0, 0xbe1cef8fac5d4899
> +        .quad 0xc08627305c26cc70, 0xbe1cf1103241cc99
> +        .quad 0xc0862732c782e788, 0xbe1cf1d35fef83fe
> +        .quad 0xc08627353223ec68, 0xbe1cef3ec8133e1d
> +        .quad 0xc08627379c0a4be8, 0xbe1cef7261daccd8
> +        .quad 0xc086273a05367688, 0xbe1cf18656c50806
> +        .quad 0xc086273c6da8dc68, 0xbe1cf1c8736e049a
> +        .quad 0xc086273ed561ed38, 0xbe1cf1f93bff4911
> +        .quad 0xc08627413c621848, 0xbe1cf188a4ea680c
> +        .quad 0xc0862743a2a9cc80, 0xbe1cf1d270930c80
> +        .quad 0xc086274608397868, 0xbe1cf25a328c28e2
> +        .quad 0xc08627486d118a28, 0xbe1cf106f90aa3b8
> +        .quad 0xc086274ad1326f80, 0xbe1cee5e9d2e885a
> +        .quad 0xc086274d349c95c0, 0xbe1cf1c0bac27228
> +        .quad 0xc086274f975069f8, 0xbe1cf1a1500f9b1c
> +        .quad 0xc0862751f94e58c0, 0xbe1cefc30663ac44
> +        .quad 0xc08627545a96ce48, 0xbe1cf17123e427a2
> +        .quad 0xc0862756bb2a3678, 0xbe1cefb92749fea4
> +        .quad 0xc08627591b08fcc0, 0xbe1cefa40e1ea74a
> +        .quad 0xc086275b7a338c40, 0xbe1cee6f4612c3e9
> +        .quad 0xc086275dd8aa4fa8, 0xbe1cf1c54a053627
> +        .quad 0xc0862760366db168, 0xbe1ceff5eb503d9e
> +        .quad 0xc0862762937e1b70, 0xbe1cf02e47f10cee
> +        .quad 0xc0862764efdbf768, 0xbe1ceeb06e1d0dad
> +        .quad 0xc08627674b87ae88, 0xbe1cf10aadd6dba5
> +        .quad 0xc0862769a681a9c0, 0xbe1cf24e9913d30f
> +        .quad 0xc086276c00ca51a0, 0xbe1cef47b301e312
> +        .quad 0xc086276e5a620e48, 0xbe1ceeb1cefc2e85
> +        .quad 0xc0862770b3494788, 0xbe1cf16f1fbbe011
> +        .quad 0xc08627730b8064e8, 0xbe1ceebdf75174c7
> +        .quad 0xc08627756307cd70, 0xbe1cf06e3871a0da
> +        .quad 0xc0862777b9dfe7f0, 0xbe1cef16799fd554
> +        .quad 0xc086277a10091ac0, 0xbe1cf248dabf5377
> +        .quad 0xc086277c6583cc00, 0xbe1cf0c78d92a2cd
> +        .quad 0xc086277eba506158, 0xbe1cf0b911b029f0
> +        .quad 0xc08627810e6f4028, 0xbe1cefdc24719766
> +        .quad 0xc086278361e0cd70, 0xbe1cefbb6562b7e7
> +        .quad 0xc0862785b4a56dd8, 0xbe1cf1e0afb349ec
> +        .quad 0xc086278806bd85c0, 0xbe1cf008292e52fc
> +        .quad 0xc086278a58297918, 0xbe1cf053073872bf
> +        .quad 0xc086278ca8e9ab88, 0xbe1cf17a0a55a947
> +        .quad 0xc086278ef8fe8068, 0xbe1ceeffb0b60234
> +        .quad 0xc086279148685aa0, 0xbe1cf162204794a8
> +        .quad 0xc086279397279ce0, 0xbe1cf24cc8cb48ac
> +        .quad 0xc0862795e53ca978, 0xbe1cf0c9be68d5c3
> +        .quad 0xc086279832a7e258, 0xbe1cf172cd3d7388
> +        .quad 0xc086279a7f69a930, 0xbe1ceea2465fbce5
> +        .quad 0xc086279ccb825f40, 0xbe1cf0a386d2500f
> +        .quad 0xc086279f16f26590, 0xbe1cf1e338ddc18a
> +        .quad 0xc08627a161ba1cd0, 0xbe1cef1f5049867f
> +        .quad 0xc08627a3abd9e548, 0xbe1cef96c1ea8b1f
> +        .quad 0xc08627a5f5521f00, 0xbe1cf138f6fd3c26
> +        .quad 0xc08627a83e2329b0, 0xbe1cf0d4fcbfdf3a
> +        .quad 0xc08627aa864d64b0, 0xbe1cf24870c12c81
> +        .quad 0xc08627accdd12f18, 0xbe1cf0ae2a56348d
> +        .quad 0xc08627af14aee7a0, 0xbe1cee8ca1a9b893
> +        .quad 0xc08627b15ae6eca8, 0xbe1cf20414d637b0
> +        .quad 0xc08627b3a0799c60, 0xbe1cf0fc6b7b12d8
> +        .quad 0xc08627b5e5675488, 0xbe1cf152d93c4a00
> +        .quad 0xc08627b829b072a0, 0xbe1cf1073f9b77c2
> +        .quad 0xc08627ba6d5553d8, 0xbe1cee694f97d5a4
> +        .quad 0xc08627bcb0565500, 0xbe1cf0456b8239d7
> +        .quad 0xc08627bef2b3d2b0, 0xbe1cf211497127e3
> +        .quad 0xc08627c1346e2930, 0xbe1cf01856c0384d
> +        .quad 0xc08627c37585b468, 0xbe1cefa7dd05479e
> +        .quad 0xc08627c5b5fad000, 0xbe1cef3ae8e50b93
> +        .quad 0xc08627c7f5cdd750, 0xbe1ceea5f32fdd3a
> +        .quad 0xc08627ca34ff2560, 0xbe1cef424caeb8d9
> +        .quad 0xc08627cc738f14f0, 0xbe1cf0194d07a81f
> +        .quad 0xc08627ceb17e0070, 0xbe1cf20f452000c1
> +        .quad 0xc08627d0eecc4210, 0xbe1cf00e356218e4
> +        .quad 0xc08627d32b7a33a0, 0xbe1cef30484b4bcb
> +        .quad 0xc08627d567882eb0, 0xbe1ceeea11a6641b
> +        .quad 0xc08627d7a2f68c80, 0xbe1cf13492d5bd7b
> +        .quad 0xc08627d9ddc5a618, 0xbe1ceeb7048fad96
> +        .quad 0xc08627dc17f5d418, 0xbe1ceef0666f0477
> +        .quad 0xc08627de51876ee8, 0xbe1cf060d4b8b5c2
> +        .quad 0xc08627e08a7acea8, 0xbe1cf0b2a4b6ff8c
> +        .quad 0xc08627e2c2d04b28, 0xbe1cf0e34809a875
> +        .quad 0xc08627e4fa883bf0, 0xbe1cf16bf74a3522
> +        .quad 0xc08627e731a2f848, 0xbe1cee6a24623d57
> +        .quad 0xc08627e96820d718, 0xbe1cefc7b4f1528e
> +        .quad 0xc08627eb9e022f18, 0xbe1cf163051f3548
> +        .quad 0xc08627edd34756b8, 0xbe1cef36b3366305
> +        .quad 0xc08627f007f0a408, 0xbe1cf18134625550
> +        .quad 0xc08627f23bfe6cf0, 0xbe1cf0ec32ec1a11
> +        .quad 0xc08627f46f710700, 0xbe1ceeb3b64f3edc
> +        .quad 0xc08627f6a248c778, 0xbe1cf0cd15805bc8
> +        .quad 0xc08627f8d4860368, 0xbe1cf20db3bddebe
> +        .quad 0xc08627fb06290f90, 0xbe1cf25188430e25
> +        .quad 0xc08627fd37324070, 0xbe1ceea1713490f9
> +        .quad 0xc08627ff67a1ea28, 0xbe1cf159521d234c
> +        .quad 0xc0862801977860b8, 0xbe1cf24dfe50783b
> +        .quad 0xc0862803c6b5f7d0, 0xbe1ceef2ef89a60b
> +        .quad 0xc0862805f55b02c8, 0xbe1cee7fc919d62c
> +        .quad 0xc08628082367d4c0, 0xbe1cf215a7fb513a
> +        .quad 0xc086280a50dcc0a8, 0xbe1cf0e4401c5ed4
> +        .quad 0xc086280c7dba1910, 0xbe1cf04ec734d256
> +        .quad 0xc086280eaa003050, 0xbe1cf010ad787fea
> +        .quad 0xc0862810d5af5880, 0xbe1cee622478393d
> +        .quad 0xc086281300c7e368, 0xbe1cf01c7482564f
> +        .quad 0xc08628152b4a22a0, 0xbe1cf0de20d33536
> +        .quad 0xc086281755366778, 0xbe1cef2edae5837d
> +        .quad 0xc08628197e8d02f0, 0xbe1cf0a345318cc9
> +        .quad 0xc086281ba74e45d8, 0xbe1cf20085aa34b8
> +        .quad 0xc086281dcf7a80c0, 0xbe1cef5fa845ad83
> +        .quad 0xc086281ff71203e0, 0xbe1cf050d1df69c4
> +        .quad 0xc08628221e151f48, 0xbe1ceffe43c035b9
> +        .quad 0xc0862824448422b8, 0xbe1cf14f3018d3c2
> +        .quad 0xc08628266a5f5dc0, 0xbe1cef0a5fbae83d
> +        .quad 0xc08628288fa71f98, 0xbe1ceff8a95b72a1
> +        .quad 0xc086282ab45bb750, 0xbe1cef073aa9849b
> +        .quad 0xc086282cd87d73a8, 0xbe1cef69b3835c02
> +        .quad 0xc086282efc0ca328, 0xbe1cf0bc139379a9
> +        .quad 0xc08628311f099420, 0xbe1cef247a9ec596
> +        .quad 0xc086283341749490, 0xbe1cef74bbcc488a
> +        .quad 0xc0862835634df248, 0xbe1cef4bc42e7b8e
> +        .quad 0xc08628378495fad0, 0xbe1cf136d4d5a810
> +        .quad 0xc0862839a54cfb80, 0xbe1cf0d290b24dd8
> +        .quad 0xc086283bc5734168, 0xbe1ceeebde8e0065
> +        .quad 0xc086283de5091950, 0xbe1cf1a09f60aa1e
> +        .quad 0xc0862840040ecfe0, 0xbe1cf0803947a234
> +        .quad 0xc08628422284b168, 0xbe1cf0abf7638127
> +        .quad 0xc0862844406b0a08, 0xbe1cf0f73ee12058
> +        .quad 0xc08628465dc225a0, 0xbe1cf2079971b26c
> +        .quad 0xc08628487a8a4fe0, 0xbe1cee74957564b1
> +        .quad 0xc086284a96c3d420, 0xbe1ceee77c1b7d43
> +        .quad 0xc086284cb26efd90, 0xbe1cf23addba6e09
> +        .quad 0xc086284ecd8c1730, 0xbe1cf199f4a1da60
> +        .quad 0xc0862850e81b6bb0, 0xbe1cf09fdea81393
> +        .quad 0xc0862853021d4588, 0xbe1cf176adb417f7
> +        .quad 0xc08628551b91ef00, 0xbe1cf0f64f84a8da
> +        .quad 0xc08628573479b220, 0xbe1ceec34cf49523
> +        .quad 0xc08628594cd4d8a8, 0xbe1cf16d60fbe0bb
> +        .quad 0xc086285b64a3ac40, 0xbe1cee8de7acfc7b
> +        .quad 0xc086285d7be67630, 0xbe1ceee6256cce8d
> +        .quad 0xc086285f929d7fa0, 0xbe1cee7d66a3d8a5
> +        .quad 0xc0862861a8c91170, 0xbe1cf0bef8265792
> +        .quad 0xc0862863be697458, 0xbe1cf097f890c6f8
> +        .quad 0xc0862865d37ef0c8, 0xbe1cf09502d5c3fc
> +        .quad 0xc0862867e809cf00, 0xbe1ceeffb239dac7
> +        .quad 0xc0862869fc0a56f8, 0xbe1cf1fbfff95c98
> +        .quad 0xc086286c0f80d090, 0xbe1cefa57ad3eef7
> +        .quad 0xc086286e226d8348, 0xbe1cf22c58b9183d
> +        .quad 0xc086287034d0b690, 0xbe1ceff262d0a248
> +        .quad 0xc086287246aab180, 0xbe1cefa7bc194186
> +        .quad 0xc086287457fbbb08, 0xbe1cf06782d784d9
> +        .quad 0xc086287668c419e0, 0xbe1cf1d44d0eaa07
> +        .quad 0xc086287879041490, 0xbe1cf034803c8a48
> +        .quad 0xc086287a88bbf158, 0xbe1cf08e84916b6f
> +        .quad 0xc086287c97ebf650, 0xbe1cf0c4d3dc1bc7
> +        .quad 0xc086287ea6946958, 0xbe1cefb1e4625943
> +        .quad 0xc0862880b4b59010, 0xbe1cf143efdd1fd0
> +        .quad 0xc0862882c24faff8, 0xbe1cee9896d016da
> +        .quad 0xc0862884cf630e38, 0xbe1cf2186072f2cc
> +        .quad 0xc0862886dbefeff0, 0xbe1cef9217633d34
> +        .quad 0xc0862888e7f699e0, 0xbe1cf05603549486
> +        .quad 0xc086288af37750b0, 0xbe1cef50fff513d3
> +        .quad 0xc086288cfe7258c0, 0xbe1cf127713b32d0
> +        .quad 0xc086288f08e7f650, 0xbe1cf05015520f3d
> +        .quad 0xc086289112d86d58, 0xbe1cf12eb458b26f
> +        .quad 0xc08628931c4401a8, 0xbe1cf22eae2887ed
> +        .quad 0xc0862895252af6e0, 0xbe1cefdd6656dd2d
> +        .quad 0xc08628972d8d9058, 0xbe1cf1048ea4e646
> +        .quad 0xc0862899356c1150, 0xbe1ceec4501167e9
> +        .quad 0xc086289b3cc6bcb8, 0xbe1cf0ad52becc3f
> +        .quad 0xc086289d439dd568, 0xbe1cf0daa4e00e35
> +        .quad 0xc086289f49f19df8, 0xbe1cf00b80de8d6a
> +        .quad 0xc08628a14fc258c8, 0xbe1cf1bcf2ea8464
> +        .quad 0xc08628a355104818, 0xbe1cf0435e2782b0
> +        .quad 0xc08628a559dbade0, 0xbe1cf0e3e1a5f56c
> +        .quad 0xc08628a75e24cbf8, 0xbe1cefed9d5a721d
> +        .quad 0xc08628a961ebe3f8, 0xbe1cf0d2d74321e2
> +        .quad 0xc08628ab65313750, 0xbe1cf24200eb55e9
> +        .quad 0xc08628ad67f50740, 0xbe1cf23e9d7cf979
> +        .quad 0xc08628af6a3794d0, 0xbe1cf23a088f421c
> +        .quad 0xc08628b16bf920e0, 0xbe1cef2c1de1ab32
> +        .quad 0xc08628b36d39ec08, 0xbe1cf1abc231f7b2
> +        .quad 0xc08628b56dfa36d0, 0xbe1cf2074d5ba303
> +        .quad 0xc08628b76e3a4180, 0xbe1cf05cd5eed880
> +        /*== Log_LA_table ==*/
> +        .align 32
> +        .quad 0x8000000000000000
> +        .quad 0xbf5ff802a9ab10e6
> +        .quad 0xbf6ff00aa2b10bc0
> +        .quad 0xbf77ee11ebd82e94
> +        .quad 0xbf7fe02a6b106789
> +        .quad 0xbf83e7295d25a7d9
> +        .quad 0xbf87dc475f810a77
> +        .quad 0xbf8bcf712c74384c
> +        .quad 0xbf8fc0a8b0fc03e4
> +        .quad 0xbf91d7f7eb9eebe7
> +        .quad 0xbf93cea44346a575
> +        .quad 0xbf95c45a51b8d389
> +        .quad 0xbf97b91b07d5b11b
> +        .quad 0xbf99ace7551cc514
> +        .quad 0xbf9b9fc027af9198
> +        .quad 0xbf9d91a66c543cc4
> +        .quad 0xbf9f829b0e783300
> +        .quad 0xbfa0b94f7c196176
> +        .quad 0xbfa1b0d98923d980
> +        .quad 0xbfa2a7ec2214e873
> +        .quad 0xbfa39e87b9febd60
> +        .quad 0xbfa494acc34d911c
> +        .quad 0xbfa58a5bafc8e4d5
> +        .quad 0xbfa67f94f094bd98
> +        .quad 0xbfa77458f632dcfc
> +        .quad 0xbfa868a83083f6cf
> +        .quad 0xbfa95c830ec8e3eb
> +        .quad 0xbfaa4fe9ffa3d235
> +        .quad 0xbfab42dd711971bf
> +        .quad 0xbfac355dd0921f2d
> +        .quad 0xbfad276b8adb0b52
> +        .quad 0xbfae19070c276016
> +        .quad 0xbfaf0a30c01162a6
> +        .quad 0xbfaffae9119b9303
> +        .quad 0xbfb075983598e471
> +        .quad 0xbfb0ed839b5526fe
> +        .quad 0xbfb16536eea37ae1
> +        .quad 0xbfb1dcb263db1944
> +        .quad 0xbfb253f62f0a1417
> +        .quad 0xbfb2cb0283f5de1f
> +        .quad 0xbfb341d7961bd1d1
> +        .quad 0xbfb3b87598b1b6ee
> +        .quad 0xbfb42edcbea646f0
> +        .quad 0xbfb4a50d3aa1b040
> +        .quad 0xbfb51b073f06183f
> +        .quad 0xbfb590cafdf01c28
> +        .quad 0xbfb60658a93750c4
> +        .quad 0xbfb67bb0726ec0fc
> +        .quad 0xbfb6f0d28ae56b4c
> +        .quad 0xbfb765bf23a6be13
> +        .quad 0xbfb7da766d7b12cd
> +        .quad 0xbfb84ef898e8282a
> +        .quad 0xbfb8c345d6319b21
> +        .quad 0xbfb9375e55595ede
> +        .quad 0xbfb9ab42462033ad
> +        .quad 0xbfba1ef1d8061cd4
> +        .quad 0xbfba926d3a4ad563
> +        .quad 0xbfbb05b49bee43fe
> +        .quad 0xbfbb78c82bb0eda1
> +        .quad 0xbfbbeba818146765
> +        .quad 0xbfbc5e548f5bc743
> +        .quad 0xbfbcd0cdbf8c13e1
> +        .quad 0xbfbd4313d66cb35d
> +        .quad 0xbfbdb5270187d927
> +        .quad 0xbfbe27076e2af2e6
> +        .quad 0xbfbe98b549671467
> +        .quad 0xbfbf0a30c01162a6
> +        .quad 0xbfbf7b79fec37ddf
> +        .quad 0xbfbfec9131dbeabb
> +        .quad 0xbfc02ebb42bf3d4b
> +        .quad 0xbfc0671512ca596e
> +        .quad 0xbfc09f561ee719c3
> +        .quad 0xbfc0d77e7cd08e59
> +        .quad 0xbfc10f8e422539b1
> +        .quad 0xbfc14785846742ac
> +        .quad 0xbfc17f6458fca611
> +        .quad 0xbfc1b72ad52f67a0
> +        .quad 0xbfc1eed90e2dc2c3
> +        .quad 0xbfc2266f190a5acb
> +        .quad 0xbfc25ded0abc6ad2
> +        .quad 0xbfc29552f81ff523
> +        .quad 0xbfc2cca0f5f5f251
> +        .quad 0xbfc303d718e47fd3
> +        .quad 0xbfc33af575770e4f
> +        .quad 0xbfc371fc201e8f74
> +        .quad 0xbfc3a8eb2d31a376
> +        .quad 0xbfc3dfc2b0ecc62a
> +        .quad 0xbfc41682bf727bc0
> +        .quad 0xbfc44d2b6ccb7d1e
> +        .quad 0xbfc483bccce6e3dd
> +        .quad 0xbfc4ba36f39a55e5
> +        .quad 0xbfc4f099f4a230b2
> +        .quad 0xbfc526e5e3a1b438
> +        .quad 0xbfc55d1ad4232d6f
> +        .quad 0xbfc59338d9982086
> +        .quad 0xbfc5c940075972b9
> +        .quad 0xbfc5ff3070a793d4
> +        .quad 0xbfc6350a28aaa758
> +        .quad 0xbfc66acd4272ad51
> +        .quad 0xbfc6a079d0f7aad2
> +        .quad 0xbfc6d60fe719d21d
> +        .quad 0xbfc70b8f97a1aa75
> +        .quad 0xbfc740f8f54037a5
> +        .quad 0xbfc7764c128f2127
> +        .quad 0xbfc7ab890210d909
> +        .quad 0xbfc7e0afd630c274
> +        .quad 0xbfc815c0a14357eb
> +        .quad 0xbfc84abb75865139
> +        .quad 0xbfc87fa06520c911
> +        .quad 0xbfc8b46f8223625b
> +        .quad 0xbfc8e928de886d41
> +        .quad 0xbfc91dcc8c340bde
> +        .quad 0xbfc9525a9cf456b4
> +        .quad 0xbfc986d3228180ca
> +        .quad 0xbfc9bb362e7dfb83
> +        .quad 0xbfc9ef83d2769a34
> +        .quad 0xbfca23bc1fe2b563
> +        .quad 0xbfca57df28244dcd
> +        .quad 0xbfca8becfc882f19
> +        .quad 0xbfcabfe5ae46124c
> +        .quad 0xbfcaf3c94e80bff3
> +        .quad 0xbfcb2797ee46320c
> +        .quad 0xbfcb5b519e8fb5a4
> +        .quad 0xbfcb8ef670420c3b
> +        .quad 0xbfcbc286742d8cd6
> +        .quad 0xbfcbf601bb0e44e2
> +        .quad 0xbfcc2968558c18c1
> +        .quad 0xbfcc5cba543ae425
> +        .quad 0xbfcc8ff7c79a9a22
> +        .quad 0xbfccc320c0176502
> +        .quad 0xbfccf6354e09c5dc
> +        .quad 0xbfcd293581b6b3e7
> +        .quad 0xbfcd5c216b4fbb91
> +        .quad 0xbfcd8ef91af31d5e
> +        .quad 0xbfcdc1bca0abec7d
> +        .quad 0xbfcdf46c0c722d2f
> +        .quad 0xbfce27076e2af2e6
> +        .quad 0xbfce598ed5a87e2f
> +        .quad 0xbfce8c0252aa5a60
> +        .quad 0xbfcebe61f4dd7b0b
> +        .quad 0xbfcef0adcbdc5936
> +        .quad 0xbfcf22e5e72f105d
> +        .quad 0xbfcf550a564b7b37
> +        .quad 0xbfcf871b28955045
> +        .quad 0xbfcfb9186d5e3e2b
> +        .quad 0xbfcfeb0233e607cc
> +        .quad 0xbfd00e6c45ad501d
> +        .quad 0xbfd0274dc16c232f
> +        .quad 0xbfd0402594b4d041
> +        .quad 0xbfd058f3c703ebc6
> +        .quad 0xbfd071b85fcd590d
> +        .quad 0xbfd08a73667c57af
> +        .quad 0xbfd0a324e27390e3
> +        .quad 0xbfd0bbccdb0d24bd
> +        .quad 0xbfd0d46b579ab74b
> +        .quad 0xbfd0ed005f657da4
> +        .quad 0xbfd1058bf9ae4ad5
> +        .quad 0xbfd11e0e2dad9cb7
> +        .quad 0xbfd136870293a8b0
> +        .quad 0xbfd14ef67f88685a
> +        .quad 0xbfd1675cababa60e
> +        .quad 0xbfd17fb98e15095d
> +        .quad 0xbfd1980d2dd4236f
> +        .quad 0xbfd1b05791f07b49
> +        .quad 0xbfd1c898c16999fb
> +        .quad 0xbfd1e0d0c33716be
> +        .quad 0xbfd1f8ff9e48a2f3
> +        .quad 0xbfd211255986160c
> +        .quad 0xbfd22941fbcf7966
> +        .quad 0xbfd241558bfd1404
> +        .quad 0xbfd2596010df763a
> +        .quad 0xbfd27161913f853d
> +        .quad 0xbfd2895a13de86a3
> +        .quad 0xbfd2a1499f762bc9
> +        .quad 0xbfd2b9303ab89d25
> +        .quad 0xbfd2d10dec508583
> +        .quad 0xbfd2e8e2bae11d31
> +        .quad 0xbfd300aead06350c
> +        .quad 0xbfd31871c9544185
> +        .quad 0xbfd3302c16586588
> +        .quad 0xbfd347dd9a987d55
> +        .quad 0xbfd35f865c93293e
> +        .quad 0xbfd3772662bfd85b
> +        .quad 0xbfd38ebdb38ed321
> +        .quad 0xbfd3a64c556945ea
> +        .quad 0xbfd3bdd24eb14b6a
> +        .quad 0xbfd3d54fa5c1f710
> +        .quad 0xbfd3ecc460ef5f50
> +        .quad 0xbfd404308686a7e4
> +        .quad 0xbfd41b941cce0bee
> +        .quad 0xbfd432ef2a04e814
> +        .quad 0xbfd44a41b463c47c
> +        .quad 0xbfd4618bc21c5ec2
> +        .quad 0xbfd478cd5959b3d9
> +        .quad 0xbfd49006804009d1
> +        .quad 0xbfd4a7373cecf997
> +        .quad 0xbfd4be5f957778a1
> +        .quad 0xbfd4d57f8fefe27f
> +        .quad 0xbfd4ec973260026a
> +        .quad 0xbfd503a682cb1cb3
> +        .quad 0xbfd51aad872df82d
> +        .quad 0xbfd531ac457ee77e
> +        .quad 0xbfd548a2c3add263
> +        .quad 0xbfd55f9107a43ee2
> +        .quad 0xbfd5767717455a6c
> +        .quad 0xbfd58d54f86e02f2
> +        .quad 0xbfd5a42ab0f4cfe2
> +        .quad 0xbfd5baf846aa1b19
> +        .quad 0xbfd5d1bdbf5809ca
> +        .quad 0xbfd5e87b20c2954a
> +        .quad 0xbfd5ff3070a793d4
> +        .quad 0xbfd615ddb4bec13c
> +        .quad 0xbfd62c82f2b9c795
> +        .quad 0x3fd61965cdb02c1f
> +        .quad 0x3fd602d08af091ec
> +        .quad 0x3fd5ec433d5c35ae
> +        .quad 0x3fd5d5bddf595f30
> +        .quad 0x3fd5bf406b543db2
> +        .quad 0x3fd5a8cadbbedfa1
> +        .quad 0x3fd5925d2b112a59
> +        .quad 0x3fd57bf753c8d1fb
> +        .quad 0x3fd565995069514c
> +        .quad 0x3fd54f431b7be1a9
> +        .quad 0x3fd538f4af8f72fe
> +        .quad 0x3fd522ae0738a3d8
> +        .quad 0x3fd50c6f1d11b97c
> +        .quad 0x3fd4f637ebba9810
> +        .quad 0x3fd4e0086dd8baca
> +        .quad 0x3fd4c9e09e172c3c
> +        .quad 0x3fd4b3c077267e9a
> +        .quad 0x3fd49da7f3bcc41f
> +        .quad 0x3fd487970e958770
> +        .quad 0x3fd4718dc271c41b
> +        .quad 0x3fd45b8c0a17df13
> +        .quad 0x3fd44591e0539f49
> +        .quad 0x3fd42f9f3ff62642
> +        .quad 0x3fd419b423d5e8c7
> +        .quad 0x3fd403d086cea79c
> +        .quad 0x3fd3edf463c1683e
> +        .quad 0x3fd3d81fb5946dba
> +        .quad 0x3fd3c25277333184
> +        .quad 0x3fd3ac8ca38e5c5f
> +        .quad 0x3fd396ce359bbf54
> +        .quad 0x3fd3811728564cb2
> +        .quad 0x3fd36b6776be1117
> +        .quad 0x3fd355bf1bd82c8b
> +        .quad 0x3fd3401e12aecba1
> +        .quad 0x3fd32a84565120a8
> +        .quad 0x3fd314f1e1d35ce4
> +        .quad 0x3fd2ff66b04ea9d4
> +        .quad 0x3fd2e9e2bce12286
> +        .quad 0x3fd2d46602adccee
> +        .quad 0x3fd2bef07cdc9354
> +        .quad 0x3fd2a982269a3dbf
> +        .quad 0x3fd2941afb186b7c
> +        .quad 0x3fd27ebaf58d8c9d
> +        .quad 0x3fd269621134db92
> +        .quad 0x3fd25410494e56c7
> +        .quad 0x3fd23ec5991eba49
> +        .quad 0x3fd22981fbef797b
> +        .quad 0x3fd214456d0eb8d4
> +        .quad 0x3fd1ff0fe7cf47a7
> +        .quad 0x3fd1e9e1678899f4
> +        .quad 0x3fd1d4b9e796c245
> +        .quad 0x3fd1bf99635a6b95
> +        .quad 0x3fd1aa7fd638d33f
> +        .quad 0x3fd1956d3b9bc2fa
> +        .quad 0x3fd180618ef18adf
> +        .quad 0x3fd16b5ccbacfb73
> +        .quad 0x3fd1565eed455fc3
> +        .quad 0x3fd14167ef367783
> +        .quad 0x3fd12c77cd00713b
> +        .quad 0x3fd1178e8227e47c
> +        .quad 0x3fd102ac0a35cc1c
> +        .quad 0x3fd0edd060b78081
> +        .quad 0x3fd0d8fb813eb1ef
> +        .quad 0x3fd0c42d676162e3
> +        .quad 0x3fd0af660eb9e279
> +        .quad 0x3fd09aa572e6c6d4
> +        .quad 0x3fd085eb8f8ae797
> +        .quad 0x3fd07138604d5862
> +        .quad 0x3fd05c8be0d9635a
> +        .quad 0x3fd047e60cde83b8
> +        .quad 0x3fd03346e0106062
> +        .quad 0x3fd01eae5626c691
> +        .quad 0x3fd00a1c6adda473
> +        .quad 0x3fcfeb2233ea07cd
> +        .quad 0x3fcfc218be620a5e
> +        .quad 0x3fcf991c6cb3b379
> +        .quad 0x3fcf702d36777df0
> +        .quad 0x3fcf474b134df229
> +        .quad 0x3fcf1e75fadf9bde
> +        .quad 0x3fcef5ade4dcffe6
> +        .quad 0x3fceccf2c8fe920a
> +        .quad 0x3fcea4449f04aaf5
> +        .quad 0x3fce7ba35eb77e2a
> +        .quad 0x3fce530effe71012
> +        .quad 0x3fce2a877a6b2c12
> +        .quad 0x3fce020cc6235ab5
> +        .quad 0x3fcdd99edaf6d7e9
> +        .quad 0x3fcdb13db0d48940
> +        .quad 0x3fcd88e93fb2f450
> +        .quad 0x3fcd60a17f903515
> +        .quad 0x3fcd38666871f465
> +        .quad 0x3fcd1037f2655e7b
> +        .quad 0x3fcce816157f1988
> +        .quad 0x3fccc000c9db3c52
> +        .quad 0x3fcc97f8079d44ec
> +        .quad 0x3fcc6ffbc6f00f71
> +        .quad 0x3fcc480c0005ccd1
> +        .quad 0x3fcc2028ab17f9b4
> +        .quad 0x3fcbf851c067555f
> +        .quad 0x3fcbd087383bd8ad
> +        .quad 0x3fcba8c90ae4ad19
> +        .quad 0x3fcb811730b823d2
> +        .quad 0x3fcb5971a213acdb
> +        .quad 0x3fcb31d8575bce3d
> +        .quad 0x3fcb0a4b48fc1b46
> +        .quad 0x3fcae2ca6f672bd4
> +        .quad 0x3fcabb55c31693ad
> +        .quad 0x3fca93ed3c8ad9e3
> +        .quad 0x3fca6c90d44b704e
> +        .quad 0x3fca454082e6ab05
> +        .quad 0x3fca1dfc40f1b7f1
> +        .quad 0x3fc9f6c407089664
> +        .quad 0x3fc9cf97cdce0ec3
> +        .quad 0x3fc9a8778debaa38
> +        .quad 0x3fc981634011aa75
> +        .quad 0x3fc95a5adcf7017f
> +        .quad 0x3fc9335e5d594989
> +        .quad 0x3fc90c6db9fcbcd9
> +        .quad 0x3fc8e588ebac2dbf
> +        .quad 0x3fc8beafeb38fe8c
> +        .quad 0x3fc897e2b17b19a5
> +        .quad 0x3fc871213750e994
> +        .quad 0x3fc84a6b759f512f
> +        .quad 0x3fc823c16551a3c2
> +        .quad 0x3fc7fd22ff599d4f
> +        .quad 0x3fc7d6903caf5ad0
> +        .quad 0x3fc7b0091651528c
> +        .quad 0x3fc7898d85444c73
> +        .quad 0x3fc7631d82935a86
> +        .quad 0x3fc73cb9074fd14d
> +        .quad 0x3fc716600c914054
> +        .quad 0x3fc6f0128b756abc
> +        .quad 0x3fc6c9d07d203fc7
> +        .quad 0x3fc6a399dabbd383
> +        .quad 0x3fc67d6e9d785771
> +        .quad 0x3fc6574ebe8c133a
> +        .quad 0x3fc6313a37335d76
> +        .quad 0x3fc60b3100b09476
> +        .quad 0x3fc5e533144c1719
> +        .quad 0x3fc5bf406b543db2
> +        .quad 0x3fc59958ff1d52f1
> +        .quad 0x3fc5737cc9018cdd
> +        .quad 0x3fc54dabc26105d2
> +        .quad 0x3fc527e5e4a1b58d
> +        .quad 0x3fc5022b292f6a45
> +        .quad 0x3fc4dc7b897bc1c8
> +        .quad 0x3fc4b6d6fefe22a4
> +        .quad 0x3fc4913d8333b561
> +        .quad 0x3fc46baf0f9f5db7
> +        .quad 0x3fc4462b9dc9b3dc
> +        .quad 0x3fc420b32740fdd4
> +        .quad 0x3fc3fb45a59928cc
> +        .quad 0x3fc3d5e3126bc27f
> +        .quad 0x3fc3b08b6757f2a9
> +        .quad 0x3fc38b3e9e027479
> +        .quad 0x3fc365fcb0159016
> +        .quad 0x3fc340c59741142e
> +        .quad 0x3fc31b994d3a4f85
> +        .quad 0x3fc2f677cbbc0a96
> +        .quad 0x3fc2d1610c86813a
> +        .quad 0x3fc2ac55095f5c59
> +        .quad 0x3fc28753bc11aba5
> +        .quad 0x3fc2625d1e6ddf57
> +        .quad 0x3fc23d712a49c202
> +        .quad 0x3fc2188fd9807263
> +        .quad 0x3fc1f3b925f25d41
> +        .quad 0x3fc1ceed09853752
> +        .quad 0x3fc1aa2b7e23f72a
> +        .quad 0x3fc185747dbecf34
> +        .quad 0x3fc160c8024b27b1
> +        .quad 0x3fc13c2605c398c3
> +        .quad 0x3fc1178e8227e47c
> +        .quad 0x3fc0f301717cf0fb
> +        .quad 0x3fc0ce7ecdccc28d
> +        .quad 0x3fc0aa06912675d5
> +        .quad 0x3fc08598b59e3a07
> +        .quad 0x3fc06135354d4b18
> +        .quad 0x3fc03cdc0a51ec0d
> +        .quad 0x3fc0188d2ecf6140
> +        .quad 0x3fbfe89139dbd566
> +        .quad 0x3fbfa01c9db57ce2
> +        .quad 0x3fbf57bc7d9005db
> +        .quad 0x3fbf0f70cdd992e3
> +        .quad 0x3fbec739830a1120
> +        .quad 0x3fbe7f1691a32d3e
> +        .quad 0x3fbe3707ee30487b
> +        .quad 0x3fbdef0d8d466db9
> +        .quad 0x3fbda727638446a2
> +        .quad 0x3fbd5f55659210e2
> +        .quad 0x3fbd179788219364
> +        .quad 0x3fbccfedbfee13a8
> +        .quad 0x3fbc885801bc4b23
> +        .quad 0x3fbc40d6425a5cb1
> +        .quad 0x3fbbf968769fca11
> +        .quad 0x3fbbb20e936d6974
> +        .quad 0x3fbb6ac88dad5b1c
> +        .quad 0x3fbb23965a52ff00
> +        .quad 0x3fbadc77ee5aea8c
> +        .quad 0x3fba956d3ecade63
> +        .quad 0x3fba4e7640b1bc38
> +        .quad 0x3fba0792e9277cac
> +        .quad 0x3fb9c0c32d4d2548
> +        .quad 0x3fb97a07024cbe74
> +        .quad 0x3fb9335e5d594989
> +        .quad 0x3fb8ecc933aeb6e8
> +        .quad 0x3fb8a6477a91dc29
> +        .quad 0x3fb85fd927506a48
> +        .quad 0x3fb8197e2f40e3f0
> +        .quad 0x3fb7d33687c293c9
> +        .quad 0x3fb78d02263d82d3
> +        .quad 0x3fb746e100226ed9
> +        .quad 0x3fb700d30aeac0e1
> +        .quad 0x3fb6bad83c1883b6
> +        .quad 0x3fb674f089365a7a
> +        .quad 0x3fb62f1be7d77743
> +        .quad 0x3fb5e95a4d9791cb
> +        .quad 0x3fb5a3abb01ade25
> +        .quad 0x3fb55e10050e0384
> +        .quad 0x3fb518874226130a
> +        .quad 0x3fb4d3115d207eac
> +        .quad 0x3fb48dae4bc31018
> +        .quad 0x3fb4485e03dbdfad
> +        .quad 0x3fb403207b414b7f
> +        .quad 0x3fb3bdf5a7d1ee64
> +        .quad 0x3fb378dd7f749714
> +        .quad 0x3fb333d7f8183f4b
> +        .quad 0x3fb2eee507b40301
> +        .quad 0x3fb2aa04a44717a5
> +        .quad 0x3fb26536c3d8c369
> +        .quad 0x3fb2207b5c78549e
> +        .quad 0x3fb1dbd2643d190b
> +        .quad 0x3fb1973bd1465567
> +        .quad 0x3fb152b799bb3cc9
> +        .quad 0x3fb10e45b3cae831
> +        .quad 0x3fb0c9e615ac4e17
> +        .quad 0x3fb08598b59e3a07
> +        .quad 0x3fb0415d89e74444
> +        .quad 0x3faffa6911ab9301
> +        .quad 0x3faf723b517fc523
> +        .quad 0x3faeea31c006b87c
> +        .quad 0x3fae624c4a0b5e1b
> +        .quad 0x3fadda8adc67ee4e
> +        .quad 0x3fad52ed6405d86f
> +        .quad 0x3faccb73cdddb2cc
> +        .quad 0x3fac441e06f72a9e
> +        .quad 0x3fabbcebfc68f420
> +        .quad 0x3fab35dd9b58baad
> +        .quad 0x3faaaef2d0fb10fc
> +        .quad 0x3faa282b8a936171
> +        .quad 0x3fa9a187b573de7c
> +        .quad 0x3fa91b073efd7314
> +        .quad 0x3fa894aa149fb343
> +        .quad 0x3fa80e7023d8ccc4
> +        .quad 0x3fa788595a3577ba
> +        .quad 0x3fa70265a550e777
> +        .quad 0x3fa67c94f2d4bb58
> +        .quad 0x3fa5f6e73078efb8
> +        .quad 0x3fa5715c4c03ceef
> +        .quad 0x3fa4ebf43349e26f
> +        .quad 0x3fa466aed42de3ea
> +        .quad 0x3fa3e18c1ca0ae92
> +        .quad 0x3fa35c8bfaa1306b
> +        .quad 0x3fa2d7ae5c3c5bae
> +        .quad 0x3fa252f32f8d183f
> +        .quad 0x3fa1ce5a62bc353a
> +        .quad 0x3fa149e3e4005a8d
> +        .quad 0x3fa0c58fa19dfaaa
> +        .quad 0x3fa0415d89e74444
> +        .quad 0x3f9f7a9b16782856
> +        .quad 0x3f9e72bf2813ce51
> +        .quad 0x3f9d6b2725979802
> +        .quad 0x3f9c63d2ec14aaf2
> +        .quad 0x3f9b5cc258b718e6
> +        .quad 0x3f9a55f548c5c43f
> +        .quad 0x3f994f6b99a24475
> +        .quad 0x3f98492528c8cabf
> +        .quad 0x3f974321d3d006d3
> +        .quad 0x3f963d6178690bd6
> +        .quad 0x3f9537e3f45f3565
> +        .quad 0x3f9432a925980cc1
> +        .quad 0x3f932db0ea132e22
> +        .quad 0x3f9228fb1fea2e28
> +        .quad 0x3f912487a5507f70
> +        .quad 0x3f90205658935847
> +        .quad 0x3f8e38ce3033310c
> +        .quad 0x3f8c317384c75f06
> +        .quad 0x3f8a2a9c6c170462
> +        .quad 0x3f882448a388a2aa
> +        .quad 0x3f861e77e8b53fc6
> +        .quad 0x3f841929f96832f0
> +        .quad 0x3f82145e939ef1e9
> +        .quad 0x3f8010157588de71
> +        .quad 0x3f7c189cbb0e27fb
> +        .quad 0x3f78121214586b54
> +        .quad 0x3f740c8a747878e2
> +        .quad 0x3f70080559588b35
> +        .quad 0x3f680904828985c0
> +        .quad 0x3f60040155d5889e
> +        .quad 0x3f50020055655889
> +        .quad 0x0000000000000000
> +        /*== poly_coeff[4] ==*/
> +        .align 32
> +        .quad 0x3fc9999CACDB4D0A, 0x3fc9999CACDB4D0A, 0x3fc9999CACDB4D0A, 0x3fc9999CACDB4D0A /* coeff4 */
> +        .quad 0xbfd0000148058EE1, 0xbfd0000148058EE1, 0xbfd0000148058EE1, 0xbfd0000148058EE1 /* coeff3 */
> +        .quad 0x3fd55555555543C5, 0x3fd55555555543C5, 0x3fd55555555543C5, 0x3fd55555555543C5 /* coeff2 */
> +        .quad 0xbfdFFFFFFFFFF81F, 0xbfdFFFFFFFFFF81F, 0xbfdFFFFFFFFFF81F, 0xbfdFFFFFFFFFF81F /* coeff1 */
> +        /*== ExpMask ==*/
> +        .align 32
> +        .quad 0x000fffffffffffff, 0x000fffffffffffff, 0x000fffffffffffff, 0x000fffffffffffff
> +        /*== Two10 ==*/
> +        .align 32
> +        .quad 0x3f50000000000000, 0x3f50000000000000, 0x3f50000000000000, 0x3f50000000000000
> +        /*== MinLog1p = -1+2^(-53) ==*/
> +        .align 32
> +        .quad 0xbfefffffffffffff, 0xbfefffffffffffff, 0xbfefffffffffffff, 0xbfefffffffffffff
> +        /*== MaxLog1p ==*/
> +        .align 32
> +        .quad 0x7f3ffffffffff000, 0x7f3ffffffffff000, 0x7f3ffffffffff000, 0x7f3ffffffffff000
> +        /*== One ==*/
> +        .align 32
> +        .quad 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000
> +        /*== SgnMask ==*/
> +        .align 32
> +        .quad 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff
> +        /*== XThreshold ==*/
> +        .align 32
> +        .quad 0x3e00000000000000, 0x3e00000000000000, 0x3e00000000000000, 0x3e00000000000000
> +        /*== XhMask ==*/
> +        .align 32
> +        .quad 0xfffffffffffffc00, 0xfffffffffffffc00, 0xfffffffffffffc00, 0xfffffffffffffc00
> +        /*== Threshold ==*/
> +        .align 32
> +        .quad 0x4086a00000000000, 0x4086a00000000000, 0x4086a00000000000, 0x4086a00000000000
> +        /*== Bias ==*/
> +        .align 32
> +        .quad 0x408ff80000000000, 0x408ff80000000000, 0x408ff80000000000, 0x408ff80000000000
> +        /*== Bias1 ==*/
> +        .align 32
> +        .quad 0x408ff00000000000, 0x408ff00000000000, 0x408ff00000000000, 0x408ff00000000000
> +        /*== ExpMask ==*/
> +        .align 32
> +        .quad 0x7ff0000000000000, 0x7ff0000000000000, 0x7ff0000000000000, 0x7ff0000000000000
> +        /*== ExpMask2 ==*/
> +        .align 32
> +        .quad 0x7f40000000000000, 0x7f40000000000000, 0x7f40000000000000, 0x7f40000000000000
> +        /*== L2L ==*/
> +        .align 32
> +        .quad 0x3fe62E42FEFA39EF, 0x3fe62E42FEFA39EF, 0x3fe62E42FEFA39EF, 0x3fe62E42FEFA39EF
> +        /*== dBigThreshold ==*/
> +        .align 32
> +        .quad 0x41D0000000000000, 0x41D0000000000000, 0x41D0000000000000, 0x41D0000000000000
> +        /*== dC2 ==*/
> +        .align 32
> +        .quad 0x3FD8000000000000, 0x3FD8000000000000, 0x3FD8000000000000, 0x3FD8000000000000
> +        /*== dC3 ==*/
> +        .align 32
> +        .quad 0x3FD4000000000000, 0x3FD4000000000000, 0x3FD4000000000000, 0x3FD4000000000000
> +        /*== dC4 ==*/
> +        .align 32
> +        .quad 0x3FD1800000000000, 0x3FD1800000000000, 0x3FD1800000000000, 0x3FD1800000000000
> +        /*== dC5 ==*/
> +        .align 32
> +        .quad 0x3FCF800000000000, 0x3FCF800000000000, 0x3FCF800000000000, 0x3FCF800000000000
> +        /*== dHalf ==*/
> +        .align 32
> +        .quad 0x3FE0000000000000, 0x3FE0000000000000, 0x3FE0000000000000, 0x3FE0000000000000
> +        /*== dLargestFinite ==*/
> +        .align 32
> +        .quad 0x7FEFFFFFFFFFFFFF, 0x7FEFFFFFFFFFFFFF, 0x7FEFFFFFFFFFFFFF, 0x7FEFFFFFFFFFFFFF
> +        /*== dLittleThreshold ==*/
> +        .align 32
> +        .quad 0x3F60000000000000, 0x3F60000000000000, 0x3F60000000000000, 0x3F60000000000000
> +        /*== dSign ==*/
> +        .align 32
> +        .quad 0x8000000000000000, 0x8000000000000000, 0x8000000000000000, 0x8000000000000000
> +        /*== dThirtyOne ==*/
> +        .align 32
> +        .quad 0x403F000000000000, 0x403F000000000000, 0x403F000000000000, 0x403F000000000000
> +        /*== dTopMask12 ==*/
> +        .align 32
> +        .quad 0xFFFFFE0000000000, 0xFFFFFE0000000000, 0xFFFFFE0000000000, 0xFFFFFE0000000000
> +        /*== dTopMask29 ==*/
> +        .align 32
> +        .quad 0xFFFFFFFFFF000000, 0xFFFFFFFFFF000000, 0xFFFFFFFFFF000000, 0xFFFFFFFFFF000000
> +        /*== XScale ==*/
> +        .align 32
> +        .quad 0x3E10000000000000, 0x3E10000000000000, 0x3E10000000000000, 0x3E10000000000000
> +        .align 32
> +        .type	__svml_dasinh_data_internal,@object
> +        .size	__svml_dasinh_data_internal,.-__svml_dasinh_data_internal
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core-avx2.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core-avx2.S
> new file mode 100644
> index 0000000000..647c73292c
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core-avx2.S
> @@ -0,0 +1,20 @@
> +/* AVX2 version of vectorized asinh, vector length is 8.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#define _ZGVeN8v_asinh _ZGVeN8v_asinh_avx2_wrapper
> +#include "../svml_d_asinh8_core.S"
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core.c b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core.c
> new file mode 100644
> index 0000000000..45e5ab72a6
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core.c
> @@ -0,0 +1,27 @@
> +/* Multiple versions of vectorized asinh, vector length is 8.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#define SYMBOL_NAME _ZGVeN8v_asinh
> +#include "ifunc-mathvec-avx512-skx.h"
> +
> +libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
> +
> +#ifdef SHARED
> +__hidden_ver1 (_ZGVeN8v_asinh, __GI__ZGVeN8v_asinh, __redirect__ZGVeN8v_asinh)
> +  __attribute__ ((visibility ("hidden")));
> +#endif
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core_avx512.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core_avx512.S
> new file mode 100644
> index 0000000000..8100e8a50a
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core_avx512.S
> @@ -0,0 +1,510 @@
> +/* Function asinh vectorized with AVX-512.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   https://www.gnu.org/licenses/.  */
> +
> +/*
> + * ALGORITHM DESCRIPTION:
> + *
> + *   Compute asinh(x) as log(x + sqrt(x*x + 1))
> + *   using RSQRT instructions for starting the
> + *   square root approximation, and small table lookups for log
> + *   that map to AVX-512 permute instructions
> + *
> + *   Special cases:
> + *
> + *   asinh(NaN) = quiet NaN, and raise invalid exception
> + *   asinh(INF) = that INF
> + *   asinh(0)   = that 0
> + *
> + */
> +
> +/* Offsets for data table __svml_dasinh_data_internal_avx512
> + */
> +#define Log_tbl_H                     	0
> +#define Log_tbl_L                     	128
> +#define One                           	256
> +#define AbsMask                       	320
> +#define SmallThreshold                	384
> +#define Threshold                     	448
> +#define LargeThreshold                	512
> +#define ca2                           	576
> +#define ca1                           	640
> +#define c4s                           	704
> +#define c3s                           	768
> +#define c2s                           	832
> +#define c1s                           	896
> +#define AddB5                         	960
> +#define RcpBitMask                    	1024
> +#define OneEighth                     	1088
> +#define Four                          	1152
> +#define poly_coeff9                   	1216
> +#define poly_coeff8                   	1280
> +#define poly_coeff7                   	1344
> +#define poly_coeff6                   	1408
> +#define poly_coeff5                   	1472
> +#define poly_coeff4                   	1536
> +#define poly_coeff3                   	1600
> +#define poly_coeff2                   	1664
> +#define poly_coeff1                   	1728
> +#define L2H                           	1792
> +#define L2L                           	1856
> +
> +#include <sysdep.h>
> +
> +        .text
> +	.section .text.evex512,"ax",@progbits
> +ENTRY(_ZGVeN8v_asinh_skx)
> +        pushq     %rbp
> +        cfi_def_cfa_offset(16)
> +        movq      %rsp, %rbp
> +        cfi_def_cfa(6, 16)
> +        cfi_offset(6, -16)
> +        andq      $-64, %rsp
> +        subq      $192, %rsp
> +        vmovaps   %zmm0, %zmm3
> +
> +/* x^2 */
> +        vmulpd    {rn-sae}, %zmm3, %zmm3, %zmm14
> +        vmovups   One+__svml_dasinh_data_internal_avx512(%rip), %zmm9
> +
> +/* polynomial computation for small inputs */
> +        vmovups   ca2+__svml_dasinh_data_internal_avx512(%rip), %zmm10
> +        vmovups   ca1+__svml_dasinh_data_internal_avx512(%rip), %zmm11
> +
> +/* not a very small input ? */
> +        vmovups   SmallThreshold+__svml_dasinh_data_internal_avx512(%rip), %zmm0
> +
> +/* A=max(x^2, 1); */
> +        vmaxpd    {sae}, %zmm14, %zmm9, %zmm4
> +
> +/* B=min(x^2, 1); */
> +        vminpd    {sae}, %zmm14, %zmm9, %zmm5
> +        vfmadd231pd {rn-sae}, %zmm14, %zmm10, %zmm11
> +
> +/* 1+x^2 */
> +        vaddpd    {rn-sae}, %zmm9, %zmm14, %zmm8
> +
> +/* |input| */
> +        vandpd    AbsMask+__svml_dasinh_data_internal_avx512(%rip), %zmm3, %zmm1
> +        vrsqrt14pd %zmm8, %zmm6
> +        vcmppd    $21, {sae}, %zmm0, %zmm1, %k2
> +
> +/* B_high */
> +        vsubpd    {rn-sae}, %zmm4, %zmm8, %zmm7
> +
> +/* sign bit */
> +        vxorpd    %zmm3, %zmm1, %zmm2
> +        vmulpd    {rn-sae}, %zmm14, %zmm11, %zmm4
> +
> +/* B_low */
> +        vsubpd    {rn-sae}, %zmm7, %zmm5, %zmm13
> +        vmovups   c2s+__svml_dasinh_data_internal_avx512(%rip), %zmm5
> +        vmovups   c1s+__svml_dasinh_data_internal_avx512(%rip), %zmm7
> +
> +/* polynomial computation for small inputs */
> +        vfmadd213pd {rn-sae}, %zmm1, %zmm1, %zmm4
> +
> +/* (x^2)_low */
> +        vmovaps   %zmm3, %zmm15
> +        vfmsub213pd {rn-sae}, %zmm14, %zmm3, %zmm15
> +
> +/* Sh ~sqrt(1+x^2) */
> +        vmulpd    {rn-sae}, %zmm6, %zmm8, %zmm14
> +
> +/* Yl = (x^2)_low + B_low */
> +        vaddpd    {rn-sae}, %zmm15, %zmm13, %zmm13
> +
> +/* very large inputs ? */
> +        vmovups   Threshold+__svml_dasinh_data_internal_avx512(%rip), %zmm15
> +
> +/* (Yh*R0)_low */
> +        vfmsub213pd {rn-sae}, %zmm14, %zmm6, %zmm8
> +        vcmppd    $21, {sae}, %zmm15, %zmm1, %k1
> +
> +/* Sl = (Yh*R0)_low+(R0*Yl) */
> +        vfmadd213pd {rn-sae}, %zmm8, %zmm6, %zmm13
> +        vmovups   LargeThreshold+__svml_dasinh_data_internal_avx512(%rip), %zmm8
> +
> +/* rel. error term: Eh=1-Sh*R0 */
> +        vmovaps   %zmm9, %zmm12
> +        vfnmadd231pd {rn-sae}, %zmm14, %zmm6, %zmm12
> +        vcmppd    $22, {sae}, %zmm8, %zmm1, %k0
> +
> +/* rel. error term: Eh=(1-Sh*R0)-Sl*R0 */
> +        vfnmadd231pd {rn-sae}, %zmm13, %zmm6, %zmm12
> +
> +/*
> + * sqrt(1+x^2) ~ Sh + Sl + Sh*Eh*poly_s
> + * poly_s = c1+c2*Eh+c3*Eh^2
> + */
> +        vmovups   c4s+__svml_dasinh_data_internal_avx512(%rip), %zmm6
> +        vmovups   c3s+__svml_dasinh_data_internal_avx512(%rip), %zmm8
> +
> +/* Sh*Eh */
> +        vmulpd    {rn-sae}, %zmm12, %zmm14, %zmm11
> +        vfmadd231pd {rn-sae}, %zmm12, %zmm6, %zmm8
> +
> +/* Sh+x */
> +        vaddpd    {rn-sae}, %zmm1, %zmm14, %zmm6
> +        kmovw     %k0, %edx
> +        vfmadd213pd {rn-sae}, %zmm5, %zmm12, %zmm8
> +        vfmadd213pd {rn-sae}, %zmm7, %zmm12, %zmm8
> +
> +/* Xh */
> +        vsubpd    {rn-sae}, %zmm14, %zmm6, %zmm12
> +
> +/* Sl + Sh*Eh*poly_s */
> +        vfmadd213pd {rn-sae}, %zmm13, %zmm8, %zmm11
> +
> +/* fixup for very large inputs */
> +        vmovups   OneEighth+__svml_dasinh_data_internal_avx512(%rip), %zmm8
> +
> +/* Xl */
> +        vsubpd    {rn-sae}, %zmm12, %zmm1, %zmm12
> +
> +/* Xin0+Sl+Sh*Eh*poly_s ~ x+sqrt(1+x^2) */
> +        vaddpd    {rn-sae}, %zmm11, %zmm6, %zmm10
> +
> +/* Sl_high */
> +        vsubpd    {rn-sae}, %zmm6, %zmm10, %zmm5
> +        vmulpd    {rn-sae}, %zmm8, %zmm1, %zmm10{%k1}
> +
> +/* Table lookups */
> +        vmovups   __svml_dasinh_data_internal_avx512(%rip), %zmm6
> +
> +/* Sl_l */
> +        vsubpd    {rn-sae}, %zmm5, %zmm11, %zmm7
> +        vrcp14pd  %zmm10, %zmm13
> +
> +/* Xin_low */
> +        vaddpd    {rn-sae}, %zmm12, %zmm7, %zmm14
> +        vmovups   Log_tbl_L+__svml_dasinh_data_internal_avx512(%rip), %zmm7
> +        vmovups   poly_coeff6+__svml_dasinh_data_internal_avx512(%rip), %zmm12
> +
> +/* round reciprocal to 1+4b mantissas */
> +        vpaddq    AddB5+__svml_dasinh_data_internal_avx512(%rip), %zmm13, %zmm11
> +
> +/* fixup for very large inputs */
> +        vxorpd    %zmm14, %zmm14, %zmm14{%k1}
> +        vmovups   poly_coeff5+__svml_dasinh_data_internal_avx512(%rip), %zmm13
> +        vandpd    RcpBitMask+__svml_dasinh_data_internal_avx512(%rip), %zmm11, %zmm15
> +        vmovups   poly_coeff7+__svml_dasinh_data_internal_avx512(%rip), %zmm11
> +
> +/* Prepare table index */
> +        vpsrlq    $48, %zmm15, %zmm5
> +
> +/* reduced argument for log(): (Rcp*Xin-1)+Rcp*Xin_low */
> +        vfmsub231pd {rn-sae}, %zmm15, %zmm10, %zmm9
> +
> +/* exponents */
> +        vgetexppd {sae}, %zmm15, %zmm8
> +        vmovups   Four+__svml_dasinh_data_internal_avx512(%rip), %zmm10
> +        vpermt2pd Log_tbl_H+64+__svml_dasinh_data_internal_avx512(%rip), %zmm5, %zmm6
> +        vpermt2pd Log_tbl_L+64+__svml_dasinh_data_internal_avx512(%rip), %zmm5, %zmm7
> +        vsubpd    {rn-sae}, %zmm10, %zmm8, %zmm8{%k1}
> +        vfmadd231pd {rn-sae}, %zmm15, %zmm14, %zmm9
> +
> +/* polynomials */
> +        vmovups   poly_coeff9+__svml_dasinh_data_internal_avx512(%rip), %zmm10
> +        vmovups   poly_coeff8+__svml_dasinh_data_internal_avx512(%rip), %zmm5
> +        vmovups   poly_coeff4+__svml_dasinh_data_internal_avx512(%rip), %zmm14
> +
> +/* -K*L2H + Th */
> +        vmovups   L2H+__svml_dasinh_data_internal_avx512(%rip), %zmm15
> +        vfmadd231pd {rn-sae}, %zmm9, %zmm10, %zmm5
> +
> +/* -K*L2L + Tl */
> +        vmovups   L2L+__svml_dasinh_data_internal_avx512(%rip), %zmm10
> +        vfnmadd231pd {rn-sae}, %zmm8, %zmm15, %zmm6
> +        vfmadd213pd {rn-sae}, %zmm11, %zmm9, %zmm5
> +        vfnmadd213pd {rn-sae}, %zmm7, %zmm10, %zmm8
> +        vmovups   poly_coeff3+__svml_dasinh_data_internal_avx512(%rip), %zmm7
> +        vmovups   poly_coeff1+__svml_dasinh_data_internal_avx512(%rip), %zmm10
> +
> +/* R^2 */
> +        vmulpd    {rn-sae}, %zmm9, %zmm9, %zmm11
> +        vfmadd213pd {rn-sae}, %zmm12, %zmm9, %zmm5
> +        vfmadd213pd {rn-sae}, %zmm13, %zmm9, %zmm5
> +        vfmadd213pd {rn-sae}, %zmm14, %zmm9, %zmm5
> +        vfmadd213pd {rn-sae}, %zmm7, %zmm9, %zmm5
> +        vmovups   poly_coeff2+__svml_dasinh_data_internal_avx512(%rip), %zmm7
> +        vfmadd213pd {rn-sae}, %zmm7, %zmm9, %zmm5
> +        vfmadd213pd {rn-sae}, %zmm10, %zmm9, %zmm5
> +
> +/* Tl + R^2*Poly */
> +        vfmadd213pd {rn-sae}, %zmm8, %zmm11, %zmm5
> +
> +/* R+Tl + R^2*Poly */
> +        vaddpd    {rn-sae}, %zmm9, %zmm5, %zmm9
> +        vaddpd    {rn-sae}, %zmm9, %zmm6, %zmm4{%k2}
> +        vxorpd    %zmm2, %zmm4, %zmm0
> +        testl     %edx, %edx
> +
> +/* Go to special inputs processing branch */
> +        jne       L(SPECIAL_VALUES_BRANCH)
> +                                # LOE rbx r12 r13 r14 r15 edx zmm0 zmm3
> +
> +/* Restore registers
> + * and exit the function
> + */
> +
> +L(EXIT):
> +        movq      %rbp, %rsp
> +        popq      %rbp
> +        cfi_def_cfa(7, 8)
> +        cfi_restore(6)
> +        ret
> +        cfi_def_cfa(6, 16)
> +        cfi_offset(6, -16)
> +
> +/* Branch to process
> + * special inputs
> + */
> +
> +L(SPECIAL_VALUES_BRANCH):
> +        vmovups   %zmm3, 64(%rsp)
> +        vmovups   %zmm0, 128(%rsp)
> +                                # LOE rbx r12 r13 r14 r15 edx zmm0
> +
> +        xorl      %eax, %eax
> +                                # LOE rbx r12 r13 r14 r15 eax edx
> +
> +        vzeroupper
> +        movq      %r12, 16(%rsp)
> +        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -176; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x50, 0xff, 0xff, 0xff, 0x22
> +        movl      %eax, %r12d
> +        movq      %r13, 8(%rsp)
> +        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -184; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x48, 0xff, 0xff, 0xff, 0x22
> +        movl      %edx, %r13d
> +        movq      %r14, (%rsp)
> +        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -192; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x40, 0xff, 0xff, 0xff, 0x22
> +                                # LOE rbx r15 r12d r13d
> +
> +/* Range mask
> + * bits check
> + */
> +
> +L(RANGEMASK_CHECK):
> +        btl       %r12d, %r13d
> +
> +/* Call scalar math function */
> +        jc        L(SCALAR_MATH_CALL)
> +                                # LOE rbx r15 r12d r13d
> +
> +/* Special inputs
> + * processing loop
> + */
> +
> +L(SPECIAL_VALUES_LOOP):
> +        incl      %r12d
> +        cmpl      $8, %r12d
> +
> +/* Check bits in range mask */
> +        jl        L(RANGEMASK_CHECK)
> +                                # LOE rbx r15 r12d r13d
> +
> +        movq      16(%rsp), %r12
> +        cfi_restore(12)
> +        movq      8(%rsp), %r13
> +        cfi_restore(13)
> +        movq      (%rsp), %r14
> +        cfi_restore(14)
> +        vmovups   128(%rsp), %zmm0
> +
> +/* Go to exit */
> +        jmp       L(EXIT)
> +        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -176; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x50, 0xff, 0xff, 0xff, 0x22
> +        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -184; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x48, 0xff, 0xff, 0xff, 0x22
> +        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -192; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x40, 0xff, 0xff, 0xff, 0x22
> +                                # LOE rbx r12 r13 r14 r15 zmm0
> +
> +/* Scalar math fucntion call
> + * to process special input
> + */
> +
> +L(SCALAR_MATH_CALL):
> +        movl      %r12d, %r14d
> +        movsd     64(%rsp,%r14,8), %xmm0
> +        call      asinh@PLT
> +                                # LOE rbx r14 r15 r12d r13d xmm0
> +
> +        movsd     %xmm0, 128(%rsp,%r14,8)
> +
> +/* Process special inputs in loop */
> +        jmp       L(SPECIAL_VALUES_LOOP)
> +                                # LOE rbx r15 r12d r13d
> +END(_ZGVeN8v_asinh_skx)
> +
> +        .section .rodata, "a"
> +        .align 64
> +
> +#ifdef __svml_dasinh_data_internal_avx512_typedef
> +typedef unsigned int VUINT32;
> +typedef struct {
> +        __declspec(align(64)) VUINT32 Log_tbl_H[16][2];
> +        __declspec(align(64)) VUINT32 Log_tbl_L[16][2];
> +        __declspec(align(64)) VUINT32 One[8][2];
> +        __declspec(align(64)) VUINT32 AbsMask[8][2];
> +        __declspec(align(64)) VUINT32 SmallThreshold[8][2];
> +        __declspec(align(64)) VUINT32 Threshold[8][2];
> +        __declspec(align(64)) VUINT32 LargeThreshold[8][2];
> +        __declspec(align(64)) VUINT32 ca2[8][2];
> +        __declspec(align(64)) VUINT32 ca1[8][2];
> +        __declspec(align(64)) VUINT32 c4s[8][2];
> +        __declspec(align(64)) VUINT32 c3s[8][2];
> +        __declspec(align(64)) VUINT32 c2s[8][2];
> +        __declspec(align(64)) VUINT32 c1s[8][2];
> +        __declspec(align(64)) VUINT32 AddB5[8][2];
> +        __declspec(align(64)) VUINT32 RcpBitMask[8][2];
> +        __declspec(align(64)) VUINT32 OneEighth[8][2];
> +        __declspec(align(64)) VUINT32 Four[8][2];
> +        __declspec(align(64)) VUINT32 poly_coeff9[8][2];
> +        __declspec(align(64)) VUINT32 poly_coeff8[8][2];
> +        __declspec(align(64)) VUINT32 poly_coeff7[8][2];
> +        __declspec(align(64)) VUINT32 poly_coeff6[8][2];
> +        __declspec(align(64)) VUINT32 poly_coeff5[8][2];
> +        __declspec(align(64)) VUINT32 poly_coeff4[8][2];
> +        __declspec(align(64)) VUINT32 poly_coeff3[8][2];
> +        __declspec(align(64)) VUINT32 poly_coeff2[8][2];
> +        __declspec(align(64)) VUINT32 poly_coeff1[8][2];
> +        __declspec(align(64)) VUINT32 L2H[8][2];
> +        __declspec(align(64)) VUINT32 L2L[8][2];
> +    } __svml_dasinh_data_internal_avx512;
> +#endif
> +__svml_dasinh_data_internal_avx512:
> +        /*== Log_tbl_H ==*/
> +        .quad 0x0000000000000000
> +        .quad 0xbfaf0a30c0120000
> +        .quad 0xbfbe27076e2b0000
> +        .quad 0xbfc5ff3070a78000
> +        .quad 0xbfcc8ff7c79a8000
> +        .quad 0xbfd1675cababc000
> +        .quad 0xbfd4618bc21c4000
> +        .quad 0xbfd739d7f6bbc000
> +        .quad 0xbfd9f323ecbf8000
> +        .quad 0xbfdc8ff7c79a8000
> +        .quad 0xbfdf128f5faf0000
> +        .quad 0xbfe0be72e4252000
> +        .quad 0xbfe1e85f5e704000
> +        .quad 0xbfe307d7334f2000
> +        .quad 0xbfe41d8fe8468000
> +        .quad 0xbfe52a2d265bc000
> +        /*== Log_tbl_L ==*/
> +        .align 64
> +        .quad 0x0000000000000000
> +        .quad 0x3d53ab33d066d1d2
> +        .quad 0x3d2a342c2af0003c
> +        .quad 0xbd43d3c873e20a07
> +        .quad 0xbd4a21ac25d81ef3
> +        .quad 0x3d59f1fc63382a8f
> +        .quad 0xbd5ec27d0b7b37b3
> +        .quad 0xbd50069ce24c53fb
> +        .quad 0xbd584bf2b68d766f
> +        .quad 0xbd5a21ac25d81ef3
> +        .quad 0xbd3bb2cd720ec44c
> +        .quad 0xbd55056d312f7668
> +        .quad 0xbd1a07bd8b34be7c
> +        .quad 0x3d5e83c094debc15
> +        .quad 0x3d5aa33736867a17
> +        .quad 0xbd46abb9df22bc57
> +        /*== One ==*/
> +        .align 64
> +        .quad 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000
> +        /*== AbsMask ==*/
> +        .align 64
> +        .quad 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff
> +        /*== SmallThreshold ==*/
> +        .align 64
> +        .quad 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000
> +        /*== Threshold ==*/
> +        .align 64
> +        .quad 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000
> +        /*== LargeThreshold ==*/
> +        .align 64
> +        .quad 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff
> +        /*== ca2 ==*/
> +        .align 64
> +        .quad 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7
> +        /*== ca1 ==*/
> +        .align 64
> +        .quad 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e
> +        /*== c4s ==*/
> +        .align 64
> +        .quad 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612
> +        /*== c3s ==*/
> +        .align 64
> +        .quad 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000
> +        /*== c2s ==*/
> +        .align 64
> +        .quad 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000
> +        /*== c1s ==*/
> +        .align 64
> +        .quad 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000
> +        /*== AddB5 ==*/
> +        .align 64
> +        .quad 0x0000800000000000, 0x0000800000000000, 0x0000800000000000, 0x0000800000000000, 0x0000800000000000, 0x0000800000000000, 0x0000800000000000, 0x0000800000000000
> +        /*== RcpBitMask ==*/
> +        .align 64
> +        .quad 0xffff000000000000, 0xffff000000000000, 0xffff000000000000, 0xffff000000000000, 0xffff000000000000, 0xffff000000000000, 0xffff000000000000, 0xffff000000000000
> +        /*==OneEighth ==*/
> +        .align 64
> +        .quad 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000
> +        /*== Four ==*/
> +        .align 64
> +        .quad 0x4010000000000000, 0x4010000000000000, 0x4010000000000000, 0x4010000000000000, 0x4010000000000000, 0x4010000000000000, 0x4010000000000000, 0x4010000000000000
> +        /*== poly_coeff9 ==*/
> +        .align 64
> +        .quad 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368
> +        /*== poly_coeff8 ==*/
> +        .align 64
> +        .quad 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778
> +        /*== poly_coeff7 ==*/
> +        .align 64
> +        .quad 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9
> +        /*== poly_coeff6 ==*/
> +        .align 64
> +        .quad 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1
> +        /*== poly_coeff5 ==*/
> +        .align 64
> +        .quad 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736
> +        /*== poly_coeff4 ==*/
> +        .align 64
> +        .quad 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af
> +        /*== poly_coeff3 ==*/
> +        .align 64
> +        .quad 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65
> +        /*== poly_coeff2 ==*/
> +        .align 64
> +        .quad 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1
> +        /*== poly_coeff1 ==*/
> +        .align 64
> +        .quad 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000
> +        /*== L2H = log(2)_high ==*/
> +        .align 64
> +        .quad 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000
> +        /*== L2L = log(2)_low ==*/
> +        .align 64
> +        .quad 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000
> +        .align 64
> +        .type	__svml_dasinh_data_internal_avx512,@object
> +        .size	__svml_dasinh_data_internal_avx512,.-__svml_dasinh_data_internal_avx512
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core-avx2.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core-avx2.S
> new file mode 100644
> index 0000000000..7dfd95e400
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core-avx2.S
> @@ -0,0 +1,20 @@
> +/* AVX2 version of vectorized asinhf.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#define _ZGVeN16v_asinhf _ZGVeN16v_asinhf_avx2_wrapper
> +#include "../svml_s_asinhf16_core.S"
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core.c b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core.c
> new file mode 100644
> index 0000000000..dc770a0e65
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core.c
> @@ -0,0 +1,28 @@
> +/* Multiple versions of vectorized asinhf, vector length is 16.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#define SYMBOL_NAME _ZGVeN16v_asinhf
> +#include "ifunc-mathvec-avx512-skx.h"
> +
> +libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
> +
> +#ifdef SHARED
> +__hidden_ver1 (_ZGVeN16v_asinhf, __GI__ZGVeN16v_asinhf,
> +	       __redirect__ZGVeN16v_asinhf)
> +  __attribute__ ((visibility ("hidden")));
> +#endif
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core_avx512.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core_avx512.S
> new file mode 100644
> index 0000000000..fc6a8e7cd3
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core_avx512.S
> @@ -0,0 +1,476 @@
> +/* Function asinhf vectorized with AVX-512.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   https://www.gnu.org/licenses/.  */
> +
> +/*
> + * ALGORITHM DESCRIPTION:
> + *
> + *   Compute asinh(x) as log(x + sqrt(x*x + 1))
> + *   using RSQRT instructions for starting the
> + *   square root approximation, and small table lookups for log
> + *   that map to AVX-512 permute instructions
> + *
> + *   Special cases:
> + *
> + *   asinh(NaN) = quiet NaN, and raise invalid exception
> + *   asinh(INF) = that INF
> + *   asinh(0)   = that 0
> + *
> + */
> +
> +/* Offsets for data table __svml_sasinh_data_internal_avx512
> + */
> +#define Log_tbl_H                     	0
> +#define Log_tbl_L                     	128
> +#define One                           	256
> +#define AbsMask                       	320
> +#define SmallThreshold                	384
> +#define Threshold                     	448
> +#define LargeThreshold                	512
> +#define ca1                           	576
> +#define c2s                           	640
> +#define c1s                           	704
> +#define AddB5                         	768
> +#define RcpBitMask                    	832
> +#define OneEighth                     	896
> +#define Four                          	960
> +#define poly_coeff3                   	1024
> +#define poly_coeff2                   	1088
> +#define poly_coeff1                   	1152
> +#define L2H                           	1216
> +#define L2L                           	1280
> +
> +#include <sysdep.h>
> +
> +        .text
> +	.section .text.exex512,"ax",@progbits
> +ENTRY(_ZGVeN16v_asinhf_skx)
> +        pushq     %rbp
> +        cfi_def_cfa_offset(16)
> +        movq      %rsp, %rbp
> +        cfi_def_cfa(6, 16)
> +        cfi_offset(6, -16)
> +        andq      $-64, %rsp
> +        subq      $192, %rsp
> +        vmovaps   %zmm0, %zmm10
> +
> +/* x^2 */
> +        vmulps    {rn-sae}, %zmm10, %zmm10, %zmm0
> +        vmovups   One+__svml_sasinh_data_internal_avx512(%rip), %zmm2
> +
> +/* polynomial computation for small inputs */
> +        vmovups   ca1+__svml_sasinh_data_internal_avx512(%rip), %zmm1
> +
> +/* not a very small input ? */
> +        vmovups   SmallThreshold+__svml_sasinh_data_internal_avx512(%rip), %zmm11
> +
> +/* 1+x^2 */
> +        vaddps    {rn-sae}, %zmm2, %zmm0, %zmm7
> +
> +/* |input| */
> +        vandps    AbsMask+__svml_sasinh_data_internal_avx512(%rip), %zmm10, %zmm12
> +
> +/* A=max(x^2, 1); */
> +        vmaxps    {sae}, %zmm0, %zmm2, %zmm14
> +        vrsqrt14ps %zmm7, %zmm8
> +
> +/* B=min(x^2, 1); */
> +        vminps    {sae}, %zmm0, %zmm2, %zmm15
> +        vcmpps    $21, {sae}, %zmm11, %zmm12, %k2
> +
> +/* B_high */
> +        vsubps    {rn-sae}, %zmm14, %zmm7, %zmm9
> +
> +/* sign bit */
> +        vxorps    %zmm10, %zmm12, %zmm13
> +
> +/* Sh ~sqrt(1+x^2) */
> +        vmulps    {rn-sae}, %zmm8, %zmm7, %zmm6
> +        vmovups   LargeThreshold+__svml_sasinh_data_internal_avx512(%rip), %zmm14
> +
> +/* B_low */
> +        vsubps    {rn-sae}, %zmm9, %zmm15, %zmm3
> +
> +/* Sh+x */
> +        vaddps    {rn-sae}, %zmm12, %zmm6, %zmm15
> +
> +/* (Yh*R0)_low */
> +        vfmsub213ps {rn-sae}, %zmm6, %zmm8, %zmm7
> +        vmulps    {rn-sae}, %zmm1, %zmm0, %zmm9
> +        vcmpps    $22, {sae}, %zmm14, %zmm12, %k0
> +        vmovups   c1s+__svml_sasinh_data_internal_avx512(%rip), %zmm1
> +
> +/* polynomial computation for small inputs */
> +        vfmadd213ps {rn-sae}, %zmm12, %zmm12, %zmm9
> +        kmovw     %k0, %edx
> +
> +/* (x^2)_low */
> +        vmovaps   %zmm10, %zmm4
> +        vfmsub213ps {rn-sae}, %zmm0, %zmm10, %zmm4
> +
> +/* Yl = (x^2)_low + B_low */
> +        vaddps    {rn-sae}, %zmm4, %zmm3, %zmm5
> +
> +/* rel. error term: Eh=1-Sh*R0 */
> +        vmovaps   %zmm2, %zmm0
> +        vfnmadd231ps {rn-sae}, %zmm6, %zmm8, %zmm0
> +
> +/* Sl = (Yh*R0)_low+(R0*Yl) */
> +        vfmadd213ps {rn-sae}, %zmm7, %zmm8, %zmm5
> +
> +/* very large inputs ? */
> +        vmovups   Threshold+__svml_sasinh_data_internal_avx512(%rip), %zmm7
> +
> +/* rel. error term: Eh=(1-Sh*R0)-Sl*R0 */
> +        vfnmadd231ps {rn-sae}, %zmm5, %zmm8, %zmm0
> +
> +/* sqrt(1+x^2) ~ Sh + Sl + Sh*Eh*poly_s */
> +        vmovups   c2s+__svml_sasinh_data_internal_avx512(%rip), %zmm8
> +        vcmpps    $21, {sae}, %zmm7, %zmm12, %k1
> +
> +/* Sh*Eh */
> +        vmulps    {rn-sae}, %zmm0, %zmm6, %zmm4
> +        vfmadd231ps {rn-sae}, %zmm0, %zmm8, %zmm1
> +
> +/* Sl + Sh*Eh*poly_s */
> +        vfmadd213ps {rn-sae}, %zmm5, %zmm1, %zmm4
> +
> +/* Xh */
> +        vsubps    {rn-sae}, %zmm6, %zmm15, %zmm5
> +
> +/* fixup for very large inputs */
> +        vmovups   OneEighth+__svml_sasinh_data_internal_avx512(%rip), %zmm6
> +
> +/* Xin0+Sl+Sh*Eh*poly_s ~ x+sqrt(1+x^2) */
> +        vaddps    {rn-sae}, %zmm4, %zmm15, %zmm3
> +
> +/* Xl */
> +        vsubps    {rn-sae}, %zmm5, %zmm12, %zmm5
> +
> +/* Sl_high */
> +        vsubps    {rn-sae}, %zmm15, %zmm3, %zmm0
> +        vmulps    {rn-sae}, %zmm6, %zmm12, %zmm3{%k1}
> +
> +/* -K*L2H + Th */
> +        vmovups   L2H+__svml_sasinh_data_internal_avx512(%rip), %zmm15
> +
> +/* Sl_l */
> +        vsubps    {rn-sae}, %zmm0, %zmm4, %zmm1
> +        vrcp14ps  %zmm3, %zmm6
> +
> +/* Table lookups */
> +        vmovups   __svml_sasinh_data_internal_avx512(%rip), %zmm0
> +
> +/* Xin_low */
> +        vaddps    {rn-sae}, %zmm5, %zmm1, %zmm7
> +
> +/* round reciprocal to 1+4b mantissas */
> +        vpaddd    AddB5+__svml_sasinh_data_internal_avx512(%rip), %zmm6, %zmm4
> +        vmovups   poly_coeff1+__svml_sasinh_data_internal_avx512(%rip), %zmm5
> +        vandps    RcpBitMask+__svml_sasinh_data_internal_avx512(%rip), %zmm4, %zmm8
> +
> +/* fixup for very large inputs */
> +        vxorps    %zmm7, %zmm7, %zmm7{%k1}
> +
> +/* polynomial */
> +        vmovups   poly_coeff3+__svml_sasinh_data_internal_avx512(%rip), %zmm4
> +
> +/* reduced argument for log(): (Rcp*Xin-1)+Rcp*Xin_low */
> +        vfmsub231ps {rn-sae}, %zmm8, %zmm3, %zmm2
> +        vmovups   Four+__svml_sasinh_data_internal_avx512(%rip), %zmm3
> +
> +/* exponents */
> +        vgetexpps {sae}, %zmm8, %zmm1
> +
> +/* Prepare table index */
> +        vpsrld    $18, %zmm8, %zmm14
> +        vfmadd231ps {rn-sae}, %zmm8, %zmm7, %zmm2
> +        vmovups   poly_coeff2+__svml_sasinh_data_internal_avx512(%rip), %zmm7
> +        vsubps    {rn-sae}, %zmm3, %zmm1, %zmm1{%k1}
> +        vpermt2ps Log_tbl_H+64+__svml_sasinh_data_internal_avx512(%rip), %zmm14, %zmm0
> +        vmovups   Log_tbl_L+__svml_sasinh_data_internal_avx512(%rip), %zmm3
> +        vfmadd231ps {rn-sae}, %zmm2, %zmm4, %zmm7
> +        vfnmadd231ps {rn-sae}, %zmm1, %zmm15, %zmm0
> +
> +/* R^2 */
> +        vmulps    {rn-sae}, %zmm2, %zmm2, %zmm6
> +        vfmadd213ps {rn-sae}, %zmm5, %zmm2, %zmm7
> +        vpermt2ps Log_tbl_L+64+__svml_sasinh_data_internal_avx512(%rip), %zmm14, %zmm3
> +
> +/* -K*L2L + Tl */
> +        vmovups   L2L+__svml_sasinh_data_internal_avx512(%rip), %zmm14
> +        vfnmadd213ps {rn-sae}, %zmm3, %zmm14, %zmm1
> +
> +/* Tl + R^2*Poly */
> +        vfmadd213ps {rn-sae}, %zmm1, %zmm6, %zmm7
> +
> +/* R+Tl + R^2*Poly */
> +        vaddps    {rn-sae}, %zmm2, %zmm7, %zmm2
> +        vaddps    {rn-sae}, %zmm2, %zmm0, %zmm9{%k2}
> +        vxorps    %zmm13, %zmm9, %zmm0
> +        testl     %edx, %edx
> +
> +/* Go to special inputs processing branch */
> +        jne       L(SPECIAL_VALUES_BRANCH)
> +                                # LOE rbx r12 r13 r14 r15 edx zmm0 zmm10
> +
> +/* Restore registers
> + * and exit the function
> + */
> +
> +L(EXIT):
> +        movq      %rbp, %rsp
> +        popq      %rbp
> +        cfi_def_cfa(7, 8)
> +        cfi_restore(6)
> +        ret
> +        cfi_def_cfa(6, 16)
> +        cfi_offset(6, -16)
> +
> +/* Branch to process
> + * special inputs
> + */
> +
> +L(SPECIAL_VALUES_BRANCH):
> +        vmovups   %zmm10, 64(%rsp)
> +        vmovups   %zmm0, 128(%rsp)
> +                                # LOE rbx r12 r13 r14 r15 edx zmm0
> +
> +        xorl      %eax, %eax
> +                                # LOE rbx r12 r13 r14 r15 eax edx
> +
> +        vzeroupper
> +        movq      %r12, 16(%rsp)
> +        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -176; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x50, 0xff, 0xff, 0xff, 0x22
> +        movl      %eax, %r12d
> +        movq      %r13, 8(%rsp)
> +        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -184; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x48, 0xff, 0xff, 0xff, 0x22
> +        movl      %edx, %r13d
> +        movq      %r14, (%rsp)
> +        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -192; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x40, 0xff, 0xff, 0xff, 0x22
> +                                # LOE rbx r15 r12d r13d
> +
> +/* Range mask
> + * bits check
> + */
> +
> +L(RANGEMASK_CHECK):
> +        btl       %r12d, %r13d
> +
> +/* Call scalar math function */
> +        jc        L(SCALAR_MATH_CALL)
> +                                # LOE rbx r15 r12d r13d
> +
> +/* Special inputs
> + * processing loop
> + */
> +
> +L(SPECIAL_VALUES_LOOP):
> +        incl      %r12d
> +        cmpl      $16, %r12d
> +
> +/* Check bits in range mask */
> +        jl        L(RANGEMASK_CHECK)
> +                                # LOE rbx r15 r12d r13d
> +
> +        movq      16(%rsp), %r12
> +        cfi_restore(12)
> +        movq      8(%rsp), %r13
> +        cfi_restore(13)
> +        movq      (%rsp), %r14
> +        cfi_restore(14)
> +        vmovups   128(%rsp), %zmm0
> +
> +/* Go to exit */
> +        jmp       L(EXIT)
> +        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -176; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x50, 0xff, 0xff, 0xff, 0x22
> +        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -184; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x48, 0xff, 0xff, 0xff, 0x22
> +        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -192; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x40, 0xff, 0xff, 0xff, 0x22
> +                                # LOE rbx r12 r13 r14 r15 zmm0
> +
> +/* Scalar math fucntion call
> + * to process special input
> + */
> +
> +L(SCALAR_MATH_CALL):
> +        movl      %r12d, %r14d
> +        movss     64(%rsp,%r14,4), %xmm0
> +        call      asinhf@PLT
> +                                # LOE rbx r14 r15 r12d r13d xmm0
> +
> +        movss     %xmm0, 128(%rsp,%r14,4)
> +
> +/* Process special inputs in loop */
> +        jmp       L(SPECIAL_VALUES_LOOP)
> +                                # LOE rbx r15 r12d r13d
> +END(_ZGVeN16v_asinhf_skx)
> +
> +        .section .rodata, "a"
> +        .align 64
> +
> +#ifdef __svml_sasinh_data_internal_avx512_typedef
> +typedef unsigned int VUINT32;
> +typedef struct {
> +        __declspec(align(64)) VUINT32 Log_tbl_H[32][1];
> +        __declspec(align(64)) VUINT32 Log_tbl_L[32][1];
> +        __declspec(align(64)) VUINT32 One[16][1];
> +        __declspec(align(64)) VUINT32 AbsMask[16][1];
> +        __declspec(align(64)) VUINT32 SmallThreshold[16][1];
> +        __declspec(align(64)) VUINT32 Threshold[16][1];
> +        __declspec(align(64)) VUINT32 LargeThreshold[16][1];
> +        __declspec(align(64)) VUINT32 ca1[16][1];
> +        __declspec(align(64)) VUINT32 c2s[16][1];
> +        __declspec(align(64)) VUINT32 c1s[16][1];
> +        __declspec(align(64)) VUINT32 AddB5[16][1];
> +        __declspec(align(64)) VUINT32 RcpBitMask[16][1];
> +        __declspec(align(64)) VUINT32 OneEighth[16][1];
> +        __declspec(align(64)) VUINT32 Four[16][1];
> +        __declspec(align(64)) VUINT32 poly_coeff3[16][1];
> +        __declspec(align(64)) VUINT32 poly_coeff2[16][1];
> +        __declspec(align(64)) VUINT32 poly_coeff1[16][1];
> +        __declspec(align(64)) VUINT32 L2H[16][1];
> +        __declspec(align(64)) VUINT32 L2L[16][1];
> +    } __svml_sasinh_data_internal_avx512;
> +#endif
> +__svml_sasinh_data_internal_avx512:
> +        /*== Log_tbl_H ==*/
> +        .long 0x00000000
> +        .long 0xbcfc0000
> +        .long 0xbd788000
> +        .long 0xbdb78000
> +        .long 0xbdf14000
> +        .long 0xbe14a000
> +        .long 0xbe300000
> +        .long 0xbe4aa000
> +        .long 0xbe648000
> +        .long 0xbe7dc000
> +        .long 0xbe8b4000
> +        .long 0xbe974000
> +        .long 0xbea31000
> +        .long 0xbeae9000
> +        .long 0xbeb9d000
> +        .long 0xbec4d000
> +        .long 0xbecfa000
> +        .long 0xbeda2000
> +        .long 0xbee48000
> +        .long 0xbeeea000
> +        .long 0xbef89000
> +        .long 0xbf012800
> +        .long 0xbf05f000
> +        .long 0xbf0aa800
> +        .long 0xbf0f4000
> +        .long 0xbf13c800
> +        .long 0xbf184000
> +        .long 0xbf1ca000
> +        .long 0xbf20f000
> +        .long 0xbf252800
> +        .long 0xbf295000
> +        .long 0xbf2d6800
> +        /*== Log_tbl_L ==*/
> +        .align 64
> +        .long 0x80000000
> +        .long 0xb726c39e
> +        .long 0x3839e7fe
> +        .long 0xb7528ae5
> +        .long 0x377891d5
> +        .long 0xb8297c10
> +        .long 0x37cf8f58
> +        .long 0x3852b186
> +        .long 0x35838656
> +        .long 0xb80c36af
> +        .long 0x38235454
> +        .long 0xb862bae1
> +        .long 0x37e87bc7
> +        .long 0x37848150
> +        .long 0x37202511
> +        .long 0xb74e1b05
> +        .long 0x385c1340
> +        .long 0xb8777bcd
> +        .long 0x36038656
> +        .long 0xb7d40984
> +        .long 0xb80f5faf
> +        .long 0xb8254b4c
> +        .long 0xb865c84a
> +        .long 0x37f0b42d
> +        .long 0xb83ebce1
> +        .long 0xb83c2513
> +        .long 0x37a332c4
> +        .long 0x3779654f
> +        .long 0x38602f73
> +        .long 0x367449f8
> +        .long 0xb7b4996f
> +        .long 0xb800986b
> +        /*== One ==*/
> +        .align 64
> +        .long 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000
> +        /*== AbsMask ==*/
> +        .align 64
> +        .long 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff
> +        /*== SmallThreshold ==*/
> +        .align 64
> +        .long 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000
> +        /*== Threshold ==*/
> +        .align 64
> +        .long 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000
> +        /*== LargeThreshold ==*/
> +        .align 64
> +        .long 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff
> +        /*== ca1 ==*/
> +        .align 64
> +        .long 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE
> +        /*== c2s ==*/
> +        .align 64
> +        .long 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000
> +        /*== c1s ==*/
> +        .align 64
> +        .long 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000
> +        /*== AddB5 ==*/
> +        .align 64
> +        .long 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000
> +        /*== RcpBitMask ==*/
> +        .align 64
> +        .long 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000
> +        /*==OneEighth ==*/
> +        .align 64
> +        .long 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000
> +        /*== Four ==*/
> +        .align 64
> +        .long 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000
> +        /*== poly_coeff3 ==*/
> +        .align 64
> +        .long 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810
> +        /*== poly_coeff2 ==*/
> +        .align 64
> +        .long 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e
> +        /*== poly_coeff1 ==*/
> +        .align 64
> +        .long 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000
> +        /*== L2H = log(2)_high ==*/
> +        .align 64
> +        .long 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000
> +        /*== L2L = log(2)_low ==*/
> +        .align 64
> +        .long 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4
> +        .align 64
> +        .type	__svml_sasinh_data_internal_avx512,@object
> +        .size	__svml_sasinh_data_internal_avx512,.-__svml_sasinh_data_internal_avx512
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core-sse2.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core-sse2.S
> new file mode 100644
> index 0000000000..52e4d2f728
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core-sse2.S
> @@ -0,0 +1,20 @@
> +/* SSE2 version of vectorized asinhf, vector length is 4.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#define _ZGVbN4v_asinhf _ZGVbN4v_asinhf_sse2
> +#include "../svml_s_asinhf4_core.S"
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core.c b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core.c
> new file mode 100644
> index 0000000000..296d5754ae
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core.c
> @@ -0,0 +1,28 @@
> +/* Multiple versions of vectorized asinhf, vector length is 4.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#define SYMBOL_NAME _ZGVbN4v_asinhf
> +#include "ifunc-mathvec-sse4_1.h"
> +
> +libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
> +
> +#ifdef SHARED
> +__hidden_ver1 (_ZGVbN4v_asinhf, __GI__ZGVbN4v_asinhf,
> +	       __redirect__ZGVbN4v_asinhf)
> +  __attribute__ ((visibility ("hidden")));
> +#endif
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core_sse4.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core_sse4.S
> new file mode 100644
> index 0000000000..1eeeb4f5af
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core_sse4.S
> @@ -0,0 +1,509 @@
> +/* Function asinhf vectorized with SSE4.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   https://www.gnu.org/licenses/.  */
> +
> +/*
> + * ALGORITHM DESCRIPTION:
> + *
> + *   Compute asinh(x) as log(x + sqrt(x*x + 1))
> + *
> + *   Special cases:
> + *
> + *   asinh(NaN) = quiet NaN, and raise invalid exception
> + *   asinh(INF) = that INF
> + *   asinh(0)   = that 0
> + *
> + */
> +
> +/* Offsets for data table __svml_sasinh_data_internal
> + */
> +#define SgnMask                       	0
> +#define sOne                          	16
> +#define sPoly                         	32
> +#define iBrkValue                     	160
> +#define iOffExpoMask                  	176
> +#define sBigThreshold                 	192
> +#define sC2                           	208
> +#define sC3                           	224
> +#define sHalf                         	240
> +#define sLargestFinite                	256
> +#define sLittleThreshold              	272
> +#define sSign                         	288
> +#define sThirtyOne                    	304
> +#define sTopMask11                    	320
> +#define sTopMask8                     	336
> +#define XScale                        	352
> +#define sLn2                          	368
> +
> +#include <sysdep.h>
> +
> +        .text
> +	.section .text.sse4,"ax",@progbits
> +ENTRY(_ZGVbN4v_asinhf_sse4)
> +        subq      $72, %rsp
> +        cfi_def_cfa_offset(80)
> +        movaps    %xmm0, %xmm8
> +
> +/*
> + * Split X into high and low parts, XHi (<= 11 bits) and XLo (<= 13 bits)
> + * We could use either X or |X| here, but it doesn't seem to matter
> + */
> +        movups    sTopMask11+__svml_sasinh_data_internal(%rip), %xmm10
> +        movaps    %xmm8, %xmm2
> +        andps     %xmm8, %xmm10
> +
> +/*
> + * Compute X^2 = (XHi + XLo)^2 = XHi^2 + XLo * (X + XHi)
> + * The two parts are shifted off by around 11 bits. So even though
> + * the low bit will not in general be exact, it's near enough
> + */
> +        movaps    %xmm10, %xmm3
> +        subps     %xmm10, %xmm2
> +        mulps     %xmm10, %xmm3
> +        addps     %xmm8, %xmm10
> +
> +/* Load the constant 1 and a sign mask */
> +        movups    sOne+__svml_sasinh_data_internal(%rip), %xmm7
> +
> +/*
> + * Finally, express Y + W = X^2 + 1 accurately where Y has <= 8 bits.
> + * If |X| <= 1 then |XHi| <= 1 and so |X2Hi| <= 1, so we can treat 1
> + * as the dominant component in the compensated summation. Otherwise,
> + * if |X| >= 1, then since X2Hi only has 22 significant bits, the basic
> + * addition will be exact anyway until we get to |X| >= 2^24. But by
> + * that time the log function is well-conditioned enough that the
> + * rounding error doesn't matter. Hence we can treat 1 as dominant even
> + * if it literally isn't.
> + */
> +        movaps    %xmm7, %xmm11
> +        movaps    %xmm7, %xmm4
> +        movups    sTopMask8+__svml_sasinh_data_internal(%rip), %xmm12
> +        addps     %xmm3, %xmm11
> +        mulps     %xmm10, %xmm2
> +        subps     %xmm11, %xmm4
> +        movaps    %xmm12, %xmm0
> +        addps     %xmm3, %xmm4
> +
> +/*
> + * Unfortunately, we can still be in trouble if |X| <= 2^-5, since
> + * the absolute error 2^-(7+24)-ish in sqrt(1 + X^2) gets scaled up
> + * by 1/X and comes close to our threshold. Hence if |X| <= 2^-4,
> + * perform an alternative computation
> + * sqrt(1 + X^2) - 1 = X^2/2 - X^4/8 + X^6/16
> + * X2 = X^2
> + */
> +        addps     %xmm2, %xmm3
> +        addps     %xmm2, %xmm4
> +        andps     %xmm11, %xmm0
> +
> +/*
> + * Compute R = 1/sqrt(Y + W) * (1 + d)
> + * Force R to <= 8 significant bits.
> + * This means that R * Y and R^2 * Y are exactly representable.
> + */
> +        rsqrtps   %xmm0, %xmm14
> +        subps     %xmm0, %xmm11
> +        andps     %xmm12, %xmm14
> +        addps     %xmm11, %xmm4
> +
> +/*
> + * Compute S = (Y/sqrt(Y + W)) * (1 + d)
> + * and T = (W/sqrt(Y + W)) * (1 + d)
> + * so that S + T = sqrt(Y + W) * (1 + d)
> + * S is exact, and the rounding error in T is OK.
> + */
> +        mulps     %xmm14, %xmm0
> +        mulps     %xmm14, %xmm4
> +
> +/*
> + * Get the absolute value of the input, since we will exploit antisymmetry
> + * and mostly assume X >= 0 in the core computation
> + */
> +        movups    SgnMask+__svml_sasinh_data_internal(%rip), %xmm6
> +
> +/*
> + * Compute e = -(2 * d + d^2)
> + * The first FMR is exact, and the rounding error in the other is acceptable
> + * since d and e are ~ 2^-8
> + */
> +        movaps    %xmm14, %xmm13
> +        andps     %xmm8, %xmm6
> +
> +/*
> + * Obtain sqrt(1 + X^2) - 1 in two pieces
> + * sqrt(1 + X^2) - 1
> + * = sqrt(Y + W) - 1
> + * = (S + T) * (1 + Corr) - 1
> + * = [S - 1] + [T + (S + T) * Corr]
> + * We need a compensated summation for the last part. We treat S - 1
> + * as the larger part; it certainly is until about X < 2^-4, and in that
> + * case, the error is affordable since X dominates over sqrt(1 + X^2) - 1
> + * Final sum is dTmp5 (hi) + dTmp7 (lo)
> + */
> +        movaps    %xmm0, %xmm1
> +
> +/*
> + * Check whether the input is finite, by checking |X| <= MaxFloat
> + * Otherwise set the rangemask so that the callout will get used.
> + * Note that this will also use the callout for NaNs since not(NaN <= MaxFloat)
> + */
> +        movaps    %xmm6, %xmm9
> +
> +/*
> + * The following computation can go wrong for very large X, basically
> + * because X^2 overflows. But for large X we have
> + * asinh(X) / log(2 X) - 1 =~= 1/(4 * X^2), so for X >= 2^30
> + * we can just later stick X back into the log and tweak up the exponent.
> + * Actually we scale X by 2^-30 and tweak the exponent up by 31,
> + * to stay in the safe range for the later log computation.
> + * Compute a flag now telling us when do do this.
> + */
> +        movaps    %xmm6, %xmm5
> +        cmpnleps  sLargestFinite+__svml_sasinh_data_internal(%rip), %xmm9
> +        cmpltps   sBigThreshold+__svml_sasinh_data_internal(%rip), %xmm5
> +        mulps     %xmm0, %xmm13
> +        addps     %xmm4, %xmm1
> +        subps     %xmm7, %xmm0
> +        mulps     %xmm4, %xmm14
> +        movmskps  %xmm9, %edx
> +        movaps    %xmm7, %xmm9
> +
> +/*
> + * Now       1 / (1 + d)
> + * = 1 / (1 + (sqrt(1 - e) - 1))
> + * = 1 / sqrt(1 - e)
> + * = 1 + 1/2 * e + 3/8 * e^2 + 5/16 * e^3 + 35/128 * e^4 + ...
> + * So compute the first three nonconstant terms of that, so that
> + * we have a relative correction (1 + Corr) to apply to S etc.
> + * C1 = 1/2
> + * C2 = 3/8
> + * C3 = 5/16
> + */
> +        movups    sC3+__svml_sasinh_data_internal(%rip), %xmm15
> +        subps     %xmm13, %xmm9
> +        movups    sHalf+__svml_sasinh_data_internal(%rip), %xmm10
> +        subps     %xmm14, %xmm9
> +
> +/* sX2over2 = X^2/2 */
> +        mulps     %xmm10, %xmm3
> +        mulps     %xmm9, %xmm15
> +
> +/* sX46 = -X^4/4 + X^6/8 */
> +        movaps    %xmm3, %xmm2
> +        movaps    %xmm3, %xmm12
> +
> +/*
> + * Now do another compensated sum to add |X| + [sqrt(1 + X^2) - 1].
> + * It's always safe to assume |X| is larger.
> + * This is the final 2-part argument to the log1p function
> + */
> +        movaps    %xmm6, %xmm14
> +        addps     sC2+__svml_sasinh_data_internal(%rip), %xmm15
> +        mulps     %xmm9, %xmm15
> +        addps     %xmm10, %xmm15
> +        mulps     %xmm15, %xmm9
> +        mulps     %xmm1, %xmm9
> +
> +/* Now multiplex to the case X = 2^-30 * input, Xl = sL = 0 in the "big" case. */
> +        movups    XScale+__svml_sasinh_data_internal(%rip), %xmm15
> +        addps     %xmm9, %xmm4
> +        movaps    %xmm4, %xmm11
> +        addps     %xmm0, %xmm11
> +        subps     %xmm11, %xmm0
> +        addps     %xmm0, %xmm4
> +
> +/* sX4over4 = X^4/4 */
> +        movaps    %xmm3, %xmm0
> +        mulps     %xmm3, %xmm0
> +        mulps     %xmm0, %xmm2
> +        subps     %xmm0, %xmm2
> +
> +/*
> + * Now we feed into the log1p code, using H in place of _VARG1 and
> + * also adding L into Xl.
> + * compute 1+x as high, low parts
> + */
> +        movaps    %xmm7, %xmm0
> +
> +/* sX46over2 = -X^4/8 + x^6/16 */
> +        mulps     %xmm2, %xmm10
> +        movaps    %xmm7, %xmm2
> +        addps     %xmm10, %xmm12
> +        subps     %xmm12, %xmm3
> +        addps     %xmm3, %xmm10
> +
> +/* Now multiplex the two possible computations */
> +        movaps    %xmm6, %xmm3
> +        cmpleps   sLittleThreshold+__svml_sasinh_data_internal(%rip), %xmm3
> +        movaps    %xmm3, %xmm13
> +        andps     %xmm3, %xmm12
> +        andnps    %xmm11, %xmm13
> +        movaps    %xmm3, %xmm1
> +        orps      %xmm12, %xmm13
> +        andnps    %xmm4, %xmm1
> +        andps     %xmm3, %xmm10
> +        movaps    %xmm6, %xmm4
> +        orps      %xmm10, %xmm1
> +        addps     %xmm13, %xmm14
> +        mulps     %xmm15, %xmm6
> +        maxps     %xmm14, %xmm0
> +        minps     %xmm14, %xmm2
> +        subps     %xmm14, %xmm4
> +        movaps    %xmm0, %xmm3
> +        addps     %xmm4, %xmm13
> +        addps     %xmm2, %xmm3
> +        addps     %xmm13, %xmm1
> +        subps     %xmm3, %xmm0
> +        movaps    %xmm5, %xmm4
> +        andps     %xmm5, %xmm3
> +        andnps    %xmm6, %xmm4
> +        addps     %xmm0, %xmm2
> +
> +/*
> + * Now resume the main code.
> + * reduction: compute r,n
> + */
> +        movdqu    iBrkValue+__svml_sasinh_data_internal(%rip), %xmm6
> +        orps      %xmm3, %xmm4
> +        psubd     %xmm6, %xmm4
> +        movaps    %xmm7, %xmm0
> +        addps     %xmm2, %xmm1
> +        movdqu    iOffExpoMask+__svml_sasinh_data_internal(%rip), %xmm2
> +        pand      %xmm4, %xmm2
> +        psrad     $23, %xmm4
> +        cvtdq2ps  %xmm4, %xmm3
> +        pslld     $23, %xmm4
> +        andps     %xmm5, %xmm1
> +        paddd     %xmm6, %xmm2
> +        psubd     %xmm4, %xmm0
> +        mulps     %xmm0, %xmm1
> +
> +/* polynomial evaluation */
> +        subps     %xmm7, %xmm2
> +        movups    sPoly+112+__svml_sasinh_data_internal(%rip), %xmm7
> +        addps     %xmm2, %xmm1
> +        mulps     %xmm1, %xmm7
> +        movaps    %xmm5, %xmm2
> +
> +/* Add 31 to the exponent in the "large" case to get log(2 * input) */
> +        movups    sThirtyOne+__svml_sasinh_data_internal(%rip), %xmm0
> +        addps     sPoly+96+__svml_sasinh_data_internal(%rip), %xmm7
> +        addps     %xmm3, %xmm0
> +        mulps     %xmm1, %xmm7
> +        andnps    %xmm0, %xmm2
> +        andps     %xmm5, %xmm3
> +        orps      %xmm3, %xmm2
> +        addps     sPoly+80+__svml_sasinh_data_internal(%rip), %xmm7
> +
> +/* final reconstruction */
> +        mulps     sLn2+__svml_sasinh_data_internal(%rip), %xmm2
> +        mulps     %xmm1, %xmm7
> +
> +/* Finally, reincorporate the original sign. */
> +        movups    sSign+__svml_sasinh_data_internal(%rip), %xmm0
> +        andps     %xmm8, %xmm0
> +        addps     sPoly+64+__svml_sasinh_data_internal(%rip), %xmm7
> +        mulps     %xmm1, %xmm7
> +        addps     sPoly+48+__svml_sasinh_data_internal(%rip), %xmm7
> +        mulps     %xmm1, %xmm7
> +        addps     sPoly+32+__svml_sasinh_data_internal(%rip), %xmm7
> +        mulps     %xmm1, %xmm7
> +        addps     sPoly+16+__svml_sasinh_data_internal(%rip), %xmm7
> +        mulps     %xmm1, %xmm7
> +        addps     sPoly+__svml_sasinh_data_internal(%rip), %xmm7
> +        mulps     %xmm1, %xmm7
> +        mulps     %xmm1, %xmm7
> +        addps     %xmm7, %xmm1
> +        addps     %xmm2, %xmm1
> +        pxor      %xmm1, %xmm0
> +        testl     %edx, %edx
> +
> +/* Go to special inputs processing branch */
> +        jne       L(SPECIAL_VALUES_BRANCH)
> +                                # LOE rbx rbp r12 r13 r14 r15 edx xmm0 xmm8
> +
> +/* Restore registers
> + * and exit the function
> + */
> +
> +L(EXIT):
> +        addq      $72, %rsp
> +        cfi_def_cfa_offset(8)
> +        ret
> +        cfi_def_cfa_offset(80)
> +
> +/* Branch to process
> + * special inputs
> + */
> +
> +L(SPECIAL_VALUES_BRANCH):
> +        movups    %xmm8, 32(%rsp)
> +        movups    %xmm0, 48(%rsp)
> +                                # LOE rbx rbp r12 r13 r14 r15 edx
> +
> +        xorl      %eax, %eax
> +        movq      %r12, 16(%rsp)
> +        cfi_offset(12, -64)
> +        movl      %eax, %r12d
> +        movq      %r13, 8(%rsp)
> +        cfi_offset(13, -72)
> +        movl      %edx, %r13d
> +        movq      %r14, (%rsp)
> +        cfi_offset(14, -80)
> +                                # LOE rbx rbp r15 r12d r13d
> +
> +/* Range mask
> + * bits check
> + */
> +
> +L(RANGEMASK_CHECK):
> +        btl       %r12d, %r13d
> +
> +/* Call scalar math function */
> +        jc        L(SCALAR_MATH_CALL)
> +                                # LOE rbx rbp r15 r12d r13d
> +
> +/* Special inputs
> + * processing loop
> + */
> +
> +L(SPECIAL_VALUES_LOOP):
> +        incl      %r12d
> +        cmpl      $4, %r12d
> +
> +/* Check bits in range mask */
> +        jl        L(RANGEMASK_CHECK)
> +                                # LOE rbx rbp r15 r12d r13d
> +
> +        movq      16(%rsp), %r12
> +        cfi_restore(12)
> +        movq      8(%rsp), %r13
> +        cfi_restore(13)
> +        movq      (%rsp), %r14
> +        cfi_restore(14)
> +        movups    48(%rsp), %xmm0
> +
> +/* Go to exit */
> +        jmp       L(EXIT)
> +        cfi_offset(12, -64)
> +        cfi_offset(13, -72)
> +        cfi_offset(14, -80)
> +                                # LOE rbx rbp r12 r13 r14 r15 xmm0
> +
> +/* Scalar math fucntion call
> + * to process special input
> + */
> +
> +L(SCALAR_MATH_CALL):
> +        movl      %r12d, %r14d
> +        movss     32(%rsp,%r14,4), %xmm0
> +        call      asinhf@PLT
> +                                # LOE rbx rbp r14 r15 r12d r13d xmm0
> +
> +        movss     %xmm0, 48(%rsp,%r14,4)
> +
> +/* Process special inputs in loop */
> +        jmp       L(SPECIAL_VALUES_LOOP)
> +                                # LOE rbx rbp r15 r12d r13d
> +END(_ZGVbN4v_asinhf_sse4)
> +
> +        .section .rodata, "a"
> +        .align 16
> +
> +#ifdef __svml_sasinh_data_internal_typedef
> +typedef unsigned int VUINT32;
> +typedef struct {
> +        __declspec(align(16)) VUINT32 SgnMask[4][1];
> +        __declspec(align(16)) VUINT32 sOne[4][1];
> +        __declspec(align(16)) VUINT32 sPoly[8][4][1];
> +        __declspec(align(16)) VUINT32 iBrkValue[4][1];
> +        __declspec(align(16)) VUINT32 iOffExpoMask[4][1];
> +        __declspec(align(16)) VUINT32 sBigThreshold[4][1];
> +        __declspec(align(16)) VUINT32 sC2[4][1];
> +        __declspec(align(16)) VUINT32 sC3[4][1];
> +        __declspec(align(16)) VUINT32 sHalf[4][1];
> +        __declspec(align(16)) VUINT32 sLargestFinite[4][1];
> +        __declspec(align(16)) VUINT32 sLittleThreshold[4][1];
> +        __declspec(align(16)) VUINT32 sSign[4][1];
> +        __declspec(align(16)) VUINT32 sThirtyOne[4][1];
> +        __declspec(align(16)) VUINT32 sTopMask11[4][1];
> +        __declspec(align(16)) VUINT32 sTopMask8[4][1];
> +        __declspec(align(16)) VUINT32 XScale[4][1];
> +        __declspec(align(16)) VUINT32 sLn2[4][1];
> +} __svml_sasinh_data_internal;
> +#endif
> +__svml_sasinh_data_internal:
> +        /*== SgnMask ==*/
> +        .long 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff
> +        /*== sOne = SP 1.0 ==*/
> +        .align 16
> +        .long 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000
> +        /*== sPoly[] = SP polynomial ==*/
> +        .align 16
> +        .long 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000 /* -5.0000000000000000000000000e-01 P0 */
> +        .long 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94 /*  3.3333265781402587890625000e-01 P1 */
> +        .long 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e /* -2.5004237890243530273437500e-01 P2 */
> +        .long 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190 /*  2.0007920265197753906250000e-01 P3 */
> +        .long 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37 /* -1.6472326219081878662109375e-01 P4 */
> +        .long 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12 /*  1.4042308926582336425781250e-01 P5 */
> +        .long 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3 /* -1.5122179687023162841796875e-01 P6 */
> +        .long 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed /*  1.3820238411426544189453125e-01 P7 */
> +        /*== iBrkValue = SP 2/3 ==*/
> +        .align 16
> +        .long 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab
> +        /*== iOffExpoMask = SP significand mask ==*/
> +        .align 16
> +        .long 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff
> +        /*== sBigThreshold ==*/
> +        .align 16
> +        .long 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000
> +        /*== sC2 ==*/
> +        .align 16
> +        .long 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000
> +        /*== sC3 ==*/
> +        .align 16
> +        .long 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000
> +        /*== sHalf ==*/
> +        .align 16
> +        .long 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000
> +        /*== sLargestFinite ==*/
> +        .align 16
> +        .long 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF
> +        /*== sLittleThreshold ==*/
> +        .align 16
> +        .long 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000
> +        /*== sSign ==*/
> +        .align 16
> +        .long 0x80000000, 0x80000000, 0x80000000, 0x80000000
> +        /*== sThirtyOne ==*/
> +        .align 16
> +        .long 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000
> +        /*== sTopMask11 ==*/
> +        .align 16
> +        .long 0xFFFFE000, 0xFFFFE000, 0xFFFFE000, 0xFFFFE000
> +        /*== sTopMask8 ==*/
> +        .align 16
> +        .long 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000
> +        /*== XScale ==*/
> +        .align 16
> +        .long 0x30800000, 0x30800000, 0x30800000, 0x30800000
> +        /*== sLn2 = SP ln(2) ==*/
> +        .align 16
> +        .long 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218
> +        .align 16
> +        .type	__svml_sasinh_data_internal,@object
> +        .size	__svml_sasinh_data_internal,.-__svml_sasinh_data_internal
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core-sse.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core-sse.S
> new file mode 100644
> index 0000000000..1a0e113e94
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core-sse.S
> @@ -0,0 +1,20 @@
> +/* SSE version of vectorized asinhf, vector length is 8.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +    Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#define _ZGVdN8v_asinhf _ZGVdN8v_asinhf_sse_wrapper
> +#include "../svml_s_asinhf8_core.S"
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core.c b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core.c
> new file mode 100644
> index 0000000000..d97097a394
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core.c
> @@ -0,0 +1,28 @@
> +/* Multiple versions of vectorized asinhf, vector length is 8.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#define SYMBOL_NAME _ZGVdN8v_asinhf
> +#include "ifunc-mathvec-avx2.h"
> +
> +libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
> +
> +#ifdef SHARED
> +__hidden_ver1 (_ZGVdN8v_asinhf, __GI__ZGVdN8v_asinhf,
> +	       __redirect__ZGVdN8v_asinhf)
> +  __attribute__ ((visibility ("hidden")));
> +#endif
> diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core_avx2.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core_avx2.S
> new file mode 100644
> index 0000000000..a966f53773
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core_avx2.S
> @@ -0,0 +1,457 @@
> +/* Function asinhf vectorized with AVX2.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   https://www.gnu.org/licenses/.  */
> +
> +/*
> + * ALGORITHM DESCRIPTION:
> + *
> + *   Compute asinh(x) as log(x + sqrt(x*x + 1))
> + *
> + *   Special cases:
> + *
> + *   asinh(NaN) = quiet NaN, and raise invalid exception
> + *   asinh(INF) = that INF
> + *   asinh(0)   = that 0
> + *
> + */
> +
> +/* Offsets for data table __svml_sasinh_data_internal
> + */
> +#define SgnMask                       	0
> +#define sOne                          	32
> +#define sPoly                         	64
> +#define iBrkValue                     	320
> +#define iOffExpoMask                  	352
> +#define sBigThreshold                 	384
> +#define sC2                           	416
> +#define sC3                           	448
> +#define sHalf                         	480
> +#define sLargestFinite                	512
> +#define sLittleThreshold              	544
> +#define sSign                         	576
> +#define sThirtyOne                    	608
> +#define sTopMask8                     	640
> +#define XScale                        	672
> +#define sLn2                          	704
> +
> +#include <sysdep.h>
> +
> +        .text
> +	.section .text.avx2,"ax",@progbits
> +ENTRY(_ZGVdN8v_asinhf_avx2)
> +        pushq     %rbp
> +        cfi_def_cfa_offset(16)
> +        movq      %rsp, %rbp
> +        cfi_def_cfa(6, 16)
> +        cfi_offset(6, -16)
> +        andq      $-32, %rsp
> +        subq      $96, %rsp
> +        vmovaps   %ymm0, %ymm9
> +
> +/* Load the constant 1 and a sign mask */
> +        vmovups   sOne+__svml_sasinh_data_internal(%rip), %ymm8
> +
> +/* No need to split X when FMA is available in hardware. */
> +        vmulps    %ymm9, %ymm9, %ymm5
> +        vmovups   sTopMask8+__svml_sasinh_data_internal(%rip), %ymm1
> +
> +/*
> + * Finally, express Y + W = X^2 + 1 accurately where Y has <= 8 bits.
> + * If |X| <= 1 then |XHi| <= 1 and so |X2Hi| <= 1, so we can treat 1
> + * as the dominant component in the compensated summation. Otherwise,
> + * if |X| >= 1, then since X2Hi only has 22 significant bits, the basic
> + * addition will be exact anyway until we get to |X| >= 2^24. But by
> + * that time the log function is well-conditioned enough that the
> + * rounding error doesn't matter. Hence we can treat 1 as dominant even
> + * if it literally isn't.
> + */
> +        vaddps    %ymm5, %ymm8, %ymm13
> +        vandps    %ymm1, %ymm13, %ymm2
> +        vmovaps   %ymm9, %ymm4
> +        vsubps    %ymm13, %ymm8, %ymm11
> +        vsubps    %ymm2, %ymm13, %ymm15
> +
> +/*
> + * Compute R = 1/sqrt(Y + W) * (1 + d)
> + * Force R to <= 8 significant bits.
> + * This means that R * Y and R^2 * Y are exactly representable.
> + */
> +        vrsqrtps  %ymm2, %ymm0
> +        vfmsub213ps %ymm5, %ymm9, %ymm4
> +        vaddps    %ymm11, %ymm5, %ymm12
> +
> +/*
> + * Get the absolute value of the input, since we will exploit antisymmetry
> + * and mostly assume X >= 0 in the core computation
> + */
> +        vandps    SgnMask+__svml_sasinh_data_internal(%rip), %ymm9, %ymm6
> +
> +/*
> + * Check whether the input is finite, by checking |X| <= MaxFloat
> + * Otherwise set the rangemask so that the callout will get used.
> + * Note that this will also use the callout for NaNs since not(NaN <= MaxFloat)
> + */
> +        vcmpnle_uqps sLargestFinite+__svml_sasinh_data_internal(%rip), %ymm6, %ymm10
> +        vaddps    %ymm12, %ymm4, %ymm14
> +
> +/*
> + * Unfortunately, we can still be in trouble if |X| <= 2^-5, since
> + * the absolute error 2^-(7+24)-ish in sqrt(1 + X^2) gets scaled up
> + * by 1/X and comes close to our threshold. Hence if |X| <= 2^-4,
> + * perform an alternative computation
> + * sqrt(1 + X^2) - 1 = X^2/2 - X^4/8 + X^6/16
> + * X2 = X^2
> + */
> +        vaddps    %ymm4, %ymm5, %ymm4
> +
> +/*
> + * The following computation can go wrong for very large X, basically
> + * because X^2 overflows. But for large X we have
> + * asinh(X) / log(2 X) - 1 =~= 1/(4 * X^2), so for X >= 2^30
> + * we can just later stick X back into the log and tweak up the exponent.
> + * Actually we scale X by 2^-30 and tweak the exponent up by 31,
> + * to stay in the safe range for the later log computation.
> + * Compute a flag now telling us when do do this.
> + */
> +        vcmplt_oqps sBigThreshold+__svml_sasinh_data_internal(%rip), %ymm6, %ymm7
> +        vaddps    %ymm15, %ymm14, %ymm3
> +
> +/*
> + * Now       1 / (1 + d)
> + * = 1 / (1 + (sqrt(1 - e) - 1))
> + * = 1 / sqrt(1 - e)
> + * = 1 + 1/2 * e + 3/8 * e^2 + 5/16 * e^3 + 35/128 * e^4 + ...
> + * So compute the first three nonconstant terms of that, so that
> + * we have a relative correction (1 + Corr) to apply to S etc.
> + * C1 = 1/2
> + * C2 = 3/8
> + * C3 = 5/16
> + */
> +        vmovups   sC3+__svml_sasinh_data_internal(%rip), %ymm12
> +        vmovmskps %ymm10, %edx
> +        vandps    %ymm1, %ymm0, %ymm10
> +
> +/*
> + * Compute S = (Y/sqrt(Y + W)) * (1 + d)
> + * and T = (W/sqrt(Y + W)) * (1 + d)
> + * so that S + T = sqrt(Y + W) * (1 + d)
> + * S is exact, and the rounding error in T is OK.
> + */
> +        vmulps    %ymm10, %ymm2, %ymm15
> +        vmulps    %ymm3, %ymm10, %ymm14
> +        vmovups   sHalf+__svml_sasinh_data_internal(%rip), %ymm3
> +        vsubps    %ymm8, %ymm15, %ymm0
> +
> +/*
> + * Obtain sqrt(1 + X^2) - 1 in two pieces
> + * sqrt(1 + X^2) - 1
> + * = sqrt(Y + W) - 1
> + * = (S + T) * (1 + Corr) - 1
> + * = [S - 1] + [T + (S + T) * Corr]
> + * We need a compensated summation for the last part. We treat S - 1
> + * as the larger part; it certainly is until about X < 2^-4, and in that
> + * case, the error is affordable since X dominates over sqrt(1 + X^2) - 1
> + * Final sum is dTmp5 (hi) + dTmp7 (lo)
> + */
> +        vaddps    %ymm14, %ymm15, %ymm13
> +
> +/*
> + * Compute e = -(2 * d + d^2)
> + * The first FMR is exact, and the rounding error in the other is acceptable
> + * since d and e are ~ 2^-8
> + */
> +        vmovaps   %ymm8, %ymm11
> +        vfnmadd231ps %ymm15, %ymm10, %ymm11
> +        vfnmadd231ps %ymm14, %ymm10, %ymm11
> +        vfmadd213ps sC2+__svml_sasinh_data_internal(%rip), %ymm11, %ymm12
> +        vfmadd213ps %ymm3, %ymm11, %ymm12
> +        vmulps    %ymm12, %ymm11, %ymm1
> +
> +/* Now multiplex the two possible computations */
> +        vcmple_oqps sLittleThreshold+__svml_sasinh_data_internal(%rip), %ymm6, %ymm11
> +        vfmadd213ps %ymm14, %ymm13, %ymm1
> +        vaddps    %ymm0, %ymm1, %ymm2
> +        vsubps    %ymm2, %ymm0, %ymm10
> +
> +/* sX2over2 = X^2/2 */
> +        vmulps    %ymm4, %ymm3, %ymm0
> +        vaddps    %ymm10, %ymm1, %ymm1
> +
> +/* sX4over4 = X^4/4 */
> +        vmulps    %ymm0, %ymm0, %ymm5
> +
> +/* sX46 = -X^4/4 + X^6/8 */
> +        vfmsub231ps %ymm0, %ymm5, %ymm5
> +
> +/* sX46over2 = -X^4/8 + x^6/16 */
> +        vmulps    %ymm5, %ymm3, %ymm3
> +        vaddps    %ymm3, %ymm0, %ymm5
> +        vblendvps %ymm11, %ymm5, %ymm2, %ymm2
> +        vsubps    %ymm5, %ymm0, %ymm4
> +
> +/*
> + * Now do another compensated sum to add |X| + [sqrt(1 + X^2) - 1].
> + * It's always safe to assume |X| is larger.
> + * This is the final 2-part argument to the log1p function
> + */
> +        vaddps    %ymm2, %ymm6, %ymm14
> +
> +/*
> + * Now resume the main code.
> + * reduction: compute r,n
> + */
> +        vmovups   iBrkValue+__svml_sasinh_data_internal(%rip), %ymm5
> +        vaddps    %ymm4, %ymm3, %ymm10
> +
> +/*
> + * Now we feed into the log1p code, using H in place of _VARG1 and
> + * also adding L into Xl.
> + * compute 1+x as high, low parts
> + */
> +        vmaxps    %ymm14, %ymm8, %ymm15
> +        vminps    %ymm14, %ymm8, %ymm0
> +        vblendvps %ymm11, %ymm10, %ymm1, %ymm12
> +        vsubps    %ymm14, %ymm6, %ymm1
> +        vaddps    %ymm0, %ymm15, %ymm3
> +
> +/* Now multiplex to the case X = 2^-30 * input, Xl = sL = 0 in the "big" case. */
> +        vmulps    XScale+__svml_sasinh_data_internal(%rip), %ymm6, %ymm6
> +        vaddps    %ymm1, %ymm2, %ymm13
> +        vsubps    %ymm3, %ymm15, %ymm15
> +        vaddps    %ymm13, %ymm12, %ymm1
> +        vaddps    %ymm15, %ymm0, %ymm2
> +        vblendvps %ymm7, %ymm3, %ymm6, %ymm0
> +        vaddps    %ymm2, %ymm1, %ymm4
> +        vpsubd    %ymm5, %ymm0, %ymm1
> +        vpsrad    $23, %ymm1, %ymm6
> +        vpand     iOffExpoMask+__svml_sasinh_data_internal(%rip), %ymm1, %ymm2
> +        vmovups   sPoly+224+__svml_sasinh_data_internal(%rip), %ymm1
> +        vpslld    $23, %ymm6, %ymm10
> +        vpaddd    %ymm5, %ymm2, %ymm13
> +        vcvtdq2ps %ymm6, %ymm0
> +        vpsubd    %ymm10, %ymm8, %ymm12
> +
> +/* polynomial evaluation */
> +        vsubps    %ymm8, %ymm13, %ymm8
> +
> +/* Add 31 to the exponent in the "large" case to get log(2 * input) */
> +        vaddps    sThirtyOne+__svml_sasinh_data_internal(%rip), %ymm0, %ymm3
> +        vandps    %ymm7, %ymm4, %ymm11
> +        vmulps    %ymm12, %ymm11, %ymm14
> +        vblendvps %ymm7, %ymm0, %ymm3, %ymm0
> +        vaddps    %ymm8, %ymm14, %ymm2
> +        vfmadd213ps sPoly+192+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
> +        vfmadd213ps sPoly+160+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
> +        vfmadd213ps sPoly+128+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
> +        vfmadd213ps sPoly+96+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
> +        vfmadd213ps sPoly+64+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
> +        vfmadd213ps sPoly+32+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
> +        vfmadd213ps sPoly+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
> +        vmulps    %ymm1, %ymm2, %ymm4
> +        vfmadd213ps %ymm2, %ymm2, %ymm4
> +
> +/* final reconstruction */
> +        vfmadd132ps sLn2+__svml_sasinh_data_internal(%rip), %ymm4, %ymm0
> +
> +/* Finally, reincorporate the original sign. */
> +        vandps    sSign+__svml_sasinh_data_internal(%rip), %ymm9, %ymm7
> +        vxorps    %ymm0, %ymm7, %ymm0
> +        testl     %edx, %edx
> +
> +/* Go to special inputs processing branch */
> +        jne       L(SPECIAL_VALUES_BRANCH)
> +                                # LOE rbx r12 r13 r14 r15 edx ymm0 ymm9
> +
> +/* Restore registers
> + * and exit the function
> + */
> +
> +L(EXIT):
> +        movq      %rbp, %rsp
> +        popq      %rbp
> +        cfi_def_cfa(7, 8)
> +        cfi_restore(6)
> +        ret
> +        cfi_def_cfa(6, 16)
> +        cfi_offset(6, -16)
> +
> +/* Branch to process
> + * special inputs
> + */
> +
> +L(SPECIAL_VALUES_BRANCH):
> +        vmovups   %ymm9, 32(%rsp)
> +        vmovups   %ymm0, 64(%rsp)
> +                                # LOE rbx r12 r13 r14 r15 edx ymm0
> +
> +        xorl      %eax, %eax
> +                                # LOE rbx r12 r13 r14 r15 eax edx
> +
> +        vzeroupper
> +        movq      %r12, 16(%rsp)
> +        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -80; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xb0, 0xff, 0xff, 0xff, 0x22
> +        movl      %eax, %r12d
> +        movq      %r13, 8(%rsp)
> +        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -88; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa8, 0xff, 0xff, 0xff, 0x22
> +        movl      %edx, %r13d
> +        movq      %r14, (%rsp)
> +        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -96; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa0, 0xff, 0xff, 0xff, 0x22
> +                                # LOE rbx r15 r12d r13d
> +
> +/* Range mask
> + * bits check
> + */
> +
> +L(RANGEMASK_CHECK):
> +        btl       %r12d, %r13d
> +
> +/* Call scalar math function */
> +        jc        L(SCALAR_MATH_CALL)
> +                                # LOE rbx r15 r12d r13d
> +
> +/* Special inputs
> + * processing loop
> + */
> +
> +L(SPECIAL_VALUES_LOOP):
> +        incl      %r12d
> +        cmpl      $8, %r12d
> +
> +/* Check bits in range mask */
> +        jl        L(RANGEMASK_CHECK)
> +                                # LOE rbx r15 r12d r13d
> +
> +        movq      16(%rsp), %r12
> +        cfi_restore(12)
> +        movq      8(%rsp), %r13
> +        cfi_restore(13)
> +        movq      (%rsp), %r14
> +        cfi_restore(14)
> +        vmovups   64(%rsp), %ymm0
> +
> +/* Go to exit */
> +        jmp       L(EXIT)
> +        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -80; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xb0, 0xff, 0xff, 0xff, 0x22
> +        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -88; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa8, 0xff, 0xff, 0xff, 0x22
> +        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -96; DW_OP_plus)  */
> +        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa0, 0xff, 0xff, 0xff, 0x22
> +                                # LOE rbx r12 r13 r14 r15 ymm0
> +
> +/* Scalar math fucntion call
> + * to process special input
> + */
> +
> +L(SCALAR_MATH_CALL):
> +        movl      %r12d, %r14d
> +        movss     32(%rsp,%r14,4), %xmm0
> +        call      asinhf@PLT
> +                                # LOE rbx r14 r15 r12d r13d xmm0
> +
> +        movss     %xmm0, 64(%rsp,%r14,4)
> +
> +/* Process special inputs in loop */
> +        jmp       L(SPECIAL_VALUES_LOOP)
> +                                # LOE rbx r15 r12d r13d
> +END(_ZGVdN8v_asinhf_avx2)
> +
> +        .section .rodata, "a"
> +        .align 32
> +
> +#ifdef __svml_sasinh_data_internal_typedef
> +typedef unsigned int VUINT32;
> +typedef struct {
> +        __declspec(align(32)) VUINT32 SgnMask[8][1];
> +        __declspec(align(32)) VUINT32 sOne[8][1];
> +        __declspec(align(32)) VUINT32 sPoly[8][8][1];
> +        __declspec(align(32)) VUINT32 iBrkValue[8][1];
> +        __declspec(align(32)) VUINT32 iOffExpoMask[8][1];
> +        __declspec(align(32)) VUINT32 sBigThreshold[8][1];
> +        __declspec(align(32)) VUINT32 sC2[8][1];
> +        __declspec(align(32)) VUINT32 sC3[8][1];
> +        __declspec(align(32)) VUINT32 sHalf[8][1];
> +        __declspec(align(32)) VUINT32 sLargestFinite[8][1];
> +        __declspec(align(32)) VUINT32 sLittleThreshold[8][1];
> +        __declspec(align(32)) VUINT32 sSign[8][1];
> +        __declspec(align(32)) VUINT32 sThirtyOne[8][1];
> +        __declspec(align(32)) VUINT32 sTopMask8[8][1];
> +        __declspec(align(32)) VUINT32 XScale[8][1];
> +        __declspec(align(32)) VUINT32 sLn2[8][1];
> +} __svml_sasinh_data_internal;
> +#endif
> +__svml_sasinh_data_internal:
> +        /*== SgnMask ==*/
> +        .long 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff
> +        /*== sOne = SP 1.0 ==*/
> +        .align 32
> +        .long 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000
> +        /*== sPoly[] = SP polynomial ==*/
> +        .align 32
> +        .long 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000 /* -5.0000000000000000000000000e-01 P0 */
> +        .long 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94 /*  3.3333265781402587890625000e-01 P1 */
> +        .long 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e /* -2.5004237890243530273437500e-01 P2 */
> +        .long 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190 /*  2.0007920265197753906250000e-01 P3 */
> +        .long 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37 /* -1.6472326219081878662109375e-01 P4 */
> +        .long 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12 /*  1.4042308926582336425781250e-01 P5 */
> +        .long 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3 /* -1.5122179687023162841796875e-01 P6 */
> +        .long 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed /*  1.3820238411426544189453125e-01 P7 */
> +        /*== iBrkValue = SP 2/3 ==*/
> +        .align 32
> +        .long 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab
> +        /*== iOffExpoMask = SP significand mask ==*/
> +        .align 32
> +        .long 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff
> +        /*== sBigThreshold ==*/
> +        .align 32
> +        .long 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000
> +        /*== sC2 ==*/
> +        .align 32
> +        .long 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000
> +        /*== sC3 ==*/
> +        .align 32
> +        .long 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000
> +        /*== sHalf ==*/
> +        .align 32
> +        .long 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000
> +        /*== sLargestFinite ==*/
> +        .align 32
> +        .long 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF
> +        /*== sLittleThreshold ==*/
> +        .align 32
> +        .long 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000
> +        /*== sSign ==*/
> +        .align 32
> +        .long 0x80000000, 0x80000000, 0x80000000, 0x80000000, 0x80000000, 0x80000000, 0x80000000, 0x80000000
> +        /*== sThirtyOne ==*/
> +        .align 32
> +        .long 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000
> +        /*== sTopMask8 ==*/
> +        .align 32
> +        .long 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000
> +        /*== XScale ==*/
> +        .align 32
> +        .long 0x30800000, 0x30800000, 0x30800000, 0x30800000, 0x30800000, 0x30800000, 0x30800000, 0x30800000
> +        /*== sLn2 = SP ln(2) ==*/
> +        .align 32
> +        .long 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218
> +        .align 32
> +        .type	__svml_sasinh_data_internal,@object
> +        .size	__svml_sasinh_data_internal,.-__svml_sasinh_data_internal
> diff --git a/sysdeps/x86_64/fpu/svml_d_asinh2_core.S b/sysdeps/x86_64/fpu/svml_d_asinh2_core.S
> new file mode 100644
> index 0000000000..60e372238a
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/svml_d_asinh2_core.S
> @@ -0,0 +1,29 @@
> +/* Function asinh vectorized with SSE2.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#include <sysdep.h>
> +#include "svml_d_wrapper_impl.h"
> +
> +	.text
> +ENTRY (_ZGVbN2v_asinh)
> +WRAPPER_IMPL_SSE2 asinh
> +END (_ZGVbN2v_asinh)
> +
> +#ifndef USE_MULTIARCH
> + libmvec_hidden_def (_ZGVbN2v_asinh)
> +#endif
> diff --git a/sysdeps/x86_64/fpu/svml_d_asinh4_core.S b/sysdeps/x86_64/fpu/svml_d_asinh4_core.S
> new file mode 100644
> index 0000000000..c7350011e1
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/svml_d_asinh4_core.S
> @@ -0,0 +1,29 @@
> +/* Function asinh vectorized with AVX2, wrapper version.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#include <sysdep.h>
> +#include "svml_d_wrapper_impl.h"
> +
> +	.text
> +ENTRY (_ZGVdN4v_asinh)
> +WRAPPER_IMPL_AVX _ZGVbN2v_asinh
> +END (_ZGVdN4v_asinh)
> +
> +#ifndef USE_MULTIARCH
> + libmvec_hidden_def (_ZGVdN4v_asinh)
> +#endif
> diff --git a/sysdeps/x86_64/fpu/svml_d_asinh4_core_avx.S b/sysdeps/x86_64/fpu/svml_d_asinh4_core_avx.S
> new file mode 100644
> index 0000000000..83aaa8c3f1
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/svml_d_asinh4_core_avx.S
> @@ -0,0 +1,25 @@
> +/* Function asinh vectorized in AVX ISA as wrapper to SSE4 ISA version.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#include <sysdep.h>
> +#include "svml_d_wrapper_impl.h"
> +
> +	.text
> +ENTRY (_ZGVcN4v_asinh)
> +WRAPPER_IMPL_AVX _ZGVbN2v_asinh
> +END (_ZGVcN4v_asinh)
> diff --git a/sysdeps/x86_64/fpu/svml_d_asinh8_core.S b/sysdeps/x86_64/fpu/svml_d_asinh8_core.S
> new file mode 100644
> index 0000000000..9597975ff6
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/svml_d_asinh8_core.S
> @@ -0,0 +1,25 @@
> +/* Function asinh vectorized with AVX-512, wrapper to AVX2.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#include <sysdep.h>
> +#include "svml_d_wrapper_impl.h"
> +
> +	.text
> +ENTRY (_ZGVeN8v_asinh)
> +WRAPPER_IMPL_AVX512 _ZGVdN4v_asinh
> +END (_ZGVeN8v_asinh)
> diff --git a/sysdeps/x86_64/fpu/svml_s_asinhf16_core.S b/sysdeps/x86_64/fpu/svml_s_asinhf16_core.S
> new file mode 100644
> index 0000000000..5b3d405f2e
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/svml_s_asinhf16_core.S
> @@ -0,0 +1,25 @@
> +/* Function asinhf vectorized with AVX-512. Wrapper to AVX2 version.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#include <sysdep.h>
> +#include "svml_s_wrapper_impl.h"
> +
> +	.text
> +ENTRY (_ZGVeN16v_asinhf)
> +WRAPPER_IMPL_AVX512 _ZGVdN8v_asinhf
> +END (_ZGVeN16v_asinhf)
> diff --git a/sysdeps/x86_64/fpu/svml_s_asinhf4_core.S b/sysdeps/x86_64/fpu/svml_s_asinhf4_core.S
> new file mode 100644
> index 0000000000..af44fa5108
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/svml_s_asinhf4_core.S
> @@ -0,0 +1,29 @@
> +/* Function asinhf vectorized with SSE2, wrapper version.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#include <sysdep.h>
> +#include "svml_s_wrapper_impl.h"
> +
> +	.text
> +ENTRY (_ZGVbN4v_asinhf)
> +WRAPPER_IMPL_SSE2 asinhf
> +END (_ZGVbN4v_asinhf)
> +
> +#ifndef USE_MULTIARCH
> + libmvec_hidden_def (_ZGVbN4v_asinhf)
> +#endif
> diff --git a/sysdeps/x86_64/fpu/svml_s_asinhf8_core.S b/sysdeps/x86_64/fpu/svml_s_asinhf8_core.S
> new file mode 100644
> index 0000000000..3bd06d8032
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/svml_s_asinhf8_core.S
> @@ -0,0 +1,29 @@
> +/* Function asinhf vectorized with AVX2, wrapper version.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#include <sysdep.h>
> +#include "svml_s_wrapper_impl.h"
> +
> +	.text
> +ENTRY (_ZGVdN8v_asinhf)
> +WRAPPER_IMPL_AVX _ZGVbN4v_asinhf
> +END (_ZGVdN8v_asinhf)
> +
> +#ifndef USE_MULTIARCH
> + libmvec_hidden_def (_ZGVdN8v_asinhf)
> +#endif
> diff --git a/sysdeps/x86_64/fpu/svml_s_asinhf8_core_avx.S b/sysdeps/x86_64/fpu/svml_s_asinhf8_core_avx.S
> new file mode 100644
> index 0000000000..f79616c0bd
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/svml_s_asinhf8_core_avx.S
> @@ -0,0 +1,25 @@
> +/* Function asinhf vectorized in AVX ISA as wrapper to SSE4 ISA version.
> +   Copyright (C) 2021 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +#include <sysdep.h>
> +#include "svml_s_wrapper_impl.h"
> +
> +        .text
> +ENTRY (_ZGVcN8v_asinhf)
> +WRAPPER_IMPL_AVX _ZGVbN4v_asinhf
> +END (_ZGVcN8v_asinhf)
> diff --git a/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx.c b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx.c
> new file mode 100644
> index 0000000000..da03528700
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx.c
> @@ -0,0 +1 @@
> +#include "test-double-libmvec-asinh.c"
> diff --git a/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx2.c b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx2.c
> new file mode 100644
> index 0000000000..da03528700
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx2.c
> @@ -0,0 +1 @@
> +#include "test-double-libmvec-asinh.c"
> diff --git a/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx512f.c b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx512f.c
> new file mode 100644
> index 0000000000..da03528700
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx512f.c
> @@ -0,0 +1 @@
> +#include "test-double-libmvec-asinh.c"
> diff --git a/sysdeps/x86_64/fpu/test-double-libmvec-asinh.c b/sysdeps/x86_64/fpu/test-double-libmvec-asinh.c
> new file mode 100644
> index 0000000000..71e6b9f578
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/test-double-libmvec-asinh.c
> @@ -0,0 +1,3 @@
> +#define LIBMVEC_TYPE double
> +#define LIBMVEC_FUNC asinh
> +#include "test-vector-abi-arg1.h"
> diff --git a/sysdeps/x86_64/fpu/test-double-vlen2-wrappers.c b/sysdeps/x86_64/fpu/test-double-vlen2-wrappers.c
> index f53bb6813e..76114772ba 100644
> --- a/sysdeps/x86_64/fpu/test-double-vlen2-wrappers.c
> +++ b/sysdeps/x86_64/fpu/test-double-vlen2-wrappers.c
> @@ -45,6 +45,7 @@ VECTOR_WRAPPER (WRAPPER_NAME (atanh), _ZGVbN2v_atanh)
>  VECTOR_WRAPPER (WRAPPER_NAME (acosh), _ZGVbN2v_acosh)
>  VECTOR_WRAPPER (WRAPPER_NAME (erf), _ZGVbN2v_erf)
>  VECTOR_WRAPPER (WRAPPER_NAME (tanh), _ZGVbN2v_tanh)
> +VECTOR_WRAPPER (WRAPPER_NAME (asinh), _ZGVbN2v_asinh)
>  
>  #define VEC_INT_TYPE __m128i
>  
> diff --git a/sysdeps/x86_64/fpu/test-double-vlen4-avx2-wrappers.c b/sysdeps/x86_64/fpu/test-double-vlen4-avx2-wrappers.c
> index 0452c3db38..1e0ee34975 100644
> --- a/sysdeps/x86_64/fpu/test-double-vlen4-avx2-wrappers.c
> +++ b/sysdeps/x86_64/fpu/test-double-vlen4-avx2-wrappers.c
> @@ -48,6 +48,7 @@ VECTOR_WRAPPER (WRAPPER_NAME (atanh), _ZGVdN4v_atanh)
>  VECTOR_WRAPPER (WRAPPER_NAME (acosh), _ZGVdN4v_acosh)
>  VECTOR_WRAPPER (WRAPPER_NAME (erf), _ZGVdN4v_erf)
>  VECTOR_WRAPPER (WRAPPER_NAME (tanh), _ZGVdN4v_tanh)
> +VECTOR_WRAPPER (WRAPPER_NAME (asinh), _ZGVdN4v_asinh)
>  
>  #ifndef __ILP32__
>  # define VEC_INT_TYPE __m256i
> diff --git a/sysdeps/x86_64/fpu/test-double-vlen4-wrappers.c b/sysdeps/x86_64/fpu/test-double-vlen4-wrappers.c
> index 197d5afc88..17c43a75d1 100644
> --- a/sysdeps/x86_64/fpu/test-double-vlen4-wrappers.c
> +++ b/sysdeps/x86_64/fpu/test-double-vlen4-wrappers.c
> @@ -45,6 +45,7 @@ VECTOR_WRAPPER (WRAPPER_NAME (atanh), _ZGVcN4v_atanh)
>  VECTOR_WRAPPER (WRAPPER_NAME (acosh), _ZGVcN4v_acosh)
>  VECTOR_WRAPPER (WRAPPER_NAME (erf), _ZGVcN4v_erf)
>  VECTOR_WRAPPER (WRAPPER_NAME (tanh), _ZGVcN4v_tanh)
> +VECTOR_WRAPPER (WRAPPER_NAME (asinh), _ZGVcN4v_asinh)
>  
>  #define VEC_INT_TYPE __m128i
>  
> diff --git a/sysdeps/x86_64/fpu/test-double-vlen8-wrappers.c b/sysdeps/x86_64/fpu/test-double-vlen8-wrappers.c
> index e56ece640c..1c6809e6e3 100644
> --- a/sysdeps/x86_64/fpu/test-double-vlen8-wrappers.c
> +++ b/sysdeps/x86_64/fpu/test-double-vlen8-wrappers.c
> @@ -45,6 +45,7 @@ VECTOR_WRAPPER (WRAPPER_NAME (atanh), _ZGVeN8v_atanh)
>  VECTOR_WRAPPER (WRAPPER_NAME (acosh), _ZGVeN8v_acosh)
>  VECTOR_WRAPPER (WRAPPER_NAME (erf), _ZGVeN8v_erf)
>  VECTOR_WRAPPER (WRAPPER_NAME (tanh), _ZGVeN8v_tanh)
> +VECTOR_WRAPPER (WRAPPER_NAME (asinh), _ZGVeN8v_asinh)
>  
>  #ifndef __ILP32__
>  # define VEC_INT_TYPE __m512i
> diff --git a/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx.c b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx.c
> new file mode 100644
> index 0000000000..77e1838bb4
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx.c
> @@ -0,0 +1 @@
> +#include "test-float-libmvec-asinhf.c"
> diff --git a/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx2.c b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx2.c
> new file mode 100644
> index 0000000000..77e1838bb4
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx2.c
> @@ -0,0 +1 @@
> +#include "test-float-libmvec-asinhf.c"
> diff --git a/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx512f.c b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx512f.c
> new file mode 100644
> index 0000000000..77e1838bb4
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx512f.c
> @@ -0,0 +1 @@
> +#include "test-float-libmvec-asinhf.c"
> diff --git a/sysdeps/x86_64/fpu/test-float-libmvec-asinhf.c b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf.c
> new file mode 100644
> index 0000000000..3353754102
> --- /dev/null
> +++ b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf.c
> @@ -0,0 +1,3 @@
> +#define LIBMVEC_TYPE float
> +#define LIBMVEC_FUNC asinhf
> +#include "test-vector-abi-arg1.h"
> diff --git a/sysdeps/x86_64/fpu/test-float-vlen16-wrappers.c b/sysdeps/x86_64/fpu/test-float-vlen16-wrappers.c
> index abbebf9993..e8ab1885a7 100644
> --- a/sysdeps/x86_64/fpu/test-float-vlen16-wrappers.c
> +++ b/sysdeps/x86_64/fpu/test-float-vlen16-wrappers.c
> @@ -45,6 +45,7 @@ VECTOR_WRAPPER (WRAPPER_NAME (atanhf), _ZGVeN16v_atanhf)
>  VECTOR_WRAPPER (WRAPPER_NAME (acoshf), _ZGVeN16v_acoshf)
>  VECTOR_WRAPPER (WRAPPER_NAME (erff), _ZGVeN16v_erff)
>  VECTOR_WRAPPER (WRAPPER_NAME (tanhf), _ZGVeN16v_tanhf)
> +VECTOR_WRAPPER (WRAPPER_NAME (asinhf), _ZGVeN16v_asinhf)
>  
>  #define VEC_INT_TYPE __m512i
>  
> diff --git a/sysdeps/x86_64/fpu/test-float-vlen4-wrappers.c b/sysdeps/x86_64/fpu/test-float-vlen4-wrappers.c
> index ae1c8b98c2..a80c5387e4 100644
> --- a/sysdeps/x86_64/fpu/test-float-vlen4-wrappers.c
> +++ b/sysdeps/x86_64/fpu/test-float-vlen4-wrappers.c
> @@ -45,6 +45,7 @@ VECTOR_WRAPPER (WRAPPER_NAME (atanhf), _ZGVbN4v_atanhf)
>  VECTOR_WRAPPER (WRAPPER_NAME (acoshf), _ZGVbN4v_acoshf)
>  VECTOR_WRAPPER (WRAPPER_NAME (erff), _ZGVbN4v_erff)
>  VECTOR_WRAPPER (WRAPPER_NAME (tanhf), _ZGVbN4v_tanhf)
> +VECTOR_WRAPPER (WRAPPER_NAME (asinhf), _ZGVbN4v_asinhf)
>  
>  #define VEC_INT_TYPE __m128i
>  
> diff --git a/sysdeps/x86_64/fpu/test-float-vlen8-avx2-wrappers.c b/sysdeps/x86_64/fpu/test-float-vlen8-avx2-wrappers.c
> index eb477a0371..c3d1d5936b 100644
> --- a/sysdeps/x86_64/fpu/test-float-vlen8-avx2-wrappers.c
> +++ b/sysdeps/x86_64/fpu/test-float-vlen8-avx2-wrappers.c
> @@ -48,6 +48,7 @@ VECTOR_WRAPPER (WRAPPER_NAME (atanhf), _ZGVdN8v_atanhf)
>  VECTOR_WRAPPER (WRAPPER_NAME (acoshf), _ZGVdN8v_acoshf)
>  VECTOR_WRAPPER (WRAPPER_NAME (erff), _ZGVdN8v_erff)
>  VECTOR_WRAPPER (WRAPPER_NAME (tanhf), _ZGVdN8v_tanhf)
> +VECTOR_WRAPPER (WRAPPER_NAME (asinhf), _ZGVdN8v_asinhf)
>  
>  /* Redefinition of wrapper to be compatible with _ZGVdN8vvv_sincosf.  */
>  #undef VECTOR_WRAPPER_fFF
> diff --git a/sysdeps/x86_64/fpu/test-float-vlen8-wrappers.c b/sysdeps/x86_64/fpu/test-float-vlen8-wrappers.c
> index 944f7f0a75..b7da0f523b 100644
> --- a/sysdeps/x86_64/fpu/test-float-vlen8-wrappers.c
> +++ b/sysdeps/x86_64/fpu/test-float-vlen8-wrappers.c
> @@ -45,6 +45,7 @@ VECTOR_WRAPPER (WRAPPER_NAME (atanhf), _ZGVcN8v_atanhf)
>  VECTOR_WRAPPER (WRAPPER_NAME (acoshf), _ZGVcN8v_acoshf)
>  VECTOR_WRAPPER (WRAPPER_NAME (erff), _ZGVcN8v_erff)
>  VECTOR_WRAPPER (WRAPPER_NAME (tanhf), _ZGVcN8v_tanhf)
> +VECTOR_WRAPPER (WRAPPER_NAME (asinhf), _ZGVcN8v_asinhf)
>  
>  #define VEC_INT_TYPE __m128i
>  
> -- 
> 2.31.1
> 

LGTM.

Reviewed-by: H.J. Lu <hjl.tools@gmail.com>

Thanks.


H.J.
  

Patch

diff --git a/bits/libm-simd-decl-stubs.h b/bits/libm-simd-decl-stubs.h
index 21f1a43232..bcaddb7a0e 100644
--- a/bits/libm-simd-decl-stubs.h
+++ b/bits/libm-simd-decl-stubs.h
@@ -296,4 +296,15 @@ 
 #define __DECL_SIMD_tanhf32x
 #define __DECL_SIMD_tanhf64x
 #define __DECL_SIMD_tanhf128x
+
+#define __DECL_SIMD_asinh
+#define __DECL_SIMD_asinhf
+#define __DECL_SIMD_asinhl
+#define __DECL_SIMD_asinhf16
+#define __DECL_SIMD_asinhf32
+#define __DECL_SIMD_asinhf64
+#define __DECL_SIMD_asinhf128
+#define __DECL_SIMD_asinhf32x
+#define __DECL_SIMD_asinhf64x
+#define __DECL_SIMD_asinhf128x
 #endif
diff --git a/math/bits/mathcalls.h b/math/bits/mathcalls.h
index 3d1c2056d5..40e055e579 100644
--- a/math/bits/mathcalls.h
+++ b/math/bits/mathcalls.h
@@ -84,7 +84,7 @@  __MATHDECL_VEC (void,sincos,,
 /* Hyperbolic arc cosine of X.  */
 __MATHCALL_VEC (acosh,, (_Mdouble_ __x));
 /* Hyperbolic arc sine of X.  */
-__MATHCALL (asinh,, (_Mdouble_ __x));
+__MATHCALL_VEC (asinh,, (_Mdouble_ __x));
 /* Hyperbolic arc tangent of X.  */
 __MATHCALL_VEC (atanh,, (_Mdouble_ __x));
 #endif
diff --git a/sysdeps/unix/sysv/linux/x86_64/libmvec.abilist b/sysdeps/unix/sysv/linux/x86_64/libmvec.abilist
index e178cef683..df265d6a12 100644
--- a/sysdeps/unix/sysv/linux/x86_64/libmvec.abilist
+++ b/sysdeps/unix/sysv/linux/x86_64/libmvec.abilist
@@ -49,6 +49,7 @@  GLIBC_2.22 _ZGVeN8vvv_sincos F
 GLIBC_2.35 _ZGVbN2v_acos F
 GLIBC_2.35 _ZGVbN2v_acosh F
 GLIBC_2.35 _ZGVbN2v_asin F
+GLIBC_2.35 _ZGVbN2v_asinh F
 GLIBC_2.35 _ZGVbN2v_atan F
 GLIBC_2.35 _ZGVbN2v_atanh F
 GLIBC_2.35 _ZGVbN2v_cbrt F
@@ -67,6 +68,7 @@  GLIBC_2.35 _ZGVbN2vv_hypot F
 GLIBC_2.35 _ZGVbN4v_acosf F
 GLIBC_2.35 _ZGVbN4v_acoshf F
 GLIBC_2.35 _ZGVbN4v_asinf F
+GLIBC_2.35 _ZGVbN4v_asinhf F
 GLIBC_2.35 _ZGVbN4v_atanf F
 GLIBC_2.35 _ZGVbN4v_atanhf F
 GLIBC_2.35 _ZGVbN4v_cbrtf F
@@ -85,6 +87,7 @@  GLIBC_2.35 _ZGVbN4vv_hypotf F
 GLIBC_2.35 _ZGVcN4v_acos F
 GLIBC_2.35 _ZGVcN4v_acosh F
 GLIBC_2.35 _ZGVcN4v_asin F
+GLIBC_2.35 _ZGVcN4v_asinh F
 GLIBC_2.35 _ZGVcN4v_atan F
 GLIBC_2.35 _ZGVcN4v_atanh F
 GLIBC_2.35 _ZGVcN4v_cbrt F
@@ -103,6 +106,7 @@  GLIBC_2.35 _ZGVcN4vv_hypot F
 GLIBC_2.35 _ZGVcN8v_acosf F
 GLIBC_2.35 _ZGVcN8v_acoshf F
 GLIBC_2.35 _ZGVcN8v_asinf F
+GLIBC_2.35 _ZGVcN8v_asinhf F
 GLIBC_2.35 _ZGVcN8v_atanf F
 GLIBC_2.35 _ZGVcN8v_atanhf F
 GLIBC_2.35 _ZGVcN8v_cbrtf F
@@ -121,6 +125,7 @@  GLIBC_2.35 _ZGVcN8vv_hypotf F
 GLIBC_2.35 _ZGVdN4v_acos F
 GLIBC_2.35 _ZGVdN4v_acosh F
 GLIBC_2.35 _ZGVdN4v_asin F
+GLIBC_2.35 _ZGVdN4v_asinh F
 GLIBC_2.35 _ZGVdN4v_atan F
 GLIBC_2.35 _ZGVdN4v_atanh F
 GLIBC_2.35 _ZGVdN4v_cbrt F
@@ -139,6 +144,7 @@  GLIBC_2.35 _ZGVdN4vv_hypot F
 GLIBC_2.35 _ZGVdN8v_acosf F
 GLIBC_2.35 _ZGVdN8v_acoshf F
 GLIBC_2.35 _ZGVdN8v_asinf F
+GLIBC_2.35 _ZGVdN8v_asinhf F
 GLIBC_2.35 _ZGVdN8v_atanf F
 GLIBC_2.35 _ZGVdN8v_atanhf F
 GLIBC_2.35 _ZGVdN8v_cbrtf F
@@ -157,6 +163,7 @@  GLIBC_2.35 _ZGVdN8vv_hypotf F
 GLIBC_2.35 _ZGVeN16v_acosf F
 GLIBC_2.35 _ZGVeN16v_acoshf F
 GLIBC_2.35 _ZGVeN16v_asinf F
+GLIBC_2.35 _ZGVeN16v_asinhf F
 GLIBC_2.35 _ZGVeN16v_atanf F
 GLIBC_2.35 _ZGVeN16v_atanhf F
 GLIBC_2.35 _ZGVeN16v_cbrtf F
@@ -175,6 +182,7 @@  GLIBC_2.35 _ZGVeN16vv_hypotf F
 GLIBC_2.35 _ZGVeN8v_acos F
 GLIBC_2.35 _ZGVeN8v_acosh F
 GLIBC_2.35 _ZGVeN8v_asin F
+GLIBC_2.35 _ZGVeN8v_asinh F
 GLIBC_2.35 _ZGVeN8v_atan F
 GLIBC_2.35 _ZGVeN8v_atanh F
 GLIBC_2.35 _ZGVeN8v_cbrt F
diff --git a/sysdeps/x86/fpu/bits/math-vector.h b/sysdeps/x86/fpu/bits/math-vector.h
index 3c657f6108..71b7d660db 100644
--- a/sysdeps/x86/fpu/bits/math-vector.h
+++ b/sysdeps/x86/fpu/bits/math-vector.h
@@ -130,6 +130,10 @@ 
 #  define __DECL_SIMD_tanh __DECL_SIMD_x86_64
 #  undef __DECL_SIMD_tanhf
 #  define __DECL_SIMD_tanhf __DECL_SIMD_x86_64
+#  undef __DECL_SIMD_asinh
+#  define __DECL_SIMD_asinh __DECL_SIMD_x86_64
+#  undef __DECL_SIMD_asinhf
+#  define __DECL_SIMD_asinhf __DECL_SIMD_x86_64
 
 # endif
 #endif
diff --git a/sysdeps/x86/fpu/finclude/math-vector-fortran.h b/sysdeps/x86/fpu/finclude/math-vector-fortran.h
index c7f81945fe..4d3afdf753 100644
--- a/sysdeps/x86/fpu/finclude/math-vector-fortran.h
+++ b/sysdeps/x86/fpu/finclude/math-vector-fortran.h
@@ -64,6 +64,8 @@ 
 !GCC$ builtin (erff) attributes simd (notinbranch) if('x86_64')
 !GCC$ builtin (tanh) attributes simd (notinbranch) if('x86_64')
 !GCC$ builtin (tanhf) attributes simd (notinbranch) if('x86_64')
+!GCC$ builtin (asinh) attributes simd (notinbranch) if('x86_64')
+!GCC$ builtin (asinhf) attributes simd (notinbranch) if('x86_64')
 
 !GCC$ builtin (cos) attributes simd (notinbranch) if('x32')
 !GCC$ builtin (cosf) attributes simd (notinbranch) if('x32')
@@ -113,3 +115,5 @@ 
 !GCC$ builtin (erff) attributes simd (notinbranch) if('x32')
 !GCC$ builtin (tanh) attributes simd (notinbranch) if('x32')
 !GCC$ builtin (tanhf) attributes simd (notinbranch) if('x32')
+!GCC$ builtin (asinh) attributes simd (notinbranch) if('x32')
+!GCC$ builtin (asinhf) attributes simd (notinbranch) if('x32')
diff --git a/sysdeps/x86_64/fpu/Makeconfig b/sysdeps/x86_64/fpu/Makeconfig
index 26df8d47bf..2ff33c7dd8 100644
--- a/sysdeps/x86_64/fpu/Makeconfig
+++ b/sysdeps/x86_64/fpu/Makeconfig
@@ -25,6 +25,7 @@  libmvec-funcs = \
   acos \
   acosh \
   asin \
+  asinh \
   atan \
   atan2 \
   atanh \
diff --git a/sysdeps/x86_64/fpu/Versions b/sysdeps/x86_64/fpu/Versions
index adcbe0fefb..e6ead13085 100644
--- a/sysdeps/x86_64/fpu/Versions
+++ b/sysdeps/x86_64/fpu/Versions
@@ -17,6 +17,7 @@  libmvec {
     _ZGVbN2v_acos; _ZGVcN4v_acos; _ZGVdN4v_acos; _ZGVeN8v_acos;
     _ZGVbN2v_acosh; _ZGVcN4v_acosh; _ZGVdN4v_acosh; _ZGVeN8v_acosh;
     _ZGVbN2v_asin; _ZGVcN4v_asin; _ZGVdN4v_asin; _ZGVeN8v_asin;
+    _ZGVbN2v_asinh; _ZGVcN4v_asinh; _ZGVdN4v_asinh; _ZGVeN8v_asinh;
     _ZGVbN2v_atan; _ZGVcN4v_atan; _ZGVdN4v_atan; _ZGVeN8v_atan;
     _ZGVbN2v_atanh; _ZGVcN4v_atanh; _ZGVdN4v_atanh; _ZGVeN8v_atanh;
     _ZGVbN2v_cbrt; _ZGVcN4v_cbrt; _ZGVdN4v_cbrt; _ZGVeN8v_cbrt;
@@ -35,6 +36,7 @@  libmvec {
     _ZGVbN4v_acosf; _ZGVcN8v_acosf; _ZGVdN8v_acosf; _ZGVeN16v_acosf;
     _ZGVbN4v_acoshf; _ZGVcN8v_acoshf; _ZGVdN8v_acoshf; _ZGVeN16v_acoshf;
     _ZGVbN4v_asinf; _ZGVcN8v_asinf; _ZGVdN8v_asinf; _ZGVeN16v_asinf;
+    _ZGVbN4v_asinhf; _ZGVcN8v_asinhf; _ZGVdN8v_asinhf; _ZGVeN16v_asinhf;
     _ZGVbN4v_atanf; _ZGVcN8v_atanf; _ZGVdN8v_atanf; _ZGVeN16v_atanf;
     _ZGVbN4v_atanhf; _ZGVcN8v_atanhf; _ZGVdN8v_atanhf; _ZGVeN16v_atanhf;
     _ZGVbN4v_cbrtf; _ZGVcN8v_cbrtf; _ZGVdN8v_cbrtf; _ZGVeN16v_cbrtf;
diff --git a/sysdeps/x86_64/fpu/libm-test-ulps b/sysdeps/x86_64/fpu/libm-test-ulps
index bfaad7acef..71e9fced02 100644
--- a/sysdeps/x86_64/fpu/libm-test-ulps
+++ b/sysdeps/x86_64/fpu/libm-test-ulps
@@ -157,6 +157,23 @@  float: 3
 float128: 4
 ldouble: 5
 
+Function: "asinh_vlen2":
+double: 1
+
+Function: "asinh_vlen4":
+double: 1
+float: 1
+
+Function: "asinh_vlen4_avx2":
+double: 1
+
+Function: "asinh_vlen8":
+double: 1
+float: 1
+
+Function: "asinh_vlen8_avx2":
+float: 1
+
 Function: "atan":
 double: 1
 float: 1
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core-sse2.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core-sse2.S
new file mode 100644
index 0000000000..ddd1c3ca24
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core-sse2.S
@@ -0,0 +1,20 @@ 
+/* SSE2 version of vectorized asinh, vector length is 2.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#define _ZGVbN2v_asinh _ZGVbN2v_asinh_sse2
+#include "../svml_d_asinh2_core.S"
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core.c b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core.c
new file mode 100644
index 0000000000..37452d0f92
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core.c
@@ -0,0 +1,27 @@ 
+/* Multiple versions of vectorized asinh, vector length is 2.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#define SYMBOL_NAME _ZGVbN2v_asinh
+#include "ifunc-mathvec-sse4_1.h"
+
+libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
+
+#ifdef SHARED
+__hidden_ver1 (_ZGVbN2v_asinh, __GI__ZGVbN2v_asinh, __redirect__ZGVbN2v_asinh)
+  __attribute__ ((visibility ("hidden")));
+#endif
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core_sse4.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core_sse4.S
new file mode 100644
index 0000000000..0fe130f20a
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh2_core_sse4.S
@@ -0,0 +1,1662 @@ 
+/* Function asinh vectorized with SSE4.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   https://www.gnu.org/licenses/.  */
+
+/*
+ * ALGORITHM DESCRIPTION:
+ *
+ *   Compute asinh(x) as log(x + sqrt(x*x + 1))
+ *
+ *   Special cases:
+ *
+ *   asinh(NaN) = quiet NaN, and raise invalid exception
+ *   asinh(INF) = that INF
+ *   asinh(0)   = that 0
+ *
+ */
+
+/* Offsets for data table __svml_dasinh_data_internal
+ */
+#define Log_HA_table                  	0
+#define Log_LA_table                  	8208
+#define poly_coeff                    	12320
+#define ExpMask                       	12384
+#define Two10                         	12400
+#define MinLog1p                      	12416
+#define MaxLog1p                      	12432
+#define One                           	12448
+#define SgnMask                       	12464
+#define XThreshold                    	12480
+#define XhMask                        	12496
+#define Threshold                     	12512
+#define Bias                          	12528
+#define Bias1                         	12544
+#define ExpMask0                      	12560
+#define ExpMask2                      	12576
+#define L2                            	12592
+#define dBigThreshold                 	12608
+#define dC2                           	12624
+#define dC3                           	12640
+#define dC4                           	12656
+#define dC5                           	12672
+#define dHalf                         	12688
+#define dLargestFinite                	12704
+#define dLittleThreshold              	12720
+#define dSign                         	12736
+#define dThirtyOne                    	12752
+#define dTopMask12                    	12768
+#define dTopMask26                    	12784
+#define dTopMask29                    	12800
+#define XScale                        	12816
+
+/* Lookup bias for data table __svml_dasinh_data_internal.  */
+#define Table_Lookup_Bias               -0x405ff0
+
+#include <sysdep.h>
+
+        .text
+	.section .text.sse4,"ax",@progbits
+ENTRY(_ZGVbN2v_asinh_sse4)
+        pushq     %rbp
+        cfi_def_cfa_offset(16)
+        movq      %rsp, %rbp
+        cfi_def_cfa(6, 16)
+        cfi_offset(6, -16)
+        andq      $-32, %rsp
+        subq      $64, %rsp
+        movaps    %xmm0, %xmm13
+
+/*
+ * Split X into high and low parts, XHi (<= 26 bits) and XLo (<= 27 bits)
+ * We could use either X or |X| here, but it doesn't seem to matter
+ */
+        movups    dTopMask26+__svml_dasinh_data_internal(%rip), %xmm15
+        movaps    %xmm13, %xmm7
+        andps     %xmm13, %xmm15
+        lea       Table_Lookup_Bias+__svml_dasinh_data_internal(%rip), %rsi
+
+/*
+ * Compute X^2 = (XHi + XLo)^2 = XHi^2 + XLo * (X + XHi)
+ * The two parts are shifted off by around 26 bits. So even though
+ * the low bit will not in general be exact, it's near enough
+ */
+        movaps    %xmm15, %xmm8
+        mulpd     %xmm15, %xmm8
+        subpd     %xmm15, %xmm7
+        addpd     %xmm13, %xmm15
+
+/* Load the constant 1 and a sign mask */
+        movups    One+__svml_dasinh_data_internal(%rip), %xmm12
+
+/*
+ * Finally, express Y + W = X^2 + 1 accurately where Y has <= 29 bits.
+ * If |X| <= 1 then |XHi| <= 1 and so |X2Hi| <= 1, so we can treat 1
+ * as the dominant component in the compensated summation. Otherwise,
+ * if |X| >= 1, then since X2Hi only has 52 significant bits, the basic
+ * addition will be exact anyway until we get to |X| >= 2^53. But by
+ * that time the log function is well-conditioned enough that the
+ * rounding error doesn't matter. Hence we can treat 1 as dominant even
+ * if it literally isn't.
+ */
+        movaps    %xmm12, %xmm3
+        movaps    %xmm12, %xmm5
+        addpd     %xmm8, %xmm3
+        mulpd     %xmm15, %xmm7
+        subpd     %xmm3, %xmm5
+        movups    dTopMask29+__svml_dasinh_data_internal(%rip), %xmm6
+        andps     %xmm3, %xmm6
+
+/*
+ * Compute R = 1/sqrt(Y + W) * (1 + d)
+ * Force R to <= 12 significant bits in case it isn't already
+ * This means that R * Y and R^2 * Y are exactly representable.
+ */
+        cvtpd2ps  %xmm6, %xmm1
+        addpd     %xmm8, %xmm5
+        subpd     %xmm6, %xmm3
+
+/*
+ * Unfortunately, we can still be in trouble if |X| <= 2^-10, since
+ * the absolute error 2^-(12+53)-ish in sqrt(1 + X^2) gets scaled up
+ * by 1/X and comes close to our threshold. Hence if |X| <= 2^-9,
+ * perform an alternative computation
+ * sqrt(1 + X^2) - 1 = X^2/2 - X^4/8 + X^6/16
+ * X2 = X^2
+ */
+        addpd     %xmm7, %xmm8
+        addpd     %xmm7, %xmm5
+        movlhps   %xmm1, %xmm1
+        rsqrtps   %xmm1, %xmm4
+        addpd     %xmm3, %xmm5
+        cvtps2pd  %xmm4, %xmm2
+        andps     dTopMask12+__svml_dasinh_data_internal(%rip), %xmm2
+
+/*
+ * Compute e = -(2 * d + d^2)
+ * The first FMR is exact, and the rounding error in the other is acceptable
+ * since d and e are ~ 2^-12
+ */
+        movaps    %xmm12, %xmm1
+
+/*
+ * Compute S = (Y/sqrt(Y + W)) * (1 + d)
+ * and T = (W/sqrt(Y + W)) * (1 + d)
+ * so that S + T = sqrt(Y + W) * (1 + d)
+ * S is exact, and the rounding error in T is OK.
+ */
+        mulpd     %xmm2, %xmm6
+        mulpd     %xmm2, %xmm5
+        movaps    %xmm2, %xmm0
+
+/*
+ * Obtain sqrt(1 + X^2) - 1 in two pieces
+ * sqrt(1 + X^2) - 1
+ * = sqrt(Y + W) - 1
+ * = (S + T) * (1 + Corr) - 1
+ * = [S - 1] + [T + (S + T) * Corr]
+ * We need a compensated summation for the last part. We treat S - 1
+ * as the larger part; it certainly is until about X < 2^-4, and in that
+ * case, the error is affordable since X dominates over sqrt(1 + X^2) - 1
+ * Final sum is dTmp5 (hi) + dTmp7 (lo)
+ */
+        movaps    %xmm6, %xmm3
+        mulpd     %xmm6, %xmm0
+        mulpd     %xmm5, %xmm2
+        subpd     %xmm0, %xmm1
+        addpd     %xmm5, %xmm3
+        subpd     %xmm12, %xmm6
+        subpd     %xmm2, %xmm1
+        movups    SgnMask+__svml_dasinh_data_internal(%rip), %xmm9
+        movaps    %xmm12, %xmm4
+
+/*
+ * Get the absolute value of the input, since we will exploit antisymmetry
+ * and mostly assume X >= 0 in the core computation
+ */
+        movaps    %xmm9, %xmm10
+        andps     %xmm13, %xmm10
+
+/*
+ * Check whether the input is finite, by checking |X| <= MaxFloat
+ * Otherwise set the rangemask so that the callout will get used.
+ * Note that this will also use the callout for NaNs since not(NaN <= MaxFloat)
+ */
+        movaps    %xmm10, %xmm14
+
+/*
+ * The following computation can go wrong for very large X, basically
+ * because X^2 overflows. But for large X we have
+ * asinh(X) / log(2 X) - 1 =~= 1/(4 * X^2), so for X >= 2^30
+ * we can just later stick X back into the log and tweak up the exponent.
+ * Actually we scale X by 2^-30 and tweak the exponent up by 31,
+ * to stay in the safe range for the later log computation.
+ * Compute a flag now telling us when do do this.
+ */
+        movaps    %xmm10, %xmm11
+        cmpnlepd  dLargestFinite+__svml_dasinh_data_internal(%rip), %xmm14
+        cmpltpd   dBigThreshold+__svml_dasinh_data_internal(%rip), %xmm11
+        movmskpd  %xmm14, %edx
+
+/*
+ * Now       1 / (1 + d)
+ * = 1 / (1 + (sqrt(1 - e) - 1))
+ * = 1 / sqrt(1 - e)
+ * = 1 + 1/2 * e + 3/8 * e^2 + 5/16 * e^3 + 35/128 * e^4 +
+ * 63/256 * e^5 + 231/1024 * e^6 + ....
+ * So compute the first five nonconstant terms of that, so that
+ * we have a relative correction (1 + Corr) to apply to S etc.
+ * C1 = 1/2
+ * C2 = 3/8
+ * C3 = 5/16
+ * C4 = 35/128
+ * C5 = 63/256
+ */
+        movups    dC5+__svml_dasinh_data_internal(%rip), %xmm14
+        movups    dHalf+__svml_dasinh_data_internal(%rip), %xmm15
+        mulpd     %xmm1, %xmm14
+
+/* dX2over2 = X^2/2 */
+        mulpd     %xmm15, %xmm8
+        addpd     dC4+__svml_dasinh_data_internal(%rip), %xmm14
+        mulpd     %xmm1, %xmm14
+        addpd     dC3+__svml_dasinh_data_internal(%rip), %xmm14
+        mulpd     %xmm1, %xmm14
+        addpd     dC2+__svml_dasinh_data_internal(%rip), %xmm14
+        mulpd     %xmm1, %xmm14
+        addpd     %xmm15, %xmm14
+        mulpd     %xmm14, %xmm1
+        mulpd     %xmm3, %xmm1
+        addpd     %xmm1, %xmm5
+        addpd     %xmm6, %xmm5
+
+/* dX4over4 = X^4/4 */
+        movaps    %xmm8, %xmm6
+
+/* dX46 = -X^4/4 + X^6/8 */
+        movaps    %xmm8, %xmm7
+        mulpd     %xmm8, %xmm6
+        mulpd     %xmm6, %xmm7
+        subpd     %xmm6, %xmm7
+
+/* dX46over2 = -X^4/8 + x^6/16 */
+        mulpd     %xmm7, %xmm15
+
+/* Now multiplex the two possible computations */
+        movaps    %xmm10, %xmm3
+        cmplepd   dLittleThreshold+__svml_dasinh_data_internal(%rip), %xmm3
+        addpd     %xmm15, %xmm8
+        movaps    %xmm3, %xmm1
+        andps     %xmm3, %xmm8
+        andnps    %xmm5, %xmm1
+
+/*
+ * Now we feed into the log1p code, using H in place of _VARG1 and
+ * also adding L into Xl.
+ * compute 1+x as high, low parts
+ */
+        movaps    %xmm12, %xmm5
+        orps      %xmm8, %xmm1
+        movaps    %xmm11, %xmm3
+
+/*
+ * Now do another compensated sum to add |X| + [sqrt(1 + X^2) - 1].
+ * It's always safe to assume |X| is larger.
+ * This is the final 2-part argument to the log1p function
+ */
+        addpd     %xmm10, %xmm1
+        maxpd     %xmm1, %xmm5
+        minpd     %xmm1, %xmm4
+
+/* Now multiplex to the case X = 2^-30 * |input|, Xl = dL = 0 in the "big" case. */
+        movups    XScale+__svml_dasinh_data_internal(%rip), %xmm8
+        andps     %xmm9, %xmm1
+        mulpd     %xmm8, %xmm10
+        cmpltpd   XThreshold+__svml_dasinh_data_internal(%rip), %xmm1
+        movaps    %xmm5, %xmm9
+        andnps    %xmm10, %xmm3
+        addpd     %xmm4, %xmm9
+        orps      XhMask+__svml_dasinh_data_internal(%rip), %xmm1
+        andps     %xmm1, %xmm9
+        subpd     %xmm9, %xmm5
+        andps     %xmm11, %xmm9
+
+/* Now resume the main code. */
+        movups    ExpMask+__svml_dasinh_data_internal(%rip), %xmm10
+        orps      %xmm9, %xmm3
+
+/* preserve mantissa, set input exponent to 2^(-10) */
+        andps     %xmm3, %xmm10
+
+/* exponent bits */
+        movaps    %xmm3, %xmm7
+        orps      Two10+__svml_dasinh_data_internal(%rip), %xmm10
+        psrlq     $20, %xmm7
+
+/* reciprocal approximation good to at least 11 bits */
+        cvtpd2ps  %xmm10, %xmm1
+        addpd     %xmm5, %xmm4
+        movlhps   %xmm1, %xmm1
+        andps     %xmm11, %xmm4
+        rcpps     %xmm1, %xmm0
+        cvtps2pd  %xmm0, %xmm0
+
+/* round reciprocal to nearest integer, will have 1+9 mantissa bits */
+        movups    .FLT_30(%rip), %xmm6
+        movaps    %xmm11, %xmm1
+        addpd     %xmm6, %xmm0
+        subpd     %xmm6, %xmm0
+
+/* exponent of X needed to scale Xl */
+        movdqu    ExpMask0+__svml_dasinh_data_internal(%rip), %xmm5
+
+/* 2^ (-10-exp(X) ) */
+        movdqu    ExpMask2+__svml_dasinh_data_internal(%rip), %xmm2
+        pand      %xmm3, %xmm5
+        psubq     %xmm5, %xmm2
+
+/* scale DblRcp */
+        mulpd     %xmm0, %xmm2
+
+/* argument reduction */
+        mulpd     %xmm2, %xmm3
+        mulpd     %xmm2, %xmm4
+        subpd     %xmm12, %xmm3
+        addpd     %xmm4, %xmm3
+
+/* polynomial */
+        movups    poly_coeff+__svml_dasinh_data_internal(%rip), %xmm12
+        movaps    %xmm3, %xmm2
+        pshufd    $221, %xmm7, %xmm8
+        mulpd     %xmm3, %xmm12
+
+/* biased exponent in DP format */
+        cvtdq2pd  %xmm8, %xmm14
+        addpd     poly_coeff+16+__svml_dasinh_data_internal(%rip), %xmm12
+        mulpd     %xmm3, %xmm2
+
+/* Add 31 to the exponent in the "large" case to get log(2 * input) */
+        movups    dThirtyOne+__svml_dasinh_data_internal(%rip), %xmm9
+
+/* exponent*log(2.0) */
+        movups    Threshold+__svml_dasinh_data_internal(%rip), %xmm5
+        addpd     %xmm14, %xmm9
+        cmpltpd   %xmm0, %xmm5
+        mulpd     %xmm2, %xmm12
+        andps     %xmm11, %xmm14
+
+/*
+ * prepare table index
+ * table lookup
+ */
+        movaps    %xmm0, %xmm11
+        movups    poly_coeff+32+__svml_dasinh_data_internal(%rip), %xmm0
+        andnps    %xmm9, %xmm1
+        mulpd     %xmm3, %xmm0
+        addpd     poly_coeff+48+__svml_dasinh_data_internal(%rip), %xmm0
+        addpd     %xmm12, %xmm0
+
+/* reconstruction */
+        mulpd     %xmm0, %xmm2
+        andps     Bias+__svml_dasinh_data_internal(%rip), %xmm5
+        psrlq     $40, %xmm11
+        orps      Bias1+__svml_dasinh_data_internal(%rip), %xmm5
+        orps      %xmm14, %xmm1
+        movd      %xmm11, %eax
+        pshufd    $2, %xmm11, %xmm11
+
+/* Finally, reincorporate the original sign. */
+        movups    dSign+__svml_dasinh_data_internal(%rip), %xmm0
+        subpd     %xmm5, %xmm1
+        addpd     %xmm2, %xmm3
+        movd      %xmm11, %ecx
+        mulpd     L2+__svml_dasinh_data_internal(%rip), %xmm1
+        movslq    %eax, %rax
+        andps     %xmm13, %xmm0
+        movslq    %ecx, %rcx
+        movsd     (%rsi,%rax), %xmm6
+        movhpd    (%rsi,%rcx), %xmm6
+        addpd     %xmm3, %xmm6
+        addpd     %xmm6, %xmm1
+        pxor      %xmm1, %xmm0
+        testl     %edx, %edx
+
+/* Go to special inputs processing branch */
+        jne       L(SPECIAL_VALUES_BRANCH)
+                                # LOE rbx r12 r13 r14 r15 edx xmm0 xmm13
+
+/* Restore registers
+ * and exit the function
+ */
+
+L(EXIT):
+        movq      %rbp, %rsp
+        popq      %rbp
+        cfi_def_cfa(7, 8)
+        cfi_restore(6)
+        ret
+        cfi_def_cfa(6, 16)
+        cfi_offset(6, -16)
+
+/* Branch to process
+ * special inputs
+ */
+
+L(SPECIAL_VALUES_BRANCH):
+        movups    %xmm13, 32(%rsp)
+        movups    %xmm0, 48(%rsp)
+                                # LOE rbx r12 r13 r14 r15 edx
+
+        xorl      %eax, %eax
+        movq      %r12, 16(%rsp)
+        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -48; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xd0, 0xff, 0xff, 0xff, 0x22
+        movl      %eax, %r12d
+        movq      %r13, 8(%rsp)
+        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -56; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xc8, 0xff, 0xff, 0xff, 0x22
+        movl      %edx, %r13d
+        movq      %r14, (%rsp)
+        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -64; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x22
+                                # LOE rbx r15 r12d r13d
+
+/* Range mask
+ * bits check
+ */
+
+L(RANGEMASK_CHECK):
+        btl       %r12d, %r13d
+
+/* Call scalar math function */
+        jc        L(SCALAR_MATH_CALL)
+                                # LOE rbx r15 r12d r13d
+
+/* Special inputs
+ * processing loop
+ */
+
+L(SPECIAL_VALUES_LOOP):
+        incl      %r12d
+        cmpl      $2, %r12d
+
+/* Check bits in range mask */
+        jl        L(RANGEMASK_CHECK)
+                                # LOE rbx r15 r12d r13d
+
+        movq      16(%rsp), %r12
+        cfi_restore(12)
+        movq      8(%rsp), %r13
+        cfi_restore(13)
+        movq      (%rsp), %r14
+        cfi_restore(14)
+        movups    48(%rsp), %xmm0
+
+/* Go to exit */
+        jmp       L(EXIT)
+        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -48; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xd0, 0xff, 0xff, 0xff, 0x22
+        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -56; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xc8, 0xff, 0xff, 0xff, 0x22
+        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -64; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x22
+                                # LOE rbx r12 r13 r14 r15 xmm0
+
+/* Scalar math fucntion call
+ * to process special input
+ */
+
+L(SCALAR_MATH_CALL):
+        movl      %r12d, %r14d
+        movsd     32(%rsp,%r14,8), %xmm0
+        call      asinh@PLT
+                                # LOE rbx r14 r15 r12d r13d xmm0
+
+        movsd     %xmm0, 48(%rsp,%r14,8)
+
+/* Process special inputs in loop */
+        jmp       L(SPECIAL_VALUES_LOOP)
+                                # LOE rbx r15 r12d r13d
+END(_ZGVbN2v_asinh_sse4)
+
+        .section .rodata, "a"
+        .align 16
+
+#ifdef __svml_dasinh_data_internal_typedef
+typedef unsigned int VUINT32;
+typedef struct {
+        __declspec(align(16)) VUINT32 Log_HA_table[(1<<10)+2][2];
+        __declspec(align(16)) VUINT32 Log_LA_table[(1<<9)+1][2];
+        __declspec(align(16)) VUINT32 poly_coeff[4][2][2];
+        __declspec(align(16)) VUINT32 ExpMask[2][2];
+        __declspec(align(16)) VUINT32 Two10[2][2];
+        __declspec(align(16)) VUINT32 MinLog1p[2][2];
+        __declspec(align(16)) VUINT32 MaxLog1p[2][2];
+        __declspec(align(16)) VUINT32 One[2][2];
+        __declspec(align(16)) VUINT32 SgnMask[2][2];
+        __declspec(align(16)) VUINT32 XThreshold[2][2];
+        __declspec(align(16)) VUINT32 XhMask[2][2];
+        __declspec(align(16)) VUINT32 Threshold[2][2];
+        __declspec(align(16)) VUINT32 Bias[2][2];
+        __declspec(align(16)) VUINT32 Bias1[2][2];
+        __declspec(align(16)) VUINT32 ExpMask0[2][2];
+        __declspec(align(16)) VUINT32 ExpMask2[2][2];
+        __declspec(align(16)) VUINT32 L2[2][2];
+        __declspec(align(16)) VUINT32 dBigThreshold[2][2];
+        __declspec(align(16)) VUINT32 dC2[2][2];
+        __declspec(align(16)) VUINT32 dC3[2][2];
+        __declspec(align(16)) VUINT32 dC4[2][2];
+        __declspec(align(16)) VUINT32 dC5[2][2];
+        __declspec(align(16)) VUINT32 dHalf[2][2];
+        __declspec(align(16)) VUINT32 dLargestFinite[2][2];
+        __declspec(align(16)) VUINT32 dLittleThreshold[2][2];
+        __declspec(align(16)) VUINT32 dSign[2][2];
+        __declspec(align(16)) VUINT32 dThirtyOne[2][2];
+        __declspec(align(16)) VUINT32 dTopMask12[2][2];
+        __declspec(align(16)) VUINT32 dTopMask26[2][2];
+        __declspec(align(16)) VUINT32 dTopMask29[2][2];
+        __declspec(align(16)) VUINT32 XScale[2][2];
+} __svml_dasinh_data_internal;
+#endif
+__svml_dasinh_data_internal:
+        /* Log_HA_table */
+        .quad 0xc086232bdd7a8300, 0xbe1ce91eef3fb100
+        .quad 0xc086232fdc7ad828, 0xbe1cefcffda73b6a
+        .quad 0xc0862333d97d2ba0, 0xbe1cef406748f1ff
+        .quad 0xc0862337d48378e0, 0xbe1cef2a9429925a
+        .quad 0xc086233bcd8fb878, 0xbe1cf138d17ebecb
+        .quad 0xc086233fc4a3e018, 0xbe1ceff2dbbbb29e
+        .quad 0xc0862343b9c1e270, 0xbe1cf1a42aae437b
+        .quad 0xc0862347acebaf68, 0xbe1cef3b152048af
+        .quad 0xc086234b9e2333f0, 0xbe1cef20e127805e
+        .quad 0xc086234f8d6a5a30, 0xbe1cf00ad6052cf4
+        .quad 0xc08623537ac30980, 0xbe1cefc4642ee597
+        .quad 0xc0862357662f2660, 0xbe1cf1f277d36e16
+        .quad 0xc086235b4fb092a0, 0xbe1ceed009e8d8e6
+        .quad 0xc086235f37492d28, 0xbe1cf1e4038cb362
+        .quad 0xc08623631cfad250, 0xbe1cf0b0873b8557
+        .quad 0xc086236700c75b98, 0xbe1cf15bb3227c0b
+        .quad 0xc086236ae2b09fe0, 0xbe1cf151ef8ca9ed
+        .quad 0xc086236ec2b87358, 0xbe1cefe1dc2cd2ed
+        .quad 0xc0862372a0e0a780, 0xbe1cf0d1eec5454f
+        .quad 0xc08623767d2b0b48, 0xbe1ceeefd570bbce
+        .quad 0xc086237a57996af0, 0xbe1cee99ae91b3a7
+        .quad 0xc086237e302d9028, 0xbe1cf0412830fbd1
+        .quad 0xc086238206e94218, 0xbe1ceee898588610
+        .quad 0xc0862385dbce4548, 0xbe1cee9a1fbcaaea
+        .quad 0xc0862389aede5bc0, 0xbe1ceed8e7cc1ad6
+        .quad 0xc086238d801b4500, 0xbe1cf10c8d059da6
+        .quad 0xc08623914f86be18, 0xbe1ceee6c63a8165
+        .quad 0xc08623951d228180, 0xbe1cf0c3592d2ff1
+        .quad 0xc0862398e8f04758, 0xbe1cf0026cc4cb1b
+        .quad 0xc086239cb2f1c538, 0xbe1cf15d48d8e670
+        .quad 0xc08623a07b28ae60, 0xbe1cef359363787c
+        .quad 0xc08623a44196b390, 0xbe1cefdf1ab2e82c
+        .quad 0xc08623a8063d8338, 0xbe1cefe43c02aa84
+        .quad 0xc08623abc91ec960, 0xbe1cf044f5ae35b7
+        .quad 0xc08623af8a3c2fb8, 0xbe1cf0b0b4001e1b
+        .quad 0xc08623b349975d98, 0xbe1cf1bae76dfbcf
+        .quad 0xc08623b70731f810, 0xbe1cef0a72e13a62
+        .quad 0xc08623bac30da1c8, 0xbe1cf184007d2b6b
+        .quad 0xc08623be7d2bfb40, 0xbe1cf16f4b239e98
+        .quad 0xc08623c2358ea2a0, 0xbe1cf0976acada87
+        .quad 0xc08623c5ec3733d0, 0xbe1cf066318a16ff
+        .quad 0xc08623c9a1274880, 0xbe1ceffaa7148798
+        .quad 0xc08623cd54607820, 0xbe1cf23ab02e9b6e
+        .quad 0xc08623d105e45800, 0xbe1cefdfef7d4fde
+        .quad 0xc08623d4b5b47b20, 0xbe1cf17fece44f2b
+        .quad 0xc08623d863d27270, 0xbe1cf18f907d0d7c
+        .quad 0xc08623dc103fccb0, 0xbe1cee61fe072c98
+        .quad 0xc08623dfbafe1668, 0xbe1cf022dd891e2f
+        .quad 0xc08623e3640eda20, 0xbe1ceecc1daf4358
+        .quad 0xc08623e70b73a028, 0xbe1cf0173c4fa380
+        .quad 0xc08623eab12deec8, 0xbe1cf16a2150c2f4
+        .quad 0xc08623ee553f4a30, 0xbe1cf1bf980b1f4b
+        .quad 0xc08623f1f7a93480, 0xbe1cef8b731663c2
+        .quad 0xc08623f5986d2dc0, 0xbe1cee9a664d7ef4
+        .quad 0xc08623f9378cb3f0, 0xbe1cf1eda2af6400
+        .quad 0xc08623fcd5094320, 0xbe1cf1923f9d68d7
+        .quad 0xc086240070e45548, 0xbe1cf0747cd3e03a
+        .quad 0xc08624040b1f6260, 0xbe1cf22ee855bd6d
+        .quad 0xc0862407a3bbe078, 0xbe1cf0d57360c00b
+        .quad 0xc086240b3abb4398, 0xbe1ceebc815cd575
+        .quad 0xc086240ed01efdd0, 0xbe1cf03bfb970951
+        .quad 0xc086241263e87f50, 0xbe1cf16e74768529
+        .quad 0xc0862415f6193658, 0xbe1cefec64b8becb
+        .quad 0xc086241986b28f30, 0xbe1cf0838d210baa
+        .quad 0xc086241d15b5f448, 0xbe1cf0ea86e75b11
+        .quad 0xc0862420a324ce28, 0xbe1cf1708d11d805
+        .quad 0xc08624242f008380, 0xbe1ceea988c5a417
+        .quad 0xc0862427b94a7910, 0xbe1cef166a7bbca5
+        .quad 0xc086242b420411d0, 0xbe1cf0c9d9e86a38
+        .quad 0xc086242ec92eaee8, 0xbe1cef0946455411
+        .quad 0xc08624324ecbaf98, 0xbe1cefea60907739
+        .quad 0xc0862435d2dc7160, 0xbe1cf1ed0934ce42
+        .quad 0xc086243955624ff8, 0xbe1cf191ba746c7d
+        .quad 0xc086243cd65ea548, 0xbe1ceeec78cf2a7e
+        .quad 0xc086244055d2c968, 0xbe1cef345284c119
+        .quad 0xc0862443d3c012b8, 0xbe1cf24f77355219
+        .quad 0xc08624475027d5e8, 0xbe1cf05bf087e114
+        .quad 0xc086244acb0b65d0, 0xbe1cef3504a32189
+        .quad 0xc086244e446c1398, 0xbe1ceff54b2a406f
+        .quad 0xc0862451bc4b2eb8, 0xbe1cf0757d54ed4f
+        .quad 0xc086245532aa04f0, 0xbe1cf0c8099fdfd5
+        .quad 0xc0862458a789e250, 0xbe1cf0b173796a31
+        .quad 0xc086245c1aec1138, 0xbe1cf11d8734540d
+        .quad 0xc086245f8cd1da60, 0xbe1cf1916a723ceb
+        .quad 0xc0862462fd3c84d8, 0xbe1cf19a911e1da7
+        .quad 0xc08624666c2d5608, 0xbe1cf23a9ef72e4f
+        .quad 0xc0862469d9a591c0, 0xbe1cef503d947663
+        .quad 0xc086246d45a67a18, 0xbe1cf0fceeb1a0b2
+        .quad 0xc0862470b0314fa8, 0xbe1cf107e27e4fbc
+        .quad 0xc086247419475160, 0xbe1cf03dd9922331
+        .quad 0xc086247780e9bc98, 0xbe1cefce1a10e129
+        .quad 0xc086247ae719cd18, 0xbe1ceea47f73c4f6
+        .quad 0xc086247e4bd8bd10, 0xbe1ceec0ac56d100
+        .quad 0xc0862481af27c528, 0xbe1cee8a6593278a
+        .quad 0xc086248511081c70, 0xbe1cf2231dd9dec7
+        .quad 0xc0862488717af888, 0xbe1cf0b4b8ed7da8
+        .quad 0xc086248bd0818d68, 0xbe1cf1bd8d835002
+        .quad 0xc086248f2e1d0d98, 0xbe1cf259acc107f4
+        .quad 0xc08624928a4eaa20, 0xbe1cee897636b00c
+        .quad 0xc0862495e5179270, 0xbe1cee757f20c326
+        .quad 0xc08624993e78f490, 0xbe1cefafd3aa54a4
+        .quad 0xc086249c9673fd10, 0xbe1cee7298d38b97
+        .quad 0xc086249fed09d6f8, 0xbe1ceedc158d4ceb
+        .quad 0xc08624a3423babe0, 0xbe1cf2282987cb2e
+        .quad 0xc08624a6960aa400, 0xbe1cefe7381ecc4b
+        .quad 0xc08624a9e877e600, 0xbe1cef328dbbce80
+        .quad 0xc08624ad39849728, 0xbe1cefde45f3cc71
+        .quad 0xc08624b08931db58, 0xbe1cefa8b89433b9
+        .quad 0xc08624b3d780d500, 0xbe1cef6773c0b139
+        .quad 0xc08624b72472a528, 0xbe1cf031c931c11f
+        .quad 0xc08624ba70086b78, 0xbe1cf088f49275e7
+        .quad 0xc08624bdba434630, 0xbe1cf17de0eaa86d
+        .quad 0xc08624c103245238, 0xbe1cefd492f1ba75
+        .quad 0xc08624c44aacab08, 0xbe1cf1253e154466
+        .quad 0xc08624c790dd6ad0, 0xbe1cf0fb09ee6d55
+        .quad 0xc08624cad5b7aa58, 0xbe1cf1f08dd048fe
+        .quad 0xc08624ce193c8120, 0xbe1ceeca0809697f
+        .quad 0xc08624d15b6d0538, 0xbe1cef8d5662d968
+        .quad 0xc08624d49c4a4b78, 0xbe1cee97b556ed78
+        .quad 0xc08624d7dbd56750, 0xbe1cf1b14b6acb75
+        .quad 0xc08624db1a0f6b00, 0xbe1cef1e860623f2
+        .quad 0xc08624de56f96758, 0xbe1ceeaf4d156f3d
+        .quad 0xc08624e192946bf0, 0xbe1ceecc12b400ed
+        .quad 0xc08624e4cce18710, 0xbe1cf180c40c794f
+        .quad 0xc08624e805e1c5c8, 0xbe1cf185a08f7f65
+        .quad 0xc08624eb3d9633d8, 0xbe1cef45fc924078
+        .quad 0xc08624ee73ffdbb0, 0xbe1cf1e4f457f32a
+        .quad 0xc08624f1a91fc6a0, 0xbe1cf040147b8a5a
+        .quad 0xc08624f4dcf6fc98, 0xbe1cf1effca0dfb2
+        .quad 0xc08624f80f868468, 0xbe1cf0470146e5bc
+        .quad 0xc08624fb40cf6390, 0xbe1cef4dd186e501
+        .quad 0xc08624fe70d29e60, 0xbe1ceebe257f66c7
+        .quad 0xc08625019f9137f0, 0xbe1ceefb7a1c395c
+        .quad 0xc0862504cd0c3220, 0xbe1cf209dedfed8c
+        .quad 0xc0862507f9448db0, 0xbe1cf082da464994
+        .quad 0xc086250b243b4a18, 0xbe1cee88694a73cf
+        .quad 0xc086250e4df165a0, 0xbe1cf0b61e8f0531
+        .quad 0xc08625117667dd78, 0xbe1cf1106599c962
+        .quad 0xc08625149d9fad98, 0xbe1ceff1ee88af1f
+        .quad 0xc0862517c399d0c8, 0xbe1cf0f746994ef6
+        .quad 0xc086251ae85740b8, 0xbe1cefe8a1d077e4
+        .quad 0xc086251e0bd8f5e0, 0xbe1cf1a1da036092
+        .quad 0xc08625212e1fe7a8, 0xbe1cf0f8a7786fcd
+        .quad 0xc08625244f2d0c48, 0xbe1cefa1174a07a7
+        .quad 0xc08625276f0158d8, 0xbe1cef1043aa5b25
+        .quad 0xc086252a8d9dc150, 0xbe1cf15d521c169d
+        .quad 0xc086252dab033898, 0xbe1cf220bba8861f
+        .quad 0xc0862530c732b078, 0xbe1cef51e310eae2
+        .quad 0xc0862533e22d1988, 0xbe1cf222fcedd8ae
+        .quad 0xc0862536fbf36370, 0xbe1cefdb4da4bda8
+        .quad 0xc086253a14867ca0, 0xbe1ceeafc1112171
+        .quad 0xc086253d2be75280, 0xbe1cee99dfb4b408
+        .quad 0xc08625404216d160, 0xbe1cf22d2536f06b
+        .quad 0xc08625435715e498, 0xbe1cef6abbf2e268
+        .quad 0xc08625466ae57648, 0xbe1cf093a14789f5
+        .quad 0xc08625497d866fa0, 0xbe1cf0f93655603c
+        .quad 0xc086254c8ef9b8b8, 0xbe1cf1cc40c9aafc
+        .quad 0xc086254f9f4038a8, 0xbe1ceeea5f4e9157
+        .quad 0xc0862552ae5ad568, 0xbe1cefa9f52d4997
+        .quad 0xc0862555bc4a7400, 0xbe1cefa490a638ff
+        .quad 0xc0862558c90ff868, 0xbe1cef7fcf797d6f
+        .quad 0xc086255bd4ac4590, 0xbe1cf1b4c51113c9
+        .quad 0xc086255edf203d78, 0xbe1cef55e5b4a55d
+        .quad 0xc0862561e86cc100, 0xbe1cf0d37a25f9dc
+        .quad 0xc0862564f092b028, 0xbe1ceebe9efc19d9
+        .quad 0xc0862567f792e9d8, 0xbe1cee8ad30a57b5
+        .quad 0xc086256afd6e4c08, 0xbe1cef4e1817b90b
+        .quad 0xc086256e0225b3b8, 0xbe1cee7fa9229996
+        .quad 0xc086257105b9fce0, 0xbe1cf0b54963d945
+        .quad 0xc0862574082c0298, 0xbe1cee5f2f3c7995
+        .quad 0xc0862577097c9ee0, 0xbe1cf0828e303a2c
+        .quad 0xc086257a09acaae0, 0xbe1cf172c3078947
+        .quad 0xc086257d08bcfec0, 0xbe1cf189252afa22
+        .quad 0xc086258006ae71b8, 0xbe1cefdb80426923
+        .quad 0xc08625830381da08, 0xbe1ceef1391a0372
+        .quad 0xc0862585ff380d00, 0xbe1cf17720c78d13
+        .quad 0xc0862588f9d1df18, 0xbe1ceef1f9027d83
+        .quad 0xc086258bf35023b8, 0xbe1cf06fac99dec9
+        .quad 0xc086258eebb3ad78, 0xbe1cf1373eeb45c0
+        .quad 0xc0862591e2fd4e00, 0xbe1cef777536bb81
+        .quad 0xc0862594d92dd600, 0xbe1cf0f43ca40766
+        .quad 0xc0862597ce461558, 0xbe1cefb2cfc6766b
+        .quad 0xc086259ac246daf0, 0xbe1ceea49e64ffa2
+        .quad 0xc086259db530f4c8, 0xbe1cf250fa457dec
+        .quad 0xc08625a0a7053018, 0xbe1cf17d8bb2a44e
+        .quad 0xc08625a397c45918, 0xbe1cf1d5906d54b7
+        .quad 0xc08625a6876f3b30, 0xbe1cf08fe7b31780
+        .quad 0xc08625a97606a0e0, 0xbe1cef13edfc9d11
+        .quad 0xc08625ac638b53c8, 0xbe1cef9d2b107219
+        .quad 0xc08625af4ffe1cb0, 0xbe1cf1ddd4ff6160
+        .quad 0xc08625b23b5fc390, 0xbe1cefa02a996495
+        .quad 0xc08625b525b10f68, 0xbe1cf166a7e37ee5
+        .quad 0xc08625b80ef2c680, 0xbe1cef0b171068a5
+        .quad 0xc08625baf725ae28, 0xbe1cf05c80779283
+        .quad 0xc08625bdde4a8af0, 0xbe1cf1bbfbffb889
+        .quad 0xc08625c0c4622090, 0xbe1cf0b8666c0124
+        .quad 0xc08625c3a96d31e0, 0xbe1cf0a8fcf47a86
+        .quad 0xc08625c68d6c80f0, 0xbe1cef46e18cb092
+        .quad 0xc08625c97060cef0, 0xbe1cf1458a350efb
+        .quad 0xc08625cc524adc58, 0xbe1ceeea1dadce12
+        .quad 0xc08625cf332b68b0, 0xbe1cf0a1bfdc44c7
+        .quad 0xc08625d2130332d0, 0xbe1cef96d02da73e
+        .quad 0xc08625d4f1d2f8a8, 0xbe1cf2451c3c7701
+        .quad 0xc08625d7cf9b7778, 0xbe1cf10d08f83812
+        .quad 0xc08625daac5d6ba0, 0xbe1ceec5b4895c5e
+        .quad 0xc08625dd881990b0, 0xbe1cf14e1325c5e4
+        .quad 0xc08625e062d0a188, 0xbe1cf21d0904be12
+        .quad 0xc08625e33c835838, 0xbe1ceed0839bcf21
+        .quad 0xc08625e615326df0, 0xbe1cf1bb944889d2
+        .quad 0xc08625e8ecde9b48, 0xbe1cee738e85eece
+        .quad 0xc08625ebc38897e0, 0xbe1cf25c2bc6ef12
+        .quad 0xc08625ee99311ac8, 0xbe1cf132b70a41ad
+        .quad 0xc08625f16dd8da28, 0xbe1cf1984236a6e3
+        .quad 0xc08625f441808b78, 0xbe1cf19ae74998f9
+        .quad 0xc08625f71428e370, 0xbe1cef3e175d61a1
+        .quad 0xc08625f9e5d295f8, 0xbe1cf101f9868fd9
+        .quad 0xc08625fcb67e5658, 0xbe1cee69db83dcd2
+        .quad 0xc08625ff862cd6f8, 0xbe1cf081b636af51
+        .quad 0xc086260254dec9a8, 0xbe1cee62c7d59b3e
+        .quad 0xc08626052294df58, 0xbe1cf1b745c57716
+        .quad 0xc0862607ef4fc868, 0xbe1cef3d2800ea23
+        .quad 0xc086260abb103458, 0xbe1cef480ff1acd2
+        .quad 0xc086260d85d6d200, 0xbe1cf2424c9a17ef
+        .quad 0xc08626104fa44f90, 0xbe1cf12cfde90fd5
+        .quad 0xc086261318795a68, 0xbe1cf21f590dd5b6
+        .quad 0xc0862615e0569f48, 0xbe1cf0c50f9cd28a
+        .quad 0xc0862618a73cca30, 0xbe1ceedbdb520545
+        .quad 0xc086261b6d2c8668, 0xbe1cf0b030396011
+        .quad 0xc086261e32267e98, 0xbe1cf19917010e96
+        .quad 0xc0862620f62b5cb0, 0xbe1cf07331355985
+        .quad 0xc0862623b93bc9e8, 0xbe1cf01ae921a1c3
+        .quad 0xc08626267b586ed0, 0xbe1cefe5cf0dbf0c
+        .quad 0xc08626293c81f348, 0xbe1cf01b258aeb50
+        .quad 0xc086262bfcb8fe88, 0xbe1cee6b9e7f4c68
+        .quad 0xc086262ebbfe3710, 0xbe1cee684a9b21c9
+        .quad 0xc08626317a5242b8, 0xbe1cf1f8bcde9a8b
+        .quad 0xc086263437b5c6c0, 0xbe1cf1d063d36238
+        .quad 0xc0862636f42967a8, 0xbe1cf1e31a19075e
+        .quad 0xc0862639afadc950, 0xbe1cf1d8efdf7e7d
+        .quad 0xc086263c6a438ef0, 0xbe1cf1812ee72dba
+        .quad 0xc086263f23eb5b18, 0xbe1cf1449a9a2279
+        .quad 0xc0862641dca5cfb8, 0xbe1cee96edce5085
+        .quad 0xc086264494738e08, 0xbe1cf06797bd03b2
+        .quad 0xc08626474b5536b8, 0xbe1cef91b9b7ffc1
+        .quad 0xc086264a014b69c0, 0xbe1cef4b6721278f
+        .quad 0xc086264cb656c678, 0xbe1cf1942925eb4a
+        .quad 0xc086264f6a77eba8, 0xbe1cefa2c7bc2e39
+        .quad 0xc08626521daf7758, 0xbe1cf252595aceb3
+        .quad 0xc0862654cffe0718, 0xbe1cee8e9ae47ec2
+        .quad 0xc0862657816437a8, 0xbe1cf1bf913828fa
+        .quad 0xc086265a31e2a558, 0xbe1cf23475d6b366
+        .quad 0xc086265ce179ebc8, 0xbe1cef8df00a922b
+        .quad 0xc086265f902aa5f0, 0xbe1cef279bfa43e0
+        .quad 0xc08626623df56e38, 0xbe1cf080e10b8365
+        .quad 0xc0862664eadade70, 0xbe1cf1a518f9b544
+        .quad 0xc086266796db8fd0, 0xbe1cef9308fed9e9
+        .quad 0xc086266a41f81ae8, 0xbe1ceea3ae6b19c9
+        .quad 0xc086266cec3117b8, 0xbe1ceef06003d4c2
+        .quad 0xc086266f95871da8, 0xbe1cf0b8457ffb0c
+        .quad 0xc08626723dfac390, 0xbe1cf0c526745ad6
+        .quad 0xc0862674e58c9fa8, 0xbe1cf0cf91ff7b5d
+        .quad 0xc08626778c3d4798, 0xbe1cefe260819380
+        .quad 0xc086267a320d5070, 0xbe1ceebd90aa27a3
+        .quad 0xc086267cd6fd4ea8, 0xbe1cf0388121dffa
+        .quad 0xc086267f7b0dd630, 0xbe1cf1a3881435f1
+        .quad 0xc08626821e3f7a68, 0xbe1cef28e9d9ac52
+        .quad 0xc0862684c092ce08, 0xbe1cf02d300062dd
+        .quad 0xc086268762086350, 0xbe1cefaee1edfa35
+        .quad 0xc086268a02a0cbe0, 0xbe1cf0a5a052e936
+        .quad 0xc086268ca25c98d8, 0xbe1cee60a4a497ed
+        .quad 0xc086268f413c5ab0, 0xbe1cf0e4a5d0cf49
+        .quad 0xc0862691df40a170, 0xbe1cf149235a4e6e
+        .quad 0xc08626947c69fc80, 0xbe1cf215180b9fcc
+        .quad 0xc086269718b8fac8, 0xbe1cef9b156a9840
+        .quad 0xc0862699b42e2a90, 0xbe1cf054c91441be
+        .quad 0xc086269c4eca19a8, 0xbe1cf13ded26512c
+        .quad 0xc086269ee88d5550, 0xbe1cf22ea4d8ac06
+        .quad 0xc08626a181786a40, 0xbe1cf2354666ee2e
+        .quad 0xc08626a4198be4a8, 0xbe1cefef936752b3
+        .quad 0xc08626a6b0c85020, 0xbe1cf1e360a9db68
+        .quad 0xc08626a9472e37d8, 0xbe1ceed6aeb812c5
+        .quad 0xc08626abdcbe2650, 0xbe1cf227340b4986
+        .quad 0xc08626ae7178a5b0, 0xbe1cf0215a0cbe0d
+        .quad 0xc08626b1055e3f70, 0xbe1cf256adf0ae26
+        .quad 0xc08626b3986f7ca8, 0xbe1ceff3c67aed06
+        .quad 0xc08626b62aace5c8, 0xbe1cf2159fb93652
+        .quad 0xc08626b8bc1702e0, 0xbe1cf01e6dbd1c7f
+        .quad 0xc08626bb4cae5b60, 0xbe1cf009e75d1c0c
+        .quad 0xc08626bddc737648, 0xbe1ceec10a020e73
+        .quad 0xc08626c06b66da08, 0xbe1cf06d5783eee7
+        .quad 0xc08626c2f9890ca0, 0xbe1cf0cb8f169ffe
+        .quad 0xc08626c586da9388, 0xbe1cef7de2452430
+        .quad 0xc08626c8135bf3b0, 0xbe1cf05da6f783ae
+        .quad 0xc08626ca9f0db198, 0xbe1cefcc877d681d
+        .quad 0xc08626cd29f05138, 0xbe1cef0531954ab3
+        .quad 0xc08626cfb4045608, 0xbe1cf06b8565ea3d
+        .quad 0xc08626d23d4a4310, 0xbe1cefdc455d9d7e
+        .quad 0xc08626d4c5c29ad0, 0xbe1ceefc47e8fa64
+        .quad 0xc08626d74d6ddf48, 0xbe1cf1872bf033f2
+        .quad 0xc08626d9d44c9210, 0xbe1cf19d91087f9d
+        .quad 0xc08626dc5a5f3438, 0xbe1cf012d444c6ab
+        .quad 0xc08626dedfa64650, 0xbe1cf0ba528ee153
+        .quad 0xc08626e164224880, 0xbe1ceeb431709788
+        .quad 0xc08626e3e7d3ba60, 0xbe1cf0b9af31a6a5
+        .quad 0xc08626e66abb1b28, 0xbe1cf168fb2e135b
+        .quad 0xc08626e8ecd8e990, 0xbe1cef9097461c93
+        .quad 0xc08626eb6e2da3d0, 0xbe1cee7a434735d8
+        .quad 0xc08626edeeb9c7a8, 0xbe1cf235732b86f2
+        .quad 0xc08626f06e7dd280, 0xbe1cefe1510b89e6
+        .quad 0xc08626f2ed7a4120, 0xbe1cf1f64b9b80ef
+        .quad 0xc08626f56baf9000, 0xbe1cf08f320ca339
+        .quad 0xc08626f7e91e3b08, 0xbe1cf1b1de2808a1
+        .quad 0xc08626fa65c6bdc0, 0xbe1cf1976d778b28
+        .quad 0xc08626fce1a99338, 0xbe1ceef40a4f076f
+        .quad 0xc08626ff5cc73600, 0xbe1cef3e45869ce3
+        .quad 0xc0862701d7202048, 0xbe1ceef601b4c9d6
+        .quad 0xc086270450b4cbc0, 0xbe1cf1eaf0b57fd6
+        .quad 0xc0862706c985b1c0, 0xbe1cef82a44990f3
+        .quad 0xc086270941934b10, 0xbe1ceefe32981f2c
+        .quad 0xc086270bb8de1018, 0xbe1cefbf6f5a0445
+        .quad 0xc086270e2f6678d0, 0xbe1cf18dba75792c
+        .quad 0xc0862710a52cfcc8, 0xbe1cf0da64ce995f
+        .quad 0xc08627131a321318, 0xbe1cef04ac0fb802
+        .quad 0xc08627158e763268, 0xbe1cee9d4e2ad9bd
+        .quad 0xc086271801f9d0f8, 0xbe1cefa9b55407b5
+        .quad 0xc086271a74bd64a0, 0xbe1cefe6bd329570
+        .quad 0xc086271ce6c162c8, 0xbe1cef0b1205dc85
+        .quad 0xc086271f58064068, 0xbe1cef092a785e3f
+        .quad 0xc0862721c88c7210, 0xbe1cf050dcdaac30
+        .quad 0xc086272438546be8, 0xbe1cf210907ded8b
+        .quad 0xc0862726a75ea1b8, 0xbe1cee760be44f99
+        .quad 0xc086272915ab86c0, 0xbe1ceeeee07c2bcc
+        .quad 0xc086272b833b8df0, 0xbe1cf06874992df5
+        .quad 0xc086272df00f29d0, 0xbe1cef8fac5d4899
+        .quad 0xc08627305c26cc70, 0xbe1cf1103241cc99
+        .quad 0xc0862732c782e788, 0xbe1cf1d35fef83fe
+        .quad 0xc08627353223ec68, 0xbe1cef3ec8133e1d
+        .quad 0xc08627379c0a4be8, 0xbe1cef7261daccd8
+        .quad 0xc086273a05367688, 0xbe1cf18656c50806
+        .quad 0xc086273c6da8dc68, 0xbe1cf1c8736e049a
+        .quad 0xc086273ed561ed38, 0xbe1cf1f93bff4911
+        .quad 0xc08627413c621848, 0xbe1cf188a4ea680c
+        .quad 0xc0862743a2a9cc80, 0xbe1cf1d270930c80
+        .quad 0xc086274608397868, 0xbe1cf25a328c28e2
+        .quad 0xc08627486d118a28, 0xbe1cf106f90aa3b8
+        .quad 0xc086274ad1326f80, 0xbe1cee5e9d2e885a
+        .quad 0xc086274d349c95c0, 0xbe1cf1c0bac27228
+        .quad 0xc086274f975069f8, 0xbe1cf1a1500f9b1c
+        .quad 0xc0862751f94e58c0, 0xbe1cefc30663ac44
+        .quad 0xc08627545a96ce48, 0xbe1cf17123e427a2
+        .quad 0xc0862756bb2a3678, 0xbe1cefb92749fea4
+        .quad 0xc08627591b08fcc0, 0xbe1cefa40e1ea74a
+        .quad 0xc086275b7a338c40, 0xbe1cee6f4612c3e9
+        .quad 0xc086275dd8aa4fa8, 0xbe1cf1c54a053627
+        .quad 0xc0862760366db168, 0xbe1ceff5eb503d9e
+        .quad 0xc0862762937e1b70, 0xbe1cf02e47f10cee
+        .quad 0xc0862764efdbf768, 0xbe1ceeb06e1d0dad
+        .quad 0xc08627674b87ae88, 0xbe1cf10aadd6dba5
+        .quad 0xc0862769a681a9c0, 0xbe1cf24e9913d30f
+        .quad 0xc086276c00ca51a0, 0xbe1cef47b301e312
+        .quad 0xc086276e5a620e48, 0xbe1ceeb1cefc2e85
+        .quad 0xc0862770b3494788, 0xbe1cf16f1fbbe011
+        .quad 0xc08627730b8064e8, 0xbe1ceebdf75174c7
+        .quad 0xc08627756307cd70, 0xbe1cf06e3871a0da
+        .quad 0xc0862777b9dfe7f0, 0xbe1cef16799fd554
+        .quad 0xc086277a10091ac0, 0xbe1cf248dabf5377
+        .quad 0xc086277c6583cc00, 0xbe1cf0c78d92a2cd
+        .quad 0xc086277eba506158, 0xbe1cf0b911b029f0
+        .quad 0xc08627810e6f4028, 0xbe1cefdc24719766
+        .quad 0xc086278361e0cd70, 0xbe1cefbb6562b7e7
+        .quad 0xc0862785b4a56dd8, 0xbe1cf1e0afb349ec
+        .quad 0xc086278806bd85c0, 0xbe1cf008292e52fc
+        .quad 0xc086278a58297918, 0xbe1cf053073872bf
+        .quad 0xc086278ca8e9ab88, 0xbe1cf17a0a55a947
+        .quad 0xc086278ef8fe8068, 0xbe1ceeffb0b60234
+        .quad 0xc086279148685aa0, 0xbe1cf162204794a8
+        .quad 0xc086279397279ce0, 0xbe1cf24cc8cb48ac
+        .quad 0xc0862795e53ca978, 0xbe1cf0c9be68d5c3
+        .quad 0xc086279832a7e258, 0xbe1cf172cd3d7388
+        .quad 0xc086279a7f69a930, 0xbe1ceea2465fbce5
+        .quad 0xc086279ccb825f40, 0xbe1cf0a386d2500f
+        .quad 0xc086279f16f26590, 0xbe1cf1e338ddc18a
+        .quad 0xc08627a161ba1cd0, 0xbe1cef1f5049867f
+        .quad 0xc08627a3abd9e548, 0xbe1cef96c1ea8b1f
+        .quad 0xc08627a5f5521f00, 0xbe1cf138f6fd3c26
+        .quad 0xc08627a83e2329b0, 0xbe1cf0d4fcbfdf3a
+        .quad 0xc08627aa864d64b0, 0xbe1cf24870c12c81
+        .quad 0xc08627accdd12f18, 0xbe1cf0ae2a56348d
+        .quad 0xc08627af14aee7a0, 0xbe1cee8ca1a9b893
+        .quad 0xc08627b15ae6eca8, 0xbe1cf20414d637b0
+        .quad 0xc08627b3a0799c60, 0xbe1cf0fc6b7b12d8
+        .quad 0xc08627b5e5675488, 0xbe1cf152d93c4a00
+        .quad 0xc08627b829b072a0, 0xbe1cf1073f9b77c2
+        .quad 0xc08627ba6d5553d8, 0xbe1cee694f97d5a4
+        .quad 0xc08627bcb0565500, 0xbe1cf0456b8239d7
+        .quad 0xc08627bef2b3d2b0, 0xbe1cf211497127e3
+        .quad 0xc08627c1346e2930, 0xbe1cf01856c0384d
+        .quad 0xc08627c37585b468, 0xbe1cefa7dd05479e
+        .quad 0xc08627c5b5fad000, 0xbe1cef3ae8e50b93
+        .quad 0xc08627c7f5cdd750, 0xbe1ceea5f32fdd3a
+        .quad 0xc08627ca34ff2560, 0xbe1cef424caeb8d9
+        .quad 0xc08627cc738f14f0, 0xbe1cf0194d07a81f
+        .quad 0xc08627ceb17e0070, 0xbe1cf20f452000c1
+        .quad 0xc08627d0eecc4210, 0xbe1cf00e356218e4
+        .quad 0xc08627d32b7a33a0, 0xbe1cef30484b4bcb
+        .quad 0xc08627d567882eb0, 0xbe1ceeea11a6641b
+        .quad 0xc08627d7a2f68c80, 0xbe1cf13492d5bd7b
+        .quad 0xc08627d9ddc5a618, 0xbe1ceeb7048fad96
+        .quad 0xc08627dc17f5d418, 0xbe1ceef0666f0477
+        .quad 0xc08627de51876ee8, 0xbe1cf060d4b8b5c2
+        .quad 0xc08627e08a7acea8, 0xbe1cf0b2a4b6ff8c
+        .quad 0xc08627e2c2d04b28, 0xbe1cf0e34809a875
+        .quad 0xc08627e4fa883bf0, 0xbe1cf16bf74a3522
+        .quad 0xc08627e731a2f848, 0xbe1cee6a24623d57
+        .quad 0xc08627e96820d718, 0xbe1cefc7b4f1528e
+        .quad 0xc08627eb9e022f18, 0xbe1cf163051f3548
+        .quad 0xc08627edd34756b8, 0xbe1cef36b3366305
+        .quad 0xc08627f007f0a408, 0xbe1cf18134625550
+        .quad 0xc08627f23bfe6cf0, 0xbe1cf0ec32ec1a11
+        .quad 0xc08627f46f710700, 0xbe1ceeb3b64f3edc
+        .quad 0xc08627f6a248c778, 0xbe1cf0cd15805bc8
+        .quad 0xc08627f8d4860368, 0xbe1cf20db3bddebe
+        .quad 0xc08627fb06290f90, 0xbe1cf25188430e25
+        .quad 0xc08627fd37324070, 0xbe1ceea1713490f9
+        .quad 0xc08627ff67a1ea28, 0xbe1cf159521d234c
+        .quad 0xc0862801977860b8, 0xbe1cf24dfe50783b
+        .quad 0xc0862803c6b5f7d0, 0xbe1ceef2ef89a60b
+        .quad 0xc0862805f55b02c8, 0xbe1cee7fc919d62c
+        .quad 0xc08628082367d4c0, 0xbe1cf215a7fb513a
+        .quad 0xc086280a50dcc0a8, 0xbe1cf0e4401c5ed4
+        .quad 0xc086280c7dba1910, 0xbe1cf04ec734d256
+        .quad 0xc086280eaa003050, 0xbe1cf010ad787fea
+        .quad 0xc0862810d5af5880, 0xbe1cee622478393d
+        .quad 0xc086281300c7e368, 0xbe1cf01c7482564f
+        .quad 0xc08628152b4a22a0, 0xbe1cf0de20d33536
+        .quad 0xc086281755366778, 0xbe1cef2edae5837d
+        .quad 0xc08628197e8d02f0, 0xbe1cf0a345318cc9
+        .quad 0xc086281ba74e45d8, 0xbe1cf20085aa34b8
+        .quad 0xc086281dcf7a80c0, 0xbe1cef5fa845ad83
+        .quad 0xc086281ff71203e0, 0xbe1cf050d1df69c4
+        .quad 0xc08628221e151f48, 0xbe1ceffe43c035b9
+        .quad 0xc0862824448422b8, 0xbe1cf14f3018d3c2
+        .quad 0xc08628266a5f5dc0, 0xbe1cef0a5fbae83d
+        .quad 0xc08628288fa71f98, 0xbe1ceff8a95b72a1
+        .quad 0xc086282ab45bb750, 0xbe1cef073aa9849b
+        .quad 0xc086282cd87d73a8, 0xbe1cef69b3835c02
+        .quad 0xc086282efc0ca328, 0xbe1cf0bc139379a9
+        .quad 0xc08628311f099420, 0xbe1cef247a9ec596
+        .quad 0xc086283341749490, 0xbe1cef74bbcc488a
+        .quad 0xc0862835634df248, 0xbe1cef4bc42e7b8e
+        .quad 0xc08628378495fad0, 0xbe1cf136d4d5a810
+        .quad 0xc0862839a54cfb80, 0xbe1cf0d290b24dd8
+        .quad 0xc086283bc5734168, 0xbe1ceeebde8e0065
+        .quad 0xc086283de5091950, 0xbe1cf1a09f60aa1e
+        .quad 0xc0862840040ecfe0, 0xbe1cf0803947a234
+        .quad 0xc08628422284b168, 0xbe1cf0abf7638127
+        .quad 0xc0862844406b0a08, 0xbe1cf0f73ee12058
+        .quad 0xc08628465dc225a0, 0xbe1cf2079971b26c
+        .quad 0xc08628487a8a4fe0, 0xbe1cee74957564b1
+        .quad 0xc086284a96c3d420, 0xbe1ceee77c1b7d43
+        .quad 0xc086284cb26efd90, 0xbe1cf23addba6e09
+        .quad 0xc086284ecd8c1730, 0xbe1cf199f4a1da60
+        .quad 0xc0862850e81b6bb0, 0xbe1cf09fdea81393
+        .quad 0xc0862853021d4588, 0xbe1cf176adb417f7
+        .quad 0xc08628551b91ef00, 0xbe1cf0f64f84a8da
+        .quad 0xc08628573479b220, 0xbe1ceec34cf49523
+        .quad 0xc08628594cd4d8a8, 0xbe1cf16d60fbe0bb
+        .quad 0xc086285b64a3ac40, 0xbe1cee8de7acfc7b
+        .quad 0xc086285d7be67630, 0xbe1ceee6256cce8d
+        .quad 0xc086285f929d7fa0, 0xbe1cee7d66a3d8a5
+        .quad 0xc0862861a8c91170, 0xbe1cf0bef8265792
+        .quad 0xc0862863be697458, 0xbe1cf097f890c6f8
+        .quad 0xc0862865d37ef0c8, 0xbe1cf09502d5c3fc
+        .quad 0xc0862867e809cf00, 0xbe1ceeffb239dac7
+        .quad 0xc0862869fc0a56f8, 0xbe1cf1fbfff95c98
+        .quad 0xc086286c0f80d090, 0xbe1cefa57ad3eef7
+        .quad 0xc086286e226d8348, 0xbe1cf22c58b9183d
+        .quad 0xc086287034d0b690, 0xbe1ceff262d0a248
+        .quad 0xc086287246aab180, 0xbe1cefa7bc194186
+        .quad 0xc086287457fbbb08, 0xbe1cf06782d784d9
+        .quad 0xc086287668c419e0, 0xbe1cf1d44d0eaa07
+        .quad 0xc086287879041490, 0xbe1cf034803c8a48
+        .quad 0xc086287a88bbf158, 0xbe1cf08e84916b6f
+        .quad 0xc086287c97ebf650, 0xbe1cf0c4d3dc1bc7
+        .quad 0xc086287ea6946958, 0xbe1cefb1e4625943
+        .quad 0xc0862880b4b59010, 0xbe1cf143efdd1fd0
+        .quad 0xc0862882c24faff8, 0xbe1cee9896d016da
+        .quad 0xc0862884cf630e38, 0xbe1cf2186072f2cc
+        .quad 0xc0862886dbefeff0, 0xbe1cef9217633d34
+        .quad 0xc0862888e7f699e0, 0xbe1cf05603549486
+        .quad 0xc086288af37750b0, 0xbe1cef50fff513d3
+        .quad 0xc086288cfe7258c0, 0xbe1cf127713b32d0
+        .quad 0xc086288f08e7f650, 0xbe1cf05015520f3d
+        .quad 0xc086289112d86d58, 0xbe1cf12eb458b26f
+        .quad 0xc08628931c4401a8, 0xbe1cf22eae2887ed
+        .quad 0xc0862895252af6e0, 0xbe1cefdd6656dd2d
+        .quad 0xc08628972d8d9058, 0xbe1cf1048ea4e646
+        .quad 0xc0862899356c1150, 0xbe1ceec4501167e9
+        .quad 0xc086289b3cc6bcb8, 0xbe1cf0ad52becc3f
+        .quad 0xc086289d439dd568, 0xbe1cf0daa4e00e35
+        .quad 0xc086289f49f19df8, 0xbe1cf00b80de8d6a
+        .quad 0xc08628a14fc258c8, 0xbe1cf1bcf2ea8464
+        .quad 0xc08628a355104818, 0xbe1cf0435e2782b0
+        .quad 0xc08628a559dbade0, 0xbe1cf0e3e1a5f56c
+        .quad 0xc08628a75e24cbf8, 0xbe1cefed9d5a721d
+        .quad 0xc08628a961ebe3f8, 0xbe1cf0d2d74321e2
+        .quad 0xc08628ab65313750, 0xbe1cf24200eb55e9
+        .quad 0xc08628ad67f50740, 0xbe1cf23e9d7cf979
+        .quad 0xc08628af6a3794d0, 0xbe1cf23a088f421c
+        .quad 0xc08628b16bf920e0, 0xbe1cef2c1de1ab32
+        .quad 0xc08628b36d39ec08, 0xbe1cf1abc231f7b2
+        .quad 0xc08628b56dfa36d0, 0xbe1cf2074d5ba303
+        .quad 0xc08628b76e3a4180, 0xbe1cf05cd5eed880
+        /*== Log_LA_table ==*/
+        .align 16
+        .quad 0x8000000000000000
+        .quad 0xbf5ff802a9ab10e6
+        .quad 0xbf6ff00aa2b10bc0
+        .quad 0xbf77ee11ebd82e94
+        .quad 0xbf7fe02a6b106789
+        .quad 0xbf83e7295d25a7d9
+        .quad 0xbf87dc475f810a77
+        .quad 0xbf8bcf712c74384c
+        .quad 0xbf8fc0a8b0fc03e4
+        .quad 0xbf91d7f7eb9eebe7
+        .quad 0xbf93cea44346a575
+        .quad 0xbf95c45a51b8d389
+        .quad 0xbf97b91b07d5b11b
+        .quad 0xbf99ace7551cc514
+        .quad 0xbf9b9fc027af9198
+        .quad 0xbf9d91a66c543cc4
+        .quad 0xbf9f829b0e783300
+        .quad 0xbfa0b94f7c196176
+        .quad 0xbfa1b0d98923d980
+        .quad 0xbfa2a7ec2214e873
+        .quad 0xbfa39e87b9febd60
+        .quad 0xbfa494acc34d911c
+        .quad 0xbfa58a5bafc8e4d5
+        .quad 0xbfa67f94f094bd98
+        .quad 0xbfa77458f632dcfc
+        .quad 0xbfa868a83083f6cf
+        .quad 0xbfa95c830ec8e3eb
+        .quad 0xbfaa4fe9ffa3d235
+        .quad 0xbfab42dd711971bf
+        .quad 0xbfac355dd0921f2d
+        .quad 0xbfad276b8adb0b52
+        .quad 0xbfae19070c276016
+        .quad 0xbfaf0a30c01162a6
+        .quad 0xbfaffae9119b9303
+        .quad 0xbfb075983598e471
+        .quad 0xbfb0ed839b5526fe
+        .quad 0xbfb16536eea37ae1
+        .quad 0xbfb1dcb263db1944
+        .quad 0xbfb253f62f0a1417
+        .quad 0xbfb2cb0283f5de1f
+        .quad 0xbfb341d7961bd1d1
+        .quad 0xbfb3b87598b1b6ee
+        .quad 0xbfb42edcbea646f0
+        .quad 0xbfb4a50d3aa1b040
+        .quad 0xbfb51b073f06183f
+        .quad 0xbfb590cafdf01c28
+        .quad 0xbfb60658a93750c4
+        .quad 0xbfb67bb0726ec0fc
+        .quad 0xbfb6f0d28ae56b4c
+        .quad 0xbfb765bf23a6be13
+        .quad 0xbfb7da766d7b12cd
+        .quad 0xbfb84ef898e8282a
+        .quad 0xbfb8c345d6319b21
+        .quad 0xbfb9375e55595ede
+        .quad 0xbfb9ab42462033ad
+        .quad 0xbfba1ef1d8061cd4
+        .quad 0xbfba926d3a4ad563
+        .quad 0xbfbb05b49bee43fe
+        .quad 0xbfbb78c82bb0eda1
+        .quad 0xbfbbeba818146765
+        .quad 0xbfbc5e548f5bc743
+        .quad 0xbfbcd0cdbf8c13e1
+        .quad 0xbfbd4313d66cb35d
+        .quad 0xbfbdb5270187d927
+        .quad 0xbfbe27076e2af2e6
+        .quad 0xbfbe98b549671467
+        .quad 0xbfbf0a30c01162a6
+        .quad 0xbfbf7b79fec37ddf
+        .quad 0xbfbfec9131dbeabb
+        .quad 0xbfc02ebb42bf3d4b
+        .quad 0xbfc0671512ca596e
+        .quad 0xbfc09f561ee719c3
+        .quad 0xbfc0d77e7cd08e59
+        .quad 0xbfc10f8e422539b1
+        .quad 0xbfc14785846742ac
+        .quad 0xbfc17f6458fca611
+        .quad 0xbfc1b72ad52f67a0
+        .quad 0xbfc1eed90e2dc2c3
+        .quad 0xbfc2266f190a5acb
+        .quad 0xbfc25ded0abc6ad2
+        .quad 0xbfc29552f81ff523
+        .quad 0xbfc2cca0f5f5f251
+        .quad 0xbfc303d718e47fd3
+        .quad 0xbfc33af575770e4f
+        .quad 0xbfc371fc201e8f74
+        .quad 0xbfc3a8eb2d31a376
+        .quad 0xbfc3dfc2b0ecc62a
+        .quad 0xbfc41682bf727bc0
+        .quad 0xbfc44d2b6ccb7d1e
+        .quad 0xbfc483bccce6e3dd
+        .quad 0xbfc4ba36f39a55e5
+        .quad 0xbfc4f099f4a230b2
+        .quad 0xbfc526e5e3a1b438
+        .quad 0xbfc55d1ad4232d6f
+        .quad 0xbfc59338d9982086
+        .quad 0xbfc5c940075972b9
+        .quad 0xbfc5ff3070a793d4
+        .quad 0xbfc6350a28aaa758
+        .quad 0xbfc66acd4272ad51
+        .quad 0xbfc6a079d0f7aad2
+        .quad 0xbfc6d60fe719d21d
+        .quad 0xbfc70b8f97a1aa75
+        .quad 0xbfc740f8f54037a5
+        .quad 0xbfc7764c128f2127
+        .quad 0xbfc7ab890210d909
+        .quad 0xbfc7e0afd630c274
+        .quad 0xbfc815c0a14357eb
+        .quad 0xbfc84abb75865139
+        .quad 0xbfc87fa06520c911
+        .quad 0xbfc8b46f8223625b
+        .quad 0xbfc8e928de886d41
+        .quad 0xbfc91dcc8c340bde
+        .quad 0xbfc9525a9cf456b4
+        .quad 0xbfc986d3228180ca
+        .quad 0xbfc9bb362e7dfb83
+        .quad 0xbfc9ef83d2769a34
+        .quad 0xbfca23bc1fe2b563
+        .quad 0xbfca57df28244dcd
+        .quad 0xbfca8becfc882f19
+        .quad 0xbfcabfe5ae46124c
+        .quad 0xbfcaf3c94e80bff3
+        .quad 0xbfcb2797ee46320c
+        .quad 0xbfcb5b519e8fb5a4
+        .quad 0xbfcb8ef670420c3b
+        .quad 0xbfcbc286742d8cd6
+        .quad 0xbfcbf601bb0e44e2
+        .quad 0xbfcc2968558c18c1
+        .quad 0xbfcc5cba543ae425
+        .quad 0xbfcc8ff7c79a9a22
+        .quad 0xbfccc320c0176502
+        .quad 0xbfccf6354e09c5dc
+        .quad 0xbfcd293581b6b3e7
+        .quad 0xbfcd5c216b4fbb91
+        .quad 0xbfcd8ef91af31d5e
+        .quad 0xbfcdc1bca0abec7d
+        .quad 0xbfcdf46c0c722d2f
+        .quad 0xbfce27076e2af2e6
+        .quad 0xbfce598ed5a87e2f
+        .quad 0xbfce8c0252aa5a60
+        .quad 0xbfcebe61f4dd7b0b
+        .quad 0xbfcef0adcbdc5936
+        .quad 0xbfcf22e5e72f105d
+        .quad 0xbfcf550a564b7b37
+        .quad 0xbfcf871b28955045
+        .quad 0xbfcfb9186d5e3e2b
+        .quad 0xbfcfeb0233e607cc
+        .quad 0xbfd00e6c45ad501d
+        .quad 0xbfd0274dc16c232f
+        .quad 0xbfd0402594b4d041
+        .quad 0xbfd058f3c703ebc6
+        .quad 0xbfd071b85fcd590d
+        .quad 0xbfd08a73667c57af
+        .quad 0xbfd0a324e27390e3
+        .quad 0xbfd0bbccdb0d24bd
+        .quad 0xbfd0d46b579ab74b
+        .quad 0xbfd0ed005f657da4
+        .quad 0xbfd1058bf9ae4ad5
+        .quad 0xbfd11e0e2dad9cb7
+        .quad 0xbfd136870293a8b0
+        .quad 0xbfd14ef67f88685a
+        .quad 0xbfd1675cababa60e
+        .quad 0xbfd17fb98e15095d
+        .quad 0xbfd1980d2dd4236f
+        .quad 0xbfd1b05791f07b49
+        .quad 0xbfd1c898c16999fb
+        .quad 0xbfd1e0d0c33716be
+        .quad 0xbfd1f8ff9e48a2f3
+        .quad 0xbfd211255986160c
+        .quad 0xbfd22941fbcf7966
+        .quad 0xbfd241558bfd1404
+        .quad 0xbfd2596010df763a
+        .quad 0xbfd27161913f853d
+        .quad 0xbfd2895a13de86a3
+        .quad 0xbfd2a1499f762bc9
+        .quad 0xbfd2b9303ab89d25
+        .quad 0xbfd2d10dec508583
+        .quad 0xbfd2e8e2bae11d31
+        .quad 0xbfd300aead06350c
+        .quad 0xbfd31871c9544185
+        .quad 0xbfd3302c16586588
+        .quad 0xbfd347dd9a987d55
+        .quad 0xbfd35f865c93293e
+        .quad 0xbfd3772662bfd85b
+        .quad 0xbfd38ebdb38ed321
+        .quad 0xbfd3a64c556945ea
+        .quad 0xbfd3bdd24eb14b6a
+        .quad 0xbfd3d54fa5c1f710
+        .quad 0xbfd3ecc460ef5f50
+        .quad 0xbfd404308686a7e4
+        .quad 0xbfd41b941cce0bee
+        .quad 0xbfd432ef2a04e814
+        .quad 0xbfd44a41b463c47c
+        .quad 0xbfd4618bc21c5ec2
+        .quad 0xbfd478cd5959b3d9
+        .quad 0xbfd49006804009d1
+        .quad 0xbfd4a7373cecf997
+        .quad 0xbfd4be5f957778a1
+        .quad 0xbfd4d57f8fefe27f
+        .quad 0xbfd4ec973260026a
+        .quad 0xbfd503a682cb1cb3
+        .quad 0xbfd51aad872df82d
+        .quad 0xbfd531ac457ee77e
+        .quad 0xbfd548a2c3add263
+        .quad 0xbfd55f9107a43ee2
+        .quad 0xbfd5767717455a6c
+        .quad 0xbfd58d54f86e02f2
+        .quad 0xbfd5a42ab0f4cfe2
+        .quad 0xbfd5baf846aa1b19
+        .quad 0xbfd5d1bdbf5809ca
+        .quad 0xbfd5e87b20c2954a
+        .quad 0xbfd5ff3070a793d4
+        .quad 0xbfd615ddb4bec13c
+        .quad 0xbfd62c82f2b9c795
+        .quad 0x3fd61965cdb02c1f
+        .quad 0x3fd602d08af091ec
+        .quad 0x3fd5ec433d5c35ae
+        .quad 0x3fd5d5bddf595f30
+        .quad 0x3fd5bf406b543db2
+        .quad 0x3fd5a8cadbbedfa1
+        .quad 0x3fd5925d2b112a59
+        .quad 0x3fd57bf753c8d1fb
+        .quad 0x3fd565995069514c
+        .quad 0x3fd54f431b7be1a9
+        .quad 0x3fd538f4af8f72fe
+        .quad 0x3fd522ae0738a3d8
+        .quad 0x3fd50c6f1d11b97c
+        .quad 0x3fd4f637ebba9810
+        .quad 0x3fd4e0086dd8baca
+        .quad 0x3fd4c9e09e172c3c
+        .quad 0x3fd4b3c077267e9a
+        .quad 0x3fd49da7f3bcc41f
+        .quad 0x3fd487970e958770
+        .quad 0x3fd4718dc271c41b
+        .quad 0x3fd45b8c0a17df13
+        .quad 0x3fd44591e0539f49
+        .quad 0x3fd42f9f3ff62642
+        .quad 0x3fd419b423d5e8c7
+        .quad 0x3fd403d086cea79c
+        .quad 0x3fd3edf463c1683e
+        .quad 0x3fd3d81fb5946dba
+        .quad 0x3fd3c25277333184
+        .quad 0x3fd3ac8ca38e5c5f
+        .quad 0x3fd396ce359bbf54
+        .quad 0x3fd3811728564cb2
+        .quad 0x3fd36b6776be1117
+        .quad 0x3fd355bf1bd82c8b
+        .quad 0x3fd3401e12aecba1
+        .quad 0x3fd32a84565120a8
+        .quad 0x3fd314f1e1d35ce4
+        .quad 0x3fd2ff66b04ea9d4
+        .quad 0x3fd2e9e2bce12286
+        .quad 0x3fd2d46602adccee
+        .quad 0x3fd2bef07cdc9354
+        .quad 0x3fd2a982269a3dbf
+        .quad 0x3fd2941afb186b7c
+        .quad 0x3fd27ebaf58d8c9d
+        .quad 0x3fd269621134db92
+        .quad 0x3fd25410494e56c7
+        .quad 0x3fd23ec5991eba49
+        .quad 0x3fd22981fbef797b
+        .quad 0x3fd214456d0eb8d4
+        .quad 0x3fd1ff0fe7cf47a7
+        .quad 0x3fd1e9e1678899f4
+        .quad 0x3fd1d4b9e796c245
+        .quad 0x3fd1bf99635a6b95
+        .quad 0x3fd1aa7fd638d33f
+        .quad 0x3fd1956d3b9bc2fa
+        .quad 0x3fd180618ef18adf
+        .quad 0x3fd16b5ccbacfb73
+        .quad 0x3fd1565eed455fc3
+        .quad 0x3fd14167ef367783
+        .quad 0x3fd12c77cd00713b
+        .quad 0x3fd1178e8227e47c
+        .quad 0x3fd102ac0a35cc1c
+        .quad 0x3fd0edd060b78081
+        .quad 0x3fd0d8fb813eb1ef
+        .quad 0x3fd0c42d676162e3
+        .quad 0x3fd0af660eb9e279
+        .quad 0x3fd09aa572e6c6d4
+        .quad 0x3fd085eb8f8ae797
+        .quad 0x3fd07138604d5862
+        .quad 0x3fd05c8be0d9635a
+        .quad 0x3fd047e60cde83b8
+        .quad 0x3fd03346e0106062
+        .quad 0x3fd01eae5626c691
+        .quad 0x3fd00a1c6adda473
+        .quad 0x3fcfeb2233ea07cd
+        .quad 0x3fcfc218be620a5e
+        .quad 0x3fcf991c6cb3b379
+        .quad 0x3fcf702d36777df0
+        .quad 0x3fcf474b134df229
+        .quad 0x3fcf1e75fadf9bde
+        .quad 0x3fcef5ade4dcffe6
+        .quad 0x3fceccf2c8fe920a
+        .quad 0x3fcea4449f04aaf5
+        .quad 0x3fce7ba35eb77e2a
+        .quad 0x3fce530effe71012
+        .quad 0x3fce2a877a6b2c12
+        .quad 0x3fce020cc6235ab5
+        .quad 0x3fcdd99edaf6d7e9
+        .quad 0x3fcdb13db0d48940
+        .quad 0x3fcd88e93fb2f450
+        .quad 0x3fcd60a17f903515
+        .quad 0x3fcd38666871f465
+        .quad 0x3fcd1037f2655e7b
+        .quad 0x3fcce816157f1988
+        .quad 0x3fccc000c9db3c52
+        .quad 0x3fcc97f8079d44ec
+        .quad 0x3fcc6ffbc6f00f71
+        .quad 0x3fcc480c0005ccd1
+        .quad 0x3fcc2028ab17f9b4
+        .quad 0x3fcbf851c067555f
+        .quad 0x3fcbd087383bd8ad
+        .quad 0x3fcba8c90ae4ad19
+        .quad 0x3fcb811730b823d2
+        .quad 0x3fcb5971a213acdb
+        .quad 0x3fcb31d8575bce3d
+        .quad 0x3fcb0a4b48fc1b46
+        .quad 0x3fcae2ca6f672bd4
+        .quad 0x3fcabb55c31693ad
+        .quad 0x3fca93ed3c8ad9e3
+        .quad 0x3fca6c90d44b704e
+        .quad 0x3fca454082e6ab05
+        .quad 0x3fca1dfc40f1b7f1
+        .quad 0x3fc9f6c407089664
+        .quad 0x3fc9cf97cdce0ec3
+        .quad 0x3fc9a8778debaa38
+        .quad 0x3fc981634011aa75
+        .quad 0x3fc95a5adcf7017f
+        .quad 0x3fc9335e5d594989
+        .quad 0x3fc90c6db9fcbcd9
+        .quad 0x3fc8e588ebac2dbf
+        .quad 0x3fc8beafeb38fe8c
+        .quad 0x3fc897e2b17b19a5
+        .quad 0x3fc871213750e994
+        .quad 0x3fc84a6b759f512f
+        .quad 0x3fc823c16551a3c2
+        .quad 0x3fc7fd22ff599d4f
+        .quad 0x3fc7d6903caf5ad0
+        .quad 0x3fc7b0091651528c
+        .quad 0x3fc7898d85444c73
+        .quad 0x3fc7631d82935a86
+        .quad 0x3fc73cb9074fd14d
+        .quad 0x3fc716600c914054
+        .quad 0x3fc6f0128b756abc
+        .quad 0x3fc6c9d07d203fc7
+        .quad 0x3fc6a399dabbd383
+        .quad 0x3fc67d6e9d785771
+        .quad 0x3fc6574ebe8c133a
+        .quad 0x3fc6313a37335d76
+        .quad 0x3fc60b3100b09476
+        .quad 0x3fc5e533144c1719
+        .quad 0x3fc5bf406b543db2
+        .quad 0x3fc59958ff1d52f1
+        .quad 0x3fc5737cc9018cdd
+        .quad 0x3fc54dabc26105d2
+        .quad 0x3fc527e5e4a1b58d
+        .quad 0x3fc5022b292f6a45
+        .quad 0x3fc4dc7b897bc1c8
+        .quad 0x3fc4b6d6fefe22a4
+        .quad 0x3fc4913d8333b561
+        .quad 0x3fc46baf0f9f5db7
+        .quad 0x3fc4462b9dc9b3dc
+        .quad 0x3fc420b32740fdd4
+        .quad 0x3fc3fb45a59928cc
+        .quad 0x3fc3d5e3126bc27f
+        .quad 0x3fc3b08b6757f2a9
+        .quad 0x3fc38b3e9e027479
+        .quad 0x3fc365fcb0159016
+        .quad 0x3fc340c59741142e
+        .quad 0x3fc31b994d3a4f85
+        .quad 0x3fc2f677cbbc0a96
+        .quad 0x3fc2d1610c86813a
+        .quad 0x3fc2ac55095f5c59
+        .quad 0x3fc28753bc11aba5
+        .quad 0x3fc2625d1e6ddf57
+        .quad 0x3fc23d712a49c202
+        .quad 0x3fc2188fd9807263
+        .quad 0x3fc1f3b925f25d41
+        .quad 0x3fc1ceed09853752
+        .quad 0x3fc1aa2b7e23f72a
+        .quad 0x3fc185747dbecf34
+        .quad 0x3fc160c8024b27b1
+        .quad 0x3fc13c2605c398c3
+        .quad 0x3fc1178e8227e47c
+        .quad 0x3fc0f301717cf0fb
+        .quad 0x3fc0ce7ecdccc28d
+        .quad 0x3fc0aa06912675d5
+        .quad 0x3fc08598b59e3a07
+        .quad 0x3fc06135354d4b18
+        .quad 0x3fc03cdc0a51ec0d
+        .quad 0x3fc0188d2ecf6140
+        .quad 0x3fbfe89139dbd566
+        .quad 0x3fbfa01c9db57ce2
+        .quad 0x3fbf57bc7d9005db
+        .quad 0x3fbf0f70cdd992e3
+        .quad 0x3fbec739830a1120
+        .quad 0x3fbe7f1691a32d3e
+        .quad 0x3fbe3707ee30487b
+        .quad 0x3fbdef0d8d466db9
+        .quad 0x3fbda727638446a2
+        .quad 0x3fbd5f55659210e2
+        .quad 0x3fbd179788219364
+        .quad 0x3fbccfedbfee13a8
+        .quad 0x3fbc885801bc4b23
+        .quad 0x3fbc40d6425a5cb1
+        .quad 0x3fbbf968769fca11
+        .quad 0x3fbbb20e936d6974
+        .quad 0x3fbb6ac88dad5b1c
+        .quad 0x3fbb23965a52ff00
+        .quad 0x3fbadc77ee5aea8c
+        .quad 0x3fba956d3ecade63
+        .quad 0x3fba4e7640b1bc38
+        .quad 0x3fba0792e9277cac
+        .quad 0x3fb9c0c32d4d2548
+        .quad 0x3fb97a07024cbe74
+        .quad 0x3fb9335e5d594989
+        .quad 0x3fb8ecc933aeb6e8
+        .quad 0x3fb8a6477a91dc29
+        .quad 0x3fb85fd927506a48
+        .quad 0x3fb8197e2f40e3f0
+        .quad 0x3fb7d33687c293c9
+        .quad 0x3fb78d02263d82d3
+        .quad 0x3fb746e100226ed9
+        .quad 0x3fb700d30aeac0e1
+        .quad 0x3fb6bad83c1883b6
+        .quad 0x3fb674f089365a7a
+        .quad 0x3fb62f1be7d77743
+        .quad 0x3fb5e95a4d9791cb
+        .quad 0x3fb5a3abb01ade25
+        .quad 0x3fb55e10050e0384
+        .quad 0x3fb518874226130a
+        .quad 0x3fb4d3115d207eac
+        .quad 0x3fb48dae4bc31018
+        .quad 0x3fb4485e03dbdfad
+        .quad 0x3fb403207b414b7f
+        .quad 0x3fb3bdf5a7d1ee64
+        .quad 0x3fb378dd7f749714
+        .quad 0x3fb333d7f8183f4b
+        .quad 0x3fb2eee507b40301
+        .quad 0x3fb2aa04a44717a5
+        .quad 0x3fb26536c3d8c369
+        .quad 0x3fb2207b5c78549e
+        .quad 0x3fb1dbd2643d190b
+        .quad 0x3fb1973bd1465567
+        .quad 0x3fb152b799bb3cc9
+        .quad 0x3fb10e45b3cae831
+        .quad 0x3fb0c9e615ac4e17
+        .quad 0x3fb08598b59e3a07
+        .quad 0x3fb0415d89e74444
+        .quad 0x3faffa6911ab9301
+        .quad 0x3faf723b517fc523
+        .quad 0x3faeea31c006b87c
+        .quad 0x3fae624c4a0b5e1b
+        .quad 0x3fadda8adc67ee4e
+        .quad 0x3fad52ed6405d86f
+        .quad 0x3faccb73cdddb2cc
+        .quad 0x3fac441e06f72a9e
+        .quad 0x3fabbcebfc68f420
+        .quad 0x3fab35dd9b58baad
+        .quad 0x3faaaef2d0fb10fc
+        .quad 0x3faa282b8a936171
+        .quad 0x3fa9a187b573de7c
+        .quad 0x3fa91b073efd7314
+        .quad 0x3fa894aa149fb343
+        .quad 0x3fa80e7023d8ccc4
+        .quad 0x3fa788595a3577ba
+        .quad 0x3fa70265a550e777
+        .quad 0x3fa67c94f2d4bb58
+        .quad 0x3fa5f6e73078efb8
+        .quad 0x3fa5715c4c03ceef
+        .quad 0x3fa4ebf43349e26f
+        .quad 0x3fa466aed42de3ea
+        .quad 0x3fa3e18c1ca0ae92
+        .quad 0x3fa35c8bfaa1306b
+        .quad 0x3fa2d7ae5c3c5bae
+        .quad 0x3fa252f32f8d183f
+        .quad 0x3fa1ce5a62bc353a
+        .quad 0x3fa149e3e4005a8d
+        .quad 0x3fa0c58fa19dfaaa
+        .quad 0x3fa0415d89e74444
+        .quad 0x3f9f7a9b16782856
+        .quad 0x3f9e72bf2813ce51
+        .quad 0x3f9d6b2725979802
+        .quad 0x3f9c63d2ec14aaf2
+        .quad 0x3f9b5cc258b718e6
+        .quad 0x3f9a55f548c5c43f
+        .quad 0x3f994f6b99a24475
+        .quad 0x3f98492528c8cabf
+        .quad 0x3f974321d3d006d3
+        .quad 0x3f963d6178690bd6
+        .quad 0x3f9537e3f45f3565
+        .quad 0x3f9432a925980cc1
+        .quad 0x3f932db0ea132e22
+        .quad 0x3f9228fb1fea2e28
+        .quad 0x3f912487a5507f70
+        .quad 0x3f90205658935847
+        .quad 0x3f8e38ce3033310c
+        .quad 0x3f8c317384c75f06
+        .quad 0x3f8a2a9c6c170462
+        .quad 0x3f882448a388a2aa
+        .quad 0x3f861e77e8b53fc6
+        .quad 0x3f841929f96832f0
+        .quad 0x3f82145e939ef1e9
+        .quad 0x3f8010157588de71
+        .quad 0x3f7c189cbb0e27fb
+        .quad 0x3f78121214586b54
+        .quad 0x3f740c8a747878e2
+        .quad 0x3f70080559588b35
+        .quad 0x3f680904828985c0
+        .quad 0x3f60040155d5889e
+        .quad 0x3f50020055655889
+        .quad 0x0000000000000000
+        /*== poly_coeff[4] ==*/
+        .align 16
+        .quad 0x3fc9999CACDB4D0A, 0x3fc9999CACDB4D0A /* coeff4 */
+        .quad 0xbfd0000148058EE1, 0xbfd0000148058EE1 /* coeff3 */
+        .quad 0x3fd55555555543C5, 0x3fd55555555543C5 /* coeff2 */
+        .quad 0xbfdFFFFFFFFFF81F, 0xbfdFFFFFFFFFF81F /* coeff1 */
+        /*== ExpMask ==*/
+        .align 16
+        .quad 0x000fffffffffffff, 0x000fffffffffffff
+        /*== Two10 ==*/
+        .align 16
+        .quad 0x3f50000000000000, 0x3f50000000000000
+        /*== MinLog1p = -1+2^(-53) ==*/
+        .align 16
+        .quad 0xbfefffffffffffff, 0xbfefffffffffffff
+        /*== MaxLog1p ==*/
+        .align 16
+        .quad 0x7f3ffffffffff000, 0x7f3ffffffffff000
+        /*== One ==*/
+        .align 16
+        .quad 0x3ff0000000000000, 0x3ff0000000000000
+        /*== SgnMask ==*/
+        .align 16
+        .quad 0x7fffffffffffffff, 0x7fffffffffffffff
+        /*== XThreshold ==*/
+        .align 16
+        .quad 0x3e00000000000000, 0x3e00000000000000
+        /*== XhMask ==*/
+        .align 16
+        .quad 0xfffffffffffffc00, 0xfffffffffffffc00
+        /*== Threshold ==*/
+        .align 16
+        .quad 0x4086a00000000000, 0x4086a00000000000
+        /*== Bias ==*/
+        .align 16
+        .quad 0x408ff80000000000, 0x408ff80000000000
+        /*== Bias1 ==*/
+        .align 16
+        .quad 0x408ff00000000000, 0x408ff00000000000
+        /*== ExpMask ==*/
+        .align 16
+        .quad 0x7ff0000000000000, 0x7ff0000000000000
+        /*== ExpMask2 ==*/
+        .align 16
+        .quad 0x7f40000000000000, 0x7f40000000000000
+        /*== L2L ==*/
+        .align 16
+        .quad 0x3fe62E42FEFA39EF, 0x3fe62E42FEFA39EF
+        /*== dBigThreshold ==*/
+        .align 16
+        .quad 0x41D0000000000000, 0x41D0000000000000
+        /*== dC2 ==*/
+        .align 16
+        .quad 0x3FD8000000000000, 0x3FD8000000000000
+        /*== dC3 ==*/
+        .align 16
+        .quad 0x3FD4000000000000, 0x3FD4000000000000
+        /*== dC4 ==*/
+        .align 16
+        .quad 0x3FD1800000000000, 0x3FD1800000000000
+        /*== dC5 ==*/
+        .align 16
+        .quad 0x3FCF800000000000, 0x3FCF800000000000
+        /*== dHalf ==*/
+        .align 16
+        .quad 0x3FE0000000000000, 0x3FE0000000000000
+        /*== dLargestFinite ==*/
+        .align 16
+        .quad 0x7FEFFFFFFFFFFFFF, 0x7FEFFFFFFFFFFFFF
+        /*== dLittleThreshold ==*/
+        .align 16
+        .quad 0x3F60000000000000, 0x3F60000000000000
+        /*== dSign ==*/
+        .align 16
+        .quad 0x8000000000000000, 0x8000000000000000
+        /*== dThirtyOne ==*/
+        .align 16
+        .quad 0x403F000000000000, 0x403F000000000000
+        /*== dTopMask12 ==*/
+        .align 16
+        .quad 0xFFFFFE0000000000, 0xFFFFFE0000000000
+        /*== dTopMask26 ==*/
+        .align 16
+        .quad 0xFFFFFFFFF8000000, 0xFFFFFFFFF8000000
+        /*== dTopMask29 ==*/
+        .align 16
+        .quad 0xFFFFFFFFFF000000, 0xFFFFFFFFFF000000
+        /*== XScale ==*/
+        .align 16
+        .quad 0x3E10000000000000, 0x3E10000000000000
+        .align 16
+        .type	__svml_dasinh_data_internal,@object
+        .size	__svml_dasinh_data_internal,.-__svml_dasinh_data_internal
+        .align 16
+
+.FLT_30:
+        .long	0x00000000,0x43380000,0x00000000,0x43380000
+        .type	.FLT_30,@object
+        .size	.FLT_30,16
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core-sse.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core-sse.S
new file mode 100644
index 0000000000..903b5f0fb5
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core-sse.S
@@ -0,0 +1,20 @@ 
+/* SSE version of vectorized asinh, vector length is 4.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#define _ZGVdN4v_asinh _ZGVdN4v_asinh_sse_wrapper
+#include "../svml_d_asinh4_core.S"
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core.c b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core.c
new file mode 100644
index 0000000000..e7acd032b5
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core.c
@@ -0,0 +1,27 @@ 
+/* Multiple versions of vectorized asinh, vector length is 4.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#define SYMBOL_NAME _ZGVdN4v_asinh
+#include "ifunc-mathvec-avx2.h"
+
+libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
+
+#ifdef SHARED
+__hidden_ver1 (_ZGVdN4v_asinh, __GI__ZGVdN4v_asinh, __redirect__ZGVdN4v_asinh)
+  __attribute__ ((visibility ("hidden")));
+#endif
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core_avx2.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core_avx2.S
new file mode 100644
index 0000000000..d691d1ec6f
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh4_core_avx2.S
@@ -0,0 +1,1601 @@ 
+/* Function asinh vectorized with AVX2.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   https://www.gnu.org/licenses/.  */
+
+/*
+ * ALGORITHM DESCRIPTION:
+ *
+ *   Compute asinh(x) as log(x + sqrt(x*x + 1))
+ *
+ *   Special cases:
+ *
+ *   asinh(NaN) = quiet NaN, and raise invalid exception
+ *   asinh(INF) = that INF
+ *   asinh(0)   = that 0
+ *
+ */
+
+/* Offsets for data table __svml_dasinh_data_internal
+ */
+#define Log_HA_table                  	0
+#define Log_LA_table                  	8224
+#define poly_coeff                    	12352
+#define ExpMask                       	12480
+#define Two10                         	12512
+#define MinLog1p                      	12544
+#define MaxLog1p                      	12576
+#define One                           	12608
+#define SgnMask                       	12640
+#define XThreshold                    	12672
+#define XhMask                        	12704
+#define Threshold                     	12736
+#define Bias                          	12768
+#define Bias1                         	12800
+#define ExpMask0                      	12832
+#define ExpMask2                      	12864
+#define L2                            	12896
+#define dBigThreshold                 	12928
+#define dC2                           	12960
+#define dC3                           	12992
+#define dC4                           	13024
+#define dC5                           	13056
+#define dHalf                         	13088
+#define dLargestFinite                	13120
+#define dLittleThreshold              	13152
+#define dSign                         	13184
+#define dThirtyOne                    	13216
+#define dTopMask12                    	13248
+#define dTopMask29                    	13280
+#define XScale                        	13312
+
+/* Lookup bias for data table __svml_dasinh_data_internal.  */
+#define Table_Lookup_Bias               -0x405fe0
+
+#include <sysdep.h>
+
+        .text
+	.section .text.avx2,"ax",@progbits
+ENTRY(_ZGVdN4v_asinh_avx2)
+        pushq     %rbp
+        cfi_def_cfa_offset(16)
+        movq      %rsp, %rbp
+        cfi_def_cfa(6, 16)
+        cfi_offset(6, -16)
+        andq      $-32, %rsp
+        subq      $96, %rsp
+        lea       Table_Lookup_Bias+__svml_dasinh_data_internal(%rip), %r8
+        vmovapd   %ymm0, %ymm13
+        vmovupd   SgnMask+__svml_dasinh_data_internal(%rip), %ymm9
+
+/* Load the constant 1 and a sign mask */
+        vmovupd   One+__svml_dasinh_data_internal(%rip), %ymm12
+
+/* No need to split X when FMA is available in hardware. */
+        vmulpd    %ymm13, %ymm13, %ymm8
+
+/*
+ * Get the absolute value of the input, since we will exploit antisymmetry
+ * and mostly assume X >= 0 in the core computation
+ */
+        vandpd    %ymm9, %ymm13, %ymm10
+
+/*
+ * Check whether the input is finite, by checking |X| <= MaxFloat
+ * Otherwise set the rangemask so that the callout will get used.
+ * Note that this will also use the callout for NaNs since not(NaN <= MaxFloat)
+ */
+        vcmpnle_uqpd dLargestFinite+__svml_dasinh_data_internal(%rip), %ymm10, %ymm14
+
+/*
+ * Finally, express Y + W = X^2 + 1 accurately where Y has <= 29 bits.
+ * If |X| <= 1 then |XHi| <= 1 and so |X2Hi| <= 1, so we can treat 1
+ * as the dominant component in the compensated summation. Otherwise,
+ * if |X| >= 1, then since X2Hi only has 52 significant bits, the basic
+ * addition will be exact anyway until we get to |X| >= 2^53. But by
+ * that time the log function is well-conditioned enough that the
+ * rounding error doesn't matter. Hence we can treat 1 as dominant even
+ * if it literally isn't.
+ */
+        vaddpd    %ymm8, %ymm12, %ymm5
+
+/*
+ * The following computation can go wrong for very large X, basically
+ * because X^2 overflows. But for large X we have
+ * asinh(X) / log(2 X) - 1 =~= 1/(4 * X^2), so for X >= 2^30
+ * we can just later stick X back into the log and tweak up the exponent.
+ * Actually we scale X by 2^-30 and tweak the exponent up by 31,
+ * to stay in the safe range for the later log computation.
+ * Compute a flag now telling us when do do this.
+ */
+        vcmplt_oqpd dBigThreshold+__svml_dasinh_data_internal(%rip), %ymm10, %ymm11
+        vsubpd    %ymm5, %ymm12, %ymm15
+        vmovmskpd %ymm14, %eax
+        vandpd    dTopMask29+__svml_dasinh_data_internal(%rip), %ymm5, %ymm14
+
+/*
+ * Compute R = 1/sqrt(Y + W) * (1 + d)
+ * Force R to <= 12 significant bits in case it isn't already
+ * This means that R * Y and R^2 * Y are exactly representable.
+ */
+        vcvtpd2ps %ymm14, %xmm1
+        vaddpd    %ymm15, %ymm8, %ymm0
+        vsubpd    %ymm14, %ymm5, %ymm2
+        vrsqrtps  %xmm1, %xmm3
+        vmovapd   %ymm13, %ymm7
+        vfmsub213pd %ymm8, %ymm13, %ymm7
+        vcvtps2pd %xmm3, %ymm6
+        vaddpd    %ymm0, %ymm7, %ymm4
+
+/*
+ * Unfortunately, we can still be in trouble if |X| <= 2^-10, since
+ * the absolute error 2^-(12+53)-ish in sqrt(1 + X^2) gets scaled up
+ * by 1/X and comes close to our threshold. Hence if |X| <= 2^-9,
+ * perform an alternative computation
+ * sqrt(1 + X^2) - 1 = X^2/2 - X^4/8 + X^6/16
+ * X2 = X^2
+ */
+        vaddpd    %ymm7, %ymm8, %ymm7
+        vaddpd    %ymm2, %ymm4, %ymm15
+
+/*
+ * Now       1 / (1 + d)
+ * = 1 / (1 + (sqrt(1 - e) - 1))
+ * = 1 / sqrt(1 - e)
+ * = 1 + 1/2 * e + 3/8 * e^2 + 5/16 * e^3 + 35/128 * e^4 +
+ * 63/256 * e^5 + 231/1024 * e^6 + ....
+ * So compute the first five nonconstant terms of that, so that
+ * we have a relative correction (1 + Corr) to apply to S etc.
+ * C1 = 1/2
+ * C2 = 3/8
+ * C3 = 5/16
+ * C4 = 35/128
+ * C5 = 63/256
+ */
+        vmovupd   dC5+__svml_dasinh_data_internal(%rip), %ymm4
+        vandpd    dTopMask12+__svml_dasinh_data_internal(%rip), %ymm6, %ymm0
+
+/*
+ * Compute S = (Y/sqrt(Y + W)) * (1 + d)
+ * and T = (W/sqrt(Y + W)) * (1 + d)
+ * so that S + T = sqrt(Y + W) * (1 + d)
+ * S is exact, and the rounding error in T is OK.
+ */
+        vmulpd    %ymm0, %ymm14, %ymm3
+        vmulpd    %ymm15, %ymm0, %ymm1
+        vmovupd   dHalf+__svml_dasinh_data_internal(%rip), %ymm6
+        vsubpd    %ymm12, %ymm3, %ymm14
+
+/*
+ * Obtain sqrt(1 + X^2) - 1 in two pieces
+ * sqrt(1 + X^2) - 1
+ * = sqrt(Y + W) - 1
+ * = (S + T) * (1 + Corr) - 1
+ * = [S - 1] + [T + (S + T) * Corr]
+ * We need a compensated summation for the last part. We treat S - 1
+ * as the larger part; it certainly is until about X < 2^-4, and in that
+ * case, the error is affordable since X dominates over sqrt(1 + X^2) - 1
+ * Final sum is dTmp5 (hi) + dTmp7 (lo)
+ */
+        vaddpd    %ymm1, %ymm3, %ymm2
+
+/*
+ * Compute e = -(2 * d + d^2)
+ * The first FMR is exact, and the rounding error in the other is acceptable
+ * since d and e are ~ 2^-12
+ */
+        vmovapd   %ymm12, %ymm5
+        vfnmadd231pd %ymm3, %ymm0, %ymm5
+        vfnmadd231pd %ymm1, %ymm0, %ymm5
+        vfmadd213pd dC4+__svml_dasinh_data_internal(%rip), %ymm5, %ymm4
+        vfmadd213pd dC3+__svml_dasinh_data_internal(%rip), %ymm5, %ymm4
+        vfmadd213pd dC2+__svml_dasinh_data_internal(%rip), %ymm5, %ymm4
+        vfmadd213pd %ymm6, %ymm5, %ymm4
+        vmulpd    %ymm4, %ymm5, %ymm0
+        vfmadd213pd %ymm1, %ymm2, %ymm0
+
+/* Now multiplex the two possible computations */
+        vcmple_oqpd dLittleThreshold+__svml_dasinh_data_internal(%rip), %ymm10, %ymm2
+        vaddpd    %ymm14, %ymm0, %ymm15
+
+/* dX2over2 = X^2/2 */
+        vmulpd    %ymm7, %ymm6, %ymm0
+
+/* dX4over4 = X^4/4 */
+        vmulpd    %ymm0, %ymm0, %ymm8
+
+/* dX46 = -X^4/4 + X^6/8 */
+        vfmsub231pd %ymm0, %ymm8, %ymm8
+
+/* dX46over2 = -X^4/8 + x^6/16 */
+        vmulpd    %ymm8, %ymm6, %ymm5
+
+/* 2^ (-10-exp(X) ) */
+        vmovupd   ExpMask2+__svml_dasinh_data_internal(%rip), %ymm8
+        vaddpd    %ymm5, %ymm0, %ymm4
+        vblendvpd %ymm2, %ymm4, %ymm15, %ymm1
+
+/*
+ * Now do another compensated sum to add |X| + [sqrt(1 + X^2) - 1].
+ * It's always safe to assume |X| is larger.
+ * This is the final 2-part argument to the log1p function
+ */
+        vaddpd    %ymm1, %ymm10, %ymm3
+
+/* Now multiplex to the case X = 2^-30 * |input|, Xl = dL = 0 in the "big" case. */
+        vmulpd    XScale+__svml_dasinh_data_internal(%rip), %ymm10, %ymm10
+
+/*
+ * Now we feed into the log1p code, using H in place of _VARG1 and
+ * also adding L into Xl.
+ * compute 1+x as high, low parts
+ */
+        vmaxpd    %ymm3, %ymm12, %ymm6
+        vminpd    %ymm3, %ymm12, %ymm7
+        vandpd    %ymm9, %ymm3, %ymm9
+        vcmplt_oqpd XThreshold+__svml_dasinh_data_internal(%rip), %ymm9, %ymm0
+        vaddpd    %ymm7, %ymm6, %ymm5
+        vorpd     XhMask+__svml_dasinh_data_internal(%rip), %ymm0, %ymm4
+        vandpd    %ymm4, %ymm5, %ymm1
+        vblendvpd %ymm11, %ymm1, %ymm10, %ymm5
+        vsubpd    %ymm1, %ymm6, %ymm2
+
+/* exponent bits */
+        vpsrlq    $20, %ymm5, %ymm10
+        vaddpd    %ymm2, %ymm7, %ymm3
+
+/*
+ * Now resume the main code.
+ * preserve mantissa, set input exponent to 2^(-10)
+ */
+        vandpd    ExpMask+__svml_dasinh_data_internal(%rip), %ymm5, %ymm0
+        vorpd     Two10+__svml_dasinh_data_internal(%rip), %ymm0, %ymm2
+
+/* reciprocal approximation good to at least 11 bits */
+        vcvtpd2ps %ymm2, %xmm6
+        vrcpps    %xmm6, %xmm7
+        vcvtps2pd %xmm7, %ymm15
+
+/* exponent of X needed to scale Xl */
+        vandps    ExpMask0+__svml_dasinh_data_internal(%rip), %ymm5, %ymm9
+        vpsubq    %ymm9, %ymm8, %ymm0
+        vandpd    %ymm11, %ymm3, %ymm4
+
+/* round reciprocal to nearest integer, will have 1+9 mantissa bits */
+        vroundpd  $0, %ymm15, %ymm3
+
+/* scale DblRcp */
+        vmulpd    %ymm0, %ymm3, %ymm2
+
+/* argument reduction */
+        vfmsub213pd %ymm12, %ymm2, %ymm5
+        vmulpd    %ymm2, %ymm4, %ymm12
+        vmovupd   poly_coeff+64+__svml_dasinh_data_internal(%rip), %ymm2
+        vaddpd    %ymm12, %ymm5, %ymm5
+        vfmadd213pd poly_coeff+96+__svml_dasinh_data_internal(%rip), %ymm5, %ymm2
+        vmulpd    %ymm5, %ymm5, %ymm4
+        vextractf128 $1, %ymm10, %xmm14
+        vshufps   $221, %xmm14, %xmm10, %xmm1
+
+/* biased exponent in DP format */
+        vcvtdq2pd %xmm1, %ymm7
+
+/* exponent*log(2.0) */
+        vmovupd   Threshold+__svml_dasinh_data_internal(%rip), %ymm10
+
+/* Add 31 to the exponent in the "large" case to get log(2 * input) */
+        vaddpd    dThirtyOne+__svml_dasinh_data_internal(%rip), %ymm7, %ymm6
+        vblendvpd %ymm11, %ymm7, %ymm6, %ymm1
+
+/*
+ * prepare table index
+ * table lookup
+ */
+        vpsrlq    $40, %ymm3, %ymm11
+        vcmplt_oqpd %ymm3, %ymm10, %ymm3
+        vandpd    Bias+__svml_dasinh_data_internal(%rip), %ymm3, %ymm14
+        vorpd     Bias1+__svml_dasinh_data_internal(%rip), %ymm14, %ymm15
+        vsubpd    %ymm15, %ymm1, %ymm1
+        vmulpd    L2+__svml_dasinh_data_internal(%rip), %ymm1, %ymm3
+
+/* polynomial */
+        vmovupd   poly_coeff+__svml_dasinh_data_internal(%rip), %ymm1
+        vfmadd213pd poly_coeff+32+__svml_dasinh_data_internal(%rip), %ymm5, %ymm1
+        vfmadd213pd %ymm2, %ymm4, %ymm1
+
+/* reconstruction */
+        vfmadd213pd %ymm5, %ymm4, %ymm1
+        vextractf128 $1, %ymm11, %xmm7
+        vmovd     %xmm11, %edx
+        vmovd     %xmm7, %esi
+        movslq    %edx, %rdx
+        vpextrd   $2, %xmm11, %ecx
+        movslq    %esi, %rsi
+        vpextrd   $2, %xmm7, %edi
+        movslq    %ecx, %rcx
+        movslq    %edi, %rdi
+        vmovsd    (%r8,%rdx), %xmm0
+        vmovsd    (%r8,%rsi), %xmm8
+        vmovhpd   (%r8,%rcx), %xmm0, %xmm6
+        vmovhpd   (%r8,%rdi), %xmm8, %xmm9
+        vinsertf128 $1, %xmm9, %ymm6, %ymm0
+        vaddpd    %ymm1, %ymm0, %ymm0
+        vaddpd    %ymm0, %ymm3, %ymm7
+
+/* Finally, reincorporate the original sign. */
+        vandpd    dSign+__svml_dasinh_data_internal(%rip), %ymm13, %ymm6
+        vxorpd    %ymm7, %ymm6, %ymm0
+        testl     %eax, %eax
+
+/* Go to special inputs processing branch */
+        jne       L(SPECIAL_VALUES_BRANCH)
+                                # LOE rbx r12 r13 r14 r15 eax ymm0 ymm13
+
+/* Restore registers
+ * and exit the function
+ */
+
+L(EXIT):
+        movq      %rbp, %rsp
+        popq      %rbp
+        cfi_def_cfa(7, 8)
+        cfi_restore(6)
+        ret
+        cfi_def_cfa(6, 16)
+        cfi_offset(6, -16)
+
+/* Branch to process
+ * special inputs
+ */
+
+L(SPECIAL_VALUES_BRANCH):
+        vmovupd   %ymm13, 32(%rsp)
+        vmovupd   %ymm0, 64(%rsp)
+                                # LOE rbx r12 r13 r14 r15 eax ymm0
+
+        xorl      %edx, %edx
+                                # LOE rbx r12 r13 r14 r15 eax edx
+
+        vzeroupper
+        movq      %r12, 16(%rsp)
+        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -80; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xb0, 0xff, 0xff, 0xff, 0x22
+        movl      %edx, %r12d
+        movq      %r13, 8(%rsp)
+        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -88; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa8, 0xff, 0xff, 0xff, 0x22
+        movl      %eax, %r13d
+        movq      %r14, (%rsp)
+        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -96; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa0, 0xff, 0xff, 0xff, 0x22
+                                # LOE rbx r15 r12d r13d
+
+/* Range mask
+ * bits check
+ */
+
+L(RANGEMASK_CHECK):
+        btl       %r12d, %r13d
+
+/* Call scalar math function */
+        jc        L(SCALAR_MATH_CALL)
+                                # LOE rbx r15 r12d r13d
+
+/* Special inputs
+ * processing loop
+ */
+
+L(SPECIAL_VALUES_LOOP):
+        incl      %r12d
+        cmpl      $4, %r12d
+
+/* Check bits in range mask */
+        jl        L(RANGEMASK_CHECK)
+                                # LOE rbx r15 r12d r13d
+
+        movq      16(%rsp), %r12
+        cfi_restore(12)
+        movq      8(%rsp), %r13
+        cfi_restore(13)
+        movq      (%rsp), %r14
+        cfi_restore(14)
+        vmovupd   64(%rsp), %ymm0
+
+/* Go to exit */
+        jmp       L(EXIT)
+        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -80; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xb0, 0xff, 0xff, 0xff, 0x22
+        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -88; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa8, 0xff, 0xff, 0xff, 0x22
+        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -96; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa0, 0xff, 0xff, 0xff, 0x22
+                                # LOE rbx r12 r13 r14 r15 ymm0
+
+/* Scalar math fucntion call
+ * to process special input
+ */
+
+L(SCALAR_MATH_CALL):
+        movl      %r12d, %r14d
+        movsd     32(%rsp,%r14,8), %xmm0
+        call      asinh@PLT
+                                # LOE rbx r14 r15 r12d r13d xmm0
+
+        movsd     %xmm0, 64(%rsp,%r14,8)
+
+/* Process special inputs in loop */
+        jmp       L(SPECIAL_VALUES_LOOP)
+                                # LOE rbx r15 r12d r13d
+END(_ZGVdN4v_asinh_avx2)
+
+        .section .rodata, "a"
+        .align 32
+
+#ifdef __svml_dasinh_data_internal_typedef
+typedef unsigned int VUINT32;
+typedef struct {
+        __declspec(align(32)) VUINT32 Log_HA_table[(1<<10)+2][2];
+        __declspec(align(32)) VUINT32 Log_LA_table[(1<<9)+1][2];
+        __declspec(align(32)) VUINT32 poly_coeff[4][4][2];
+        __declspec(align(32)) VUINT32 ExpMask[4][2];
+        __declspec(align(32)) VUINT32 Two10[4][2];
+        __declspec(align(32)) VUINT32 MinLog1p[4][2];
+        __declspec(align(32)) VUINT32 MaxLog1p[4][2];
+        __declspec(align(32)) VUINT32 One[4][2];
+        __declspec(align(32)) VUINT32 SgnMask[4][2];
+        __declspec(align(32)) VUINT32 XThreshold[4][2];
+        __declspec(align(32)) VUINT32 XhMask[4][2];
+        __declspec(align(32)) VUINT32 Threshold[4][2];
+        __declspec(align(32)) VUINT32 Bias[4][2];
+        __declspec(align(32)) VUINT32 Bias1[4][2];
+        __declspec(align(32)) VUINT32 ExpMask0[4][2];
+        __declspec(align(32)) VUINT32 ExpMask2[4][2];
+        __declspec(align(32)) VUINT32 L2[4][2];
+        __declspec(align(32)) VUINT32 dBigThreshold[4][2];
+        __declspec(align(32)) VUINT32 dC2[4][2];
+        __declspec(align(32)) VUINT32 dC3[4][2];
+        __declspec(align(32)) VUINT32 dC4[4][2];
+        __declspec(align(32)) VUINT32 dC5[4][2];
+        __declspec(align(32)) VUINT32 dHalf[4][2];
+        __declspec(align(32)) VUINT32 dLargestFinite[4][2];
+        __declspec(align(32)) VUINT32 dLittleThreshold[4][2];
+        __declspec(align(32)) VUINT32 dSign[4][2];
+        __declspec(align(32)) VUINT32 dThirtyOne[4][2];
+        __declspec(align(32)) VUINT32 dTopMask12[4][2];
+        __declspec(align(32)) VUINT32 dTopMask29[4][2];
+        __declspec(align(32)) VUINT32 XScale[4][2];
+} __svml_dasinh_data_internal;
+#endif
+__svml_dasinh_data_internal:
+        /* Log_HA_table */
+        .quad 0xc086232bdd7a8300, 0xbe1ce91eef3fb100
+        .quad 0xc086232fdc7ad828, 0xbe1cefcffda73b6a
+        .quad 0xc0862333d97d2ba0, 0xbe1cef406748f1ff
+        .quad 0xc0862337d48378e0, 0xbe1cef2a9429925a
+        .quad 0xc086233bcd8fb878, 0xbe1cf138d17ebecb
+        .quad 0xc086233fc4a3e018, 0xbe1ceff2dbbbb29e
+        .quad 0xc0862343b9c1e270, 0xbe1cf1a42aae437b
+        .quad 0xc0862347acebaf68, 0xbe1cef3b152048af
+        .quad 0xc086234b9e2333f0, 0xbe1cef20e127805e
+        .quad 0xc086234f8d6a5a30, 0xbe1cf00ad6052cf4
+        .quad 0xc08623537ac30980, 0xbe1cefc4642ee597
+        .quad 0xc0862357662f2660, 0xbe1cf1f277d36e16
+        .quad 0xc086235b4fb092a0, 0xbe1ceed009e8d8e6
+        .quad 0xc086235f37492d28, 0xbe1cf1e4038cb362
+        .quad 0xc08623631cfad250, 0xbe1cf0b0873b8557
+        .quad 0xc086236700c75b98, 0xbe1cf15bb3227c0b
+        .quad 0xc086236ae2b09fe0, 0xbe1cf151ef8ca9ed
+        .quad 0xc086236ec2b87358, 0xbe1cefe1dc2cd2ed
+        .quad 0xc0862372a0e0a780, 0xbe1cf0d1eec5454f
+        .quad 0xc08623767d2b0b48, 0xbe1ceeefd570bbce
+        .quad 0xc086237a57996af0, 0xbe1cee99ae91b3a7
+        .quad 0xc086237e302d9028, 0xbe1cf0412830fbd1
+        .quad 0xc086238206e94218, 0xbe1ceee898588610
+        .quad 0xc0862385dbce4548, 0xbe1cee9a1fbcaaea
+        .quad 0xc0862389aede5bc0, 0xbe1ceed8e7cc1ad6
+        .quad 0xc086238d801b4500, 0xbe1cf10c8d059da6
+        .quad 0xc08623914f86be18, 0xbe1ceee6c63a8165
+        .quad 0xc08623951d228180, 0xbe1cf0c3592d2ff1
+        .quad 0xc0862398e8f04758, 0xbe1cf0026cc4cb1b
+        .quad 0xc086239cb2f1c538, 0xbe1cf15d48d8e670
+        .quad 0xc08623a07b28ae60, 0xbe1cef359363787c
+        .quad 0xc08623a44196b390, 0xbe1cefdf1ab2e82c
+        .quad 0xc08623a8063d8338, 0xbe1cefe43c02aa84
+        .quad 0xc08623abc91ec960, 0xbe1cf044f5ae35b7
+        .quad 0xc08623af8a3c2fb8, 0xbe1cf0b0b4001e1b
+        .quad 0xc08623b349975d98, 0xbe1cf1bae76dfbcf
+        .quad 0xc08623b70731f810, 0xbe1cef0a72e13a62
+        .quad 0xc08623bac30da1c8, 0xbe1cf184007d2b6b
+        .quad 0xc08623be7d2bfb40, 0xbe1cf16f4b239e98
+        .quad 0xc08623c2358ea2a0, 0xbe1cf0976acada87
+        .quad 0xc08623c5ec3733d0, 0xbe1cf066318a16ff
+        .quad 0xc08623c9a1274880, 0xbe1ceffaa7148798
+        .quad 0xc08623cd54607820, 0xbe1cf23ab02e9b6e
+        .quad 0xc08623d105e45800, 0xbe1cefdfef7d4fde
+        .quad 0xc08623d4b5b47b20, 0xbe1cf17fece44f2b
+        .quad 0xc08623d863d27270, 0xbe1cf18f907d0d7c
+        .quad 0xc08623dc103fccb0, 0xbe1cee61fe072c98
+        .quad 0xc08623dfbafe1668, 0xbe1cf022dd891e2f
+        .quad 0xc08623e3640eda20, 0xbe1ceecc1daf4358
+        .quad 0xc08623e70b73a028, 0xbe1cf0173c4fa380
+        .quad 0xc08623eab12deec8, 0xbe1cf16a2150c2f4
+        .quad 0xc08623ee553f4a30, 0xbe1cf1bf980b1f4b
+        .quad 0xc08623f1f7a93480, 0xbe1cef8b731663c2
+        .quad 0xc08623f5986d2dc0, 0xbe1cee9a664d7ef4
+        .quad 0xc08623f9378cb3f0, 0xbe1cf1eda2af6400
+        .quad 0xc08623fcd5094320, 0xbe1cf1923f9d68d7
+        .quad 0xc086240070e45548, 0xbe1cf0747cd3e03a
+        .quad 0xc08624040b1f6260, 0xbe1cf22ee855bd6d
+        .quad 0xc0862407a3bbe078, 0xbe1cf0d57360c00b
+        .quad 0xc086240b3abb4398, 0xbe1ceebc815cd575
+        .quad 0xc086240ed01efdd0, 0xbe1cf03bfb970951
+        .quad 0xc086241263e87f50, 0xbe1cf16e74768529
+        .quad 0xc0862415f6193658, 0xbe1cefec64b8becb
+        .quad 0xc086241986b28f30, 0xbe1cf0838d210baa
+        .quad 0xc086241d15b5f448, 0xbe1cf0ea86e75b11
+        .quad 0xc0862420a324ce28, 0xbe1cf1708d11d805
+        .quad 0xc08624242f008380, 0xbe1ceea988c5a417
+        .quad 0xc0862427b94a7910, 0xbe1cef166a7bbca5
+        .quad 0xc086242b420411d0, 0xbe1cf0c9d9e86a38
+        .quad 0xc086242ec92eaee8, 0xbe1cef0946455411
+        .quad 0xc08624324ecbaf98, 0xbe1cefea60907739
+        .quad 0xc0862435d2dc7160, 0xbe1cf1ed0934ce42
+        .quad 0xc086243955624ff8, 0xbe1cf191ba746c7d
+        .quad 0xc086243cd65ea548, 0xbe1ceeec78cf2a7e
+        .quad 0xc086244055d2c968, 0xbe1cef345284c119
+        .quad 0xc0862443d3c012b8, 0xbe1cf24f77355219
+        .quad 0xc08624475027d5e8, 0xbe1cf05bf087e114
+        .quad 0xc086244acb0b65d0, 0xbe1cef3504a32189
+        .quad 0xc086244e446c1398, 0xbe1ceff54b2a406f
+        .quad 0xc0862451bc4b2eb8, 0xbe1cf0757d54ed4f
+        .quad 0xc086245532aa04f0, 0xbe1cf0c8099fdfd5
+        .quad 0xc0862458a789e250, 0xbe1cf0b173796a31
+        .quad 0xc086245c1aec1138, 0xbe1cf11d8734540d
+        .quad 0xc086245f8cd1da60, 0xbe1cf1916a723ceb
+        .quad 0xc0862462fd3c84d8, 0xbe1cf19a911e1da7
+        .quad 0xc08624666c2d5608, 0xbe1cf23a9ef72e4f
+        .quad 0xc0862469d9a591c0, 0xbe1cef503d947663
+        .quad 0xc086246d45a67a18, 0xbe1cf0fceeb1a0b2
+        .quad 0xc0862470b0314fa8, 0xbe1cf107e27e4fbc
+        .quad 0xc086247419475160, 0xbe1cf03dd9922331
+        .quad 0xc086247780e9bc98, 0xbe1cefce1a10e129
+        .quad 0xc086247ae719cd18, 0xbe1ceea47f73c4f6
+        .quad 0xc086247e4bd8bd10, 0xbe1ceec0ac56d100
+        .quad 0xc0862481af27c528, 0xbe1cee8a6593278a
+        .quad 0xc086248511081c70, 0xbe1cf2231dd9dec7
+        .quad 0xc0862488717af888, 0xbe1cf0b4b8ed7da8
+        .quad 0xc086248bd0818d68, 0xbe1cf1bd8d835002
+        .quad 0xc086248f2e1d0d98, 0xbe1cf259acc107f4
+        .quad 0xc08624928a4eaa20, 0xbe1cee897636b00c
+        .quad 0xc0862495e5179270, 0xbe1cee757f20c326
+        .quad 0xc08624993e78f490, 0xbe1cefafd3aa54a4
+        .quad 0xc086249c9673fd10, 0xbe1cee7298d38b97
+        .quad 0xc086249fed09d6f8, 0xbe1ceedc158d4ceb
+        .quad 0xc08624a3423babe0, 0xbe1cf2282987cb2e
+        .quad 0xc08624a6960aa400, 0xbe1cefe7381ecc4b
+        .quad 0xc08624a9e877e600, 0xbe1cef328dbbce80
+        .quad 0xc08624ad39849728, 0xbe1cefde45f3cc71
+        .quad 0xc08624b08931db58, 0xbe1cefa8b89433b9
+        .quad 0xc08624b3d780d500, 0xbe1cef6773c0b139
+        .quad 0xc08624b72472a528, 0xbe1cf031c931c11f
+        .quad 0xc08624ba70086b78, 0xbe1cf088f49275e7
+        .quad 0xc08624bdba434630, 0xbe1cf17de0eaa86d
+        .quad 0xc08624c103245238, 0xbe1cefd492f1ba75
+        .quad 0xc08624c44aacab08, 0xbe1cf1253e154466
+        .quad 0xc08624c790dd6ad0, 0xbe1cf0fb09ee6d55
+        .quad 0xc08624cad5b7aa58, 0xbe1cf1f08dd048fe
+        .quad 0xc08624ce193c8120, 0xbe1ceeca0809697f
+        .quad 0xc08624d15b6d0538, 0xbe1cef8d5662d968
+        .quad 0xc08624d49c4a4b78, 0xbe1cee97b556ed78
+        .quad 0xc08624d7dbd56750, 0xbe1cf1b14b6acb75
+        .quad 0xc08624db1a0f6b00, 0xbe1cef1e860623f2
+        .quad 0xc08624de56f96758, 0xbe1ceeaf4d156f3d
+        .quad 0xc08624e192946bf0, 0xbe1ceecc12b400ed
+        .quad 0xc08624e4cce18710, 0xbe1cf180c40c794f
+        .quad 0xc08624e805e1c5c8, 0xbe1cf185a08f7f65
+        .quad 0xc08624eb3d9633d8, 0xbe1cef45fc924078
+        .quad 0xc08624ee73ffdbb0, 0xbe1cf1e4f457f32a
+        .quad 0xc08624f1a91fc6a0, 0xbe1cf040147b8a5a
+        .quad 0xc08624f4dcf6fc98, 0xbe1cf1effca0dfb2
+        .quad 0xc08624f80f868468, 0xbe1cf0470146e5bc
+        .quad 0xc08624fb40cf6390, 0xbe1cef4dd186e501
+        .quad 0xc08624fe70d29e60, 0xbe1ceebe257f66c7
+        .quad 0xc08625019f9137f0, 0xbe1ceefb7a1c395c
+        .quad 0xc0862504cd0c3220, 0xbe1cf209dedfed8c
+        .quad 0xc0862507f9448db0, 0xbe1cf082da464994
+        .quad 0xc086250b243b4a18, 0xbe1cee88694a73cf
+        .quad 0xc086250e4df165a0, 0xbe1cf0b61e8f0531
+        .quad 0xc08625117667dd78, 0xbe1cf1106599c962
+        .quad 0xc08625149d9fad98, 0xbe1ceff1ee88af1f
+        .quad 0xc0862517c399d0c8, 0xbe1cf0f746994ef6
+        .quad 0xc086251ae85740b8, 0xbe1cefe8a1d077e4
+        .quad 0xc086251e0bd8f5e0, 0xbe1cf1a1da036092
+        .quad 0xc08625212e1fe7a8, 0xbe1cf0f8a7786fcd
+        .quad 0xc08625244f2d0c48, 0xbe1cefa1174a07a7
+        .quad 0xc08625276f0158d8, 0xbe1cef1043aa5b25
+        .quad 0xc086252a8d9dc150, 0xbe1cf15d521c169d
+        .quad 0xc086252dab033898, 0xbe1cf220bba8861f
+        .quad 0xc0862530c732b078, 0xbe1cef51e310eae2
+        .quad 0xc0862533e22d1988, 0xbe1cf222fcedd8ae
+        .quad 0xc0862536fbf36370, 0xbe1cefdb4da4bda8
+        .quad 0xc086253a14867ca0, 0xbe1ceeafc1112171
+        .quad 0xc086253d2be75280, 0xbe1cee99dfb4b408
+        .quad 0xc08625404216d160, 0xbe1cf22d2536f06b
+        .quad 0xc08625435715e498, 0xbe1cef6abbf2e268
+        .quad 0xc08625466ae57648, 0xbe1cf093a14789f5
+        .quad 0xc08625497d866fa0, 0xbe1cf0f93655603c
+        .quad 0xc086254c8ef9b8b8, 0xbe1cf1cc40c9aafc
+        .quad 0xc086254f9f4038a8, 0xbe1ceeea5f4e9157
+        .quad 0xc0862552ae5ad568, 0xbe1cefa9f52d4997
+        .quad 0xc0862555bc4a7400, 0xbe1cefa490a638ff
+        .quad 0xc0862558c90ff868, 0xbe1cef7fcf797d6f
+        .quad 0xc086255bd4ac4590, 0xbe1cf1b4c51113c9
+        .quad 0xc086255edf203d78, 0xbe1cef55e5b4a55d
+        .quad 0xc0862561e86cc100, 0xbe1cf0d37a25f9dc
+        .quad 0xc0862564f092b028, 0xbe1ceebe9efc19d9
+        .quad 0xc0862567f792e9d8, 0xbe1cee8ad30a57b5
+        .quad 0xc086256afd6e4c08, 0xbe1cef4e1817b90b
+        .quad 0xc086256e0225b3b8, 0xbe1cee7fa9229996
+        .quad 0xc086257105b9fce0, 0xbe1cf0b54963d945
+        .quad 0xc0862574082c0298, 0xbe1cee5f2f3c7995
+        .quad 0xc0862577097c9ee0, 0xbe1cf0828e303a2c
+        .quad 0xc086257a09acaae0, 0xbe1cf172c3078947
+        .quad 0xc086257d08bcfec0, 0xbe1cf189252afa22
+        .quad 0xc086258006ae71b8, 0xbe1cefdb80426923
+        .quad 0xc08625830381da08, 0xbe1ceef1391a0372
+        .quad 0xc0862585ff380d00, 0xbe1cf17720c78d13
+        .quad 0xc0862588f9d1df18, 0xbe1ceef1f9027d83
+        .quad 0xc086258bf35023b8, 0xbe1cf06fac99dec9
+        .quad 0xc086258eebb3ad78, 0xbe1cf1373eeb45c0
+        .quad 0xc0862591e2fd4e00, 0xbe1cef777536bb81
+        .quad 0xc0862594d92dd600, 0xbe1cf0f43ca40766
+        .quad 0xc0862597ce461558, 0xbe1cefb2cfc6766b
+        .quad 0xc086259ac246daf0, 0xbe1ceea49e64ffa2
+        .quad 0xc086259db530f4c8, 0xbe1cf250fa457dec
+        .quad 0xc08625a0a7053018, 0xbe1cf17d8bb2a44e
+        .quad 0xc08625a397c45918, 0xbe1cf1d5906d54b7
+        .quad 0xc08625a6876f3b30, 0xbe1cf08fe7b31780
+        .quad 0xc08625a97606a0e0, 0xbe1cef13edfc9d11
+        .quad 0xc08625ac638b53c8, 0xbe1cef9d2b107219
+        .quad 0xc08625af4ffe1cb0, 0xbe1cf1ddd4ff6160
+        .quad 0xc08625b23b5fc390, 0xbe1cefa02a996495
+        .quad 0xc08625b525b10f68, 0xbe1cf166a7e37ee5
+        .quad 0xc08625b80ef2c680, 0xbe1cef0b171068a5
+        .quad 0xc08625baf725ae28, 0xbe1cf05c80779283
+        .quad 0xc08625bdde4a8af0, 0xbe1cf1bbfbffb889
+        .quad 0xc08625c0c4622090, 0xbe1cf0b8666c0124
+        .quad 0xc08625c3a96d31e0, 0xbe1cf0a8fcf47a86
+        .quad 0xc08625c68d6c80f0, 0xbe1cef46e18cb092
+        .quad 0xc08625c97060cef0, 0xbe1cf1458a350efb
+        .quad 0xc08625cc524adc58, 0xbe1ceeea1dadce12
+        .quad 0xc08625cf332b68b0, 0xbe1cf0a1bfdc44c7
+        .quad 0xc08625d2130332d0, 0xbe1cef96d02da73e
+        .quad 0xc08625d4f1d2f8a8, 0xbe1cf2451c3c7701
+        .quad 0xc08625d7cf9b7778, 0xbe1cf10d08f83812
+        .quad 0xc08625daac5d6ba0, 0xbe1ceec5b4895c5e
+        .quad 0xc08625dd881990b0, 0xbe1cf14e1325c5e4
+        .quad 0xc08625e062d0a188, 0xbe1cf21d0904be12
+        .quad 0xc08625e33c835838, 0xbe1ceed0839bcf21
+        .quad 0xc08625e615326df0, 0xbe1cf1bb944889d2
+        .quad 0xc08625e8ecde9b48, 0xbe1cee738e85eece
+        .quad 0xc08625ebc38897e0, 0xbe1cf25c2bc6ef12
+        .quad 0xc08625ee99311ac8, 0xbe1cf132b70a41ad
+        .quad 0xc08625f16dd8da28, 0xbe1cf1984236a6e3
+        .quad 0xc08625f441808b78, 0xbe1cf19ae74998f9
+        .quad 0xc08625f71428e370, 0xbe1cef3e175d61a1
+        .quad 0xc08625f9e5d295f8, 0xbe1cf101f9868fd9
+        .quad 0xc08625fcb67e5658, 0xbe1cee69db83dcd2
+        .quad 0xc08625ff862cd6f8, 0xbe1cf081b636af51
+        .quad 0xc086260254dec9a8, 0xbe1cee62c7d59b3e
+        .quad 0xc08626052294df58, 0xbe1cf1b745c57716
+        .quad 0xc0862607ef4fc868, 0xbe1cef3d2800ea23
+        .quad 0xc086260abb103458, 0xbe1cef480ff1acd2
+        .quad 0xc086260d85d6d200, 0xbe1cf2424c9a17ef
+        .quad 0xc08626104fa44f90, 0xbe1cf12cfde90fd5
+        .quad 0xc086261318795a68, 0xbe1cf21f590dd5b6
+        .quad 0xc0862615e0569f48, 0xbe1cf0c50f9cd28a
+        .quad 0xc0862618a73cca30, 0xbe1ceedbdb520545
+        .quad 0xc086261b6d2c8668, 0xbe1cf0b030396011
+        .quad 0xc086261e32267e98, 0xbe1cf19917010e96
+        .quad 0xc0862620f62b5cb0, 0xbe1cf07331355985
+        .quad 0xc0862623b93bc9e8, 0xbe1cf01ae921a1c3
+        .quad 0xc08626267b586ed0, 0xbe1cefe5cf0dbf0c
+        .quad 0xc08626293c81f348, 0xbe1cf01b258aeb50
+        .quad 0xc086262bfcb8fe88, 0xbe1cee6b9e7f4c68
+        .quad 0xc086262ebbfe3710, 0xbe1cee684a9b21c9
+        .quad 0xc08626317a5242b8, 0xbe1cf1f8bcde9a8b
+        .quad 0xc086263437b5c6c0, 0xbe1cf1d063d36238
+        .quad 0xc0862636f42967a8, 0xbe1cf1e31a19075e
+        .quad 0xc0862639afadc950, 0xbe1cf1d8efdf7e7d
+        .quad 0xc086263c6a438ef0, 0xbe1cf1812ee72dba
+        .quad 0xc086263f23eb5b18, 0xbe1cf1449a9a2279
+        .quad 0xc0862641dca5cfb8, 0xbe1cee96edce5085
+        .quad 0xc086264494738e08, 0xbe1cf06797bd03b2
+        .quad 0xc08626474b5536b8, 0xbe1cef91b9b7ffc1
+        .quad 0xc086264a014b69c0, 0xbe1cef4b6721278f
+        .quad 0xc086264cb656c678, 0xbe1cf1942925eb4a
+        .quad 0xc086264f6a77eba8, 0xbe1cefa2c7bc2e39
+        .quad 0xc08626521daf7758, 0xbe1cf252595aceb3
+        .quad 0xc0862654cffe0718, 0xbe1cee8e9ae47ec2
+        .quad 0xc0862657816437a8, 0xbe1cf1bf913828fa
+        .quad 0xc086265a31e2a558, 0xbe1cf23475d6b366
+        .quad 0xc086265ce179ebc8, 0xbe1cef8df00a922b
+        .quad 0xc086265f902aa5f0, 0xbe1cef279bfa43e0
+        .quad 0xc08626623df56e38, 0xbe1cf080e10b8365
+        .quad 0xc0862664eadade70, 0xbe1cf1a518f9b544
+        .quad 0xc086266796db8fd0, 0xbe1cef9308fed9e9
+        .quad 0xc086266a41f81ae8, 0xbe1ceea3ae6b19c9
+        .quad 0xc086266cec3117b8, 0xbe1ceef06003d4c2
+        .quad 0xc086266f95871da8, 0xbe1cf0b8457ffb0c
+        .quad 0xc08626723dfac390, 0xbe1cf0c526745ad6
+        .quad 0xc0862674e58c9fa8, 0xbe1cf0cf91ff7b5d
+        .quad 0xc08626778c3d4798, 0xbe1cefe260819380
+        .quad 0xc086267a320d5070, 0xbe1ceebd90aa27a3
+        .quad 0xc086267cd6fd4ea8, 0xbe1cf0388121dffa
+        .quad 0xc086267f7b0dd630, 0xbe1cf1a3881435f1
+        .quad 0xc08626821e3f7a68, 0xbe1cef28e9d9ac52
+        .quad 0xc0862684c092ce08, 0xbe1cf02d300062dd
+        .quad 0xc086268762086350, 0xbe1cefaee1edfa35
+        .quad 0xc086268a02a0cbe0, 0xbe1cf0a5a052e936
+        .quad 0xc086268ca25c98d8, 0xbe1cee60a4a497ed
+        .quad 0xc086268f413c5ab0, 0xbe1cf0e4a5d0cf49
+        .quad 0xc0862691df40a170, 0xbe1cf149235a4e6e
+        .quad 0xc08626947c69fc80, 0xbe1cf215180b9fcc
+        .quad 0xc086269718b8fac8, 0xbe1cef9b156a9840
+        .quad 0xc0862699b42e2a90, 0xbe1cf054c91441be
+        .quad 0xc086269c4eca19a8, 0xbe1cf13ded26512c
+        .quad 0xc086269ee88d5550, 0xbe1cf22ea4d8ac06
+        .quad 0xc08626a181786a40, 0xbe1cf2354666ee2e
+        .quad 0xc08626a4198be4a8, 0xbe1cefef936752b3
+        .quad 0xc08626a6b0c85020, 0xbe1cf1e360a9db68
+        .quad 0xc08626a9472e37d8, 0xbe1ceed6aeb812c5
+        .quad 0xc08626abdcbe2650, 0xbe1cf227340b4986
+        .quad 0xc08626ae7178a5b0, 0xbe1cf0215a0cbe0d
+        .quad 0xc08626b1055e3f70, 0xbe1cf256adf0ae26
+        .quad 0xc08626b3986f7ca8, 0xbe1ceff3c67aed06
+        .quad 0xc08626b62aace5c8, 0xbe1cf2159fb93652
+        .quad 0xc08626b8bc1702e0, 0xbe1cf01e6dbd1c7f
+        .quad 0xc08626bb4cae5b60, 0xbe1cf009e75d1c0c
+        .quad 0xc08626bddc737648, 0xbe1ceec10a020e73
+        .quad 0xc08626c06b66da08, 0xbe1cf06d5783eee7
+        .quad 0xc08626c2f9890ca0, 0xbe1cf0cb8f169ffe
+        .quad 0xc08626c586da9388, 0xbe1cef7de2452430
+        .quad 0xc08626c8135bf3b0, 0xbe1cf05da6f783ae
+        .quad 0xc08626ca9f0db198, 0xbe1cefcc877d681d
+        .quad 0xc08626cd29f05138, 0xbe1cef0531954ab3
+        .quad 0xc08626cfb4045608, 0xbe1cf06b8565ea3d
+        .quad 0xc08626d23d4a4310, 0xbe1cefdc455d9d7e
+        .quad 0xc08626d4c5c29ad0, 0xbe1ceefc47e8fa64
+        .quad 0xc08626d74d6ddf48, 0xbe1cf1872bf033f2
+        .quad 0xc08626d9d44c9210, 0xbe1cf19d91087f9d
+        .quad 0xc08626dc5a5f3438, 0xbe1cf012d444c6ab
+        .quad 0xc08626dedfa64650, 0xbe1cf0ba528ee153
+        .quad 0xc08626e164224880, 0xbe1ceeb431709788
+        .quad 0xc08626e3e7d3ba60, 0xbe1cf0b9af31a6a5
+        .quad 0xc08626e66abb1b28, 0xbe1cf168fb2e135b
+        .quad 0xc08626e8ecd8e990, 0xbe1cef9097461c93
+        .quad 0xc08626eb6e2da3d0, 0xbe1cee7a434735d8
+        .quad 0xc08626edeeb9c7a8, 0xbe1cf235732b86f2
+        .quad 0xc08626f06e7dd280, 0xbe1cefe1510b89e6
+        .quad 0xc08626f2ed7a4120, 0xbe1cf1f64b9b80ef
+        .quad 0xc08626f56baf9000, 0xbe1cf08f320ca339
+        .quad 0xc08626f7e91e3b08, 0xbe1cf1b1de2808a1
+        .quad 0xc08626fa65c6bdc0, 0xbe1cf1976d778b28
+        .quad 0xc08626fce1a99338, 0xbe1ceef40a4f076f
+        .quad 0xc08626ff5cc73600, 0xbe1cef3e45869ce3
+        .quad 0xc0862701d7202048, 0xbe1ceef601b4c9d6
+        .quad 0xc086270450b4cbc0, 0xbe1cf1eaf0b57fd6
+        .quad 0xc0862706c985b1c0, 0xbe1cef82a44990f3
+        .quad 0xc086270941934b10, 0xbe1ceefe32981f2c
+        .quad 0xc086270bb8de1018, 0xbe1cefbf6f5a0445
+        .quad 0xc086270e2f6678d0, 0xbe1cf18dba75792c
+        .quad 0xc0862710a52cfcc8, 0xbe1cf0da64ce995f
+        .quad 0xc08627131a321318, 0xbe1cef04ac0fb802
+        .quad 0xc08627158e763268, 0xbe1cee9d4e2ad9bd
+        .quad 0xc086271801f9d0f8, 0xbe1cefa9b55407b5
+        .quad 0xc086271a74bd64a0, 0xbe1cefe6bd329570
+        .quad 0xc086271ce6c162c8, 0xbe1cef0b1205dc85
+        .quad 0xc086271f58064068, 0xbe1cef092a785e3f
+        .quad 0xc0862721c88c7210, 0xbe1cf050dcdaac30
+        .quad 0xc086272438546be8, 0xbe1cf210907ded8b
+        .quad 0xc0862726a75ea1b8, 0xbe1cee760be44f99
+        .quad 0xc086272915ab86c0, 0xbe1ceeeee07c2bcc
+        .quad 0xc086272b833b8df0, 0xbe1cf06874992df5
+        .quad 0xc086272df00f29d0, 0xbe1cef8fac5d4899
+        .quad 0xc08627305c26cc70, 0xbe1cf1103241cc99
+        .quad 0xc0862732c782e788, 0xbe1cf1d35fef83fe
+        .quad 0xc08627353223ec68, 0xbe1cef3ec8133e1d
+        .quad 0xc08627379c0a4be8, 0xbe1cef7261daccd8
+        .quad 0xc086273a05367688, 0xbe1cf18656c50806
+        .quad 0xc086273c6da8dc68, 0xbe1cf1c8736e049a
+        .quad 0xc086273ed561ed38, 0xbe1cf1f93bff4911
+        .quad 0xc08627413c621848, 0xbe1cf188a4ea680c
+        .quad 0xc0862743a2a9cc80, 0xbe1cf1d270930c80
+        .quad 0xc086274608397868, 0xbe1cf25a328c28e2
+        .quad 0xc08627486d118a28, 0xbe1cf106f90aa3b8
+        .quad 0xc086274ad1326f80, 0xbe1cee5e9d2e885a
+        .quad 0xc086274d349c95c0, 0xbe1cf1c0bac27228
+        .quad 0xc086274f975069f8, 0xbe1cf1a1500f9b1c
+        .quad 0xc0862751f94e58c0, 0xbe1cefc30663ac44
+        .quad 0xc08627545a96ce48, 0xbe1cf17123e427a2
+        .quad 0xc0862756bb2a3678, 0xbe1cefb92749fea4
+        .quad 0xc08627591b08fcc0, 0xbe1cefa40e1ea74a
+        .quad 0xc086275b7a338c40, 0xbe1cee6f4612c3e9
+        .quad 0xc086275dd8aa4fa8, 0xbe1cf1c54a053627
+        .quad 0xc0862760366db168, 0xbe1ceff5eb503d9e
+        .quad 0xc0862762937e1b70, 0xbe1cf02e47f10cee
+        .quad 0xc0862764efdbf768, 0xbe1ceeb06e1d0dad
+        .quad 0xc08627674b87ae88, 0xbe1cf10aadd6dba5
+        .quad 0xc0862769a681a9c0, 0xbe1cf24e9913d30f
+        .quad 0xc086276c00ca51a0, 0xbe1cef47b301e312
+        .quad 0xc086276e5a620e48, 0xbe1ceeb1cefc2e85
+        .quad 0xc0862770b3494788, 0xbe1cf16f1fbbe011
+        .quad 0xc08627730b8064e8, 0xbe1ceebdf75174c7
+        .quad 0xc08627756307cd70, 0xbe1cf06e3871a0da
+        .quad 0xc0862777b9dfe7f0, 0xbe1cef16799fd554
+        .quad 0xc086277a10091ac0, 0xbe1cf248dabf5377
+        .quad 0xc086277c6583cc00, 0xbe1cf0c78d92a2cd
+        .quad 0xc086277eba506158, 0xbe1cf0b911b029f0
+        .quad 0xc08627810e6f4028, 0xbe1cefdc24719766
+        .quad 0xc086278361e0cd70, 0xbe1cefbb6562b7e7
+        .quad 0xc0862785b4a56dd8, 0xbe1cf1e0afb349ec
+        .quad 0xc086278806bd85c0, 0xbe1cf008292e52fc
+        .quad 0xc086278a58297918, 0xbe1cf053073872bf
+        .quad 0xc086278ca8e9ab88, 0xbe1cf17a0a55a947
+        .quad 0xc086278ef8fe8068, 0xbe1ceeffb0b60234
+        .quad 0xc086279148685aa0, 0xbe1cf162204794a8
+        .quad 0xc086279397279ce0, 0xbe1cf24cc8cb48ac
+        .quad 0xc0862795e53ca978, 0xbe1cf0c9be68d5c3
+        .quad 0xc086279832a7e258, 0xbe1cf172cd3d7388
+        .quad 0xc086279a7f69a930, 0xbe1ceea2465fbce5
+        .quad 0xc086279ccb825f40, 0xbe1cf0a386d2500f
+        .quad 0xc086279f16f26590, 0xbe1cf1e338ddc18a
+        .quad 0xc08627a161ba1cd0, 0xbe1cef1f5049867f
+        .quad 0xc08627a3abd9e548, 0xbe1cef96c1ea8b1f
+        .quad 0xc08627a5f5521f00, 0xbe1cf138f6fd3c26
+        .quad 0xc08627a83e2329b0, 0xbe1cf0d4fcbfdf3a
+        .quad 0xc08627aa864d64b0, 0xbe1cf24870c12c81
+        .quad 0xc08627accdd12f18, 0xbe1cf0ae2a56348d
+        .quad 0xc08627af14aee7a0, 0xbe1cee8ca1a9b893
+        .quad 0xc08627b15ae6eca8, 0xbe1cf20414d637b0
+        .quad 0xc08627b3a0799c60, 0xbe1cf0fc6b7b12d8
+        .quad 0xc08627b5e5675488, 0xbe1cf152d93c4a00
+        .quad 0xc08627b829b072a0, 0xbe1cf1073f9b77c2
+        .quad 0xc08627ba6d5553d8, 0xbe1cee694f97d5a4
+        .quad 0xc08627bcb0565500, 0xbe1cf0456b8239d7
+        .quad 0xc08627bef2b3d2b0, 0xbe1cf211497127e3
+        .quad 0xc08627c1346e2930, 0xbe1cf01856c0384d
+        .quad 0xc08627c37585b468, 0xbe1cefa7dd05479e
+        .quad 0xc08627c5b5fad000, 0xbe1cef3ae8e50b93
+        .quad 0xc08627c7f5cdd750, 0xbe1ceea5f32fdd3a
+        .quad 0xc08627ca34ff2560, 0xbe1cef424caeb8d9
+        .quad 0xc08627cc738f14f0, 0xbe1cf0194d07a81f
+        .quad 0xc08627ceb17e0070, 0xbe1cf20f452000c1
+        .quad 0xc08627d0eecc4210, 0xbe1cf00e356218e4
+        .quad 0xc08627d32b7a33a0, 0xbe1cef30484b4bcb
+        .quad 0xc08627d567882eb0, 0xbe1ceeea11a6641b
+        .quad 0xc08627d7a2f68c80, 0xbe1cf13492d5bd7b
+        .quad 0xc08627d9ddc5a618, 0xbe1ceeb7048fad96
+        .quad 0xc08627dc17f5d418, 0xbe1ceef0666f0477
+        .quad 0xc08627de51876ee8, 0xbe1cf060d4b8b5c2
+        .quad 0xc08627e08a7acea8, 0xbe1cf0b2a4b6ff8c
+        .quad 0xc08627e2c2d04b28, 0xbe1cf0e34809a875
+        .quad 0xc08627e4fa883bf0, 0xbe1cf16bf74a3522
+        .quad 0xc08627e731a2f848, 0xbe1cee6a24623d57
+        .quad 0xc08627e96820d718, 0xbe1cefc7b4f1528e
+        .quad 0xc08627eb9e022f18, 0xbe1cf163051f3548
+        .quad 0xc08627edd34756b8, 0xbe1cef36b3366305
+        .quad 0xc08627f007f0a408, 0xbe1cf18134625550
+        .quad 0xc08627f23bfe6cf0, 0xbe1cf0ec32ec1a11
+        .quad 0xc08627f46f710700, 0xbe1ceeb3b64f3edc
+        .quad 0xc08627f6a248c778, 0xbe1cf0cd15805bc8
+        .quad 0xc08627f8d4860368, 0xbe1cf20db3bddebe
+        .quad 0xc08627fb06290f90, 0xbe1cf25188430e25
+        .quad 0xc08627fd37324070, 0xbe1ceea1713490f9
+        .quad 0xc08627ff67a1ea28, 0xbe1cf159521d234c
+        .quad 0xc0862801977860b8, 0xbe1cf24dfe50783b
+        .quad 0xc0862803c6b5f7d0, 0xbe1ceef2ef89a60b
+        .quad 0xc0862805f55b02c8, 0xbe1cee7fc919d62c
+        .quad 0xc08628082367d4c0, 0xbe1cf215a7fb513a
+        .quad 0xc086280a50dcc0a8, 0xbe1cf0e4401c5ed4
+        .quad 0xc086280c7dba1910, 0xbe1cf04ec734d256
+        .quad 0xc086280eaa003050, 0xbe1cf010ad787fea
+        .quad 0xc0862810d5af5880, 0xbe1cee622478393d
+        .quad 0xc086281300c7e368, 0xbe1cf01c7482564f
+        .quad 0xc08628152b4a22a0, 0xbe1cf0de20d33536
+        .quad 0xc086281755366778, 0xbe1cef2edae5837d
+        .quad 0xc08628197e8d02f0, 0xbe1cf0a345318cc9
+        .quad 0xc086281ba74e45d8, 0xbe1cf20085aa34b8
+        .quad 0xc086281dcf7a80c0, 0xbe1cef5fa845ad83
+        .quad 0xc086281ff71203e0, 0xbe1cf050d1df69c4
+        .quad 0xc08628221e151f48, 0xbe1ceffe43c035b9
+        .quad 0xc0862824448422b8, 0xbe1cf14f3018d3c2
+        .quad 0xc08628266a5f5dc0, 0xbe1cef0a5fbae83d
+        .quad 0xc08628288fa71f98, 0xbe1ceff8a95b72a1
+        .quad 0xc086282ab45bb750, 0xbe1cef073aa9849b
+        .quad 0xc086282cd87d73a8, 0xbe1cef69b3835c02
+        .quad 0xc086282efc0ca328, 0xbe1cf0bc139379a9
+        .quad 0xc08628311f099420, 0xbe1cef247a9ec596
+        .quad 0xc086283341749490, 0xbe1cef74bbcc488a
+        .quad 0xc0862835634df248, 0xbe1cef4bc42e7b8e
+        .quad 0xc08628378495fad0, 0xbe1cf136d4d5a810
+        .quad 0xc0862839a54cfb80, 0xbe1cf0d290b24dd8
+        .quad 0xc086283bc5734168, 0xbe1ceeebde8e0065
+        .quad 0xc086283de5091950, 0xbe1cf1a09f60aa1e
+        .quad 0xc0862840040ecfe0, 0xbe1cf0803947a234
+        .quad 0xc08628422284b168, 0xbe1cf0abf7638127
+        .quad 0xc0862844406b0a08, 0xbe1cf0f73ee12058
+        .quad 0xc08628465dc225a0, 0xbe1cf2079971b26c
+        .quad 0xc08628487a8a4fe0, 0xbe1cee74957564b1
+        .quad 0xc086284a96c3d420, 0xbe1ceee77c1b7d43
+        .quad 0xc086284cb26efd90, 0xbe1cf23addba6e09
+        .quad 0xc086284ecd8c1730, 0xbe1cf199f4a1da60
+        .quad 0xc0862850e81b6bb0, 0xbe1cf09fdea81393
+        .quad 0xc0862853021d4588, 0xbe1cf176adb417f7
+        .quad 0xc08628551b91ef00, 0xbe1cf0f64f84a8da
+        .quad 0xc08628573479b220, 0xbe1ceec34cf49523
+        .quad 0xc08628594cd4d8a8, 0xbe1cf16d60fbe0bb
+        .quad 0xc086285b64a3ac40, 0xbe1cee8de7acfc7b
+        .quad 0xc086285d7be67630, 0xbe1ceee6256cce8d
+        .quad 0xc086285f929d7fa0, 0xbe1cee7d66a3d8a5
+        .quad 0xc0862861a8c91170, 0xbe1cf0bef8265792
+        .quad 0xc0862863be697458, 0xbe1cf097f890c6f8
+        .quad 0xc0862865d37ef0c8, 0xbe1cf09502d5c3fc
+        .quad 0xc0862867e809cf00, 0xbe1ceeffb239dac7
+        .quad 0xc0862869fc0a56f8, 0xbe1cf1fbfff95c98
+        .quad 0xc086286c0f80d090, 0xbe1cefa57ad3eef7
+        .quad 0xc086286e226d8348, 0xbe1cf22c58b9183d
+        .quad 0xc086287034d0b690, 0xbe1ceff262d0a248
+        .quad 0xc086287246aab180, 0xbe1cefa7bc194186
+        .quad 0xc086287457fbbb08, 0xbe1cf06782d784d9
+        .quad 0xc086287668c419e0, 0xbe1cf1d44d0eaa07
+        .quad 0xc086287879041490, 0xbe1cf034803c8a48
+        .quad 0xc086287a88bbf158, 0xbe1cf08e84916b6f
+        .quad 0xc086287c97ebf650, 0xbe1cf0c4d3dc1bc7
+        .quad 0xc086287ea6946958, 0xbe1cefb1e4625943
+        .quad 0xc0862880b4b59010, 0xbe1cf143efdd1fd0
+        .quad 0xc0862882c24faff8, 0xbe1cee9896d016da
+        .quad 0xc0862884cf630e38, 0xbe1cf2186072f2cc
+        .quad 0xc0862886dbefeff0, 0xbe1cef9217633d34
+        .quad 0xc0862888e7f699e0, 0xbe1cf05603549486
+        .quad 0xc086288af37750b0, 0xbe1cef50fff513d3
+        .quad 0xc086288cfe7258c0, 0xbe1cf127713b32d0
+        .quad 0xc086288f08e7f650, 0xbe1cf05015520f3d
+        .quad 0xc086289112d86d58, 0xbe1cf12eb458b26f
+        .quad 0xc08628931c4401a8, 0xbe1cf22eae2887ed
+        .quad 0xc0862895252af6e0, 0xbe1cefdd6656dd2d
+        .quad 0xc08628972d8d9058, 0xbe1cf1048ea4e646
+        .quad 0xc0862899356c1150, 0xbe1ceec4501167e9
+        .quad 0xc086289b3cc6bcb8, 0xbe1cf0ad52becc3f
+        .quad 0xc086289d439dd568, 0xbe1cf0daa4e00e35
+        .quad 0xc086289f49f19df8, 0xbe1cf00b80de8d6a
+        .quad 0xc08628a14fc258c8, 0xbe1cf1bcf2ea8464
+        .quad 0xc08628a355104818, 0xbe1cf0435e2782b0
+        .quad 0xc08628a559dbade0, 0xbe1cf0e3e1a5f56c
+        .quad 0xc08628a75e24cbf8, 0xbe1cefed9d5a721d
+        .quad 0xc08628a961ebe3f8, 0xbe1cf0d2d74321e2
+        .quad 0xc08628ab65313750, 0xbe1cf24200eb55e9
+        .quad 0xc08628ad67f50740, 0xbe1cf23e9d7cf979
+        .quad 0xc08628af6a3794d0, 0xbe1cf23a088f421c
+        .quad 0xc08628b16bf920e0, 0xbe1cef2c1de1ab32
+        .quad 0xc08628b36d39ec08, 0xbe1cf1abc231f7b2
+        .quad 0xc08628b56dfa36d0, 0xbe1cf2074d5ba303
+        .quad 0xc08628b76e3a4180, 0xbe1cf05cd5eed880
+        /*== Log_LA_table ==*/
+        .align 32
+        .quad 0x8000000000000000
+        .quad 0xbf5ff802a9ab10e6
+        .quad 0xbf6ff00aa2b10bc0
+        .quad 0xbf77ee11ebd82e94
+        .quad 0xbf7fe02a6b106789
+        .quad 0xbf83e7295d25a7d9
+        .quad 0xbf87dc475f810a77
+        .quad 0xbf8bcf712c74384c
+        .quad 0xbf8fc0a8b0fc03e4
+        .quad 0xbf91d7f7eb9eebe7
+        .quad 0xbf93cea44346a575
+        .quad 0xbf95c45a51b8d389
+        .quad 0xbf97b91b07d5b11b
+        .quad 0xbf99ace7551cc514
+        .quad 0xbf9b9fc027af9198
+        .quad 0xbf9d91a66c543cc4
+        .quad 0xbf9f829b0e783300
+        .quad 0xbfa0b94f7c196176
+        .quad 0xbfa1b0d98923d980
+        .quad 0xbfa2a7ec2214e873
+        .quad 0xbfa39e87b9febd60
+        .quad 0xbfa494acc34d911c
+        .quad 0xbfa58a5bafc8e4d5
+        .quad 0xbfa67f94f094bd98
+        .quad 0xbfa77458f632dcfc
+        .quad 0xbfa868a83083f6cf
+        .quad 0xbfa95c830ec8e3eb
+        .quad 0xbfaa4fe9ffa3d235
+        .quad 0xbfab42dd711971bf
+        .quad 0xbfac355dd0921f2d
+        .quad 0xbfad276b8adb0b52
+        .quad 0xbfae19070c276016
+        .quad 0xbfaf0a30c01162a6
+        .quad 0xbfaffae9119b9303
+        .quad 0xbfb075983598e471
+        .quad 0xbfb0ed839b5526fe
+        .quad 0xbfb16536eea37ae1
+        .quad 0xbfb1dcb263db1944
+        .quad 0xbfb253f62f0a1417
+        .quad 0xbfb2cb0283f5de1f
+        .quad 0xbfb341d7961bd1d1
+        .quad 0xbfb3b87598b1b6ee
+        .quad 0xbfb42edcbea646f0
+        .quad 0xbfb4a50d3aa1b040
+        .quad 0xbfb51b073f06183f
+        .quad 0xbfb590cafdf01c28
+        .quad 0xbfb60658a93750c4
+        .quad 0xbfb67bb0726ec0fc
+        .quad 0xbfb6f0d28ae56b4c
+        .quad 0xbfb765bf23a6be13
+        .quad 0xbfb7da766d7b12cd
+        .quad 0xbfb84ef898e8282a
+        .quad 0xbfb8c345d6319b21
+        .quad 0xbfb9375e55595ede
+        .quad 0xbfb9ab42462033ad
+        .quad 0xbfba1ef1d8061cd4
+        .quad 0xbfba926d3a4ad563
+        .quad 0xbfbb05b49bee43fe
+        .quad 0xbfbb78c82bb0eda1
+        .quad 0xbfbbeba818146765
+        .quad 0xbfbc5e548f5bc743
+        .quad 0xbfbcd0cdbf8c13e1
+        .quad 0xbfbd4313d66cb35d
+        .quad 0xbfbdb5270187d927
+        .quad 0xbfbe27076e2af2e6
+        .quad 0xbfbe98b549671467
+        .quad 0xbfbf0a30c01162a6
+        .quad 0xbfbf7b79fec37ddf
+        .quad 0xbfbfec9131dbeabb
+        .quad 0xbfc02ebb42bf3d4b
+        .quad 0xbfc0671512ca596e
+        .quad 0xbfc09f561ee719c3
+        .quad 0xbfc0d77e7cd08e59
+        .quad 0xbfc10f8e422539b1
+        .quad 0xbfc14785846742ac
+        .quad 0xbfc17f6458fca611
+        .quad 0xbfc1b72ad52f67a0
+        .quad 0xbfc1eed90e2dc2c3
+        .quad 0xbfc2266f190a5acb
+        .quad 0xbfc25ded0abc6ad2
+        .quad 0xbfc29552f81ff523
+        .quad 0xbfc2cca0f5f5f251
+        .quad 0xbfc303d718e47fd3
+        .quad 0xbfc33af575770e4f
+        .quad 0xbfc371fc201e8f74
+        .quad 0xbfc3a8eb2d31a376
+        .quad 0xbfc3dfc2b0ecc62a
+        .quad 0xbfc41682bf727bc0
+        .quad 0xbfc44d2b6ccb7d1e
+        .quad 0xbfc483bccce6e3dd
+        .quad 0xbfc4ba36f39a55e5
+        .quad 0xbfc4f099f4a230b2
+        .quad 0xbfc526e5e3a1b438
+        .quad 0xbfc55d1ad4232d6f
+        .quad 0xbfc59338d9982086
+        .quad 0xbfc5c940075972b9
+        .quad 0xbfc5ff3070a793d4
+        .quad 0xbfc6350a28aaa758
+        .quad 0xbfc66acd4272ad51
+        .quad 0xbfc6a079d0f7aad2
+        .quad 0xbfc6d60fe719d21d
+        .quad 0xbfc70b8f97a1aa75
+        .quad 0xbfc740f8f54037a5
+        .quad 0xbfc7764c128f2127
+        .quad 0xbfc7ab890210d909
+        .quad 0xbfc7e0afd630c274
+        .quad 0xbfc815c0a14357eb
+        .quad 0xbfc84abb75865139
+        .quad 0xbfc87fa06520c911
+        .quad 0xbfc8b46f8223625b
+        .quad 0xbfc8e928de886d41
+        .quad 0xbfc91dcc8c340bde
+        .quad 0xbfc9525a9cf456b4
+        .quad 0xbfc986d3228180ca
+        .quad 0xbfc9bb362e7dfb83
+        .quad 0xbfc9ef83d2769a34
+        .quad 0xbfca23bc1fe2b563
+        .quad 0xbfca57df28244dcd
+        .quad 0xbfca8becfc882f19
+        .quad 0xbfcabfe5ae46124c
+        .quad 0xbfcaf3c94e80bff3
+        .quad 0xbfcb2797ee46320c
+        .quad 0xbfcb5b519e8fb5a4
+        .quad 0xbfcb8ef670420c3b
+        .quad 0xbfcbc286742d8cd6
+        .quad 0xbfcbf601bb0e44e2
+        .quad 0xbfcc2968558c18c1
+        .quad 0xbfcc5cba543ae425
+        .quad 0xbfcc8ff7c79a9a22
+        .quad 0xbfccc320c0176502
+        .quad 0xbfccf6354e09c5dc
+        .quad 0xbfcd293581b6b3e7
+        .quad 0xbfcd5c216b4fbb91
+        .quad 0xbfcd8ef91af31d5e
+        .quad 0xbfcdc1bca0abec7d
+        .quad 0xbfcdf46c0c722d2f
+        .quad 0xbfce27076e2af2e6
+        .quad 0xbfce598ed5a87e2f
+        .quad 0xbfce8c0252aa5a60
+        .quad 0xbfcebe61f4dd7b0b
+        .quad 0xbfcef0adcbdc5936
+        .quad 0xbfcf22e5e72f105d
+        .quad 0xbfcf550a564b7b37
+        .quad 0xbfcf871b28955045
+        .quad 0xbfcfb9186d5e3e2b
+        .quad 0xbfcfeb0233e607cc
+        .quad 0xbfd00e6c45ad501d
+        .quad 0xbfd0274dc16c232f
+        .quad 0xbfd0402594b4d041
+        .quad 0xbfd058f3c703ebc6
+        .quad 0xbfd071b85fcd590d
+        .quad 0xbfd08a73667c57af
+        .quad 0xbfd0a324e27390e3
+        .quad 0xbfd0bbccdb0d24bd
+        .quad 0xbfd0d46b579ab74b
+        .quad 0xbfd0ed005f657da4
+        .quad 0xbfd1058bf9ae4ad5
+        .quad 0xbfd11e0e2dad9cb7
+        .quad 0xbfd136870293a8b0
+        .quad 0xbfd14ef67f88685a
+        .quad 0xbfd1675cababa60e
+        .quad 0xbfd17fb98e15095d
+        .quad 0xbfd1980d2dd4236f
+        .quad 0xbfd1b05791f07b49
+        .quad 0xbfd1c898c16999fb
+        .quad 0xbfd1e0d0c33716be
+        .quad 0xbfd1f8ff9e48a2f3
+        .quad 0xbfd211255986160c
+        .quad 0xbfd22941fbcf7966
+        .quad 0xbfd241558bfd1404
+        .quad 0xbfd2596010df763a
+        .quad 0xbfd27161913f853d
+        .quad 0xbfd2895a13de86a3
+        .quad 0xbfd2a1499f762bc9
+        .quad 0xbfd2b9303ab89d25
+        .quad 0xbfd2d10dec508583
+        .quad 0xbfd2e8e2bae11d31
+        .quad 0xbfd300aead06350c
+        .quad 0xbfd31871c9544185
+        .quad 0xbfd3302c16586588
+        .quad 0xbfd347dd9a987d55
+        .quad 0xbfd35f865c93293e
+        .quad 0xbfd3772662bfd85b
+        .quad 0xbfd38ebdb38ed321
+        .quad 0xbfd3a64c556945ea
+        .quad 0xbfd3bdd24eb14b6a
+        .quad 0xbfd3d54fa5c1f710
+        .quad 0xbfd3ecc460ef5f50
+        .quad 0xbfd404308686a7e4
+        .quad 0xbfd41b941cce0bee
+        .quad 0xbfd432ef2a04e814
+        .quad 0xbfd44a41b463c47c
+        .quad 0xbfd4618bc21c5ec2
+        .quad 0xbfd478cd5959b3d9
+        .quad 0xbfd49006804009d1
+        .quad 0xbfd4a7373cecf997
+        .quad 0xbfd4be5f957778a1
+        .quad 0xbfd4d57f8fefe27f
+        .quad 0xbfd4ec973260026a
+        .quad 0xbfd503a682cb1cb3
+        .quad 0xbfd51aad872df82d
+        .quad 0xbfd531ac457ee77e
+        .quad 0xbfd548a2c3add263
+        .quad 0xbfd55f9107a43ee2
+        .quad 0xbfd5767717455a6c
+        .quad 0xbfd58d54f86e02f2
+        .quad 0xbfd5a42ab0f4cfe2
+        .quad 0xbfd5baf846aa1b19
+        .quad 0xbfd5d1bdbf5809ca
+        .quad 0xbfd5e87b20c2954a
+        .quad 0xbfd5ff3070a793d4
+        .quad 0xbfd615ddb4bec13c
+        .quad 0xbfd62c82f2b9c795
+        .quad 0x3fd61965cdb02c1f
+        .quad 0x3fd602d08af091ec
+        .quad 0x3fd5ec433d5c35ae
+        .quad 0x3fd5d5bddf595f30
+        .quad 0x3fd5bf406b543db2
+        .quad 0x3fd5a8cadbbedfa1
+        .quad 0x3fd5925d2b112a59
+        .quad 0x3fd57bf753c8d1fb
+        .quad 0x3fd565995069514c
+        .quad 0x3fd54f431b7be1a9
+        .quad 0x3fd538f4af8f72fe
+        .quad 0x3fd522ae0738a3d8
+        .quad 0x3fd50c6f1d11b97c
+        .quad 0x3fd4f637ebba9810
+        .quad 0x3fd4e0086dd8baca
+        .quad 0x3fd4c9e09e172c3c
+        .quad 0x3fd4b3c077267e9a
+        .quad 0x3fd49da7f3bcc41f
+        .quad 0x3fd487970e958770
+        .quad 0x3fd4718dc271c41b
+        .quad 0x3fd45b8c0a17df13
+        .quad 0x3fd44591e0539f49
+        .quad 0x3fd42f9f3ff62642
+        .quad 0x3fd419b423d5e8c7
+        .quad 0x3fd403d086cea79c
+        .quad 0x3fd3edf463c1683e
+        .quad 0x3fd3d81fb5946dba
+        .quad 0x3fd3c25277333184
+        .quad 0x3fd3ac8ca38e5c5f
+        .quad 0x3fd396ce359bbf54
+        .quad 0x3fd3811728564cb2
+        .quad 0x3fd36b6776be1117
+        .quad 0x3fd355bf1bd82c8b
+        .quad 0x3fd3401e12aecba1
+        .quad 0x3fd32a84565120a8
+        .quad 0x3fd314f1e1d35ce4
+        .quad 0x3fd2ff66b04ea9d4
+        .quad 0x3fd2e9e2bce12286
+        .quad 0x3fd2d46602adccee
+        .quad 0x3fd2bef07cdc9354
+        .quad 0x3fd2a982269a3dbf
+        .quad 0x3fd2941afb186b7c
+        .quad 0x3fd27ebaf58d8c9d
+        .quad 0x3fd269621134db92
+        .quad 0x3fd25410494e56c7
+        .quad 0x3fd23ec5991eba49
+        .quad 0x3fd22981fbef797b
+        .quad 0x3fd214456d0eb8d4
+        .quad 0x3fd1ff0fe7cf47a7
+        .quad 0x3fd1e9e1678899f4
+        .quad 0x3fd1d4b9e796c245
+        .quad 0x3fd1bf99635a6b95
+        .quad 0x3fd1aa7fd638d33f
+        .quad 0x3fd1956d3b9bc2fa
+        .quad 0x3fd180618ef18adf
+        .quad 0x3fd16b5ccbacfb73
+        .quad 0x3fd1565eed455fc3
+        .quad 0x3fd14167ef367783
+        .quad 0x3fd12c77cd00713b
+        .quad 0x3fd1178e8227e47c
+        .quad 0x3fd102ac0a35cc1c
+        .quad 0x3fd0edd060b78081
+        .quad 0x3fd0d8fb813eb1ef
+        .quad 0x3fd0c42d676162e3
+        .quad 0x3fd0af660eb9e279
+        .quad 0x3fd09aa572e6c6d4
+        .quad 0x3fd085eb8f8ae797
+        .quad 0x3fd07138604d5862
+        .quad 0x3fd05c8be0d9635a
+        .quad 0x3fd047e60cde83b8
+        .quad 0x3fd03346e0106062
+        .quad 0x3fd01eae5626c691
+        .quad 0x3fd00a1c6adda473
+        .quad 0x3fcfeb2233ea07cd
+        .quad 0x3fcfc218be620a5e
+        .quad 0x3fcf991c6cb3b379
+        .quad 0x3fcf702d36777df0
+        .quad 0x3fcf474b134df229
+        .quad 0x3fcf1e75fadf9bde
+        .quad 0x3fcef5ade4dcffe6
+        .quad 0x3fceccf2c8fe920a
+        .quad 0x3fcea4449f04aaf5
+        .quad 0x3fce7ba35eb77e2a
+        .quad 0x3fce530effe71012
+        .quad 0x3fce2a877a6b2c12
+        .quad 0x3fce020cc6235ab5
+        .quad 0x3fcdd99edaf6d7e9
+        .quad 0x3fcdb13db0d48940
+        .quad 0x3fcd88e93fb2f450
+        .quad 0x3fcd60a17f903515
+        .quad 0x3fcd38666871f465
+        .quad 0x3fcd1037f2655e7b
+        .quad 0x3fcce816157f1988
+        .quad 0x3fccc000c9db3c52
+        .quad 0x3fcc97f8079d44ec
+        .quad 0x3fcc6ffbc6f00f71
+        .quad 0x3fcc480c0005ccd1
+        .quad 0x3fcc2028ab17f9b4
+        .quad 0x3fcbf851c067555f
+        .quad 0x3fcbd087383bd8ad
+        .quad 0x3fcba8c90ae4ad19
+        .quad 0x3fcb811730b823d2
+        .quad 0x3fcb5971a213acdb
+        .quad 0x3fcb31d8575bce3d
+        .quad 0x3fcb0a4b48fc1b46
+        .quad 0x3fcae2ca6f672bd4
+        .quad 0x3fcabb55c31693ad
+        .quad 0x3fca93ed3c8ad9e3
+        .quad 0x3fca6c90d44b704e
+        .quad 0x3fca454082e6ab05
+        .quad 0x3fca1dfc40f1b7f1
+        .quad 0x3fc9f6c407089664
+        .quad 0x3fc9cf97cdce0ec3
+        .quad 0x3fc9a8778debaa38
+        .quad 0x3fc981634011aa75
+        .quad 0x3fc95a5adcf7017f
+        .quad 0x3fc9335e5d594989
+        .quad 0x3fc90c6db9fcbcd9
+        .quad 0x3fc8e588ebac2dbf
+        .quad 0x3fc8beafeb38fe8c
+        .quad 0x3fc897e2b17b19a5
+        .quad 0x3fc871213750e994
+        .quad 0x3fc84a6b759f512f
+        .quad 0x3fc823c16551a3c2
+        .quad 0x3fc7fd22ff599d4f
+        .quad 0x3fc7d6903caf5ad0
+        .quad 0x3fc7b0091651528c
+        .quad 0x3fc7898d85444c73
+        .quad 0x3fc7631d82935a86
+        .quad 0x3fc73cb9074fd14d
+        .quad 0x3fc716600c914054
+        .quad 0x3fc6f0128b756abc
+        .quad 0x3fc6c9d07d203fc7
+        .quad 0x3fc6a399dabbd383
+        .quad 0x3fc67d6e9d785771
+        .quad 0x3fc6574ebe8c133a
+        .quad 0x3fc6313a37335d76
+        .quad 0x3fc60b3100b09476
+        .quad 0x3fc5e533144c1719
+        .quad 0x3fc5bf406b543db2
+        .quad 0x3fc59958ff1d52f1
+        .quad 0x3fc5737cc9018cdd
+        .quad 0x3fc54dabc26105d2
+        .quad 0x3fc527e5e4a1b58d
+        .quad 0x3fc5022b292f6a45
+        .quad 0x3fc4dc7b897bc1c8
+        .quad 0x3fc4b6d6fefe22a4
+        .quad 0x3fc4913d8333b561
+        .quad 0x3fc46baf0f9f5db7
+        .quad 0x3fc4462b9dc9b3dc
+        .quad 0x3fc420b32740fdd4
+        .quad 0x3fc3fb45a59928cc
+        .quad 0x3fc3d5e3126bc27f
+        .quad 0x3fc3b08b6757f2a9
+        .quad 0x3fc38b3e9e027479
+        .quad 0x3fc365fcb0159016
+        .quad 0x3fc340c59741142e
+        .quad 0x3fc31b994d3a4f85
+        .quad 0x3fc2f677cbbc0a96
+        .quad 0x3fc2d1610c86813a
+        .quad 0x3fc2ac55095f5c59
+        .quad 0x3fc28753bc11aba5
+        .quad 0x3fc2625d1e6ddf57
+        .quad 0x3fc23d712a49c202
+        .quad 0x3fc2188fd9807263
+        .quad 0x3fc1f3b925f25d41
+        .quad 0x3fc1ceed09853752
+        .quad 0x3fc1aa2b7e23f72a
+        .quad 0x3fc185747dbecf34
+        .quad 0x3fc160c8024b27b1
+        .quad 0x3fc13c2605c398c3
+        .quad 0x3fc1178e8227e47c
+        .quad 0x3fc0f301717cf0fb
+        .quad 0x3fc0ce7ecdccc28d
+        .quad 0x3fc0aa06912675d5
+        .quad 0x3fc08598b59e3a07
+        .quad 0x3fc06135354d4b18
+        .quad 0x3fc03cdc0a51ec0d
+        .quad 0x3fc0188d2ecf6140
+        .quad 0x3fbfe89139dbd566
+        .quad 0x3fbfa01c9db57ce2
+        .quad 0x3fbf57bc7d9005db
+        .quad 0x3fbf0f70cdd992e3
+        .quad 0x3fbec739830a1120
+        .quad 0x3fbe7f1691a32d3e
+        .quad 0x3fbe3707ee30487b
+        .quad 0x3fbdef0d8d466db9
+        .quad 0x3fbda727638446a2
+        .quad 0x3fbd5f55659210e2
+        .quad 0x3fbd179788219364
+        .quad 0x3fbccfedbfee13a8
+        .quad 0x3fbc885801bc4b23
+        .quad 0x3fbc40d6425a5cb1
+        .quad 0x3fbbf968769fca11
+        .quad 0x3fbbb20e936d6974
+        .quad 0x3fbb6ac88dad5b1c
+        .quad 0x3fbb23965a52ff00
+        .quad 0x3fbadc77ee5aea8c
+        .quad 0x3fba956d3ecade63
+        .quad 0x3fba4e7640b1bc38
+        .quad 0x3fba0792e9277cac
+        .quad 0x3fb9c0c32d4d2548
+        .quad 0x3fb97a07024cbe74
+        .quad 0x3fb9335e5d594989
+        .quad 0x3fb8ecc933aeb6e8
+        .quad 0x3fb8a6477a91dc29
+        .quad 0x3fb85fd927506a48
+        .quad 0x3fb8197e2f40e3f0
+        .quad 0x3fb7d33687c293c9
+        .quad 0x3fb78d02263d82d3
+        .quad 0x3fb746e100226ed9
+        .quad 0x3fb700d30aeac0e1
+        .quad 0x3fb6bad83c1883b6
+        .quad 0x3fb674f089365a7a
+        .quad 0x3fb62f1be7d77743
+        .quad 0x3fb5e95a4d9791cb
+        .quad 0x3fb5a3abb01ade25
+        .quad 0x3fb55e10050e0384
+        .quad 0x3fb518874226130a
+        .quad 0x3fb4d3115d207eac
+        .quad 0x3fb48dae4bc31018
+        .quad 0x3fb4485e03dbdfad
+        .quad 0x3fb403207b414b7f
+        .quad 0x3fb3bdf5a7d1ee64
+        .quad 0x3fb378dd7f749714
+        .quad 0x3fb333d7f8183f4b
+        .quad 0x3fb2eee507b40301
+        .quad 0x3fb2aa04a44717a5
+        .quad 0x3fb26536c3d8c369
+        .quad 0x3fb2207b5c78549e
+        .quad 0x3fb1dbd2643d190b
+        .quad 0x3fb1973bd1465567
+        .quad 0x3fb152b799bb3cc9
+        .quad 0x3fb10e45b3cae831
+        .quad 0x3fb0c9e615ac4e17
+        .quad 0x3fb08598b59e3a07
+        .quad 0x3fb0415d89e74444
+        .quad 0x3faffa6911ab9301
+        .quad 0x3faf723b517fc523
+        .quad 0x3faeea31c006b87c
+        .quad 0x3fae624c4a0b5e1b
+        .quad 0x3fadda8adc67ee4e
+        .quad 0x3fad52ed6405d86f
+        .quad 0x3faccb73cdddb2cc
+        .quad 0x3fac441e06f72a9e
+        .quad 0x3fabbcebfc68f420
+        .quad 0x3fab35dd9b58baad
+        .quad 0x3faaaef2d0fb10fc
+        .quad 0x3faa282b8a936171
+        .quad 0x3fa9a187b573de7c
+        .quad 0x3fa91b073efd7314
+        .quad 0x3fa894aa149fb343
+        .quad 0x3fa80e7023d8ccc4
+        .quad 0x3fa788595a3577ba
+        .quad 0x3fa70265a550e777
+        .quad 0x3fa67c94f2d4bb58
+        .quad 0x3fa5f6e73078efb8
+        .quad 0x3fa5715c4c03ceef
+        .quad 0x3fa4ebf43349e26f
+        .quad 0x3fa466aed42de3ea
+        .quad 0x3fa3e18c1ca0ae92
+        .quad 0x3fa35c8bfaa1306b
+        .quad 0x3fa2d7ae5c3c5bae
+        .quad 0x3fa252f32f8d183f
+        .quad 0x3fa1ce5a62bc353a
+        .quad 0x3fa149e3e4005a8d
+        .quad 0x3fa0c58fa19dfaaa
+        .quad 0x3fa0415d89e74444
+        .quad 0x3f9f7a9b16782856
+        .quad 0x3f9e72bf2813ce51
+        .quad 0x3f9d6b2725979802
+        .quad 0x3f9c63d2ec14aaf2
+        .quad 0x3f9b5cc258b718e6
+        .quad 0x3f9a55f548c5c43f
+        .quad 0x3f994f6b99a24475
+        .quad 0x3f98492528c8cabf
+        .quad 0x3f974321d3d006d3
+        .quad 0x3f963d6178690bd6
+        .quad 0x3f9537e3f45f3565
+        .quad 0x3f9432a925980cc1
+        .quad 0x3f932db0ea132e22
+        .quad 0x3f9228fb1fea2e28
+        .quad 0x3f912487a5507f70
+        .quad 0x3f90205658935847
+        .quad 0x3f8e38ce3033310c
+        .quad 0x3f8c317384c75f06
+        .quad 0x3f8a2a9c6c170462
+        .quad 0x3f882448a388a2aa
+        .quad 0x3f861e77e8b53fc6
+        .quad 0x3f841929f96832f0
+        .quad 0x3f82145e939ef1e9
+        .quad 0x3f8010157588de71
+        .quad 0x3f7c189cbb0e27fb
+        .quad 0x3f78121214586b54
+        .quad 0x3f740c8a747878e2
+        .quad 0x3f70080559588b35
+        .quad 0x3f680904828985c0
+        .quad 0x3f60040155d5889e
+        .quad 0x3f50020055655889
+        .quad 0x0000000000000000
+        /*== poly_coeff[4] ==*/
+        .align 32
+        .quad 0x3fc9999CACDB4D0A, 0x3fc9999CACDB4D0A, 0x3fc9999CACDB4D0A, 0x3fc9999CACDB4D0A /* coeff4 */
+        .quad 0xbfd0000148058EE1, 0xbfd0000148058EE1, 0xbfd0000148058EE1, 0xbfd0000148058EE1 /* coeff3 */
+        .quad 0x3fd55555555543C5, 0x3fd55555555543C5, 0x3fd55555555543C5, 0x3fd55555555543C5 /* coeff2 */
+        .quad 0xbfdFFFFFFFFFF81F, 0xbfdFFFFFFFFFF81F, 0xbfdFFFFFFFFFF81F, 0xbfdFFFFFFFFFF81F /* coeff1 */
+        /*== ExpMask ==*/
+        .align 32
+        .quad 0x000fffffffffffff, 0x000fffffffffffff, 0x000fffffffffffff, 0x000fffffffffffff
+        /*== Two10 ==*/
+        .align 32
+        .quad 0x3f50000000000000, 0x3f50000000000000, 0x3f50000000000000, 0x3f50000000000000
+        /*== MinLog1p = -1+2^(-53) ==*/
+        .align 32
+        .quad 0xbfefffffffffffff, 0xbfefffffffffffff, 0xbfefffffffffffff, 0xbfefffffffffffff
+        /*== MaxLog1p ==*/
+        .align 32
+        .quad 0x7f3ffffffffff000, 0x7f3ffffffffff000, 0x7f3ffffffffff000, 0x7f3ffffffffff000
+        /*== One ==*/
+        .align 32
+        .quad 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000
+        /*== SgnMask ==*/
+        .align 32
+        .quad 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff
+        /*== XThreshold ==*/
+        .align 32
+        .quad 0x3e00000000000000, 0x3e00000000000000, 0x3e00000000000000, 0x3e00000000000000
+        /*== XhMask ==*/
+        .align 32
+        .quad 0xfffffffffffffc00, 0xfffffffffffffc00, 0xfffffffffffffc00, 0xfffffffffffffc00
+        /*== Threshold ==*/
+        .align 32
+        .quad 0x4086a00000000000, 0x4086a00000000000, 0x4086a00000000000, 0x4086a00000000000
+        /*== Bias ==*/
+        .align 32
+        .quad 0x408ff80000000000, 0x408ff80000000000, 0x408ff80000000000, 0x408ff80000000000
+        /*== Bias1 ==*/
+        .align 32
+        .quad 0x408ff00000000000, 0x408ff00000000000, 0x408ff00000000000, 0x408ff00000000000
+        /*== ExpMask ==*/
+        .align 32
+        .quad 0x7ff0000000000000, 0x7ff0000000000000, 0x7ff0000000000000, 0x7ff0000000000000
+        /*== ExpMask2 ==*/
+        .align 32
+        .quad 0x7f40000000000000, 0x7f40000000000000, 0x7f40000000000000, 0x7f40000000000000
+        /*== L2L ==*/
+        .align 32
+        .quad 0x3fe62E42FEFA39EF, 0x3fe62E42FEFA39EF, 0x3fe62E42FEFA39EF, 0x3fe62E42FEFA39EF
+        /*== dBigThreshold ==*/
+        .align 32
+        .quad 0x41D0000000000000, 0x41D0000000000000, 0x41D0000000000000, 0x41D0000000000000
+        /*== dC2 ==*/
+        .align 32
+        .quad 0x3FD8000000000000, 0x3FD8000000000000, 0x3FD8000000000000, 0x3FD8000000000000
+        /*== dC3 ==*/
+        .align 32
+        .quad 0x3FD4000000000000, 0x3FD4000000000000, 0x3FD4000000000000, 0x3FD4000000000000
+        /*== dC4 ==*/
+        .align 32
+        .quad 0x3FD1800000000000, 0x3FD1800000000000, 0x3FD1800000000000, 0x3FD1800000000000
+        /*== dC5 ==*/
+        .align 32
+        .quad 0x3FCF800000000000, 0x3FCF800000000000, 0x3FCF800000000000, 0x3FCF800000000000
+        /*== dHalf ==*/
+        .align 32
+        .quad 0x3FE0000000000000, 0x3FE0000000000000, 0x3FE0000000000000, 0x3FE0000000000000
+        /*== dLargestFinite ==*/
+        .align 32
+        .quad 0x7FEFFFFFFFFFFFFF, 0x7FEFFFFFFFFFFFFF, 0x7FEFFFFFFFFFFFFF, 0x7FEFFFFFFFFFFFFF
+        /*== dLittleThreshold ==*/
+        .align 32
+        .quad 0x3F60000000000000, 0x3F60000000000000, 0x3F60000000000000, 0x3F60000000000000
+        /*== dSign ==*/
+        .align 32
+        .quad 0x8000000000000000, 0x8000000000000000, 0x8000000000000000, 0x8000000000000000
+        /*== dThirtyOne ==*/
+        .align 32
+        .quad 0x403F000000000000, 0x403F000000000000, 0x403F000000000000, 0x403F000000000000
+        /*== dTopMask12 ==*/
+        .align 32
+        .quad 0xFFFFFE0000000000, 0xFFFFFE0000000000, 0xFFFFFE0000000000, 0xFFFFFE0000000000
+        /*== dTopMask29 ==*/
+        .align 32
+        .quad 0xFFFFFFFFFF000000, 0xFFFFFFFFFF000000, 0xFFFFFFFFFF000000, 0xFFFFFFFFFF000000
+        /*== XScale ==*/
+        .align 32
+        .quad 0x3E10000000000000, 0x3E10000000000000, 0x3E10000000000000, 0x3E10000000000000
+        .align 32
+        .type	__svml_dasinh_data_internal,@object
+        .size	__svml_dasinh_data_internal,.-__svml_dasinh_data_internal
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core-avx2.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core-avx2.S
new file mode 100644
index 0000000000..647c73292c
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core-avx2.S
@@ -0,0 +1,20 @@ 
+/* AVX2 version of vectorized asinh, vector length is 8.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#define _ZGVeN8v_asinh _ZGVeN8v_asinh_avx2_wrapper
+#include "../svml_d_asinh8_core.S"
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core.c b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core.c
new file mode 100644
index 0000000000..45e5ab72a6
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core.c
@@ -0,0 +1,27 @@ 
+/* Multiple versions of vectorized asinh, vector length is 8.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#define SYMBOL_NAME _ZGVeN8v_asinh
+#include "ifunc-mathvec-avx512-skx.h"
+
+libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
+
+#ifdef SHARED
+__hidden_ver1 (_ZGVeN8v_asinh, __GI__ZGVeN8v_asinh, __redirect__ZGVeN8v_asinh)
+  __attribute__ ((visibility ("hidden")));
+#endif
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core_avx512.S b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core_avx512.S
new file mode 100644
index 0000000000..8100e8a50a
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_d_asinh8_core_avx512.S
@@ -0,0 +1,510 @@ 
+/* Function asinh vectorized with AVX-512.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   https://www.gnu.org/licenses/.  */
+
+/*
+ * ALGORITHM DESCRIPTION:
+ *
+ *   Compute asinh(x) as log(x + sqrt(x*x + 1))
+ *   using RSQRT instructions for starting the
+ *   square root approximation, and small table lookups for log
+ *   that map to AVX-512 permute instructions
+ *
+ *   Special cases:
+ *
+ *   asinh(NaN) = quiet NaN, and raise invalid exception
+ *   asinh(INF) = that INF
+ *   asinh(0)   = that 0
+ *
+ */
+
+/* Offsets for data table __svml_dasinh_data_internal_avx512
+ */
+#define Log_tbl_H                     	0
+#define Log_tbl_L                     	128
+#define One                           	256
+#define AbsMask                       	320
+#define SmallThreshold                	384
+#define Threshold                     	448
+#define LargeThreshold                	512
+#define ca2                           	576
+#define ca1                           	640
+#define c4s                           	704
+#define c3s                           	768
+#define c2s                           	832
+#define c1s                           	896
+#define AddB5                         	960
+#define RcpBitMask                    	1024
+#define OneEighth                     	1088
+#define Four                          	1152
+#define poly_coeff9                   	1216
+#define poly_coeff8                   	1280
+#define poly_coeff7                   	1344
+#define poly_coeff6                   	1408
+#define poly_coeff5                   	1472
+#define poly_coeff4                   	1536
+#define poly_coeff3                   	1600
+#define poly_coeff2                   	1664
+#define poly_coeff1                   	1728
+#define L2H                           	1792
+#define L2L                           	1856
+
+#include <sysdep.h>
+
+        .text
+	.section .text.evex512,"ax",@progbits
+ENTRY(_ZGVeN8v_asinh_skx)
+        pushq     %rbp
+        cfi_def_cfa_offset(16)
+        movq      %rsp, %rbp
+        cfi_def_cfa(6, 16)
+        cfi_offset(6, -16)
+        andq      $-64, %rsp
+        subq      $192, %rsp
+        vmovaps   %zmm0, %zmm3
+
+/* x^2 */
+        vmulpd    {rn-sae}, %zmm3, %zmm3, %zmm14
+        vmovups   One+__svml_dasinh_data_internal_avx512(%rip), %zmm9
+
+/* polynomial computation for small inputs */
+        vmovups   ca2+__svml_dasinh_data_internal_avx512(%rip), %zmm10
+        vmovups   ca1+__svml_dasinh_data_internal_avx512(%rip), %zmm11
+
+/* not a very small input ? */
+        vmovups   SmallThreshold+__svml_dasinh_data_internal_avx512(%rip), %zmm0
+
+/* A=max(x^2, 1); */
+        vmaxpd    {sae}, %zmm14, %zmm9, %zmm4
+
+/* B=min(x^2, 1); */
+        vminpd    {sae}, %zmm14, %zmm9, %zmm5
+        vfmadd231pd {rn-sae}, %zmm14, %zmm10, %zmm11
+
+/* 1+x^2 */
+        vaddpd    {rn-sae}, %zmm9, %zmm14, %zmm8
+
+/* |input| */
+        vandpd    AbsMask+__svml_dasinh_data_internal_avx512(%rip), %zmm3, %zmm1
+        vrsqrt14pd %zmm8, %zmm6
+        vcmppd    $21, {sae}, %zmm0, %zmm1, %k2
+
+/* B_high */
+        vsubpd    {rn-sae}, %zmm4, %zmm8, %zmm7
+
+/* sign bit */
+        vxorpd    %zmm3, %zmm1, %zmm2
+        vmulpd    {rn-sae}, %zmm14, %zmm11, %zmm4
+
+/* B_low */
+        vsubpd    {rn-sae}, %zmm7, %zmm5, %zmm13
+        vmovups   c2s+__svml_dasinh_data_internal_avx512(%rip), %zmm5
+        vmovups   c1s+__svml_dasinh_data_internal_avx512(%rip), %zmm7
+
+/* polynomial computation for small inputs */
+        vfmadd213pd {rn-sae}, %zmm1, %zmm1, %zmm4
+
+/* (x^2)_low */
+        vmovaps   %zmm3, %zmm15
+        vfmsub213pd {rn-sae}, %zmm14, %zmm3, %zmm15
+
+/* Sh ~sqrt(1+x^2) */
+        vmulpd    {rn-sae}, %zmm6, %zmm8, %zmm14
+
+/* Yl = (x^2)_low + B_low */
+        vaddpd    {rn-sae}, %zmm15, %zmm13, %zmm13
+
+/* very large inputs ? */
+        vmovups   Threshold+__svml_dasinh_data_internal_avx512(%rip), %zmm15
+
+/* (Yh*R0)_low */
+        vfmsub213pd {rn-sae}, %zmm14, %zmm6, %zmm8
+        vcmppd    $21, {sae}, %zmm15, %zmm1, %k1
+
+/* Sl = (Yh*R0)_low+(R0*Yl) */
+        vfmadd213pd {rn-sae}, %zmm8, %zmm6, %zmm13
+        vmovups   LargeThreshold+__svml_dasinh_data_internal_avx512(%rip), %zmm8
+
+/* rel. error term: Eh=1-Sh*R0 */
+        vmovaps   %zmm9, %zmm12
+        vfnmadd231pd {rn-sae}, %zmm14, %zmm6, %zmm12
+        vcmppd    $22, {sae}, %zmm8, %zmm1, %k0
+
+/* rel. error term: Eh=(1-Sh*R0)-Sl*R0 */
+        vfnmadd231pd {rn-sae}, %zmm13, %zmm6, %zmm12
+
+/*
+ * sqrt(1+x^2) ~ Sh + Sl + Sh*Eh*poly_s
+ * poly_s = c1+c2*Eh+c3*Eh^2
+ */
+        vmovups   c4s+__svml_dasinh_data_internal_avx512(%rip), %zmm6
+        vmovups   c3s+__svml_dasinh_data_internal_avx512(%rip), %zmm8
+
+/* Sh*Eh */
+        vmulpd    {rn-sae}, %zmm12, %zmm14, %zmm11
+        vfmadd231pd {rn-sae}, %zmm12, %zmm6, %zmm8
+
+/* Sh+x */
+        vaddpd    {rn-sae}, %zmm1, %zmm14, %zmm6
+        kmovw     %k0, %edx
+        vfmadd213pd {rn-sae}, %zmm5, %zmm12, %zmm8
+        vfmadd213pd {rn-sae}, %zmm7, %zmm12, %zmm8
+
+/* Xh */
+        vsubpd    {rn-sae}, %zmm14, %zmm6, %zmm12
+
+/* Sl + Sh*Eh*poly_s */
+        vfmadd213pd {rn-sae}, %zmm13, %zmm8, %zmm11
+
+/* fixup for very large inputs */
+        vmovups   OneEighth+__svml_dasinh_data_internal_avx512(%rip), %zmm8
+
+/* Xl */
+        vsubpd    {rn-sae}, %zmm12, %zmm1, %zmm12
+
+/* Xin0+Sl+Sh*Eh*poly_s ~ x+sqrt(1+x^2) */
+        vaddpd    {rn-sae}, %zmm11, %zmm6, %zmm10
+
+/* Sl_high */
+        vsubpd    {rn-sae}, %zmm6, %zmm10, %zmm5
+        vmulpd    {rn-sae}, %zmm8, %zmm1, %zmm10{%k1}
+
+/* Table lookups */
+        vmovups   __svml_dasinh_data_internal_avx512(%rip), %zmm6
+
+/* Sl_l */
+        vsubpd    {rn-sae}, %zmm5, %zmm11, %zmm7
+        vrcp14pd  %zmm10, %zmm13
+
+/* Xin_low */
+        vaddpd    {rn-sae}, %zmm12, %zmm7, %zmm14
+        vmovups   Log_tbl_L+__svml_dasinh_data_internal_avx512(%rip), %zmm7
+        vmovups   poly_coeff6+__svml_dasinh_data_internal_avx512(%rip), %zmm12
+
+/* round reciprocal to 1+4b mantissas */
+        vpaddq    AddB5+__svml_dasinh_data_internal_avx512(%rip), %zmm13, %zmm11
+
+/* fixup for very large inputs */
+        vxorpd    %zmm14, %zmm14, %zmm14{%k1}
+        vmovups   poly_coeff5+__svml_dasinh_data_internal_avx512(%rip), %zmm13
+        vandpd    RcpBitMask+__svml_dasinh_data_internal_avx512(%rip), %zmm11, %zmm15
+        vmovups   poly_coeff7+__svml_dasinh_data_internal_avx512(%rip), %zmm11
+
+/* Prepare table index */
+        vpsrlq    $48, %zmm15, %zmm5
+
+/* reduced argument for log(): (Rcp*Xin-1)+Rcp*Xin_low */
+        vfmsub231pd {rn-sae}, %zmm15, %zmm10, %zmm9
+
+/* exponents */
+        vgetexppd {sae}, %zmm15, %zmm8
+        vmovups   Four+__svml_dasinh_data_internal_avx512(%rip), %zmm10
+        vpermt2pd Log_tbl_H+64+__svml_dasinh_data_internal_avx512(%rip), %zmm5, %zmm6
+        vpermt2pd Log_tbl_L+64+__svml_dasinh_data_internal_avx512(%rip), %zmm5, %zmm7
+        vsubpd    {rn-sae}, %zmm10, %zmm8, %zmm8{%k1}
+        vfmadd231pd {rn-sae}, %zmm15, %zmm14, %zmm9
+
+/* polynomials */
+        vmovups   poly_coeff9+__svml_dasinh_data_internal_avx512(%rip), %zmm10
+        vmovups   poly_coeff8+__svml_dasinh_data_internal_avx512(%rip), %zmm5
+        vmovups   poly_coeff4+__svml_dasinh_data_internal_avx512(%rip), %zmm14
+
+/* -K*L2H + Th */
+        vmovups   L2H+__svml_dasinh_data_internal_avx512(%rip), %zmm15
+        vfmadd231pd {rn-sae}, %zmm9, %zmm10, %zmm5
+
+/* -K*L2L + Tl */
+        vmovups   L2L+__svml_dasinh_data_internal_avx512(%rip), %zmm10
+        vfnmadd231pd {rn-sae}, %zmm8, %zmm15, %zmm6
+        vfmadd213pd {rn-sae}, %zmm11, %zmm9, %zmm5
+        vfnmadd213pd {rn-sae}, %zmm7, %zmm10, %zmm8
+        vmovups   poly_coeff3+__svml_dasinh_data_internal_avx512(%rip), %zmm7
+        vmovups   poly_coeff1+__svml_dasinh_data_internal_avx512(%rip), %zmm10
+
+/* R^2 */
+        vmulpd    {rn-sae}, %zmm9, %zmm9, %zmm11
+        vfmadd213pd {rn-sae}, %zmm12, %zmm9, %zmm5
+        vfmadd213pd {rn-sae}, %zmm13, %zmm9, %zmm5
+        vfmadd213pd {rn-sae}, %zmm14, %zmm9, %zmm5
+        vfmadd213pd {rn-sae}, %zmm7, %zmm9, %zmm5
+        vmovups   poly_coeff2+__svml_dasinh_data_internal_avx512(%rip), %zmm7
+        vfmadd213pd {rn-sae}, %zmm7, %zmm9, %zmm5
+        vfmadd213pd {rn-sae}, %zmm10, %zmm9, %zmm5
+
+/* Tl + R^2*Poly */
+        vfmadd213pd {rn-sae}, %zmm8, %zmm11, %zmm5
+
+/* R+Tl + R^2*Poly */
+        vaddpd    {rn-sae}, %zmm9, %zmm5, %zmm9
+        vaddpd    {rn-sae}, %zmm9, %zmm6, %zmm4{%k2}
+        vxorpd    %zmm2, %zmm4, %zmm0
+        testl     %edx, %edx
+
+/* Go to special inputs processing branch */
+        jne       L(SPECIAL_VALUES_BRANCH)
+                                # LOE rbx r12 r13 r14 r15 edx zmm0 zmm3
+
+/* Restore registers
+ * and exit the function
+ */
+
+L(EXIT):
+        movq      %rbp, %rsp
+        popq      %rbp
+        cfi_def_cfa(7, 8)
+        cfi_restore(6)
+        ret
+        cfi_def_cfa(6, 16)
+        cfi_offset(6, -16)
+
+/* Branch to process
+ * special inputs
+ */
+
+L(SPECIAL_VALUES_BRANCH):
+        vmovups   %zmm3, 64(%rsp)
+        vmovups   %zmm0, 128(%rsp)
+                                # LOE rbx r12 r13 r14 r15 edx zmm0
+
+        xorl      %eax, %eax
+                                # LOE rbx r12 r13 r14 r15 eax edx
+
+        vzeroupper
+        movq      %r12, 16(%rsp)
+        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -176; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x50, 0xff, 0xff, 0xff, 0x22
+        movl      %eax, %r12d
+        movq      %r13, 8(%rsp)
+        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -184; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x48, 0xff, 0xff, 0xff, 0x22
+        movl      %edx, %r13d
+        movq      %r14, (%rsp)
+        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -192; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x40, 0xff, 0xff, 0xff, 0x22
+                                # LOE rbx r15 r12d r13d
+
+/* Range mask
+ * bits check
+ */
+
+L(RANGEMASK_CHECK):
+        btl       %r12d, %r13d
+
+/* Call scalar math function */
+        jc        L(SCALAR_MATH_CALL)
+                                # LOE rbx r15 r12d r13d
+
+/* Special inputs
+ * processing loop
+ */
+
+L(SPECIAL_VALUES_LOOP):
+        incl      %r12d
+        cmpl      $8, %r12d
+
+/* Check bits in range mask */
+        jl        L(RANGEMASK_CHECK)
+                                # LOE rbx r15 r12d r13d
+
+        movq      16(%rsp), %r12
+        cfi_restore(12)
+        movq      8(%rsp), %r13
+        cfi_restore(13)
+        movq      (%rsp), %r14
+        cfi_restore(14)
+        vmovups   128(%rsp), %zmm0
+
+/* Go to exit */
+        jmp       L(EXIT)
+        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -176; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x50, 0xff, 0xff, 0xff, 0x22
+        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -184; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x48, 0xff, 0xff, 0xff, 0x22
+        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -192; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x40, 0xff, 0xff, 0xff, 0x22
+                                # LOE rbx r12 r13 r14 r15 zmm0
+
+/* Scalar math fucntion call
+ * to process special input
+ */
+
+L(SCALAR_MATH_CALL):
+        movl      %r12d, %r14d
+        movsd     64(%rsp,%r14,8), %xmm0
+        call      asinh@PLT
+                                # LOE rbx r14 r15 r12d r13d xmm0
+
+        movsd     %xmm0, 128(%rsp,%r14,8)
+
+/* Process special inputs in loop */
+        jmp       L(SPECIAL_VALUES_LOOP)
+                                # LOE rbx r15 r12d r13d
+END(_ZGVeN8v_asinh_skx)
+
+        .section .rodata, "a"
+        .align 64
+
+#ifdef __svml_dasinh_data_internal_avx512_typedef
+typedef unsigned int VUINT32;
+typedef struct {
+        __declspec(align(64)) VUINT32 Log_tbl_H[16][2];
+        __declspec(align(64)) VUINT32 Log_tbl_L[16][2];
+        __declspec(align(64)) VUINT32 One[8][2];
+        __declspec(align(64)) VUINT32 AbsMask[8][2];
+        __declspec(align(64)) VUINT32 SmallThreshold[8][2];
+        __declspec(align(64)) VUINT32 Threshold[8][2];
+        __declspec(align(64)) VUINT32 LargeThreshold[8][2];
+        __declspec(align(64)) VUINT32 ca2[8][2];
+        __declspec(align(64)) VUINT32 ca1[8][2];
+        __declspec(align(64)) VUINT32 c4s[8][2];
+        __declspec(align(64)) VUINT32 c3s[8][2];
+        __declspec(align(64)) VUINT32 c2s[8][2];
+        __declspec(align(64)) VUINT32 c1s[8][2];
+        __declspec(align(64)) VUINT32 AddB5[8][2];
+        __declspec(align(64)) VUINT32 RcpBitMask[8][2];
+        __declspec(align(64)) VUINT32 OneEighth[8][2];
+        __declspec(align(64)) VUINT32 Four[8][2];
+        __declspec(align(64)) VUINT32 poly_coeff9[8][2];
+        __declspec(align(64)) VUINT32 poly_coeff8[8][2];
+        __declspec(align(64)) VUINT32 poly_coeff7[8][2];
+        __declspec(align(64)) VUINT32 poly_coeff6[8][2];
+        __declspec(align(64)) VUINT32 poly_coeff5[8][2];
+        __declspec(align(64)) VUINT32 poly_coeff4[8][2];
+        __declspec(align(64)) VUINT32 poly_coeff3[8][2];
+        __declspec(align(64)) VUINT32 poly_coeff2[8][2];
+        __declspec(align(64)) VUINT32 poly_coeff1[8][2];
+        __declspec(align(64)) VUINT32 L2H[8][2];
+        __declspec(align(64)) VUINT32 L2L[8][2];
+    } __svml_dasinh_data_internal_avx512;
+#endif
+__svml_dasinh_data_internal_avx512:
+        /*== Log_tbl_H ==*/
+        .quad 0x0000000000000000
+        .quad 0xbfaf0a30c0120000
+        .quad 0xbfbe27076e2b0000
+        .quad 0xbfc5ff3070a78000
+        .quad 0xbfcc8ff7c79a8000
+        .quad 0xbfd1675cababc000
+        .quad 0xbfd4618bc21c4000
+        .quad 0xbfd739d7f6bbc000
+        .quad 0xbfd9f323ecbf8000
+        .quad 0xbfdc8ff7c79a8000
+        .quad 0xbfdf128f5faf0000
+        .quad 0xbfe0be72e4252000
+        .quad 0xbfe1e85f5e704000
+        .quad 0xbfe307d7334f2000
+        .quad 0xbfe41d8fe8468000
+        .quad 0xbfe52a2d265bc000
+        /*== Log_tbl_L ==*/
+        .align 64
+        .quad 0x0000000000000000
+        .quad 0x3d53ab33d066d1d2
+        .quad 0x3d2a342c2af0003c
+        .quad 0xbd43d3c873e20a07
+        .quad 0xbd4a21ac25d81ef3
+        .quad 0x3d59f1fc63382a8f
+        .quad 0xbd5ec27d0b7b37b3
+        .quad 0xbd50069ce24c53fb
+        .quad 0xbd584bf2b68d766f
+        .quad 0xbd5a21ac25d81ef3
+        .quad 0xbd3bb2cd720ec44c
+        .quad 0xbd55056d312f7668
+        .quad 0xbd1a07bd8b34be7c
+        .quad 0x3d5e83c094debc15
+        .quad 0x3d5aa33736867a17
+        .quad 0xbd46abb9df22bc57
+        /*== One ==*/
+        .align 64
+        .quad 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000, 0x3ff0000000000000
+        /*== AbsMask ==*/
+        .align 64
+        .quad 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff, 0x7fffffffffffffff
+        /*== SmallThreshold ==*/
+        .align 64
+        .quad 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000, 0x3f70000000000000
+        /*== Threshold ==*/
+        .align 64
+        .quad 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000, 0x5fe0000000000000
+        /*== LargeThreshold ==*/
+        .align 64
+        .quad 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff, 0x7fefffffffffffff
+        /*== ca2 ==*/
+        .align 64
+        .quad 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7, 0x3fb333220eaf02e7
+        /*== ca1 ==*/
+        .align 64
+        .quad 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e, 0xbfc5555555521e7e
+        /*== c4s ==*/
+        .align 64
+        .quad 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612, 0x3fd1800001943612
+        /*== c3s ==*/
+        .align 64
+        .quad 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000, 0x3fd40000013b0000
+        /*== c2s ==*/
+        .align 64
+        .quad 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000, 0x3fd8000000000000
+        /*== c1s ==*/
+        .align 64
+        .quad 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000, 0x3fe0000000000000
+        /*== AddB5 ==*/
+        .align 64
+        .quad 0x0000800000000000, 0x0000800000000000, 0x0000800000000000, 0x0000800000000000, 0x0000800000000000, 0x0000800000000000, 0x0000800000000000, 0x0000800000000000
+        /*== RcpBitMask ==*/
+        .align 64
+        .quad 0xffff000000000000, 0xffff000000000000, 0xffff000000000000, 0xffff000000000000, 0xffff000000000000, 0xffff000000000000, 0xffff000000000000, 0xffff000000000000
+        /*==OneEighth ==*/
+        .align 64
+        .quad 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000, 0x3fc0000000000000
+        /*== Four ==*/
+        .align 64
+        .quad 0x4010000000000000, 0x4010000000000000, 0x4010000000000000, 0x4010000000000000, 0x4010000000000000, 0x4010000000000000, 0x4010000000000000, 0x4010000000000000
+        /*== poly_coeff9 ==*/
+        .align 64
+        .quad 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368, 0xbfb9a9b040214368
+        /*== poly_coeff8 ==*/
+        .align 64
+        .quad 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778, 0x3fbc80666e249778
+        /*== poly_coeff7 ==*/
+        .align 64
+        .quad 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9, 0xbfbffffb8a054bc9
+        /*== poly_coeff6 ==*/
+        .align 64
+        .quad 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1, 0x3fc24922f71256f1
+        /*== poly_coeff5 ==*/
+        .align 64
+        .quad 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736, 0xbfc55555559ba736
+        /*== poly_coeff4 ==*/
+        .align 64
+        .quad 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af, 0x3fc9999999be77af
+        /*== poly_coeff3 ==*/
+        .align 64
+        .quad 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65, 0xbfcffffffffffc65
+        /*== poly_coeff2 ==*/
+        .align 64
+        .quad 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1, 0x3fd55555555554c1
+        /*== poly_coeff1 ==*/
+        .align 64
+        .quad 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000, 0xbfe0000000000000
+        /*== L2H = log(2)_high ==*/
+        .align 64
+        .quad 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000, 0x3fe62E42FEFA0000
+        /*== L2L = log(2)_low ==*/
+        .align 64
+        .quad 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000, 0x3d7cf79abc9e0000
+        .align 64
+        .type	__svml_dasinh_data_internal_avx512,@object
+        .size	__svml_dasinh_data_internal_avx512,.-__svml_dasinh_data_internal_avx512
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core-avx2.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core-avx2.S
new file mode 100644
index 0000000000..7dfd95e400
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core-avx2.S
@@ -0,0 +1,20 @@ 
+/* AVX2 version of vectorized asinhf.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#define _ZGVeN16v_asinhf _ZGVeN16v_asinhf_avx2_wrapper
+#include "../svml_s_asinhf16_core.S"
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core.c b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core.c
new file mode 100644
index 0000000000..dc770a0e65
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core.c
@@ -0,0 +1,28 @@ 
+/* Multiple versions of vectorized asinhf, vector length is 16.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#define SYMBOL_NAME _ZGVeN16v_asinhf
+#include "ifunc-mathvec-avx512-skx.h"
+
+libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
+
+#ifdef SHARED
+__hidden_ver1 (_ZGVeN16v_asinhf, __GI__ZGVeN16v_asinhf,
+	       __redirect__ZGVeN16v_asinhf)
+  __attribute__ ((visibility ("hidden")));
+#endif
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core_avx512.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core_avx512.S
new file mode 100644
index 0000000000..fc6a8e7cd3
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf16_core_avx512.S
@@ -0,0 +1,476 @@ 
+/* Function asinhf vectorized with AVX-512.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   https://www.gnu.org/licenses/.  */
+
+/*
+ * ALGORITHM DESCRIPTION:
+ *
+ *   Compute asinh(x) as log(x + sqrt(x*x + 1))
+ *   using RSQRT instructions for starting the
+ *   square root approximation, and small table lookups for log
+ *   that map to AVX-512 permute instructions
+ *
+ *   Special cases:
+ *
+ *   asinh(NaN) = quiet NaN, and raise invalid exception
+ *   asinh(INF) = that INF
+ *   asinh(0)   = that 0
+ *
+ */
+
+/* Offsets for data table __svml_sasinh_data_internal_avx512
+ */
+#define Log_tbl_H                     	0
+#define Log_tbl_L                     	128
+#define One                           	256
+#define AbsMask                       	320
+#define SmallThreshold                	384
+#define Threshold                     	448
+#define LargeThreshold                	512
+#define ca1                           	576
+#define c2s                           	640
+#define c1s                           	704
+#define AddB5                         	768
+#define RcpBitMask                    	832
+#define OneEighth                     	896
+#define Four                          	960
+#define poly_coeff3                   	1024
+#define poly_coeff2                   	1088
+#define poly_coeff1                   	1152
+#define L2H                           	1216
+#define L2L                           	1280
+
+#include <sysdep.h>
+
+        .text
+	.section .text.exex512,"ax",@progbits
+ENTRY(_ZGVeN16v_asinhf_skx)
+        pushq     %rbp
+        cfi_def_cfa_offset(16)
+        movq      %rsp, %rbp
+        cfi_def_cfa(6, 16)
+        cfi_offset(6, -16)
+        andq      $-64, %rsp
+        subq      $192, %rsp
+        vmovaps   %zmm0, %zmm10
+
+/* x^2 */
+        vmulps    {rn-sae}, %zmm10, %zmm10, %zmm0
+        vmovups   One+__svml_sasinh_data_internal_avx512(%rip), %zmm2
+
+/* polynomial computation for small inputs */
+        vmovups   ca1+__svml_sasinh_data_internal_avx512(%rip), %zmm1
+
+/* not a very small input ? */
+        vmovups   SmallThreshold+__svml_sasinh_data_internal_avx512(%rip), %zmm11
+
+/* 1+x^2 */
+        vaddps    {rn-sae}, %zmm2, %zmm0, %zmm7
+
+/* |input| */
+        vandps    AbsMask+__svml_sasinh_data_internal_avx512(%rip), %zmm10, %zmm12
+
+/* A=max(x^2, 1); */
+        vmaxps    {sae}, %zmm0, %zmm2, %zmm14
+        vrsqrt14ps %zmm7, %zmm8
+
+/* B=min(x^2, 1); */
+        vminps    {sae}, %zmm0, %zmm2, %zmm15
+        vcmpps    $21, {sae}, %zmm11, %zmm12, %k2
+
+/* B_high */
+        vsubps    {rn-sae}, %zmm14, %zmm7, %zmm9
+
+/* sign bit */
+        vxorps    %zmm10, %zmm12, %zmm13
+
+/* Sh ~sqrt(1+x^2) */
+        vmulps    {rn-sae}, %zmm8, %zmm7, %zmm6
+        vmovups   LargeThreshold+__svml_sasinh_data_internal_avx512(%rip), %zmm14
+
+/* B_low */
+        vsubps    {rn-sae}, %zmm9, %zmm15, %zmm3
+
+/* Sh+x */
+        vaddps    {rn-sae}, %zmm12, %zmm6, %zmm15
+
+/* (Yh*R0)_low */
+        vfmsub213ps {rn-sae}, %zmm6, %zmm8, %zmm7
+        vmulps    {rn-sae}, %zmm1, %zmm0, %zmm9
+        vcmpps    $22, {sae}, %zmm14, %zmm12, %k0
+        vmovups   c1s+__svml_sasinh_data_internal_avx512(%rip), %zmm1
+
+/* polynomial computation for small inputs */
+        vfmadd213ps {rn-sae}, %zmm12, %zmm12, %zmm9
+        kmovw     %k0, %edx
+
+/* (x^2)_low */
+        vmovaps   %zmm10, %zmm4
+        vfmsub213ps {rn-sae}, %zmm0, %zmm10, %zmm4
+
+/* Yl = (x^2)_low + B_low */
+        vaddps    {rn-sae}, %zmm4, %zmm3, %zmm5
+
+/* rel. error term: Eh=1-Sh*R0 */
+        vmovaps   %zmm2, %zmm0
+        vfnmadd231ps {rn-sae}, %zmm6, %zmm8, %zmm0
+
+/* Sl = (Yh*R0)_low+(R0*Yl) */
+        vfmadd213ps {rn-sae}, %zmm7, %zmm8, %zmm5
+
+/* very large inputs ? */
+        vmovups   Threshold+__svml_sasinh_data_internal_avx512(%rip), %zmm7
+
+/* rel. error term: Eh=(1-Sh*R0)-Sl*R0 */
+        vfnmadd231ps {rn-sae}, %zmm5, %zmm8, %zmm0
+
+/* sqrt(1+x^2) ~ Sh + Sl + Sh*Eh*poly_s */
+        vmovups   c2s+__svml_sasinh_data_internal_avx512(%rip), %zmm8
+        vcmpps    $21, {sae}, %zmm7, %zmm12, %k1
+
+/* Sh*Eh */
+        vmulps    {rn-sae}, %zmm0, %zmm6, %zmm4
+        vfmadd231ps {rn-sae}, %zmm0, %zmm8, %zmm1
+
+/* Sl + Sh*Eh*poly_s */
+        vfmadd213ps {rn-sae}, %zmm5, %zmm1, %zmm4
+
+/* Xh */
+        vsubps    {rn-sae}, %zmm6, %zmm15, %zmm5
+
+/* fixup for very large inputs */
+        vmovups   OneEighth+__svml_sasinh_data_internal_avx512(%rip), %zmm6
+
+/* Xin0+Sl+Sh*Eh*poly_s ~ x+sqrt(1+x^2) */
+        vaddps    {rn-sae}, %zmm4, %zmm15, %zmm3
+
+/* Xl */
+        vsubps    {rn-sae}, %zmm5, %zmm12, %zmm5
+
+/* Sl_high */
+        vsubps    {rn-sae}, %zmm15, %zmm3, %zmm0
+        vmulps    {rn-sae}, %zmm6, %zmm12, %zmm3{%k1}
+
+/* -K*L2H + Th */
+        vmovups   L2H+__svml_sasinh_data_internal_avx512(%rip), %zmm15
+
+/* Sl_l */
+        vsubps    {rn-sae}, %zmm0, %zmm4, %zmm1
+        vrcp14ps  %zmm3, %zmm6
+
+/* Table lookups */
+        vmovups   __svml_sasinh_data_internal_avx512(%rip), %zmm0
+
+/* Xin_low */
+        vaddps    {rn-sae}, %zmm5, %zmm1, %zmm7
+
+/* round reciprocal to 1+4b mantissas */
+        vpaddd    AddB5+__svml_sasinh_data_internal_avx512(%rip), %zmm6, %zmm4
+        vmovups   poly_coeff1+__svml_sasinh_data_internal_avx512(%rip), %zmm5
+        vandps    RcpBitMask+__svml_sasinh_data_internal_avx512(%rip), %zmm4, %zmm8
+
+/* fixup for very large inputs */
+        vxorps    %zmm7, %zmm7, %zmm7{%k1}
+
+/* polynomial */
+        vmovups   poly_coeff3+__svml_sasinh_data_internal_avx512(%rip), %zmm4
+
+/* reduced argument for log(): (Rcp*Xin-1)+Rcp*Xin_low */
+        vfmsub231ps {rn-sae}, %zmm8, %zmm3, %zmm2
+        vmovups   Four+__svml_sasinh_data_internal_avx512(%rip), %zmm3
+
+/* exponents */
+        vgetexpps {sae}, %zmm8, %zmm1
+
+/* Prepare table index */
+        vpsrld    $18, %zmm8, %zmm14
+        vfmadd231ps {rn-sae}, %zmm8, %zmm7, %zmm2
+        vmovups   poly_coeff2+__svml_sasinh_data_internal_avx512(%rip), %zmm7
+        vsubps    {rn-sae}, %zmm3, %zmm1, %zmm1{%k1}
+        vpermt2ps Log_tbl_H+64+__svml_sasinh_data_internal_avx512(%rip), %zmm14, %zmm0
+        vmovups   Log_tbl_L+__svml_sasinh_data_internal_avx512(%rip), %zmm3
+        vfmadd231ps {rn-sae}, %zmm2, %zmm4, %zmm7
+        vfnmadd231ps {rn-sae}, %zmm1, %zmm15, %zmm0
+
+/* R^2 */
+        vmulps    {rn-sae}, %zmm2, %zmm2, %zmm6
+        vfmadd213ps {rn-sae}, %zmm5, %zmm2, %zmm7
+        vpermt2ps Log_tbl_L+64+__svml_sasinh_data_internal_avx512(%rip), %zmm14, %zmm3
+
+/* -K*L2L + Tl */
+        vmovups   L2L+__svml_sasinh_data_internal_avx512(%rip), %zmm14
+        vfnmadd213ps {rn-sae}, %zmm3, %zmm14, %zmm1
+
+/* Tl + R^2*Poly */
+        vfmadd213ps {rn-sae}, %zmm1, %zmm6, %zmm7
+
+/* R+Tl + R^2*Poly */
+        vaddps    {rn-sae}, %zmm2, %zmm7, %zmm2
+        vaddps    {rn-sae}, %zmm2, %zmm0, %zmm9{%k2}
+        vxorps    %zmm13, %zmm9, %zmm0
+        testl     %edx, %edx
+
+/* Go to special inputs processing branch */
+        jne       L(SPECIAL_VALUES_BRANCH)
+                                # LOE rbx r12 r13 r14 r15 edx zmm0 zmm10
+
+/* Restore registers
+ * and exit the function
+ */
+
+L(EXIT):
+        movq      %rbp, %rsp
+        popq      %rbp
+        cfi_def_cfa(7, 8)
+        cfi_restore(6)
+        ret
+        cfi_def_cfa(6, 16)
+        cfi_offset(6, -16)
+
+/* Branch to process
+ * special inputs
+ */
+
+L(SPECIAL_VALUES_BRANCH):
+        vmovups   %zmm10, 64(%rsp)
+        vmovups   %zmm0, 128(%rsp)
+                                # LOE rbx r12 r13 r14 r15 edx zmm0
+
+        xorl      %eax, %eax
+                                # LOE rbx r12 r13 r14 r15 eax edx
+
+        vzeroupper
+        movq      %r12, 16(%rsp)
+        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -176; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x50, 0xff, 0xff, 0xff, 0x22
+        movl      %eax, %r12d
+        movq      %r13, 8(%rsp)
+        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -184; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x48, 0xff, 0xff, 0xff, 0x22
+        movl      %edx, %r13d
+        movq      %r14, (%rsp)
+        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -192; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x40, 0xff, 0xff, 0xff, 0x22
+                                # LOE rbx r15 r12d r13d
+
+/* Range mask
+ * bits check
+ */
+
+L(RANGEMASK_CHECK):
+        btl       %r12d, %r13d
+
+/* Call scalar math function */
+        jc        L(SCALAR_MATH_CALL)
+                                # LOE rbx r15 r12d r13d
+
+/* Special inputs
+ * processing loop
+ */
+
+L(SPECIAL_VALUES_LOOP):
+        incl      %r12d
+        cmpl      $16, %r12d
+
+/* Check bits in range mask */
+        jl        L(RANGEMASK_CHECK)
+                                # LOE rbx r15 r12d r13d
+
+        movq      16(%rsp), %r12
+        cfi_restore(12)
+        movq      8(%rsp), %r13
+        cfi_restore(13)
+        movq      (%rsp), %r14
+        cfi_restore(14)
+        vmovups   128(%rsp), %zmm0
+
+/* Go to exit */
+        jmp       L(EXIT)
+        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -176; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x50, 0xff, 0xff, 0xff, 0x22
+        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -184; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x48, 0xff, 0xff, 0xff, 0x22
+        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -64; DW_OP_and; DW_OP_const4s: -192; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xc0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0x40, 0xff, 0xff, 0xff, 0x22
+                                # LOE rbx r12 r13 r14 r15 zmm0
+
+/* Scalar math fucntion call
+ * to process special input
+ */
+
+L(SCALAR_MATH_CALL):
+        movl      %r12d, %r14d
+        movss     64(%rsp,%r14,4), %xmm0
+        call      asinhf@PLT
+                                # LOE rbx r14 r15 r12d r13d xmm0
+
+        movss     %xmm0, 128(%rsp,%r14,4)
+
+/* Process special inputs in loop */
+        jmp       L(SPECIAL_VALUES_LOOP)
+                                # LOE rbx r15 r12d r13d
+END(_ZGVeN16v_asinhf_skx)
+
+        .section .rodata, "a"
+        .align 64
+
+#ifdef __svml_sasinh_data_internal_avx512_typedef
+typedef unsigned int VUINT32;
+typedef struct {
+        __declspec(align(64)) VUINT32 Log_tbl_H[32][1];
+        __declspec(align(64)) VUINT32 Log_tbl_L[32][1];
+        __declspec(align(64)) VUINT32 One[16][1];
+        __declspec(align(64)) VUINT32 AbsMask[16][1];
+        __declspec(align(64)) VUINT32 SmallThreshold[16][1];
+        __declspec(align(64)) VUINT32 Threshold[16][1];
+        __declspec(align(64)) VUINT32 LargeThreshold[16][1];
+        __declspec(align(64)) VUINT32 ca1[16][1];
+        __declspec(align(64)) VUINT32 c2s[16][1];
+        __declspec(align(64)) VUINT32 c1s[16][1];
+        __declspec(align(64)) VUINT32 AddB5[16][1];
+        __declspec(align(64)) VUINT32 RcpBitMask[16][1];
+        __declspec(align(64)) VUINT32 OneEighth[16][1];
+        __declspec(align(64)) VUINT32 Four[16][1];
+        __declspec(align(64)) VUINT32 poly_coeff3[16][1];
+        __declspec(align(64)) VUINT32 poly_coeff2[16][1];
+        __declspec(align(64)) VUINT32 poly_coeff1[16][1];
+        __declspec(align(64)) VUINT32 L2H[16][1];
+        __declspec(align(64)) VUINT32 L2L[16][1];
+    } __svml_sasinh_data_internal_avx512;
+#endif
+__svml_sasinh_data_internal_avx512:
+        /*== Log_tbl_H ==*/
+        .long 0x00000000
+        .long 0xbcfc0000
+        .long 0xbd788000
+        .long 0xbdb78000
+        .long 0xbdf14000
+        .long 0xbe14a000
+        .long 0xbe300000
+        .long 0xbe4aa000
+        .long 0xbe648000
+        .long 0xbe7dc000
+        .long 0xbe8b4000
+        .long 0xbe974000
+        .long 0xbea31000
+        .long 0xbeae9000
+        .long 0xbeb9d000
+        .long 0xbec4d000
+        .long 0xbecfa000
+        .long 0xbeda2000
+        .long 0xbee48000
+        .long 0xbeeea000
+        .long 0xbef89000
+        .long 0xbf012800
+        .long 0xbf05f000
+        .long 0xbf0aa800
+        .long 0xbf0f4000
+        .long 0xbf13c800
+        .long 0xbf184000
+        .long 0xbf1ca000
+        .long 0xbf20f000
+        .long 0xbf252800
+        .long 0xbf295000
+        .long 0xbf2d6800
+        /*== Log_tbl_L ==*/
+        .align 64
+        .long 0x80000000
+        .long 0xb726c39e
+        .long 0x3839e7fe
+        .long 0xb7528ae5
+        .long 0x377891d5
+        .long 0xb8297c10
+        .long 0x37cf8f58
+        .long 0x3852b186
+        .long 0x35838656
+        .long 0xb80c36af
+        .long 0x38235454
+        .long 0xb862bae1
+        .long 0x37e87bc7
+        .long 0x37848150
+        .long 0x37202511
+        .long 0xb74e1b05
+        .long 0x385c1340
+        .long 0xb8777bcd
+        .long 0x36038656
+        .long 0xb7d40984
+        .long 0xb80f5faf
+        .long 0xb8254b4c
+        .long 0xb865c84a
+        .long 0x37f0b42d
+        .long 0xb83ebce1
+        .long 0xb83c2513
+        .long 0x37a332c4
+        .long 0x3779654f
+        .long 0x38602f73
+        .long 0x367449f8
+        .long 0xb7b4996f
+        .long 0xb800986b
+        /*== One ==*/
+        .align 64
+        .long 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000
+        /*== AbsMask ==*/
+        .align 64
+        .long 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff
+        /*== SmallThreshold ==*/
+        .align 64
+        .long 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000, 0x3c800000
+        /*== Threshold ==*/
+        .align 64
+        .long 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000, 0x5f000000
+        /*== LargeThreshold ==*/
+        .align 64
+        .long 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff, 0x7f7fffff
+        /*== ca1 ==*/
+        .align 64
+        .long 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE, 0xbe2AA5DE
+        /*== c2s ==*/
+        .align 64
+        .long 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000, 0x3ec00000
+        /*== c1s ==*/
+        .align 64
+        .long 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000, 0x3f000000
+        /*== AddB5 ==*/
+        .align 64
+        .long 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000, 0x00020000
+        /*== RcpBitMask ==*/
+        .align 64
+        .long 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000, 0xfffc0000
+        /*==OneEighth ==*/
+        .align 64
+        .long 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000, 0x3e000000
+        /*== Four ==*/
+        .align 64
+        .long 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000, 0x40800000
+        /*== poly_coeff3 ==*/
+        .align 64
+        .long 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810, 0xbe800810
+        /*== poly_coeff2 ==*/
+        .align 64
+        .long 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e, 0x3eaab11e
+        /*== poly_coeff1 ==*/
+        .align 64
+        .long 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000
+        /*== L2H = log(2)_high ==*/
+        .align 64
+        .long 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000, 0x3f317000
+        /*== L2L = log(2)_low ==*/
+        .align 64
+        .long 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4, 0x3805fdf4
+        .align 64
+        .type	__svml_sasinh_data_internal_avx512,@object
+        .size	__svml_sasinh_data_internal_avx512,.-__svml_sasinh_data_internal_avx512
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core-sse2.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core-sse2.S
new file mode 100644
index 0000000000..52e4d2f728
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core-sse2.S
@@ -0,0 +1,20 @@ 
+/* SSE2 version of vectorized asinhf, vector length is 4.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#define _ZGVbN4v_asinhf _ZGVbN4v_asinhf_sse2
+#include "../svml_s_asinhf4_core.S"
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core.c b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core.c
new file mode 100644
index 0000000000..296d5754ae
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core.c
@@ -0,0 +1,28 @@ 
+/* Multiple versions of vectorized asinhf, vector length is 4.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#define SYMBOL_NAME _ZGVbN4v_asinhf
+#include "ifunc-mathvec-sse4_1.h"
+
+libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
+
+#ifdef SHARED
+__hidden_ver1 (_ZGVbN4v_asinhf, __GI__ZGVbN4v_asinhf,
+	       __redirect__ZGVbN4v_asinhf)
+  __attribute__ ((visibility ("hidden")));
+#endif
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core_sse4.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core_sse4.S
new file mode 100644
index 0000000000..1eeeb4f5af
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf4_core_sse4.S
@@ -0,0 +1,509 @@ 
+/* Function asinhf vectorized with SSE4.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   https://www.gnu.org/licenses/.  */
+
+/*
+ * ALGORITHM DESCRIPTION:
+ *
+ *   Compute asinh(x) as log(x + sqrt(x*x + 1))
+ *
+ *   Special cases:
+ *
+ *   asinh(NaN) = quiet NaN, and raise invalid exception
+ *   asinh(INF) = that INF
+ *   asinh(0)   = that 0
+ *
+ */
+
+/* Offsets for data table __svml_sasinh_data_internal
+ */
+#define SgnMask                       	0
+#define sOne                          	16
+#define sPoly                         	32
+#define iBrkValue                     	160
+#define iOffExpoMask                  	176
+#define sBigThreshold                 	192
+#define sC2                           	208
+#define sC3                           	224
+#define sHalf                         	240
+#define sLargestFinite                	256
+#define sLittleThreshold              	272
+#define sSign                         	288
+#define sThirtyOne                    	304
+#define sTopMask11                    	320
+#define sTopMask8                     	336
+#define XScale                        	352
+#define sLn2                          	368
+
+#include <sysdep.h>
+
+        .text
+	.section .text.sse4,"ax",@progbits
+ENTRY(_ZGVbN4v_asinhf_sse4)
+        subq      $72, %rsp
+        cfi_def_cfa_offset(80)
+        movaps    %xmm0, %xmm8
+
+/*
+ * Split X into high and low parts, XHi (<= 11 bits) and XLo (<= 13 bits)
+ * We could use either X or |X| here, but it doesn't seem to matter
+ */
+        movups    sTopMask11+__svml_sasinh_data_internal(%rip), %xmm10
+        movaps    %xmm8, %xmm2
+        andps     %xmm8, %xmm10
+
+/*
+ * Compute X^2 = (XHi + XLo)^2 = XHi^2 + XLo * (X + XHi)
+ * The two parts are shifted off by around 11 bits. So even though
+ * the low bit will not in general be exact, it's near enough
+ */
+        movaps    %xmm10, %xmm3
+        subps     %xmm10, %xmm2
+        mulps     %xmm10, %xmm3
+        addps     %xmm8, %xmm10
+
+/* Load the constant 1 and a sign mask */
+        movups    sOne+__svml_sasinh_data_internal(%rip), %xmm7
+
+/*
+ * Finally, express Y + W = X^2 + 1 accurately where Y has <= 8 bits.
+ * If |X| <= 1 then |XHi| <= 1 and so |X2Hi| <= 1, so we can treat 1
+ * as the dominant component in the compensated summation. Otherwise,
+ * if |X| >= 1, then since X2Hi only has 22 significant bits, the basic
+ * addition will be exact anyway until we get to |X| >= 2^24. But by
+ * that time the log function is well-conditioned enough that the
+ * rounding error doesn't matter. Hence we can treat 1 as dominant even
+ * if it literally isn't.
+ */
+        movaps    %xmm7, %xmm11
+        movaps    %xmm7, %xmm4
+        movups    sTopMask8+__svml_sasinh_data_internal(%rip), %xmm12
+        addps     %xmm3, %xmm11
+        mulps     %xmm10, %xmm2
+        subps     %xmm11, %xmm4
+        movaps    %xmm12, %xmm0
+        addps     %xmm3, %xmm4
+
+/*
+ * Unfortunately, we can still be in trouble if |X| <= 2^-5, since
+ * the absolute error 2^-(7+24)-ish in sqrt(1 + X^2) gets scaled up
+ * by 1/X and comes close to our threshold. Hence if |X| <= 2^-4,
+ * perform an alternative computation
+ * sqrt(1 + X^2) - 1 = X^2/2 - X^4/8 + X^6/16
+ * X2 = X^2
+ */
+        addps     %xmm2, %xmm3
+        addps     %xmm2, %xmm4
+        andps     %xmm11, %xmm0
+
+/*
+ * Compute R = 1/sqrt(Y + W) * (1 + d)
+ * Force R to <= 8 significant bits.
+ * This means that R * Y and R^2 * Y are exactly representable.
+ */
+        rsqrtps   %xmm0, %xmm14
+        subps     %xmm0, %xmm11
+        andps     %xmm12, %xmm14
+        addps     %xmm11, %xmm4
+
+/*
+ * Compute S = (Y/sqrt(Y + W)) * (1 + d)
+ * and T = (W/sqrt(Y + W)) * (1 + d)
+ * so that S + T = sqrt(Y + W) * (1 + d)
+ * S is exact, and the rounding error in T is OK.
+ */
+        mulps     %xmm14, %xmm0
+        mulps     %xmm14, %xmm4
+
+/*
+ * Get the absolute value of the input, since we will exploit antisymmetry
+ * and mostly assume X >= 0 in the core computation
+ */
+        movups    SgnMask+__svml_sasinh_data_internal(%rip), %xmm6
+
+/*
+ * Compute e = -(2 * d + d^2)
+ * The first FMR is exact, and the rounding error in the other is acceptable
+ * since d and e are ~ 2^-8
+ */
+        movaps    %xmm14, %xmm13
+        andps     %xmm8, %xmm6
+
+/*
+ * Obtain sqrt(1 + X^2) - 1 in two pieces
+ * sqrt(1 + X^2) - 1
+ * = sqrt(Y + W) - 1
+ * = (S + T) * (1 + Corr) - 1
+ * = [S - 1] + [T + (S + T) * Corr]
+ * We need a compensated summation for the last part. We treat S - 1
+ * as the larger part; it certainly is until about X < 2^-4, and in that
+ * case, the error is affordable since X dominates over sqrt(1 + X^2) - 1
+ * Final sum is dTmp5 (hi) + dTmp7 (lo)
+ */
+        movaps    %xmm0, %xmm1
+
+/*
+ * Check whether the input is finite, by checking |X| <= MaxFloat
+ * Otherwise set the rangemask so that the callout will get used.
+ * Note that this will also use the callout for NaNs since not(NaN <= MaxFloat)
+ */
+        movaps    %xmm6, %xmm9
+
+/*
+ * The following computation can go wrong for very large X, basically
+ * because X^2 overflows. But for large X we have
+ * asinh(X) / log(2 X) - 1 =~= 1/(4 * X^2), so for X >= 2^30
+ * we can just later stick X back into the log and tweak up the exponent.
+ * Actually we scale X by 2^-30 and tweak the exponent up by 31,
+ * to stay in the safe range for the later log computation.
+ * Compute a flag now telling us when do do this.
+ */
+        movaps    %xmm6, %xmm5
+        cmpnleps  sLargestFinite+__svml_sasinh_data_internal(%rip), %xmm9
+        cmpltps   sBigThreshold+__svml_sasinh_data_internal(%rip), %xmm5
+        mulps     %xmm0, %xmm13
+        addps     %xmm4, %xmm1
+        subps     %xmm7, %xmm0
+        mulps     %xmm4, %xmm14
+        movmskps  %xmm9, %edx
+        movaps    %xmm7, %xmm9
+
+/*
+ * Now       1 / (1 + d)
+ * = 1 / (1 + (sqrt(1 - e) - 1))
+ * = 1 / sqrt(1 - e)
+ * = 1 + 1/2 * e + 3/8 * e^2 + 5/16 * e^3 + 35/128 * e^4 + ...
+ * So compute the first three nonconstant terms of that, so that
+ * we have a relative correction (1 + Corr) to apply to S etc.
+ * C1 = 1/2
+ * C2 = 3/8
+ * C3 = 5/16
+ */
+        movups    sC3+__svml_sasinh_data_internal(%rip), %xmm15
+        subps     %xmm13, %xmm9
+        movups    sHalf+__svml_sasinh_data_internal(%rip), %xmm10
+        subps     %xmm14, %xmm9
+
+/* sX2over2 = X^2/2 */
+        mulps     %xmm10, %xmm3
+        mulps     %xmm9, %xmm15
+
+/* sX46 = -X^4/4 + X^6/8 */
+        movaps    %xmm3, %xmm2
+        movaps    %xmm3, %xmm12
+
+/*
+ * Now do another compensated sum to add |X| + [sqrt(1 + X^2) - 1].
+ * It's always safe to assume |X| is larger.
+ * This is the final 2-part argument to the log1p function
+ */
+        movaps    %xmm6, %xmm14
+        addps     sC2+__svml_sasinh_data_internal(%rip), %xmm15
+        mulps     %xmm9, %xmm15
+        addps     %xmm10, %xmm15
+        mulps     %xmm15, %xmm9
+        mulps     %xmm1, %xmm9
+
+/* Now multiplex to the case X = 2^-30 * input, Xl = sL = 0 in the "big" case. */
+        movups    XScale+__svml_sasinh_data_internal(%rip), %xmm15
+        addps     %xmm9, %xmm4
+        movaps    %xmm4, %xmm11
+        addps     %xmm0, %xmm11
+        subps     %xmm11, %xmm0
+        addps     %xmm0, %xmm4
+
+/* sX4over4 = X^4/4 */
+        movaps    %xmm3, %xmm0
+        mulps     %xmm3, %xmm0
+        mulps     %xmm0, %xmm2
+        subps     %xmm0, %xmm2
+
+/*
+ * Now we feed into the log1p code, using H in place of _VARG1 and
+ * also adding L into Xl.
+ * compute 1+x as high, low parts
+ */
+        movaps    %xmm7, %xmm0
+
+/* sX46over2 = -X^4/8 + x^6/16 */
+        mulps     %xmm2, %xmm10
+        movaps    %xmm7, %xmm2
+        addps     %xmm10, %xmm12
+        subps     %xmm12, %xmm3
+        addps     %xmm3, %xmm10
+
+/* Now multiplex the two possible computations */
+        movaps    %xmm6, %xmm3
+        cmpleps   sLittleThreshold+__svml_sasinh_data_internal(%rip), %xmm3
+        movaps    %xmm3, %xmm13
+        andps     %xmm3, %xmm12
+        andnps    %xmm11, %xmm13
+        movaps    %xmm3, %xmm1
+        orps      %xmm12, %xmm13
+        andnps    %xmm4, %xmm1
+        andps     %xmm3, %xmm10
+        movaps    %xmm6, %xmm4
+        orps      %xmm10, %xmm1
+        addps     %xmm13, %xmm14
+        mulps     %xmm15, %xmm6
+        maxps     %xmm14, %xmm0
+        minps     %xmm14, %xmm2
+        subps     %xmm14, %xmm4
+        movaps    %xmm0, %xmm3
+        addps     %xmm4, %xmm13
+        addps     %xmm2, %xmm3
+        addps     %xmm13, %xmm1
+        subps     %xmm3, %xmm0
+        movaps    %xmm5, %xmm4
+        andps     %xmm5, %xmm3
+        andnps    %xmm6, %xmm4
+        addps     %xmm0, %xmm2
+
+/*
+ * Now resume the main code.
+ * reduction: compute r,n
+ */
+        movdqu    iBrkValue+__svml_sasinh_data_internal(%rip), %xmm6
+        orps      %xmm3, %xmm4
+        psubd     %xmm6, %xmm4
+        movaps    %xmm7, %xmm0
+        addps     %xmm2, %xmm1
+        movdqu    iOffExpoMask+__svml_sasinh_data_internal(%rip), %xmm2
+        pand      %xmm4, %xmm2
+        psrad     $23, %xmm4
+        cvtdq2ps  %xmm4, %xmm3
+        pslld     $23, %xmm4
+        andps     %xmm5, %xmm1
+        paddd     %xmm6, %xmm2
+        psubd     %xmm4, %xmm0
+        mulps     %xmm0, %xmm1
+
+/* polynomial evaluation */
+        subps     %xmm7, %xmm2
+        movups    sPoly+112+__svml_sasinh_data_internal(%rip), %xmm7
+        addps     %xmm2, %xmm1
+        mulps     %xmm1, %xmm7
+        movaps    %xmm5, %xmm2
+
+/* Add 31 to the exponent in the "large" case to get log(2 * input) */
+        movups    sThirtyOne+__svml_sasinh_data_internal(%rip), %xmm0
+        addps     sPoly+96+__svml_sasinh_data_internal(%rip), %xmm7
+        addps     %xmm3, %xmm0
+        mulps     %xmm1, %xmm7
+        andnps    %xmm0, %xmm2
+        andps     %xmm5, %xmm3
+        orps      %xmm3, %xmm2
+        addps     sPoly+80+__svml_sasinh_data_internal(%rip), %xmm7
+
+/* final reconstruction */
+        mulps     sLn2+__svml_sasinh_data_internal(%rip), %xmm2
+        mulps     %xmm1, %xmm7
+
+/* Finally, reincorporate the original sign. */
+        movups    sSign+__svml_sasinh_data_internal(%rip), %xmm0
+        andps     %xmm8, %xmm0
+        addps     sPoly+64+__svml_sasinh_data_internal(%rip), %xmm7
+        mulps     %xmm1, %xmm7
+        addps     sPoly+48+__svml_sasinh_data_internal(%rip), %xmm7
+        mulps     %xmm1, %xmm7
+        addps     sPoly+32+__svml_sasinh_data_internal(%rip), %xmm7
+        mulps     %xmm1, %xmm7
+        addps     sPoly+16+__svml_sasinh_data_internal(%rip), %xmm7
+        mulps     %xmm1, %xmm7
+        addps     sPoly+__svml_sasinh_data_internal(%rip), %xmm7
+        mulps     %xmm1, %xmm7
+        mulps     %xmm1, %xmm7
+        addps     %xmm7, %xmm1
+        addps     %xmm2, %xmm1
+        pxor      %xmm1, %xmm0
+        testl     %edx, %edx
+
+/* Go to special inputs processing branch */
+        jne       L(SPECIAL_VALUES_BRANCH)
+                                # LOE rbx rbp r12 r13 r14 r15 edx xmm0 xmm8
+
+/* Restore registers
+ * and exit the function
+ */
+
+L(EXIT):
+        addq      $72, %rsp
+        cfi_def_cfa_offset(8)
+        ret
+        cfi_def_cfa_offset(80)
+
+/* Branch to process
+ * special inputs
+ */
+
+L(SPECIAL_VALUES_BRANCH):
+        movups    %xmm8, 32(%rsp)
+        movups    %xmm0, 48(%rsp)
+                                # LOE rbx rbp r12 r13 r14 r15 edx
+
+        xorl      %eax, %eax
+        movq      %r12, 16(%rsp)
+        cfi_offset(12, -64)
+        movl      %eax, %r12d
+        movq      %r13, 8(%rsp)
+        cfi_offset(13, -72)
+        movl      %edx, %r13d
+        movq      %r14, (%rsp)
+        cfi_offset(14, -80)
+                                # LOE rbx rbp r15 r12d r13d
+
+/* Range mask
+ * bits check
+ */
+
+L(RANGEMASK_CHECK):
+        btl       %r12d, %r13d
+
+/* Call scalar math function */
+        jc        L(SCALAR_MATH_CALL)
+                                # LOE rbx rbp r15 r12d r13d
+
+/* Special inputs
+ * processing loop
+ */
+
+L(SPECIAL_VALUES_LOOP):
+        incl      %r12d
+        cmpl      $4, %r12d
+
+/* Check bits in range mask */
+        jl        L(RANGEMASK_CHECK)
+                                # LOE rbx rbp r15 r12d r13d
+
+        movq      16(%rsp), %r12
+        cfi_restore(12)
+        movq      8(%rsp), %r13
+        cfi_restore(13)
+        movq      (%rsp), %r14
+        cfi_restore(14)
+        movups    48(%rsp), %xmm0
+
+/* Go to exit */
+        jmp       L(EXIT)
+        cfi_offset(12, -64)
+        cfi_offset(13, -72)
+        cfi_offset(14, -80)
+                                # LOE rbx rbp r12 r13 r14 r15 xmm0
+
+/* Scalar math fucntion call
+ * to process special input
+ */
+
+L(SCALAR_MATH_CALL):
+        movl      %r12d, %r14d
+        movss     32(%rsp,%r14,4), %xmm0
+        call      asinhf@PLT
+                                # LOE rbx rbp r14 r15 r12d r13d xmm0
+
+        movss     %xmm0, 48(%rsp,%r14,4)
+
+/* Process special inputs in loop */
+        jmp       L(SPECIAL_VALUES_LOOP)
+                                # LOE rbx rbp r15 r12d r13d
+END(_ZGVbN4v_asinhf_sse4)
+
+        .section .rodata, "a"
+        .align 16
+
+#ifdef __svml_sasinh_data_internal_typedef
+typedef unsigned int VUINT32;
+typedef struct {
+        __declspec(align(16)) VUINT32 SgnMask[4][1];
+        __declspec(align(16)) VUINT32 sOne[4][1];
+        __declspec(align(16)) VUINT32 sPoly[8][4][1];
+        __declspec(align(16)) VUINT32 iBrkValue[4][1];
+        __declspec(align(16)) VUINT32 iOffExpoMask[4][1];
+        __declspec(align(16)) VUINT32 sBigThreshold[4][1];
+        __declspec(align(16)) VUINT32 sC2[4][1];
+        __declspec(align(16)) VUINT32 sC3[4][1];
+        __declspec(align(16)) VUINT32 sHalf[4][1];
+        __declspec(align(16)) VUINT32 sLargestFinite[4][1];
+        __declspec(align(16)) VUINT32 sLittleThreshold[4][1];
+        __declspec(align(16)) VUINT32 sSign[4][1];
+        __declspec(align(16)) VUINT32 sThirtyOne[4][1];
+        __declspec(align(16)) VUINT32 sTopMask11[4][1];
+        __declspec(align(16)) VUINT32 sTopMask8[4][1];
+        __declspec(align(16)) VUINT32 XScale[4][1];
+        __declspec(align(16)) VUINT32 sLn2[4][1];
+} __svml_sasinh_data_internal;
+#endif
+__svml_sasinh_data_internal:
+        /*== SgnMask ==*/
+        .long 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff
+        /*== sOne = SP 1.0 ==*/
+        .align 16
+        .long 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000
+        /*== sPoly[] = SP polynomial ==*/
+        .align 16
+        .long 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000 /* -5.0000000000000000000000000e-01 P0 */
+        .long 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94 /*  3.3333265781402587890625000e-01 P1 */
+        .long 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e /* -2.5004237890243530273437500e-01 P2 */
+        .long 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190 /*  2.0007920265197753906250000e-01 P3 */
+        .long 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37 /* -1.6472326219081878662109375e-01 P4 */
+        .long 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12 /*  1.4042308926582336425781250e-01 P5 */
+        .long 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3 /* -1.5122179687023162841796875e-01 P6 */
+        .long 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed /*  1.3820238411426544189453125e-01 P7 */
+        /*== iBrkValue = SP 2/3 ==*/
+        .align 16
+        .long 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab
+        /*== iOffExpoMask = SP significand mask ==*/
+        .align 16
+        .long 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff
+        /*== sBigThreshold ==*/
+        .align 16
+        .long 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000
+        /*== sC2 ==*/
+        .align 16
+        .long 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000
+        /*== sC3 ==*/
+        .align 16
+        .long 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000
+        /*== sHalf ==*/
+        .align 16
+        .long 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000
+        /*== sLargestFinite ==*/
+        .align 16
+        .long 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF
+        /*== sLittleThreshold ==*/
+        .align 16
+        .long 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000
+        /*== sSign ==*/
+        .align 16
+        .long 0x80000000, 0x80000000, 0x80000000, 0x80000000
+        /*== sThirtyOne ==*/
+        .align 16
+        .long 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000
+        /*== sTopMask11 ==*/
+        .align 16
+        .long 0xFFFFE000, 0xFFFFE000, 0xFFFFE000, 0xFFFFE000
+        /*== sTopMask8 ==*/
+        .align 16
+        .long 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000
+        /*== XScale ==*/
+        .align 16
+        .long 0x30800000, 0x30800000, 0x30800000, 0x30800000
+        /*== sLn2 = SP ln(2) ==*/
+        .align 16
+        .long 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218
+        .align 16
+        .type	__svml_sasinh_data_internal,@object
+        .size	__svml_sasinh_data_internal,.-__svml_sasinh_data_internal
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core-sse.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core-sse.S
new file mode 100644
index 0000000000..1a0e113e94
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core-sse.S
@@ -0,0 +1,20 @@ 
+/* SSE version of vectorized asinhf, vector length is 8.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+    Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#define _ZGVdN8v_asinhf _ZGVdN8v_asinhf_sse_wrapper
+#include "../svml_s_asinhf8_core.S"
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core.c b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core.c
new file mode 100644
index 0000000000..d97097a394
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core.c
@@ -0,0 +1,28 @@ 
+/* Multiple versions of vectorized asinhf, vector length is 8.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#define SYMBOL_NAME _ZGVdN8v_asinhf
+#include "ifunc-mathvec-avx2.h"
+
+libc_ifunc_redirected (REDIRECT_NAME, SYMBOL_NAME, IFUNC_SELECTOR ());
+
+#ifdef SHARED
+__hidden_ver1 (_ZGVdN8v_asinhf, __GI__ZGVdN8v_asinhf,
+	       __redirect__ZGVdN8v_asinhf)
+  __attribute__ ((visibility ("hidden")));
+#endif
diff --git a/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core_avx2.S b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core_avx2.S
new file mode 100644
index 0000000000..a966f53773
--- /dev/null
+++ b/sysdeps/x86_64/fpu/multiarch/svml_s_asinhf8_core_avx2.S
@@ -0,0 +1,457 @@ 
+/* Function asinhf vectorized with AVX2.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   https://www.gnu.org/licenses/.  */
+
+/*
+ * ALGORITHM DESCRIPTION:
+ *
+ *   Compute asinh(x) as log(x + sqrt(x*x + 1))
+ *
+ *   Special cases:
+ *
+ *   asinh(NaN) = quiet NaN, and raise invalid exception
+ *   asinh(INF) = that INF
+ *   asinh(0)   = that 0
+ *
+ */
+
+/* Offsets for data table __svml_sasinh_data_internal
+ */
+#define SgnMask                       	0
+#define sOne                          	32
+#define sPoly                         	64
+#define iBrkValue                     	320
+#define iOffExpoMask                  	352
+#define sBigThreshold                 	384
+#define sC2                           	416
+#define sC3                           	448
+#define sHalf                         	480
+#define sLargestFinite                	512
+#define sLittleThreshold              	544
+#define sSign                         	576
+#define sThirtyOne                    	608
+#define sTopMask8                     	640
+#define XScale                        	672
+#define sLn2                          	704
+
+#include <sysdep.h>
+
+        .text
+	.section .text.avx2,"ax",@progbits
+ENTRY(_ZGVdN8v_asinhf_avx2)
+        pushq     %rbp
+        cfi_def_cfa_offset(16)
+        movq      %rsp, %rbp
+        cfi_def_cfa(6, 16)
+        cfi_offset(6, -16)
+        andq      $-32, %rsp
+        subq      $96, %rsp
+        vmovaps   %ymm0, %ymm9
+
+/* Load the constant 1 and a sign mask */
+        vmovups   sOne+__svml_sasinh_data_internal(%rip), %ymm8
+
+/* No need to split X when FMA is available in hardware. */
+        vmulps    %ymm9, %ymm9, %ymm5
+        vmovups   sTopMask8+__svml_sasinh_data_internal(%rip), %ymm1
+
+/*
+ * Finally, express Y + W = X^2 + 1 accurately where Y has <= 8 bits.
+ * If |X| <= 1 then |XHi| <= 1 and so |X2Hi| <= 1, so we can treat 1
+ * as the dominant component in the compensated summation. Otherwise,
+ * if |X| >= 1, then since X2Hi only has 22 significant bits, the basic
+ * addition will be exact anyway until we get to |X| >= 2^24. But by
+ * that time the log function is well-conditioned enough that the
+ * rounding error doesn't matter. Hence we can treat 1 as dominant even
+ * if it literally isn't.
+ */
+        vaddps    %ymm5, %ymm8, %ymm13
+        vandps    %ymm1, %ymm13, %ymm2
+        vmovaps   %ymm9, %ymm4
+        vsubps    %ymm13, %ymm8, %ymm11
+        vsubps    %ymm2, %ymm13, %ymm15
+
+/*
+ * Compute R = 1/sqrt(Y + W) * (1 + d)
+ * Force R to <= 8 significant bits.
+ * This means that R * Y and R^2 * Y are exactly representable.
+ */
+        vrsqrtps  %ymm2, %ymm0
+        vfmsub213ps %ymm5, %ymm9, %ymm4
+        vaddps    %ymm11, %ymm5, %ymm12
+
+/*
+ * Get the absolute value of the input, since we will exploit antisymmetry
+ * and mostly assume X >= 0 in the core computation
+ */
+        vandps    SgnMask+__svml_sasinh_data_internal(%rip), %ymm9, %ymm6
+
+/*
+ * Check whether the input is finite, by checking |X| <= MaxFloat
+ * Otherwise set the rangemask so that the callout will get used.
+ * Note that this will also use the callout for NaNs since not(NaN <= MaxFloat)
+ */
+        vcmpnle_uqps sLargestFinite+__svml_sasinh_data_internal(%rip), %ymm6, %ymm10
+        vaddps    %ymm12, %ymm4, %ymm14
+
+/*
+ * Unfortunately, we can still be in trouble if |X| <= 2^-5, since
+ * the absolute error 2^-(7+24)-ish in sqrt(1 + X^2) gets scaled up
+ * by 1/X and comes close to our threshold. Hence if |X| <= 2^-4,
+ * perform an alternative computation
+ * sqrt(1 + X^2) - 1 = X^2/2 - X^4/8 + X^6/16
+ * X2 = X^2
+ */
+        vaddps    %ymm4, %ymm5, %ymm4
+
+/*
+ * The following computation can go wrong for very large X, basically
+ * because X^2 overflows. But for large X we have
+ * asinh(X) / log(2 X) - 1 =~= 1/(4 * X^2), so for X >= 2^30
+ * we can just later stick X back into the log and tweak up the exponent.
+ * Actually we scale X by 2^-30 and tweak the exponent up by 31,
+ * to stay in the safe range for the later log computation.
+ * Compute a flag now telling us when do do this.
+ */
+        vcmplt_oqps sBigThreshold+__svml_sasinh_data_internal(%rip), %ymm6, %ymm7
+        vaddps    %ymm15, %ymm14, %ymm3
+
+/*
+ * Now       1 / (1 + d)
+ * = 1 / (1 + (sqrt(1 - e) - 1))
+ * = 1 / sqrt(1 - e)
+ * = 1 + 1/2 * e + 3/8 * e^2 + 5/16 * e^3 + 35/128 * e^4 + ...
+ * So compute the first three nonconstant terms of that, so that
+ * we have a relative correction (1 + Corr) to apply to S etc.
+ * C1 = 1/2
+ * C2 = 3/8
+ * C3 = 5/16
+ */
+        vmovups   sC3+__svml_sasinh_data_internal(%rip), %ymm12
+        vmovmskps %ymm10, %edx
+        vandps    %ymm1, %ymm0, %ymm10
+
+/*
+ * Compute S = (Y/sqrt(Y + W)) * (1 + d)
+ * and T = (W/sqrt(Y + W)) * (1 + d)
+ * so that S + T = sqrt(Y + W) * (1 + d)
+ * S is exact, and the rounding error in T is OK.
+ */
+        vmulps    %ymm10, %ymm2, %ymm15
+        vmulps    %ymm3, %ymm10, %ymm14
+        vmovups   sHalf+__svml_sasinh_data_internal(%rip), %ymm3
+        vsubps    %ymm8, %ymm15, %ymm0
+
+/*
+ * Obtain sqrt(1 + X^2) - 1 in two pieces
+ * sqrt(1 + X^2) - 1
+ * = sqrt(Y + W) - 1
+ * = (S + T) * (1 + Corr) - 1
+ * = [S - 1] + [T + (S + T) * Corr]
+ * We need a compensated summation for the last part. We treat S - 1
+ * as the larger part; it certainly is until about X < 2^-4, and in that
+ * case, the error is affordable since X dominates over sqrt(1 + X^2) - 1
+ * Final sum is dTmp5 (hi) + dTmp7 (lo)
+ */
+        vaddps    %ymm14, %ymm15, %ymm13
+
+/*
+ * Compute e = -(2 * d + d^2)
+ * The first FMR is exact, and the rounding error in the other is acceptable
+ * since d and e are ~ 2^-8
+ */
+        vmovaps   %ymm8, %ymm11
+        vfnmadd231ps %ymm15, %ymm10, %ymm11
+        vfnmadd231ps %ymm14, %ymm10, %ymm11
+        vfmadd213ps sC2+__svml_sasinh_data_internal(%rip), %ymm11, %ymm12
+        vfmadd213ps %ymm3, %ymm11, %ymm12
+        vmulps    %ymm12, %ymm11, %ymm1
+
+/* Now multiplex the two possible computations */
+        vcmple_oqps sLittleThreshold+__svml_sasinh_data_internal(%rip), %ymm6, %ymm11
+        vfmadd213ps %ymm14, %ymm13, %ymm1
+        vaddps    %ymm0, %ymm1, %ymm2
+        vsubps    %ymm2, %ymm0, %ymm10
+
+/* sX2over2 = X^2/2 */
+        vmulps    %ymm4, %ymm3, %ymm0
+        vaddps    %ymm10, %ymm1, %ymm1
+
+/* sX4over4 = X^4/4 */
+        vmulps    %ymm0, %ymm0, %ymm5
+
+/* sX46 = -X^4/4 + X^6/8 */
+        vfmsub231ps %ymm0, %ymm5, %ymm5
+
+/* sX46over2 = -X^4/8 + x^6/16 */
+        vmulps    %ymm5, %ymm3, %ymm3
+        vaddps    %ymm3, %ymm0, %ymm5
+        vblendvps %ymm11, %ymm5, %ymm2, %ymm2
+        vsubps    %ymm5, %ymm0, %ymm4
+
+/*
+ * Now do another compensated sum to add |X| + [sqrt(1 + X^2) - 1].
+ * It's always safe to assume |X| is larger.
+ * This is the final 2-part argument to the log1p function
+ */
+        vaddps    %ymm2, %ymm6, %ymm14
+
+/*
+ * Now resume the main code.
+ * reduction: compute r,n
+ */
+        vmovups   iBrkValue+__svml_sasinh_data_internal(%rip), %ymm5
+        vaddps    %ymm4, %ymm3, %ymm10
+
+/*
+ * Now we feed into the log1p code, using H in place of _VARG1 and
+ * also adding L into Xl.
+ * compute 1+x as high, low parts
+ */
+        vmaxps    %ymm14, %ymm8, %ymm15
+        vminps    %ymm14, %ymm8, %ymm0
+        vblendvps %ymm11, %ymm10, %ymm1, %ymm12
+        vsubps    %ymm14, %ymm6, %ymm1
+        vaddps    %ymm0, %ymm15, %ymm3
+
+/* Now multiplex to the case X = 2^-30 * input, Xl = sL = 0 in the "big" case. */
+        vmulps    XScale+__svml_sasinh_data_internal(%rip), %ymm6, %ymm6
+        vaddps    %ymm1, %ymm2, %ymm13
+        vsubps    %ymm3, %ymm15, %ymm15
+        vaddps    %ymm13, %ymm12, %ymm1
+        vaddps    %ymm15, %ymm0, %ymm2
+        vblendvps %ymm7, %ymm3, %ymm6, %ymm0
+        vaddps    %ymm2, %ymm1, %ymm4
+        vpsubd    %ymm5, %ymm0, %ymm1
+        vpsrad    $23, %ymm1, %ymm6
+        vpand     iOffExpoMask+__svml_sasinh_data_internal(%rip), %ymm1, %ymm2
+        vmovups   sPoly+224+__svml_sasinh_data_internal(%rip), %ymm1
+        vpslld    $23, %ymm6, %ymm10
+        vpaddd    %ymm5, %ymm2, %ymm13
+        vcvtdq2ps %ymm6, %ymm0
+        vpsubd    %ymm10, %ymm8, %ymm12
+
+/* polynomial evaluation */
+        vsubps    %ymm8, %ymm13, %ymm8
+
+/* Add 31 to the exponent in the "large" case to get log(2 * input) */
+        vaddps    sThirtyOne+__svml_sasinh_data_internal(%rip), %ymm0, %ymm3
+        vandps    %ymm7, %ymm4, %ymm11
+        vmulps    %ymm12, %ymm11, %ymm14
+        vblendvps %ymm7, %ymm0, %ymm3, %ymm0
+        vaddps    %ymm8, %ymm14, %ymm2
+        vfmadd213ps sPoly+192+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
+        vfmadd213ps sPoly+160+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
+        vfmadd213ps sPoly+128+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
+        vfmadd213ps sPoly+96+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
+        vfmadd213ps sPoly+64+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
+        vfmadd213ps sPoly+32+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
+        vfmadd213ps sPoly+__svml_sasinh_data_internal(%rip), %ymm2, %ymm1
+        vmulps    %ymm1, %ymm2, %ymm4
+        vfmadd213ps %ymm2, %ymm2, %ymm4
+
+/* final reconstruction */
+        vfmadd132ps sLn2+__svml_sasinh_data_internal(%rip), %ymm4, %ymm0
+
+/* Finally, reincorporate the original sign. */
+        vandps    sSign+__svml_sasinh_data_internal(%rip), %ymm9, %ymm7
+        vxorps    %ymm0, %ymm7, %ymm0
+        testl     %edx, %edx
+
+/* Go to special inputs processing branch */
+        jne       L(SPECIAL_VALUES_BRANCH)
+                                # LOE rbx r12 r13 r14 r15 edx ymm0 ymm9
+
+/* Restore registers
+ * and exit the function
+ */
+
+L(EXIT):
+        movq      %rbp, %rsp
+        popq      %rbp
+        cfi_def_cfa(7, 8)
+        cfi_restore(6)
+        ret
+        cfi_def_cfa(6, 16)
+        cfi_offset(6, -16)
+
+/* Branch to process
+ * special inputs
+ */
+
+L(SPECIAL_VALUES_BRANCH):
+        vmovups   %ymm9, 32(%rsp)
+        vmovups   %ymm0, 64(%rsp)
+                                # LOE rbx r12 r13 r14 r15 edx ymm0
+
+        xorl      %eax, %eax
+                                # LOE rbx r12 r13 r14 r15 eax edx
+
+        vzeroupper
+        movq      %r12, 16(%rsp)
+        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -80; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xb0, 0xff, 0xff, 0xff, 0x22
+        movl      %eax, %r12d
+        movq      %r13, 8(%rsp)
+        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -88; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa8, 0xff, 0xff, 0xff, 0x22
+        movl      %edx, %r13d
+        movq      %r14, (%rsp)
+        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -96; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa0, 0xff, 0xff, 0xff, 0x22
+                                # LOE rbx r15 r12d r13d
+
+/* Range mask
+ * bits check
+ */
+
+L(RANGEMASK_CHECK):
+        btl       %r12d, %r13d
+
+/* Call scalar math function */
+        jc        L(SCALAR_MATH_CALL)
+                                # LOE rbx r15 r12d r13d
+
+/* Special inputs
+ * processing loop
+ */
+
+L(SPECIAL_VALUES_LOOP):
+        incl      %r12d
+        cmpl      $8, %r12d
+
+/* Check bits in range mask */
+        jl        L(RANGEMASK_CHECK)
+                                # LOE rbx r15 r12d r13d
+
+        movq      16(%rsp), %r12
+        cfi_restore(12)
+        movq      8(%rsp), %r13
+        cfi_restore(13)
+        movq      (%rsp), %r14
+        cfi_restore(14)
+        vmovups   64(%rsp), %ymm0
+
+/* Go to exit */
+        jmp       L(EXIT)
+        /*  DW_CFA_expression: r12 (r12) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -80; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0c, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xb0, 0xff, 0xff, 0xff, 0x22
+        /*  DW_CFA_expression: r13 (r13) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -88; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0d, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa8, 0xff, 0xff, 0xff, 0x22
+        /*  DW_CFA_expression: r14 (r14) (DW_OP_lit8; DW_OP_minus; DW_OP_const4s: -32; DW_OP_and; DW_OP_const4s: -96; DW_OP_plus)  */
+        .cfi_escape 0x10, 0x0e, 0x0e, 0x38, 0x1c, 0x0d, 0xe0, 0xff, 0xff, 0xff, 0x1a, 0x0d, 0xa0, 0xff, 0xff, 0xff, 0x22
+                                # LOE rbx r12 r13 r14 r15 ymm0
+
+/* Scalar math fucntion call
+ * to process special input
+ */
+
+L(SCALAR_MATH_CALL):
+        movl      %r12d, %r14d
+        movss     32(%rsp,%r14,4), %xmm0
+        call      asinhf@PLT
+                                # LOE rbx r14 r15 r12d r13d xmm0
+
+        movss     %xmm0, 64(%rsp,%r14,4)
+
+/* Process special inputs in loop */
+        jmp       L(SPECIAL_VALUES_LOOP)
+                                # LOE rbx r15 r12d r13d
+END(_ZGVdN8v_asinhf_avx2)
+
+        .section .rodata, "a"
+        .align 32
+
+#ifdef __svml_sasinh_data_internal_typedef
+typedef unsigned int VUINT32;
+typedef struct {
+        __declspec(align(32)) VUINT32 SgnMask[8][1];
+        __declspec(align(32)) VUINT32 sOne[8][1];
+        __declspec(align(32)) VUINT32 sPoly[8][8][1];
+        __declspec(align(32)) VUINT32 iBrkValue[8][1];
+        __declspec(align(32)) VUINT32 iOffExpoMask[8][1];
+        __declspec(align(32)) VUINT32 sBigThreshold[8][1];
+        __declspec(align(32)) VUINT32 sC2[8][1];
+        __declspec(align(32)) VUINT32 sC3[8][1];
+        __declspec(align(32)) VUINT32 sHalf[8][1];
+        __declspec(align(32)) VUINT32 sLargestFinite[8][1];
+        __declspec(align(32)) VUINT32 sLittleThreshold[8][1];
+        __declspec(align(32)) VUINT32 sSign[8][1];
+        __declspec(align(32)) VUINT32 sThirtyOne[8][1];
+        __declspec(align(32)) VUINT32 sTopMask8[8][1];
+        __declspec(align(32)) VUINT32 XScale[8][1];
+        __declspec(align(32)) VUINT32 sLn2[8][1];
+} __svml_sasinh_data_internal;
+#endif
+__svml_sasinh_data_internal:
+        /*== SgnMask ==*/
+        .long 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff, 0x7fffffff
+        /*== sOne = SP 1.0 ==*/
+        .align 32
+        .long 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000, 0x3f800000
+        /*== sPoly[] = SP polynomial ==*/
+        .align 32
+        .long 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000, 0xbf000000 /* -5.0000000000000000000000000e-01 P0 */
+        .long 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94, 0x3eaaaa94 /*  3.3333265781402587890625000e-01 P1 */
+        .long 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e, 0xbe80058e /* -2.5004237890243530273437500e-01 P2 */
+        .long 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190, 0x3e4ce190 /*  2.0007920265197753906250000e-01 P3 */
+        .long 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37, 0xbe28ad37 /* -1.6472326219081878662109375e-01 P4 */
+        .long 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12, 0x3e0fcb12 /*  1.4042308926582336425781250e-01 P5 */
+        .long 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3, 0xbe1ad9e3 /* -1.5122179687023162841796875e-01 P6 */
+        .long 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed, 0x3e0d84ed /*  1.3820238411426544189453125e-01 P7 */
+        /*== iBrkValue = SP 2/3 ==*/
+        .align 32
+        .long 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab, 0x3f2aaaab
+        /*== iOffExpoMask = SP significand mask ==*/
+        .align 32
+        .long 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff, 0x007fffff
+        /*== sBigThreshold ==*/
+        .align 32
+        .long 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000, 0x4E800000
+        /*== sC2 ==*/
+        .align 32
+        .long 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000, 0x3EC00000
+        /*== sC3 ==*/
+        .align 32
+        .long 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000, 0x3EA00000
+        /*== sHalf ==*/
+        .align 32
+        .long 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000, 0x3F000000
+        /*== sLargestFinite ==*/
+        .align 32
+        .long 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF, 0x7F7FFFFF
+        /*== sLittleThreshold ==*/
+        .align 32
+        .long 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000, 0x3D800000
+        /*== sSign ==*/
+        .align 32
+        .long 0x80000000, 0x80000000, 0x80000000, 0x80000000, 0x80000000, 0x80000000, 0x80000000, 0x80000000
+        /*== sThirtyOne ==*/
+        .align 32
+        .long 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000, 0x41F80000
+        /*== sTopMask8 ==*/
+        .align 32
+        .long 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000, 0xFFFF0000
+        /*== XScale ==*/
+        .align 32
+        .long 0x30800000, 0x30800000, 0x30800000, 0x30800000, 0x30800000, 0x30800000, 0x30800000, 0x30800000
+        /*== sLn2 = SP ln(2) ==*/
+        .align 32
+        .long 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218, 0x3f317218
+        .align 32
+        .type	__svml_sasinh_data_internal,@object
+        .size	__svml_sasinh_data_internal,.-__svml_sasinh_data_internal
diff --git a/sysdeps/x86_64/fpu/svml_d_asinh2_core.S b/sysdeps/x86_64/fpu/svml_d_asinh2_core.S
new file mode 100644
index 0000000000..60e372238a
--- /dev/null
+++ b/sysdeps/x86_64/fpu/svml_d_asinh2_core.S
@@ -0,0 +1,29 @@ 
+/* Function asinh vectorized with SSE2.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#include <sysdep.h>
+#include "svml_d_wrapper_impl.h"
+
+	.text
+ENTRY (_ZGVbN2v_asinh)
+WRAPPER_IMPL_SSE2 asinh
+END (_ZGVbN2v_asinh)
+
+#ifndef USE_MULTIARCH
+ libmvec_hidden_def (_ZGVbN2v_asinh)
+#endif
diff --git a/sysdeps/x86_64/fpu/svml_d_asinh4_core.S b/sysdeps/x86_64/fpu/svml_d_asinh4_core.S
new file mode 100644
index 0000000000..c7350011e1
--- /dev/null
+++ b/sysdeps/x86_64/fpu/svml_d_asinh4_core.S
@@ -0,0 +1,29 @@ 
+/* Function asinh vectorized with AVX2, wrapper version.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#include <sysdep.h>
+#include "svml_d_wrapper_impl.h"
+
+	.text
+ENTRY (_ZGVdN4v_asinh)
+WRAPPER_IMPL_AVX _ZGVbN2v_asinh
+END (_ZGVdN4v_asinh)
+
+#ifndef USE_MULTIARCH
+ libmvec_hidden_def (_ZGVdN4v_asinh)
+#endif
diff --git a/sysdeps/x86_64/fpu/svml_d_asinh4_core_avx.S b/sysdeps/x86_64/fpu/svml_d_asinh4_core_avx.S
new file mode 100644
index 0000000000..83aaa8c3f1
--- /dev/null
+++ b/sysdeps/x86_64/fpu/svml_d_asinh4_core_avx.S
@@ -0,0 +1,25 @@ 
+/* Function asinh vectorized in AVX ISA as wrapper to SSE4 ISA version.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#include <sysdep.h>
+#include "svml_d_wrapper_impl.h"
+
+	.text
+ENTRY (_ZGVcN4v_asinh)
+WRAPPER_IMPL_AVX _ZGVbN2v_asinh
+END (_ZGVcN4v_asinh)
diff --git a/sysdeps/x86_64/fpu/svml_d_asinh8_core.S b/sysdeps/x86_64/fpu/svml_d_asinh8_core.S
new file mode 100644
index 0000000000..9597975ff6
--- /dev/null
+++ b/sysdeps/x86_64/fpu/svml_d_asinh8_core.S
@@ -0,0 +1,25 @@ 
+/* Function asinh vectorized with AVX-512, wrapper to AVX2.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#include <sysdep.h>
+#include "svml_d_wrapper_impl.h"
+
+	.text
+ENTRY (_ZGVeN8v_asinh)
+WRAPPER_IMPL_AVX512 _ZGVdN4v_asinh
+END (_ZGVeN8v_asinh)
diff --git a/sysdeps/x86_64/fpu/svml_s_asinhf16_core.S b/sysdeps/x86_64/fpu/svml_s_asinhf16_core.S
new file mode 100644
index 0000000000..5b3d405f2e
--- /dev/null
+++ b/sysdeps/x86_64/fpu/svml_s_asinhf16_core.S
@@ -0,0 +1,25 @@ 
+/* Function asinhf vectorized with AVX-512. Wrapper to AVX2 version.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#include <sysdep.h>
+#include "svml_s_wrapper_impl.h"
+
+	.text
+ENTRY (_ZGVeN16v_asinhf)
+WRAPPER_IMPL_AVX512 _ZGVdN8v_asinhf
+END (_ZGVeN16v_asinhf)
diff --git a/sysdeps/x86_64/fpu/svml_s_asinhf4_core.S b/sysdeps/x86_64/fpu/svml_s_asinhf4_core.S
new file mode 100644
index 0000000000..af44fa5108
--- /dev/null
+++ b/sysdeps/x86_64/fpu/svml_s_asinhf4_core.S
@@ -0,0 +1,29 @@ 
+/* Function asinhf vectorized with SSE2, wrapper version.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#include <sysdep.h>
+#include "svml_s_wrapper_impl.h"
+
+	.text
+ENTRY (_ZGVbN4v_asinhf)
+WRAPPER_IMPL_SSE2 asinhf
+END (_ZGVbN4v_asinhf)
+
+#ifndef USE_MULTIARCH
+ libmvec_hidden_def (_ZGVbN4v_asinhf)
+#endif
diff --git a/sysdeps/x86_64/fpu/svml_s_asinhf8_core.S b/sysdeps/x86_64/fpu/svml_s_asinhf8_core.S
new file mode 100644
index 0000000000..3bd06d8032
--- /dev/null
+++ b/sysdeps/x86_64/fpu/svml_s_asinhf8_core.S
@@ -0,0 +1,29 @@ 
+/* Function asinhf vectorized with AVX2, wrapper version.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#include <sysdep.h>
+#include "svml_s_wrapper_impl.h"
+
+	.text
+ENTRY (_ZGVdN8v_asinhf)
+WRAPPER_IMPL_AVX _ZGVbN4v_asinhf
+END (_ZGVdN8v_asinhf)
+
+#ifndef USE_MULTIARCH
+ libmvec_hidden_def (_ZGVdN8v_asinhf)
+#endif
diff --git a/sysdeps/x86_64/fpu/svml_s_asinhf8_core_avx.S b/sysdeps/x86_64/fpu/svml_s_asinhf8_core_avx.S
new file mode 100644
index 0000000000..f79616c0bd
--- /dev/null
+++ b/sysdeps/x86_64/fpu/svml_s_asinhf8_core_avx.S
@@ -0,0 +1,25 @@ 
+/* Function asinhf vectorized in AVX ISA as wrapper to SSE4 ISA version.
+   Copyright (C) 2021 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+#include <sysdep.h>
+#include "svml_s_wrapper_impl.h"
+
+        .text
+ENTRY (_ZGVcN8v_asinhf)
+WRAPPER_IMPL_AVX _ZGVbN4v_asinhf
+END (_ZGVcN8v_asinhf)
diff --git a/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx.c b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx.c
new file mode 100644
index 0000000000..da03528700
--- /dev/null
+++ b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx.c
@@ -0,0 +1 @@ 
+#include "test-double-libmvec-asinh.c"
diff --git a/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx2.c b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx2.c
new file mode 100644
index 0000000000..da03528700
--- /dev/null
+++ b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx2.c
@@ -0,0 +1 @@ 
+#include "test-double-libmvec-asinh.c"
diff --git a/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx512f.c b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx512f.c
new file mode 100644
index 0000000000..da03528700
--- /dev/null
+++ b/sysdeps/x86_64/fpu/test-double-libmvec-asinh-avx512f.c
@@ -0,0 +1 @@ 
+#include "test-double-libmvec-asinh.c"
diff --git a/sysdeps/x86_64/fpu/test-double-libmvec-asinh.c b/sysdeps/x86_64/fpu/test-double-libmvec-asinh.c
new file mode 100644
index 0000000000..71e6b9f578
--- /dev/null
+++ b/sysdeps/x86_64/fpu/test-double-libmvec-asinh.c
@@ -0,0 +1,3 @@ 
+#define LIBMVEC_TYPE double
+#define LIBMVEC_FUNC asinh
+#include "test-vector-abi-arg1.h"
diff --git a/sysdeps/x86_64/fpu/test-double-vlen2-wrappers.c b/sysdeps/x86_64/fpu/test-double-vlen2-wrappers.c
index f53bb6813e..76114772ba 100644
--- a/sysdeps/x86_64/fpu/test-double-vlen2-wrappers.c
+++ b/sysdeps/x86_64/fpu/test-double-vlen2-wrappers.c
@@ -45,6 +45,7 @@  VECTOR_WRAPPER (WRAPPER_NAME (atanh), _ZGVbN2v_atanh)
 VECTOR_WRAPPER (WRAPPER_NAME (acosh), _ZGVbN2v_acosh)
 VECTOR_WRAPPER (WRAPPER_NAME (erf), _ZGVbN2v_erf)
 VECTOR_WRAPPER (WRAPPER_NAME (tanh), _ZGVbN2v_tanh)
+VECTOR_WRAPPER (WRAPPER_NAME (asinh), _ZGVbN2v_asinh)
 
 #define VEC_INT_TYPE __m128i
 
diff --git a/sysdeps/x86_64/fpu/test-double-vlen4-avx2-wrappers.c b/sysdeps/x86_64/fpu/test-double-vlen4-avx2-wrappers.c
index 0452c3db38..1e0ee34975 100644
--- a/sysdeps/x86_64/fpu/test-double-vlen4-avx2-wrappers.c
+++ b/sysdeps/x86_64/fpu/test-double-vlen4-avx2-wrappers.c
@@ -48,6 +48,7 @@  VECTOR_WRAPPER (WRAPPER_NAME (atanh), _ZGVdN4v_atanh)
 VECTOR_WRAPPER (WRAPPER_NAME (acosh), _ZGVdN4v_acosh)
 VECTOR_WRAPPER (WRAPPER_NAME (erf), _ZGVdN4v_erf)
 VECTOR_WRAPPER (WRAPPER_NAME (tanh), _ZGVdN4v_tanh)
+VECTOR_WRAPPER (WRAPPER_NAME (asinh), _ZGVdN4v_asinh)
 
 #ifndef __ILP32__
 # define VEC_INT_TYPE __m256i
diff --git a/sysdeps/x86_64/fpu/test-double-vlen4-wrappers.c b/sysdeps/x86_64/fpu/test-double-vlen4-wrappers.c
index 197d5afc88..17c43a75d1 100644
--- a/sysdeps/x86_64/fpu/test-double-vlen4-wrappers.c
+++ b/sysdeps/x86_64/fpu/test-double-vlen4-wrappers.c
@@ -45,6 +45,7 @@  VECTOR_WRAPPER (WRAPPER_NAME (atanh), _ZGVcN4v_atanh)
 VECTOR_WRAPPER (WRAPPER_NAME (acosh), _ZGVcN4v_acosh)
 VECTOR_WRAPPER (WRAPPER_NAME (erf), _ZGVcN4v_erf)
 VECTOR_WRAPPER (WRAPPER_NAME (tanh), _ZGVcN4v_tanh)
+VECTOR_WRAPPER (WRAPPER_NAME (asinh), _ZGVcN4v_asinh)
 
 #define VEC_INT_TYPE __m128i
 
diff --git a/sysdeps/x86_64/fpu/test-double-vlen8-wrappers.c b/sysdeps/x86_64/fpu/test-double-vlen8-wrappers.c
index e56ece640c..1c6809e6e3 100644
--- a/sysdeps/x86_64/fpu/test-double-vlen8-wrappers.c
+++ b/sysdeps/x86_64/fpu/test-double-vlen8-wrappers.c
@@ -45,6 +45,7 @@  VECTOR_WRAPPER (WRAPPER_NAME (atanh), _ZGVeN8v_atanh)
 VECTOR_WRAPPER (WRAPPER_NAME (acosh), _ZGVeN8v_acosh)
 VECTOR_WRAPPER (WRAPPER_NAME (erf), _ZGVeN8v_erf)
 VECTOR_WRAPPER (WRAPPER_NAME (tanh), _ZGVeN8v_tanh)
+VECTOR_WRAPPER (WRAPPER_NAME (asinh), _ZGVeN8v_asinh)
 
 #ifndef __ILP32__
 # define VEC_INT_TYPE __m512i
diff --git a/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx.c b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx.c
new file mode 100644
index 0000000000..77e1838bb4
--- /dev/null
+++ b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx.c
@@ -0,0 +1 @@ 
+#include "test-float-libmvec-asinhf.c"
diff --git a/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx2.c b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx2.c
new file mode 100644
index 0000000000..77e1838bb4
--- /dev/null
+++ b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx2.c
@@ -0,0 +1 @@ 
+#include "test-float-libmvec-asinhf.c"
diff --git a/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx512f.c b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx512f.c
new file mode 100644
index 0000000000..77e1838bb4
--- /dev/null
+++ b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf-avx512f.c
@@ -0,0 +1 @@ 
+#include "test-float-libmvec-asinhf.c"
diff --git a/sysdeps/x86_64/fpu/test-float-libmvec-asinhf.c b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf.c
new file mode 100644
index 0000000000..3353754102
--- /dev/null
+++ b/sysdeps/x86_64/fpu/test-float-libmvec-asinhf.c
@@ -0,0 +1,3 @@ 
+#define LIBMVEC_TYPE float
+#define LIBMVEC_FUNC asinhf
+#include "test-vector-abi-arg1.h"
diff --git a/sysdeps/x86_64/fpu/test-float-vlen16-wrappers.c b/sysdeps/x86_64/fpu/test-float-vlen16-wrappers.c
index abbebf9993..e8ab1885a7 100644
--- a/sysdeps/x86_64/fpu/test-float-vlen16-wrappers.c
+++ b/sysdeps/x86_64/fpu/test-float-vlen16-wrappers.c
@@ -45,6 +45,7 @@  VECTOR_WRAPPER (WRAPPER_NAME (atanhf), _ZGVeN16v_atanhf)
 VECTOR_WRAPPER (WRAPPER_NAME (acoshf), _ZGVeN16v_acoshf)
 VECTOR_WRAPPER (WRAPPER_NAME (erff), _ZGVeN16v_erff)
 VECTOR_WRAPPER (WRAPPER_NAME (tanhf), _ZGVeN16v_tanhf)
+VECTOR_WRAPPER (WRAPPER_NAME (asinhf), _ZGVeN16v_asinhf)
 
 #define VEC_INT_TYPE __m512i
 
diff --git a/sysdeps/x86_64/fpu/test-float-vlen4-wrappers.c b/sysdeps/x86_64/fpu/test-float-vlen4-wrappers.c
index ae1c8b98c2..a80c5387e4 100644
--- a/sysdeps/x86_64/fpu/test-float-vlen4-wrappers.c
+++ b/sysdeps/x86_64/fpu/test-float-vlen4-wrappers.c
@@ -45,6 +45,7 @@  VECTOR_WRAPPER (WRAPPER_NAME (atanhf), _ZGVbN4v_atanhf)
 VECTOR_WRAPPER (WRAPPER_NAME (acoshf), _ZGVbN4v_acoshf)
 VECTOR_WRAPPER (WRAPPER_NAME (erff), _ZGVbN4v_erff)
 VECTOR_WRAPPER (WRAPPER_NAME (tanhf), _ZGVbN4v_tanhf)
+VECTOR_WRAPPER (WRAPPER_NAME (asinhf), _ZGVbN4v_asinhf)
 
 #define VEC_INT_TYPE __m128i
 
diff --git a/sysdeps/x86_64/fpu/test-float-vlen8-avx2-wrappers.c b/sysdeps/x86_64/fpu/test-float-vlen8-avx2-wrappers.c
index eb477a0371..c3d1d5936b 100644
--- a/sysdeps/x86_64/fpu/test-float-vlen8-avx2-wrappers.c
+++ b/sysdeps/x86_64/fpu/test-float-vlen8-avx2-wrappers.c
@@ -48,6 +48,7 @@  VECTOR_WRAPPER (WRAPPER_NAME (atanhf), _ZGVdN8v_atanhf)
 VECTOR_WRAPPER (WRAPPER_NAME (acoshf), _ZGVdN8v_acoshf)
 VECTOR_WRAPPER (WRAPPER_NAME (erff), _ZGVdN8v_erff)
 VECTOR_WRAPPER (WRAPPER_NAME (tanhf), _ZGVdN8v_tanhf)
+VECTOR_WRAPPER (WRAPPER_NAME (asinhf), _ZGVdN8v_asinhf)
 
 /* Redefinition of wrapper to be compatible with _ZGVdN8vvv_sincosf.  */
 #undef VECTOR_WRAPPER_fFF
diff --git a/sysdeps/x86_64/fpu/test-float-vlen8-wrappers.c b/sysdeps/x86_64/fpu/test-float-vlen8-wrappers.c
index 944f7f0a75..b7da0f523b 100644
--- a/sysdeps/x86_64/fpu/test-float-vlen8-wrappers.c
+++ b/sysdeps/x86_64/fpu/test-float-vlen8-wrappers.c
@@ -45,6 +45,7 @@  VECTOR_WRAPPER (WRAPPER_NAME (atanhf), _ZGVcN8v_atanhf)
 VECTOR_WRAPPER (WRAPPER_NAME (acoshf), _ZGVcN8v_acoshf)
 VECTOR_WRAPPER (WRAPPER_NAME (erff), _ZGVcN8v_erff)
 VECTOR_WRAPPER (WRAPPER_NAME (tanhf), _ZGVcN8v_tanhf)
+VECTOR_WRAPPER (WRAPPER_NAME (asinhf), _ZGVcN8v_asinhf)
 
 #define VEC_INT_TYPE __m128i