[v3,1/2] x86: Move mem{p}{mov|cpy}_{chk_}erms to its own file

Message ID 20220629230716.1264249-1-goldstein.w.n@gmail.com
State Committed
Commit 21925f64730d52eb7d8b2fb62b412f8ab92b0caf
Headers
Series [v3,1/2] x86: Move mem{p}{mov|cpy}_{chk_}erms to its own file |

Checks

Context Check Description
dj/TryBot-apply_patch success Patch applied to master at the time it was sent

Commit Message

Noah Goldstein June 29, 2022, 11:07 p.m. UTC
  The primary memmove_{impl}_unaligned_erms implementations don't
interact with this function. Putting them in same file both
wastes space and unnecessarily bloats a hot code section.
---
 sysdeps/x86_64/multiarch/Makefile             |  1 +
 sysdeps/x86_64/multiarch/memmove-erms.S       | 72 +++++++++++++++++++
 .../multiarch/memmove-vec-unaligned-erms.S    | 50 -------------
 3 files changed, 73 insertions(+), 50 deletions(-)
 create mode 100644 sysdeps/x86_64/multiarch/memmove-erms.S
  

Comments

H.J. Lu June 29, 2022, 11:30 p.m. UTC | #1
On Wed, Jun 29, 2022 at 4:07 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
>
> The primary memmove_{impl}_unaligned_erms implementations don't
> interact with this function. Putting them in same file both
> wastes space and unnecessarily bloats a hot code section.
> ---
>  sysdeps/x86_64/multiarch/Makefile             |  1 +
>  sysdeps/x86_64/multiarch/memmove-erms.S       | 72 +++++++++++++++++++
>  .../multiarch/memmove-vec-unaligned-erms.S    | 50 -------------
>  3 files changed, 73 insertions(+), 50 deletions(-)
>  create mode 100644 sysdeps/x86_64/multiarch/memmove-erms.S
>
> diff --git a/sysdeps/x86_64/multiarch/Makefile b/sysdeps/x86_64/multiarch/Makefile
> index 666ee4d5d6..62a4d96fb8 100644
> --- a/sysdeps/x86_64/multiarch/Makefile
> +++ b/sysdeps/x86_64/multiarch/Makefile
> @@ -18,6 +18,7 @@ sysdep_routines += \
>    memmove-avx-unaligned-erms-rtm \
>    memmove-avx512-no-vzeroupper \
>    memmove-avx512-unaligned-erms \
> +  memmove-erms \
>    memmove-evex-unaligned-erms \
>    memmove-sse2-unaligned-erms \
>    memmove-ssse3 \
> diff --git a/sysdeps/x86_64/multiarch/memmove-erms.S b/sysdeps/x86_64/multiarch/memmove-erms.S
> new file mode 100644
> index 0000000000..2d3a6ccb76
> --- /dev/null
> +++ b/sysdeps/x86_64/multiarch/memmove-erms.S
> @@ -0,0 +1,72 @@
> +/* memcpy/mempcpy/memmove implement with rep movsb
> +   Copyright (C) 2022 Free Software Foundation, Inc.
> +   This file is part of the GNU C Library.
> +
> +   The GNU C Library is free software; you can redistribute it and/or
> +   modify it under the terms of the GNU Lesser General Public
> +   License as published by the Free Software Foundation; either
> +   version 2.1 of the License, or (at your option) any later version.
> +
> +   The GNU C Library is distributed in the hope that it will be useful,
> +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> +   Lesser General Public License for more details.
> +
> +   You should have received a copy of the GNU Lesser General Public
> +   License along with the GNU C Library; if not, see
> +   <https://www.gnu.org/licenses/>.  */
> +
> +
> +#include <sysdep.h>
> +
> +#if defined USE_MULTIARCH && IS_IN (libc)
> +       .text
> +ENTRY (__mempcpy_chk_erms)
> +       cmp     %RDX_LP, %RCX_LP
> +       jb      HIDDEN_JUMPTARGET (__chk_fail)
> +END (__mempcpy_chk_erms)
> +
> +/* Only used to measure performance of REP MOVSB.  */
> +ENTRY (__mempcpy_erms)
> +       mov     %RDI_LP, %RAX_LP
> +       /* Skip zero length.  */
> +       test    %RDX_LP, %RDX_LP
> +       jz      2f
> +       add     %RDX_LP, %RAX_LP
> +       jmp     L(start_movsb)
> +END (__mempcpy_erms)
> +
> +ENTRY (__memmove_chk_erms)
> +       cmp     %RDX_LP, %RCX_LP
> +       jb      HIDDEN_JUMPTARGET (__chk_fail)
> +END (__memmove_chk_erms)
> +
> +ENTRY (__memmove_erms)
> +       movq    %rdi, %rax
> +       /* Skip zero length.  */
> +       test    %RDX_LP, %RDX_LP
> +       jz      2f
> +L(start_movsb):
> +       mov     %RDX_LP, %RCX_LP
> +       cmp     %RSI_LP, %RDI_LP
> +       jb      1f
> +       /* Source == destination is less common.  */
> +       je      2f
> +       lea     (%rsi,%rcx), %RDX_LP
> +       cmp     %RDX_LP, %RDI_LP
> +       jb      L(movsb_backward)
> +1:
> +       rep movsb
> +2:
> +       ret
> +L(movsb_backward):
> +       leaq    -1(%rdi,%rcx), %rdi
> +       leaq    -1(%rsi,%rcx), %rsi
> +       std
> +       rep movsb
> +       cld
> +       ret
> +END (__memmove_erms)
> +strong_alias (__memmove_erms, __memcpy_erms)
> +strong_alias (__memmove_chk_erms, __memcpy_chk_erms)
> +#endif
> diff --git a/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S b/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
> index d1518b8bab..04747133b7 100644
> --- a/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
> +++ b/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
> @@ -239,56 +239,6 @@ L(start):
>  #endif
>  #if defined USE_MULTIARCH && IS_IN (libc)
>  END (MEMMOVE_SYMBOL (__memmove, unaligned))
> -# if VEC_SIZE == 16
> -ENTRY (__mempcpy_chk_erms)
> -       cmp     %RDX_LP, %RCX_LP
> -       jb      HIDDEN_JUMPTARGET (__chk_fail)
> -END (__mempcpy_chk_erms)
> -
> -/* Only used to measure performance of REP MOVSB.  */
> -ENTRY (__mempcpy_erms)
> -       mov     %RDI_LP, %RAX_LP
> -       /* Skip zero length.  */
> -       test    %RDX_LP, %RDX_LP
> -       jz      2f
> -       add     %RDX_LP, %RAX_LP
> -       jmp     L(start_movsb)
> -END (__mempcpy_erms)
> -
> -ENTRY (__memmove_chk_erms)
> -       cmp     %RDX_LP, %RCX_LP
> -       jb      HIDDEN_JUMPTARGET (__chk_fail)
> -END (__memmove_chk_erms)
> -
> -ENTRY (__memmove_erms)
> -       movq    %rdi, %rax
> -       /* Skip zero length.  */
> -       test    %RDX_LP, %RDX_LP
> -       jz      2f
> -L(start_movsb):
> -       mov     %RDX_LP, %RCX_LP
> -       cmp     %RSI_LP, %RDI_LP
> -       jb      1f
> -       /* Source == destination is less common.  */
> -       je      2f
> -       lea     (%rsi,%rcx), %RDX_LP
> -       cmp     %RDX_LP, %RDI_LP
> -       jb      L(movsb_backward)
> -1:
> -       rep movsb
> -2:
> -       ret
> -L(movsb_backward):
> -       leaq    -1(%rdi,%rcx), %rdi
> -       leaq    -1(%rsi,%rcx), %rsi
> -       std
> -       rep movsb
> -       cld
> -       ret
> -END (__memmove_erms)
> -strong_alias (__memmove_erms, __memcpy_erms)
> -strong_alias (__memmove_chk_erms, __memcpy_chk_erms)
> -# endif
>
>  # ifdef SHARED
>  ENTRY (MEMMOVE_CHK_SYMBOL (__mempcpy_chk, unaligned_erms))
> --
> 2.34.1
>

LGTM.

Thanks.
  
Sunil Pandey July 14, 2022, 3:05 a.m. UTC | #2
On Wed, Jun 29, 2022 at 4:31 PM H.J. Lu via Libc-alpha
<libc-alpha@sourceware.org> wrote:
>
> On Wed, Jun 29, 2022 at 4:07 PM Noah Goldstein <goldstein.w.n@gmail.com> wrote:
> >
> > The primary memmove_{impl}_unaligned_erms implementations don't
> > interact with this function. Putting them in same file both
> > wastes space and unnecessarily bloats a hot code section.
> > ---
> >  sysdeps/x86_64/multiarch/Makefile             |  1 +
> >  sysdeps/x86_64/multiarch/memmove-erms.S       | 72 +++++++++++++++++++
> >  .../multiarch/memmove-vec-unaligned-erms.S    | 50 -------------
> >  3 files changed, 73 insertions(+), 50 deletions(-)
> >  create mode 100644 sysdeps/x86_64/multiarch/memmove-erms.S
> >
> > diff --git a/sysdeps/x86_64/multiarch/Makefile b/sysdeps/x86_64/multiarch/Makefile
> > index 666ee4d5d6..62a4d96fb8 100644
> > --- a/sysdeps/x86_64/multiarch/Makefile
> > +++ b/sysdeps/x86_64/multiarch/Makefile
> > @@ -18,6 +18,7 @@ sysdep_routines += \
> >    memmove-avx-unaligned-erms-rtm \
> >    memmove-avx512-no-vzeroupper \
> >    memmove-avx512-unaligned-erms \
> > +  memmove-erms \
> >    memmove-evex-unaligned-erms \
> >    memmove-sse2-unaligned-erms \
> >    memmove-ssse3 \
> > diff --git a/sysdeps/x86_64/multiarch/memmove-erms.S b/sysdeps/x86_64/multiarch/memmove-erms.S
> > new file mode 100644
> > index 0000000000..2d3a6ccb76
> > --- /dev/null
> > +++ b/sysdeps/x86_64/multiarch/memmove-erms.S
> > @@ -0,0 +1,72 @@
> > +/* memcpy/mempcpy/memmove implement with rep movsb
> > +   Copyright (C) 2022 Free Software Foundation, Inc.
> > +   This file is part of the GNU C Library.
> > +
> > +   The GNU C Library is free software; you can redistribute it and/or
> > +   modify it under the terms of the GNU Lesser General Public
> > +   License as published by the Free Software Foundation; either
> > +   version 2.1 of the License, or (at your option) any later version.
> > +
> > +   The GNU C Library is distributed in the hope that it will be useful,
> > +   but WITHOUT ANY WARRANTY; without even the implied warranty of
> > +   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> > +   Lesser General Public License for more details.
> > +
> > +   You should have received a copy of the GNU Lesser General Public
> > +   License along with the GNU C Library; if not, see
> > +   <https://www.gnu.org/licenses/>.  */
> > +
> > +
> > +#include <sysdep.h>
> > +
> > +#if defined USE_MULTIARCH && IS_IN (libc)
> > +       .text
> > +ENTRY (__mempcpy_chk_erms)
> > +       cmp     %RDX_LP, %RCX_LP
> > +       jb      HIDDEN_JUMPTARGET (__chk_fail)
> > +END (__mempcpy_chk_erms)
> > +
> > +/* Only used to measure performance of REP MOVSB.  */
> > +ENTRY (__mempcpy_erms)
> > +       mov     %RDI_LP, %RAX_LP
> > +       /* Skip zero length.  */
> > +       test    %RDX_LP, %RDX_LP
> > +       jz      2f
> > +       add     %RDX_LP, %RAX_LP
> > +       jmp     L(start_movsb)
> > +END (__mempcpy_erms)
> > +
> > +ENTRY (__memmove_chk_erms)
> > +       cmp     %RDX_LP, %RCX_LP
> > +       jb      HIDDEN_JUMPTARGET (__chk_fail)
> > +END (__memmove_chk_erms)
> > +
> > +ENTRY (__memmove_erms)
> > +       movq    %rdi, %rax
> > +       /* Skip zero length.  */
> > +       test    %RDX_LP, %RDX_LP
> > +       jz      2f
> > +L(start_movsb):
> > +       mov     %RDX_LP, %RCX_LP
> > +       cmp     %RSI_LP, %RDI_LP
> > +       jb      1f
> > +       /* Source == destination is less common.  */
> > +       je      2f
> > +       lea     (%rsi,%rcx), %RDX_LP
> > +       cmp     %RDX_LP, %RDI_LP
> > +       jb      L(movsb_backward)
> > +1:
> > +       rep movsb
> > +2:
> > +       ret
> > +L(movsb_backward):
> > +       leaq    -1(%rdi,%rcx), %rdi
> > +       leaq    -1(%rsi,%rcx), %rsi
> > +       std
> > +       rep movsb
> > +       cld
> > +       ret
> > +END (__memmove_erms)
> > +strong_alias (__memmove_erms, __memcpy_erms)
> > +strong_alias (__memmove_chk_erms, __memcpy_chk_erms)
> > +#endif
> > diff --git a/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S b/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
> > index d1518b8bab..04747133b7 100644
> > --- a/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
> > +++ b/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
> > @@ -239,56 +239,6 @@ L(start):
> >  #endif
> >  #if defined USE_MULTIARCH && IS_IN (libc)
> >  END (MEMMOVE_SYMBOL (__memmove, unaligned))
> > -# if VEC_SIZE == 16
> > -ENTRY (__mempcpy_chk_erms)
> > -       cmp     %RDX_LP, %RCX_LP
> > -       jb      HIDDEN_JUMPTARGET (__chk_fail)
> > -END (__mempcpy_chk_erms)
> > -
> > -/* Only used to measure performance of REP MOVSB.  */
> > -ENTRY (__mempcpy_erms)
> > -       mov     %RDI_LP, %RAX_LP
> > -       /* Skip zero length.  */
> > -       test    %RDX_LP, %RDX_LP
> > -       jz      2f
> > -       add     %RDX_LP, %RAX_LP
> > -       jmp     L(start_movsb)
> > -END (__mempcpy_erms)
> > -
> > -ENTRY (__memmove_chk_erms)
> > -       cmp     %RDX_LP, %RCX_LP
> > -       jb      HIDDEN_JUMPTARGET (__chk_fail)
> > -END (__memmove_chk_erms)
> > -
> > -ENTRY (__memmove_erms)
> > -       movq    %rdi, %rax
> > -       /* Skip zero length.  */
> > -       test    %RDX_LP, %RDX_LP
> > -       jz      2f
> > -L(start_movsb):
> > -       mov     %RDX_LP, %RCX_LP
> > -       cmp     %RSI_LP, %RDI_LP
> > -       jb      1f
> > -       /* Source == destination is less common.  */
> > -       je      2f
> > -       lea     (%rsi,%rcx), %RDX_LP
> > -       cmp     %RDX_LP, %RDI_LP
> > -       jb      L(movsb_backward)
> > -1:
> > -       rep movsb
> > -2:
> > -       ret
> > -L(movsb_backward):
> > -       leaq    -1(%rdi,%rcx), %rdi
> > -       leaq    -1(%rsi,%rcx), %rsi
> > -       std
> > -       rep movsb
> > -       cld
> > -       ret
> > -END (__memmove_erms)
> > -strong_alias (__memmove_erms, __memcpy_erms)
> > -strong_alias (__memmove_chk_erms, __memcpy_chk_erms)
> > -# endif
> >
> >  # ifdef SHARED
> >  ENTRY (MEMMOVE_CHK_SYMBOL (__mempcpy_chk, unaligned_erms))
> > --
> > 2.34.1
> >
>
> LGTM.
>
> Thanks.
>
> --
> H.J.

I would like to backport this patch to release branches.
Any comments or objections?

--Sunil
  

Patch

diff --git a/sysdeps/x86_64/multiarch/Makefile b/sysdeps/x86_64/multiarch/Makefile
index 666ee4d5d6..62a4d96fb8 100644
--- a/sysdeps/x86_64/multiarch/Makefile
+++ b/sysdeps/x86_64/multiarch/Makefile
@@ -18,6 +18,7 @@  sysdep_routines += \
   memmove-avx-unaligned-erms-rtm \
   memmove-avx512-no-vzeroupper \
   memmove-avx512-unaligned-erms \
+  memmove-erms \
   memmove-evex-unaligned-erms \
   memmove-sse2-unaligned-erms \
   memmove-ssse3 \
diff --git a/sysdeps/x86_64/multiarch/memmove-erms.S b/sysdeps/x86_64/multiarch/memmove-erms.S
new file mode 100644
index 0000000000..2d3a6ccb76
--- /dev/null
+++ b/sysdeps/x86_64/multiarch/memmove-erms.S
@@ -0,0 +1,72 @@ 
+/* memcpy/mempcpy/memmove implement with rep movsb
+   Copyright (C) 2022 Free Software Foundation, Inc.
+   This file is part of the GNU C Library.
+
+   The GNU C Library is free software; you can redistribute it and/or
+   modify it under the terms of the GNU Lesser General Public
+   License as published by the Free Software Foundation; either
+   version 2.1 of the License, or (at your option) any later version.
+
+   The GNU C Library is distributed in the hope that it will be useful,
+   but WITHOUT ANY WARRANTY; without even the implied warranty of
+   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+   Lesser General Public License for more details.
+
+   You should have received a copy of the GNU Lesser General Public
+   License along with the GNU C Library; if not, see
+   <https://www.gnu.org/licenses/>.  */
+
+
+#include <sysdep.h>
+
+#if defined USE_MULTIARCH && IS_IN (libc)
+	.text
+ENTRY (__mempcpy_chk_erms)
+	cmp	%RDX_LP, %RCX_LP
+	jb	HIDDEN_JUMPTARGET (__chk_fail)
+END (__mempcpy_chk_erms)
+
+/* Only used to measure performance of REP MOVSB.  */
+ENTRY (__mempcpy_erms)
+	mov	%RDI_LP, %RAX_LP
+	/* Skip zero length.  */
+	test	%RDX_LP, %RDX_LP
+	jz	2f
+	add	%RDX_LP, %RAX_LP
+	jmp	L(start_movsb)
+END (__mempcpy_erms)
+
+ENTRY (__memmove_chk_erms)
+	cmp	%RDX_LP, %RCX_LP
+	jb	HIDDEN_JUMPTARGET (__chk_fail)
+END (__memmove_chk_erms)
+
+ENTRY (__memmove_erms)
+	movq	%rdi, %rax
+	/* Skip zero length.  */
+	test	%RDX_LP, %RDX_LP
+	jz	2f
+L(start_movsb):
+	mov	%RDX_LP, %RCX_LP
+	cmp	%RSI_LP, %RDI_LP
+	jb	1f
+	/* Source == destination is less common.  */
+	je	2f
+	lea	(%rsi,%rcx), %RDX_LP
+	cmp	%RDX_LP, %RDI_LP
+	jb	L(movsb_backward)
+1:
+	rep movsb
+2:
+	ret
+L(movsb_backward):
+	leaq	-1(%rdi,%rcx), %rdi
+	leaq	-1(%rsi,%rcx), %rsi
+	std
+	rep movsb
+	cld
+	ret
+END (__memmove_erms)
+strong_alias (__memmove_erms, __memcpy_erms)
+strong_alias (__memmove_chk_erms, __memcpy_chk_erms)
+#endif
diff --git a/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S b/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
index d1518b8bab..04747133b7 100644
--- a/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
+++ b/sysdeps/x86_64/multiarch/memmove-vec-unaligned-erms.S
@@ -239,56 +239,6 @@  L(start):
 #endif
 #if defined USE_MULTIARCH && IS_IN (libc)
 END (MEMMOVE_SYMBOL (__memmove, unaligned))
-# if VEC_SIZE == 16
-ENTRY (__mempcpy_chk_erms)
-	cmp	%RDX_LP, %RCX_LP
-	jb	HIDDEN_JUMPTARGET (__chk_fail)
-END (__mempcpy_chk_erms)
-
-/* Only used to measure performance of REP MOVSB.  */
-ENTRY (__mempcpy_erms)
-	mov	%RDI_LP, %RAX_LP
-	/* Skip zero length.  */
-	test	%RDX_LP, %RDX_LP
-	jz	2f
-	add	%RDX_LP, %RAX_LP
-	jmp	L(start_movsb)
-END (__mempcpy_erms)
-
-ENTRY (__memmove_chk_erms)
-	cmp	%RDX_LP, %RCX_LP
-	jb	HIDDEN_JUMPTARGET (__chk_fail)
-END (__memmove_chk_erms)
-
-ENTRY (__memmove_erms)
-	movq	%rdi, %rax
-	/* Skip zero length.  */
-	test	%RDX_LP, %RDX_LP
-	jz	2f
-L(start_movsb):
-	mov	%RDX_LP, %RCX_LP
-	cmp	%RSI_LP, %RDI_LP
-	jb	1f
-	/* Source == destination is less common.  */
-	je	2f
-	lea	(%rsi,%rcx), %RDX_LP
-	cmp	%RDX_LP, %RDI_LP
-	jb	L(movsb_backward)
-1:
-	rep movsb
-2:
-	ret
-L(movsb_backward):
-	leaq	-1(%rdi,%rcx), %rdi
-	leaq	-1(%rsi,%rcx), %rsi
-	std
-	rep movsb
-	cld
-	ret
-END (__memmove_erms)
-strong_alias (__memmove_erms, __memcpy_erms)
-strong_alias (__memmove_chk_erms, __memcpy_chk_erms)
-# endif
 
 # ifdef SHARED
 ENTRY (MEMMOVE_CHK_SYMBOL (__mempcpy_chk, unaligned_erms))