From patchwork Mon Mar 19 12:45:43 2018 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Andrew Senkevich X-Patchwork-Id: 26374 Received: (qmail 68441 invoked by alias); 19 Mar 2018 12:46:50 -0000 Mailing-List: contact libc-alpha-help@sourceware.org; run by ezmlm Precedence: bulk List-Id: List-Unsubscribe: List-Subscribe: List-Archive: List-Post: List-Help: , Sender: libc-alpha-owner@sourceware.org Delivered-To: mailing list libc-alpha@sourceware.org Received: (qmail 59582 invoked by uid 89); 19 Mar 2018 12:46:17 -0000 Authentication-Results: sourceware.org; auth=none X-Virus-Found: No X-Spam-SWARE-Status: No, score=-24.7 required=5.0 tests=AWL, BAYES_00, FREEMAIL_FROM, GIT_PATCH_0, GIT_PATCH_1, GIT_PATCH_2, GIT_PATCH_3, RCVD_IN_DNSWL_NONE, SPF_PASS autolearn=ham version=3.3.2 spammy=H*Ad:U*max X-HELO: mail-lf0-f66.google.com X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:in-reply-to:references:from:date :message-id:subject:to:cc; bh=xahtsImZWTNcOnrxFghiEHKu94PrbqHl8KgTNcbhF9w=; b=Dj2Z3eFqqHPobsx10XgDEJS8g5N5ukGbXODIGxGKWN7emRz1byYsQB+iErz2wSS3IQ ER5Vf1oRgbubki8YmVlztm+E2AFh7vMkiywvmaUDQdtcCvyyudpVOXwwZ97p0K1b/aZU smtPGNiH5Y00DpLgUREopIsXltk/oVjwiDibypBwgAFVaXopLXCgPWBFiSGVZQnaT2px aoTfmhTAWcbKFrbZsCgyp/K4uZHtaGEo474YKX/+MlAWwFbBE8DphACc6Esgl7YsIqaM KQ/zPiZbMZ6ebjNuInntIwlk67SiawzCLhYx11snjjlNxDqpBLtwXqrepXu/Pb9fKXqO ZWUA== X-Gm-Message-State: AElRT7FZmk7m4OdAJh6v2oY4yqWg+qVizmir+5Q039L2A03a8uTgpD1c XeREFRwJiymZgQSCSi//8wUQoOn7skJ0+HSuqaY= X-Google-Smtp-Source: AG47ELsXh2tKEzKkOUmghc8W2ff93T+eCq/+tizcAEEka4+YIEEVbE4BIozNoLgHR57HlVLpHfKg61FC+ZfxFAoL8aI= X-Received: by 2002:a19:3bc6:: with SMTP id d67-v6mr8299272lfl.100.1521463573346; Mon, 19 Mar 2018 05:46:13 -0700 (PDT) MIME-Version: 1.0 In-Reply-To: References: From: Andrew Senkevich Date: Mon, 19 Mar 2018 13:45:43 +0100 Message-ID: Subject: Re: [PATCH] Fix i386 memmove issue [BZ #22644] To: "H.J. Lu" Cc: Andreas Schwab , libc-alpha , Max Horn , thomas@grindinggear.com 2018-03-14 15:59 GMT+01:00 H.J. Lu : > On Wed, Mar 14, 2018 at 7:43 AM, Andrew Senkevich > wrote: >> 2018-02-19 11:13 GMT+01:00 Andreas Schwab : >>> On Feb 19 2018, Andrew Senkevich wrote: >>> >>>> diff --git a/string/test-memmove.c b/string/test-memmove.c >>>> index edc7a4c..8dc152b >>>> --- a/string/test-memmove.c >>>> +++ b/string/test-memmove.c >>>> @@ -245,6 +245,49 @@ do_random_tests (void) >>>> } >>>> } >>>> >>>> +#if __SIZEOF_POINTER__ == 4 >>>> +static void >>>> +do_test2 (void) >>>> +{ >>>> + uint32_t i; >>>> + uint32_t num = 0x20000000; >>>> + uint32_t * large_buf = mmap (0, sizeof(uint32_t) * num, PROT_READ | >>>> PROT_WRITE, >>>> + MAP_PRIVATE | MAP_ANON, -1, 0); >>>> + if (large_buf == MAP_FAILED) >>>> + error (EXIT_FAILURE, errno, "large mmap failed"); >>>> + >>>> + if (!((uint32_t)(large_buf) < (0x80000000 - 128) && (0x80000000 + >>>> 128) < (uint32_t)(&large_buf[num]))) >>>> + { >>>> + error (0, 0,"allocated large memory doesn't cross 0x80000000 boundary"); >>>> + ret = 1; >>>> + return; >>> >>> Please properly fold long lines, and remove the redundant parens. Also, >>> there is no guarantee that the address range is unallocated. >> >> Thanks, updated patch below. Any comment or it is Ok for trunk? > > Please also test crossing 0x80000000 boundary on 64-bit systems. Hi, I extended test for 64-bit using MAP_FIXED which lets to hardcode allocated address. Manual says it is less portable, but without MAP_FIXED I was unable to get proper address for 64bits. I checked what test fails without fix of memcpy-sse2-unaligned implementation. Is it Ok for trunk? --- WBR, Andrew diff --git a/string/test-memmove.c b/string/test-memmove.c index edc7a4c..5920652 100644 --- a/string/test-memmove.c +++ b/string/test-memmove.c @@ -24,6 +24,7 @@ # define TEST_NAME "memmove" #endif #include "test-string.h" +#include char *simple_memmove (char *, const char *, size_t); @@ -245,6 +246,57 @@ do_random_tests (void) } } +#if __SIZEOF_POINTER__ == 4 +# define ptr_type uint32_t +#else +# define ptr_type uint64_t +#endif + +static void +do_test2 (void) +{ + uint32_t num = 0x20000000; + uint32_t * large_buf; + + large_buf = mmap ((void*)0x70000000, num, PROT_READ | PROT_WRITE, + MAP_PRIVATE | MAP_ANON | MAP_FIXED, -1, 0); + + if (large_buf == MAP_FAILED) + error (EXIT_FAILURE, errno, "Large mmap failed"); + + uint32_t bytes_move = 0x80000000 - (ptr_type)large_buf; + uint32_t arr_size = bytes_move / sizeof(uint32_t); + uint32_t i; + + FOR_EACH_IMPL (impl, 0) + { + for (i = 0; i < arr_size; i++) + large_buf[i] = i; + + uint32_t * dst = &large_buf[33]; + +#ifdef TEST_BCOPY + CALL (impl, (char *)large_buf, (char *)dst, bytes_move); +#else + CALL (impl, (char *)dst, (char *)large_buf, bytes_move); +#endif + + for (i = 0; i < arr_size; i++) + { + if (dst[i] != i) + { + error (0, 0, + "Wrong result in function %s dst \"%p\" src \"%p\" offset \"%d\"", + impl->name, dst, large_buf, i); + ret = 1; + break; + } + } + } + + munmap((void *)large_buf, sizeof(uint32_t) * num); +} + int test_main (void) { @@ -284,6 +336,9 @@ test_main (void) } do_random_tests (); + + do_test2 (); + return ret; } diff --git a/sysdeps/i386/i686/multiarch/memcpy-sse2-unaligned.S b/sysdeps/i386/i686/multiarch/memcpy-sse2-unaligned.S index 9c3bbe7..9aa17de 100644 --- a/sysdeps/i386/i686/multiarch/memcpy-sse2-unaligned.S +++ b/sysdeps/i386/i686/multiarch/memcpy-sse2-unaligned.S @@ -72,7 +72,7 @@ ENTRY (MEMCPY) cmp %edx, %eax # ifdef USE_AS_MEMMOVE - jg L(check_forward) + ja L(check_forward) L(mm_len_0_or_more_backward): /* Now do checks for lengths. We do [0..16], [16..32], [32..64], [64..128] @@ -81,7 +81,7 @@ L(mm_len_0_or_more_backward): jbe L(mm_len_0_16_bytes_backward) cmpl $32, %ecx - jg L(mm_len_32_or_more_backward) + ja L(mm_len_32_or_more_backward) /* Copy [0..32] and return. */ movdqu (%eax), %xmm0 @@ -92,7 +92,7 @@ L(mm_len_0_or_more_backward): L(mm_len_32_or_more_backward): cmpl $64, %ecx - jg L(mm_len_64_or_more_backward) + ja L(mm_len_64_or_more_backward) /* Copy [0..64] and return. */ movdqu (%eax), %xmm0 @@ -107,7 +107,7 @@ L(mm_len_32_or_more_backward): L(mm_len_64_or_more_backward): cmpl $128, %ecx - jg L(mm_len_128_or_more_backward) + ja L(mm_len_128_or_more_backward) /* Copy [0..128] and return. */ movdqu (%eax), %xmm0 @@ -132,7 +132,7 @@ L(mm_len_128_or_more_backward): add %ecx, %eax cmp %edx, %eax movl SRC(%esp), %eax - jle L(forward) + jbe L(forward) PUSH (%esi) PUSH (%edi) PUSH (%ebx) @@ -269,7 +269,7 @@ L(check_forward): add %edx, %ecx cmp %eax, %ecx movl LEN(%esp), %ecx - jle L(forward) + jbe L(forward) /* Now do checks for lengths. We do [0..16], [0..32], [0..64], [0..128] separately. */