From patchwork Sun Apr 21 05:01:35 2024 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: "Li, Pan2" X-Patchwork-Id: 88808 Return-Path: X-Original-To: patchwork@sourceware.org Delivered-To: patchwork@sourceware.org Received: from server2.sourceware.org (localhost [IPv6:::1]) by sourceware.org (Postfix) with ESMTP id E79AB3858404 for ; Sun, 21 Apr 2024 05:02:09 +0000 (GMT) X-Original-To: gcc-patches@gcc.gnu.org Delivered-To: gcc-patches@gcc.gnu.org Received: from mgamail.intel.com (mgamail.intel.com [192.198.163.15]) by sourceware.org (Postfix) with ESMTPS id CF0EA3858D28 for ; Sun, 21 Apr 2024 05:01:41 +0000 (GMT) DMARC-Filter: OpenDMARC Filter v1.4.2 sourceware.org CF0EA3858D28 Authentication-Results: sourceware.org; dmarc=pass (p=none dis=none) header.from=intel.com Authentication-Results: sourceware.org; spf=pass smtp.mailfrom=intel.com ARC-Filter: OpenARC Filter v1.0.0 sourceware.org CF0EA3858D28 Authentication-Results: server2.sourceware.org; arc=none smtp.remote-ip=192.198.163.15 ARC-Seal: i=1; a=rsa-sha256; d=sourceware.org; s=key; t=1713675705; cv=none; b=PRyHTYZp7i86w6R5TKpCjLBuFeAFtHDy3x28nBmEqE8QG2kymI0wCovZgvEgicRNowpPoG4ZV58K3aXgXS+yG/nNIa3rnFOLzTj3qsYVhzC3MMKKONKsvVwcq/X0E1dNNLSi3M+b3YqHv1THSI+bjI2s3z/NL+Qnt+gScnYo4d8= ARC-Message-Signature: i=1; a=rsa-sha256; d=sourceware.org; s=key; t=1713675705; c=relaxed/simple; bh=E2wKcP9J1QBqYF3nA9agQeCjElblME1GtnxqGl1tc94=; h=DKIM-Signature:From:To:Subject:Date:Message-Id:MIME-Version; b=YEZTYAA8oigaWKX8qC9akZSlOnyjMW2Pkhec5C4Pwd5Z3W++yiRJd0M8ocFgOZrZq2+WFB3lK9RrM3BgN1b/ZJJ+5wchD8AkT/lVEAvOMamyZgXyQGun1rz1kjsxu27vtbGYn0rHhdi47p7VlQLMj9MjoZTAX9DKXnf4IEr9w+M= ARC-Authentication-Results: i=1; server2.sourceware.org DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1713675702; x=1745211702; h=from:to:cc:subject:date:message-id:mime-version: content-transfer-encoding; bh=E2wKcP9J1QBqYF3nA9agQeCjElblME1GtnxqGl1tc94=; b=INu/OfSPIPI5nxTnXoLUgiu0MnfYZItUe9pYOVNaYDPNLKu5WUBHCnoB zH6K83b4/J5uEkTeqZ1zJv9NhHegvsTff63PYrcwqR3tCLigRkKXcdhhK agD37lYyUDyI3qUgHCUnaDo8Q+s9U7fN5tm5JwrqZ17JX8tm0oRhVP3x2 f4/NldhBmbY8szoFgSdD4BTdR1Pv+frOLL+51yAYVVpPUPo0tKW7XC72A GjjIJA5snDeDSmfgHUwA2RdNDCfiK/R+JJ4AEf4ToTqqxF90WadYJSPFc rENxTdD8S/uABucClb4PAm2l1wfJpN9a6vaqsIXAj9DWBi/Xh1jQWRTxq w==; X-CSE-ConnectionGUID: 3eWC4d/aQ1efqQT8jZUNWA== X-CSE-MsgGUID: tXjkRtMAQ2qha7OQ+D/rbw== X-IronPort-AV: E=McAfee;i="6600,9927,11050"; a="9407347" X-IronPort-AV: E=Sophos;i="6.07,217,1708416000"; d="scan'208";a="9407347" Received: from orviesa003.jf.intel.com ([10.64.159.143]) by fmvoesa109.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 20 Apr 2024 22:01:40 -0700 X-CSE-ConnectionGUID: L4eT2Mh0RNSriCIu3O6xIA== X-CSE-MsgGUID: Tmg8mjkOTc+Dl+cdTOvypg== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="6.07,217,1708416000"; d="scan'208";a="28358738" Received: from shvmail02.sh.intel.com ([10.239.244.9]) by orviesa003.jf.intel.com with ESMTP; 20 Apr 2024 22:01:38 -0700 Received: from pli-ubuntu.sh.intel.com (pli-ubuntu.sh.intel.com [10.239.159.47]) by shvmail02.sh.intel.com (Postfix) with ESMTP id B0B0E1006F25; Sun, 21 Apr 2024 13:01:36 +0800 (CST) From: pan2.li@intel.com To: gcc-patches@gcc.gnu.org Cc: juzhe.zhong@rivai.ai, kito.cheng@gmail.com, rdapp.gcc@gmail.com, Pan Li Subject: [PATCH v1] RISC-V: Add xfail test case for highpart register overlap of vx/vf widen Date: Sun, 21 Apr 2024 13:01:35 +0800 Message-Id: <20240421050135.3977233-1-pan2.li@intel.com> X-Mailer: git-send-email 2.34.1 MIME-Version: 1.0 X-Spam-Status: No, score=-11.7 required=5.0 tests=BAYES_00, DKIMWL_WL_HIGH, DKIM_SIGNED, DKIM_VALID, DKIM_VALID_AU, DKIM_VALID_EF, GIT_PATCH_0, KAM_SHORT, SPF_HELO_NONE, SPF_NONE, TXREP autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on server2.sourceware.org X-BeenThere: gcc-patches@gcc.gnu.org X-Mailman-Version: 2.1.30 Precedence: list List-Id: Gcc-patches mailing list List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: gcc-patches-bounces+patchwork=sourceware.org@gcc.gnu.org From: Pan Li We reverted below patch for register group overlap, add the related insn test and mark it as xfail. And we will remove the xfail after we support the register overlap in GCC-15. a23415d7572 RISC-V: Support highpart register overlap for widen vx/vf instructions The below test suites are passed. * The rv64gcv fully regression test. gcc/testsuite/ChangeLog: * gcc.target/riscv/rvv/base/pr112431-22.c: New test. * gcc.target/riscv/rvv/base/pr112431-23.c: New test. * gcc.target/riscv/rvv/base/pr112431-24.c: New test. * gcc.target/riscv/rvv/base/pr112431-25.c: New test. * gcc.target/riscv/rvv/base/pr112431-26.c: New test. * gcc.target/riscv/rvv/base/pr112431-27.c: New test. Signed-off-by: Pan Li Signed-off-by: Pan Li Signed-off-by: Pan Li > --- .../gcc.target/riscv/rvv/base/pr112431-22.c | 188 ++++++++++++++++++ .../gcc.target/riscv/rvv/base/pr112431-23.c | 119 +++++++++++ .../gcc.target/riscv/rvv/base/pr112431-24.c | 86 ++++++++ .../gcc.target/riscv/rvv/base/pr112431-25.c | 104 ++++++++++ .../gcc.target/riscv/rvv/base/pr112431-26.c | 68 +++++++ .../gcc.target/riscv/rvv/base/pr112431-27.c | 51 +++++ 6 files changed, 616 insertions(+) create mode 100644 gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-22.c create mode 100644 gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-23.c create mode 100644 gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-24.c create mode 100644 gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-25.c create mode 100644 gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-26.c create mode 100644 gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-27.c diff --git a/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-22.c b/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-22.c new file mode 100644 index 00000000000..ac56703c75c --- /dev/null +++ b/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-22.c @@ -0,0 +1,188 @@ +/* { dg-do compile } */ +/* { dg-options "-march=rv64gcv -mabi=lp64d -O3" } */ + +#include "riscv_vector.h" + +size_t __attribute__ ((noinline)) +sumation (size_t sum0, size_t sum1, size_t sum2, size_t sum3, size_t sum4, + size_t sum5, size_t sum6, size_t sum7, size_t sum8, size_t sum9, + size_t sum10, size_t sum11, size_t sum12, size_t sum13, size_t sum14, + size_t sum15) +{ + return sum0 + sum1 + sum2 + sum3 + sum4 + sum5 + sum6 + sum7 + sum8 + sum9 + + sum10 + sum11 + sum12 + sum13 + sum14 + sum15; +} + +size_t +foo (char const *buf, size_t len) +{ + size_t sum = 0; + size_t vl = __riscv_vsetvlmax_e8m8 (); + size_t step = vl * 4; + const char *it = buf, *end = buf + len; + for (; it + step <= end;) + { + vint8m1_t v0 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v1 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v2 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v3 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v4 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v5 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v6 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v7 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v8 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v9 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v10 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v11 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v12 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v13 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v14 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v15 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + + asm volatile("nop" ::: "memory"); + vint16m2_t vw0 = __riscv_vwadd_vx_i16m2 (v0, 33, vl); + vint16m2_t vw1 = __riscv_vwadd_vx_i16m2 (v1, 33, vl); + vint16m2_t vw2 = __riscv_vwadd_vx_i16m2 (v2, 33, vl); + vint16m2_t vw3 = __riscv_vwadd_vx_i16m2 (v3, 33, vl); + vint16m2_t vw4 = __riscv_vwadd_vx_i16m2 (v4, 33, vl); + vint16m2_t vw5 = __riscv_vwadd_vx_i16m2 (v5, 33, vl); + vint16m2_t vw6 = __riscv_vwadd_vx_i16m2 (v6, 33, vl); + vint16m2_t vw7 = __riscv_vwadd_vx_i16m2 (v7, 33, vl); + vint16m2_t vw8 = __riscv_vwadd_vx_i16m2 (v8, 33, vl); + vint16m2_t vw9 = __riscv_vwadd_vx_i16m2 (v9, 33, vl); + vint16m2_t vw10 = __riscv_vwadd_vx_i16m2 (v10, 33, vl); + vint16m2_t vw11 = __riscv_vwadd_vx_i16m2 (v11, 33, vl); + vint16m2_t vw12 = __riscv_vwadd_vx_i16m2 (v12, 33, vl); + vint16m2_t vw13 = __riscv_vwadd_vx_i16m2 (v13, 33, vl); + vint16m2_t vw14 = __riscv_vwadd_vx_i16m2 (v14, 33, vl); + vint16m2_t vw15 = __riscv_vwadd_vx_i16m2 (v15, 33, vl); + + asm volatile("nop" ::: "memory"); + size_t sum0 = __riscv_vmv_x_s_i16m2_i16 (vw0); + size_t sum1 = __riscv_vmv_x_s_i16m2_i16 (vw1); + size_t sum2 = __riscv_vmv_x_s_i16m2_i16 (vw2); + size_t sum3 = __riscv_vmv_x_s_i16m2_i16 (vw3); + size_t sum4 = __riscv_vmv_x_s_i16m2_i16 (vw4); + size_t sum5 = __riscv_vmv_x_s_i16m2_i16 (vw5); + size_t sum6 = __riscv_vmv_x_s_i16m2_i16 (vw6); + size_t sum7 = __riscv_vmv_x_s_i16m2_i16 (vw7); + size_t sum8 = __riscv_vmv_x_s_i16m2_i16 (vw8); + size_t sum9 = __riscv_vmv_x_s_i16m2_i16 (vw9); + size_t sum10 = __riscv_vmv_x_s_i16m2_i16 (vw10); + size_t sum11 = __riscv_vmv_x_s_i16m2_i16 (vw11); + size_t sum12 = __riscv_vmv_x_s_i16m2_i16 (vw12); + size_t sum13 = __riscv_vmv_x_s_i16m2_i16 (vw13); + size_t sum14 = __riscv_vmv_x_s_i16m2_i16 (vw14); + size_t sum15 = __riscv_vmv_x_s_i16m2_i16 (vw15); + + sum += sumation (sum0, sum1, sum2, sum3, sum4, sum5, sum6, sum7, sum8, + sum9, sum10, sum11, sum12, sum13, sum14, sum15); + } + return sum; +} + +size_t +foo2 (char const *buf, size_t len) +{ + size_t sum = 0; + size_t vl = __riscv_vsetvlmax_e8m8 (); + size_t step = vl * 4; + const char *it = buf, *end = buf + len; + for (; it + step <= end;) + { + vint8m1_t v0 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v1 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v2 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v3 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v4 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v5 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v6 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v7 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v8 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v9 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v10 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v11 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v12 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v13 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v14 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + vint8m1_t v15 = __riscv_vle8_v_i8m1 ((void *) it, vl); + it += vl; + + asm volatile("nop" ::: "memory"); + vint16m2_t vw0 = __riscv_vwmulsu_vx_i16m2 (v0, 33, vl); + vint16m2_t vw1 = __riscv_vwmulsu_vx_i16m2 (v1, 33, vl); + vint16m2_t vw2 = __riscv_vwmulsu_vx_i16m2 (v2, 33, vl); + vint16m2_t vw3 = __riscv_vwmulsu_vx_i16m2 (v3, 33, vl); + vint16m2_t vw4 = __riscv_vwmulsu_vx_i16m2 (v4, 33, vl); + vint16m2_t vw5 = __riscv_vwmulsu_vx_i16m2 (v5, 33, vl); + vint16m2_t vw6 = __riscv_vwmulsu_vx_i16m2 (v6, 33, vl); + vint16m2_t vw7 = __riscv_vwmulsu_vx_i16m2 (v7, 33, vl); + vint16m2_t vw8 = __riscv_vwmulsu_vx_i16m2 (v8, 33, vl); + vint16m2_t vw9 = __riscv_vwmulsu_vx_i16m2 (v9, 33, vl); + vint16m2_t vw10 = __riscv_vwmulsu_vx_i16m2 (v10, 33, vl); + vint16m2_t vw11 = __riscv_vwmulsu_vx_i16m2 (v11, 33, vl); + vint16m2_t vw12 = __riscv_vwmulsu_vx_i16m2 (v12, 33, vl); + vint16m2_t vw13 = __riscv_vwmulsu_vx_i16m2 (v13, 33, vl); + vint16m2_t vw14 = __riscv_vwmulsu_vx_i16m2 (v14, 33, vl); + vint16m2_t vw15 = __riscv_vwmulsu_vx_i16m2 (v15, 33, vl); + + asm volatile("nop" ::: "memory"); + size_t sum0 = __riscv_vmv_x_s_i16m2_i16 (vw0); + size_t sum1 = __riscv_vmv_x_s_i16m2_i16 (vw1); + size_t sum2 = __riscv_vmv_x_s_i16m2_i16 (vw2); + size_t sum3 = __riscv_vmv_x_s_i16m2_i16 (vw3); + size_t sum4 = __riscv_vmv_x_s_i16m2_i16 (vw4); + size_t sum5 = __riscv_vmv_x_s_i16m2_i16 (vw5); + size_t sum6 = __riscv_vmv_x_s_i16m2_i16 (vw6); + size_t sum7 = __riscv_vmv_x_s_i16m2_i16 (vw7); + size_t sum8 = __riscv_vmv_x_s_i16m2_i16 (vw8); + size_t sum9 = __riscv_vmv_x_s_i16m2_i16 (vw9); + size_t sum10 = __riscv_vmv_x_s_i16m2_i16 (vw10); + size_t sum11 = __riscv_vmv_x_s_i16m2_i16 (vw11); + size_t sum12 = __riscv_vmv_x_s_i16m2_i16 (vw12); + size_t sum13 = __riscv_vmv_x_s_i16m2_i16 (vw13); + size_t sum14 = __riscv_vmv_x_s_i16m2_i16 (vw14); + size_t sum15 = __riscv_vmv_x_s_i16m2_i16 (vw15); + + sum += sumation (sum0, sum1, sum2, sum3, sum4, sum5, sum6, sum7, sum8, + sum9, sum10, sum11, sum12, sum13, sum14, sum15); + } + return sum; +} + +/* { dg-final { scan-assembler-not {vmv1r} } } */ +/* { dg-final { scan-assembler-not {vmv2r} } } */ +/* { dg-final { scan-assembler-not {vmv4r} } } */ +/* { dg-final { scan-assembler-not {vmv8r} } } */ +/* { dg-final { scan-assembler-not {csrr} { xfail riscv*-*-* } } } */ diff --git a/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-23.c b/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-23.c new file mode 100644 index 00000000000..f91119307f4 --- /dev/null +++ b/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-23.c @@ -0,0 +1,119 @@ +/* { dg-do compile } */ +/* { dg-options "-march=rv64gcv -mabi=lp64d -O3" } */ + +#include "riscv_vector.h" + +size_t __attribute__ ((noinline)) +sumation (size_t sum0, size_t sum1, size_t sum2, size_t sum3, size_t sum4, + size_t sum5, size_t sum6, size_t sum7) +{ + return sum0 + sum1 + sum2 + sum3 + sum4 + sum5 + sum6 + sum7; +} + +size_t +foo (char const *buf, size_t len) +{ + size_t sum = 0; + size_t vl = __riscv_vsetvlmax_e8m8 (); + size_t step = vl * 4; + const char *it = buf, *end = buf + len; + for (; it + step <= end;) + { + vint8m2_t v0 = __riscv_vle8_v_i8m2 ((void *) it, vl); + it += vl; + vint8m2_t v1 = __riscv_vle8_v_i8m2 ((void *) it, vl); + it += vl; + vint8m2_t v2 = __riscv_vle8_v_i8m2 ((void *) it, vl); + it += vl; + vint8m2_t v3 = __riscv_vle8_v_i8m2 ((void *) it, vl); + it += vl; + vint8m2_t v4 = __riscv_vle8_v_i8m2 ((void *) it, vl); + it += vl; + vint8m2_t v5 = __riscv_vle8_v_i8m2 ((void *) it, vl); + it += vl; + vint8m2_t v6 = __riscv_vle8_v_i8m2 ((void *) it, vl); + it += vl; + vint8m2_t v7 = __riscv_vle8_v_i8m2 ((void *) it, vl); + it += vl; + + asm volatile("nop" ::: "memory"); + vint16m4_t vw0 = __riscv_vwadd_vx_i16m4 (v0, 55, vl); + vint16m4_t vw1 = __riscv_vwadd_vx_i16m4 (v1, 55, vl); + vint16m4_t vw2 = __riscv_vwadd_vx_i16m4 (v2, 55, vl); + vint16m4_t vw3 = __riscv_vwadd_vx_i16m4 (v3, 55, vl); + vint16m4_t vw4 = __riscv_vwadd_vx_i16m4 (v4, 55, vl); + vint16m4_t vw5 = __riscv_vwadd_vx_i16m4 (v5, 55, vl); + vint16m4_t vw6 = __riscv_vwadd_vx_i16m4 (v6, 55, vl); + vint16m4_t vw7 = __riscv_vwadd_vx_i16m4 (v7, 55, vl); + + asm volatile("nop" ::: "memory"); + size_t sum0 = __riscv_vmv_x_s_i16m4_i16 (vw0); + size_t sum1 = __riscv_vmv_x_s_i16m4_i16 (vw1); + size_t sum2 = __riscv_vmv_x_s_i16m4_i16 (vw2); + size_t sum3 = __riscv_vmv_x_s_i16m4_i16 (vw3); + size_t sum4 = __riscv_vmv_x_s_i16m4_i16 (vw4); + size_t sum5 = __riscv_vmv_x_s_i16m4_i16 (vw5); + size_t sum6 = __riscv_vmv_x_s_i16m4_i16 (vw6); + size_t sum7 = __riscv_vmv_x_s_i16m4_i16 (vw7); + + sum += sumation (sum0, sum1, sum2, sum3, sum4, sum5, sum6, sum7); + } + return sum; +} + +size_t +foo2 (char const *buf, size_t len) +{ + size_t sum = 0; + size_t vl = __riscv_vsetvlmax_e8m8 (); + size_t step = vl * 4; + const char *it = buf, *end = buf + len; + for (; it + step <= end;) + { + vint8m2_t v0 = __riscv_vle8_v_i8m2 ((void *) it, vl); + it += vl; + vint8m2_t v1 = __riscv_vle8_v_i8m2 ((void *) it, vl); + it += vl; + vint8m2_t v2 = __riscv_vle8_v_i8m2 ((void *) it, vl); + it += vl; + vint8m2_t v3 = __riscv_vle8_v_i8m2 ((void *) it, vl); + it += vl; + vint8m2_t v4 = __riscv_vle8_v_i8m2 ((void *) it, vl); + it += vl; + vint8m2_t v5 = __riscv_vle8_v_i8m2 ((void *) it, vl); + it += vl; + vint8m2_t v6 = __riscv_vle8_v_i8m2 ((void *) it, vl); + it += vl; + vint8m2_t v7 = __riscv_vle8_v_i8m2 ((void *) it, vl); + it += vl; + + asm volatile("nop" ::: "memory"); + vint16m4_t vw0 = __riscv_vwmulsu_vx_i16m4 (v0, 55, vl); + vint16m4_t vw1 = __riscv_vwmulsu_vx_i16m4 (v1, 55, vl); + vint16m4_t vw2 = __riscv_vwmulsu_vx_i16m4 (v2, 55, vl); + vint16m4_t vw3 = __riscv_vwmulsu_vx_i16m4 (v3, 55, vl); + vint16m4_t vw4 = __riscv_vwmulsu_vx_i16m4 (v4, 55, vl); + vint16m4_t vw5 = __riscv_vwmulsu_vx_i16m4 (v5, 55, vl); + vint16m4_t vw6 = __riscv_vwmulsu_vx_i16m4 (v6, 55, vl); + vint16m4_t vw7 = __riscv_vwmulsu_vx_i16m4 (v7, 55, vl); + + asm volatile("nop" ::: "memory"); + size_t sum0 = __riscv_vmv_x_s_i16m4_i16 (vw0); + size_t sum1 = __riscv_vmv_x_s_i16m4_i16 (vw1); + size_t sum2 = __riscv_vmv_x_s_i16m4_i16 (vw2); + size_t sum3 = __riscv_vmv_x_s_i16m4_i16 (vw3); + size_t sum4 = __riscv_vmv_x_s_i16m4_i16 (vw4); + size_t sum5 = __riscv_vmv_x_s_i16m4_i16 (vw5); + size_t sum6 = __riscv_vmv_x_s_i16m4_i16 (vw6); + size_t sum7 = __riscv_vmv_x_s_i16m4_i16 (vw7); + + sum += sumation (sum0, sum1, sum2, sum3, sum4, sum5, sum6, sum7); + } + return sum; +} + +/* { dg-final { scan-assembler-not {vmv1r} } } */ +/* { dg-final { scan-assembler-not {vmv2r} } } */ +/* { dg-final { scan-assembler-not {vmv4r} } } */ +/* { dg-final { scan-assembler-not {vmv8r} } } */ +/* { dg-final { scan-assembler-not {csrr} { xfail riscv*-*-* } } } */ diff --git a/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-24.c b/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-24.c new file mode 100644 index 00000000000..bcd60c0a7c3 --- /dev/null +++ b/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-24.c @@ -0,0 +1,86 @@ +/* { dg-do compile } */ +/* { dg-options "-march=rv64gcv -mabi=lp64d -O3" } */ + +#include "riscv_vector.h" + +size_t __attribute__ ((noinline)) +sumation (size_t sum0, size_t sum1, size_t sum2, size_t sum3) +{ + return sum0 + sum1 + sum2 + sum3; +} + +size_t +foo (char const *buf, size_t len) +{ + size_t sum = 0; + size_t vl = __riscv_vsetvlmax_e8m8 (); + size_t step = vl * 4; + const char *it = buf, *end = buf + len; + for (; it + step <= end;) + { + vint8m4_t v0 = __riscv_vle8_v_i8m4 ((void *) it, vl); + it += vl; + vint8m4_t v1 = __riscv_vle8_v_i8m4 ((void *) it, vl); + it += vl; + vint8m4_t v2 = __riscv_vle8_v_i8m4 ((void *) it, vl); + it += vl; + vint8m4_t v3 = __riscv_vle8_v_i8m4 ((void *) it, vl); + it += vl; + + asm volatile("nop" ::: "memory"); + vint16m8_t vw0 = __riscv_vwadd_vx_i16m8 (v0, 66, vl); + vint16m8_t vw1 = __riscv_vwadd_vx_i16m8 (v1, 66, vl); + vint16m8_t vw2 = __riscv_vwadd_vx_i16m8 (v2, 66, vl); + vint16m8_t vw3 = __riscv_vwadd_vx_i16m8 (v3, 66, vl); + + asm volatile("nop" ::: "memory"); + size_t sum0 = __riscv_vmv_x_s_i16m8_i16 (vw0); + size_t sum1 = __riscv_vmv_x_s_i16m8_i16 (vw1); + size_t sum2 = __riscv_vmv_x_s_i16m8_i16 (vw2); + size_t sum3 = __riscv_vmv_x_s_i16m8_i16 (vw3); + + sum += sumation (sum0, sum1, sum2, sum3); + } + return sum; +} + +size_t +foo2 (char const *buf, size_t len) +{ + size_t sum = 0; + size_t vl = __riscv_vsetvlmax_e8m8 (); + size_t step = vl * 4; + const char *it = buf, *end = buf + len; + for (; it + step <= end;) + { + vint8m4_t v0 = __riscv_vle8_v_i8m4 ((void *) it, vl); + it += vl; + vint8m4_t v1 = __riscv_vle8_v_i8m4 ((void *) it, vl); + it += vl; + vint8m4_t v2 = __riscv_vle8_v_i8m4 ((void *) it, vl); + it += vl; + vint8m4_t v3 = __riscv_vle8_v_i8m4 ((void *) it, vl); + it += vl; + + asm volatile("nop" ::: "memory"); + vint16m8_t vw0 = __riscv_vwmulsu_vx_i16m8 (v0, 66, vl); + vint16m8_t vw1 = __riscv_vwmulsu_vx_i16m8 (v1, 66, vl); + vint16m8_t vw2 = __riscv_vwmulsu_vx_i16m8 (v2, 66, vl); + vint16m8_t vw3 = __riscv_vwmulsu_vx_i16m8 (v3, 66, vl); + + asm volatile("nop" ::: "memory"); + size_t sum0 = __riscv_vmv_x_s_i16m8_i16 (vw0); + size_t sum1 = __riscv_vmv_x_s_i16m8_i16 (vw1); + size_t sum2 = __riscv_vmv_x_s_i16m8_i16 (vw2); + size_t sum3 = __riscv_vmv_x_s_i16m8_i16 (vw3); + + sum += sumation (sum0, sum1, sum2, sum3); + } + return sum; +} + +/* { dg-final { scan-assembler-not {vmv1r} } } */ +/* { dg-final { scan-assembler-not {vmv2r} } } */ +/* { dg-final { scan-assembler-not {vmv4r} } } */ +/* { dg-final { scan-assembler-not {vmv8r} } } */ +/* { dg-final { scan-assembler-not {csrr} { xfail riscv*-*-* } } } */ diff --git a/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-25.c b/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-25.c new file mode 100644 index 00000000000..57a8ef28486 --- /dev/null +++ b/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-25.c @@ -0,0 +1,104 @@ +/* { dg-do compile } */ +/* { dg-options "-march=rv64gcv -mabi=lp64d -O3" } */ + +#include "riscv_vector.h" + +size_t __attribute__ ((noinline)) +sumation (size_t sum0, size_t sum1, size_t sum2, size_t sum3, size_t sum4, + size_t sum5, size_t sum6, size_t sum7, size_t sum8, size_t sum9, + size_t sum10, size_t sum11, size_t sum12, size_t sum13, size_t sum14, + size_t sum15) +{ + return sum0 + sum1 + sum2 + sum3 + sum4 + sum5 + sum6 + sum7 + sum8 + sum9 + + sum10 + sum11 + sum12 + sum13 + sum14 + sum15; +} + +size_t +foo (char const *buf, size_t len) +{ + size_t sum = 0; + size_t vl = __riscv_vsetvlmax_e8m8 (); + size_t step = vl * 4; + const char *it = buf, *end = buf + len; + for (; it + step <= end;) + { + vfloat32m1_t v0 = __riscv_vle32_v_f32m1 ((void *) it, vl); + it += vl; + vfloat32m1_t v1 = __riscv_vle32_v_f32m1 ((void *) it, vl); + it += vl; + vfloat32m1_t v2 = __riscv_vle32_v_f32m1 ((void *) it, vl); + it += vl; + vfloat32m1_t v3 = __riscv_vle32_v_f32m1 ((void *) it, vl); + it += vl; + vfloat32m1_t v4 = __riscv_vle32_v_f32m1 ((void *) it, vl); + it += vl; + vfloat32m1_t v5 = __riscv_vle32_v_f32m1 ((void *) it, vl); + it += vl; + vfloat32m1_t v6 = __riscv_vle32_v_f32m1 ((void *) it, vl); + it += vl; + vfloat32m1_t v7 = __riscv_vle32_v_f32m1 ((void *) it, vl); + it += vl; + vfloat32m1_t v8 = __riscv_vle32_v_f32m1 ((void *) it, vl); + it += vl; + vfloat32m1_t v9 = __riscv_vle32_v_f32m1 ((void *) it, vl); + it += vl; + vfloat32m1_t v10 = __riscv_vle32_v_f32m1 ((void *) it, vl); + it += vl; + vfloat32m1_t v11 = __riscv_vle32_v_f32m1 ((void *) it, vl); + it += vl; + vfloat32m1_t v12 = __riscv_vle32_v_f32m1 ((void *) it, vl); + it += vl; + vfloat32m1_t v13 = __riscv_vle32_v_f32m1 ((void *) it, vl); + it += vl; + vfloat32m1_t v14 = __riscv_vle32_v_f32m1 ((void *) it, vl); + it += vl; + vfloat32m1_t v15 = __riscv_vle32_v_f32m1 ((void *) it, vl); + it += vl; + + asm volatile("nop" ::: "memory"); + vfloat64m2_t vw0 = __riscv_vfwadd_vf_f64m2 (v0, 33, vl); + vfloat64m2_t vw1 = __riscv_vfwadd_vf_f64m2 (v1, 33, vl); + vfloat64m2_t vw2 = __riscv_vfwadd_vf_f64m2 (v2, 33, vl); + vfloat64m2_t vw3 = __riscv_vfwadd_vf_f64m2 (v3, 33, vl); + vfloat64m2_t vw4 = __riscv_vfwadd_vf_f64m2 (v4, 33, vl); + vfloat64m2_t vw5 = __riscv_vfwadd_vf_f64m2 (v5, 33, vl); + vfloat64m2_t vw6 = __riscv_vfwadd_vf_f64m2 (v6, 33, vl); + vfloat64m2_t vw7 = __riscv_vfwadd_vf_f64m2 (v7, 33, vl); + vfloat64m2_t vw8 = __riscv_vfwadd_vf_f64m2 (v8, 33, vl); + vfloat64m2_t vw9 = __riscv_vfwadd_vf_f64m2 (v9, 33, vl); + vfloat64m2_t vw10 = __riscv_vfwadd_vf_f64m2 (v10, 33, vl); + vfloat64m2_t vw11 = __riscv_vfwadd_vf_f64m2 (v11, 33, vl); + vfloat64m2_t vw12 = __riscv_vfwadd_vf_f64m2 (v12, 33, vl); + vfloat64m2_t vw13 = __riscv_vfwadd_vf_f64m2 (v13, 33, vl); + vfloat64m2_t vw14 = __riscv_vfwadd_vf_f64m2 (v14, 33, vl); + vfloat64m2_t vw15 = __riscv_vfwadd_vf_f64m2 (v15, 33, vl); + + asm volatile("nop" ::: "memory"); + size_t sum0 = __riscv_vfmv_f_s_f64m2_f64 (vw0); + size_t sum1 = __riscv_vfmv_f_s_f64m2_f64 (vw1); + size_t sum2 = __riscv_vfmv_f_s_f64m2_f64 (vw2); + size_t sum3 = __riscv_vfmv_f_s_f64m2_f64 (vw3); + size_t sum4 = __riscv_vfmv_f_s_f64m2_f64 (vw4); + size_t sum5 = __riscv_vfmv_f_s_f64m2_f64 (vw5); + size_t sum6 = __riscv_vfmv_f_s_f64m2_f64 (vw6); + size_t sum7 = __riscv_vfmv_f_s_f64m2_f64 (vw7); + size_t sum8 = __riscv_vfmv_f_s_f64m2_f64 (vw8); + size_t sum9 = __riscv_vfmv_f_s_f64m2_f64 (vw9); + size_t sum10 = __riscv_vfmv_f_s_f64m2_f64 (vw10); + size_t sum11 = __riscv_vfmv_f_s_f64m2_f64 (vw11); + size_t sum12 = __riscv_vfmv_f_s_f64m2_f64 (vw12); + size_t sum13 = __riscv_vfmv_f_s_f64m2_f64 (vw13); + size_t sum14 = __riscv_vfmv_f_s_f64m2_f64 (vw14); + size_t sum15 = __riscv_vfmv_f_s_f64m2_f64 (vw15); + + sum += sumation (sum0, sum1, sum2, sum3, sum4, sum5, sum6, sum7, sum8, + sum9, sum10, sum11, sum12, sum13, sum14, sum15); + } + return sum; +} + +/* { dg-final { scan-assembler-not {vmv1r} } } */ +/* { dg-final { scan-assembler-not {vmv2r} } } */ +/* { dg-final { scan-assembler-not {vmv4r} } } */ +/* { dg-final { scan-assembler-not {vmv8r} } } */ +/* { dg-final { scan-assembler-not {csrr} { xfail riscv*-*-* } } } */ diff --git a/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-26.c b/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-26.c new file mode 100644 index 00000000000..0f05e2b2e1d --- /dev/null +++ b/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-26.c @@ -0,0 +1,68 @@ +/* { dg-do compile } */ +/* { dg-options "-march=rv64gcv -mabi=lp64d -O3" } */ + +#include "riscv_vector.h" + +size_t __attribute__ ((noinline)) +sumation (size_t sum0, size_t sum1, size_t sum2, size_t sum3, size_t sum4, + size_t sum5, size_t sum6, size_t sum7) +{ + return sum0 + sum1 + sum2 + sum3 + sum4 + sum5 + sum6 + sum7; +} + +size_t +foo (char const *buf, size_t len) +{ + size_t sum = 0; + size_t vl = __riscv_vsetvlmax_e8m8 (); + size_t step = vl * 4; + const char *it = buf, *end = buf + len; + for (; it + step <= end;) + { + vfloat32m2_t v0 = __riscv_vle32_v_f32m2 ((void *) it, vl); + it += vl; + vfloat32m2_t v1 = __riscv_vle32_v_f32m2 ((void *) it, vl); + it += vl; + vfloat32m2_t v2 = __riscv_vle32_v_f32m2 ((void *) it, vl); + it += vl; + vfloat32m2_t v3 = __riscv_vle32_v_f32m2 ((void *) it, vl); + it += vl; + vfloat32m2_t v4 = __riscv_vle32_v_f32m2 ((void *) it, vl); + it += vl; + vfloat32m2_t v5 = __riscv_vle32_v_f32m2 ((void *) it, vl); + it += vl; + vfloat32m2_t v6 = __riscv_vle32_v_f32m2 ((void *) it, vl); + it += vl; + vfloat32m2_t v7 = __riscv_vle32_v_f32m2 ((void *) it, vl); + it += vl; + + asm volatile("nop" ::: "memory"); + vfloat64m4_t vw0 = __riscv_vfwadd_vf_f64m4 (v0, 33, vl); + vfloat64m4_t vw1 = __riscv_vfwadd_vf_f64m4 (v1, 33, vl); + vfloat64m4_t vw2 = __riscv_vfwadd_vf_f64m4 (v2, 33, vl); + vfloat64m4_t vw3 = __riscv_vfwadd_vf_f64m4 (v3, 33, vl); + vfloat64m4_t vw4 = __riscv_vfwadd_vf_f64m4 (v4, 33, vl); + vfloat64m4_t vw5 = __riscv_vfwadd_vf_f64m4 (v5, 33, vl); + vfloat64m4_t vw6 = __riscv_vfwadd_vf_f64m4 (v6, 33, vl); + vfloat64m4_t vw7 = __riscv_vfwadd_vf_f64m4 (v7, 33, vl); + + asm volatile("nop" ::: "memory"); + size_t sum0 = __riscv_vfmv_f_s_f64m4_f64 (vw0); + size_t sum1 = __riscv_vfmv_f_s_f64m4_f64 (vw1); + size_t sum2 = __riscv_vfmv_f_s_f64m4_f64 (vw2); + size_t sum3 = __riscv_vfmv_f_s_f64m4_f64 (vw3); + size_t sum4 = __riscv_vfmv_f_s_f64m4_f64 (vw4); + size_t sum5 = __riscv_vfmv_f_s_f64m4_f64 (vw5); + size_t sum6 = __riscv_vfmv_f_s_f64m4_f64 (vw6); + size_t sum7 = __riscv_vfmv_f_s_f64m4_f64 (vw7); + + sum += sumation (sum0, sum1, sum2, sum3, sum4, sum5, sum6, sum7); + } + return sum; +} + +/* { dg-final { scan-assembler-not {vmv1r} } } */ +/* { dg-final { scan-assembler-not {vmv2r} } } */ +/* { dg-final { scan-assembler-not {vmv4r} } } */ +/* { dg-final { scan-assembler-not {vmv8r} } } */ +/* { dg-final { scan-assembler-not {csrr} { xfail riscv*-*-* } } } */ diff --git a/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-27.c b/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-27.c new file mode 100644 index 00000000000..d640bcf74ba --- /dev/null +++ b/gcc/testsuite/gcc.target/riscv/rvv/base/pr112431-27.c @@ -0,0 +1,51 @@ +/* { dg-do compile } */ +/* { dg-options "-march=rv64gcv -mabi=lp64d -O3" } */ + +#include "riscv_vector.h" + +size_t __attribute__ ((noinline)) +sumation (size_t sum0, size_t sum1, size_t sum2, size_t sum3) +{ + return sum0 + sum1 + sum2 + sum3; +} + +size_t +foo (char const *buf, size_t len) +{ + size_t sum = 0; + size_t vl = __riscv_vsetvlmax_e8m8 (); + size_t step = vl * 4; + const char *it = buf, *end = buf + len; + for (; it + step <= end;) + { + vfloat32m4_t v0 = __riscv_vle32_v_f32m4 ((void *) it, vl); + it += vl; + vfloat32m4_t v1 = __riscv_vle32_v_f32m4 ((void *) it, vl); + it += vl; + vfloat32m4_t v2 = __riscv_vle32_v_f32m4 ((void *) it, vl); + it += vl; + vfloat32m4_t v3 = __riscv_vle32_v_f32m4 ((void *) it, vl); + it += vl; + + asm volatile("nop" ::: "memory"); + vfloat64m8_t vw0 = __riscv_vfwadd_vf_f64m8 (v0, 33, vl); + vfloat64m8_t vw1 = __riscv_vfwadd_vf_f64m8 (v1, 33, vl); + vfloat64m8_t vw2 = __riscv_vfwadd_vf_f64m8 (v2, 33, vl); + vfloat64m8_t vw3 = __riscv_vfwadd_vf_f64m8 (v3, 33, vl); + + asm volatile("nop" ::: "memory"); + size_t sum0 = __riscv_vfmv_f_s_f64m8_f64 (vw0); + size_t sum1 = __riscv_vfmv_f_s_f64m8_f64 (vw1); + size_t sum2 = __riscv_vfmv_f_s_f64m8_f64 (vw2); + size_t sum3 = __riscv_vfmv_f_s_f64m8_f64 (vw3); + + sum += sumation (sum0, sum1, sum2, sum3); + } + return sum; +} + +/* { dg-final { scan-assembler-not {vmv1r} } } */ +/* { dg-final { scan-assembler-not {vmv2r} } } */ +/* { dg-final { scan-assembler-not {vmv4r} } } */ +/* { dg-final { scan-assembler-not {vmv8r} } } */ +/* { dg-final { scan-assembler-not {csrr} { xfail riscv*-*-* } } } */