aarch64: Fix costing of manual bfi instructions
Checks
Context |
Check |
Description |
linaro-tcwg-bot/tcwg_gcc_build--master-arm |
success
|
Testing passed
|
linaro-tcwg-bot/tcwg_gcc_build--master-aarch64 |
success
|
Testing passed
|
linaro-tcwg-bot/tcwg_gcc_check--master-aarch64 |
success
|
Testing passed
|
linaro-tcwg-bot/tcwg_gcc_check--master-arm |
success
|
Testing passed
|
Commit Message
This fixes the cost model for BFI instructions which don't
use directly zero_extract on the LHS.
aarch64_bfi_rtx_p does the heavy lifting by matching of
the patterns.
Note this alone does not fix PR 107270, it is a step in the right
direction. There we get z zero_extend for the non-shifted part
which we don't currently match.
Built and tested on aarch64-linux-gnu with no regressions.
gcc/ChangeLog:
* config/aarch64/aarch64.cc (aarch64_bfi_rtx_p): New function.
(aarch64_rtx_costs): For IOR, try calling aarch64_bfi_rtx_p.
Signed-off-by: Andrew Pinski <quic_apinski@quicinc.com>
---
gcc/config/aarch64/aarch64.cc | 94 +++++++++++++++++++++++++++++++++++
1 file changed, 94 insertions(+)
Comments
Andrew Pinski <quic_apinski@quicinc.com> writes:
> This fixes the cost model for BFI instructions which don't
> use directly zero_extract on the LHS.
> aarch64_bfi_rtx_p does the heavy lifting by matching of
> the patterns.
>
> Note this alone does not fix PR 107270, it is a step in the right
> direction. There we get z zero_extend for the non-shifted part
> which we don't currently match.
>
> Built and tested on aarch64-linux-gnu with no regressions.
>
> gcc/ChangeLog:
>
> * config/aarch64/aarch64.cc (aarch64_bfi_rtx_p): New function.
> (aarch64_rtx_costs): For IOR, try calling aarch64_bfi_rtx_p.
>
> Signed-off-by: Andrew Pinski <quic_apinski@quicinc.com>
> ---
> gcc/config/aarch64/aarch64.cc | 94 +++++++++++++++++++++++++++++++++++
> 1 file changed, 94 insertions(+)
>
> diff --git a/gcc/config/aarch64/aarch64.cc b/gcc/config/aarch64/aarch64.cc
> index 3d8341c17fe..dc5c5c23cb3 100644
> --- a/gcc/config/aarch64/aarch64.cc
> +++ b/gcc/config/aarch64/aarch64.cc
> @@ -13776,6 +13776,90 @@ aarch64_extr_rtx_p (rtx x, rtx *res_op0, rtx *res_op1)
> return false;
> }
>
> +/* Return true iff X is an rtx that will match an bfi instruction
> + i.e. as described in the *aarch64_bfi<mode>5 family of patterns.
> + OP0 and OP1 will be set to the operands of the insert involved
> + on success and will be NULL_RTX otherwise. */
> +
> +static bool
> +aarch64_bfi_rtx_p (rtx x, rtx *res_op0, rtx *res_op1)
I think it'd be slightly neater to pass an XEXP index in as well,
and use it...
> +{
> + rtx op0, op1;
> + scalar_int_mode mode;
> +
> + *res_op0 = NULL_RTX;
> + *res_op1 = NULL_RTX;
> + if (!is_a <scalar_int_mode> (GET_MODE (x), &mode))
> + return false;
> +
> + if (GET_CODE (x) != IOR)
> + return false;
> +
> + unsigned HOST_WIDE_INT mask1;
> + unsigned HOST_WIDE_INT shft_amnt;
> + unsigned HOST_WIDE_INT mask2;
> + rtx shiftop;
> +
> + rtx iop0 = XEXP (x, 0);
> + rtx iop1 = XEXP (x, 1);
...here as opno and 1 - opno. That way we don't need to...
> +
> + if (GET_CODE (iop0) == AND
> + && CONST_INT_P (XEXP (iop0, 1))
> + && GET_CODE (XEXP (iop0, 0)) != ASHIFT)
> + {
> + op0 = XEXP (iop0, 0);
> + mask1 = UINTVAL (XEXP (iop0, 1));
> + shiftop = iop1;
> + }
> + else if (GET_CODE (iop1) == AND
> + && CONST_INT_P (XEXP (iop1, 1))
> + && GET_CODE (XEXP (iop1, 0)) != ASHIFT)
> + {
> + op0 = XEXP (iop1, 0);
> + mask1 = UINTVAL (XEXP (iop1, 1));
> + shiftop = iop0;
> + }
> + else
> + return false;
...handle this both ways, and don't need to exclude ASHIFT.
Maybe some variation on "insert_op" would be better than "shiftop",
since the operand might not include a shift.
Looks generally good to me otherwise FWIW, but obviously GCC 15 material.
Thanks,
Richard
> +
> + /* Shifted with no mask. */
> + if (GET_CODE (shiftop) == ASHIFT
> + && CONST_INT_P (XEXP (shiftop, 1)))
> + {
> + shft_amnt = UINTVAL (XEXP (shiftop, 1));
> + mask2 = HOST_WIDE_INT_M1U << shft_amnt;
> + op1 = XEXP (shiftop, 0);
> + }
> + else if (GET_CODE (shiftop) == AND
> + && CONST_INT_P (XEXP (shiftop, 1)))
> + {
> + mask2 = UINTVAL (XEXP (shiftop, 1));
> + if (GET_CODE (XEXP (shiftop, 0)) == ASHIFT
> + && CONST_INT_P (XEXP (XEXP (shiftop, 0), 1)))
> + {
> + op1 = XEXP (XEXP (shiftop, 0), 0);
> + shft_amnt = UINTVAL (XEXP (XEXP (shiftop, 0), 1));
> + }
> + else
> + {
> + op1 = XEXP (shiftop, 0);
> + shft_amnt = 0;
> + }
> + }
> + else
> + return false;
> +
> + if (shft_amnt >= GET_MODE_BITSIZE (mode))
> + return false;
> +
> + if (!aarch64_masks_and_shift_for_bfi_p (mode, mask1, shft_amnt, mask2))
> + return false;
> +
> + *res_op0 = op0;
> + *res_op1 = op1;
> + return true;
> +}
> +
> /* Calculate the cost of calculating (if_then_else (OP0) (OP1) (OP2)),
> storing it in *COST. Result is true if the total cost of the operation
> has now been calculated. */
> @@ -14662,6 +14746,16 @@ cost_plus:
> return true;
> }
>
> + if (aarch64_bfi_rtx_p (x, &op0, &op1))
> + {
> + *cost += rtx_cost (op0, mode, IOR, 0, speed);
> + *cost += rtx_cost (op0, mode, IOR, 1, speed);
> + if (speed)
> + *cost += extra_cost->alu.bfi;
> +
> + return true;
> + }
> +
> if (aarch64_extr_rtx_p (x, &op0, &op1))
> {
> *cost += rtx_cost (op0, mode, IOR, 0, speed);
@@ -13776,6 +13776,90 @@ aarch64_extr_rtx_p (rtx x, rtx *res_op0, rtx *res_op1)
return false;
}
+/* Return true iff X is an rtx that will match an bfi instruction
+ i.e. as described in the *aarch64_bfi<mode>5 family of patterns.
+ OP0 and OP1 will be set to the operands of the insert involved
+ on success and will be NULL_RTX otherwise. */
+
+static bool
+aarch64_bfi_rtx_p (rtx x, rtx *res_op0, rtx *res_op1)
+{
+ rtx op0, op1;
+ scalar_int_mode mode;
+
+ *res_op0 = NULL_RTX;
+ *res_op1 = NULL_RTX;
+ if (!is_a <scalar_int_mode> (GET_MODE (x), &mode))
+ return false;
+
+ if (GET_CODE (x) != IOR)
+ return false;
+
+ unsigned HOST_WIDE_INT mask1;
+ unsigned HOST_WIDE_INT shft_amnt;
+ unsigned HOST_WIDE_INT mask2;
+ rtx shiftop;
+
+ rtx iop0 = XEXP (x, 0);
+ rtx iop1 = XEXP (x, 1);
+
+ if (GET_CODE (iop0) == AND
+ && CONST_INT_P (XEXP (iop0, 1))
+ && GET_CODE (XEXP (iop0, 0)) != ASHIFT)
+ {
+ op0 = XEXP (iop0, 0);
+ mask1 = UINTVAL (XEXP (iop0, 1));
+ shiftop = iop1;
+ }
+ else if (GET_CODE (iop1) == AND
+ && CONST_INT_P (XEXP (iop1, 1))
+ && GET_CODE (XEXP (iop1, 0)) != ASHIFT)
+ {
+ op0 = XEXP (iop1, 0);
+ mask1 = UINTVAL (XEXP (iop1, 1));
+ shiftop = iop0;
+ }
+ else
+ return false;
+
+ /* Shifted with no mask. */
+ if (GET_CODE (shiftop) == ASHIFT
+ && CONST_INT_P (XEXP (shiftop, 1)))
+ {
+ shft_amnt = UINTVAL (XEXP (shiftop, 1));
+ mask2 = HOST_WIDE_INT_M1U << shft_amnt;
+ op1 = XEXP (shiftop, 0);
+ }
+ else if (GET_CODE (shiftop) == AND
+ && CONST_INT_P (XEXP (shiftop, 1)))
+ {
+ mask2 = UINTVAL (XEXP (shiftop, 1));
+ if (GET_CODE (XEXP (shiftop, 0)) == ASHIFT
+ && CONST_INT_P (XEXP (XEXP (shiftop, 0), 1)))
+ {
+ op1 = XEXP (XEXP (shiftop, 0), 0);
+ shft_amnt = UINTVAL (XEXP (XEXP (shiftop, 0), 1));
+ }
+ else
+ {
+ op1 = XEXP (shiftop, 0);
+ shft_amnt = 0;
+ }
+ }
+ else
+ return false;
+
+ if (shft_amnt >= GET_MODE_BITSIZE (mode))
+ return false;
+
+ if (!aarch64_masks_and_shift_for_bfi_p (mode, mask1, shft_amnt, mask2))
+ return false;
+
+ *res_op0 = op0;
+ *res_op1 = op1;
+ return true;
+}
+
/* Calculate the cost of calculating (if_then_else (OP0) (OP1) (OP2)),
storing it in *COST. Result is true if the total cost of the operation
has now been calculated. */
@@ -14662,6 +14746,16 @@ cost_plus:
return true;
}
+ if (aarch64_bfi_rtx_p (x, &op0, &op1))
+ {
+ *cost += rtx_cost (op0, mode, IOR, 0, speed);
+ *cost += rtx_cost (op0, mode, IOR, 1, speed);
+ if (speed)
+ *cost += extra_cost->alu.bfi;
+
+ return true;
+ }
+
if (aarch64_extr_rtx_p (x, &op0, &op1))
{
*cost += rtx_cost (op0, mode, IOR, 0, speed);