[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[PATCH v2 40/48] tcg/optimize: Expand fold_addsub2_i32 to 64-bit ops
From: |
Richard Henderson |
Subject: |
[PATCH v2 40/48] tcg/optimize: Expand fold_addsub2_i32 to 64-bit ops |
Date: |
Thu, 7 Oct 2021 12:54:48 -0700 |
Rename to fold_addsub2.
Use Int128 to implement the wider operation.
Reviewed-by: Philippe Mathieu-Daudé <f4bug@amsat.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
tcg/optimize.c | 64 +++++++++++++++++++++++++++++++++-----------------
1 file changed, 43 insertions(+), 21 deletions(-)
diff --git a/tcg/optimize.c b/tcg/optimize.c
index 0011ac31ec..5e662ad8f7 100644
--- a/tcg/optimize.c
+++ b/tcg/optimize.c
@@ -838,37 +838,59 @@ static bool fold_add(OptContext *ctx, TCGOp *op)
return false;
}
-static bool fold_addsub2_i32(OptContext *ctx, TCGOp *op, bool add)
+static bool fold_addsub2(OptContext *ctx, TCGOp *op, bool add)
{
if (arg_is_const(op->args[2]) && arg_is_const(op->args[3]) &&
arg_is_const(op->args[4]) && arg_is_const(op->args[5])) {
- uint32_t al = arg_info(op->args[2])->val;
- uint32_t ah = arg_info(op->args[3])->val;
- uint32_t bl = arg_info(op->args[4])->val;
- uint32_t bh = arg_info(op->args[5])->val;
- uint64_t a = ((uint64_t)ah << 32) | al;
- uint64_t b = ((uint64_t)bh << 32) | bl;
+ uint64_t al = arg_info(op->args[2])->val;
+ uint64_t ah = arg_info(op->args[3])->val;
+ uint64_t bl = arg_info(op->args[4])->val;
+ uint64_t bh = arg_info(op->args[5])->val;
TCGArg rl, rh;
- TCGOp *op2 = tcg_op_insert_before(ctx->tcg, op, INDEX_op_mov_i32);
+ TCGOp *op2;
- if (add) {
- a += b;
+ if (ctx->type == TCG_TYPE_I32) {
+ uint64_t a = deposit64(al, 32, 32, ah);
+ uint64_t b = deposit64(bl, 32, 32, bh);
+
+ if (add) {
+ a += b;
+ } else {
+ a -= b;
+ }
+
+ al = sextract64(a, 0, 32);
+ ah = sextract64(a, 32, 32);
} else {
- a -= b;
+ Int128 a = int128_make128(al, ah);
+ Int128 b = int128_make128(bl, bh);
+
+ if (add) {
+ a = int128_add(a, b);
+ } else {
+ a = int128_sub(a, b);
+ }
+
+ al = int128_getlo(a);
+ ah = int128_gethi(a);
}
rl = op->args[0];
rh = op->args[1];
- tcg_opt_gen_movi(ctx, op, rl, (int32_t)a);
- tcg_opt_gen_movi(ctx, op2, rh, (int32_t)(a >> 32));
+
+ /* The proper opcode is supplied by tcg_opt_gen_mov. */
+ op2 = tcg_op_insert_before(ctx->tcg, op, 0);
+
+ tcg_opt_gen_movi(ctx, op, rl, al);
+ tcg_opt_gen_movi(ctx, op2, rh, ah);
return true;
}
return false;
}
-static bool fold_add2_i32(OptContext *ctx, TCGOp *op)
+static bool fold_add2(OptContext *ctx, TCGOp *op)
{
- return fold_addsub2_i32(ctx, op, true);
+ return fold_addsub2(ctx, op, true);
}
static bool fold_and(OptContext *ctx, TCGOp *op)
@@ -1707,9 +1729,9 @@ static bool fold_sub(OptContext *ctx, TCGOp *op)
return false;
}
-static bool fold_sub2_i32(OptContext *ctx, TCGOp *op)
+static bool fold_sub2(OptContext *ctx, TCGOp *op)
{
- return fold_addsub2_i32(ctx, op, false);
+ return fold_addsub2(ctx, op, false);
}
static bool fold_tcg_ld(OptContext *ctx, TCGOp *op)
@@ -1855,8 +1877,8 @@ void tcg_optimize(TCGContext *s)
CASE_OP_32_64_VEC(add):
done = fold_add(&ctx, op);
break;
- case INDEX_op_add2_i32:
- done = fold_add2_i32(&ctx, op);
+ CASE_OP_32_64(add2):
+ done = fold_add2(&ctx, op);
break;
CASE_OP_32_64_VEC(and):
done = fold_and(&ctx, op);
@@ -1991,8 +2013,8 @@ void tcg_optimize(TCGContext *s)
CASE_OP_32_64_VEC(sub):
done = fold_sub(&ctx, op);
break;
- case INDEX_op_sub2_i32:
- done = fold_sub2_i32(&ctx, op);
+ CASE_OP_32_64(sub2):
+ done = fold_sub2(&ctx, op);
break;
CASE_OP_32_64_VEC(xor):
done = fold_xor(&ctx, op);
--
2.25.1
- [PATCH v2 29/48] tcg/optimize: Split out fold_mov, (continued)
- [PATCH v2 29/48] tcg/optimize: Split out fold_mov, Richard Henderson, 2021/10/07
- [PATCH v2 20/48] tcg/optimize: Split out fold_mulu2_i32, Richard Henderson, 2021/10/07
- [PATCH v2 36/48] tcg/optimize: Split out fold_xi_to_x, Richard Henderson, 2021/10/07
- [PATCH v2 37/48] tcg/optimize: Split out fold_ix_to_i, Richard Henderson, 2021/10/07
- [PATCH v2 38/48] tcg/optimize: Split out fold_masks, Richard Henderson, 2021/10/07
- [PATCH v2 31/48] tcg/optimize: Split out fold_xx_to_x, Richard Henderson, 2021/10/07
- [PATCH v2 13/48] tcg/optimize: Use a boolean to avoid a mass of continues, Richard Henderson, 2021/10/07
- [PATCH v2 42/48] tcg/optimize: Add more simplifications for orc, Richard Henderson, 2021/10/07
- [PATCH v2 43/48] tcg/optimize: Stop forcing z_mask to "garbage" for 32-bit values, Richard Henderson, 2021/10/07
- [PATCH v2 40/48] tcg/optimize: Expand fold_addsub2_i32 to 64-bit ops,
Richard Henderson <=
- [PATCH v2 41/48] tcg/optimize: Sink commutative operand swapping into fold functions, Richard Henderson, 2021/10/07
- [PATCH v2 45/48] tcg/optimize: Propagate sign info for logical operations, Richard Henderson, 2021/10/07
- [PATCH v2 44/48] tcg/optimize: Optimize sign extensions, Richard Henderson, 2021/10/07
- [PATCH v2 46/48] tcg/optimize: Propagate sign info for setcond, Richard Henderson, 2021/10/07
- [PATCH v2 47/48] tcg/optimize: Propagate sign info for bit counting, Richard Henderson, 2021/10/07
- [PATCH v2 48/48] tcg/optimize: Propagate sign info for shifting, Richard Henderson, 2021/10/07
- Re: [PATCH v2 00/48] tcg: optimize redundant sign extensions, Richard Henderson, 2021/10/15
- Re: [PATCH v2 00/48] tcg: optimize redundant sign extensions, Alex Bennée, 2021/10/20