[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [PATCH v2 29/93] tcg/tci: Remove TCG_CONST
From: |
Alex Bennée |
Subject: |
Re: [PATCH v2 29/93] tcg/tci: Remove TCG_CONST |
Date: |
Thu, 04 Feb 2021 15:39:22 +0000 |
User-agent: |
mu4e 1.5.7; emacs 28.0.50 |
Richard Henderson <richard.henderson@linaro.org> writes:
> Only allow registers or constants, but not both, in any
> given position.
Aren't we switching to all registers (there are no more _i functions
after this)? I guess you mean the registers can have constants in them?
> Removing this difference in input will
> allow more code to be shared between 32-bit and 64-bit.
>
> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
> ---
> tcg/tci/tcg-target-con-set.h | 6 +-
> tcg/tci/tcg-target.h | 3 -
> tcg/tci.c | 189 +++++++++++++----------------------
> tcg/tci/tcg-target.c.inc | 82 ++++-----------
> 4 files changed, 89 insertions(+), 191 deletions(-)
>
> diff --git a/tcg/tci/tcg-target-con-set.h b/tcg/tci/tcg-target-con-set.h
> index 38e82f7535..f51b7bcb13 100644
> --- a/tcg/tci/tcg-target-con-set.h
> +++ b/tcg/tci/tcg-target-con-set.h
> @@ -10,16 +10,12 @@
> * tcg-target-con-str.h; the constraint combination is inclusive or.
> */
> C_O0_I2(r, r)
> -C_O0_I2(r, ri)
> C_O0_I3(r, r, r)
> -C_O0_I4(r, r, ri, ri)
> C_O0_I4(r, r, r, r)
> C_O1_I1(r, r)
> C_O1_I2(r, 0, r)
> -C_O1_I2(r, ri, ri)
> C_O1_I2(r, r, r)
> -C_O1_I2(r, r, ri)
> -C_O1_I4(r, r, r, ri, ri)
> +C_O1_I4(r, r, r, r, r)
> C_O2_I1(r, r, r)
> C_O2_I2(r, r, r, r)
> C_O2_I4(r, r, r, r, r, r)
> diff --git a/tcg/tci/tcg-target.h b/tcg/tci/tcg-target.h
> index 8f7ed676fc..9c0021a26f 100644
> --- a/tcg/tci/tcg-target.h
> +++ b/tcg/tci/tcg-target.h
> @@ -157,9 +157,6 @@ typedef enum {
>
> TCG_AREG0 = TCG_REG_R14,
> TCG_REG_CALL_STACK = TCG_REG_R15,
> -
> - /* Special value UINT8_MAX is used by TCI to encode constant values. */
> - TCG_CONST = UINT8_MAX
> } TCGReg;
>
> /* Used for function call generation. */
> diff --git a/tcg/tci.c b/tcg/tci.c
> index 935eb87330..fb3c97aaf1 100644
> --- a/tcg/tci.c
> +++ b/tcg/tci.c
> @@ -255,61 +255,6 @@ tci_read_ulong(const tcg_target_ulong *regs, const
> uint8_t **tb_ptr)
> return taddr;
> }
>
> -/* Read indexed register or constant (native size) from bytecode. */
> -static tcg_target_ulong
> -tci_read_ri(const tcg_target_ulong *regs, const uint8_t **tb_ptr)
> -{
> - tcg_target_ulong value;
> - TCGReg r = **tb_ptr;
> - *tb_ptr += 1;
> - if (r == TCG_CONST) {
> - value = tci_read_i(tb_ptr);
> - } else {
> - value = tci_read_reg(regs, r);
> - }
> - return value;
> -}
> -
> -/* Read indexed register or constant (32 bit) from bytecode. */
> -static uint32_t tci_read_ri32(const tcg_target_ulong *regs,
> - const uint8_t **tb_ptr)
> -{
> - uint32_t value;
> - TCGReg r = **tb_ptr;
> - *tb_ptr += 1;
> - if (r == TCG_CONST) {
> - value = tci_read_i32(tb_ptr);
> - } else {
> - value = tci_read_reg32(regs, r);
> - }
> - return value;
> -}
> -
> -#if TCG_TARGET_REG_BITS == 32
> -/* Read two indexed registers or constants (2 * 32 bit) from bytecode. */
> -static uint64_t tci_read_ri64(const tcg_target_ulong *regs,
> - const uint8_t **tb_ptr)
> -{
> - uint32_t low = tci_read_ri32(regs, tb_ptr);
> - return tci_uint64(tci_read_ri32(regs, tb_ptr), low);
> -}
> -#elif TCG_TARGET_REG_BITS == 64
> -/* Read indexed register or constant (64 bit) from bytecode. */
> -static uint64_t tci_read_ri64(const tcg_target_ulong *regs,
> - const uint8_t **tb_ptr)
> -{
> - uint64_t value;
> - TCGReg r = **tb_ptr;
> - *tb_ptr += 1;
> - if (r == TCG_CONST) {
> - value = tci_read_i64(tb_ptr);
> - } else {
> - value = tci_read_reg64(regs, r);
> - }
> - return value;
> -}
> -#endif
> -
> static tcg_target_ulong tci_read_label(const uint8_t **tb_ptr)
> {
> tcg_target_ulong label = tci_read_i(tb_ptr);
> @@ -504,7 +449,7 @@ uintptr_t QEMU_DISABLE_CFI tcg_qemu_tb_exec(CPUArchState
> *env,
>
> switch (opc) {
> case INDEX_op_call:
> - t0 = tci_read_ri(regs, &tb_ptr);
> + t0 = tci_read_i(&tb_ptr);
> tci_tb_ptr = (uintptr_t)tb_ptr;
> #if TCG_TARGET_REG_BITS == 32
> tmp64 = ((helper_function)t0)(tci_read_reg(regs, TCG_REG_R0),
> @@ -539,7 +484,7 @@ uintptr_t QEMU_DISABLE_CFI tcg_qemu_tb_exec(CPUArchState
> *env,
> case INDEX_op_setcond_i32:
> t0 = *tb_ptr++;
> t1 = tci_read_r32(regs, &tb_ptr);
> - t2 = tci_read_ri32(regs, &tb_ptr);
> + t2 = tci_read_r32(regs, &tb_ptr);
> condition = *tb_ptr++;
> tci_write_reg(regs, t0, tci_compare32(t1, t2, condition));
> break;
> @@ -547,7 +492,7 @@ uintptr_t QEMU_DISABLE_CFI tcg_qemu_tb_exec(CPUArchState
> *env,
> case INDEX_op_setcond2_i32:
> t0 = *tb_ptr++;
> tmp64 = tci_read_r64(regs, &tb_ptr);
> - v64 = tci_read_ri64(regs, &tb_ptr);
> + v64 = tci_read_r64(regs, &tb_ptr);
> condition = *tb_ptr++;
> tci_write_reg(regs, t0, tci_compare64(tmp64, v64, condition));
> break;
> @@ -555,7 +500,7 @@ uintptr_t QEMU_DISABLE_CFI tcg_qemu_tb_exec(CPUArchState
> *env,
> case INDEX_op_setcond_i64:
> t0 = *tb_ptr++;
> t1 = tci_read_r64(regs, &tb_ptr);
> - t2 = tci_read_ri64(regs, &tb_ptr);
> + t2 = tci_read_r64(regs, &tb_ptr);
> condition = *tb_ptr++;
> tci_write_reg(regs, t0, tci_compare64(t1, t2, condition));
> break;
> @@ -628,62 +573,62 @@ uintptr_t QEMU_DISABLE_CFI
> tcg_qemu_tb_exec(CPUArchState *env,
>
> case INDEX_op_add_i32:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri32(regs, &tb_ptr);
> - t2 = tci_read_ri32(regs, &tb_ptr);
> + t1 = tci_read_r32(regs, &tb_ptr);
> + t2 = tci_read_r32(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 + t2);
> break;
> case INDEX_op_sub_i32:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri32(regs, &tb_ptr);
> - t2 = tci_read_ri32(regs, &tb_ptr);
> + t1 = tci_read_r32(regs, &tb_ptr);
> + t2 = tci_read_r32(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 - t2);
> break;
> case INDEX_op_mul_i32:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri32(regs, &tb_ptr);
> - t2 = tci_read_ri32(regs, &tb_ptr);
> + t1 = tci_read_r32(regs, &tb_ptr);
> + t2 = tci_read_r32(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 * t2);
> break;
> case INDEX_op_div_i32:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri32(regs, &tb_ptr);
> - t2 = tci_read_ri32(regs, &tb_ptr);
> + t1 = tci_read_r32(regs, &tb_ptr);
> + t2 = tci_read_r32(regs, &tb_ptr);
> tci_write_reg(regs, t0, (int32_t)t1 / (int32_t)t2);
> break;
> case INDEX_op_divu_i32:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri32(regs, &tb_ptr);
> - t2 = tci_read_ri32(regs, &tb_ptr);
> + t1 = tci_read_r32(regs, &tb_ptr);
> + t2 = tci_read_r32(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 / t2);
> break;
> case INDEX_op_rem_i32:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri32(regs, &tb_ptr);
> - t2 = tci_read_ri32(regs, &tb_ptr);
> + t1 = tci_read_r32(regs, &tb_ptr);
> + t2 = tci_read_r32(regs, &tb_ptr);
> tci_write_reg(regs, t0, (int32_t)t1 % (int32_t)t2);
> break;
> case INDEX_op_remu_i32:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri32(regs, &tb_ptr);
> - t2 = tci_read_ri32(regs, &tb_ptr);
> + t1 = tci_read_r32(regs, &tb_ptr);
> + t2 = tci_read_r32(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 % t2);
> break;
> case INDEX_op_and_i32:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri32(regs, &tb_ptr);
> - t2 = tci_read_ri32(regs, &tb_ptr);
> + t1 = tci_read_r32(regs, &tb_ptr);
> + t2 = tci_read_r32(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 & t2);
> break;
> case INDEX_op_or_i32:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri32(regs, &tb_ptr);
> - t2 = tci_read_ri32(regs, &tb_ptr);
> + t1 = tci_read_r32(regs, &tb_ptr);
> + t2 = tci_read_r32(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 | t2);
> break;
> case INDEX_op_xor_i32:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri32(regs, &tb_ptr);
> - t2 = tci_read_ri32(regs, &tb_ptr);
> + t1 = tci_read_r32(regs, &tb_ptr);
> + t2 = tci_read_r32(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 ^ t2);
> break;
>
> @@ -691,33 +636,33 @@ uintptr_t QEMU_DISABLE_CFI
> tcg_qemu_tb_exec(CPUArchState *env,
>
> case INDEX_op_shl_i32:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri32(regs, &tb_ptr);
> - t2 = tci_read_ri32(regs, &tb_ptr);
> + t1 = tci_read_r32(regs, &tb_ptr);
> + t2 = tci_read_r32(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 << (t2 & 31));
> break;
> case INDEX_op_shr_i32:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri32(regs, &tb_ptr);
> - t2 = tci_read_ri32(regs, &tb_ptr);
> + t1 = tci_read_r32(regs, &tb_ptr);
> + t2 = tci_read_r32(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 >> (t2 & 31));
> break;
> case INDEX_op_sar_i32:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri32(regs, &tb_ptr);
> - t2 = tci_read_ri32(regs, &tb_ptr);
> + t1 = tci_read_r32(regs, &tb_ptr);
> + t2 = tci_read_r32(regs, &tb_ptr);
> tci_write_reg(regs, t0, ((int32_t)t1 >> (t2 & 31)));
> break;
> #if TCG_TARGET_HAS_rot_i32
> case INDEX_op_rotl_i32:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri32(regs, &tb_ptr);
> - t2 = tci_read_ri32(regs, &tb_ptr);
> + t1 = tci_read_r32(regs, &tb_ptr);
> + t2 = tci_read_r32(regs, &tb_ptr);
> tci_write_reg(regs, t0, rol32(t1, t2 & 31));
> break;
> case INDEX_op_rotr_i32:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri32(regs, &tb_ptr);
> - t2 = tci_read_ri32(regs, &tb_ptr);
> + t1 = tci_read_r32(regs, &tb_ptr);
> + t2 = tci_read_r32(regs, &tb_ptr);
> tci_write_reg(regs, t0, ror32(t1, t2 & 31));
> break;
> #endif
> @@ -734,7 +679,7 @@ uintptr_t QEMU_DISABLE_CFI tcg_qemu_tb_exec(CPUArchState
> *env,
> #endif
> case INDEX_op_brcond_i32:
> t0 = tci_read_r32(regs, &tb_ptr);
> - t1 = tci_read_ri32(regs, &tb_ptr);
> + t1 = tci_read_r32(regs, &tb_ptr);
> condition = *tb_ptr++;
> label = tci_read_label(&tb_ptr);
> if (tci_compare32(t0, t1, condition)) {
> @@ -760,7 +705,7 @@ uintptr_t QEMU_DISABLE_CFI tcg_qemu_tb_exec(CPUArchState
> *env,
> break;
> case INDEX_op_brcond2_i32:
> tmp64 = tci_read_r64(regs, &tb_ptr);
> - v64 = tci_read_ri64(regs, &tb_ptr);
> + v64 = tci_read_r64(regs, &tb_ptr);
> condition = *tb_ptr++;
> label = tci_read_label(&tb_ptr);
> if (tci_compare64(tmp64, v64, condition)) {
> @@ -870,62 +815,62 @@ uintptr_t QEMU_DISABLE_CFI
> tcg_qemu_tb_exec(CPUArchState *env,
>
> case INDEX_op_add_i64:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri64(regs, &tb_ptr);
> - t2 = tci_read_ri64(regs, &tb_ptr);
> + t1 = tci_read_r64(regs, &tb_ptr);
> + t2 = tci_read_r64(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 + t2);
> break;
> case INDEX_op_sub_i64:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri64(regs, &tb_ptr);
> - t2 = tci_read_ri64(regs, &tb_ptr);
> + t1 = tci_read_r64(regs, &tb_ptr);
> + t2 = tci_read_r64(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 - t2);
> break;
> case INDEX_op_mul_i64:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri64(regs, &tb_ptr);
> - t2 = tci_read_ri64(regs, &tb_ptr);
> + t1 = tci_read_r64(regs, &tb_ptr);
> + t2 = tci_read_r64(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 * t2);
> break;
> case INDEX_op_div_i64:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri64(regs, &tb_ptr);
> - t2 = tci_read_ri64(regs, &tb_ptr);
> + t1 = tci_read_r64(regs, &tb_ptr);
> + t2 = tci_read_r64(regs, &tb_ptr);
> tci_write_reg(regs, t0, (int64_t)t1 / (int64_t)t2);
> break;
> case INDEX_op_divu_i64:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri64(regs, &tb_ptr);
> - t2 = tci_read_ri64(regs, &tb_ptr);
> + t1 = tci_read_r64(regs, &tb_ptr);
> + t2 = tci_read_r64(regs, &tb_ptr);
> tci_write_reg(regs, t0, (uint64_t)t1 / (uint64_t)t2);
> break;
> case INDEX_op_rem_i64:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri64(regs, &tb_ptr);
> - t2 = tci_read_ri64(regs, &tb_ptr);
> + t1 = tci_read_r64(regs, &tb_ptr);
> + t2 = tci_read_r64(regs, &tb_ptr);
> tci_write_reg(regs, t0, (int64_t)t1 % (int64_t)t2);
> break;
> case INDEX_op_remu_i64:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri64(regs, &tb_ptr);
> - t2 = tci_read_ri64(regs, &tb_ptr);
> + t1 = tci_read_r64(regs, &tb_ptr);
> + t2 = tci_read_r64(regs, &tb_ptr);
> tci_write_reg(regs, t0, (uint64_t)t1 % (uint64_t)t2);
> break;
> case INDEX_op_and_i64:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri64(regs, &tb_ptr);
> - t2 = tci_read_ri64(regs, &tb_ptr);
> + t1 = tci_read_r64(regs, &tb_ptr);
> + t2 = tci_read_r64(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 & t2);
> break;
> case INDEX_op_or_i64:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri64(regs, &tb_ptr);
> - t2 = tci_read_ri64(regs, &tb_ptr);
> + t1 = tci_read_r64(regs, &tb_ptr);
> + t2 = tci_read_r64(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 | t2);
> break;
> case INDEX_op_xor_i64:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri64(regs, &tb_ptr);
> - t2 = tci_read_ri64(regs, &tb_ptr);
> + t1 = tci_read_r64(regs, &tb_ptr);
> + t2 = tci_read_r64(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 ^ t2);
> break;
>
> @@ -933,33 +878,33 @@ uintptr_t QEMU_DISABLE_CFI
> tcg_qemu_tb_exec(CPUArchState *env,
>
> case INDEX_op_shl_i64:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri64(regs, &tb_ptr);
> - t2 = tci_read_ri64(regs, &tb_ptr);
> + t1 = tci_read_r64(regs, &tb_ptr);
> + t2 = tci_read_r64(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 << (t2 & 63));
> break;
> case INDEX_op_shr_i64:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri64(regs, &tb_ptr);
> - t2 = tci_read_ri64(regs, &tb_ptr);
> + t1 = tci_read_r64(regs, &tb_ptr);
> + t2 = tci_read_r64(regs, &tb_ptr);
> tci_write_reg(regs, t0, t1 >> (t2 & 63));
> break;
> case INDEX_op_sar_i64:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri64(regs, &tb_ptr);
> - t2 = tci_read_ri64(regs, &tb_ptr);
> + t1 = tci_read_r64(regs, &tb_ptr);
> + t2 = tci_read_r64(regs, &tb_ptr);
> tci_write_reg(regs, t0, ((int64_t)t1 >> (t2 & 63)));
> break;
> #if TCG_TARGET_HAS_rot_i64
> case INDEX_op_rotl_i64:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri64(regs, &tb_ptr);
> - t2 = tci_read_ri64(regs, &tb_ptr);
> + t1 = tci_read_r64(regs, &tb_ptr);
> + t2 = tci_read_r64(regs, &tb_ptr);
> tci_write_reg(regs, t0, rol64(t1, t2 & 63));
> break;
> case INDEX_op_rotr_i64:
> t0 = *tb_ptr++;
> - t1 = tci_read_ri64(regs, &tb_ptr);
> - t2 = tci_read_ri64(regs, &tb_ptr);
> + t1 = tci_read_r64(regs, &tb_ptr);
> + t2 = tci_read_r64(regs, &tb_ptr);
> tci_write_reg(regs, t0, ror64(t1, t2 & 63));
> break;
> #endif
> @@ -976,7 +921,7 @@ uintptr_t QEMU_DISABLE_CFI tcg_qemu_tb_exec(CPUArchState
> *env,
> #endif
> case INDEX_op_brcond_i64:
> t0 = tci_read_r64(regs, &tb_ptr);
> - t1 = tci_read_ri64(regs, &tb_ptr);
> + t1 = tci_read_r64(regs, &tb_ptr);
> condition = *tb_ptr++;
> label = tci_read_label(&tb_ptr);
> if (tci_compare64(t0, t1, condition)) {
> diff --git a/tcg/tci/tcg-target.c.inc b/tcg/tci/tcg-target.c.inc
> index 1b66368c94..feac4659cc 100644
> --- a/tcg/tci/tcg-target.c.inc
> +++ b/tcg/tci/tcg-target.c.inc
> @@ -92,8 +92,6 @@ static TCGConstraintSetIndex tcg_target_op_def(TCGOpcode op)
> case INDEX_op_rem_i64:
> case INDEX_op_remu_i32:
> case INDEX_op_remu_i64:
> - return C_O1_I2(r, r, r);
> -
> case INDEX_op_add_i32:
> case INDEX_op_add_i64:
> case INDEX_op_sub_i32:
> @@ -126,8 +124,9 @@ static TCGConstraintSetIndex tcg_target_op_def(TCGOpcode
> op)
> case INDEX_op_rotl_i64:
> case INDEX_op_rotr_i32:
> case INDEX_op_rotr_i64:
> - /* TODO: Does R, RI, RI result in faster code than R, R, RI? */
> - return C_O1_I2(r, ri, ri);
> + case INDEX_op_setcond_i32:
> + case INDEX_op_setcond_i64:
> + return C_O1_I2(r, r, r);
>
> case INDEX_op_deposit_i32:
> case INDEX_op_deposit_i64:
> @@ -135,11 +134,7 @@ static TCGConstraintSetIndex tcg_target_op_def(TCGOpcode
> op)
>
> case INDEX_op_brcond_i32:
> case INDEX_op_brcond_i64:
> - return C_O0_I2(r, ri);
> -
> - case INDEX_op_setcond_i32:
> - case INDEX_op_setcond_i64:
> - return C_O1_I2(r, r, ri);
> + return C_O0_I2(r, r);
>
> #if TCG_TARGET_REG_BITS == 32
> /* TODO: Support R, R, R, R, RI, RI? Will it be faster? */
> @@ -147,11 +142,11 @@ static TCGConstraintSetIndex
> tcg_target_op_def(TCGOpcode op)
> case INDEX_op_sub2_i32:
> return C_O2_I4(r, r, r, r, r, r);
> case INDEX_op_brcond2_i32:
> - return C_O0_I4(r, r, ri, ri);
> + return C_O0_I4(r, r, r, r);
> case INDEX_op_mulu2_i32:
> return C_O2_I2(r, r, r, r);
> case INDEX_op_setcond2_i32:
> - return C_O1_I4(r, r, r, ri, ri);
> + return C_O1_I4(r, r, r, r, r);
> #endif
>
> case INDEX_op_qemu_ld_i32:
> @@ -294,41 +289,6 @@ static void tcg_out_r(TCGContext *s, TCGArg t0)
> tcg_out8(s, t0);
> }
>
> -/* Write register or constant (native size). */
> -static void tcg_out_ri(TCGContext *s, bool const_arg, TCGArg arg)
> -{
> - if (const_arg) {
> - tcg_out8(s, TCG_CONST);
> - tcg_out_i(s, arg);
> - } else {
> - tcg_out_r(s, arg);
> - }
> -}
> -
> -/* Write register or constant (32 bit). */
> -static void tcg_out_ri32(TCGContext *s, bool const_arg, TCGArg arg)
> -{
> - if (const_arg) {
> - tcg_out8(s, TCG_CONST);
> - tcg_out32(s, arg);
> - } else {
> - tcg_out_r(s, arg);
> - }
> -}
> -
> -#if TCG_TARGET_REG_BITS == 64
> -/* Write register or constant (64 bit). */
> -static void tcg_out_ri64(TCGContext *s, bool const_arg, TCGArg arg)
> -{
> - if (const_arg) {
> - tcg_out8(s, TCG_CONST);
> - tcg_out64(s, arg);
> - } else {
> - tcg_out_r(s, arg);
> - }
> -}
> -#endif
> -
> /* Write label. */
> static void tci_out_label(TCGContext *s, TCGLabel *label)
> {
> @@ -416,7 +376,7 @@ static inline void tcg_out_call(TCGContext *s, const
> tcg_insn_unit *arg)
> {
> uint8_t *old_code_ptr = s->code_ptr;
> tcg_out_op_t(s, INDEX_op_call);
> - tcg_out_ri(s, 1, (uintptr_t)arg);
> + tcg_out_i(s, (uintptr_t)arg);
> old_code_ptr[1] = s->code_ptr - old_code_ptr;
> }
>
> @@ -450,7 +410,7 @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
> const TCGArg *args,
> case INDEX_op_setcond_i32:
> tcg_out_r(s, args[0]);
> tcg_out_r(s, args[1]);
> - tcg_out_ri32(s, const_args[2], args[2]);
> + tcg_out_r(s, args[2]);
> tcg_out8(s, args[3]); /* condition */
> break;
> #if TCG_TARGET_REG_BITS == 32
> @@ -459,15 +419,15 @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
> const TCGArg *args,
> tcg_out_r(s, args[0]);
> tcg_out_r(s, args[1]);
> tcg_out_r(s, args[2]);
> - tcg_out_ri32(s, const_args[3], args[3]);
> - tcg_out_ri32(s, const_args[4], args[4]);
> + tcg_out_r(s, args[3]);
> + tcg_out_r(s, args[4]);
> tcg_out8(s, args[5]); /* condition */
> break;
> #elif TCG_TARGET_REG_BITS == 64
> case INDEX_op_setcond_i64:
> tcg_out_r(s, args[0]);
> tcg_out_r(s, args[1]);
> - tcg_out_ri64(s, const_args[2], args[2]);
> + tcg_out_r(s, args[2]);
> tcg_out8(s, args[3]); /* condition */
> break;
> #endif
> @@ -513,8 +473,8 @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
> const TCGArg *args,
> case INDEX_op_rotl_i32: /* Optional (TCG_TARGET_HAS_rot_i32). */
> case INDEX_op_rotr_i32: /* Optional (TCG_TARGET_HAS_rot_i32). */
> tcg_out_r(s, args[0]);
> - tcg_out_ri32(s, const_args[1], args[1]);
> - tcg_out_ri32(s, const_args[2], args[2]);
> + tcg_out_r(s, args[1]);
> + tcg_out_r(s, args[2]);
> break;
> case INDEX_op_deposit_i32: /* Optional (TCG_TARGET_HAS_deposit_i32). */
> tcg_out_r(s, args[0]);
> @@ -548,8 +508,8 @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
> const TCGArg *args,
> case INDEX_op_rem_i64: /* Optional (TCG_TARGET_HAS_div_i64). */
> case INDEX_op_remu_i64: /* Optional (TCG_TARGET_HAS_div_i64). */
> tcg_out_r(s, args[0]);
> - tcg_out_ri64(s, const_args[1], args[1]);
> - tcg_out_ri64(s, const_args[2], args[2]);
> + tcg_out_r(s, args[1]);
> + tcg_out_r(s, args[2]);
> break;
> case INDEX_op_deposit_i64: /* Optional (TCG_TARGET_HAS_deposit_i64). */
> tcg_out_r(s, args[0]);
> @@ -562,7 +522,7 @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
> const TCGArg *args,
> break;
> case INDEX_op_brcond_i64:
> tcg_out_r(s, args[0]);
> - tcg_out_ri64(s, const_args[1], args[1]);
> + tcg_out_r(s, args[1]);
> tcg_out8(s, args[2]); /* condition */
> tci_out_label(s, arg_label(args[3]));
> break;
> @@ -596,8 +556,8 @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
> const TCGArg *args,
> case INDEX_op_rem_i32: /* Optional (TCG_TARGET_HAS_div_i32). */
> case INDEX_op_remu_i32: /* Optional (TCG_TARGET_HAS_div_i32). */
> tcg_out_r(s, args[0]);
> - tcg_out_ri32(s, const_args[1], args[1]);
> - tcg_out_ri32(s, const_args[2], args[2]);
> + tcg_out_r(s, args[1]);
> + tcg_out_r(s, args[2]);
> break;
> #if TCG_TARGET_REG_BITS == 32
> case INDEX_op_add2_i32:
> @@ -612,8 +572,8 @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
> const TCGArg *args,
> case INDEX_op_brcond2_i32:
> tcg_out_r(s, args[0]);
> tcg_out_r(s, args[1]);
> - tcg_out_ri32(s, const_args[2], args[2]);
> - tcg_out_ri32(s, const_args[3], args[3]);
> + tcg_out_r(s, args[2]);
> + tcg_out_r(s, args[3]);
> tcg_out8(s, args[4]); /* condition */
> tci_out_label(s, arg_label(args[5]));
> break;
> @@ -626,7 +586,7 @@ static void tcg_out_op(TCGContext *s, TCGOpcode opc,
> const TCGArg *args,
> #endif
> case INDEX_op_brcond_i32:
> tcg_out_r(s, args[0]);
> - tcg_out_ri32(s, const_args[1], args[1]);
> + tcg_out_r(s, args[1]);
> tcg_out8(s, args[2]); /* condition */
> tci_out_label(s, arg_label(args[3]));
> break;
--
Alex Bennée
- [PATCH v2 35/93] tcg/tci: Remove tci_read_r32s, (continued)
- [PATCH v2 35/93] tcg/tci: Remove tci_read_r32s, Richard Henderson, 2021/02/03
- [PATCH v2 28/93] tcg/tci: Use bool in tcg_out_ri*, Richard Henderson, 2021/02/03
- [PATCH v2 39/93] tcg/tci: Remove ifdefs for TCG_TARGET_HAS_ext32[us]_i64, Richard Henderson, 2021/02/03
- [PATCH v2 41/93] tcg/tci: Merge mov, not and neg operations, Richard Henderson, 2021/02/03
- [PATCH v2 30/93] tcg/tci: Merge identical cases in generation, Richard Henderson, 2021/02/03
- [PATCH v2 38/93] tcg/tci: Merge extension operations, Richard Henderson, 2021/02/03
- [PATCH v2 43/93] tcg/tci: Split out tci_args_rrs, Richard Henderson, 2021/02/03
- [PATCH v2 29/93] tcg/tci: Remove TCG_CONST, Richard Henderson, 2021/02/03
- Re: [PATCH v2 29/93] tcg/tci: Remove TCG_CONST,
Alex Bennée <=
- [PATCH v2 45/93] tcg/tci: Split out tci_args_rrr, Richard Henderson, 2021/02/03
- [PATCH v2 51/93] tcg/tci: Reuse tci_args_l for calls., Richard Henderson, 2021/02/03
- [PATCH v2 53/93] tcg/tci: Reuse tci_args_l for goto_tb, Richard Henderson, 2021/02/03
- [PATCH v2 55/93] tcg/tci: Split out tci_args_rrrr, Richard Henderson, 2021/02/03
- [PATCH v2 32/93] tcg/tci: Remove tci_read_r8s, Richard Henderson, 2021/02/03
- [PATCH v2 34/93] tcg/tci: Remove tci_read_r16s, Richard Henderson, 2021/02/03
- [PATCH v2 33/93] tcg/tci: Remove tci_read_r16, Richard Henderson, 2021/02/03
- [PATCH v2 36/93] tcg/tci: Reduce use of tci_read_r64, Richard Henderson, 2021/02/03