[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[PATCH v3 04/20] target/arm: Use tcg_gen_qemu_ld_i128 for LDXP
|
From: |
Richard Henderson |
|
Subject: |
[PATCH v3 04/20] target/arm: Use tcg_gen_qemu_ld_i128 for LDXP |
|
Date: |
Tue, 30 May 2023 12:14:22 -0700 |
While we don't require 16-byte atomicity here, using a single larger
load simplifies the code, and makes it a closer match to STXP.
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
target/arm/tcg/translate-a64.c | 31 ++++++++++++++++++++-----------
1 file changed, 20 insertions(+), 11 deletions(-)
diff --git a/target/arm/tcg/translate-a64.c b/target/arm/tcg/translate-a64.c
index 967400ed68..0e720f2612 100644
--- a/target/arm/tcg/translate-a64.c
+++ b/target/arm/tcg/translate-a64.c
@@ -2398,14 +2398,14 @@ static void gen_load_exclusive(DisasContext *s, int rt,
int rt2,
TCGv_i64 addr, int size, bool is_pair)
{
int idx = get_mem_index(s);
- MemOp memop = s->be_data;
+ MemOp memop;
g_assert(size <= 3);
if (is_pair) {
g_assert(size >= 2);
if (size == 2) {
/* The pair must be single-copy atomic for the doubleword. */
- memop |= MO_64 | MO_ALIGN;
+ memop = finalize_memop(s, MO_64 | MO_ALIGN);
tcg_gen_qemu_ld_i64(cpu_exclusive_val, addr, idx, memop);
if (s->be_data == MO_LE) {
tcg_gen_extract_i64(cpu_reg(s, rt), cpu_exclusive_val, 0, 32);
@@ -2415,21 +2415,30 @@ static void gen_load_exclusive(DisasContext *s, int rt,
int rt2,
tcg_gen_extract_i64(cpu_reg(s, rt2), cpu_exclusive_val, 0, 32);
}
} else {
- /* The pair must be single-copy atomic for *each* doubleword, not
- the entire quadword, however it must be quadword aligned. */
- memop |= MO_64;
- tcg_gen_qemu_ld_i64(cpu_exclusive_val, addr, idx,
- memop | MO_ALIGN_16);
+ /*
+ * The pair must be single-copy atomic for *each* doubleword, not
+ * the entire quadword, however it must be quadword aligned.
+ * Expose the complete load to tcg, for ease of tlb lookup,
+ * but indicate that only 8-byte atomicity is required.
+ */
+ TCGv_i128 t16 = tcg_temp_new_i128();
- TCGv_i64 addr2 = tcg_temp_new_i64();
- tcg_gen_addi_i64(addr2, addr, 8);
- tcg_gen_qemu_ld_i64(cpu_exclusive_high, addr2, idx, memop);
+ memop = finalize_memop_atom(s, MO_128 | MO_ALIGN_16,
+ MO_ATOM_IFALIGN_PAIR);
+ tcg_gen_qemu_ld_i128(t16, addr, idx, memop);
+ if (s->be_data == MO_LE) {
+ tcg_gen_extr_i128_i64(cpu_exclusive_val,
+ cpu_exclusive_high, t16);
+ } else {
+ tcg_gen_extr_i128_i64(cpu_exclusive_high,
+ cpu_exclusive_val, t16);
+ }
tcg_gen_mov_i64(cpu_reg(s, rt), cpu_exclusive_val);
tcg_gen_mov_i64(cpu_reg(s, rt2), cpu_exclusive_high);
}
} else {
- memop |= size | MO_ALIGN;
+ memop = finalize_memop(s, size | MO_ALIGN);
tcg_gen_qemu_ld_i64(cpu_exclusive_val, addr, idx, memop);
tcg_gen_mov_i64(cpu_reg(s, rt), cpu_exclusive_val);
}
--
2.34.1
- [PATCH v3 00/20] target/arm: Implement FEAT_LSE2, Richard Henderson, 2023/05/30
- [PATCH v3 02/20] target/arm: Add feature test for FEAT_LSE2, Richard Henderson, 2023/05/30
- [PATCH v3 01/20] target/arm: Add commentary for CPUARMState.exclusive_high, Richard Henderson, 2023/05/30
- [PATCH v3 03/20] target/arm: Introduce finalize_memop_{atom,pair}, Richard Henderson, 2023/05/30
- [PATCH v3 05/20] target/arm: Use tcg_gen_qemu_{st, ld}_i128 for do_fp_{st, ld}, Richard Henderson, 2023/05/30
- [PATCH v3 09/20] target/arm: Load/store integer pair with one tcg operation, Richard Henderson, 2023/05/30
- [PATCH v3 13/20] target/arm: Pass single_memop to gen_mte_checkN, Richard Henderson, 2023/05/30
- [PATCH v3 07/20] target/arm: Use tcg_gen_qemu_{ld, st}_i128 in gen_sve_{ld, st}r, Richard Henderson, 2023/05/30
- [PATCH v3 04/20] target/arm: Use tcg_gen_qemu_ld_i128 for LDXP,
Richard Henderson <=
- [PATCH v3 06/20] target/arm: Use tcg_gen_qemu_st_i128 for STZG, STZ2G, Richard Henderson, 2023/05/30
- [PATCH v3 11/20] target/arm: Hoist finalize_memop out of do_fp_{ld, st}, Richard Henderson, 2023/05/30
- [PATCH v3 08/20] target/arm: Sink gen_mte_check1 into load/store_exclusive, Richard Henderson, 2023/05/30
- [PATCH v3 10/20] target/arm: Hoist finalize_memop out of do_gpr_{ld, st}, Richard Henderson, 2023/05/30
- [PATCH v3 12/20] target/arm: Pass memop to gen_mte_check1*, Richard Henderson, 2023/05/30
- [PATCH v3 16/20] target/arm: Relax ordered/atomic alignment checks for LSE2, Richard Henderson, 2023/05/30
- [PATCH v3 17/20] target/arm: Move mte check for store-exclusive, Richard Henderson, 2023/05/30
- [PATCH v3 14/20] target/arm: Check alignment in helper_mte_check, Richard Henderson, 2023/05/30
- [PATCH v3 20/20] target/arm: Enable FEAT_LSE2 for -cpu max, Richard Henderson, 2023/05/30
- [PATCH v3 18/20] tests/tcg/aarch64: Use stz2g in mte-7.c, Richard Henderson, 2023/05/30