[Top][All Lists]
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [Qemu-devel] [PATCH 5/8] tcg: Add interpreter for bytecode
From: |
Blue Swirl |
Subject: |
Re: [Qemu-devel] [PATCH 5/8] tcg: Add interpreter for bytecode |
Date: |
Sun, 18 Sep 2011 10:18:32 +0000 |
On Sat, Sep 17, 2011 at 8:00 PM, Stefan Weil <address@hidden> wrote:
> Signed-off-by: Stefan Weil <address@hidden>
> ---
> tcg/tcg.h | 4 +-
> tcg/tci.c | 1200
> +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
> 2 files changed, 1203 insertions(+), 1 deletions(-)
> create mode 100644 tcg/tci.c
>
> diff --git a/tcg/tcg.h b/tcg/tcg.h
> index 1859fae..c99c7ea 100644
> --- a/tcg/tcg.h
> +++ b/tcg/tcg.h
> @@ -577,7 +577,9 @@ TCGv_i32 tcg_const_local_i32(int32_t val);
> TCGv_i64 tcg_const_local_i64(int64_t val);
>
> extern uint8_t code_gen_prologue[];
> -#if defined(_ARCH_PPC) && !defined(_ARCH_PPC64)
> +#if defined(CONFIG_TCG_INTERPRETER)
> +unsigned long tcg_qemu_tb_exec(CPUState *env, uint8_t *tb_ptr);
> +#elif defined(_ARCH_PPC) && !defined(_ARCH_PPC64)
> #define tcg_qemu_tb_exec(env, tb_ptr) \
> ((long REGPARM __attribute__ ((longcall)) (*)(void *, void
> *))code_gen_prologue)(env, tb_ptr)
> #else
> diff --git a/tcg/tci.c b/tcg/tci.c
> new file mode 100644
> index 0000000..eea9992
> --- /dev/null
> +++ b/tcg/tci.c
> @@ -0,0 +1,1200 @@
> +/*
> + * Tiny Code Interpreter for QEMU
> + *
> + * Copyright (c) 2009, 2011 Stefan Weil
> + *
> + * This program is free software: you can redistribute it and/or modify
> + * it under the terms of the GNU General Public License as published by
> + * the Free Software Foundation, either version 2 of the License, or
> + * (at your option) any later version.
> + *
> + * This program is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
> + * GNU General Public License for more details.
> + *
> + * You should have received a copy of the GNU General Public License
> + * along with this program. If not, see <http://www.gnu.org/licenses/>.
> + */
> +
> +#include "config.h"
> +#include "qemu-common.h"
> +#include "exec-all.h" /* MAX_OPC_PARAM_IARGS */
> +#include "tcg-op.h"
> +
> +/* Marker for missing code. */
> +#define TODO() \
> + do { \
> + fprintf(stderr, "TODO %s:%u: %s()\n", \
> + __FILE__, __LINE__, __func__); \
> + tcg_abort(); \
> + } while (0)
> +
> +/* Trace message to see program flow. */
> +#if defined(CONFIG_DEBUG_TCG_INTERPRETER)
> +#define TRACE() \
> + loglevel \
> + ? fprintf(stderr, "TCG %s:%u: %s()\n", __FILE__, __LINE__, __func__) \
> + : (void)0
> +#else
> +#define TRACE() ((void)0)
> +#endif
The macros must use do { } while(0), but inline function would be
preferred over those as well as tracepoints.
> +
> +#if MAX_OPC_PARAM_IARGS != 4
> +# error Fix needed, number of supported input arguments changed!
> +#endif
> +#if TCG_TARGET_REG_BITS == 32
> +typedef uint64_t (*helper_function)(tcg_target_ulong, tcg_target_ulong,
> + tcg_target_ulong, tcg_target_ulong,
> + tcg_target_ulong, tcg_target_ulong,
> + tcg_target_ulong, tcg_target_ulong);
> +#else
> +typedef uint64_t (*helper_function)(tcg_target_ulong, tcg_target_ulong,
> + tcg_target_ulong, tcg_target_ulong);
> +#endif
> +
> +CPUState *env;
> +
> +/* Alpha and SH4 user mode emulations call GETPC(), so they need tci_tb_ptr.
> */
> +#if defined(CONFIG_SOFTMMU) || defined(TARGET_ALPHA) || defined(TARGET_SH4)
> +# define NEEDS_TB_PTR
This should be #defined in target-alpha/cpu.h etc. or dyngen-exec.h.
> +#endif
> +
> +#ifdef NEEDS_TB_PTR
> +uint8_t *tci_tb_ptr;
> +#endif
> +
> +static tcg_target_ulong tci_reg[TCG_TARGET_NB_REGS];
> +
> +static tcg_target_ulong tci_read_reg(TCGRegister index)
> +{
> + assert(index < ARRAY_SIZE(tci_reg));
> + return tci_reg[index];
> +}
> +
> +#if TCG_TARGET_HAS_ext8s_i32 || TCG_TARGET_HAS_ext8s_i64
> +static int8_t tci_read_reg8s(TCGRegister index)
> +{
> + return (int8_t)tci_read_reg(index);
> +}
> +#endif
> +
> +#if TCG_TARGET_HAS_ext16s_i32 || TCG_TARGET_HAS_ext16s_i64
> +static int16_t tci_read_reg16s(TCGRegister index)
> +{
> + return (int16_t)tci_read_reg(index);
> +}
> +#endif
> +
> +#if TCG_TARGET_REG_BITS == 64
> +static int32_t tci_read_reg32s(TCGRegister index)
> +{
> + return (int32_t)tci_read_reg(index);
> +}
> +#endif
> +
> +static uint8_t tci_read_reg8(TCGRegister index)
> +{
> + return (uint8_t)tci_read_reg(index);
> +}
> +
> +static uint16_t tci_read_reg16(TCGRegister index)
> +{
> + return (uint16_t)tci_read_reg(index);
> +}
> +
> +static uint32_t tci_read_reg32(TCGRegister index)
> +{
> + return (uint32_t)tci_read_reg(index);
> +}
> +
> +#if TCG_TARGET_REG_BITS == 64
> +static uint64_t tci_read_reg64(TCGRegister index)
> +{
> + return tci_read_reg(index);
> +}
> +#endif
> +
> +static void tci_write_reg(TCGRegister index, tcg_target_ulong value)
> +{
> + assert(index < ARRAY_SIZE(tci_reg));
> + assert(index != TCG_AREG0);
> + tci_reg[index] = value;
> +}
> +
> +static void tci_write_reg8s(TCGRegister index, int8_t value)
> +{
> + tci_write_reg(index, value);
> +}
> +
> +static void tci_write_reg16s(TCGRegister index, int16_t value)
> +{
> + tci_write_reg(index, value);
> +}
> +
> +#if TCG_TARGET_REG_BITS == 64
> +static void tci_write_reg32s(TCGRegister index, int32_t value)
> +{
> + tci_write_reg(index, value);
> +}
> +#endif
> +
> +static void tci_write_reg8(TCGRegister index, uint8_t value)
> +{
> + tci_write_reg(index, value);
> +}
> +
> +static void tci_write_reg16(TCGRegister index, uint16_t value)
> +{
> + tci_write_reg(index, value);
> +}
> +
> +static void tci_write_reg32(TCGRegister index, uint32_t value)
> +{
> + tci_write_reg(index, value);
> +}
> +
> +#if TCG_TARGET_REG_BITS == 32
> +static void tci_write_reg64(uint32_t high_index, uint32_t low_index,
> + uint64_t value)
> +{
> + tci_write_reg(low_index, value);
> + tci_write_reg(high_index, value >> 32);
> +}
> +#elif TCG_TARGET_REG_BITS == 64
> +static void tci_write_reg64(TCGRegister index, uint64_t value)
> +{
> + tci_write_reg(index, value);
> +}
> +#endif
> +
> +#if TCG_TARGET_REG_BITS == 32
> +/* Create a 64 bit value from two 32 bit values. */
> +static uint64_t tci_uint64(uint32_t high, uint32_t low)
> +{
> + return ((uint64_t)high << 32) + low;
> +}
> +#endif
> +
> +/* Read constant (native size) from bytecode. */
> +static tcg_target_ulong tci_read_i(uint8_t **tb_ptr)
> +{
> + tcg_target_ulong value = *(tcg_target_ulong *)(*tb_ptr);
> + *tb_ptr += sizeof(value);
> + return value;
> +}
> +
> +/* Read constant (32 bit) from bytecode. */
> +static uint32_t tci_read_i32(uint8_t **tb_ptr)
> +{
> + uint32_t value = *(uint32_t *)(*tb_ptr);
> + *tb_ptr += sizeof(value);
> + return value;
> +}
> +
> +#if TCG_TARGET_REG_BITS == 64
> +/* Read constant (64 bit) from bytecode. */
> +static uint64_t tci_read_i64(uint8_t **tb_ptr)
> +{
> + uint64_t value = *(uint64_t *)(*tb_ptr);
> + *tb_ptr += sizeof(value);
> + return value;
> +}
> +#endif
> +
> +/* Read indexed register (native size) from bytecode. */
> +static tcg_target_ulong tci_read_r(uint8_t **tb_ptr)
> +{
> + tcg_target_ulong value = tci_read_reg(**tb_ptr);
> + *tb_ptr += 1;
> + return value;
> +}
> +
> +/* Read indexed register (8 bit) from bytecode. */
> +static uint8_t tci_read_r8(uint8_t **tb_ptr)
> +{
> + uint8_t value = tci_read_reg8(**tb_ptr);
> + *tb_ptr += 1;
> + return value;
> +}
> +
> +#if TCG_TARGET_HAS_ext8s_i32 || TCG_TARGET_HAS_ext8s_i64
> +/* Read indexed register (8 bit signed) from bytecode. */
> +static int8_t tci_read_r8s(uint8_t **tb_ptr)
> +{
> + int8_t value = tci_read_reg8s(**tb_ptr);
> + *tb_ptr += 1;
> + return value;
> +}
> +#endif
> +
> +/* Read indexed register (16 bit) from bytecode. */
> +static uint16_t tci_read_r16(uint8_t **tb_ptr)
> +{
> + uint16_t value = tci_read_reg16(**tb_ptr);
> + *tb_ptr += 1;
> + return value;
> +}
> +
> +#if TCG_TARGET_HAS_ext16s_i32 || TCG_TARGET_HAS_ext16s_i64
> +/* Read indexed register (16 bit signed) from bytecode. */
> +static int16_t tci_read_r16s(uint8_t **tb_ptr)
> +{
> + int16_t value = tci_read_reg16s(**tb_ptr);
> + *tb_ptr += 1;
> + return value;
> +}
> +#endif
> +
> +/* Read indexed register (32 bit) from bytecode. */
> +static uint32_t tci_read_r32(uint8_t **tb_ptr)
> +{
> + uint32_t value = tci_read_reg32(**tb_ptr);
> + *tb_ptr += 1;
> + return value;
> +}
> +
> +#if TCG_TARGET_REG_BITS == 32
> +/* Read two indexed registers (2 * 32 bit) from bytecode. */
> +static uint64_t tci_read_r64(uint8_t **tb_ptr)
> +{
> + uint32_t low = tci_read_r32(tb_ptr);
> + return tci_uint64(tci_read_r32(tb_ptr), low);
> +}
> +#elif TCG_TARGET_REG_BITS == 64
> +/* Read indexed register (32 bit signed) from bytecode. */
> +static int32_t tci_read_r32s(uint8_t **tb_ptr)
> +{
> + int32_t value = tci_read_reg32s(**tb_ptr);
> + *tb_ptr += 1;
> + return value;
> +}
> +
> +/* Read indexed register (64 bit) from bytecode. */
> +static uint64_t tci_read_r64(uint8_t **tb_ptr)
> +{
> + uint64_t value = tci_read_reg64(**tb_ptr);
> + *tb_ptr += 1;
> + return value;
> +}
> +#endif
> +
> +/* Read indexed register(s) with target address from bytecode. */
> +static target_ulong tci_read_ulong(uint8_t **tb_ptr)
> +{
> + target_ulong taddr = tci_read_r(tb_ptr);
> +#if TARGET_LONG_BITS > TCG_TARGET_REG_BITS
> + taddr += (uint64_t)tci_read_r(tb_ptr) << 32;
> +#endif
> + return taddr;
> +}
> +
> +/* Read indexed register or constant (native size) from bytecode. */
> +static tcg_target_ulong tci_read_ri(uint8_t **tb_ptr)
> +{
> + tcg_target_ulong value;
> + TCGRegister r = **tb_ptr;
> + *tb_ptr += 1;
> + if (r == TCG_CONST) {
> + value = tci_read_i(tb_ptr);
> + } else {
> + value = tci_read_reg(r);
> + }
> + return value;
> +}
> +
> +/* Read indexed register or constant (32 bit) from bytecode. */
> +static uint32_t tci_read_ri32(uint8_t **tb_ptr)
> +{
> + uint32_t value;
> + TCGRegister r = **tb_ptr;
> + *tb_ptr += 1;
> + if (r == TCG_CONST) {
> + value = tci_read_i32(tb_ptr);
> + } else {
> + value = tci_read_reg32(r);
> + }
> + return value;
> +}
> +
> +#if TCG_TARGET_REG_BITS == 32
> +/* Read two indexed registers or constants (2 * 32 bit) from bytecode. */
> +static uint64_t tci_read_ri64(uint8_t **tb_ptr)
> +{
> + uint32_t low = tci_read_ri32(tb_ptr);
> + return tci_uint64(tci_read_ri32(tb_ptr), low);
> +}
> +#elif TCG_TARGET_REG_BITS == 64
> +/* Read indexed register or constant (64 bit) from bytecode. */
> +static uint64_t tci_read_ri64(uint8_t **tb_ptr)
> +{
> + uint64_t value;
> + TCGRegister r = **tb_ptr;
> + *tb_ptr += 1;
> + if (r == TCG_CONST) {
> + value = tci_read_i64(tb_ptr);
> + } else {
> + value = tci_read_reg64(r);
> + }
> + return value;
> +}
> +#endif
> +
> +static target_ulong tci_read_label(uint8_t **tb_ptr)
> +{
> + target_ulong label = tci_read_i(tb_ptr);
> + assert(label != 0);
> + return label;
> +}
> +
> +static bool tci_compare32(uint32_t u0, uint32_t u1, TCGCond condition)
> +{
> + bool result = false;
> + int32_t i0 = u0;
> + int32_t i1 = u1;
> + switch (condition) {
> + case TCG_COND_EQ:
> + result = (u0 == u1);
> + break;
> + case TCG_COND_NE:
> + result = (u0 != u1);
> + break;
> + case TCG_COND_LT:
> + result = (i0 < i1);
> + break;
> + case TCG_COND_GE:
> + result = (i0 >= i1);
> + break;
> + case TCG_COND_LE:
> + result = (i0 <= i1);
> + break;
> + case TCG_COND_GT:
> + result = (i0 > i1);
> + break;
> + case TCG_COND_LTU:
> + result = (u0 < u1);
> + break;
> + case TCG_COND_GEU:
> + result = (u0 >= u1);
> + break;
> + case TCG_COND_LEU:
> + result = (u0 <= u1);
> + break;
> + case TCG_COND_GTU:
> + result = (u0 > u1);
> + break;
> + default:
> + TODO();
> + }
> + return result;
> +}
> +
> +static bool tci_compare64(uint64_t u0, uint64_t u1, TCGCond condition)
> +{
> + bool result = false;
> + int64_t i0 = u0;
> + int64_t i1 = u1;
> + switch (condition) {
> + case TCG_COND_EQ:
> + result = (u0 == u1);
> + break;
> + case TCG_COND_NE:
> + result = (u0 != u1);
> + break;
> + case TCG_COND_LT:
> + result = (i0 < i1);
> + break;
> + case TCG_COND_GE:
> + result = (i0 >= i1);
> + break;
> + case TCG_COND_LE:
> + result = (i0 <= i1);
> + break;
> + case TCG_COND_GT:
> + result = (i0 > i1);
> + break;
> + case TCG_COND_LTU:
> + result = (u0 < u1);
> + break;
> + case TCG_COND_GEU:
> + result = (u0 >= u1);
> + break;
> + case TCG_COND_LEU:
> + result = (u0 <= u1);
> + break;
> + case TCG_COND_GTU:
> + result = (u0 > u1);
> + break;
> + default:
> + TODO();
> + }
> + return result;
> +}
> +
> +/* Interpret pseudo code in tb. */
> +unsigned long tcg_qemu_tb_exec(CPUState *cpustate, uint8_t *tb_ptr)
> +{
> + unsigned long next_tb = 0;
> +
> + TRACE();
> +
> + env = cpustate;
> + tci_reg[TCG_AREG0] = (tcg_target_ulong)cpustate;
> + assert(tb_ptr);
> +
> + for (;;) {
> +#ifdef NEEDS_TB_PTR
> + tci_tb_ptr = tb_ptr;
> +#endif
> + uint8_t *old_code_ptr = tb_ptr;
> + TCGOpcode opc = *tb_ptr++;
> + uint8_t op_size = *tb_ptr++;
> + tcg_target_ulong t0;
> + tcg_target_ulong t1;
> + tcg_target_ulong t2;
> + tcg_target_ulong label;
> + TCGCond condition;
> + target_ulong taddr;
> +#ifndef CONFIG_SOFTMMU
> + tcg_target_ulong host_addr;
> +#endif
> + uint8_t u8;
> + uint16_t u16;
> + uint32_t u32;
> + uint64_t u64;
The above variable names could conflict with Linux kernel headers if
they leaked the types used there. s/u/val/?
> +#if TCG_TARGET_REG_BITS == 32
> + uint64_t v64;
> +#endif
> +
> +#if defined(CONFIG_DEBUG_TCG_INTERPRETER)
> + if (loglevel) {
> + tci_disas(opc);
> + }
> +#endif
> +
> + switch (opc) {
> + case INDEX_op_end:
> + case INDEX_op_nop:
> + break;
> + case INDEX_op_nop1:
> + case INDEX_op_nop2:
> + case INDEX_op_nop3:
> + case INDEX_op_nopn:
> + case INDEX_op_discard:
> + TODO();
> + break;
> + case INDEX_op_set_label:
> + TODO();
> + break;
> + case INDEX_op_call:
> + t0 = tci_read_ri(&tb_ptr);
> +#if TCG_TARGET_REG_BITS == 32
> + u64 = ((helper_function)t0)(tci_read_reg(TCG_REG_R0),
> + tci_read_reg(TCG_REG_R1),
> + tci_read_reg(TCG_REG_R2),
> + tci_read_reg(TCG_REG_R3),
> + tci_read_reg(TCG_REG_R5),
> + tci_read_reg(TCG_REG_R6),
> + tci_read_reg(TCG_REG_R7),
> + tci_read_reg(TCG_REG_R8));
> + tci_write_reg(TCG_REG_R0, u64);
> + tci_write_reg(TCG_REG_R1, u64 >> 32);
> +#else
> + u64 = ((helper_function)t0)(tci_read_reg(TCG_REG_R0),
> + tci_read_reg(TCG_REG_R1),
> + tci_read_reg(TCG_REG_R2),
> + tci_read_reg(TCG_REG_R3));
> + tci_write_reg(TCG_REG_R0, u64);
> +#endif
> + break;
> + case INDEX_op_jmp:
> + case INDEX_op_br:
> + label = tci_read_label(&tb_ptr);
> + assert(tb_ptr == old_code_ptr + op_size);
> + tb_ptr = (uint8_t *)label;
> + continue;
> + case INDEX_op_setcond_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r32(&tb_ptr);
> + t2 = tci_read_ri32(&tb_ptr);
> + condition = *tb_ptr++;
> + tci_write_reg32(t0, tci_compare32(t1, t2, condition));
> + break;
> +#if TCG_TARGET_REG_BITS == 32
> + case INDEX_op_setcond2_i32:
> + t0 = *tb_ptr++;
> + u64 = tci_read_r64(&tb_ptr);
> + v64 = tci_read_ri64(&tb_ptr);
> + condition = *tb_ptr++;
> + tci_write_reg32(t0, tci_compare64(u64, v64, condition));
> + break;
> +#elif TCG_TARGET_REG_BITS == 64
> + case INDEX_op_setcond_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r64(&tb_ptr);
> + t2 = tci_read_ri64(&tb_ptr);
> + condition = *tb_ptr++;
> + tci_write_reg64(t0, tci_compare64(t1, t2, condition));
> + break;
> +#endif
> + case INDEX_op_mov_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r32(&tb_ptr);
> + tci_write_reg32(t0, t1);
> + break;
> + case INDEX_op_movi_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_i32(&tb_ptr);
> + tci_write_reg32(t0, t1);
> + break;
> + /* Load/store operations. */
> + case INDEX_op_ld8u_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r(&tb_ptr);
> + t2 = tci_read_i32(&tb_ptr);
> + tci_write_reg8(t0, *(uint8_t *)(t1 + t2));
> + break;
> + case INDEX_op_ld8s_i32:
> + case INDEX_op_ld16u_i32:
> + TODO();
> + break;
> + case INDEX_op_ld16s_i32:
> + TODO();
> + break;
> + case INDEX_op_ld_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r(&tb_ptr);
> + t2 = tci_read_i32(&tb_ptr);
> + tci_write_reg32(t0, *(uint32_t *)(t1 + t2));
> + break;
> + case INDEX_op_st8_i32:
> + t0 = tci_read_r8(&tb_ptr);
> + t1 = tci_read_r(&tb_ptr);
> + t2 = tci_read_i32(&tb_ptr);
> + *(uint8_t *)(t1 + t2) = t0;
> + break;
> + case INDEX_op_st16_i32:
> + t0 = tci_read_r16(&tb_ptr);
> + t1 = tci_read_r(&tb_ptr);
> + t2 = tci_read_i32(&tb_ptr);
> + *(uint16_t *)(t1 + t2) = t0;
> + break;
> + case INDEX_op_st_i32:
> + t0 = tci_read_r32(&tb_ptr);
> + t1 = tci_read_r(&tb_ptr);
> + t2 = tci_read_i32(&tb_ptr);
> + *(uint32_t *)(t1 + t2) = t0;
> + break;
> + /* Arithmetic operations. */
> + case INDEX_op_add_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri32(&tb_ptr);
> + t2 = tci_read_ri32(&tb_ptr);
> + tci_write_reg32(t0, t1 + t2);
> + break;
> + case INDEX_op_sub_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri32(&tb_ptr);
> + t2 = tci_read_ri32(&tb_ptr);
> + tci_write_reg32(t0, t1 - t2);
> + break;
> + case INDEX_op_mul_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri32(&tb_ptr);
> + t2 = tci_read_ri32(&tb_ptr);
> + tci_write_reg32(t0, t1 * t2);
> + break;
> +#if TCG_TARGET_HAS_div_i32
> + case INDEX_op_div_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri32(&tb_ptr);
> + t2 = tci_read_ri32(&tb_ptr);
> + tci_write_reg32(t0, (int32_t)t1 / (int32_t)t2);
> + break;
> + case INDEX_op_divu_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri32(&tb_ptr);
> + t2 = tci_read_ri32(&tb_ptr);
> + tci_write_reg32(t0, t1 / t2);
> + break;
> + case INDEX_op_rem_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri32(&tb_ptr);
> + t2 = tci_read_ri32(&tb_ptr);
> + tci_write_reg32(t0, (int32_t)t1 % (int32_t)t2);
> + break;
> + case INDEX_op_remu_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri32(&tb_ptr);
> + t2 = tci_read_ri32(&tb_ptr);
> + tci_write_reg32(t0, t1 % t2);
> + break;
> +#elif TCG_TARGET_HAS_div2_i32
> + case INDEX_op_div2_i32:
> + case INDEX_op_divu2_i32:
> + TODO();
> + break;
> +#endif
> + case INDEX_op_and_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri32(&tb_ptr);
> + t2 = tci_read_ri32(&tb_ptr);
> + tci_write_reg32(t0, t1 & t2);
> + break;
> + case INDEX_op_or_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri32(&tb_ptr);
> + t2 = tci_read_ri32(&tb_ptr);
> + tci_write_reg32(t0, t1 | t2);
> + break;
> + case INDEX_op_xor_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri32(&tb_ptr);
> + t2 = tci_read_ri32(&tb_ptr);
> + tci_write_reg32(t0, t1 ^ t2);
> + break;
> + /* Shift/rotate operations. */
> + case INDEX_op_shl_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri32(&tb_ptr);
> + t2 = tci_read_ri32(&tb_ptr);
> + tci_write_reg32(t0, t1 << t2);
> + break;
> + case INDEX_op_shr_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri32(&tb_ptr);
> + t2 = tci_read_ri32(&tb_ptr);
> + tci_write_reg32(t0, t1 >> t2);
> + break;
> + case INDEX_op_sar_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri32(&tb_ptr);
> + t2 = tci_read_ri32(&tb_ptr);
> + tci_write_reg32(t0, ((int32_t)t1 >> t2));
> + break;
> +#if TCG_TARGET_HAS_rot_i32
> + case INDEX_op_rotl_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri32(&tb_ptr);
> + t2 = tci_read_ri32(&tb_ptr);
> + tci_write_reg32(t0, (t1 << t2) | (t1 >> (32 - t2)));
> + break;
> + case INDEX_op_rotr_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri32(&tb_ptr);
> + t2 = tci_read_ri32(&tb_ptr);
> + tci_write_reg32(t0, (t1 >> t2) | (t1 << (32 - t2)));
> + break;
> +#endif
> + case INDEX_op_brcond_i32:
> + t0 = tci_read_r32(&tb_ptr);
> + t1 = tci_read_ri32(&tb_ptr);
> + condition = *tb_ptr++;
> + label = tci_read_label(&tb_ptr);
> + if (tci_compare32(t0, t1, condition)) {
> + assert(tb_ptr == old_code_ptr + op_size);
> + tb_ptr = (uint8_t *)label;
> + continue;
> + }
> + break;
> +#if TCG_TARGET_REG_BITS == 32
> + case INDEX_op_add2_i32:
> + t0 = *tb_ptr++;
> + t1 = *tb_ptr++;
> + u64 = tci_read_r64(&tb_ptr);
> + u64 += tci_read_r64(&tb_ptr);
> + tci_write_reg64(t1, t0, u64);
> + break;
> + case INDEX_op_sub2_i32:
> + t0 = *tb_ptr++;
> + t1 = *tb_ptr++;
> + u64 = tci_read_r64(&tb_ptr);
> + u64 -= tci_read_r64(&tb_ptr);
> + tci_write_reg64(t1, t0, u64);
> + break;
> + case INDEX_op_brcond2_i32:
> + u64 = tci_read_r64(&tb_ptr);
> + v64 = tci_read_ri64(&tb_ptr);
> + condition = *tb_ptr++;
> + label = tci_read_label(&tb_ptr);
> + if (tci_compare64(u64, v64, condition)) {
> + assert(tb_ptr == old_code_ptr + op_size);
> + tb_ptr = (uint8_t *)label;
> + continue;
> + }
> + break;
> + case INDEX_op_mulu2_i32:
> + t0 = *tb_ptr++;
> + t1 = *tb_ptr++;
> + t2 = tci_read_r32(&tb_ptr);
> + u64 = tci_read_r32(&tb_ptr);
> + tci_write_reg64(t1, t0, t2 * u64);
> + break;
> +#endif /* TCG_TARGET_REG_BITS == 32 */
> +#if TCG_TARGET_HAS_ext8s_i32
> + case INDEX_op_ext8s_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r8s(&tb_ptr);
> + tci_write_reg32(t0, t1);
> + break;
> +#endif
> +#if TCG_TARGET_HAS_ext16s_i32
> + case INDEX_op_ext16s_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r16s(&tb_ptr);
> + tci_write_reg32(t0, t1);
> + break;
> +#endif
> +#if TCG_TARGET_HAS_ext8u_i32
> + case INDEX_op_ext8u_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r8(&tb_ptr);
> + tci_write_reg32(t0, t1);
> + break;
> +#endif
> +#if TCG_TARGET_HAS_ext16u_i32
> + case INDEX_op_ext16u_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r16(&tb_ptr);
> + tci_write_reg32(t0, t1);
> + break;
> +#endif
> +#if TCG_TARGET_HAS_bswap16_i32
> + case INDEX_op_bswap16_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r16(&tb_ptr);
> + tci_write_reg32(t0, bswap16(t1));
> + break;
> +#endif
> +#if TCG_TARGET_HAS_bswap32_i32
> + case INDEX_op_bswap32_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r32(&tb_ptr);
> + tci_write_reg32(t0, bswap32(t1));
> + break;
> +#endif
> +#if TCG_TARGET_HAS_not_i32
> + case INDEX_op_not_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r32(&tb_ptr);
> + tci_write_reg32(t0, ~t1);
> + break;
> +#endif
> +#if TCG_TARGET_HAS_neg_i32
> + case INDEX_op_neg_i32:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r32(&tb_ptr);
> + tci_write_reg32(t0, -t1);
> + break;
> +#endif
> +#if TCG_TARGET_REG_BITS == 64
> + case INDEX_op_mov_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r64(&tb_ptr);
> + tci_write_reg64(t0, t1);
> + break;
> + case INDEX_op_movi_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_i64(&tb_ptr);
> + tci_write_reg64(t0, t1);
> + break;
> + /* Load/store operations. */
> + case INDEX_op_ld8u_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r(&tb_ptr);
> + t2 = tci_read_i32(&tb_ptr);
> + tci_write_reg8(t0, *(uint8_t *)(t1 + t2));
> + break;
> + case INDEX_op_ld8s_i64:
> + case INDEX_op_ld16u_i64:
> + case INDEX_op_ld16s_i64:
> + TODO();
> + break;
> + case INDEX_op_ld32u_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r(&tb_ptr);
> + t2 = tci_read_i32(&tb_ptr);
> + tci_write_reg32(t0, *(uint32_t *)(t1 + t2));
> + break;
> + case INDEX_op_ld32s_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r(&tb_ptr);
> + t2 = tci_read_i32(&tb_ptr);
> + tci_write_reg32s(t0, *(int32_t *)(t1 + t2));
> + break;
> + case INDEX_op_ld_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r(&tb_ptr);
> + t2 = tci_read_i32(&tb_ptr);
> + tci_write_reg64(t0, *(uint64_t *)(t1 + t2));
> + break;
> + case INDEX_op_st8_i64:
> + t0 = tci_read_r8(&tb_ptr);
> + t1 = tci_read_r(&tb_ptr);
> + t2 = tci_read_i32(&tb_ptr);
> + *(uint8_t *)(t1 + t2) = t0;
> + break;
> + case INDEX_op_st16_i64:
> + t0 = tci_read_r16(&tb_ptr);
> + t1 = tci_read_r(&tb_ptr);
> + t2 = tci_read_i32(&tb_ptr);
> + *(uint16_t *)(t1 + t2) = t0;
> + break;
> + case INDEX_op_st32_i64:
> + t0 = tci_read_r32(&tb_ptr);
> + t1 = tci_read_r(&tb_ptr);
> + t2 = tci_read_i32(&tb_ptr);
> + *(uint32_t *)(t1 + t2) = t0;
> + break;
> + case INDEX_op_st_i64:
> + t0 = tci_read_r64(&tb_ptr);
> + t1 = tci_read_r(&tb_ptr);
> + t2 = tci_read_i32(&tb_ptr);
> + *(uint64_t *)(t1 + t2) = t0;
> + break;
> + /* Arithmetic operations. */
> + case INDEX_op_add_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri64(&tb_ptr);
> + t2 = tci_read_ri64(&tb_ptr);
> + tci_write_reg64(t0, t1 + t2);
> + break;
> + case INDEX_op_sub_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri64(&tb_ptr);
> + t2 = tci_read_ri64(&tb_ptr);
> + tci_write_reg64(t0, t1 - t2);
> + break;
> + case INDEX_op_mul_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri64(&tb_ptr);
> + t2 = tci_read_ri64(&tb_ptr);
> + tci_write_reg64(t0, t1 * t2);
> + break;
> +#if TCG_TARGET_HAS_div_i64
> + case INDEX_op_div_i64:
> + case INDEX_op_divu_i64:
> + case INDEX_op_rem_i64:
> + case INDEX_op_remu_i64:
> + TODO();
> + break;
> +#elif TCG_TARGET_HAS_div2_i64
> + case INDEX_op_div2_i64:
> + case INDEX_op_divu2_i64:
> + TODO();
> + break;
> +#endif
> + case INDEX_op_and_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri64(&tb_ptr);
> + t2 = tci_read_ri64(&tb_ptr);
> + tci_write_reg64(t0, t1 & t2);
> + break;
> + case INDEX_op_or_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri64(&tb_ptr);
> + t2 = tci_read_ri64(&tb_ptr);
> + tci_write_reg64(t0, t1 | t2);
> + break;
> + case INDEX_op_xor_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri64(&tb_ptr);
> + t2 = tci_read_ri64(&tb_ptr);
> + tci_write_reg64(t0, t1 ^ t2);
> + break;
> + /* Shift/rotate operations. */
> + case INDEX_op_shl_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri64(&tb_ptr);
> + t2 = tci_read_ri64(&tb_ptr);
> + tci_write_reg64(t0, t1 << t2);
> + break;
> + case INDEX_op_shr_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri64(&tb_ptr);
> + t2 = tci_read_ri64(&tb_ptr);
> + tci_write_reg64(t0, t1 >> t2);
> + break;
> + case INDEX_op_sar_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_ri64(&tb_ptr);
> + t2 = tci_read_ri64(&tb_ptr);
> + tci_write_reg64(t0, ((int64_t)t1 >> t2));
> + break;
> +#if TCG_TARGET_HAS_rot_i64
> + case INDEX_op_rotl_i64:
> + case INDEX_op_rotr_i64:
> + TODO();
> + break;
> +#endif
> + case INDEX_op_brcond_i64:
> + t0 = tci_read_r64(&tb_ptr);
> + t1 = tci_read_ri64(&tb_ptr);
> + condition = *tb_ptr++;
> + label = tci_read_label(&tb_ptr);
> + if (tci_compare64(t0, t1, condition)) {
> + assert(tb_ptr == old_code_ptr + op_size);
> + tb_ptr = (uint8_t *)label;
> + continue;
> + }
> + break;
> +#if TCG_TARGET_HAS_ext8u_i64
> + case INDEX_op_ext8u_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r8(&tb_ptr);
> + tci_write_reg64(t0, t1);
> + break;
> +#endif
> +#if TCG_TARGET_HAS_ext8s_i64
> + case INDEX_op_ext8s_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r8s(&tb_ptr);
> + tci_write_reg64(t0, t1);
> + break;
> +#endif
> +#if TCG_TARGET_HAS_ext16s_i64
> + case INDEX_op_ext16s_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r16s(&tb_ptr);
> + tci_write_reg64(t0, t1);
> + break;
> +#endif
> +#if TCG_TARGET_HAS_ext16u_i64
> + case INDEX_op_ext16u_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r16(&tb_ptr);
> + tci_write_reg64(t0, t1);
> + break;
> +#endif
> +#if TCG_TARGET_HAS_ext32s_i64
> + case INDEX_op_ext32s_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r32s(&tb_ptr);
> + tci_write_reg64(t0, t1);
> + break;
> +#endif
> +#if TCG_TARGET_HAS_ext32u_i64
> + case INDEX_op_ext32u_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r32(&tb_ptr);
> + tci_write_reg64(t0, t1);
> + break;
> +#endif
> +#if TCG_TARGET_HAS_bswap16_i64
> + case INDEX_op_bswap16_i64:
> + TODO();
> + t0 = *tb_ptr++;
> + t1 = tci_read_r16(&tb_ptr);
> + tci_write_reg64(t0, bswap16(t1));
> + break;
> +#endif
> +#if TCG_TARGET_HAS_bswap32_i64
> + case INDEX_op_bswap32_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r32(&tb_ptr);
> + tci_write_reg64(t0, bswap32(t1));
> + break;
> +#endif
> +#if TCG_TARGET_HAS_bswap64_i64
> + case INDEX_op_bswap64_i64:
> + TODO();
> + t0 = *tb_ptr++;
> + t1 = tci_read_r64(&tb_ptr);
> + tci_write_reg64(t0, bswap64(t1));
> + break;
> +#endif
> +#if TCG_TARGET_HAS_not_i64
> + case INDEX_op_not_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r64(&tb_ptr);
> + tci_write_reg64(t0, ~t1);
> + break;
> +#endif
> +#if TCG_TARGET_HAS_neg_i64
> + case INDEX_op_neg_i64:
> + t0 = *tb_ptr++;
> + t1 = tci_read_r64(&tb_ptr);
> + tci_write_reg64(t0, -t1);
> + break;
> +#endif
> +#endif /* TCG_TARGET_REG_BITS == 64 */
> + /* QEMU specific */
> +#if TARGET_LONG_BITS > TCG_TARGET_REG_BITS
> + case INDEX_op_debug_insn_start:
> + TODO();
> + break;
> +#else
> + case INDEX_op_debug_insn_start:
> + TODO();
> + break;
> +#endif
> + case INDEX_op_exit_tb:
> + next_tb = *(uint64_t *)tb_ptr;
> + goto exit;
> + break;
> + case INDEX_op_goto_tb:
> + t0 = tci_read_i32(&tb_ptr);
> + assert(tb_ptr == old_code_ptr + op_size);
> + tb_ptr += (int32_t)t0;
> + continue;
> + case INDEX_op_qemu_ld8u:
> + t0 = *tb_ptr++;
> + taddr = tci_read_ulong(&tb_ptr);
> +#ifdef CONFIG_SOFTMMU
> + u8 = __ldb_mmu(taddr, tci_read_i(&tb_ptr));
> +#else
> + host_addr = (tcg_target_ulong)taddr;
> + assert(taddr == host_addr);
> + u8 = *(uint8_t *)(host_addr + GUEST_BASE);
> +#endif
> + tci_write_reg8(t0, u8);
> + break;
> + case INDEX_op_qemu_ld8s:
> + t0 = *tb_ptr++;
> + taddr = tci_read_ulong(&tb_ptr);
> +#ifdef CONFIG_SOFTMMU
> + u8 = __ldb_mmu(taddr, tci_read_i(&tb_ptr));
> +#else
> + host_addr = (tcg_target_ulong)taddr;
> + assert(taddr == host_addr);
> + u8 = *(uint8_t *)(host_addr + GUEST_BASE);
> +#endif
> + tci_write_reg8s(t0, u8);
> + break;
> + case INDEX_op_qemu_ld16u:
> + t0 = *tb_ptr++;
> + taddr = tci_read_ulong(&tb_ptr);
> +#ifdef CONFIG_SOFTMMU
> + u16 = __ldw_mmu(taddr, tci_read_i(&tb_ptr));
> +#else
> + host_addr = (tcg_target_ulong)taddr;
> + assert(taddr == host_addr);
> + u16 = tswap16(*(uint16_t *)(host_addr + GUEST_BASE));
> +#endif
> + tci_write_reg16(t0, u16);
> + break;
> + case INDEX_op_qemu_ld16s:
> + t0 = *tb_ptr++;
> + taddr = tci_read_ulong(&tb_ptr);
> +#ifdef CONFIG_SOFTMMU
> + u16 = __ldw_mmu(taddr, tci_read_i(&tb_ptr));
> +#else
> + host_addr = (tcg_target_ulong)taddr;
> + assert(taddr == host_addr);
> + u16 = tswap16(*(uint16_t *)(host_addr + GUEST_BASE));
> +#endif
> + tci_write_reg16s(t0, u16);
> + break;
> +#if TCG_TARGET_REG_BITS == 64
> + case INDEX_op_qemu_ld32u:
> + t0 = *tb_ptr++;
> + taddr = tci_read_ulong(&tb_ptr);
> +#ifdef CONFIG_SOFTMMU
> + u32 = __ldl_mmu(taddr, tci_read_i(&tb_ptr));
> +#else
> + host_addr = (tcg_target_ulong)taddr;
> + assert(taddr == host_addr);
> + u32 = tswap32(*(uint32_t *)(host_addr + GUEST_BASE));
> +#endif
> + tci_write_reg32(t0, u32);
> + break;
> + case INDEX_op_qemu_ld32s:
> + t0 = *tb_ptr++;
> + taddr = tci_read_ulong(&tb_ptr);
> +#ifdef CONFIG_SOFTMMU
> + u32 = __ldl_mmu(taddr, tci_read_i(&tb_ptr));
> +#else
> + host_addr = (tcg_target_ulong)taddr;
> + assert(taddr == host_addr);
> + u32 = tswap32(*(uint32_t *)(host_addr + GUEST_BASE));
> +#endif
> + tci_write_reg32s(t0, u32);
> + break;
> +#endif /* TCG_TARGET_REG_BITS == 64 */
> + case INDEX_op_qemu_ld32:
> + t0 = *tb_ptr++;
> + taddr = tci_read_ulong(&tb_ptr);
> +#ifdef CONFIG_SOFTMMU
> + u32 = __ldl_mmu(taddr, tci_read_i(&tb_ptr));
> +#else
> + host_addr = (tcg_target_ulong)taddr;
> + assert(taddr == host_addr);
> + u32 = tswap32(*(uint32_t *)(host_addr + GUEST_BASE));
> +#endif
> + tci_write_reg32(t0, u32);
> + break;
> + case INDEX_op_qemu_ld64:
> + t0 = *tb_ptr++;
> +#if TCG_TARGET_REG_BITS == 32
> + t1 = *tb_ptr++;
> +#endif
> + taddr = tci_read_ulong(&tb_ptr);
> +#ifdef CONFIG_SOFTMMU
> + u64 = __ldq_mmu(taddr, tci_read_i(&tb_ptr));
> +#else
> + host_addr = (tcg_target_ulong)taddr;
> + assert(taddr == host_addr);
> + u64 = tswap64(*(uint64_t *)(host_addr + GUEST_BASE));
> +#endif
> + tci_write_reg(t0, u64);
> +#if TCG_TARGET_REG_BITS == 32
> + tci_write_reg(t1, u64 >> 32);
> +#endif
> + break;
> + case INDEX_op_qemu_st8:
> + t0 = tci_read_r8(&tb_ptr);
> + taddr = tci_read_ulong(&tb_ptr);
> +#ifdef CONFIG_SOFTMMU
> + t2 = tci_read_i(&tb_ptr);
> + __stb_mmu(taddr, t0, t2);
> +#else
> + host_addr = (tcg_target_ulong)taddr;
> + assert(taddr == host_addr);
> + *(uint8_t *)(host_addr + GUEST_BASE) = t0;
> +#endif
> + break;
> + case INDEX_op_qemu_st16:
> + t0 = tci_read_r16(&tb_ptr);
> + taddr = tci_read_ulong(&tb_ptr);
> +#ifdef CONFIG_SOFTMMU
> + t2 = tci_read_i(&tb_ptr);
> + __stw_mmu(taddr, t0, t2);
> +#else
> + host_addr = (tcg_target_ulong)taddr;
> + assert(taddr == host_addr);
> + *(uint16_t *)(host_addr + GUEST_BASE) = tswap16(t0);
> +#endif
> + break;
> + case INDEX_op_qemu_st32:
> + t0 = tci_read_r32(&tb_ptr);
> + taddr = tci_read_ulong(&tb_ptr);
> +#ifdef CONFIG_SOFTMMU
> + t2 = tci_read_i(&tb_ptr);
> + __stl_mmu(taddr, t0, t2);
> +#else
> + host_addr = (tcg_target_ulong)taddr;
> + assert(taddr == host_addr);
> + *(uint32_t *)(host_addr + GUEST_BASE) = tswap32(t0);
> +#endif
> + break;
> + case INDEX_op_qemu_st64:
> + u64 = tci_read_r64(&tb_ptr);
> + taddr = tci_read_ulong(&tb_ptr);
> +#ifdef CONFIG_SOFTMMU
> + t2 = tci_read_i(&tb_ptr);
> + __stq_mmu(taddr, u64, t2);
> +#else
> + host_addr = (tcg_target_ulong)taddr;
> + assert(taddr == host_addr);
> + *(uint64_t *)(host_addr + GUEST_BASE) = tswap64(u64);
> +#endif
> + break;
> + default:
> + TODO();
> + break;
> + }
> + assert(tb_ptr == old_code_ptr + op_size);
> + }
> +exit:
> + return next_tb;
> +}
> --
> 1.7.2.5
>
>
>
- Re: [Qemu-devel] [PATCH 6/8] tcg: Add bytecode generator for tcg interpreter, (continued)
- [Qemu-devel] [PATCH 5/8] tcg: Add interpreter for bytecode, Stefan Weil, 2011/09/17
- Re: [Qemu-devel] [PATCH 5/8] tcg: Add interpreter for bytecode, Andi Kleen, 2011/09/18
- Re: [Qemu-devel] [PATCH 5/8] tcg: Add interpreter for bytecode, Stefan Weil, 2011/09/18
- Re: [Qemu-devel] [PATCH 5/8] tcg: Add interpreter for bytecode, Paolo Bonzini, 2011/09/18
- Re: [Qemu-devel] [PATCH 5/8] tcg: Add interpreter for bytecode, Avi Kivity, 2011/09/18
- Re: [Qemu-devel] [PATCH 5/8] tcg: Add interpreter for bytecode, Andi Kleen, 2011/09/19
- Re: [Qemu-devel] [PATCH 5/8] tcg: Add interpreter for bytecode, Avi Kivity, 2011/09/19
- Re: [Qemu-devel] [PATCH 5/8] tcg: Add interpreter for bytecode, Andi Kleen, 2011/09/19
Re: [Qemu-devel] [PATCH 5/8] tcg: Add interpreter for bytecode,
Blue Swirl <=
Re: [Qemu-devel] [PATCH 5/8] tcg: Add interpreter for bytecode, Richard Henderson, 2011/09/19
Re: [Qemu-devel] [PATCH 5/8] tcg: Add interpreter for bytecode, Stuart Brady, 2011/09/19
Re: [Qemu-devel] [PATCH 0/8] tcg/interpreter: Add TCG + interpreter for bytecode (virtual machine), Blue Swirl, 2011/09/18
- Re: [Qemu-devel] [PATCH 0/8] tcg/interpreter: Add TCG + interpreter for bytecode (virtual machine), malc, 2011/09/18
- Re: [Qemu-devel] [PATCH 0/8] tcg/interpreter: Add TCG + interpreter for bytecode (virtual machine), Blue Swirl, 2011/09/18
- Re: [Qemu-devel] [PATCH 0/8] tcg/interpreter: Add TCG + interpreter for bytecode (virtual machine), malc, 2011/09/18
- Re: [Qemu-devel] [PATCH 0/8] tcg/interpreter: Add TCG + interpreter for bytecode (virtual machine), Blue Swirl, 2011/09/18
- Re: [Qemu-devel] [PATCH 0/8] tcg/interpreter: Add TCG + interpreter for bytecode (virtual machine), malc, 2011/09/18
- Re: [Qemu-devel] [PATCH 0/8] tcg/interpreter: Add TCG + interpreter for bytecode (virtual machine), Blue Swirl, 2011/09/18
- Re: [Qemu-devel] [PATCH 0/8] tcg/interpreter: Add TCG + interpreter for bytecode (virtual machine), Stefan Weil, 2011/09/25