accel/tcg: Standardize atomic helpers on softmmu api
Reduce the amount of code duplication by always passing the TCGMemOpIdx argument to helper_atomic_*. This is not currently used for user-only, but it's easy to ignore. Tested-by: Cole Robinson <crobinso@redhat.com> Reviewed-by: Peter Maydell <peter.maydell@linaro.org> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
This commit is contained in:
parent
be9568b4e0
commit
e28a866438
@ -52,3 +52,73 @@ void atomic_trace_st_post(CPUArchState *env, target_ulong addr, uint16_t info)
|
|||||||
{
|
{
|
||||||
qemu_plugin_vcpu_mem_cb(env_cpu(env), addr, info);
|
qemu_plugin_vcpu_mem_cb(env_cpu(env), addr, info);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Atomic helpers callable from TCG.
|
||||||
|
* These have a common interface and all defer to cpu_atomic_*
|
||||||
|
* using the host return address from GETPC().
|
||||||
|
*/
|
||||||
|
|
||||||
|
#define CMPXCHG_HELPER(OP, TYPE) \
|
||||||
|
TYPE HELPER(atomic_##OP)(CPUArchState *env, target_ulong addr, \
|
||||||
|
TYPE oldv, TYPE newv, uint32_t oi) \
|
||||||
|
{ return cpu_atomic_##OP##_mmu(env, addr, oldv, newv, oi, GETPC()); }
|
||||||
|
|
||||||
|
CMPXCHG_HELPER(cmpxchgb, uint32_t)
|
||||||
|
CMPXCHG_HELPER(cmpxchgw_be, uint32_t)
|
||||||
|
CMPXCHG_HELPER(cmpxchgw_le, uint32_t)
|
||||||
|
CMPXCHG_HELPER(cmpxchgl_be, uint32_t)
|
||||||
|
CMPXCHG_HELPER(cmpxchgl_le, uint32_t)
|
||||||
|
|
||||||
|
#ifdef CONFIG_ATOMIC64
|
||||||
|
CMPXCHG_HELPER(cmpxchgq_be, uint64_t)
|
||||||
|
CMPXCHG_HELPER(cmpxchgq_le, uint64_t)
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#undef CMPXCHG_HELPER
|
||||||
|
|
||||||
|
#define ATOMIC_HELPER(OP, TYPE) \
|
||||||
|
TYPE HELPER(glue(atomic_,OP))(CPUArchState *env, target_ulong addr, \
|
||||||
|
TYPE val, uint32_t oi) \
|
||||||
|
{ return glue(glue(cpu_atomic_,OP),_mmu)(env, addr, val, oi, GETPC()); }
|
||||||
|
|
||||||
|
#ifdef CONFIG_ATOMIC64
|
||||||
|
#define GEN_ATOMIC_HELPERS(OP) \
|
||||||
|
ATOMIC_HELPER(glue(OP,b), uint32_t) \
|
||||||
|
ATOMIC_HELPER(glue(OP,w_be), uint32_t) \
|
||||||
|
ATOMIC_HELPER(glue(OP,w_le), uint32_t) \
|
||||||
|
ATOMIC_HELPER(glue(OP,l_be), uint32_t) \
|
||||||
|
ATOMIC_HELPER(glue(OP,l_le), uint32_t) \
|
||||||
|
ATOMIC_HELPER(glue(OP,q_be), uint64_t) \
|
||||||
|
ATOMIC_HELPER(glue(OP,q_le), uint64_t)
|
||||||
|
#else
|
||||||
|
#define GEN_ATOMIC_HELPERS(OP) \
|
||||||
|
ATOMIC_HELPER(glue(OP,b), uint32_t) \
|
||||||
|
ATOMIC_HELPER(glue(OP,w_be), uint32_t) \
|
||||||
|
ATOMIC_HELPER(glue(OP,w_le), uint32_t) \
|
||||||
|
ATOMIC_HELPER(glue(OP,l_be), uint32_t) \
|
||||||
|
ATOMIC_HELPER(glue(OP,l_le), uint32_t)
|
||||||
|
#endif
|
||||||
|
|
||||||
|
GEN_ATOMIC_HELPERS(fetch_add)
|
||||||
|
GEN_ATOMIC_HELPERS(fetch_and)
|
||||||
|
GEN_ATOMIC_HELPERS(fetch_or)
|
||||||
|
GEN_ATOMIC_HELPERS(fetch_xor)
|
||||||
|
GEN_ATOMIC_HELPERS(fetch_smin)
|
||||||
|
GEN_ATOMIC_HELPERS(fetch_umin)
|
||||||
|
GEN_ATOMIC_HELPERS(fetch_smax)
|
||||||
|
GEN_ATOMIC_HELPERS(fetch_umax)
|
||||||
|
|
||||||
|
GEN_ATOMIC_HELPERS(add_fetch)
|
||||||
|
GEN_ATOMIC_HELPERS(and_fetch)
|
||||||
|
GEN_ATOMIC_HELPERS(or_fetch)
|
||||||
|
GEN_ATOMIC_HELPERS(xor_fetch)
|
||||||
|
GEN_ATOMIC_HELPERS(smin_fetch)
|
||||||
|
GEN_ATOMIC_HELPERS(umin_fetch)
|
||||||
|
GEN_ATOMIC_HELPERS(smax_fetch)
|
||||||
|
GEN_ATOMIC_HELPERS(umax_fetch)
|
||||||
|
|
||||||
|
GEN_ATOMIC_HELPERS(xchg)
|
||||||
|
|
||||||
|
#undef ATOMIC_HELPER
|
||||||
|
#undef GEN_ATOMIC_HELPERS
|
||||||
|
@ -2725,38 +2725,6 @@ void cpu_stq_le_data(CPUArchState *env, target_ulong ptr, uint64_t val)
|
|||||||
#include "atomic_template.h"
|
#include "atomic_template.h"
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
/* Second set of helpers are directly callable from TCG as helpers. */
|
|
||||||
|
|
||||||
#undef EXTRA_ARGS
|
|
||||||
#undef ATOMIC_NAME
|
|
||||||
#undef ATOMIC_MMU_LOOKUP_RW
|
|
||||||
#undef ATOMIC_MMU_LOOKUP_R
|
|
||||||
#undef ATOMIC_MMU_LOOKUP_W
|
|
||||||
|
|
||||||
#define EXTRA_ARGS , TCGMemOpIdx oi
|
|
||||||
#define ATOMIC_NAME(X) HELPER(glue(glue(atomic_ ## X, SUFFIX), END))
|
|
||||||
#define ATOMIC_MMU_LOOKUP_RW \
|
|
||||||
atomic_mmu_lookup(env, addr, oi, DATA_SIZE, PAGE_READ | PAGE_WRITE, GETPC())
|
|
||||||
#define ATOMIC_MMU_LOOKUP_R \
|
|
||||||
atomic_mmu_lookup(env, addr, oi, DATA_SIZE, PAGE_READ, GETPC())
|
|
||||||
#define ATOMIC_MMU_LOOKUP_W \
|
|
||||||
atomic_mmu_lookup(env, addr, oi, DATA_SIZE, PAGE_WRITE, GETPC())
|
|
||||||
|
|
||||||
#define DATA_SIZE 1
|
|
||||||
#include "atomic_template.h"
|
|
||||||
|
|
||||||
#define DATA_SIZE 2
|
|
||||||
#include "atomic_template.h"
|
|
||||||
|
|
||||||
#define DATA_SIZE 4
|
|
||||||
#include "atomic_template.h"
|
|
||||||
|
|
||||||
#ifdef CONFIG_ATOMIC64
|
|
||||||
#define DATA_SIZE 8
|
|
||||||
#include "atomic_template.h"
|
|
||||||
#endif
|
|
||||||
#undef ATOMIC_MMU_IDX
|
|
||||||
|
|
||||||
/* Code access functions. */
|
/* Code access functions. */
|
||||||
|
|
||||||
static uint64_t full_ldub_code(CPUArchState *env, target_ulong addr,
|
static uint64_t full_ldub_code(CPUArchState *env, target_ulong addr,
|
||||||
|
@ -39,8 +39,6 @@ DEF_HELPER_FLAGS_1(exit_atomic, TCG_CALL_NO_WG, noreturn, env)
|
|||||||
DEF_HELPER_FLAGS_3(memset, TCG_CALL_NO_RWG, ptr, ptr, int, ptr)
|
DEF_HELPER_FLAGS_3(memset, TCG_CALL_NO_RWG, ptr, ptr, int, ptr)
|
||||||
#endif /* IN_HELPER_PROTO */
|
#endif /* IN_HELPER_PROTO */
|
||||||
|
|
||||||
#ifdef CONFIG_SOFTMMU
|
|
||||||
|
|
||||||
DEF_HELPER_FLAGS_5(atomic_cmpxchgb, TCG_CALL_NO_WG,
|
DEF_HELPER_FLAGS_5(atomic_cmpxchgb, TCG_CALL_NO_WG,
|
||||||
i32, env, tl, i32, i32, i32)
|
i32, env, tl, i32, i32, i32)
|
||||||
DEF_HELPER_FLAGS_5(atomic_cmpxchgw_be, TCG_CALL_NO_WG,
|
DEF_HELPER_FLAGS_5(atomic_cmpxchgw_be, TCG_CALL_NO_WG,
|
||||||
@ -88,50 +86,6 @@ DEF_HELPER_FLAGS_5(atomic_cmpxchgq_le, TCG_CALL_NO_WG,
|
|||||||
TCG_CALL_NO_WG, i32, env, tl, i32, i32)
|
TCG_CALL_NO_WG, i32, env, tl, i32, i32)
|
||||||
#endif /* CONFIG_ATOMIC64 */
|
#endif /* CONFIG_ATOMIC64 */
|
||||||
|
|
||||||
#else
|
|
||||||
|
|
||||||
DEF_HELPER_FLAGS_4(atomic_cmpxchgb, TCG_CALL_NO_WG, i32, env, tl, i32, i32)
|
|
||||||
DEF_HELPER_FLAGS_4(atomic_cmpxchgw_be, TCG_CALL_NO_WG, i32, env, tl, i32, i32)
|
|
||||||
DEF_HELPER_FLAGS_4(atomic_cmpxchgw_le, TCG_CALL_NO_WG, i32, env, tl, i32, i32)
|
|
||||||
DEF_HELPER_FLAGS_4(atomic_cmpxchgl_be, TCG_CALL_NO_WG, i32, env, tl, i32, i32)
|
|
||||||
DEF_HELPER_FLAGS_4(atomic_cmpxchgl_le, TCG_CALL_NO_WG, i32, env, tl, i32, i32)
|
|
||||||
#ifdef CONFIG_ATOMIC64
|
|
||||||
DEF_HELPER_FLAGS_4(atomic_cmpxchgq_be, TCG_CALL_NO_WG, i64, env, tl, i64, i64)
|
|
||||||
DEF_HELPER_FLAGS_4(atomic_cmpxchgq_le, TCG_CALL_NO_WG, i64, env, tl, i64, i64)
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#ifdef CONFIG_ATOMIC64
|
|
||||||
#define GEN_ATOMIC_HELPERS(NAME) \
|
|
||||||
DEF_HELPER_FLAGS_3(glue(glue(atomic_, NAME), b), \
|
|
||||||
TCG_CALL_NO_WG, i32, env, tl, i32) \
|
|
||||||
DEF_HELPER_FLAGS_3(glue(glue(atomic_, NAME), w_le), \
|
|
||||||
TCG_CALL_NO_WG, i32, env, tl, i32) \
|
|
||||||
DEF_HELPER_FLAGS_3(glue(glue(atomic_, NAME), w_be), \
|
|
||||||
TCG_CALL_NO_WG, i32, env, tl, i32) \
|
|
||||||
DEF_HELPER_FLAGS_3(glue(glue(atomic_, NAME), l_le), \
|
|
||||||
TCG_CALL_NO_WG, i32, env, tl, i32) \
|
|
||||||
DEF_HELPER_FLAGS_3(glue(glue(atomic_, NAME), l_be), \
|
|
||||||
TCG_CALL_NO_WG, i32, env, tl, i32) \
|
|
||||||
DEF_HELPER_FLAGS_3(glue(glue(atomic_, NAME), q_le), \
|
|
||||||
TCG_CALL_NO_WG, i64, env, tl, i64) \
|
|
||||||
DEF_HELPER_FLAGS_3(glue(glue(atomic_, NAME), q_be), \
|
|
||||||
TCG_CALL_NO_WG, i64, env, tl, i64)
|
|
||||||
#else
|
|
||||||
#define GEN_ATOMIC_HELPERS(NAME) \
|
|
||||||
DEF_HELPER_FLAGS_3(glue(glue(atomic_, NAME), b), \
|
|
||||||
TCG_CALL_NO_WG, i32, env, tl, i32) \
|
|
||||||
DEF_HELPER_FLAGS_3(glue(glue(atomic_, NAME), w_le), \
|
|
||||||
TCG_CALL_NO_WG, i32, env, tl, i32) \
|
|
||||||
DEF_HELPER_FLAGS_3(glue(glue(atomic_, NAME), w_be), \
|
|
||||||
TCG_CALL_NO_WG, i32, env, tl, i32) \
|
|
||||||
DEF_HELPER_FLAGS_3(glue(glue(atomic_, NAME), l_le), \
|
|
||||||
TCG_CALL_NO_WG, i32, env, tl, i32) \
|
|
||||||
DEF_HELPER_FLAGS_3(glue(glue(atomic_, NAME), l_be), \
|
|
||||||
TCG_CALL_NO_WG, i32, env, tl, i32)
|
|
||||||
#endif /* CONFIG_ATOMIC64 */
|
|
||||||
|
|
||||||
#endif /* CONFIG_SOFTMMU */
|
|
||||||
|
|
||||||
GEN_ATOMIC_HELPERS(fetch_add)
|
GEN_ATOMIC_HELPERS(fetch_add)
|
||||||
GEN_ATOMIC_HELPERS(fetch_and)
|
GEN_ATOMIC_HELPERS(fetch_and)
|
||||||
GEN_ATOMIC_HELPERS(fetch_or)
|
GEN_ATOMIC_HELPERS(fetch_or)
|
||||||
|
@ -1269,29 +1269,3 @@ static void *atomic_mmu_lookup(CPUArchState *env, target_ulong addr,
|
|||||||
#define DATA_SIZE 16
|
#define DATA_SIZE 16
|
||||||
#include "atomic_template.h"
|
#include "atomic_template.h"
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
/*
|
|
||||||
* Second set of functions is directly callable from TCG.
|
|
||||||
*/
|
|
||||||
|
|
||||||
#undef EXTRA_ARGS
|
|
||||||
#undef ATOMIC_NAME
|
|
||||||
#undef ATOMIC_MMU_DECLS
|
|
||||||
|
|
||||||
#define ATOMIC_NAME(X) HELPER(glue(glue(atomic_ ## X, SUFFIX), END))
|
|
||||||
#define EXTRA_ARGS
|
|
||||||
#define ATOMIC_MMU_DECLS uintptr_t retaddr = GETPC()
|
|
||||||
|
|
||||||
#define DATA_SIZE 1
|
|
||||||
#include "atomic_template.h"
|
|
||||||
|
|
||||||
#define DATA_SIZE 2
|
|
||||||
#include "atomic_template.h"
|
|
||||||
|
|
||||||
#define DATA_SIZE 4
|
|
||||||
#include "atomic_template.h"
|
|
||||||
|
|
||||||
#ifdef CONFIG_ATOMIC64
|
|
||||||
#define DATA_SIZE 8
|
|
||||||
#include "atomic_template.h"
|
|
||||||
#endif
|
|
||||||
|
51
tcg/tcg-op.c
51
tcg/tcg-op.c
@ -3084,7 +3084,6 @@ static void tcg_gen_ext_i64(TCGv_i64 ret, TCGv_i64 val, MemOp opc)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_SOFTMMU
|
|
||||||
typedef void (*gen_atomic_cx_i32)(TCGv_i32, TCGv_env, TCGv,
|
typedef void (*gen_atomic_cx_i32)(TCGv_i32, TCGv_env, TCGv,
|
||||||
TCGv_i32, TCGv_i32, TCGv_i32);
|
TCGv_i32, TCGv_i32, TCGv_i32);
|
||||||
typedef void (*gen_atomic_cx_i64)(TCGv_i64, TCGv_env, TCGv,
|
typedef void (*gen_atomic_cx_i64)(TCGv_i64, TCGv_env, TCGv,
|
||||||
@ -3093,12 +3092,6 @@ typedef void (*gen_atomic_op_i32)(TCGv_i32, TCGv_env, TCGv,
|
|||||||
TCGv_i32, TCGv_i32);
|
TCGv_i32, TCGv_i32);
|
||||||
typedef void (*gen_atomic_op_i64)(TCGv_i64, TCGv_env, TCGv,
|
typedef void (*gen_atomic_op_i64)(TCGv_i64, TCGv_env, TCGv,
|
||||||
TCGv_i64, TCGv_i32);
|
TCGv_i64, TCGv_i32);
|
||||||
#else
|
|
||||||
typedef void (*gen_atomic_cx_i32)(TCGv_i32, TCGv_env, TCGv, TCGv_i32, TCGv_i32);
|
|
||||||
typedef void (*gen_atomic_cx_i64)(TCGv_i64, TCGv_env, TCGv, TCGv_i64, TCGv_i64);
|
|
||||||
typedef void (*gen_atomic_op_i32)(TCGv_i32, TCGv_env, TCGv, TCGv_i32);
|
|
||||||
typedef void (*gen_atomic_op_i64)(TCGv_i64, TCGv_env, TCGv, TCGv_i64);
|
|
||||||
#endif
|
|
||||||
|
|
||||||
#ifdef CONFIG_ATOMIC64
|
#ifdef CONFIG_ATOMIC64
|
||||||
# define WITH_ATOMIC64(X) X,
|
# define WITH_ATOMIC64(X) X,
|
||||||
@ -3140,18 +3133,13 @@ void tcg_gen_atomic_cmpxchg_i32(TCGv_i32 retv, TCGv addr, TCGv_i32 cmpv,
|
|||||||
tcg_temp_free_i32(t1);
|
tcg_temp_free_i32(t1);
|
||||||
} else {
|
} else {
|
||||||
gen_atomic_cx_i32 gen;
|
gen_atomic_cx_i32 gen;
|
||||||
|
TCGMemOpIdx oi;
|
||||||
|
|
||||||
gen = table_cmpxchg[memop & (MO_SIZE | MO_BSWAP)];
|
gen = table_cmpxchg[memop & (MO_SIZE | MO_BSWAP)];
|
||||||
tcg_debug_assert(gen != NULL);
|
tcg_debug_assert(gen != NULL);
|
||||||
|
|
||||||
#ifdef CONFIG_SOFTMMU
|
oi = make_memop_idx(memop & ~MO_SIGN, idx);
|
||||||
{
|
gen(retv, cpu_env, addr, cmpv, newv, tcg_constant_i32(oi));
|
||||||
TCGMemOpIdx oi = make_memop_idx(memop & ~MO_SIGN, idx);
|
|
||||||
gen(retv, cpu_env, addr, cmpv, newv, tcg_constant_i32(oi));
|
|
||||||
}
|
|
||||||
#else
|
|
||||||
gen(retv, cpu_env, addr, cmpv, newv);
|
|
||||||
#endif
|
|
||||||
|
|
||||||
if (memop & MO_SIGN) {
|
if (memop & MO_SIGN) {
|
||||||
tcg_gen_ext_i32(retv, retv, memop);
|
tcg_gen_ext_i32(retv, retv, memop);
|
||||||
@ -3184,18 +3172,13 @@ void tcg_gen_atomic_cmpxchg_i64(TCGv_i64 retv, TCGv addr, TCGv_i64 cmpv,
|
|||||||
} else if ((memop & MO_SIZE) == MO_64) {
|
} else if ((memop & MO_SIZE) == MO_64) {
|
||||||
#ifdef CONFIG_ATOMIC64
|
#ifdef CONFIG_ATOMIC64
|
||||||
gen_atomic_cx_i64 gen;
|
gen_atomic_cx_i64 gen;
|
||||||
|
TCGMemOpIdx oi;
|
||||||
|
|
||||||
gen = table_cmpxchg[memop & (MO_SIZE | MO_BSWAP)];
|
gen = table_cmpxchg[memop & (MO_SIZE | MO_BSWAP)];
|
||||||
tcg_debug_assert(gen != NULL);
|
tcg_debug_assert(gen != NULL);
|
||||||
|
|
||||||
#ifdef CONFIG_SOFTMMU
|
oi = make_memop_idx(memop, idx);
|
||||||
{
|
gen(retv, cpu_env, addr, cmpv, newv, tcg_constant_i32(oi));
|
||||||
TCGMemOpIdx oi = make_memop_idx(memop, idx);
|
|
||||||
gen(retv, cpu_env, addr, cmpv, newv, tcg_constant_i32(oi));
|
|
||||||
}
|
|
||||||
#else
|
|
||||||
gen(retv, cpu_env, addr, cmpv, newv);
|
|
||||||
#endif
|
|
||||||
#else
|
#else
|
||||||
gen_helper_exit_atomic(cpu_env);
|
gen_helper_exit_atomic(cpu_env);
|
||||||
/* Produce a result, so that we have a well-formed opcode stream
|
/* Produce a result, so that we have a well-formed opcode stream
|
||||||
@ -3245,20 +3228,15 @@ static void do_atomic_op_i32(TCGv_i32 ret, TCGv addr, TCGv_i32 val,
|
|||||||
TCGArg idx, MemOp memop, void * const table[])
|
TCGArg idx, MemOp memop, void * const table[])
|
||||||
{
|
{
|
||||||
gen_atomic_op_i32 gen;
|
gen_atomic_op_i32 gen;
|
||||||
|
TCGMemOpIdx oi;
|
||||||
|
|
||||||
memop = tcg_canonicalize_memop(memop, 0, 0);
|
memop = tcg_canonicalize_memop(memop, 0, 0);
|
||||||
|
|
||||||
gen = table[memop & (MO_SIZE | MO_BSWAP)];
|
gen = table[memop & (MO_SIZE | MO_BSWAP)];
|
||||||
tcg_debug_assert(gen != NULL);
|
tcg_debug_assert(gen != NULL);
|
||||||
|
|
||||||
#ifdef CONFIG_SOFTMMU
|
oi = make_memop_idx(memop & ~MO_SIGN, idx);
|
||||||
{
|
gen(ret, cpu_env, addr, val, tcg_constant_i32(oi));
|
||||||
TCGMemOpIdx oi = make_memop_idx(memop & ~MO_SIGN, idx);
|
|
||||||
gen(ret, cpu_env, addr, val, tcg_constant_i32(oi));
|
|
||||||
}
|
|
||||||
#else
|
|
||||||
gen(ret, cpu_env, addr, val);
|
|
||||||
#endif
|
|
||||||
|
|
||||||
if (memop & MO_SIGN) {
|
if (memop & MO_SIGN) {
|
||||||
tcg_gen_ext_i32(ret, ret, memop);
|
tcg_gen_ext_i32(ret, ret, memop);
|
||||||
@ -3292,18 +3270,13 @@ static void do_atomic_op_i64(TCGv_i64 ret, TCGv addr, TCGv_i64 val,
|
|||||||
if ((memop & MO_SIZE) == MO_64) {
|
if ((memop & MO_SIZE) == MO_64) {
|
||||||
#ifdef CONFIG_ATOMIC64
|
#ifdef CONFIG_ATOMIC64
|
||||||
gen_atomic_op_i64 gen;
|
gen_atomic_op_i64 gen;
|
||||||
|
TCGMemOpIdx oi;
|
||||||
|
|
||||||
gen = table[memop & (MO_SIZE | MO_BSWAP)];
|
gen = table[memop & (MO_SIZE | MO_BSWAP)];
|
||||||
tcg_debug_assert(gen != NULL);
|
tcg_debug_assert(gen != NULL);
|
||||||
|
|
||||||
#ifdef CONFIG_SOFTMMU
|
oi = make_memop_idx(memop & ~MO_SIGN, idx);
|
||||||
{
|
gen(ret, cpu_env, addr, val, tcg_constant_i32(oi));
|
||||||
TCGMemOpIdx oi = make_memop_idx(memop & ~MO_SIGN, idx);
|
|
||||||
gen(ret, cpu_env, addr, val, tcg_constant_i32(oi));
|
|
||||||
}
|
|
||||||
#else
|
|
||||||
gen(ret, cpu_env, addr, val);
|
|
||||||
#endif
|
|
||||||
#else
|
#else
|
||||||
gen_helper_exit_atomic(cpu_env);
|
gen_helper_exit_atomic(cpu_env);
|
||||||
/* Produce a result, so that we have a well-formed opcode stream
|
/* Produce a result, so that we have a well-formed opcode stream
|
||||||
|
Loading…
Reference in New Issue
Block a user