target/arm: Implement LDG, STG, ST2G instructions
Reviewed-by: Peter Maydell <peter.maydell@linaro.org> Signed-off-by: Richard Henderson <richard.henderson@linaro.org> Message-id: 20200626033144.790098-16-richard.henderson@linaro.org Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
This commit is contained in:
parent
0d1762e931
commit
c15294c1e3
@ -106,3 +106,10 @@ DEF_HELPER_FLAGS_2(xpacd, TCG_CALL_NO_RWG_SE, i64, env, i64)
|
||||
|
||||
DEF_HELPER_FLAGS_3(irg, TCG_CALL_NO_RWG, i64, env, i64, i64)
|
||||
DEF_HELPER_FLAGS_4(addsubg, TCG_CALL_NO_RWG_SE, i64, env, i64, s32, i32)
|
||||
DEF_HELPER_FLAGS_3(ldg, TCG_CALL_NO_WG, i64, env, i64, i64)
|
||||
DEF_HELPER_FLAGS_3(stg, TCG_CALL_NO_WG, void, env, i64, i64)
|
||||
DEF_HELPER_FLAGS_3(stg_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
|
||||
DEF_HELPER_FLAGS_2(stg_stub, TCG_CALL_NO_WG, void, env, i64)
|
||||
DEF_HELPER_FLAGS_3(st2g, TCG_CALL_NO_WG, void, env, i64, i64)
|
||||
DEF_HELPER_FLAGS_3(st2g_parallel, TCG_CALL_NO_WG, void, env, i64, i64)
|
||||
DEF_HELPER_FLAGS_2(st2g_stub, TCG_CALL_NO_WG, void, env, i64)
|
||||
|
@ -96,6 +96,8 @@ DEF_HELPER_FLAGS_1(rebuild_hflags_a32_newel, TCG_CALL_NO_RWG, void, env)
|
||||
DEF_HELPER_FLAGS_2(rebuild_hflags_a32, TCG_CALL_NO_RWG, void, env, int)
|
||||
DEF_HELPER_FLAGS_2(rebuild_hflags_a64, TCG_CALL_NO_RWG, void, env, int)
|
||||
|
||||
DEF_HELPER_FLAGS_5(probe_access, TCG_CALL_NO_WG, void, env, tl, i32, i32, i32)
|
||||
|
||||
DEF_HELPER_1(vfp_get_fpscr, i32, env)
|
||||
DEF_HELPER_2(vfp_set_fpscr, void, env, i32)
|
||||
|
||||
|
@ -44,6 +44,40 @@ static int choose_nonexcluded_tag(int tag, int offset, uint16_t exclude)
|
||||
return tag;
|
||||
}
|
||||
|
||||
/**
|
||||
* allocation_tag_mem:
|
||||
* @env: the cpu environment
|
||||
* @ptr_mmu_idx: the addressing regime to use for the virtual address
|
||||
* @ptr: the virtual address for which to look up tag memory
|
||||
* @ptr_access: the access to use for the virtual address
|
||||
* @ptr_size: the number of bytes in the normal memory access
|
||||
* @tag_access: the access to use for the tag memory
|
||||
* @tag_size: the number of bytes in the tag memory access
|
||||
* @ra: the return address for exception handling
|
||||
*
|
||||
* Our tag memory is formatted as a sequence of little-endian nibbles.
|
||||
* That is, the byte at (addr >> (LOG2_TAG_GRANULE + 1)) contains two
|
||||
* tags, with the tag at [3:0] for the lower addr and the tag at [7:4]
|
||||
* for the higher addr.
|
||||
*
|
||||
* Here, resolve the physical address from the virtual address, and return
|
||||
* a pointer to the corresponding tag byte. Exit with exception if the
|
||||
* virtual address is not accessible for @ptr_access.
|
||||
*
|
||||
* The @ptr_size and @tag_size values may not have an obvious relation
|
||||
* due to the alignment of @ptr, and the number of tag checks required.
|
||||
*
|
||||
* If there is no tag storage corresponding to @ptr, return NULL.
|
||||
*/
|
||||
static uint8_t *allocation_tag_mem(CPUARMState *env, int ptr_mmu_idx,
|
||||
uint64_t ptr, MMUAccessType ptr_access,
|
||||
int ptr_size, MMUAccessType tag_access,
|
||||
int tag_size, uintptr_t ra)
|
||||
{
|
||||
/* Tag storage not implemented. */
|
||||
return NULL;
|
||||
}
|
||||
|
||||
uint64_t HELPER(irg)(CPUARMState *env, uint64_t rn, uint64_t rm)
|
||||
{
|
||||
int rtag;
|
||||
@ -80,3 +114,163 @@ uint64_t HELPER(addsubg)(CPUARMState *env, uint64_t ptr,
|
||||
|
||||
return address_with_allocation_tag(ptr + offset, rtag);
|
||||
}
|
||||
|
||||
static int load_tag1(uint64_t ptr, uint8_t *mem)
|
||||
{
|
||||
int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
|
||||
return extract32(*mem, ofs, 4);
|
||||
}
|
||||
|
||||
uint64_t HELPER(ldg)(CPUARMState *env, uint64_t ptr, uint64_t xt)
|
||||
{
|
||||
int mmu_idx = cpu_mmu_index(env, false);
|
||||
uint8_t *mem;
|
||||
int rtag = 0;
|
||||
|
||||
/* Trap if accessing an invalid page. */
|
||||
mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_LOAD, 1,
|
||||
MMU_DATA_LOAD, 1, GETPC());
|
||||
|
||||
/* Load if page supports tags. */
|
||||
if (mem) {
|
||||
rtag = load_tag1(ptr, mem);
|
||||
}
|
||||
|
||||
return address_with_allocation_tag(xt, rtag);
|
||||
}
|
||||
|
||||
static void check_tag_aligned(CPUARMState *env, uint64_t ptr, uintptr_t ra)
|
||||
{
|
||||
if (unlikely(!QEMU_IS_ALIGNED(ptr, TAG_GRANULE))) {
|
||||
arm_cpu_do_unaligned_access(env_cpu(env), ptr, MMU_DATA_STORE,
|
||||
cpu_mmu_index(env, false), ra);
|
||||
g_assert_not_reached();
|
||||
}
|
||||
}
|
||||
|
||||
/* For use in a non-parallel context, store to the given nibble. */
|
||||
static void store_tag1(uint64_t ptr, uint8_t *mem, int tag)
|
||||
{
|
||||
int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
|
||||
*mem = deposit32(*mem, ofs, 4, tag);
|
||||
}
|
||||
|
||||
/* For use in a parallel context, atomically store to the given nibble. */
|
||||
static void store_tag1_parallel(uint64_t ptr, uint8_t *mem, int tag)
|
||||
{
|
||||
int ofs = extract32(ptr, LOG2_TAG_GRANULE, 1) * 4;
|
||||
uint8_t old = atomic_read(mem);
|
||||
|
||||
while (1) {
|
||||
uint8_t new = deposit32(old, ofs, 4, tag);
|
||||
uint8_t cmp = atomic_cmpxchg(mem, old, new);
|
||||
if (likely(cmp == old)) {
|
||||
return;
|
||||
}
|
||||
old = cmp;
|
||||
}
|
||||
}
|
||||
|
||||
typedef void stg_store1(uint64_t, uint8_t *, int);
|
||||
|
||||
static inline void do_stg(CPUARMState *env, uint64_t ptr, uint64_t xt,
|
||||
uintptr_t ra, stg_store1 store1)
|
||||
{
|
||||
int mmu_idx = cpu_mmu_index(env, false);
|
||||
uint8_t *mem;
|
||||
|
||||
check_tag_aligned(env, ptr, ra);
|
||||
|
||||
/* Trap if accessing an invalid page. */
|
||||
mem = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE, TAG_GRANULE,
|
||||
MMU_DATA_STORE, 1, ra);
|
||||
|
||||
/* Store if page supports tags. */
|
||||
if (mem) {
|
||||
store1(ptr, mem, allocation_tag_from_addr(xt));
|
||||
}
|
||||
}
|
||||
|
||||
void HELPER(stg)(CPUARMState *env, uint64_t ptr, uint64_t xt)
|
||||
{
|
||||
do_stg(env, ptr, xt, GETPC(), store_tag1);
|
||||
}
|
||||
|
||||
void HELPER(stg_parallel)(CPUARMState *env, uint64_t ptr, uint64_t xt)
|
||||
{
|
||||
do_stg(env, ptr, xt, GETPC(), store_tag1_parallel);
|
||||
}
|
||||
|
||||
void HELPER(stg_stub)(CPUARMState *env, uint64_t ptr)
|
||||
{
|
||||
int mmu_idx = cpu_mmu_index(env, false);
|
||||
uintptr_t ra = GETPC();
|
||||
|
||||
check_tag_aligned(env, ptr, ra);
|
||||
probe_write(env, ptr, TAG_GRANULE, mmu_idx, ra);
|
||||
}
|
||||
|
||||
static inline void do_st2g(CPUARMState *env, uint64_t ptr, uint64_t xt,
|
||||
uintptr_t ra, stg_store1 store1)
|
||||
{
|
||||
int mmu_idx = cpu_mmu_index(env, false);
|
||||
int tag = allocation_tag_from_addr(xt);
|
||||
uint8_t *mem1, *mem2;
|
||||
|
||||
check_tag_aligned(env, ptr, ra);
|
||||
|
||||
/*
|
||||
* Trap if accessing an invalid page(s).
|
||||
* This takes priority over !allocation_tag_access_enabled.
|
||||
*/
|
||||
if (ptr & TAG_GRANULE) {
|
||||
/* Two stores unaligned mod TAG_GRANULE*2 -- modify two bytes. */
|
||||
mem1 = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
|
||||
TAG_GRANULE, MMU_DATA_STORE, 1, ra);
|
||||
mem2 = allocation_tag_mem(env, mmu_idx, ptr + TAG_GRANULE,
|
||||
MMU_DATA_STORE, TAG_GRANULE,
|
||||
MMU_DATA_STORE, 1, ra);
|
||||
|
||||
/* Store if page(s) support tags. */
|
||||
if (mem1) {
|
||||
store1(TAG_GRANULE, mem1, tag);
|
||||
}
|
||||
if (mem2) {
|
||||
store1(0, mem2, tag);
|
||||
}
|
||||
} else {
|
||||
/* Two stores aligned mod TAG_GRANULE*2 -- modify one byte. */
|
||||
mem1 = allocation_tag_mem(env, mmu_idx, ptr, MMU_DATA_STORE,
|
||||
2 * TAG_GRANULE, MMU_DATA_STORE, 1, ra);
|
||||
if (mem1) {
|
||||
tag |= tag << 4;
|
||||
atomic_set(mem1, tag);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
void HELPER(st2g)(CPUARMState *env, uint64_t ptr, uint64_t xt)
|
||||
{
|
||||
do_st2g(env, ptr, xt, GETPC(), store_tag1);
|
||||
}
|
||||
|
||||
void HELPER(st2g_parallel)(CPUARMState *env, uint64_t ptr, uint64_t xt)
|
||||
{
|
||||
do_st2g(env, ptr, xt, GETPC(), store_tag1_parallel);
|
||||
}
|
||||
|
||||
void HELPER(st2g_stub)(CPUARMState *env, uint64_t ptr)
|
||||
{
|
||||
int mmu_idx = cpu_mmu_index(env, false);
|
||||
uintptr_t ra = GETPC();
|
||||
int in_page = -(ptr | TARGET_PAGE_MASK);
|
||||
|
||||
check_tag_aligned(env, ptr, ra);
|
||||
|
||||
if (likely(in_page >= 2 * TAG_GRANULE)) {
|
||||
probe_write(env, ptr, 2 * TAG_GRANULE, mmu_idx, ra);
|
||||
} else {
|
||||
probe_write(env, ptr, TAG_GRANULE, mmu_idx, ra);
|
||||
probe_write(env, ptr + TAG_GRANULE, TAG_GRANULE, mmu_idx, ra);
|
||||
}
|
||||
}
|
||||
|
@ -935,3 +935,19 @@ uint32_t HELPER(ror_cc)(CPUARMState *env, uint32_t x, uint32_t i)
|
||||
return ((uint32_t)x >> shift) | (x << (32 - shift));
|
||||
}
|
||||
}
|
||||
|
||||
void HELPER(probe_access)(CPUARMState *env, target_ulong ptr,
|
||||
uint32_t access_type, uint32_t mmu_idx,
|
||||
uint32_t size)
|
||||
{
|
||||
uint32_t in_page = -((uint32_t)ptr | TARGET_PAGE_SIZE);
|
||||
uintptr_t ra = GETPC();
|
||||
|
||||
if (likely(size <= in_page)) {
|
||||
probe_access(env, ptr, size, access_type, mmu_idx, ra);
|
||||
} else {
|
||||
probe_access(env, ptr, in_page, access_type, mmu_idx, ra);
|
||||
probe_access(env, ptr + in_page, size - in_page,
|
||||
access_type, mmu_idx, ra);
|
||||
}
|
||||
}
|
||||
|
@ -232,6 +232,19 @@ static void gen_address_with_allocation_tag0(TCGv_i64 dst, TCGv_i64 src)
|
||||
tcg_gen_andi_i64(dst, src, ~MAKE_64BIT_MASK(56, 4));
|
||||
}
|
||||
|
||||
static void gen_probe_access(DisasContext *s, TCGv_i64 ptr,
|
||||
MMUAccessType acc, int log2_size)
|
||||
{
|
||||
TCGv_i32 t_acc = tcg_const_i32(acc);
|
||||
TCGv_i32 t_idx = tcg_const_i32(get_mem_index(s));
|
||||
TCGv_i32 t_size = tcg_const_i32(1 << log2_size);
|
||||
|
||||
gen_helper_probe_access(cpu_env, ptr, t_acc, t_idx, t_size);
|
||||
tcg_temp_free_i32(t_acc);
|
||||
tcg_temp_free_i32(t_idx);
|
||||
tcg_temp_free_i32(t_size);
|
||||
}
|
||||
|
||||
typedef struct DisasCompare64 {
|
||||
TCGCond cond;
|
||||
TCGv_i64 value;
|
||||
@ -3685,6 +3698,154 @@ static void disas_ldst_single_struct(DisasContext *s, uint32_t insn)
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* Load/Store memory tags
|
||||
*
|
||||
* 31 30 29 24 22 21 12 10 5 0
|
||||
* +-----+-------------+-----+---+------+-----+------+------+
|
||||
* | 1 1 | 0 1 1 0 0 1 | op1 | 1 | imm9 | op2 | Rn | Rt |
|
||||
* +-----+-------------+-----+---+------+-----+------+------+
|
||||
*/
|
||||
static void disas_ldst_tag(DisasContext *s, uint32_t insn)
|
||||
{
|
||||
int rt = extract32(insn, 0, 5);
|
||||
int rn = extract32(insn, 5, 5);
|
||||
uint64_t offset = sextract64(insn, 12, 9) << LOG2_TAG_GRANULE;
|
||||
int op2 = extract32(insn, 10, 2);
|
||||
int op1 = extract32(insn, 22, 2);
|
||||
bool is_load = false, is_pair = false, is_zero = false;
|
||||
int index = 0;
|
||||
TCGv_i64 addr, clean_addr, tcg_rt;
|
||||
|
||||
/* We checked insn bits [29:24,21] in the caller. */
|
||||
if (extract32(insn, 30, 2) != 3) {
|
||||
goto do_unallocated;
|
||||
}
|
||||
|
||||
/*
|
||||
* @index is a tri-state variable which has 3 states:
|
||||
* < 0 : post-index, writeback
|
||||
* = 0 : signed offset
|
||||
* > 0 : pre-index, writeback
|
||||
*/
|
||||
switch (op1) {
|
||||
case 0:
|
||||
if (op2 != 0) {
|
||||
/* STG */
|
||||
index = op2 - 2;
|
||||
break;
|
||||
}
|
||||
goto do_unallocated;
|
||||
case 1:
|
||||
if (op2 != 0) {
|
||||
/* STZG */
|
||||
is_zero = true;
|
||||
index = op2 - 2;
|
||||
} else {
|
||||
/* LDG */
|
||||
is_load = true;
|
||||
}
|
||||
break;
|
||||
case 2:
|
||||
if (op2 != 0) {
|
||||
/* ST2G */
|
||||
is_pair = true;
|
||||
index = op2 - 2;
|
||||
break;
|
||||
}
|
||||
goto do_unallocated;
|
||||
case 3:
|
||||
if (op2 != 0) {
|
||||
/* STZ2G */
|
||||
is_pair = is_zero = true;
|
||||
index = op2 - 2;
|
||||
break;
|
||||
}
|
||||
goto do_unallocated;
|
||||
|
||||
default:
|
||||
do_unallocated:
|
||||
unallocated_encoding(s);
|
||||
return;
|
||||
}
|
||||
|
||||
if (!dc_isar_feature(aa64_mte_insn_reg, s)) {
|
||||
goto do_unallocated;
|
||||
}
|
||||
|
||||
if (rn == 31) {
|
||||
gen_check_sp_alignment(s);
|
||||
}
|
||||
|
||||
addr = read_cpu_reg_sp(s, rn, true);
|
||||
if (index >= 0) {
|
||||
/* pre-index or signed offset */
|
||||
tcg_gen_addi_i64(addr, addr, offset);
|
||||
}
|
||||
|
||||
if (is_load) {
|
||||
tcg_gen_andi_i64(addr, addr, -TAG_GRANULE);
|
||||
tcg_rt = cpu_reg(s, rt);
|
||||
if (s->ata) {
|
||||
gen_helper_ldg(tcg_rt, cpu_env, addr, tcg_rt);
|
||||
} else {
|
||||
clean_addr = clean_data_tbi(s, addr);
|
||||
gen_probe_access(s, clean_addr, MMU_DATA_LOAD, MO_8);
|
||||
gen_address_with_allocation_tag0(tcg_rt, addr);
|
||||
}
|
||||
} else {
|
||||
tcg_rt = cpu_reg_sp(s, rt);
|
||||
if (!s->ata) {
|
||||
/*
|
||||
* For STG and ST2G, we need to check alignment and probe memory.
|
||||
* TODO: For STZG and STZ2G, we could rely on the stores below,
|
||||
* at least for system mode; user-only won't enforce alignment.
|
||||
*/
|
||||
if (is_pair) {
|
||||
gen_helper_st2g_stub(cpu_env, addr);
|
||||
} else {
|
||||
gen_helper_stg_stub(cpu_env, addr);
|
||||
}
|
||||
} else if (tb_cflags(s->base.tb) & CF_PARALLEL) {
|
||||
if (is_pair) {
|
||||
gen_helper_st2g_parallel(cpu_env, addr, tcg_rt);
|
||||
} else {
|
||||
gen_helper_stg_parallel(cpu_env, addr, tcg_rt);
|
||||
}
|
||||
} else {
|
||||
if (is_pair) {
|
||||
gen_helper_st2g(cpu_env, addr, tcg_rt);
|
||||
} else {
|
||||
gen_helper_stg(cpu_env, addr, tcg_rt);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if (is_zero) {
|
||||
TCGv_i64 clean_addr = clean_data_tbi(s, addr);
|
||||
TCGv_i64 tcg_zero = tcg_const_i64(0);
|
||||
int mem_index = get_mem_index(s);
|
||||
int i, n = (1 + is_pair) << LOG2_TAG_GRANULE;
|
||||
|
||||
tcg_gen_qemu_st_i64(tcg_zero, clean_addr, mem_index,
|
||||
MO_Q | MO_ALIGN_16);
|
||||
for (i = 8; i < n; i += 8) {
|
||||
tcg_gen_addi_i64(clean_addr, clean_addr, 8);
|
||||
tcg_gen_qemu_st_i64(tcg_zero, clean_addr, mem_index, MO_Q);
|
||||
}
|
||||
tcg_temp_free_i64(tcg_zero);
|
||||
}
|
||||
|
||||
if (index != 0) {
|
||||
/* pre-index or post-index */
|
||||
if (index < 0) {
|
||||
/* post-index */
|
||||
tcg_gen_addi_i64(addr, addr, offset);
|
||||
}
|
||||
tcg_gen_mov_i64(cpu_reg_sp(s, rn), addr);
|
||||
}
|
||||
}
|
||||
|
||||
/* Loads and stores */
|
||||
static void disas_ldst(DisasContext *s, uint32_t insn)
|
||||
{
|
||||
@ -3709,13 +3870,14 @@ static void disas_ldst(DisasContext *s, uint32_t insn)
|
||||
case 0x0d: /* AdvSIMD load/store single structure */
|
||||
disas_ldst_single_struct(s, insn);
|
||||
break;
|
||||
case 0x19: /* LDAPR/STLR (unscaled immediate) */
|
||||
if (extract32(insn, 10, 2) != 0 ||
|
||||
extract32(insn, 21, 1) != 0) {
|
||||
case 0x19:
|
||||
if (extract32(insn, 21, 1) != 0) {
|
||||
disas_ldst_tag(s, insn);
|
||||
} else if (extract32(insn, 10, 2) == 0) {
|
||||
disas_ldst_ldapr_stlr(s, insn);
|
||||
} else {
|
||||
unallocated_encoding(s);
|
||||
break;
|
||||
}
|
||||
disas_ldst_ldapr_stlr(s, insn);
|
||||
break;
|
||||
default:
|
||||
unallocated_encoding(s);
|
||||
|
Loading…
Reference in New Issue
Block a user