962a145cdc
Almost all of the disas_log implementations are identical. Unify them within translator_loop. Drop extra Priv/Virt logging from target/riscv. Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
362 lines
9.3 KiB
C
362 lines
9.3 KiB
C
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
|
/*
|
|
* LoongArch emulation for QEMU - main translation routines.
|
|
*
|
|
* Copyright (c) 2021 Loongson Technology Corporation Limited
|
|
*/
|
|
|
|
#include "qemu/osdep.h"
|
|
#include "cpu.h"
|
|
#include "tcg/tcg-op.h"
|
|
#include "tcg/tcg-op-gvec.h"
|
|
#include "exec/translation-block.h"
|
|
#include "exec/translator.h"
|
|
#include "exec/helper-proto.h"
|
|
#include "exec/helper-gen.h"
|
|
#include "exec/log.h"
|
|
#include "qemu/qemu-print.h"
|
|
#include "fpu/softfloat.h"
|
|
#include "translate.h"
|
|
#include "internals.h"
|
|
#include "vec.h"
|
|
|
|
/* Global register indices */
|
|
TCGv cpu_gpr[32], cpu_pc;
|
|
static TCGv cpu_lladdr, cpu_llval;
|
|
|
|
#define HELPER_H "helper.h"
|
|
#include "exec/helper-info.c.inc"
|
|
#undef HELPER_H
|
|
|
|
#define DISAS_STOP DISAS_TARGET_0
|
|
#define DISAS_EXIT DISAS_TARGET_1
|
|
#define DISAS_EXIT_UPDATE DISAS_TARGET_2
|
|
|
|
static inline int vec_full_offset(int regno)
|
|
{
|
|
return offsetof(CPULoongArchState, fpr[regno]);
|
|
}
|
|
|
|
static inline int vec_reg_offset(int regno, int index, MemOp mop)
|
|
{
|
|
const uint8_t size = 1 << mop;
|
|
int offs = index * size;
|
|
|
|
if (HOST_BIG_ENDIAN && size < 8 ) {
|
|
offs ^= (8 - size);
|
|
}
|
|
|
|
return offs + vec_full_offset(regno);
|
|
}
|
|
|
|
static inline void get_vreg64(TCGv_i64 dest, int regno, int index)
|
|
{
|
|
tcg_gen_ld_i64(dest, tcg_env,
|
|
offsetof(CPULoongArchState, fpr[regno].vreg.D(index)));
|
|
}
|
|
|
|
static inline void set_vreg64(TCGv_i64 src, int regno, int index)
|
|
{
|
|
tcg_gen_st_i64(src, tcg_env,
|
|
offsetof(CPULoongArchState, fpr[regno].vreg.D(index)));
|
|
}
|
|
|
|
static inline int plus_1(DisasContext *ctx, int x)
|
|
{
|
|
return x + 1;
|
|
}
|
|
|
|
static inline int shl_1(DisasContext *ctx, int x)
|
|
{
|
|
return x << 1;
|
|
}
|
|
|
|
static inline int shl_2(DisasContext *ctx, int x)
|
|
{
|
|
return x << 2;
|
|
}
|
|
|
|
static inline int shl_3(DisasContext *ctx, int x)
|
|
{
|
|
return x << 3;
|
|
}
|
|
|
|
/*
|
|
* LoongArch the upper 32 bits are undefined ("can be any value").
|
|
* QEMU chooses to nanbox, because it is most likely to show guest bugs early.
|
|
*/
|
|
static void gen_nanbox_s(TCGv_i64 out, TCGv_i64 in)
|
|
{
|
|
tcg_gen_ori_i64(out, in, MAKE_64BIT_MASK(32, 32));
|
|
}
|
|
|
|
void generate_exception(DisasContext *ctx, int excp)
|
|
{
|
|
tcg_gen_movi_tl(cpu_pc, ctx->base.pc_next);
|
|
gen_helper_raise_exception(tcg_env, tcg_constant_i32(excp));
|
|
ctx->base.is_jmp = DISAS_NORETURN;
|
|
}
|
|
|
|
static inline void gen_goto_tb(DisasContext *ctx, int n, target_ulong dest)
|
|
{
|
|
if (ctx->va32) {
|
|
dest = (uint32_t) dest;
|
|
}
|
|
|
|
if (translator_use_goto_tb(&ctx->base, dest)) {
|
|
tcg_gen_goto_tb(n);
|
|
tcg_gen_movi_tl(cpu_pc, dest);
|
|
tcg_gen_exit_tb(ctx->base.tb, n);
|
|
} else {
|
|
tcg_gen_movi_tl(cpu_pc, dest);
|
|
tcg_gen_lookup_and_goto_ptr();
|
|
}
|
|
}
|
|
|
|
static void loongarch_tr_init_disas_context(DisasContextBase *dcbase,
|
|
CPUState *cs)
|
|
{
|
|
int64_t bound;
|
|
CPULoongArchState *env = cpu_env(cs);
|
|
DisasContext *ctx = container_of(dcbase, DisasContext, base);
|
|
|
|
ctx->page_start = ctx->base.pc_first & TARGET_PAGE_MASK;
|
|
ctx->plv = ctx->base.tb->flags & HW_FLAGS_PLV_MASK;
|
|
if (ctx->base.tb->flags & HW_FLAGS_CRMD_PG) {
|
|
ctx->mem_idx = ctx->plv;
|
|
} else {
|
|
ctx->mem_idx = MMU_DA_IDX;
|
|
}
|
|
|
|
/* Bound the number of insns to execute to those left on the page. */
|
|
bound = -(ctx->base.pc_first | TARGET_PAGE_MASK) / 4;
|
|
ctx->base.max_insns = MIN(ctx->base.max_insns, bound);
|
|
|
|
if (FIELD_EX64(env->cpucfg[2], CPUCFG2, LSX)) {
|
|
ctx->vl = LSX_LEN;
|
|
}
|
|
|
|
if (FIELD_EX64(env->cpucfg[2], CPUCFG2, LASX)) {
|
|
ctx->vl = LASX_LEN;
|
|
}
|
|
|
|
ctx->la64 = is_la64(env);
|
|
ctx->va32 = (ctx->base.tb->flags & HW_FLAGS_VA32) != 0;
|
|
|
|
ctx->zero = tcg_constant_tl(0);
|
|
|
|
ctx->cpucfg1 = env->cpucfg[1];
|
|
ctx->cpucfg2 = env->cpucfg[2];
|
|
}
|
|
|
|
static void loongarch_tr_tb_start(DisasContextBase *dcbase, CPUState *cs)
|
|
{
|
|
}
|
|
|
|
static void loongarch_tr_insn_start(DisasContextBase *dcbase, CPUState *cs)
|
|
{
|
|
DisasContext *ctx = container_of(dcbase, DisasContext, base);
|
|
|
|
tcg_gen_insn_start(ctx->base.pc_next);
|
|
}
|
|
|
|
/*
|
|
* Wrappers for getting reg values.
|
|
*
|
|
* The $zero register does not have cpu_gpr[0] allocated -- we supply the
|
|
* constant zero as a source, and an uninitialized sink as destination.
|
|
*
|
|
* Further, we may provide an extension for word operations.
|
|
*/
|
|
static TCGv gpr_src(DisasContext *ctx, int reg_num, DisasExtend src_ext)
|
|
{
|
|
TCGv t;
|
|
|
|
if (reg_num == 0) {
|
|
return ctx->zero;
|
|
}
|
|
|
|
switch (src_ext) {
|
|
case EXT_NONE:
|
|
return cpu_gpr[reg_num];
|
|
case EXT_SIGN:
|
|
t = tcg_temp_new();
|
|
tcg_gen_ext32s_tl(t, cpu_gpr[reg_num]);
|
|
return t;
|
|
case EXT_ZERO:
|
|
t = tcg_temp_new();
|
|
tcg_gen_ext32u_tl(t, cpu_gpr[reg_num]);
|
|
return t;
|
|
}
|
|
g_assert_not_reached();
|
|
}
|
|
|
|
static TCGv gpr_dst(DisasContext *ctx, int reg_num, DisasExtend dst_ext)
|
|
{
|
|
if (reg_num == 0 || dst_ext) {
|
|
return tcg_temp_new();
|
|
}
|
|
return cpu_gpr[reg_num];
|
|
}
|
|
|
|
static void gen_set_gpr(int reg_num, TCGv t, DisasExtend dst_ext)
|
|
{
|
|
if (reg_num != 0) {
|
|
switch (dst_ext) {
|
|
case EXT_NONE:
|
|
tcg_gen_mov_tl(cpu_gpr[reg_num], t);
|
|
break;
|
|
case EXT_SIGN:
|
|
tcg_gen_ext32s_tl(cpu_gpr[reg_num], t);
|
|
break;
|
|
case EXT_ZERO:
|
|
tcg_gen_ext32u_tl(cpu_gpr[reg_num], t);
|
|
break;
|
|
default:
|
|
g_assert_not_reached();
|
|
}
|
|
}
|
|
}
|
|
|
|
static TCGv get_fpr(DisasContext *ctx, int reg_num)
|
|
{
|
|
TCGv t = tcg_temp_new();
|
|
tcg_gen_ld_i64(t, tcg_env,
|
|
offsetof(CPULoongArchState, fpr[reg_num].vreg.D(0)));
|
|
return t;
|
|
}
|
|
|
|
static void set_fpr(int reg_num, TCGv val)
|
|
{
|
|
tcg_gen_st_i64(val, tcg_env,
|
|
offsetof(CPULoongArchState, fpr[reg_num].vreg.D(0)));
|
|
}
|
|
|
|
static TCGv make_address_x(DisasContext *ctx, TCGv base, TCGv addend)
|
|
{
|
|
TCGv temp = NULL;
|
|
|
|
if (addend || ctx->va32) {
|
|
temp = tcg_temp_new();
|
|
}
|
|
if (addend) {
|
|
tcg_gen_add_tl(temp, base, addend);
|
|
base = temp;
|
|
}
|
|
if (ctx->va32) {
|
|
tcg_gen_ext32u_tl(temp, base);
|
|
base = temp;
|
|
}
|
|
return base;
|
|
}
|
|
|
|
static TCGv make_address_i(DisasContext *ctx, TCGv base, target_long ofs)
|
|
{
|
|
TCGv addend = ofs ? tcg_constant_tl(ofs) : NULL;
|
|
return make_address_x(ctx, base, addend);
|
|
}
|
|
|
|
static uint64_t make_address_pc(DisasContext *ctx, uint64_t addr)
|
|
{
|
|
if (ctx->va32) {
|
|
addr = (int32_t)addr;
|
|
}
|
|
return addr;
|
|
}
|
|
|
|
#include "decode-insns.c.inc"
|
|
#include "insn_trans/trans_arith.c.inc"
|
|
#include "insn_trans/trans_shift.c.inc"
|
|
#include "insn_trans/trans_bit.c.inc"
|
|
#include "insn_trans/trans_memory.c.inc"
|
|
#include "insn_trans/trans_atomic.c.inc"
|
|
#include "insn_trans/trans_extra.c.inc"
|
|
#include "insn_trans/trans_farith.c.inc"
|
|
#include "insn_trans/trans_fcmp.c.inc"
|
|
#include "insn_trans/trans_fcnv.c.inc"
|
|
#include "insn_trans/trans_fmov.c.inc"
|
|
#include "insn_trans/trans_fmemory.c.inc"
|
|
#include "insn_trans/trans_branch.c.inc"
|
|
#include "insn_trans/trans_privileged.c.inc"
|
|
#include "insn_trans/trans_vec.c.inc"
|
|
|
|
static void loongarch_tr_translate_insn(DisasContextBase *dcbase, CPUState *cs)
|
|
{
|
|
DisasContext *ctx = container_of(dcbase, DisasContext, base);
|
|
|
|
ctx->opcode = translator_ldl(cpu_env(cs), &ctx->base, ctx->base.pc_next);
|
|
|
|
if (!decode(ctx, ctx->opcode)) {
|
|
qemu_log_mask(LOG_UNIMP, "Error: unknown opcode. "
|
|
TARGET_FMT_lx ": 0x%x\n",
|
|
ctx->base.pc_next, ctx->opcode);
|
|
generate_exception(ctx, EXCCODE_INE);
|
|
}
|
|
|
|
ctx->base.pc_next += 4;
|
|
|
|
if (ctx->va32) {
|
|
ctx->base.pc_next = (uint32_t)ctx->base.pc_next;
|
|
}
|
|
}
|
|
|
|
static void loongarch_tr_tb_stop(DisasContextBase *dcbase, CPUState *cs)
|
|
{
|
|
DisasContext *ctx = container_of(dcbase, DisasContext, base);
|
|
|
|
switch (ctx->base.is_jmp) {
|
|
case DISAS_STOP:
|
|
tcg_gen_movi_tl(cpu_pc, ctx->base.pc_next);
|
|
tcg_gen_lookup_and_goto_ptr();
|
|
break;
|
|
case DISAS_TOO_MANY:
|
|
gen_goto_tb(ctx, 0, ctx->base.pc_next);
|
|
break;
|
|
case DISAS_NORETURN:
|
|
break;
|
|
case DISAS_EXIT_UPDATE:
|
|
tcg_gen_movi_tl(cpu_pc, ctx->base.pc_next);
|
|
QEMU_FALLTHROUGH;
|
|
case DISAS_EXIT:
|
|
tcg_gen_exit_tb(NULL, 0);
|
|
break;
|
|
default:
|
|
g_assert_not_reached();
|
|
}
|
|
}
|
|
|
|
static const TranslatorOps loongarch_tr_ops = {
|
|
.init_disas_context = loongarch_tr_init_disas_context,
|
|
.tb_start = loongarch_tr_tb_start,
|
|
.insn_start = loongarch_tr_insn_start,
|
|
.translate_insn = loongarch_tr_translate_insn,
|
|
.tb_stop = loongarch_tr_tb_stop,
|
|
};
|
|
|
|
void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int *max_insns,
|
|
vaddr pc, void *host_pc)
|
|
{
|
|
DisasContext ctx;
|
|
|
|
translator_loop(cs, tb, max_insns, pc, host_pc,
|
|
&loongarch_tr_ops, &ctx.base);
|
|
}
|
|
|
|
void loongarch_translate_init(void)
|
|
{
|
|
int i;
|
|
|
|
cpu_gpr[0] = NULL;
|
|
for (i = 1; i < 32; i++) {
|
|
cpu_gpr[i] = tcg_global_mem_new(tcg_env,
|
|
offsetof(CPULoongArchState, gpr[i]),
|
|
regnames[i]);
|
|
}
|
|
|
|
cpu_pc = tcg_global_mem_new(tcg_env, offsetof(CPULoongArchState, pc), "pc");
|
|
cpu_lladdr = tcg_global_mem_new(tcg_env,
|
|
offsetof(CPULoongArchState, lladdr), "lladdr");
|
|
cpu_llval = tcg_global_mem_new(tcg_env,
|
|
offsetof(CPULoongArchState, llval), "llval");
|
|
}
|