qemu/target/tilegx/translate.c
Richard Henderson 8b86d6d258 tcg: Hoist max_insns computation to tb_gen_code
In order to handle TB's that translate to too much code, we
need to place the control of the length of the translation
in the hands of the code gen master loop.

Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
Reviewed-by: Philippe Mathieu-Daudé <philmd@redhat.com>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
2019-04-24 13:04:33 -07:00

2444 lines
76 KiB
C

/*
* QEMU TILE-Gx CPU
*
* Copyright (c) 2015 Chen Gang
*
* This library is free software; you can redistribute it and/or
* modify it under the terms of the GNU Lesser General Public
* License as published by the Free Software Foundation; either
* version 2.1 of the License, or (at your option) any later version.
*
* This library is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* Lesser General Public License for more details.
*
* You should have received a copy of the GNU Lesser General Public
* License along with this library; if not, see
* <http://www.gnu.org/licenses/lgpl-2.1.html>
*/
#include "qemu/osdep.h"
#include "cpu.h"
#include "qemu/log.h"
#include "exec/log.h"
#include "disas/disas.h"
#include "exec/exec-all.h"
#include "tcg-op.h"
#include "exec/cpu_ldst.h"
#include "linux-user/syscall_defs.h"
#include "opcode_tilegx.h"
#include "spr_def_64.h"
#define FMT64X "%016" PRIx64
static TCGv cpu_pc;
static TCGv cpu_regs[TILEGX_R_COUNT];
static const char * const reg_names[64] = {
"r0", "r1", "r2", "r3", "r4", "r5", "r6", "r7",
"r8", "r9", "r10", "r11", "r12", "r13", "r14", "r15",
"r16", "r17", "r18", "r19", "r20", "r21", "r22", "r23",
"r24", "r25", "r26", "r27", "r28", "r29", "r30", "r31",
"r32", "r33", "r34", "r35", "r36", "r37", "r38", "r39",
"r40", "r41", "r42", "r43", "r44", "r45", "r46", "r47",
"r48", "r49", "r50", "r51", "bp", "tp", "sp", "lr",
"sn", "idn0", "idn1", "udn0", "udn1", "udn2", "udn2", "zero"
};
/* Modified registers are cached in temporaries until the end of the bundle. */
typedef struct {
unsigned reg;
TCGv val;
} DisasContextTemp;
#define MAX_WRITEBACK 4
/* This is the state at translation time. */
typedef struct {
uint64_t pc; /* Current pc */
TCGv zero; /* For zero register */
DisasContextTemp wb[MAX_WRITEBACK];
int num_wb;
int mmuidx;
bool exit_tb;
TileExcp atomic_excp;
struct {
TCGCond cond; /* branch condition */
TCGv dest; /* branch destination */
TCGv val1; /* value to be compared against zero, for cond */
} jmp; /* Jump object, only once in each TB block */
} DisasContext;
#include "exec/gen-icount.h"
/* Differentiate the various pipe encodings. */
#define TY_X0 0
#define TY_X1 1
#define TY_Y0 2
#define TY_Y1 3
/* Remerge the base opcode and extension fields for switching.
The X opcode fields are 3 bits; Y0/Y1 opcode fields are 4 bits;
Y2 opcode field is 2 bits. */
#define OE(OP, EXT, XY) (TY_##XY + OP * 4 + EXT * 64)
/* Similar, but for Y2 only. */
#define OEY2(OP, MODE) (OP + MODE * 4)
/* Similar, but make sure opcode names match up. */
#define OE_RR_X0(E) OE(RRR_0_OPCODE_X0, E##_UNARY_OPCODE_X0, X0)
#define OE_RR_X1(E) OE(RRR_0_OPCODE_X1, E##_UNARY_OPCODE_X1, X1)
#define OE_RR_Y0(E) OE(RRR_1_OPCODE_Y0, E##_UNARY_OPCODE_Y0, Y0)
#define OE_RR_Y1(E) OE(RRR_1_OPCODE_Y1, E##_UNARY_OPCODE_Y1, Y1)
#define OE_RRR(E,N,XY) OE(RRR_##N##_OPCODE_##XY, E##_RRR_##N##_OPCODE_##XY, XY)
#define OE_IM(E,XY) OE(IMM8_OPCODE_##XY, E##_IMM8_OPCODE_##XY, XY)
#define OE_SH(E,XY) OE(SHIFT_OPCODE_##XY, E##_SHIFT_OPCODE_##XY, XY)
#define V1_IMM(X) (((X) & 0xff) * 0x0101010101010101ull)
#define V2_IMM(X) (((X) & 0xffff) * 0x0001000100010001ull)
static void gen_exception(DisasContext *dc, TileExcp num)
{
TCGv_i32 tmp;
tcg_gen_movi_tl(cpu_pc, dc->pc + TILEGX_BUNDLE_SIZE_IN_BYTES);
tmp = tcg_const_i32(num);
gen_helper_exception(cpu_env, tmp);
tcg_temp_free_i32(tmp);
dc->exit_tb = true;
}
static bool check_gr(DisasContext *dc, uint8_t reg)
{
if (likely(reg < TILEGX_R_COUNT)) {
return true;
}
switch (reg) {
case TILEGX_R_SN:
case TILEGX_R_ZERO:
break;
case TILEGX_R_IDN0:
case TILEGX_R_IDN1:
gen_exception(dc, TILEGX_EXCP_REG_IDN_ACCESS);
break;
case TILEGX_R_UDN0:
case TILEGX_R_UDN1:
case TILEGX_R_UDN2:
case TILEGX_R_UDN3:
gen_exception(dc, TILEGX_EXCP_REG_UDN_ACCESS);
break;
default:
g_assert_not_reached();
}
return false;
}
static TCGv load_zero(DisasContext *dc)
{
if (!dc->zero) {
dc->zero = tcg_const_i64(0);
}
return dc->zero;
}
static TCGv load_gr(DisasContext *dc, unsigned reg)
{
if (check_gr(dc, reg)) {
return cpu_regs[reg];
}
return load_zero(dc);
}
static TCGv dest_gr(DisasContext *dc, unsigned reg)
{
int n;
/* Skip the result, mark the exception if necessary, and continue */
check_gr(dc, reg);
n = dc->num_wb++;
dc->wb[n].reg = reg;
return dc->wb[n].val = tcg_temp_new_i64();
}
static void gen_saturate_op(TCGv tdest, TCGv tsrca, TCGv tsrcb,
void (*operate)(TCGv, TCGv, TCGv))
{
TCGv t0 = tcg_temp_new();
tcg_gen_ext32s_tl(tdest, tsrca);
tcg_gen_ext32s_tl(t0, tsrcb);
operate(tdest, tdest, t0);
tcg_gen_movi_tl(t0, 0x7fffffff);
tcg_gen_movcond_tl(TCG_COND_GT, tdest, tdest, t0, t0, tdest);
tcg_gen_movi_tl(t0, -0x80000000LL);
tcg_gen_movcond_tl(TCG_COND_LT, tdest, tdest, t0, t0, tdest);
tcg_temp_free(t0);
}
static void gen_atomic_excp(DisasContext *dc, unsigned dest, TCGv tdest,
TCGv tsrca, TCGv tsrcb, TileExcp excp)
{
#ifdef CONFIG_USER_ONLY
TCGv_i32 t;
tcg_gen_st_tl(tsrca, cpu_env, offsetof(CPUTLGState, atomic_srca));
tcg_gen_st_tl(tsrcb, cpu_env, offsetof(CPUTLGState, atomic_srcb));
t = tcg_const_i32(dest);
tcg_gen_st_i32(t, cpu_env, offsetof(CPUTLGState, atomic_dstr));
tcg_temp_free_i32(t);
/* We're going to write the real result in the exception. But in
the meantime we've already created a writeback register, and
we don't want that to remain uninitialized. */
tcg_gen_movi_tl(tdest, 0);
/* Note that we need to delay issuing the exception that implements
the atomic operation until after writing back the results of the
instruction occupying the X0 pipe. */
dc->atomic_excp = excp;
#else
gen_exception(dc, TILEGX_EXCP_OPCODE_UNIMPLEMENTED);
#endif
}
/* Shift the 128-bit value TSRCA:TSRCD right by the number of bytes
specified by the bottom 3 bits of TSRCB, and set TDEST to the
low 64 bits of the resulting value. */
static void gen_dblalign(TCGv tdest, TCGv tsrcd, TCGv tsrca, TCGv tsrcb)
{
TCGv t0 = tcg_temp_new();
tcg_gen_andi_tl(t0, tsrcb, 7);
tcg_gen_shli_tl(t0, t0, 3);
tcg_gen_shr_tl(tdest, tsrcd, t0);
/* We want to do "t0 = tsrca << (64 - t0)". Two's complement
arithmetic on a 6-bit field tells us that 64 - t0 is equal
to (t0 ^ 63) + 1. So we can do the shift in two parts,
neither of which will be an invalid shift by 64. */
tcg_gen_xori_tl(t0, t0, 63);
tcg_gen_shl_tl(t0, tsrca, t0);
tcg_gen_shli_tl(t0, t0, 1);
tcg_gen_or_tl(tdest, tdest, t0);
tcg_temp_free(t0);
}
/* Similarly, except that the 128-bit value is TSRCA:TSRCB, and the
right shift is an immediate. */
static void gen_dblaligni(TCGv tdest, TCGv tsrca, TCGv tsrcb, int shr)
{
TCGv t0 = tcg_temp_new();
tcg_gen_shri_tl(t0, tsrcb, shr);
tcg_gen_shli_tl(tdest, tsrca, 64 - shr);
tcg_gen_or_tl(tdest, tdest, t0);
tcg_temp_free(t0);
}
typedef enum {
LU, LS, HU, HS
} MulHalf;
static void gen_ext_half(TCGv d, TCGv s, MulHalf h)
{
switch (h) {
case LU:
tcg_gen_ext32u_tl(d, s);
break;
case LS:
tcg_gen_ext32s_tl(d, s);
break;
case HU:
tcg_gen_shri_tl(d, s, 32);
break;
case HS:
tcg_gen_sari_tl(d, s, 32);
break;
}
}
static void gen_mul_half(TCGv tdest, TCGv tsrca, TCGv tsrcb,
MulHalf ha, MulHalf hb)
{
TCGv t = tcg_temp_new();
gen_ext_half(t, tsrca, ha);
gen_ext_half(tdest, tsrcb, hb);
tcg_gen_mul_tl(tdest, tdest, t);
tcg_temp_free(t);
}
static void gen_cmul2(TCGv tdest, TCGv tsrca, TCGv tsrcb, int sh, int rd)
{
TCGv_i32 tsh = tcg_const_i32(sh);
TCGv_i32 trd = tcg_const_i32(rd);
gen_helper_cmul2(tdest, tsrca, tsrcb, tsh, trd);
tcg_temp_free_i32(tsh);
tcg_temp_free_i32(trd);
}
static TileExcp gen_st_opcode(DisasContext *dc, unsigned dest, unsigned srca,
unsigned srcb, TCGMemOp memop, const char *name)
{
if (dest) {
return TILEGX_EXCP_OPCODE_UNKNOWN;
}
tcg_gen_qemu_st_tl(load_gr(dc, srcb), load_gr(dc, srca),
dc->mmuidx, memop);
qemu_log_mask(CPU_LOG_TB_IN_ASM, "%s %s, %s", name,
reg_names[srca], reg_names[srcb]);
return TILEGX_EXCP_NONE;
}
static TileExcp gen_st_add_opcode(DisasContext *dc, unsigned srca, unsigned srcb,
int imm, TCGMemOp memop, const char *name)
{
TCGv tsrca = load_gr(dc, srca);
TCGv tsrcb = load_gr(dc, srcb);
tcg_gen_qemu_st_tl(tsrcb, tsrca, dc->mmuidx, memop);
tcg_gen_addi_tl(dest_gr(dc, srca), tsrca, imm);
qemu_log_mask(CPU_LOG_TB_IN_ASM, "%s %s, %s, %d", name,
reg_names[srca], reg_names[srcb], imm);
return TILEGX_EXCP_NONE;
}
/* Equality comparison with zero can be done quickly and efficiently. */
static void gen_v1cmpeq0(TCGv v)
{
TCGv m = tcg_const_tl(V1_IMM(0x7f));
TCGv c = tcg_temp_new();
/* ~(((v & m) + m) | m | v). Sets the msb for each byte == 0. */
tcg_gen_and_tl(c, v, m);
tcg_gen_add_tl(c, c, m);
tcg_gen_or_tl(c, c, m);
tcg_gen_nor_tl(c, c, v);
tcg_temp_free(m);
/* Shift the msb down to form the lsb boolean result. */
tcg_gen_shri_tl(v, c, 7);
tcg_temp_free(c);
}
static void gen_v1cmpne0(TCGv v)
{
TCGv m = tcg_const_tl(V1_IMM(0x7f));
TCGv c = tcg_temp_new();
/* (((v & m) + m) | v) & ~m. Sets the msb for each byte != 0. */
tcg_gen_and_tl(c, v, m);
tcg_gen_add_tl(c, c, m);
tcg_gen_or_tl(c, c, v);
tcg_gen_andc_tl(c, c, m);
tcg_temp_free(m);
/* Shift the msb down to form the lsb boolean result. */
tcg_gen_shri_tl(v, c, 7);
tcg_temp_free(c);
}
/* Vector addition can be performed via arithmetic plus masking. It is
efficient this way only for 4 or more elements. */
static void gen_v12add(TCGv tdest, TCGv tsrca, TCGv tsrcb, uint64_t sign)
{
TCGv tmask = tcg_const_tl(~sign);
TCGv t0 = tcg_temp_new();
TCGv t1 = tcg_temp_new();
/* ((a & ~sign) + (b & ~sign)) ^ ((a ^ b) & sign). */
tcg_gen_and_tl(t0, tsrca, tmask);
tcg_gen_and_tl(t1, tsrcb, tmask);
tcg_gen_add_tl(tdest, t0, t1);
tcg_gen_xor_tl(t0, tsrca, tsrcb);
tcg_gen_andc_tl(t0, t0, tmask);
tcg_gen_xor_tl(tdest, tdest, t0);
tcg_temp_free(t1);
tcg_temp_free(t0);
tcg_temp_free(tmask);
}
/* Similarly for vector subtraction. */
static void gen_v12sub(TCGv tdest, TCGv tsrca, TCGv tsrcb, uint64_t sign)
{
TCGv tsign = tcg_const_tl(sign);
TCGv t0 = tcg_temp_new();
TCGv t1 = tcg_temp_new();
/* ((a | sign) - (b & ~sign)) ^ ((a ^ ~b) & sign). */
tcg_gen_or_tl(t0, tsrca, tsign);
tcg_gen_andc_tl(t1, tsrcb, tsign);
tcg_gen_sub_tl(tdest, t0, t1);
tcg_gen_eqv_tl(t0, tsrca, tsrcb);
tcg_gen_and_tl(t0, t0, tsign);
tcg_gen_xor_tl(tdest, tdest, t0);
tcg_temp_free(t1);
tcg_temp_free(t0);
tcg_temp_free(tsign);
}
static void gen_v4sh(TCGv d64, TCGv a64, TCGv b64,
void (*generate)(TCGv_i32, TCGv_i32, TCGv_i32))
{
TCGv_i32 al = tcg_temp_new_i32();
TCGv_i32 ah = tcg_temp_new_i32();
TCGv_i32 bl = tcg_temp_new_i32();
tcg_gen_extr_i64_i32(al, ah, a64);
tcg_gen_extrl_i64_i32(bl, b64);
tcg_gen_andi_i32(bl, bl, 31);
generate(al, al, bl);
generate(ah, ah, bl);
tcg_gen_concat_i32_i64(d64, al, ah);
tcg_temp_free_i32(al);
tcg_temp_free_i32(ah);
tcg_temp_free_i32(bl);
}
static void gen_v4op(TCGv d64, TCGv a64, TCGv b64,
void (*generate)(TCGv_i32, TCGv_i32, TCGv_i32))
{
TCGv_i32 al = tcg_temp_new_i32();
TCGv_i32 ah = tcg_temp_new_i32();
TCGv_i32 bl = tcg_temp_new_i32();
TCGv_i32 bh = tcg_temp_new_i32();
tcg_gen_extr_i64_i32(al, ah, a64);
tcg_gen_extr_i64_i32(bl, bh, b64);
generate(al, al, bl);
generate(ah, ah, bh);
tcg_gen_concat_i32_i64(d64, al, ah);
tcg_temp_free_i32(al);
tcg_temp_free_i32(ah);
tcg_temp_free_i32(bl);
tcg_temp_free_i32(bh);
}
static TileExcp gen_signal(DisasContext *dc, int signo, int sigcode,
const char *mnemonic)
{
TCGv_i32 t0 = tcg_const_i32(signo);
TCGv_i32 t1 = tcg_const_i32(sigcode);
tcg_gen_st_i32(t0, cpu_env, offsetof(CPUTLGState, signo));
tcg_gen_st_i32(t1, cpu_env, offsetof(CPUTLGState, sigcode));
tcg_temp_free_i32(t1);
tcg_temp_free_i32(t0);
qemu_log_mask(CPU_LOG_TB_IN_ASM, "%s", mnemonic);
return TILEGX_EXCP_SIGNAL;
}
static bool parse_from_addli(uint64_t bundle, int *signo, int *sigcode)
{
int imm;
if ((get_Opcode_X0(bundle) != ADDLI_OPCODE_X0)
|| (get_Dest_X0(bundle) != TILEGX_R_ZERO)
|| (get_SrcA_X0(bundle) != TILEGX_R_ZERO)) {
return false;
}
imm = get_Imm16_X0(bundle);
*signo = imm & 0x3f;
*sigcode = (imm >> 6) & 0xf;
/* ??? The linux kernel validates both signo and the sigcode vs the
known max for each signal. Don't bother here. */
return true;
}
static TileExcp gen_specill(DisasContext *dc, unsigned dest, unsigned srca,
uint64_t bundle)
{
const char *mnemonic;
int signo;
int sigcode;
if (dest == 0x1c && srca == 0x25) {
signo = TARGET_SIGTRAP;
sigcode = TARGET_TRAP_BRKPT;
mnemonic = "bpt";
} else if (dest == 0x1d && srca == 0x25
&& parse_from_addli(bundle, &signo, &sigcode)) {
mnemonic = "raise";
} else {
signo = TARGET_SIGILL;
sigcode = TARGET_ILL_ILLOPC;
mnemonic = "ill";
}
return gen_signal(dc, signo, sigcode, mnemonic);
}
static TileExcp gen_rr_opcode(DisasContext *dc, unsigned opext,
unsigned dest, unsigned srca, uint64_t bundle)
{
TCGv tdest, tsrca;
const char *mnemonic;
TCGMemOp memop;
TileExcp ret = TILEGX_EXCP_NONE;
bool prefetch_nofault = false;
/* Eliminate instructions with no output before doing anything else. */
switch (opext) {
case OE_RR_Y0(NOP):
case OE_RR_Y1(NOP):
case OE_RR_X0(NOP):
case OE_RR_X1(NOP):
mnemonic = "nop";
goto done0;
case OE_RR_Y0(FNOP):
case OE_RR_Y1(FNOP):
case OE_RR_X0(FNOP):
case OE_RR_X1(FNOP):
mnemonic = "fnop";
goto done0;
case OE_RR_X1(DRAIN):
mnemonic = "drain";
goto done0;
case OE_RR_X1(FLUSHWB):
mnemonic = "flushwb";
goto done0;
case OE_RR_X1(ILL):
return gen_specill(dc, dest, srca, bundle);
case OE_RR_Y1(ILL):
return gen_signal(dc, TARGET_SIGILL, TARGET_ILL_ILLOPC, "ill");
case OE_RR_X1(MF):
mnemonic = "mf";
goto done0;
case OE_RR_X1(NAP):
/* ??? This should yield, especially in system mode. */
mnemonic = "nap";
goto done0;
case OE_RR_X1(IRET):
gen_helper_ext01_ics(cpu_env);
dc->jmp.cond = TCG_COND_ALWAYS;
dc->jmp.dest = tcg_temp_new();
tcg_gen_ld_tl(dc->jmp.dest, cpu_env,
offsetof(CPUTLGState, spregs[TILEGX_SPR_EX_CONTEXT_0_0]));
tcg_gen_andi_tl(dc->jmp.dest, dc->jmp.dest, ~7);
mnemonic = "iret";
goto done0;
case OE_RR_X1(SWINT0):
case OE_RR_X1(SWINT2):
case OE_RR_X1(SWINT3):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RR_X1(SWINT1):
ret = TILEGX_EXCP_SYSCALL;
mnemonic = "swint1";
done0:
if (srca || dest) {
return TILEGX_EXCP_OPCODE_UNKNOWN;
}
qemu_log_mask(CPU_LOG_TB_IN_ASM, "%s", mnemonic);
return ret;
case OE_RR_X1(DTLBPR):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RR_X1(FINV):
mnemonic = "finv";
goto done1;
case OE_RR_X1(FLUSH):
mnemonic = "flush";
goto done1;
case OE_RR_X1(ICOH):
mnemonic = "icoh";
goto done1;
case OE_RR_X1(INV):
mnemonic = "inv";
goto done1;
case OE_RR_X1(WH64):
mnemonic = "wh64";
goto done1;
case OE_RR_X1(JRP):
case OE_RR_Y1(JRP):
mnemonic = "jrp";
goto do_jr;
case OE_RR_X1(JR):
case OE_RR_Y1(JR):
mnemonic = "jr";
goto do_jr;
case OE_RR_X1(JALRP):
case OE_RR_Y1(JALRP):
mnemonic = "jalrp";
goto do_jalr;
case OE_RR_X1(JALR):
case OE_RR_Y1(JALR):
mnemonic = "jalr";
do_jalr:
tcg_gen_movi_tl(dest_gr(dc, TILEGX_R_LR),
dc->pc + TILEGX_BUNDLE_SIZE_IN_BYTES);
do_jr:
dc->jmp.cond = TCG_COND_ALWAYS;
dc->jmp.dest = tcg_temp_new();
tcg_gen_andi_tl(dc->jmp.dest, load_gr(dc, srca), ~7);
done1:
if (dest) {
return TILEGX_EXCP_OPCODE_UNKNOWN;
}
qemu_log_mask(CPU_LOG_TB_IN_ASM, "%s %s", mnemonic, reg_names[srca]);
return ret;
}
tdest = dest_gr(dc, dest);
tsrca = load_gr(dc, srca);
switch (opext) {
case OE_RR_X0(CNTLZ):
case OE_RR_Y0(CNTLZ):
tcg_gen_clzi_tl(tdest, tsrca, TARGET_LONG_BITS);
mnemonic = "cntlz";
break;
case OE_RR_X0(CNTTZ):
case OE_RR_Y0(CNTTZ):
tcg_gen_ctzi_tl(tdest, tsrca, TARGET_LONG_BITS);
mnemonic = "cnttz";
break;
case OE_RR_X0(FSINGLE_PACK1):
case OE_RR_Y0(FSINGLE_PACK1):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RR_X1(LD1S):
memop = MO_SB;
mnemonic = "ld1s"; /* prefetch_l1_fault */
goto do_load;
case OE_RR_X1(LD1U):
memop = MO_UB;
mnemonic = "ld1u"; /* prefetch, prefetch_l1 */
prefetch_nofault = (dest == TILEGX_R_ZERO);
goto do_load;
case OE_RR_X1(LD2S):
memop = MO_TESW;
mnemonic = "ld2s"; /* prefetch_l2_fault */
goto do_load;
case OE_RR_X1(LD2U):
memop = MO_TEUW;
mnemonic = "ld2u"; /* prefetch_l2 */
prefetch_nofault = (dest == TILEGX_R_ZERO);
goto do_load;
case OE_RR_X1(LD4S):
memop = MO_TESL;
mnemonic = "ld4s"; /* prefetch_l3_fault */
goto do_load;
case OE_RR_X1(LD4U):
memop = MO_TEUL;
mnemonic = "ld4u"; /* prefetch_l3 */
prefetch_nofault = (dest == TILEGX_R_ZERO);
goto do_load;
case OE_RR_X1(LDNT1S):
memop = MO_SB;
mnemonic = "ldnt1s";
goto do_load;
case OE_RR_X1(LDNT1U):
memop = MO_UB;
mnemonic = "ldnt1u";
goto do_load;
case OE_RR_X1(LDNT2S):
memop = MO_TESW;
mnemonic = "ldnt2s";
goto do_load;
case OE_RR_X1(LDNT2U):
memop = MO_TEUW;
mnemonic = "ldnt2u";
goto do_load;
case OE_RR_X1(LDNT4S):
memop = MO_TESL;
mnemonic = "ldnt4s";
goto do_load;
case OE_RR_X1(LDNT4U):
memop = MO_TEUL;
mnemonic = "ldnt4u";
goto do_load;
case OE_RR_X1(LDNT):
memop = MO_TEQ;
mnemonic = "ldnt";
goto do_load;
case OE_RR_X1(LD):
memop = MO_TEQ;
mnemonic = "ld";
do_load:
if (!prefetch_nofault) {
tcg_gen_qemu_ld_tl(tdest, tsrca, dc->mmuidx, memop);
}
break;
case OE_RR_X1(LDNA):
tcg_gen_andi_tl(tdest, tsrca, ~7);
tcg_gen_qemu_ld_tl(tdest, tdest, dc->mmuidx, MO_TEQ);
mnemonic = "ldna";
break;
case OE_RR_X1(LNK):
case OE_RR_Y1(LNK):
if (srca) {
return TILEGX_EXCP_OPCODE_UNKNOWN;
}
tcg_gen_movi_tl(tdest, dc->pc + TILEGX_BUNDLE_SIZE_IN_BYTES);
mnemonic = "lnk";
break;
case OE_RR_X0(PCNT):
case OE_RR_Y0(PCNT):
tcg_gen_ctpop_tl(tdest, tsrca);
mnemonic = "pcnt";
break;
case OE_RR_X0(REVBITS):
case OE_RR_Y0(REVBITS):
gen_helper_revbits(tdest, tsrca);
mnemonic = "revbits";
break;
case OE_RR_X0(REVBYTES):
case OE_RR_Y0(REVBYTES):
tcg_gen_bswap64_tl(tdest, tsrca);
mnemonic = "revbytes";
break;
case OE_RR_X0(TBLIDXB0):
case OE_RR_Y0(TBLIDXB0):
tcg_gen_deposit_tl(tdest, load_gr(dc, dest), tsrca, 2, 8);
mnemonic = "tblidxb0";
break;
case OE_RR_X0(TBLIDXB1):
case OE_RR_Y0(TBLIDXB1):
tcg_gen_shri_tl(tdest, tsrca, 8);
tcg_gen_deposit_tl(tdest, load_gr(dc, dest), tdest, 2, 8);
mnemonic = "tblidxb1";
break;
case OE_RR_X0(TBLIDXB2):
case OE_RR_Y0(TBLIDXB2):
tcg_gen_shri_tl(tdest, tsrca, 16);
tcg_gen_deposit_tl(tdest, load_gr(dc, dest), tdest, 2, 8);
mnemonic = "tblidxb2";
break;
case OE_RR_X0(TBLIDXB3):
case OE_RR_Y0(TBLIDXB3):
tcg_gen_shri_tl(tdest, tsrca, 24);
tcg_gen_deposit_tl(tdest, load_gr(dc, dest), tdest, 2, 8);
mnemonic = "tblidxb3";
break;
default:
return TILEGX_EXCP_OPCODE_UNKNOWN;
}
qemu_log_mask(CPU_LOG_TB_IN_ASM, "%s %s, %s", mnemonic,
reg_names[dest], reg_names[srca]);
return ret;
}
static TileExcp gen_rrr_opcode(DisasContext *dc, unsigned opext,
unsigned dest, unsigned srca, unsigned srcb)
{
TCGv tdest = dest_gr(dc, dest);
TCGv tsrca = load_gr(dc, srca);
TCGv tsrcb = load_gr(dc, srcb);
TCGv t0;
const char *mnemonic;
switch (opext) {
case OE_RRR(ADDXSC, 0, X0):
case OE_RRR(ADDXSC, 0, X1):
gen_saturate_op(tdest, tsrca, tsrcb, tcg_gen_add_tl);
mnemonic = "addxsc";
break;
case OE_RRR(ADDX, 0, X0):
case OE_RRR(ADDX, 0, X1):
case OE_RRR(ADDX, 0, Y0):
case OE_RRR(ADDX, 0, Y1):
tcg_gen_add_tl(tdest, tsrca, tsrcb);
tcg_gen_ext32s_tl(tdest, tdest);
mnemonic = "addx";
break;
case OE_RRR(ADD, 0, X0):
case OE_RRR(ADD, 0, X1):
case OE_RRR(ADD, 0, Y0):
case OE_RRR(ADD, 0, Y1):
tcg_gen_add_tl(tdest, tsrca, tsrcb);
mnemonic = "add";
break;
case OE_RRR(AND, 0, X0):
case OE_RRR(AND, 0, X1):
case OE_RRR(AND, 5, Y0):
case OE_RRR(AND, 5, Y1):
tcg_gen_and_tl(tdest, tsrca, tsrcb);
mnemonic = "and";
break;
case OE_RRR(CMOVEQZ, 0, X0):
case OE_RRR(CMOVEQZ, 4, Y0):
tcg_gen_movcond_tl(TCG_COND_EQ, tdest, tsrca, load_zero(dc),
tsrcb, load_gr(dc, dest));
mnemonic = "cmoveqz";
break;
case OE_RRR(CMOVNEZ, 0, X0):
case OE_RRR(CMOVNEZ, 4, Y0):
tcg_gen_movcond_tl(TCG_COND_NE, tdest, tsrca, load_zero(dc),
tsrcb, load_gr(dc, dest));
mnemonic = "cmovnez";
break;
case OE_RRR(CMPEQ, 0, X0):
case OE_RRR(CMPEQ, 0, X1):
case OE_RRR(CMPEQ, 3, Y0):
case OE_RRR(CMPEQ, 3, Y1):
tcg_gen_setcond_tl(TCG_COND_EQ, tdest, tsrca, tsrcb);
mnemonic = "cmpeq";
break;
case OE_RRR(CMPEXCH4, 0, X1):
gen_atomic_excp(dc, dest, tdest, tsrca, tsrcb,
TILEGX_EXCP_OPCODE_CMPEXCH4);
mnemonic = "cmpexch4";
break;
case OE_RRR(CMPEXCH, 0, X1):
gen_atomic_excp(dc, dest, tdest, tsrca, tsrcb,
TILEGX_EXCP_OPCODE_CMPEXCH);
mnemonic = "cmpexch";
break;
case OE_RRR(CMPLES, 0, X0):
case OE_RRR(CMPLES, 0, X1):
case OE_RRR(CMPLES, 2, Y0):
case OE_RRR(CMPLES, 2, Y1):
tcg_gen_setcond_tl(TCG_COND_LE, tdest, tsrca, tsrcb);
mnemonic = "cmples";
break;
case OE_RRR(CMPLEU, 0, X0):
case OE_RRR(CMPLEU, 0, X1):
case OE_RRR(CMPLEU, 2, Y0):
case OE_RRR(CMPLEU, 2, Y1):
tcg_gen_setcond_tl(TCG_COND_LEU, tdest, tsrca, tsrcb);
mnemonic = "cmpleu";
break;
case OE_RRR(CMPLTS, 0, X0):
case OE_RRR(CMPLTS, 0, X1):
case OE_RRR(CMPLTS, 2, Y0):
case OE_RRR(CMPLTS, 2, Y1):
tcg_gen_setcond_tl(TCG_COND_LT, tdest, tsrca, tsrcb);
mnemonic = "cmplts";
break;
case OE_RRR(CMPLTU, 0, X0):
case OE_RRR(CMPLTU, 0, X1):
case OE_RRR(CMPLTU, 2, Y0):
case OE_RRR(CMPLTU, 2, Y1):
tcg_gen_setcond_tl(TCG_COND_LTU, tdest, tsrca, tsrcb);
mnemonic = "cmpltu";
break;
case OE_RRR(CMPNE, 0, X0):
case OE_RRR(CMPNE, 0, X1):
case OE_RRR(CMPNE, 3, Y0):
case OE_RRR(CMPNE, 3, Y1):
tcg_gen_setcond_tl(TCG_COND_NE, tdest, tsrca, tsrcb);
mnemonic = "cmpne";
break;
case OE_RRR(CMULAF, 0, X0):
gen_helper_cmulaf(tdest, load_gr(dc, dest), tsrca, tsrcb);
mnemonic = "cmulaf";
break;
case OE_RRR(CMULA, 0, X0):
gen_helper_cmula(tdest, load_gr(dc, dest), tsrca, tsrcb);
mnemonic = "cmula";
break;
case OE_RRR(CMULFR, 0, X0):
gen_cmul2(tdest, tsrca, tsrcb, 15, 1 << 14);
mnemonic = "cmulfr";
break;
case OE_RRR(CMULF, 0, X0):
gen_cmul2(tdest, tsrca, tsrcb, 15, 0);
mnemonic = "cmulf";
break;
case OE_RRR(CMULHR, 0, X0):
gen_cmul2(tdest, tsrca, tsrcb, 16, 1 << 15);
mnemonic = "cmulhr";
break;
case OE_RRR(CMULH, 0, X0):
gen_cmul2(tdest, tsrca, tsrcb, 16, 0);
mnemonic = "cmulh";
break;
case OE_RRR(CMUL, 0, X0):
gen_helper_cmula(tdest, load_zero(dc), tsrca, tsrcb);
mnemonic = "cmul";
break;
case OE_RRR(CRC32_32, 0, X0):
gen_helper_crc32_32(tdest, tsrca, tsrcb);
mnemonic = "crc32_32";
break;
case OE_RRR(CRC32_8, 0, X0):
gen_helper_crc32_8(tdest, tsrca, tsrcb);
mnemonic = "crc32_8";
break;
case OE_RRR(DBLALIGN2, 0, X0):
case OE_RRR(DBLALIGN2, 0, X1):
gen_dblaligni(tdest, tsrca, tsrcb, 16);
mnemonic = "dblalign2";
break;
case OE_RRR(DBLALIGN4, 0, X0):
case OE_RRR(DBLALIGN4, 0, X1):
gen_dblaligni(tdest, tsrca, tsrcb, 32);
mnemonic = "dblalign4";
break;
case OE_RRR(DBLALIGN6, 0, X0):
case OE_RRR(DBLALIGN6, 0, X1):
gen_dblaligni(tdest, tsrca, tsrcb, 48);
mnemonic = "dblalign6";
break;
case OE_RRR(DBLALIGN, 0, X0):
gen_dblalign(tdest, load_gr(dc, dest), tsrca, tsrcb);
mnemonic = "dblalign";
break;
case OE_RRR(EXCH4, 0, X1):
gen_atomic_excp(dc, dest, tdest, tsrca, tsrcb,
TILEGX_EXCP_OPCODE_EXCH4);
mnemonic = "exch4";
break;
case OE_RRR(EXCH, 0, X1):
gen_atomic_excp(dc, dest, tdest, tsrca, tsrcb,
TILEGX_EXCP_OPCODE_EXCH);
mnemonic = "exch";
break;
case OE_RRR(FDOUBLE_ADDSUB, 0, X0):
case OE_RRR(FDOUBLE_ADD_FLAGS, 0, X0):
case OE_RRR(FDOUBLE_MUL_FLAGS, 0, X0):
case OE_RRR(FDOUBLE_PACK1, 0, X0):
case OE_RRR(FDOUBLE_PACK2, 0, X0):
case OE_RRR(FDOUBLE_SUB_FLAGS, 0, X0):
case OE_RRR(FDOUBLE_UNPACK_MAX, 0, X0):
case OE_RRR(FDOUBLE_UNPACK_MIN, 0, X0):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(FETCHADD4, 0, X1):
gen_atomic_excp(dc, dest, tdest, tsrca, tsrcb,
TILEGX_EXCP_OPCODE_FETCHADD4);
mnemonic = "fetchadd4";
break;
case OE_RRR(FETCHADDGEZ4, 0, X1):
gen_atomic_excp(dc, dest, tdest, tsrca, tsrcb,
TILEGX_EXCP_OPCODE_FETCHADDGEZ4);
mnemonic = "fetchaddgez4";
break;
case OE_RRR(FETCHADDGEZ, 0, X1):
gen_atomic_excp(dc, dest, tdest, tsrca, tsrcb,
TILEGX_EXCP_OPCODE_FETCHADDGEZ);
mnemonic = "fetchaddgez";
break;
case OE_RRR(FETCHADD, 0, X1):
gen_atomic_excp(dc, dest, tdest, tsrca, tsrcb,
TILEGX_EXCP_OPCODE_FETCHADD);
mnemonic = "fetchadd";
break;
case OE_RRR(FETCHAND4, 0, X1):
gen_atomic_excp(dc, dest, tdest, tsrca, tsrcb,
TILEGX_EXCP_OPCODE_FETCHAND4);
mnemonic = "fetchand4";
break;
case OE_RRR(FETCHAND, 0, X1):
gen_atomic_excp(dc, dest, tdest, tsrca, tsrcb,
TILEGX_EXCP_OPCODE_FETCHAND);
mnemonic = "fetchand";
break;
case OE_RRR(FETCHOR4, 0, X1):
gen_atomic_excp(dc, dest, tdest, tsrca, tsrcb,
TILEGX_EXCP_OPCODE_FETCHOR4);
mnemonic = "fetchor4";
break;
case OE_RRR(FETCHOR, 0, X1):
gen_atomic_excp(dc, dest, tdest, tsrca, tsrcb,
TILEGX_EXCP_OPCODE_FETCHOR);
mnemonic = "fetchor";
break;
case OE_RRR(FSINGLE_ADD1, 0, X0):
case OE_RRR(FSINGLE_ADDSUB2, 0, X0):
case OE_RRR(FSINGLE_MUL1, 0, X0):
case OE_RRR(FSINGLE_MUL2, 0, X0):
case OE_RRR(FSINGLE_PACK2, 0, X0):
case OE_RRR(FSINGLE_SUB1, 0, X0):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(MNZ, 0, X0):
case OE_RRR(MNZ, 0, X1):
case OE_RRR(MNZ, 4, Y0):
case OE_RRR(MNZ, 4, Y1):
t0 = load_zero(dc);
tcg_gen_movcond_tl(TCG_COND_NE, tdest, tsrca, t0, tsrcb, t0);
mnemonic = "mnz";
break;
case OE_RRR(MULAX, 0, X0):
case OE_RRR(MULAX, 3, Y0):
tcg_gen_mul_tl(tdest, tsrca, tsrcb);
tcg_gen_add_tl(tdest, tdest, load_gr(dc, dest));
tcg_gen_ext32s_tl(tdest, tdest);
mnemonic = "mulax";
break;
case OE_RRR(MULA_HS_HS, 0, X0):
case OE_RRR(MULA_HS_HS, 9, Y0):
gen_mul_half(tdest, tsrca, tsrcb, HS, HS);
tcg_gen_add_tl(tdest, tdest, load_gr(dc, dest));
mnemonic = "mula_hs_hs";
break;
case OE_RRR(MULA_HS_HU, 0, X0):
gen_mul_half(tdest, tsrca, tsrcb, HS, HU);
tcg_gen_add_tl(tdest, tdest, load_gr(dc, dest));
mnemonic = "mula_hs_hu";
break;
case OE_RRR(MULA_HS_LS, 0, X0):
gen_mul_half(tdest, tsrca, tsrcb, HS, LS);
tcg_gen_add_tl(tdest, tdest, load_gr(dc, dest));
mnemonic = "mula_hs_ls";
break;
case OE_RRR(MULA_HS_LU, 0, X0):
gen_mul_half(tdest, tsrca, tsrcb, HS, LU);
tcg_gen_add_tl(tdest, tdest, load_gr(dc, dest));
mnemonic = "mula_hs_lu";
break;
case OE_RRR(MULA_HU_HU, 0, X0):
case OE_RRR(MULA_HU_HU, 9, Y0):
gen_mul_half(tdest, tsrca, tsrcb, HU, HU);
tcg_gen_add_tl(tdest, tdest, load_gr(dc, dest));
mnemonic = "mula_hu_hu";
break;
case OE_RRR(MULA_HU_LS, 0, X0):
gen_mul_half(tdest, tsrca, tsrcb, HU, LS);
tcg_gen_add_tl(tdest, tdest, load_gr(dc, dest));
mnemonic = "mula_hu_ls";
break;
case OE_RRR(MULA_HU_LU, 0, X0):
gen_mul_half(tdest, tsrca, tsrcb, HU, LU);
tcg_gen_add_tl(tdest, tdest, load_gr(dc, dest));
mnemonic = "mula_hu_lu";
break;
case OE_RRR(MULA_LS_LS, 0, X0):
case OE_RRR(MULA_LS_LS, 9, Y0):
gen_mul_half(tdest, tsrca, tsrcb, LS, LS);
tcg_gen_add_tl(tdest, tdest, load_gr(dc, dest));
mnemonic = "mula_ls_ls";
break;
case OE_RRR(MULA_LS_LU, 0, X0):
gen_mul_half(tdest, tsrca, tsrcb, LS, LU);
tcg_gen_add_tl(tdest, tdest, load_gr(dc, dest));
mnemonic = "mula_ls_lu";
break;
case OE_RRR(MULA_LU_LU, 0, X0):
case OE_RRR(MULA_LU_LU, 9, Y0):
gen_mul_half(tdest, tsrca, tsrcb, LU, LU);
tcg_gen_add_tl(tdest, tdest, load_gr(dc, dest));
mnemonic = "mula_lu_lu";
break;
case OE_RRR(MULX, 0, X0):
case OE_RRR(MULX, 3, Y0):
tcg_gen_mul_tl(tdest, tsrca, tsrcb);
tcg_gen_ext32s_tl(tdest, tdest);
mnemonic = "mulx";
break;
case OE_RRR(MUL_HS_HS, 0, X0):
case OE_RRR(MUL_HS_HS, 8, Y0):
gen_mul_half(tdest, tsrca, tsrcb, HS, HS);
mnemonic = "mul_hs_hs";
break;
case OE_RRR(MUL_HS_HU, 0, X0):
gen_mul_half(tdest, tsrca, tsrcb, HS, HU);
mnemonic = "mul_hs_hu";
break;
case OE_RRR(MUL_HS_LS, 0, X0):
gen_mul_half(tdest, tsrca, tsrcb, HS, LS);
mnemonic = "mul_hs_ls";
break;
case OE_RRR(MUL_HS_LU, 0, X0):
gen_mul_half(tdest, tsrca, tsrcb, HS, LU);
mnemonic = "mul_hs_lu";
break;
case OE_RRR(MUL_HU_HU, 0, X0):
case OE_RRR(MUL_HU_HU, 8, Y0):
gen_mul_half(tdest, tsrca, tsrcb, HU, HU);
mnemonic = "mul_hu_hu";
break;
case OE_RRR(MUL_HU_LS, 0, X0):
gen_mul_half(tdest, tsrca, tsrcb, HU, LS);
mnemonic = "mul_hu_ls";
break;
case OE_RRR(MUL_HU_LU, 0, X0):
gen_mul_half(tdest, tsrca, tsrcb, HU, LU);
mnemonic = "mul_hu_lu";
break;
case OE_RRR(MUL_LS_LS, 0, X0):
case OE_RRR(MUL_LS_LS, 8, Y0):
gen_mul_half(tdest, tsrca, tsrcb, LS, LS);
mnemonic = "mul_ls_ls";
break;
case OE_RRR(MUL_LS_LU, 0, X0):
gen_mul_half(tdest, tsrca, tsrcb, LS, LU);
mnemonic = "mul_ls_lu";
break;
case OE_RRR(MUL_LU_LU, 0, X0):
case OE_RRR(MUL_LU_LU, 8, Y0):
gen_mul_half(tdest, tsrca, tsrcb, LU, LU);
mnemonic = "mul_lu_lu";
break;
case OE_RRR(MZ, 0, X0):
case OE_RRR(MZ, 0, X1):
case OE_RRR(MZ, 4, Y0):
case OE_RRR(MZ, 4, Y1):
t0 = load_zero(dc);
tcg_gen_movcond_tl(TCG_COND_EQ, tdest, tsrca, t0, tsrcb, t0);
mnemonic = "mz";
break;
case OE_RRR(NOR, 0, X0):
case OE_RRR(NOR, 0, X1):
case OE_RRR(NOR, 5, Y0):
case OE_RRR(NOR, 5, Y1):
tcg_gen_nor_tl(tdest, tsrca, tsrcb);
mnemonic = "nor";
break;
case OE_RRR(OR, 0, X0):
case OE_RRR(OR, 0, X1):
case OE_RRR(OR, 5, Y0):
case OE_RRR(OR, 5, Y1):
tcg_gen_or_tl(tdest, tsrca, tsrcb);
mnemonic = "or";
break;
case OE_RRR(ROTL, 0, X0):
case OE_RRR(ROTL, 0, X1):
case OE_RRR(ROTL, 6, Y0):
case OE_RRR(ROTL, 6, Y1):
tcg_gen_andi_tl(tdest, tsrcb, 63);
tcg_gen_rotl_tl(tdest, tsrca, tdest);
mnemonic = "rotl";
break;
case OE_RRR(SHL1ADDX, 0, X0):
case OE_RRR(SHL1ADDX, 0, X1):
case OE_RRR(SHL1ADDX, 7, Y0):
case OE_RRR(SHL1ADDX, 7, Y1):
tcg_gen_shli_tl(tdest, tsrca, 1);
tcg_gen_add_tl(tdest, tdest, tsrcb);
tcg_gen_ext32s_tl(tdest, tdest);
mnemonic = "shl1addx";
break;
case OE_RRR(SHL1ADD, 0, X0):
case OE_RRR(SHL1ADD, 0, X1):
case OE_RRR(SHL1ADD, 1, Y0):
case OE_RRR(SHL1ADD, 1, Y1):
tcg_gen_shli_tl(tdest, tsrca, 1);
tcg_gen_add_tl(tdest, tdest, tsrcb);
mnemonic = "shl1add";
break;
case OE_RRR(SHL2ADDX, 0, X0):
case OE_RRR(SHL2ADDX, 0, X1):
case OE_RRR(SHL2ADDX, 7, Y0):
case OE_RRR(SHL2ADDX, 7, Y1):
tcg_gen_shli_tl(tdest, tsrca, 2);
tcg_gen_add_tl(tdest, tdest, tsrcb);
tcg_gen_ext32s_tl(tdest, tdest);
mnemonic = "shl2addx";
break;
case OE_RRR(SHL2ADD, 0, X0):
case OE_RRR(SHL2ADD, 0, X1):
case OE_RRR(SHL2ADD, 1, Y0):
case OE_RRR(SHL2ADD, 1, Y1):
tcg_gen_shli_tl(tdest, tsrca, 2);
tcg_gen_add_tl(tdest, tdest, tsrcb);
mnemonic = "shl2add";
break;
case OE_RRR(SHL3ADDX, 0, X0):
case OE_RRR(SHL3ADDX, 0, X1):
case OE_RRR(SHL3ADDX, 7, Y0):
case OE_RRR(SHL3ADDX, 7, Y1):
tcg_gen_shli_tl(tdest, tsrca, 3);
tcg_gen_add_tl(tdest, tdest, tsrcb);
tcg_gen_ext32s_tl(tdest, tdest);
mnemonic = "shl3addx";
break;
case OE_RRR(SHL3ADD, 0, X0):
case OE_RRR(SHL3ADD, 0, X1):
case OE_RRR(SHL3ADD, 1, Y0):
case OE_RRR(SHL3ADD, 1, Y1):
tcg_gen_shli_tl(tdest, tsrca, 3);
tcg_gen_add_tl(tdest, tdest, tsrcb);
mnemonic = "shl3add";
break;
case OE_RRR(SHLX, 0, X0):
case OE_RRR(SHLX, 0, X1):
tcg_gen_andi_tl(tdest, tsrcb, 31);
tcg_gen_shl_tl(tdest, tsrca, tdest);
tcg_gen_ext32s_tl(tdest, tdest);
mnemonic = "shlx";
break;
case OE_RRR(SHL, 0, X0):
case OE_RRR(SHL, 0, X1):
case OE_RRR(SHL, 6, Y0):
case OE_RRR(SHL, 6, Y1):
tcg_gen_andi_tl(tdest, tsrcb, 63);
tcg_gen_shl_tl(tdest, tsrca, tdest);
mnemonic = "shl";
break;
case OE_RRR(SHRS, 0, X0):
case OE_RRR(SHRS, 0, X1):
case OE_RRR(SHRS, 6, Y0):
case OE_RRR(SHRS, 6, Y1):
tcg_gen_andi_tl(tdest, tsrcb, 63);
tcg_gen_sar_tl(tdest, tsrca, tdest);
mnemonic = "shrs";
break;
case OE_RRR(SHRUX, 0, X0):
case OE_RRR(SHRUX, 0, X1):
t0 = tcg_temp_new();
tcg_gen_andi_tl(t0, tsrcb, 31);
tcg_gen_ext32u_tl(tdest, tsrca);
tcg_gen_shr_tl(tdest, tdest, t0);
tcg_gen_ext32s_tl(tdest, tdest);
tcg_temp_free(t0);
mnemonic = "shrux";
break;
case OE_RRR(SHRU, 0, X0):
case OE_RRR(SHRU, 0, X1):
case OE_RRR(SHRU, 6, Y0):
case OE_RRR(SHRU, 6, Y1):
tcg_gen_andi_tl(tdest, tsrcb, 63);
tcg_gen_shr_tl(tdest, tsrca, tdest);
mnemonic = "shru";
break;
case OE_RRR(SHUFFLEBYTES, 0, X0):
gen_helper_shufflebytes(tdest, load_gr(dc, dest), tsrca, tsrca);
mnemonic = "shufflebytes";
break;
case OE_RRR(SUBXSC, 0, X0):
case OE_RRR(SUBXSC, 0, X1):
gen_saturate_op(tdest, tsrca, tsrcb, tcg_gen_sub_tl);
mnemonic = "subxsc";
break;
case OE_RRR(SUBX, 0, X0):
case OE_RRR(SUBX, 0, X1):
case OE_RRR(SUBX, 0, Y0):
case OE_RRR(SUBX, 0, Y1):
tcg_gen_sub_tl(tdest, tsrca, tsrcb);
tcg_gen_ext32s_tl(tdest, tdest);
mnemonic = "subx";
break;
case OE_RRR(SUB, 0, X0):
case OE_RRR(SUB, 0, X1):
case OE_RRR(SUB, 0, Y0):
case OE_RRR(SUB, 0, Y1):
tcg_gen_sub_tl(tdest, tsrca, tsrcb);
mnemonic = "sub";
break;
case OE_RRR(V1ADDUC, 0, X0):
case OE_RRR(V1ADDUC, 0, X1):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(V1ADD, 0, X0):
case OE_RRR(V1ADD, 0, X1):
gen_v12add(tdest, tsrca, tsrcb, V1_IMM(0x80));
mnemonic = "v1add";
break;
case OE_RRR(V1ADIFFU, 0, X0):
case OE_RRR(V1AVGU, 0, X0):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(V1CMPEQ, 0, X0):
case OE_RRR(V1CMPEQ, 0, X1):
tcg_gen_xor_tl(tdest, tsrca, tsrcb);
gen_v1cmpeq0(tdest);
mnemonic = "v1cmpeq";
break;
case OE_RRR(V1CMPLES, 0, X0):
case OE_RRR(V1CMPLES, 0, X1):
case OE_RRR(V1CMPLEU, 0, X0):
case OE_RRR(V1CMPLEU, 0, X1):
case OE_RRR(V1CMPLTS, 0, X0):
case OE_RRR(V1CMPLTS, 0, X1):
case OE_RRR(V1CMPLTU, 0, X0):
case OE_RRR(V1CMPLTU, 0, X1):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(V1CMPNE, 0, X0):
case OE_RRR(V1CMPNE, 0, X1):
tcg_gen_xor_tl(tdest, tsrca, tsrcb);
gen_v1cmpne0(tdest);
mnemonic = "v1cmpne";
break;
case OE_RRR(V1DDOTPUA, 0, X0):
case OE_RRR(V1DDOTPUSA, 0, X0):
case OE_RRR(V1DDOTPUS, 0, X0):
case OE_RRR(V1DDOTPU, 0, X0):
case OE_RRR(V1DOTPA, 0, X0):
case OE_RRR(V1DOTPUA, 0, X0):
case OE_RRR(V1DOTPUSA, 0, X0):
case OE_RRR(V1DOTPUS, 0, X0):
case OE_RRR(V1DOTPU, 0, X0):
case OE_RRR(V1DOTP, 0, X0):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(V1INT_H, 0, X0):
case OE_RRR(V1INT_H, 0, X1):
gen_helper_v1int_h(tdest, tsrca, tsrcb);
mnemonic = "v1int_h";
break;
case OE_RRR(V1INT_L, 0, X0):
case OE_RRR(V1INT_L, 0, X1):
gen_helper_v1int_l(tdest, tsrca, tsrcb);
mnemonic = "v1int_l";
break;
case OE_RRR(V1MAXU, 0, X0):
case OE_RRR(V1MAXU, 0, X1):
case OE_RRR(V1MINU, 0, X0):
case OE_RRR(V1MINU, 0, X1):
case OE_RRR(V1MNZ, 0, X0):
case OE_RRR(V1MNZ, 0, X1):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(V1MULTU, 0, X0):
gen_helper_v1multu(tdest, tsrca, tsrcb);
mnemonic = "v1multu";
break;
case OE_RRR(V1MULUS, 0, X0):
case OE_RRR(V1MULU, 0, X0):
case OE_RRR(V1MZ, 0, X0):
case OE_RRR(V1MZ, 0, X1):
case OE_RRR(V1SADAU, 0, X0):
case OE_RRR(V1SADU, 0, X0):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(V1SHL, 0, X0):
case OE_RRR(V1SHL, 0, X1):
gen_helper_v1shl(tdest, tsrca, tsrcb);
mnemonic = "v1shl";
break;
case OE_RRR(V1SHRS, 0, X0):
case OE_RRR(V1SHRS, 0, X1):
gen_helper_v1shrs(tdest, tsrca, tsrcb);
mnemonic = "v1shrs";
break;
case OE_RRR(V1SHRU, 0, X0):
case OE_RRR(V1SHRU, 0, X1):
gen_helper_v1shru(tdest, tsrca, tsrcb);
mnemonic = "v1shru";
break;
case OE_RRR(V1SUBUC, 0, X0):
case OE_RRR(V1SUBUC, 0, X1):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(V1SUB, 0, X0):
case OE_RRR(V1SUB, 0, X1):
gen_v12sub(tdest, tsrca, tsrcb, V1_IMM(0x80));
mnemonic = "v1sub";
break;
case OE_RRR(V2ADDSC, 0, X0):
case OE_RRR(V2ADDSC, 0, X1):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(V2ADD, 0, X0):
case OE_RRR(V2ADD, 0, X1):
gen_v12add(tdest, tsrca, tsrcb, V2_IMM(0x8000));
mnemonic = "v2add";
break;
case OE_RRR(V2ADIFFS, 0, X0):
case OE_RRR(V2AVGS, 0, X0):
case OE_RRR(V2CMPEQ, 0, X0):
case OE_RRR(V2CMPEQ, 0, X1):
case OE_RRR(V2CMPLES, 0, X0):
case OE_RRR(V2CMPLES, 0, X1):
case OE_RRR(V2CMPLEU, 0, X0):
case OE_RRR(V2CMPLEU, 0, X1):
case OE_RRR(V2CMPLTS, 0, X0):
case OE_RRR(V2CMPLTS, 0, X1):
case OE_RRR(V2CMPLTU, 0, X0):
case OE_RRR(V2CMPLTU, 0, X1):
case OE_RRR(V2CMPNE, 0, X0):
case OE_RRR(V2CMPNE, 0, X1):
case OE_RRR(V2DOTPA, 0, X0):
case OE_RRR(V2DOTP, 0, X0):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(V2INT_H, 0, X0):
case OE_RRR(V2INT_H, 0, X1):
gen_helper_v2int_h(tdest, tsrca, tsrcb);
mnemonic = "v2int_h";
break;
case OE_RRR(V2INT_L, 0, X0):
case OE_RRR(V2INT_L, 0, X1):
gen_helper_v2int_l(tdest, tsrca, tsrcb);
mnemonic = "v2int_l";
break;
case OE_RRR(V2MAXS, 0, X0):
case OE_RRR(V2MAXS, 0, X1):
case OE_RRR(V2MINS, 0, X0):
case OE_RRR(V2MINS, 0, X1):
case OE_RRR(V2MNZ, 0, X0):
case OE_RRR(V2MNZ, 0, X1):
case OE_RRR(V2MULFSC, 0, X0):
case OE_RRR(V2MULS, 0, X0):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(V2MULTS, 0, X0):
gen_helper_v2mults(tdest, tsrca, tsrcb);
mnemonic = "v2mults";
break;
case OE_RRR(V2MZ, 0, X0):
case OE_RRR(V2MZ, 0, X1):
case OE_RRR(V2PACKH, 0, X0):
case OE_RRR(V2PACKH, 0, X1):
case OE_RRR(V2PACKL, 0, X0):
case OE_RRR(V2PACKL, 0, X1):
case OE_RRR(V2PACKUC, 0, X0):
case OE_RRR(V2PACKUC, 0, X1):
case OE_RRR(V2SADAS, 0, X0):
case OE_RRR(V2SADAU, 0, X0):
case OE_RRR(V2SADS, 0, X0):
case OE_RRR(V2SADU, 0, X0):
case OE_RRR(V2SHLSC, 0, X0):
case OE_RRR(V2SHLSC, 0, X1):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(V2SHL, 0, X0):
case OE_RRR(V2SHL, 0, X1):
gen_helper_v2shl(tdest, tsrca, tsrcb);
mnemonic = "v2shl";
break;
case OE_RRR(V2SHRS, 0, X0):
case OE_RRR(V2SHRS, 0, X1):
gen_helper_v2shrs(tdest, tsrca, tsrcb);
mnemonic = "v2shrs";
break;
case OE_RRR(V2SHRU, 0, X0):
case OE_RRR(V2SHRU, 0, X1):
gen_helper_v2shru(tdest, tsrca, tsrcb);
mnemonic = "v2shru";
break;
case OE_RRR(V2SUBSC, 0, X0):
case OE_RRR(V2SUBSC, 0, X1):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(V2SUB, 0, X0):
case OE_RRR(V2SUB, 0, X1):
gen_v12sub(tdest, tsrca, tsrcb, V2_IMM(0x8000));
mnemonic = "v2sub";
break;
case OE_RRR(V4ADDSC, 0, X0):
case OE_RRR(V4ADDSC, 0, X1):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(V4ADD, 0, X0):
case OE_RRR(V4ADD, 0, X1):
gen_v4op(tdest, tsrca, tsrcb, tcg_gen_add_i32);
mnemonic = "v4add";
break;
case OE_RRR(V4INT_H, 0, X0):
case OE_RRR(V4INT_H, 0, X1):
tcg_gen_shri_tl(tdest, tsrcb, 32);
tcg_gen_deposit_tl(tdest, tsrca, tdest, 0, 32);
mnemonic = "v4int_h";
break;
case OE_RRR(V4INT_L, 0, X0):
case OE_RRR(V4INT_L, 0, X1):
tcg_gen_deposit_tl(tdest, tsrcb, tsrca, 32, 32);
mnemonic = "v4int_l";
break;
case OE_RRR(V4PACKSC, 0, X0):
case OE_RRR(V4PACKSC, 0, X1):
case OE_RRR(V4SHLSC, 0, X0):
case OE_RRR(V4SHLSC, 0, X1):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(V4SHL, 0, X0):
case OE_RRR(V4SHL, 0, X1):
gen_v4sh(tdest, tsrca, tsrcb, tcg_gen_shl_i32);
mnemonic = "v4shl";
break;
case OE_RRR(V4SHRS, 0, X0):
case OE_RRR(V4SHRS, 0, X1):
gen_v4sh(tdest, tsrca, tsrcb, tcg_gen_sar_i32);
mnemonic = "v4shrs";
break;
case OE_RRR(V4SHRU, 0, X0):
case OE_RRR(V4SHRU, 0, X1):
gen_v4sh(tdest, tsrca, tsrcb, tcg_gen_shr_i32);
mnemonic = "v4shru";
break;
case OE_RRR(V4SUBSC, 0, X0):
case OE_RRR(V4SUBSC, 0, X1):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_RRR(V4SUB, 0, X0):
case OE_RRR(V4SUB, 0, X1):
gen_v4op(tdest, tsrca, tsrcb, tcg_gen_sub_i32);
mnemonic = "v2sub";
break;
case OE_RRR(XOR, 0, X0):
case OE_RRR(XOR, 0, X1):
case OE_RRR(XOR, 5, Y0):
case OE_RRR(XOR, 5, Y1):
tcg_gen_xor_tl(tdest, tsrca, tsrcb);
mnemonic = "xor";
break;
default:
return TILEGX_EXCP_OPCODE_UNKNOWN;
}
qemu_log_mask(CPU_LOG_TB_IN_ASM, "%s %s, %s, %s", mnemonic,
reg_names[dest], reg_names[srca], reg_names[srcb]);
return TILEGX_EXCP_NONE;
}
static TileExcp gen_rri_opcode(DisasContext *dc, unsigned opext,
unsigned dest, unsigned srca, int imm)
{
TCGv tdest = dest_gr(dc, dest);
TCGv tsrca = load_gr(dc, srca);
bool prefetch_nofault = false;
const char *mnemonic;
TCGMemOp memop;
int i2, i3;
TCGv t0;
switch (opext) {
case OE(ADDI_OPCODE_Y0, 0, Y0):
case OE(ADDI_OPCODE_Y1, 0, Y1):
case OE_IM(ADDI, X0):
case OE_IM(ADDI, X1):
tcg_gen_addi_tl(tdest, tsrca, imm);
mnemonic = "addi";
break;
case OE(ADDXI_OPCODE_Y0, 0, Y0):
case OE(ADDXI_OPCODE_Y1, 0, Y1):
case OE_IM(ADDXI, X0):
case OE_IM(ADDXI, X1):
tcg_gen_addi_tl(tdest, tsrca, imm);
tcg_gen_ext32s_tl(tdest, tdest);
mnemonic = "addxi";
break;
case OE(ANDI_OPCODE_Y0, 0, Y0):
case OE(ANDI_OPCODE_Y1, 0, Y1):
case OE_IM(ANDI, X0):
case OE_IM(ANDI, X1):
tcg_gen_andi_tl(tdest, tsrca, imm);
mnemonic = "andi";
break;
case OE(CMPEQI_OPCODE_Y0, 0, Y0):
case OE(CMPEQI_OPCODE_Y1, 0, Y1):
case OE_IM(CMPEQI, X0):
case OE_IM(CMPEQI, X1):
tcg_gen_setcondi_tl(TCG_COND_EQ, tdest, tsrca, imm);
mnemonic = "cmpeqi";
break;
case OE(CMPLTSI_OPCODE_Y0, 0, Y0):
case OE(CMPLTSI_OPCODE_Y1, 0, Y1):
case OE_IM(CMPLTSI, X0):
case OE_IM(CMPLTSI, X1):
tcg_gen_setcondi_tl(TCG_COND_LT, tdest, tsrca, imm);
mnemonic = "cmpltsi";
break;
case OE_IM(CMPLTUI, X0):
case OE_IM(CMPLTUI, X1):
tcg_gen_setcondi_tl(TCG_COND_LTU, tdest, tsrca, imm);
mnemonic = "cmpltui";
break;
case OE_IM(LD1S_ADD, X1):
memop = MO_SB;
mnemonic = "ld1s_add"; /* prefetch_add_l1_fault */
goto do_load_add;
case OE_IM(LD1U_ADD, X1):
memop = MO_UB;
mnemonic = "ld1u_add"; /* prefetch_add_l1 */
prefetch_nofault = (dest == TILEGX_R_ZERO);
goto do_load_add;
case OE_IM(LD2S_ADD, X1):
memop = MO_TESW;
mnemonic = "ld2s_add"; /* prefetch_add_l2_fault */
goto do_load_add;
case OE_IM(LD2U_ADD, X1):
memop = MO_TEUW;
mnemonic = "ld2u_add"; /* prefetch_add_l2 */
prefetch_nofault = (dest == TILEGX_R_ZERO);
goto do_load_add;
case OE_IM(LD4S_ADD, X1):
memop = MO_TESL;
mnemonic = "ld4s_add"; /* prefetch_add_l3_fault */
goto do_load_add;
case OE_IM(LD4U_ADD, X1):
memop = MO_TEUL;
mnemonic = "ld4u_add"; /* prefetch_add_l3 */
prefetch_nofault = (dest == TILEGX_R_ZERO);
goto do_load_add;
case OE_IM(LDNT1S_ADD, X1):
memop = MO_SB;
mnemonic = "ldnt1s_add";
goto do_load_add;
case OE_IM(LDNT1U_ADD, X1):
memop = MO_UB;
mnemonic = "ldnt1u_add";
goto do_load_add;
case OE_IM(LDNT2S_ADD, X1):
memop = MO_TESW;
mnemonic = "ldnt2s_add";
goto do_load_add;
case OE_IM(LDNT2U_ADD, X1):
memop = MO_TEUW;
mnemonic = "ldnt2u_add";
goto do_load_add;
case OE_IM(LDNT4S_ADD, X1):
memop = MO_TESL;
mnemonic = "ldnt4s_add";
goto do_load_add;
case OE_IM(LDNT4U_ADD, X1):
memop = MO_TEUL;
mnemonic = "ldnt4u_add";
goto do_load_add;
case OE_IM(LDNT_ADD, X1):
memop = MO_TEQ;
mnemonic = "ldnt_add";
goto do_load_add;
case OE_IM(LD_ADD, X1):
memop = MO_TEQ;
mnemonic = "ld_add";
do_load_add:
if (!prefetch_nofault) {
tcg_gen_qemu_ld_tl(tdest, tsrca, dc->mmuidx, memop);
}
tcg_gen_addi_tl(dest_gr(dc, srca), tsrca, imm);
break;
case OE_IM(LDNA_ADD, X1):
tcg_gen_andi_tl(tdest, tsrca, ~7);
tcg_gen_qemu_ld_tl(tdest, tdest, dc->mmuidx, MO_TEQ);
tcg_gen_addi_tl(dest_gr(dc, srca), tsrca, imm);
mnemonic = "ldna_add";
break;
case OE_IM(ORI, X0):
case OE_IM(ORI, X1):
tcg_gen_ori_tl(tdest, tsrca, imm);
mnemonic = "ori";
break;
case OE_IM(V1ADDI, X0):
case OE_IM(V1ADDI, X1):
t0 = tcg_const_tl(V1_IMM(imm));
gen_v12add(tdest, tsrca, t0, V1_IMM(0x80));
tcg_temp_free(t0);
mnemonic = "v1addi";
break;
case OE_IM(V1CMPEQI, X0):
case OE_IM(V1CMPEQI, X1):
tcg_gen_xori_tl(tdest, tsrca, V1_IMM(imm));
gen_v1cmpeq0(tdest);
mnemonic = "v1cmpeqi";
break;
case OE_IM(V1CMPLTSI, X0):
case OE_IM(V1CMPLTSI, X1):
case OE_IM(V1CMPLTUI, X0):
case OE_IM(V1CMPLTUI, X1):
case OE_IM(V1MAXUI, X0):
case OE_IM(V1MAXUI, X1):
case OE_IM(V1MINUI, X0):
case OE_IM(V1MINUI, X1):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_IM(V2ADDI, X0):
case OE_IM(V2ADDI, X1):
t0 = tcg_const_tl(V2_IMM(imm));
gen_v12add(tdest, tsrca, t0, V2_IMM(0x8000));
tcg_temp_free(t0);
mnemonic = "v2addi";
break;
case OE_IM(V2CMPEQI, X0):
case OE_IM(V2CMPEQI, X1):
case OE_IM(V2CMPLTSI, X0):
case OE_IM(V2CMPLTSI, X1):
case OE_IM(V2CMPLTUI, X0):
case OE_IM(V2CMPLTUI, X1):
case OE_IM(V2MAXSI, X0):
case OE_IM(V2MAXSI, X1):
case OE_IM(V2MINSI, X0):
case OE_IM(V2MINSI, X1):
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
case OE_IM(XORI, X0):
case OE_IM(XORI, X1):
tcg_gen_xori_tl(tdest, tsrca, imm);
mnemonic = "xori";
break;
case OE_SH(ROTLI, X0):
case OE_SH(ROTLI, X1):
case OE_SH(ROTLI, Y0):
case OE_SH(ROTLI, Y1):
tcg_gen_rotli_tl(tdest, tsrca, imm);
mnemonic = "rotli";
break;
case OE_SH(SHLI, X0):
case OE_SH(SHLI, X1):
case OE_SH(SHLI, Y0):
case OE_SH(SHLI, Y1):
tcg_gen_shli_tl(tdest, tsrca, imm);
mnemonic = "shli";
break;
case OE_SH(SHLXI, X0):
case OE_SH(SHLXI, X1):
tcg_gen_shli_tl(tdest, tsrca, imm & 31);
tcg_gen_ext32s_tl(tdest, tdest);
mnemonic = "shlxi";
break;
case OE_SH(SHRSI, X0):
case OE_SH(SHRSI, X1):
case OE_SH(SHRSI, Y0):
case OE_SH(SHRSI, Y1):
tcg_gen_sari_tl(tdest, tsrca, imm);
mnemonic = "shrsi";
break;
case OE_SH(SHRUI, X0):
case OE_SH(SHRUI, X1):
case OE_SH(SHRUI, Y0):
case OE_SH(SHRUI, Y1):
tcg_gen_shri_tl(tdest, tsrca, imm);
mnemonic = "shrui";
break;
case OE_SH(SHRUXI, X0):
case OE_SH(SHRUXI, X1):
if ((imm & 31) == 0) {
tcg_gen_ext32s_tl(tdest, tsrca);
} else {
tcg_gen_ext32u_tl(tdest, tsrca);
tcg_gen_shri_tl(tdest, tdest, imm & 31);
}
mnemonic = "shlxi";
break;
case OE_SH(V1SHLI, X0):
case OE_SH(V1SHLI, X1):
i2 = imm & 7;
i3 = 0xff >> i2;
tcg_gen_andi_tl(tdest, tsrca, V1_IMM(i3));
tcg_gen_shli_tl(tdest, tdest, i2);
mnemonic = "v1shli";
break;
case OE_SH(V1SHRSI, X0):
case OE_SH(V1SHRSI, X1):
t0 = tcg_const_tl(imm & 7);
gen_helper_v1shrs(tdest, tsrca, t0);
tcg_temp_free(t0);
mnemonic = "v1shrsi";
break;
case OE_SH(V1SHRUI, X0):
case OE_SH(V1SHRUI, X1):
i2 = imm & 7;
i3 = (0xff << i2) & 0xff;
tcg_gen_andi_tl(tdest, tsrca, V1_IMM(i3));
tcg_gen_shri_tl(tdest, tdest, i2);
mnemonic = "v1shrui";
break;
case OE_SH(V2SHLI, X0):
case OE_SH(V2SHLI, X1):
i2 = imm & 15;
i3 = 0xffff >> i2;
tcg_gen_andi_tl(tdest, tsrca, V2_IMM(i3));
tcg_gen_shli_tl(tdest, tdest, i2);
mnemonic = "v2shli";
break;
case OE_SH(V2SHRSI, X0):
case OE_SH(V2SHRSI, X1):
t0 = tcg_const_tl(imm & 15);
gen_helper_v2shrs(tdest, tsrca, t0);
tcg_temp_free(t0);
mnemonic = "v2shrsi";
break;
case OE_SH(V2SHRUI, X0):
case OE_SH(V2SHRUI, X1):
i2 = imm & 15;
i3 = (0xffff << i2) & 0xffff;
tcg_gen_andi_tl(tdest, tsrca, V2_IMM(i3));
tcg_gen_shri_tl(tdest, tdest, i2);
mnemonic = "v2shrui";
break;
case OE(ADDLI_OPCODE_X0, 0, X0):
case OE(ADDLI_OPCODE_X1, 0, X1):
tcg_gen_addi_tl(tdest, tsrca, imm);
mnemonic = "addli";
break;
case OE(ADDXLI_OPCODE_X0, 0, X0):
case OE(ADDXLI_OPCODE_X1, 0, X1):
tcg_gen_addi_tl(tdest, tsrca, imm);
tcg_gen_ext32s_tl(tdest, tdest);
mnemonic = "addxli";
break;
case OE(SHL16INSLI_OPCODE_X0, 0, X0):
case OE(SHL16INSLI_OPCODE_X1, 0, X1):
tcg_gen_shli_tl(tdest, tsrca, 16);
tcg_gen_ori_tl(tdest, tdest, imm & 0xffff);
mnemonic = "shl16insli";
break;
default:
return TILEGX_EXCP_OPCODE_UNKNOWN;
}
qemu_log_mask(CPU_LOG_TB_IN_ASM, "%s %s, %s, %d", mnemonic,
reg_names[dest], reg_names[srca], imm);
return TILEGX_EXCP_NONE;
}
static TileExcp gen_bf_opcode_x0(DisasContext *dc, unsigned ext,
unsigned dest, unsigned srca,
unsigned bfs, unsigned bfe)
{
TCGv tdest = dest_gr(dc, dest);
TCGv tsrca = load_gr(dc, srca);
TCGv tsrcd;
int len;
const char *mnemonic;
/* The bitfield is either between E and S inclusive,
or up from S and down from E inclusive. */
if (bfs <= bfe) {
len = bfe - bfs + 1;
} else {
len = (64 - bfs) + (bfe + 1);
}
switch (ext) {
case BFEXTU_BF_OPCODE_X0:
if (bfs == 0 && bfe == 7) {
tcg_gen_ext8u_tl(tdest, tsrca);
} else if (bfs == 0 && bfe == 15) {
tcg_gen_ext16u_tl(tdest, tsrca);
} else if (bfs == 0 && bfe == 31) {
tcg_gen_ext32u_tl(tdest, tsrca);
} else {
int rol = 63 - bfe;
if (bfs <= bfe) {
tcg_gen_shli_tl(tdest, tsrca, rol);
} else {
tcg_gen_rotli_tl(tdest, tsrca, rol);
}
tcg_gen_shri_tl(tdest, tdest, (bfs + rol) & 63);
}
mnemonic = "bfextu";
break;
case BFEXTS_BF_OPCODE_X0:
if (bfs == 0 && bfe == 7) {
tcg_gen_ext8s_tl(tdest, tsrca);
} else if (bfs == 0 && bfe == 15) {
tcg_gen_ext16s_tl(tdest, tsrca);
} else if (bfs == 0 && bfe == 31) {
tcg_gen_ext32s_tl(tdest, tsrca);
} else {
int rol = 63 - bfe;
if (bfs <= bfe) {
tcg_gen_shli_tl(tdest, tsrca, rol);
} else {
tcg_gen_rotli_tl(tdest, tsrca, rol);
}
tcg_gen_sari_tl(tdest, tdest, (bfs + rol) & 63);
}
mnemonic = "bfexts";
break;
case BFINS_BF_OPCODE_X0:
tsrcd = load_gr(dc, dest);
if (bfs <= bfe) {
tcg_gen_deposit_tl(tdest, tsrcd, tsrca, bfs, len);
} else {
tcg_gen_rotri_tl(tdest, tsrcd, bfs);
tcg_gen_deposit_tl(tdest, tdest, tsrca, 0, len);
tcg_gen_rotli_tl(tdest, tdest, bfs);
}
mnemonic = "bfins";
break;
case MM_BF_OPCODE_X0:
tsrcd = load_gr(dc, dest);
if (bfs == 0) {
tcg_gen_deposit_tl(tdest, tsrca, tsrcd, 0, len);
} else {
uint64_t mask = len == 64 ? -1 : rol64((1ULL << len) - 1, bfs);
TCGv tmp = tcg_const_tl(mask);
tcg_gen_and_tl(tdest, tsrcd, tmp);
tcg_gen_andc_tl(tmp, tsrca, tmp);
tcg_gen_or_tl(tdest, tdest, tmp);
tcg_temp_free(tmp);
}
mnemonic = "mm";
break;
default:
return TILEGX_EXCP_OPCODE_UNKNOWN;
}
qemu_log_mask(CPU_LOG_TB_IN_ASM, "%s %s, %s, %u, %u", mnemonic,
reg_names[dest], reg_names[srca], bfs, bfe);
return TILEGX_EXCP_NONE;
}
static TileExcp gen_branch_opcode_x1(DisasContext *dc, unsigned ext,
unsigned srca, int off)
{
target_ulong tgt = dc->pc + off * TILEGX_BUNDLE_SIZE_IN_BYTES;
const char *mnemonic;
dc->jmp.dest = tcg_const_tl(tgt);
dc->jmp.val1 = tcg_temp_new();
tcg_gen_mov_tl(dc->jmp.val1, load_gr(dc, srca));
/* Note that the "predict taken" opcodes have bit 0 clear.
Therefore, fold the two cases together by setting bit 0. */
switch (ext | 1) {
case BEQZ_BRANCH_OPCODE_X1:
dc->jmp.cond = TCG_COND_EQ;
mnemonic = "beqz";
break;
case BNEZ_BRANCH_OPCODE_X1:
dc->jmp.cond = TCG_COND_NE;
mnemonic = "bnez";
break;
case BGEZ_BRANCH_OPCODE_X1:
dc->jmp.cond = TCG_COND_GE;
mnemonic = "bgez";
break;
case BGTZ_BRANCH_OPCODE_X1:
dc->jmp.cond = TCG_COND_GT;
mnemonic = "bgtz";
break;
case BLEZ_BRANCH_OPCODE_X1:
dc->jmp.cond = TCG_COND_LE;
mnemonic = "blez";
break;
case BLTZ_BRANCH_OPCODE_X1:
dc->jmp.cond = TCG_COND_LT;
mnemonic = "bltz";
break;
case BLBC_BRANCH_OPCODE_X1:
dc->jmp.cond = TCG_COND_EQ;
tcg_gen_andi_tl(dc->jmp.val1, dc->jmp.val1, 1);
mnemonic = "blbc";
break;
case BLBS_BRANCH_OPCODE_X1:
dc->jmp.cond = TCG_COND_NE;
tcg_gen_andi_tl(dc->jmp.val1, dc->jmp.val1, 1);
mnemonic = "blbs";
break;
default:
return TILEGX_EXCP_OPCODE_UNKNOWN;
}
if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)) {
qemu_log("%s%s %s, " TARGET_FMT_lx " <%s>",
mnemonic, ext & 1 ? "" : "t",
reg_names[srca], tgt, lookup_symbol(tgt));
}
return TILEGX_EXCP_NONE;
}
static TileExcp gen_jump_opcode_x1(DisasContext *dc, unsigned ext, int off)
{
target_ulong tgt = dc->pc + off * TILEGX_BUNDLE_SIZE_IN_BYTES;
const char *mnemonic = "j";
/* The extension field is 1 bit, therefore we only have JAL and J. */
if (ext == JAL_JUMP_OPCODE_X1) {
tcg_gen_movi_tl(dest_gr(dc, TILEGX_R_LR),
dc->pc + TILEGX_BUNDLE_SIZE_IN_BYTES);
mnemonic = "jal";
}
dc->jmp.cond = TCG_COND_ALWAYS;
dc->jmp.dest = tcg_const_tl(tgt);
if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)) {
qemu_log("%s " TARGET_FMT_lx " <%s>",
mnemonic, tgt, lookup_symbol(tgt));
}
return TILEGX_EXCP_NONE;
}
typedef struct {
const char *name;
intptr_t offset;
void (*get)(TCGv, TCGv_ptr);
void (*put)(TCGv_ptr, TCGv);
} TileSPR;
static const TileSPR *find_spr(unsigned spr)
{
/* Allow the compiler to construct the binary search tree. */
#define D(N, O, G, P) \
case SPR_##N: { static const TileSPR x = { #N, O, G, P }; return &x; }
switch (spr) {
D(CMPEXCH_VALUE,
offsetof(CPUTLGState, spregs[TILEGX_SPR_CMPEXCH]), 0, 0)
D(INTERRUPT_CRITICAL_SECTION,
offsetof(CPUTLGState, spregs[TILEGX_SPR_CRITICAL_SEC]), 0, 0)
D(SIM_CONTROL,
offsetof(CPUTLGState, spregs[TILEGX_SPR_SIM_CONTROL]), 0, 0)
D(EX_CONTEXT_0_0,
offsetof(CPUTLGState, spregs[TILEGX_SPR_EX_CONTEXT_0_0]), 0, 0)
D(EX_CONTEXT_0_1,
offsetof(CPUTLGState, spregs[TILEGX_SPR_EX_CONTEXT_0_1]), 0, 0)
}
#undef D
qemu_log_mask(LOG_UNIMP, "UNIMP SPR %u\n", spr);
return NULL;
}
static TileExcp gen_mtspr_x1(DisasContext *dc, unsigned spr, unsigned srca)
{
const TileSPR *def = find_spr(spr);
TCGv tsrca;
if (def == NULL) {
qemu_log_mask(CPU_LOG_TB_IN_ASM, "mtspr spr[%u], %s", spr, reg_names[srca]);
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
}
tsrca = load_gr(dc, srca);
if (def->put) {
def->put(cpu_env, tsrca);
} else {
tcg_gen_st_tl(tsrca, cpu_env, def->offset);
}
qemu_log_mask(CPU_LOG_TB_IN_ASM, "mtspr %s, %s", def->name, reg_names[srca]);
return TILEGX_EXCP_NONE;
}
static TileExcp gen_mfspr_x1(DisasContext *dc, unsigned dest, unsigned spr)
{
const TileSPR *def = find_spr(spr);
TCGv tdest;
if (def == NULL) {
qemu_log_mask(CPU_LOG_TB_IN_ASM, "mtspr %s, spr[%u]", reg_names[dest], spr);
return TILEGX_EXCP_OPCODE_UNIMPLEMENTED;
}
tdest = dest_gr(dc, dest);
if (def->get) {
def->get(tdest, cpu_env);
} else {
tcg_gen_ld_tl(tdest, cpu_env, def->offset);
}
qemu_log_mask(CPU_LOG_TB_IN_ASM, "mfspr %s, %s", reg_names[dest], def->name);
return TILEGX_EXCP_NONE;
}
static TileExcp decode_y0(DisasContext *dc, tilegx_bundle_bits bundle)
{
unsigned opc = get_Opcode_Y0(bundle);
unsigned ext = get_RRROpcodeExtension_Y0(bundle);
unsigned dest = get_Dest_Y0(bundle);
unsigned srca = get_SrcA_Y0(bundle);
unsigned srcb;
int imm;
switch (opc) {
case RRR_1_OPCODE_Y0:
if (ext == UNARY_RRR_1_OPCODE_Y0) {
ext = get_UnaryOpcodeExtension_Y0(bundle);
return gen_rr_opcode(dc, OE(opc, ext, Y0), dest, srca, bundle);
}
/* fallthru */
case RRR_0_OPCODE_Y0:
case RRR_2_OPCODE_Y0:
case RRR_3_OPCODE_Y0:
case RRR_4_OPCODE_Y0:
case RRR_5_OPCODE_Y0:
case RRR_6_OPCODE_Y0:
case RRR_7_OPCODE_Y0:
case RRR_8_OPCODE_Y0:
case RRR_9_OPCODE_Y0:
srcb = get_SrcB_Y0(bundle);
return gen_rrr_opcode(dc, OE(opc, ext, Y0), dest, srca, srcb);
case SHIFT_OPCODE_Y0:
ext = get_ShiftOpcodeExtension_Y0(bundle);
imm = get_ShAmt_Y0(bundle);
return gen_rri_opcode(dc, OE(opc, ext, Y0), dest, srca, imm);
case ADDI_OPCODE_Y0:
case ADDXI_OPCODE_Y0:
case ANDI_OPCODE_Y0:
case CMPEQI_OPCODE_Y0:
case CMPLTSI_OPCODE_Y0:
imm = (int8_t)get_Imm8_Y0(bundle);
return gen_rri_opcode(dc, OE(opc, 0, Y0), dest, srca, imm);
default:
return TILEGX_EXCP_OPCODE_UNKNOWN;
}
}
static TileExcp decode_y1(DisasContext *dc, tilegx_bundle_bits bundle)
{
unsigned opc = get_Opcode_Y1(bundle);
unsigned ext = get_RRROpcodeExtension_Y1(bundle);
unsigned dest = get_Dest_Y1(bundle);
unsigned srca = get_SrcA_Y1(bundle);
unsigned srcb;
int imm;
switch (get_Opcode_Y1(bundle)) {
case RRR_1_OPCODE_Y1:
if (ext == UNARY_RRR_1_OPCODE_Y0) {
ext = get_UnaryOpcodeExtension_Y1(bundle);
return gen_rr_opcode(dc, OE(opc, ext, Y1), dest, srca, bundle);
}
/* fallthru */
case RRR_0_OPCODE_Y1:
case RRR_2_OPCODE_Y1:
case RRR_3_OPCODE_Y1:
case RRR_4_OPCODE_Y1:
case RRR_5_OPCODE_Y1:
case RRR_6_OPCODE_Y1:
case RRR_7_OPCODE_Y1:
srcb = get_SrcB_Y1(bundle);
return gen_rrr_opcode(dc, OE(opc, ext, Y1), dest, srca, srcb);
case SHIFT_OPCODE_Y1:
ext = get_ShiftOpcodeExtension_Y1(bundle);
imm = get_ShAmt_Y1(bundle);
return gen_rri_opcode(dc, OE(opc, ext, Y1), dest, srca, imm);
case ADDI_OPCODE_Y1:
case ADDXI_OPCODE_Y1:
case ANDI_OPCODE_Y1:
case CMPEQI_OPCODE_Y1:
case CMPLTSI_OPCODE_Y1:
imm = (int8_t)get_Imm8_Y1(bundle);
return gen_rri_opcode(dc, OE(opc, 0, Y1), dest, srca, imm);
default:
return TILEGX_EXCP_OPCODE_UNKNOWN;
}
}
static TileExcp decode_y2(DisasContext *dc, tilegx_bundle_bits bundle)
{
unsigned mode = get_Mode(bundle);
unsigned opc = get_Opcode_Y2(bundle);
unsigned srca = get_SrcA_Y2(bundle);
unsigned srcbdest = get_SrcBDest_Y2(bundle);
const char *mnemonic;
TCGMemOp memop;
bool prefetch_nofault = false;
switch (OEY2(opc, mode)) {
case OEY2(LD1S_OPCODE_Y2, MODE_OPCODE_YA2):
memop = MO_SB;
mnemonic = "ld1s"; /* prefetch_l1_fault */
goto do_load;
case OEY2(LD1U_OPCODE_Y2, MODE_OPCODE_YA2):
memop = MO_UB;
mnemonic = "ld1u"; /* prefetch, prefetch_l1 */
prefetch_nofault = (srcbdest == TILEGX_R_ZERO);
goto do_load;
case OEY2(LD2S_OPCODE_Y2, MODE_OPCODE_YA2):
memop = MO_TESW;
mnemonic = "ld2s"; /* prefetch_l2_fault */
goto do_load;
case OEY2(LD2U_OPCODE_Y2, MODE_OPCODE_YA2):
memop = MO_TEUW;
mnemonic = "ld2u"; /* prefetch_l2 */
prefetch_nofault = (srcbdest == TILEGX_R_ZERO);
goto do_load;
case OEY2(LD4S_OPCODE_Y2, MODE_OPCODE_YB2):
memop = MO_TESL;
mnemonic = "ld4s"; /* prefetch_l3_fault */
goto do_load;
case OEY2(LD4U_OPCODE_Y2, MODE_OPCODE_YB2):
memop = MO_TEUL;
mnemonic = "ld4u"; /* prefetch_l3 */
prefetch_nofault = (srcbdest == TILEGX_R_ZERO);
goto do_load;
case OEY2(LD_OPCODE_Y2, MODE_OPCODE_YB2):
memop = MO_TEQ;
mnemonic = "ld";
do_load:
if (!prefetch_nofault) {
tcg_gen_qemu_ld_tl(dest_gr(dc, srcbdest), load_gr(dc, srca),
dc->mmuidx, memop);
}
qemu_log_mask(CPU_LOG_TB_IN_ASM, "%s %s, %s", mnemonic,
reg_names[srcbdest], reg_names[srca]);
return TILEGX_EXCP_NONE;
case OEY2(ST1_OPCODE_Y2, MODE_OPCODE_YC2):
return gen_st_opcode(dc, 0, srca, srcbdest, MO_UB, "st1");
case OEY2(ST2_OPCODE_Y2, MODE_OPCODE_YC2):
return gen_st_opcode(dc, 0, srca, srcbdest, MO_TEUW, "st2");
case OEY2(ST4_OPCODE_Y2, MODE_OPCODE_YC2):
return gen_st_opcode(dc, 0, srca, srcbdest, MO_TEUL, "st4");
case OEY2(ST_OPCODE_Y2, MODE_OPCODE_YC2):
return gen_st_opcode(dc, 0, srca, srcbdest, MO_TEQ, "st");
default:
return TILEGX_EXCP_OPCODE_UNKNOWN;
}
}
static TileExcp decode_x0(DisasContext *dc, tilegx_bundle_bits bundle)
{
unsigned opc = get_Opcode_X0(bundle);
unsigned dest = get_Dest_X0(bundle);
unsigned srca = get_SrcA_X0(bundle);
unsigned ext, srcb, bfs, bfe;
int imm;
switch (opc) {
case RRR_0_OPCODE_X0:
ext = get_RRROpcodeExtension_X0(bundle);
if (ext == UNARY_RRR_0_OPCODE_X0) {
ext = get_UnaryOpcodeExtension_X0(bundle);
return gen_rr_opcode(dc, OE(opc, ext, X0), dest, srca, bundle);
}
srcb = get_SrcB_X0(bundle);
return gen_rrr_opcode(dc, OE(opc, ext, X0), dest, srca, srcb);
case SHIFT_OPCODE_X0:
ext = get_ShiftOpcodeExtension_X0(bundle);
imm = get_ShAmt_X0(bundle);
return gen_rri_opcode(dc, OE(opc, ext, X0), dest, srca, imm);
case IMM8_OPCODE_X0:
ext = get_Imm8OpcodeExtension_X0(bundle);
imm = (int8_t)get_Imm8_X0(bundle);
return gen_rri_opcode(dc, OE(opc, ext, X0), dest, srca, imm);
case BF_OPCODE_X0:
ext = get_BFOpcodeExtension_X0(bundle);
bfs = get_BFStart_X0(bundle);
bfe = get_BFEnd_X0(bundle);
return gen_bf_opcode_x0(dc, ext, dest, srca, bfs, bfe);
case ADDLI_OPCODE_X0:
case SHL16INSLI_OPCODE_X0:
case ADDXLI_OPCODE_X0:
imm = (int16_t)get_Imm16_X0(bundle);
return gen_rri_opcode(dc, OE(opc, 0, X0), dest, srca, imm);
default:
return TILEGX_EXCP_OPCODE_UNKNOWN;
}
}
static TileExcp decode_x1(DisasContext *dc, tilegx_bundle_bits bundle)
{
unsigned opc = get_Opcode_X1(bundle);
unsigned dest = get_Dest_X1(bundle);
unsigned srca = get_SrcA_X1(bundle);
unsigned ext, srcb;
int imm;
switch (opc) {
case RRR_0_OPCODE_X1:
ext = get_RRROpcodeExtension_X1(bundle);
srcb = get_SrcB_X1(bundle);
switch (ext) {
case UNARY_RRR_0_OPCODE_X1:
ext = get_UnaryOpcodeExtension_X1(bundle);
return gen_rr_opcode(dc, OE(opc, ext, X1), dest, srca, bundle);
case ST1_RRR_0_OPCODE_X1:
return gen_st_opcode(dc, dest, srca, srcb, MO_UB, "st1");
case ST2_RRR_0_OPCODE_X1:
return gen_st_opcode(dc, dest, srca, srcb, MO_TEUW, "st2");
case ST4_RRR_0_OPCODE_X1:
return gen_st_opcode(dc, dest, srca, srcb, MO_TEUL, "st4");
case STNT1_RRR_0_OPCODE_X1:
return gen_st_opcode(dc, dest, srca, srcb, MO_UB, "stnt1");
case STNT2_RRR_0_OPCODE_X1:
return gen_st_opcode(dc, dest, srca, srcb, MO_TEUW, "stnt2");
case STNT4_RRR_0_OPCODE_X1:
return gen_st_opcode(dc, dest, srca, srcb, MO_TEUL, "stnt4");
case STNT_RRR_0_OPCODE_X1:
return gen_st_opcode(dc, dest, srca, srcb, MO_TEQ, "stnt");
case ST_RRR_0_OPCODE_X1:
return gen_st_opcode(dc, dest, srca, srcb, MO_TEQ, "st");
}
return gen_rrr_opcode(dc, OE(opc, ext, X1), dest, srca, srcb);
case SHIFT_OPCODE_X1:
ext = get_ShiftOpcodeExtension_X1(bundle);
imm = get_ShAmt_X1(bundle);
return gen_rri_opcode(dc, OE(opc, ext, X1), dest, srca, imm);
case IMM8_OPCODE_X1:
ext = get_Imm8OpcodeExtension_X1(bundle);
imm = (int8_t)get_Dest_Imm8_X1(bundle);
srcb = get_SrcB_X1(bundle);
switch (ext) {
case ST1_ADD_IMM8_OPCODE_X1:
return gen_st_add_opcode(dc, srca, srcb, imm, MO_UB, "st1_add");
case ST2_ADD_IMM8_OPCODE_X1:
return gen_st_add_opcode(dc, srca, srcb, imm, MO_TEUW, "st2_add");
case ST4_ADD_IMM8_OPCODE_X1:
return gen_st_add_opcode(dc, srca, srcb, imm, MO_TEUL, "st4_add");
case STNT1_ADD_IMM8_OPCODE_X1:
return gen_st_add_opcode(dc, srca, srcb, imm, MO_UB, "stnt1_add");
case STNT2_ADD_IMM8_OPCODE_X1:
return gen_st_add_opcode(dc, srca, srcb, imm, MO_TEUW, "stnt2_add");
case STNT4_ADD_IMM8_OPCODE_X1:
return gen_st_add_opcode(dc, srca, srcb, imm, MO_TEUL, "stnt4_add");
case STNT_ADD_IMM8_OPCODE_X1:
return gen_st_add_opcode(dc, srca, srcb, imm, MO_TEQ, "stnt_add");
case ST_ADD_IMM8_OPCODE_X1:
return gen_st_add_opcode(dc, srca, srcb, imm, MO_TEQ, "st_add");
case MFSPR_IMM8_OPCODE_X1:
return gen_mfspr_x1(dc, dest, get_MF_Imm14_X1(bundle));
case MTSPR_IMM8_OPCODE_X1:
return gen_mtspr_x1(dc, get_MT_Imm14_X1(bundle), srca);
}
imm = (int8_t)get_Imm8_X1(bundle);
return gen_rri_opcode(dc, OE(opc, ext, X1), dest, srca, imm);
case BRANCH_OPCODE_X1:
ext = get_BrType_X1(bundle);
imm = sextract32(get_BrOff_X1(bundle), 0, 17);
return gen_branch_opcode_x1(dc, ext, srca, imm);
case JUMP_OPCODE_X1:
ext = get_JumpOpcodeExtension_X1(bundle);
imm = sextract32(get_JumpOff_X1(bundle), 0, 27);
return gen_jump_opcode_x1(dc, ext, imm);
case ADDLI_OPCODE_X1:
case SHL16INSLI_OPCODE_X1:
case ADDXLI_OPCODE_X1:
imm = (int16_t)get_Imm16_X1(bundle);
return gen_rri_opcode(dc, OE(opc, 0, X1), dest, srca, imm);
default:
return TILEGX_EXCP_OPCODE_UNKNOWN;
}
}
static void notice_excp(DisasContext *dc, uint64_t bundle,
const char *type, TileExcp excp)
{
if (likely(excp == TILEGX_EXCP_NONE)) {
return;
}
gen_exception(dc, excp);
switch (excp) {
case TILEGX_EXCP_OPCODE_UNIMPLEMENTED:
qemu_log_mask(LOG_UNIMP, "UNIMP %s, [" FMT64X "]\n", type, bundle);
break;
case TILEGX_EXCP_OPCODE_UNKNOWN:
qemu_log_mask(LOG_UNIMP, "UNKNOWN %s, [" FMT64X "]\n", type, bundle);
break;
default:
break;
}
}
static void translate_one_bundle(DisasContext *dc, uint64_t bundle)
{
int i;
for (i = 0; i < ARRAY_SIZE(dc->wb); i++) {
DisasContextTemp *wb = &dc->wb[i];
wb->reg = TILEGX_R_NOREG;
wb->val = NULL;
}
dc->num_wb = 0;
qemu_log_mask(CPU_LOG_TB_IN_ASM, " %" PRIx64 ": { ", dc->pc);
if (get_Mode(bundle)) {
notice_excp(dc, bundle, "y0", decode_y0(dc, bundle));
qemu_log_mask(CPU_LOG_TB_IN_ASM, " ; ");
notice_excp(dc, bundle, "y1", decode_y1(dc, bundle));
qemu_log_mask(CPU_LOG_TB_IN_ASM, " ; ");
notice_excp(dc, bundle, "y2", decode_y2(dc, bundle));
} else {
notice_excp(dc, bundle, "x0", decode_x0(dc, bundle));
qemu_log_mask(CPU_LOG_TB_IN_ASM, " ; ");
notice_excp(dc, bundle, "x1", decode_x1(dc, bundle));
}
qemu_log_mask(CPU_LOG_TB_IN_ASM, " }\n");
for (i = dc->num_wb - 1; i >= 0; --i) {
DisasContextTemp *wb = &dc->wb[i];
if (wb->reg < TILEGX_R_COUNT) {
tcg_gen_mov_i64(cpu_regs[wb->reg], wb->val);
}
tcg_temp_free_i64(wb->val);
}
if (dc->jmp.cond != TCG_COND_NEVER) {
if (dc->jmp.cond == TCG_COND_ALWAYS) {
tcg_gen_mov_i64(cpu_pc, dc->jmp.dest);
} else {
TCGv next = tcg_const_i64(dc->pc + TILEGX_BUNDLE_SIZE_IN_BYTES);
tcg_gen_movcond_i64(dc->jmp.cond, cpu_pc,
dc->jmp.val1, load_zero(dc),
dc->jmp.dest, next);
tcg_temp_free_i64(dc->jmp.val1);
tcg_temp_free_i64(next);
}
tcg_temp_free_i64(dc->jmp.dest);
tcg_gen_exit_tb(NULL, 0);
dc->exit_tb = true;
} else if (dc->atomic_excp != TILEGX_EXCP_NONE) {
gen_exception(dc, dc->atomic_excp);
}
}
void gen_intermediate_code(CPUState *cs, TranslationBlock *tb, int max_insns)
{
CPUTLGState *env = cs->env_ptr;
DisasContext ctx;
DisasContext *dc = &ctx;
uint64_t pc_start = tb->pc;
uint64_t page_start = pc_start & TARGET_PAGE_MASK;
int num_insns = 0;
dc->pc = pc_start;
dc->mmuidx = 0;
dc->exit_tb = false;
dc->atomic_excp = TILEGX_EXCP_NONE;
dc->jmp.cond = TCG_COND_NEVER;
dc->jmp.dest = NULL;
dc->jmp.val1 = NULL;
dc->zero = NULL;
if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)) {
qemu_log_lock();
qemu_log("IN: %s\n", lookup_symbol(pc_start));
}
gen_tb_start(tb);
while (1) {
tcg_gen_insn_start(dc->pc);
num_insns++;
translate_one_bundle(dc, cpu_ldq_data(env, dc->pc));
if (dc->exit_tb) {
/* PC updated and EXIT_TB/GOTO_TB/exception emitted. */
break;
}
dc->pc += TILEGX_BUNDLE_SIZE_IN_BYTES;
if (num_insns >= max_insns
|| (dc->pc - page_start >= TARGET_PAGE_SIZE)
|| tcg_op_buf_full()) {
/* Ending the TB due to TB size or page boundary. Set PC. */
tcg_gen_movi_tl(cpu_pc, dc->pc);
tcg_gen_exit_tb(NULL, 0);
break;
}
}
gen_tb_end(tb, num_insns);
tb->size = dc->pc - pc_start;
tb->icount = num_insns;
if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM)) {
qemu_log("\n");
qemu_log_unlock();
}
}
void restore_state_to_opc(CPUTLGState *env, TranslationBlock *tb,
target_ulong *data)
{
env->pc = data[0];
}
void tilegx_tcg_init(void)
{
int i;
cpu_pc = tcg_global_mem_new_i64(cpu_env, offsetof(CPUTLGState, pc), "pc");
for (i = 0; i < TILEGX_R_COUNT; i++) {
cpu_regs[i] = tcg_global_mem_new_i64(cpu_env,
offsetof(CPUTLGState, regs[i]),
reg_names[i]);
}
}