target/ppc: Move VMX storage access instructions to decodetree

Moving the following instructions to decodetree specification :

	{l,st}ve{b,h,w}x,
	{l,st}v{x,xl},
	lvs{l,r}		: X-form

The changes were verified by validating that the tcg ops generated by those
instructions remain the same, which were captured using the '-d in_asm,op' flag.

Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
Signed-off-by: Chinmay Rath <rathc@linux.ibm.com>
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
This commit is contained in:
Chinmay Rath 2024-04-29 10:43:15 +05:30 committed by Nicholas Piggin
parent 948e257c48
commit 21b5f5464f
6 changed files with 119 additions and 162 deletions

View File

@ -267,12 +267,12 @@ DEF_HELPER_5(VMSUMSHS, void, env, avr, avr, avr, avr)
DEF_HELPER_FLAGS_5(VMLADDUHM, TCG_CALL_NO_RWG, void, avr, avr, avr, avr, i32) DEF_HELPER_FLAGS_5(VMLADDUHM, TCG_CALL_NO_RWG, void, avr, avr, avr, avr, i32)
DEF_HELPER_FLAGS_2(mtvscr, TCG_CALL_NO_RWG, void, env, i32) DEF_HELPER_FLAGS_2(mtvscr, TCG_CALL_NO_RWG, void, env, i32)
DEF_HELPER_FLAGS_1(mfvscr, TCG_CALL_NO_RWG, i32, env) DEF_HELPER_FLAGS_1(mfvscr, TCG_CALL_NO_RWG, i32, env)
DEF_HELPER_3(lvebx, void, env, avr, tl) DEF_HELPER_3(LVEBX, void, env, avr, tl)
DEF_HELPER_3(lvehx, void, env, avr, tl) DEF_HELPER_3(LVEHX, void, env, avr, tl)
DEF_HELPER_3(lvewx, void, env, avr, tl) DEF_HELPER_3(LVEWX, void, env, avr, tl)
DEF_HELPER_3(stvebx, void, env, avr, tl) DEF_HELPER_3(STVEBX, void, env, avr, tl)
DEF_HELPER_3(stvehx, void, env, avr, tl) DEF_HELPER_3(STVEHX, void, env, avr, tl)
DEF_HELPER_3(stvewx, void, env, avr, tl) DEF_HELPER_3(STVEWX, void, env, avr, tl)
#if defined(TARGET_PPC64) #if defined(TARGET_PPC64)
DEF_HELPER_4(lxvl, void, env, tl, vsr, tl) DEF_HELPER_4(lxvl, void, env, tl, vsr, tl)
DEF_HELPER_4(lxvll, void, env, tl, vsr, tl) DEF_HELPER_4(lxvll, void, env, tl, vsr, tl)

View File

@ -661,6 +661,23 @@ DSCRIQ 111111 ..... ..... ...... 001100010 . @Z22_tap_sh_rc
VPMSUMD 000100 ..... ..... ..... 10011001000 @VX VPMSUMD 000100 ..... ..... ..... 10011001000 @VX
## Vector Load/Store Instructions
LVEBX 011111 ..... ..... ..... 0000000111 - @X
LVEHX 011111 ..... ..... ..... 0000100111 - @X
LVEWX 011111 ..... ..... ..... 0001000111 - @X
LVX 011111 ..... ..... ..... 0001100111 - @X
LVXL 011111 ..... ..... ..... 0101100111 - @X
STVEBX 011111 ..... ..... ..... 0010000111 - @X
STVEHX 011111 ..... ..... ..... 0010100111 - @X
STVEWX 011111 ..... ..... ..... 0011000111 - @X
STVX 011111 ..... ..... ..... 0011100111 - @X
STVXL 011111 ..... ..... ..... 0111100111 - @X
LVSL 011111 ..... ..... ..... 0000000110 - @X
LVSR 011111 ..... ..... ..... 0000100110 - @X
## Vector Integer Instructions ## Vector Integer Instructions
VCMPEQUB 000100 ..... ..... ..... . 0000000110 @VC VCMPEQUB 000100 ..... ..... ..... . 0000000110 @VC

View File

@ -404,9 +404,9 @@ target_ulong helper_lscbx(CPUPPCState *env, target_ulong addr, uint32_t reg,
} \ } \
} }
#define I(x) (x) #define I(x) (x)
LVE(lvebx, cpu_ldub_data_ra, I, u8) LVE(LVEBX, cpu_ldub_data_ra, I, u8)
LVE(lvehx, cpu_lduw_data_ra, bswap16, u16) LVE(LVEHX, cpu_lduw_data_ra, bswap16, u16)
LVE(lvewx, cpu_ldl_data_ra, bswap32, u32) LVE(LVEWX, cpu_ldl_data_ra, bswap32, u32)
#undef I #undef I
#undef LVE #undef LVE
@ -432,9 +432,9 @@ LVE(lvewx, cpu_ldl_data_ra, bswap32, u32)
} \ } \
} }
#define I(x) (x) #define I(x) (x)
STVE(stvebx, cpu_stb_data_ra, I, u8) STVE(STVEBX, cpu_stb_data_ra, I, u8)
STVE(stvehx, cpu_stw_data_ra, bswap16, u16) STVE(STVEHX, cpu_stw_data_ra, bswap16, u16)
STVE(stvewx, cpu_stl_data_ra, bswap32, u32) STVE(STVEWX, cpu_stl_data_ra, bswap32, u32)
#undef I #undef I
#undef LVE #undef LVE

View File

@ -5790,8 +5790,6 @@ GEN_HANDLER2_E(icbt_440, "icbt", 0x1F, 0x16, 0x00, 0x03E00001,
PPC_BOOKE, PPC2_BOOKE206), PPC_BOOKE, PPC2_BOOKE206),
GEN_HANDLER2(icbt_440, "icbt", 0x1F, 0x06, 0x08, 0x03E00001, GEN_HANDLER2(icbt_440, "icbt", 0x1F, 0x06, 0x08, 0x03E00001,
PPC_440_SPEC), PPC_440_SPEC),
GEN_HANDLER(lvsl, 0x1f, 0x06, 0x00, 0x00000001, PPC_ALTIVEC),
GEN_HANDLER(lvsr, 0x1f, 0x06, 0x01, 0x00000001, PPC_ALTIVEC),
GEN_HANDLER(mfvscr, 0x04, 0x2, 0x18, 0x001ff800, PPC_ALTIVEC), GEN_HANDLER(mfvscr, 0x04, 0x2, 0x18, 0x001ff800, PPC_ALTIVEC),
GEN_HANDLER(mtvscr, 0x04, 0x2, 0x19, 0x03ff0000, PPC_ALTIVEC), GEN_HANDLER(mtvscr, 0x04, 0x2, 0x19, 0x03ff0000, PPC_ALTIVEC),

View File

@ -14,125 +14,88 @@ static inline TCGv_ptr gen_avr_ptr(int reg)
return r; return r;
} }
#define GEN_VR_LDX(name, opc2, opc3) \ static bool trans_LVX(DisasContext *ctx, arg_X *a)
static void glue(gen_, name)(DisasContext *ctx) \ {
{ \ TCGv EA;
TCGv EA; \ TCGv_i64 avr;
TCGv_i64 avr; \ REQUIRE_INSNS_FLAGS(ctx, ALTIVEC);
if (unlikely(!ctx->altivec_enabled)) { \ REQUIRE_VECTOR(ctx);
gen_exception(ctx, POWERPC_EXCP_VPU); \ gen_set_access_type(ctx, ACCESS_INT);
return; \ avr = tcg_temp_new_i64();
} \ EA = do_ea_calc(ctx, a->ra, cpu_gpr[a->rb]);
gen_set_access_type(ctx, ACCESS_INT); \ tcg_gen_andi_tl(EA, EA, ~0xf);
avr = tcg_temp_new_i64(); \ /*
EA = tcg_temp_new(); \ * We only need to swap high and low halves. gen_qemu_ld64_i64
gen_addr_reg_index(ctx, EA); \ * does necessary 64-bit byteswap already.
tcg_gen_andi_tl(EA, EA, ~0xf); \ */
/* \ gen_qemu_ld64_i64(ctx, avr, EA);
* We only need to swap high and low halves. gen_qemu_ld64_i64 \ set_avr64(a->rt, avr, !ctx->le_mode);
* does necessary 64-bit byteswap already. \ tcg_gen_addi_tl(EA, EA, 8);
*/ \ gen_qemu_ld64_i64(ctx, avr, EA);
if (ctx->le_mode) { \ set_avr64(a->rt, avr, ctx->le_mode);
gen_qemu_ld64_i64(ctx, avr, EA); \ return true;
set_avr64(rD(ctx->opcode), avr, false); \
tcg_gen_addi_tl(EA, EA, 8); \
gen_qemu_ld64_i64(ctx, avr, EA); \
set_avr64(rD(ctx->opcode), avr, true); \
} else { \
gen_qemu_ld64_i64(ctx, avr, EA); \
set_avr64(rD(ctx->opcode), avr, true); \
tcg_gen_addi_tl(EA, EA, 8); \
gen_qemu_ld64_i64(ctx, avr, EA); \
set_avr64(rD(ctx->opcode), avr, false); \
} \
} }
#define GEN_VR_STX(name, opc2, opc3) \
static void gen_st##name(DisasContext *ctx) \
{ \
TCGv EA; \
TCGv_i64 avr; \
if (unlikely(!ctx->altivec_enabled)) { \
gen_exception(ctx, POWERPC_EXCP_VPU); \
return; \
} \
gen_set_access_type(ctx, ACCESS_INT); \
avr = tcg_temp_new_i64(); \
EA = tcg_temp_new(); \
gen_addr_reg_index(ctx, EA); \
tcg_gen_andi_tl(EA, EA, ~0xf); \
/* \
* We only need to swap high and low halves. gen_qemu_st64_i64 \
* does necessary 64-bit byteswap already. \
*/ \
if (ctx->le_mode) { \
get_avr64(avr, rD(ctx->opcode), false); \
gen_qemu_st64_i64(ctx, avr, EA); \
tcg_gen_addi_tl(EA, EA, 8); \
get_avr64(avr, rD(ctx->opcode), true); \
gen_qemu_st64_i64(ctx, avr, EA); \
} else { \
get_avr64(avr, rD(ctx->opcode), true); \
gen_qemu_st64_i64(ctx, avr, EA); \
tcg_gen_addi_tl(EA, EA, 8); \
get_avr64(avr, rD(ctx->opcode), false); \
gen_qemu_st64_i64(ctx, avr, EA); \
} \
}
#define GEN_VR_LVE(name, opc2, opc3, size) \
static void gen_lve##name(DisasContext *ctx) \
{ \
TCGv EA; \
TCGv_ptr rs; \
if (unlikely(!ctx->altivec_enabled)) { \
gen_exception(ctx, POWERPC_EXCP_VPU); \
return; \
} \
gen_set_access_type(ctx, ACCESS_INT); \
EA = tcg_temp_new(); \
gen_addr_reg_index(ctx, EA); \
if (size > 1) { \
tcg_gen_andi_tl(EA, EA, ~(size - 1)); \
} \
rs = gen_avr_ptr(rS(ctx->opcode)); \
gen_helper_lve##name(tcg_env, rs, EA); \
}
#define GEN_VR_STVE(name, opc2, opc3, size) \
static void gen_stve##name(DisasContext *ctx) \
{ \
TCGv EA; \
TCGv_ptr rs; \
if (unlikely(!ctx->altivec_enabled)) { \
gen_exception(ctx, POWERPC_EXCP_VPU); \
return; \
} \
gen_set_access_type(ctx, ACCESS_INT); \
EA = tcg_temp_new(); \
gen_addr_reg_index(ctx, EA); \
if (size > 1) { \
tcg_gen_andi_tl(EA, EA, ~(size - 1)); \
} \
rs = gen_avr_ptr(rS(ctx->opcode)); \
gen_helper_stve##name(tcg_env, rs, EA); \
}
GEN_VR_LDX(lvx, 0x07, 0x03);
/* As we don't emulate the cache, lvxl is strictly equivalent to lvx */ /* As we don't emulate the cache, lvxl is strictly equivalent to lvx */
GEN_VR_LDX(lvxl, 0x07, 0x0B); QEMU_FLATTEN
static bool trans_LVXL(DisasContext *ctx, arg_LVXL *a)
{
return trans_LVX(ctx, a);
}
GEN_VR_LVE(bx, 0x07, 0x00, 1); static bool trans_STVX(DisasContext *ctx, arg_STVX *a)
GEN_VR_LVE(hx, 0x07, 0x01, 2); {
GEN_VR_LVE(wx, 0x07, 0x02, 4); TCGv EA;
TCGv_i64 avr;
REQUIRE_INSNS_FLAGS(ctx, ALTIVEC);
REQUIRE_VECTOR(ctx);
gen_set_access_type(ctx, ACCESS_INT);
avr = tcg_temp_new_i64();
EA = do_ea_calc(ctx, a->ra, cpu_gpr[a->rb]);
tcg_gen_andi_tl(EA, EA, ~0xf);
/*
* We only need to swap high and low halves. gen_qemu_st64_i64
* does necessary 64-bit byteswap already.
*/
get_avr64(avr, a->rt, !ctx->le_mode);
gen_qemu_st64_i64(ctx, avr, EA);
tcg_gen_addi_tl(EA, EA, 8);
get_avr64(avr, a->rt, ctx->le_mode);
gen_qemu_st64_i64(ctx, avr, EA);
return true;
}
GEN_VR_STX(svx, 0x07, 0x07);
/* As we don't emulate the cache, stvxl is strictly equivalent to stvx */ /* As we don't emulate the cache, stvxl is strictly equivalent to stvx */
GEN_VR_STX(svxl, 0x07, 0x0F); QEMU_FLATTEN
static bool trans_STVXL(DisasContext *ctx, arg_STVXL *a)
{
return trans_STVX(ctx, a);
}
GEN_VR_STVE(bx, 0x07, 0x04, 1); static bool do_ldst_ve_X(DisasContext *ctx, arg_X *a, int size,
GEN_VR_STVE(hx, 0x07, 0x05, 2); void (*helper)(TCGv_env, TCGv_ptr, TCGv))
GEN_VR_STVE(wx, 0x07, 0x06, 4); {
TCGv EA;
TCGv_ptr vrt;
REQUIRE_INSNS_FLAGS(ctx, ALTIVEC);
REQUIRE_VECTOR(ctx);
gen_set_access_type(ctx, ACCESS_INT);
EA = do_ea_calc(ctx, a->ra, cpu_gpr[a->rb]);
if (size > 1) {
tcg_gen_andi_tl(EA, EA, ~(size - 1));
}
vrt = gen_avr_ptr(a->rt);
helper(tcg_env, vrt, EA);
return true;
}
TRANS(LVEBX, do_ldst_ve_X, 1, gen_helper_LVEBX);
TRANS(LVEHX, do_ldst_ve_X, 2, gen_helper_LVEHX);
TRANS(LVEWX, do_ldst_ve_X, 4, gen_helper_LVEWX);
TRANS(STVEBX, do_ldst_ve_X, 1, gen_helper_STVEBX);
TRANS(STVEHX, do_ldst_ve_X, 2, gen_helper_STVEHX);
TRANS(STVEWX, do_ldst_ve_X, 4, gen_helper_STVEWX);
static void gen_mfvscr(DisasContext *ctx) static void gen_mfvscr(DisasContext *ctx)
{ {
@ -460,15 +423,17 @@ static void trans_vmrgow(DisasContext *ctx)
* Let X be the 32-byte value 0x00 || 0x01 || 0x02 || ... || 0x1E || 0x1F. * Let X be the 32-byte value 0x00 || 0x01 || 0x02 || ... || 0x1E || 0x1F.
* Bytes sh:sh+15 of X are placed into vD. * Bytes sh:sh+15 of X are placed into vD.
*/ */
static void trans_lvsl(DisasContext *ctx) static bool trans_LVSL(DisasContext *ctx, arg_LVSL *a)
{ {
int VT = rD(ctx->opcode);
TCGv_i64 result = tcg_temp_new_i64(); TCGv_i64 result = tcg_temp_new_i64();
TCGv_i64 sh = tcg_temp_new_i64(); TCGv_i64 sh = tcg_temp_new_i64();
TCGv EA = tcg_temp_new(); TCGv EA = tcg_temp_new();
REQUIRE_INSNS_FLAGS(ctx, ALTIVEC);
REQUIRE_VECTOR(ctx);
/* Get sh(from description) by anding EA with 0xf. */ /* Get sh(from description) by anding EA with 0xf. */
gen_addr_reg_index(ctx, EA); EA = do_ea_calc(ctx, a->ra, cpu_gpr[a->rb]);
tcg_gen_extu_tl_i64(sh, EA); tcg_gen_extu_tl_i64(sh, EA);
tcg_gen_andi_i64(sh, sh, 0xfULL); tcg_gen_andi_i64(sh, sh, 0xfULL);
@ -478,13 +443,14 @@ static void trans_lvsl(DisasContext *ctx)
*/ */
tcg_gen_muli_i64(sh, sh, 0x0101010101010101ULL); tcg_gen_muli_i64(sh, sh, 0x0101010101010101ULL);
tcg_gen_addi_i64(result, sh, 0x0001020304050607ull); tcg_gen_addi_i64(result, sh, 0x0001020304050607ull);
set_avr64(VT, result, true); set_avr64(a->rt, result, true);
/* /*
* Create bytes sh+8:sh+15 of X(from description) and place them in * Create bytes sh+8:sh+15 of X(from description) and place them in
* lower doubleword of vD. * lower doubleword of vD.
*/ */
tcg_gen_addi_i64(result, sh, 0x08090a0b0c0d0e0fULL); tcg_gen_addi_i64(result, sh, 0x08090a0b0c0d0e0fULL);
set_avr64(VT, result, false); set_avr64(a->rt, result, false);
return true;
} }
/* /*
@ -494,16 +460,17 @@ static void trans_lvsl(DisasContext *ctx)
* Let X be the 32-byte value 0x00 || 0x01 || 0x02 || ... || 0x1E || 0x1F. * Let X be the 32-byte value 0x00 || 0x01 || 0x02 || ... || 0x1E || 0x1F.
* Bytes (16-sh):(31-sh) of X are placed into vD. * Bytes (16-sh):(31-sh) of X are placed into vD.
*/ */
static void trans_lvsr(DisasContext *ctx) static bool trans_LVSR(DisasContext *ctx, arg_LVSR *a)
{ {
int VT = rD(ctx->opcode);
TCGv_i64 result = tcg_temp_new_i64(); TCGv_i64 result = tcg_temp_new_i64();
TCGv_i64 sh = tcg_temp_new_i64(); TCGv_i64 sh = tcg_temp_new_i64();
TCGv EA = tcg_temp_new(); TCGv EA = tcg_temp_new();
REQUIRE_INSNS_FLAGS(ctx, ALTIVEC);
REQUIRE_VECTOR(ctx);
/* Get sh(from description) by anding EA with 0xf. */ /* Get sh(from description) by anding EA with 0xf. */
gen_addr_reg_index(ctx, EA); EA = do_ea_calc(ctx, a->ra, cpu_gpr[a->rb]);
tcg_gen_extu_tl_i64(sh, EA); tcg_gen_extu_tl_i64(sh, EA);
tcg_gen_andi_i64(sh, sh, 0xfULL); tcg_gen_andi_i64(sh, sh, 0xfULL);
@ -513,13 +480,14 @@ static void trans_lvsr(DisasContext *ctx)
*/ */
tcg_gen_muli_i64(sh, sh, 0x0101010101010101ULL); tcg_gen_muli_i64(sh, sh, 0x0101010101010101ULL);
tcg_gen_subfi_i64(result, 0x1011121314151617ULL, sh); tcg_gen_subfi_i64(result, 0x1011121314151617ULL, sh);
set_avr64(VT, result, true); set_avr64(a->rt, result, true);
/* /*
* Create bytes (24-sh):(32-sh) of X(from description) and place them in * Create bytes (24-sh):(32-sh) of X(from description) and place them in
* lower doubleword of vD. * lower doubleword of vD.
*/ */
tcg_gen_subfi_i64(result, 0x18191a1b1c1d1e1fULL, sh); tcg_gen_subfi_i64(result, 0x18191a1b1c1d1e1fULL, sh);
set_avr64(VT, result, false); set_avr64(a->rt, result, false);
return true;
} }
/* /*
@ -1158,8 +1126,6 @@ GEN_VXFORM_TRANS_DUAL(vmrgow, PPC_NONE, PPC2_ALTIVEC_207,
GEN_VXFORM_HETRO(vextubrx, 6, 28) GEN_VXFORM_HETRO(vextubrx, 6, 28)
GEN_VXFORM_HETRO(vextuhrx, 6, 29) GEN_VXFORM_HETRO(vextuhrx, 6, 29)
GEN_VXFORM_HETRO(vextuwrx, 6, 30) GEN_VXFORM_HETRO(vextuwrx, 6, 30)
GEN_VXFORM_TRANS(lvsl, 6, 31)
GEN_VXFORM_TRANS(lvsr, 6, 32)
GEN_VXFORM_TRANS_DUAL(vmrgew, PPC_NONE, PPC2_ALTIVEC_207, GEN_VXFORM_TRANS_DUAL(vmrgew, PPC_NONE, PPC2_ALTIVEC_207,
vextuwrx, PPC_NONE, PPC2_ISA300) vextuwrx, PPC_NONE, PPC2_ISA300)
@ -3365,11 +3331,6 @@ TRANS_FLAGS2(ISA310, VMODUQ, do_vx_helper, gen_helper_VMODUQ)
#undef DIVS64 #undef DIVS64
#undef DIVU64 #undef DIVU64
#undef GEN_VR_LDX
#undef GEN_VR_STX
#undef GEN_VR_LVE
#undef GEN_VR_STVE
#undef GEN_VX_LOGICAL #undef GEN_VX_LOGICAL
#undef GEN_VX_LOGICAL_207 #undef GEN_VX_LOGICAL_207
#undef GEN_VXFORM #undef GEN_VXFORM

View File

@ -1,22 +1,3 @@
#define GEN_VR_LDX(name, opc2, opc3) \
GEN_HANDLER(name, 0x1F, opc2, opc3, 0x00000001, PPC_ALTIVEC)
#define GEN_VR_STX(name, opc2, opc3) \
GEN_HANDLER(st##name, 0x1F, opc2, opc3, 0x00000001, PPC_ALTIVEC)
#define GEN_VR_LVE(name, opc2, opc3) \
GEN_HANDLER(lve##name, 0x1F, opc2, opc3, 0x00000001, PPC_ALTIVEC)
#define GEN_VR_STVE(name, opc2, opc3) \
GEN_HANDLER(stve##name, 0x1F, opc2, opc3, 0x00000001, PPC_ALTIVEC)
GEN_VR_LDX(lvx, 0x07, 0x03),
GEN_VR_LDX(lvxl, 0x07, 0x0B),
GEN_VR_LVE(bx, 0x07, 0x00),
GEN_VR_LVE(hx, 0x07, 0x01),
GEN_VR_LVE(wx, 0x07, 0x02),
GEN_VR_STX(svx, 0x07, 0x07),
GEN_VR_STX(svxl, 0x07, 0x0F),
GEN_VR_STVE(bx, 0x07, 0x04),
GEN_VR_STVE(hx, 0x07, 0x05),
GEN_VR_STVE(wx, 0x07, 0x06),
#define GEN_VX_LOGICAL(name, tcg_op, opc2, opc3) \ #define GEN_VX_LOGICAL(name, tcg_op, opc2, opc3) \
GEN_HANDLER(name, 0x04, opc2, opc3, 0x00000000, PPC_ALTIVEC) GEN_HANDLER(name, 0x04, opc2, opc3, 0x00000000, PPC_ALTIVEC)