2022-10-08 19:51:30 +03:00
|
|
|
/////////////////////////////////////////////////////////////////////////
|
|
|
|
// $Id$
|
|
|
|
/////////////////////////////////////////////////////////////////////////
|
|
|
|
//
|
|
|
|
// Copyright (c) 2022 Stanislav Shwartsman
|
|
|
|
// Written by Stanislav Shwartsman [sshwarts at sourceforge net]
|
|
|
|
//
|
|
|
|
// This library is free software; you can redistribute it and/or
|
|
|
|
// modify it under the terms of the GNU Lesser General Public
|
|
|
|
// License as published by the Free Software Foundation; either
|
|
|
|
// version 2 of the License, or (at your option) any later version.
|
|
|
|
//
|
|
|
|
// This library is distributed in the hope that it will be useful,
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
// Lesser General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU Lesser General Public
|
|
|
|
// License along with this library; if not, write to the Free Software
|
|
|
|
// Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA B 02110-1301 USA
|
|
|
|
/////////////////////////////////////////////////////////////////////////
|
|
|
|
|
|
|
|
#define NEED_CPU_REG_SHORTCUTS 1
|
|
|
|
#include "bochs.h"
|
|
|
|
#include "cpu.h"
|
|
|
|
#define LOG_THIS BX_CPU_THIS_PTR
|
|
|
|
|
|
|
|
#if BX_SUPPORT_AVX
|
|
|
|
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPBEXADD_EqGqBq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
|
|
if (laddr & 7) {
|
|
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
|
|
}
|
|
|
|
|
2023-11-18 00:18:43 +03:00
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
2022-10-08 19:51:30 +03:00
|
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
|
|
write_RMW_linear_qword((get_CF() || get_ZF()) ? op1_64 + op3_64 : op1_64);
|
|
|
|
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
|
|
|
|
BX_NEXT_INSTR(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPBXADD_EqGqBq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
|
|
if (laddr & 7) {
|
|
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
|
|
}
|
|
|
|
|
2023-11-18 00:18:43 +03:00
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
2022-10-08 19:51:30 +03:00
|
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
|
|
write_RMW_linear_qword(get_CF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
|
|
|
|
BX_NEXT_INSTR(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPLEXADD_EqGqBq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
|
|
if (laddr & 7) {
|
|
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
|
|
}
|
|
|
|
|
2023-11-18 00:18:43 +03:00
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
2022-10-08 19:51:30 +03:00
|
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
|
|
write_RMW_linear_qword((get_ZF() || getB_SF() != getB_OF()) ? op1_64 + op3_64 : op1_64);
|
|
|
|
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
|
|
|
|
BX_NEXT_INSTR(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPLXADD_EqGqBq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
|
|
if (laddr & 7) {
|
|
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
|
|
}
|
|
|
|
|
2023-11-18 00:18:43 +03:00
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
2022-10-08 19:51:30 +03:00
|
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
|
|
write_RMW_linear_qword((getB_SF() != getB_OF()) ? op1_64 + op3_64 : op1_64);
|
|
|
|
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
|
|
|
|
BX_NEXT_INSTR(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPNBEXADD_EqGqBq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
|
|
if (laddr & 7) {
|
|
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
|
|
}
|
|
|
|
|
2023-11-18 00:18:43 +03:00
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
2022-10-08 19:51:30 +03:00
|
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
|
|
write_RMW_linear_qword((!get_CF() && !get_ZF()) ? op1_64 + op3_64 : op1_64);
|
|
|
|
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
|
|
|
|
BX_NEXT_INSTR(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPNBXADD_EqGqBq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
|
|
if (laddr & 7) {
|
|
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
|
|
}
|
|
|
|
|
2023-11-18 00:18:43 +03:00
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
2022-10-08 19:51:30 +03:00
|
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
|
|
write_RMW_linear_qword(!get_CF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
|
|
|
|
BX_NEXT_INSTR(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPNLEXADD_EqGqBq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
|
|
if (laddr & 7) {
|
|
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
|
|
}
|
|
|
|
|
2023-11-18 00:18:43 +03:00
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
2022-10-08 19:51:30 +03:00
|
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
|
|
write_RMW_linear_qword((!get_ZF() && getB_SF() == getB_OF()) ? op1_64 + op3_64 : op1_64);
|
|
|
|
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
|
|
|
|
BX_NEXT_INSTR(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPNLXADD_EqGqBq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
|
|
if (laddr & 7) {
|
|
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
|
|
}
|
|
|
|
|
2023-11-18 00:18:43 +03:00
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
2022-10-08 19:51:30 +03:00
|
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
|
|
write_RMW_linear_qword((getB_SF() == getB_OF()) ? op1_64 + op3_64 : op1_64);
|
|
|
|
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
|
|
|
|
BX_NEXT_INSTR(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPNOXADD_EqGqBq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
|
|
if (laddr & 7) {
|
|
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
|
|
}
|
|
|
|
|
2023-11-18 00:18:43 +03:00
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
2022-10-08 19:51:30 +03:00
|
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
|
|
write_RMW_linear_qword(!get_OF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
|
|
|
|
BX_NEXT_INSTR(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPNPXADD_EqGqBq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
|
|
if (laddr & 7) {
|
|
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
|
|
}
|
|
|
|
|
2023-11-18 00:18:43 +03:00
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
2022-10-08 19:51:30 +03:00
|
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
|
|
write_RMW_linear_qword(!get_PF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
|
|
|
|
BX_NEXT_INSTR(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPNSXADD_EqGqBq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
|
|
if (laddr & 7) {
|
|
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
|
|
}
|
|
|
|
|
2023-11-18 00:18:43 +03:00
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
2022-10-08 19:51:30 +03:00
|
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
|
|
write_RMW_linear_qword(!get_SF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
|
|
|
|
BX_NEXT_INSTR(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPNZXADD_EqGqBq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
|
|
if (laddr & 7) {
|
|
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
|
|
}
|
|
|
|
|
2023-11-18 00:18:43 +03:00
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
2022-10-08 19:51:30 +03:00
|
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
|
|
write_RMW_linear_qword((!get_ZF()) ? op1_64 + op3_64 : op1_64);
|
|
|
|
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
|
|
|
|
BX_NEXT_INSTR(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPOXADD_EqGqBq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
|
|
if (laddr & 7) {
|
|
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
|
|
}
|
|
|
|
|
2023-11-18 00:18:43 +03:00
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
2022-10-08 19:51:30 +03:00
|
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
|
|
write_RMW_linear_qword(get_OF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
|
|
|
|
BX_NEXT_INSTR(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPPXADD_EqGqBq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
|
|
if (laddr & 7) {
|
|
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
|
|
}
|
|
|
|
|
2023-11-18 00:18:43 +03:00
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
2022-10-08 19:51:30 +03:00
|
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
|
|
write_RMW_linear_qword(get_PF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
|
|
|
|
BX_NEXT_INSTR(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPSXADD_EqGqBq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
|
|
if (laddr & 7) {
|
|
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
|
|
}
|
|
|
|
|
2023-11-18 00:18:43 +03:00
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
2022-10-08 19:51:30 +03:00
|
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
|
|
write_RMW_linear_qword(get_SF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
|
|
|
|
BX_NEXT_INSTR(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPZXADD_EqGqBq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
|
|
if (laddr & 7) {
|
|
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
|
|
}
|
|
|
|
|
2023-11-18 00:18:43 +03:00
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
2022-10-08 19:51:30 +03:00
|
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
2022-10-08 20:35:36 +03:00
|
|
|
write_RMW_linear_qword(get_ZF() ? op1_64 + op3_64 : op1_64);
|
2022-10-08 19:51:30 +03:00
|
|
|
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
|
|
|
|
BX_NEXT_INSTR(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif
|