0730ff4c4a
these are include LOCKed RMW of course and also a lot of others in the future it will be very hard to find all the cases that must be atomic so better to start marking them already now try to mark every RMW case for atomicity requirements no code changes, only comments
367 lines
12 KiB
C++
367 lines
12 KiB
C++
/////////////////////////////////////////////////////////////////////////
|
|
// $Id$
|
|
/////////////////////////////////////////////////////////////////////////
|
|
//
|
|
// Copyright (c) 2022 Stanislav Shwartsman
|
|
// Written by Stanislav Shwartsman [sshwarts at sourceforge net]
|
|
//
|
|
// This library is free software; you can redistribute it and/or
|
|
// modify it under the terms of the GNU Lesser General Public
|
|
// License as published by the Free Software Foundation; either
|
|
// version 2 of the License, or (at your option) any later version.
|
|
//
|
|
// This library is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
// Lesser General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Lesser General Public
|
|
// License along with this library; if not, write to the Free Software
|
|
// Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA B 02110-1301 USA
|
|
/////////////////////////////////////////////////////////////////////////
|
|
|
|
#define NEED_CPU_REG_SHORTCUTS 1
|
|
#include "bochs.h"
|
|
#include "cpu.h"
|
|
#define LOG_THIS BX_CPU_THIS_PTR
|
|
|
|
#if BX_SUPPORT_AVX
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPBEXADD_EqGqBq(bxInstruction_c *i)
|
|
{
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
if (laddr & 7) {
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
}
|
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
write_RMW_linear_qword((get_CF() || get_ZF()) ? op1_64 + op3_64 : op1_64);
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
BX_NEXT_INSTR(i);
|
|
}
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPBXADD_EqGqBq(bxInstruction_c *i)
|
|
{
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
if (laddr & 7) {
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
}
|
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
write_RMW_linear_qword(get_CF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
BX_NEXT_INSTR(i);
|
|
}
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPLEXADD_EqGqBq(bxInstruction_c *i)
|
|
{
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
if (laddr & 7) {
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
}
|
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
write_RMW_linear_qword((get_ZF() || getB_SF() != getB_OF()) ? op1_64 + op3_64 : op1_64);
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
BX_NEXT_INSTR(i);
|
|
}
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPLXADD_EqGqBq(bxInstruction_c *i)
|
|
{
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
if (laddr & 7) {
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
}
|
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
write_RMW_linear_qword((getB_SF() != getB_OF()) ? op1_64 + op3_64 : op1_64);
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
BX_NEXT_INSTR(i);
|
|
}
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPNBEXADD_EqGqBq(bxInstruction_c *i)
|
|
{
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
if (laddr & 7) {
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
}
|
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
write_RMW_linear_qword((!get_CF() && !get_ZF()) ? op1_64 + op3_64 : op1_64);
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
BX_NEXT_INSTR(i);
|
|
}
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPNBXADD_EqGqBq(bxInstruction_c *i)
|
|
{
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
if (laddr & 7) {
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
}
|
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
write_RMW_linear_qword(!get_CF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
BX_NEXT_INSTR(i);
|
|
}
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPNLEXADD_EqGqBq(bxInstruction_c *i)
|
|
{
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
if (laddr & 7) {
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
}
|
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
write_RMW_linear_qword((!get_ZF() && getB_SF() == getB_OF()) ? op1_64 + op3_64 : op1_64);
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
BX_NEXT_INSTR(i);
|
|
}
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPNLXADD_EqGqBq(bxInstruction_c *i)
|
|
{
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
if (laddr & 7) {
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
}
|
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
write_RMW_linear_qword((getB_SF() == getB_OF()) ? op1_64 + op3_64 : op1_64);
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
BX_NEXT_INSTR(i);
|
|
}
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPNOXADD_EqGqBq(bxInstruction_c *i)
|
|
{
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
if (laddr & 7) {
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
}
|
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
write_RMW_linear_qword(!get_OF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
BX_NEXT_INSTR(i);
|
|
}
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPNPXADD_EqGqBq(bxInstruction_c *i)
|
|
{
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
if (laddr & 7) {
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
}
|
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
write_RMW_linear_qword(!get_PF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
BX_NEXT_INSTR(i);
|
|
}
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPNSXADD_EqGqBq(bxInstruction_c *i)
|
|
{
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
if (laddr & 7) {
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
}
|
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
write_RMW_linear_qword(!get_SF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
BX_NEXT_INSTR(i);
|
|
}
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPNZXADD_EqGqBq(bxInstruction_c *i)
|
|
{
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
if (laddr & 7) {
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
}
|
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
write_RMW_linear_qword((!get_ZF()) ? op1_64 + op3_64 : op1_64);
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
BX_NEXT_INSTR(i);
|
|
}
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPOXADD_EqGqBq(bxInstruction_c *i)
|
|
{
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
if (laddr & 7) {
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
}
|
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
write_RMW_linear_qword(get_OF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
BX_NEXT_INSTR(i);
|
|
}
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPPXADD_EqGqBq(bxInstruction_c *i)
|
|
{
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
if (laddr & 7) {
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
}
|
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
write_RMW_linear_qword(get_PF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
BX_NEXT_INSTR(i);
|
|
}
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPSXADD_EqGqBq(bxInstruction_c *i)
|
|
{
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
if (laddr & 7) {
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
}
|
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
write_RMW_linear_qword(get_SF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
BX_NEXT_INSTR(i);
|
|
}
|
|
|
|
void BX_CPP_AttrRegparmN(1) BX_CPU_C::CMPZXADD_EqGqBq(bxInstruction_c *i)
|
|
{
|
|
Bit64u op2_64 = BX_READ_64BIT_REG(i->src1());
|
|
Bit64u op3_64 = BX_READ_64BIT_REG(i->src2());
|
|
|
|
bx_address eaddr = BX_CPU_RESOLVE_ADDR_64(i), laddr = get_laddr64(i->seg(), eaddr);
|
|
if (laddr & 7) {
|
|
BX_ERROR(("%s: #GP misaligned access", i->getIaOpcodeNameShort()));
|
|
exception(BX_GP_EXCEPTION, 0);
|
|
}
|
|
|
|
Bit64u op1_64 = read_RMW_linear_qword(i->seg(), laddr); // implicit lock
|
|
Bit64u diff_64 = op1_64 - op2_64;
|
|
SET_FLAGS_OSZAPC_SUB_64(op1_64, op2_64, diff_64);
|
|
write_RMW_linear_qword(get_ZF() ? op1_64 + op3_64 : op1_64);
|
|
|
|
BX_WRITE_64BIT_REG(i->src1(), op1_64);
|
|
|
|
BX_NEXT_INSTR(i);
|
|
}
|
|
|
|
#endif
|