2001-10-03 17:10:38 +04:00
|
|
|
/////////////////////////////////////////////////////////////////////////
|
2005-08-02 01:40:17 +04:00
|
|
|
// $Id: string.cc,v 1.31 2005-08-01 21:40:17 sshwarts Exp $
|
2001-10-03 17:10:38 +04:00
|
|
|
/////////////////////////////////////////////////////////////////////////
|
|
|
|
//
|
2001-04-10 06:20:02 +04:00
|
|
|
// Copyright (C) 2001 MandrakeSoft S.A.
|
2001-04-10 05:04:59 +04:00
|
|
|
//
|
|
|
|
// MandrakeSoft S.A.
|
|
|
|
// 43, rue d'Aboukir
|
|
|
|
// 75002 Paris - France
|
|
|
|
// http://www.linux-mandrake.com/
|
|
|
|
// http://www.mandrakesoft.com/
|
|
|
|
//
|
|
|
|
// This library is free software; you can redistribute it and/or
|
|
|
|
// modify it under the terms of the GNU Lesser General Public
|
|
|
|
// License as published by the Free Software Foundation; either
|
|
|
|
// version 2 of the License, or (at your option) any later version.
|
|
|
|
//
|
|
|
|
// This library is distributed in the hope that it will be useful,
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
// Lesser General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU Lesser General Public
|
|
|
|
// License along with this library; if not, write to the Free Software
|
|
|
|
// Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
|
|
|
|
|
|
|
|
|
|
|
2001-05-24 22:46:34 +04:00
|
|
|
#define NEED_CPU_REG_SHORTCUTS 1
|
2001-04-10 05:04:59 +04:00
|
|
|
#include "bochs.h"
|
merge in BRANCH-io-cleanup.
To see the commit logs for this use either cvsweb or
cvs update -r BRANCH-io-cleanup and then 'cvs log' the various files.
In general this provides a generic interface for logging.
logfunctions:: is a class that is inherited by some classes, and also
. allocated as a standalone global called 'genlog'. All logging uses
. one of the ::info(), ::error(), ::ldebug(), ::panic() methods of this
. class through 'BX_INFO(), BX_ERROR(), BX_DEBUG(), BX_PANIC()' macros
. respectively.
.
. An example usage:
. BX_INFO(("Hello, World!\n"));
iofunctions:: is a class that is allocated once by default, and assigned
as the iofunction of each logfunctions instance. It is this class that
maintains the file descriptor and other output related code, at this
point using vfprintf(). At some future point, someone may choose to
write a gui 'console' for bochs to which messages would be redirected
simply by assigning a different iofunction class to the various logfunctions
objects.
More cleanup is coming, but this works for now. If you want to see alot
of debugging output, in main.cc, change onoff[LOGLEV_DEBUG]=0 to =1.
Comments, bugs, flames, to me: todd@fries.net
2001-05-15 18:49:57 +04:00
|
|
|
#define LOG_THIS BX_CPU_THIS_PTR
|
2001-04-10 05:04:59 +04:00
|
|
|
|
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
#if BX_SUPPORT_X86_64==0
|
|
|
|
#define RSI ESI
|
|
|
|
#define RDI EDI
|
|
|
|
#define RAX EAX
|
|
|
|
#endif
|
|
|
|
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
|
2005-07-04 21:44:08 +04:00
|
|
|
#if BX_SupportRepeatSpeedups
|
|
|
|
Bit32u BX_CPU_C::FastRepMOVSB(bxInstruction_c *i, unsigned srcSeg, bx_address srcOff, unsigned dstSeg, bx_address dstOff, Bit32u count)
|
|
|
|
{
|
|
|
|
Bit32u bytesFitSrc, bytesFitDst;
|
|
|
|
signed int pointerDelta;
|
|
|
|
bx_address laddrDst, laddrSrc;
|
|
|
|
Bit32u paddrDst, paddrSrc;
|
|
|
|
|
|
|
|
bx_segment_reg_t *srcSegPtr = &BX_CPU_THIS_PTR sregs[srcSeg];
|
|
|
|
bx_segment_reg_t *dstSegPtr = &BX_CPU_THIS_PTR sregs[dstSeg];
|
|
|
|
|
|
|
|
// Do segment checks for the 1st byte. We do not want to
|
|
|
|
// trip an exception beyond this, because the address would
|
|
|
|
// be incorrect. After we know how many bytes we will directly
|
|
|
|
// transfer, we can do the full segment limit check ourselves
|
|
|
|
// without generating an exception.
|
|
|
|
read_virtual_checks(srcSegPtr, srcOff, 1);
|
|
|
|
laddrSrc = BX_CPU_THIS_PTR get_segment_base(srcSeg) + srcOff;
|
|
|
|
if (BX_CPU_THIS_PTR cr0.pg) {
|
|
|
|
paddrSrc = dtranslate_linear(laddrSrc, CPL==3, BX_READ);
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
paddrSrc = laddrSrc;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If we want to write directly into the physical memory array,
|
|
|
|
// we need the A20 address.
|
|
|
|
paddrSrc = A20ADDR(paddrSrc);
|
|
|
|
Bit8u *hostAddrSrc = BX_CPU_THIS_PTR mem->getHostMemAddr(BX_CPU_THIS,
|
|
|
|
paddrSrc, BX_READ);
|
|
|
|
|
|
|
|
if (! hostAddrSrc) return 0;
|
|
|
|
|
|
|
|
write_virtual_checks(dstSegPtr, dstOff, 1);
|
|
|
|
laddrDst = BX_CPU_THIS_PTR get_segment_base(dstSeg) + dstOff;
|
|
|
|
if (BX_CPU_THIS_PTR cr0.pg) {
|
|
|
|
paddrDst = dtranslate_linear(laddrDst, CPL==3, BX_WRITE);
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
paddrDst = laddrDst;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If we want to write directly into the physical memory array,
|
|
|
|
// we need the A20 address.
|
|
|
|
paddrDst = A20ADDR(paddrDst);
|
|
|
|
Bit8u *hostAddrDst = BX_CPU_THIS_PTR mem->getHostMemAddr(BX_CPU_THIS,
|
|
|
|
paddrDst, BX_WRITE);
|
|
|
|
|
|
|
|
if (! hostAddrDst) return 0;
|
|
|
|
|
|
|
|
// See how many bytes can fit in the rest of this page.
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
// Counting downward.
|
|
|
|
bytesFitSrc = 1 + (paddrSrc & 0xfff);
|
|
|
|
bytesFitDst = 1 + (paddrDst & 0xfff);
|
|
|
|
pointerDelta = (signed int) -1;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
// Counting upward.
|
|
|
|
bytesFitSrc = (0x1000 - (paddrSrc & 0xfff));
|
|
|
|
bytesFitDst = (0x1000 - (paddrDst & 0xfff));
|
|
|
|
pointerDelta = (signed int) 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Restrict word count to the number that will fit in either
|
|
|
|
// source or dest pages.
|
|
|
|
if (count > bytesFitSrc)
|
|
|
|
count = bytesFitSrc;
|
|
|
|
if (count > bytesFitDst)
|
|
|
|
count = bytesFitDst;
|
|
|
|
if (count > bx_pc_system.getNumCpuTicksLeftNextEvent())
|
|
|
|
count = bx_pc_system.getNumCpuTicksLeftNextEvent();
|
|
|
|
|
|
|
|
// If after all the restrictions, there is anything left to do...
|
|
|
|
if (count) {
|
|
|
|
// Before we copy memory, we need to make sure that the segments
|
|
|
|
// allow the accesses up to the given source and dest offset. If
|
|
|
|
// the cache.valid bits have SegAccessWOK and ROK, we know that
|
|
|
|
// the cache is valid for those operations, and that the segments
|
|
|
|
// are non expand-down (thus we can make a simple limit check).
|
|
|
|
if ( !(srcSegPtr->cache.valid & SegAccessROK) ||
|
|
|
|
!(dstSegPtr->cache.valid & SegAccessWOK) )
|
|
|
|
{
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2005-08-02 01:40:17 +04:00
|
|
|
if (BX_CPU_THIS_PTR cpu_mode != BX_MODE_LONG_64)
|
2005-07-04 21:44:08 +04:00
|
|
|
{
|
|
|
|
Bit32u srcSegLimit = srcSegPtr->cache.u.segment.limit_scaled;
|
|
|
|
Bit32u dstSegLimit = dstSegPtr->cache.u.segment.limit_scaled;
|
|
|
|
|
|
|
|
if (! i->as32L()) {
|
|
|
|
// For 16-bit addressing mode, clamp the segment limits to 16bits
|
|
|
|
// so we don't have to worry about computations using si/di
|
|
|
|
// rolling over 16-bit boundaries.
|
|
|
|
if (srcSegLimit > 0xffff)
|
|
|
|
srcSegLimit = 0xffff;
|
|
|
|
if (dstSegLimit > 0xffff)
|
|
|
|
dstSegLimit = 0xffff;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now make sure transfer will fit within the constraints of the
|
|
|
|
// segment boundaries, 0..limit for non expand-down. We know
|
|
|
|
// count >= 1 here.
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
Bit32u minOffset = (count-1);
|
|
|
|
if ( srcOff < minOffset )
|
|
|
|
return 0;
|
|
|
|
if ( dstOff < minOffset )
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
// Counting upward.
|
|
|
|
Bit32u srcMaxOffset = (srcSegLimit - count) + 1;
|
|
|
|
Bit32u dstMaxOffset = (dstSegLimit - count) + 1;
|
|
|
|
if ( srcOff > srcMaxOffset )
|
|
|
|
return 0;
|
|
|
|
if ( dstOff > dstMaxOffset )
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Transfer data directly using host addresses
|
|
|
|
for (unsigned j=0; j<count; j++) {
|
|
|
|
* (Bit8u *) hostAddrDst = * (Bit8u *) hostAddrSrc;
|
|
|
|
hostAddrDst += pointerDelta;
|
|
|
|
hostAddrSrc += pointerDelta;
|
|
|
|
}
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
Bit32u BX_CPU_C::FastRepMOVSW(bxInstruction_c *i, unsigned srcSeg, bx_address srcOff, unsigned dstSeg, bx_address dstOff, Bit32u count)
|
|
|
|
{
|
|
|
|
Bit32u wordsFitSrc, wordsFitDst;
|
|
|
|
signed int pointerDelta;
|
|
|
|
bx_address laddrDst, laddrSrc;
|
|
|
|
Bit32u paddrDst, paddrSrc;
|
|
|
|
|
|
|
|
bx_segment_reg_t *srcSegPtr = &BX_CPU_THIS_PTR sregs[srcSeg];
|
|
|
|
bx_segment_reg_t *dstSegPtr = &BX_CPU_THIS_PTR sregs[dstSeg];
|
|
|
|
|
|
|
|
// Do segment checks for the 1st word. We do not want to
|
|
|
|
// trip an exception beyond this, because the address would
|
|
|
|
// be incorrect. After we know how many bytes we will directly
|
|
|
|
// transfer, we can do the full segment limit check ourselves
|
|
|
|
// without generating an exception.
|
|
|
|
read_virtual_checks(srcSegPtr, srcOff, 2);
|
|
|
|
laddrSrc = BX_CPU_THIS_PTR get_segment_base(srcSeg) + srcOff;
|
|
|
|
if (BX_CPU_THIS_PTR cr0.pg) {
|
|
|
|
paddrSrc = dtranslate_linear(laddrSrc, CPL==3, BX_READ);
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
paddrSrc = laddrSrc;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If we want to write directly into the physical memory array,
|
|
|
|
// we need the A20 address.
|
|
|
|
paddrSrc = A20ADDR(paddrSrc);
|
|
|
|
Bit8u *hostAddrSrc = BX_CPU_THIS_PTR mem->getHostMemAddr(BX_CPU_THIS,
|
|
|
|
paddrSrc, BX_READ);
|
|
|
|
|
|
|
|
if (! hostAddrSrc) return 0;
|
|
|
|
|
|
|
|
write_virtual_checks(dstSegPtr, dstOff, 2);
|
|
|
|
laddrDst = BX_CPU_THIS_PTR get_segment_base(dstSeg) + dstOff;
|
|
|
|
if (BX_CPU_THIS_PTR cr0.pg) {
|
|
|
|
paddrDst = dtranslate_linear(laddrDst, CPL==3, BX_WRITE);
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
paddrDst = laddrDst;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If we want to write directly into the physical memory array,
|
|
|
|
// we need the A20 address.
|
|
|
|
paddrDst = A20ADDR(paddrDst);
|
|
|
|
Bit8u *hostAddrDst = BX_CPU_THIS_PTR mem->getHostMemAddr(BX_CPU_THIS,
|
|
|
|
paddrDst, BX_WRITE);
|
|
|
|
|
|
|
|
if (! hostAddrDst) return 0;
|
|
|
|
|
|
|
|
// See how many words can fit in the rest of this page.
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
// Counting downward.
|
|
|
|
// Note: 1st word must not cross page boundary.
|
|
|
|
if ( ((paddrSrc & 0xfff) > 0xffe) || ((paddrDst & 0xfff) > 0xffe) )
|
|
|
|
return 0;
|
|
|
|
wordsFitSrc = (2 + (paddrSrc & 0xfff)) >> 1;
|
|
|
|
wordsFitDst = (2 + (paddrDst & 0xfff)) >> 1;
|
|
|
|
pointerDelta = (signed int) -2;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
// Counting upward.
|
|
|
|
wordsFitSrc = (0x1000 - (paddrSrc & 0xfff)) >> 1;
|
|
|
|
wordsFitDst = (0x1000 - (paddrDst & 0xfff)) >> 1;
|
|
|
|
pointerDelta = (signed int) 2;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Restrict word count to the number that will fit in either
|
|
|
|
// source or dest pages.
|
|
|
|
if (count > wordsFitSrc)
|
|
|
|
count = wordsFitSrc;
|
|
|
|
if (count > wordsFitDst)
|
|
|
|
count = wordsFitDst;
|
|
|
|
if (count > bx_pc_system.getNumCpuTicksLeftNextEvent())
|
|
|
|
count = bx_pc_system.getNumCpuTicksLeftNextEvent();
|
|
|
|
|
|
|
|
// If after all the restrictions, there is anything left to do...
|
|
|
|
if (count) {
|
|
|
|
// Before we copy memory, we need to make sure that the segments
|
|
|
|
// allow the accesses up to the given source and dest offset. If
|
|
|
|
// the cache.valid bits have SegAccessWOK and ROK, we know that
|
|
|
|
// the cache is valid for those operations, and that the segments
|
|
|
|
// are non expand-down (thus we can make a simple limit check).
|
|
|
|
if ( !(srcSegPtr->cache.valid & SegAccessROK) ||
|
|
|
|
!(dstSegPtr->cache.valid & SegAccessWOK) )
|
|
|
|
{
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2005-08-02 01:40:17 +04:00
|
|
|
if (BX_CPU_THIS_PTR cpu_mode != BX_MODE_LONG_64)
|
2005-07-04 21:44:08 +04:00
|
|
|
{
|
|
|
|
Bit32u srcSegLimit = srcSegPtr->cache.u.segment.limit_scaled;
|
|
|
|
Bit32u dstSegLimit = dstSegPtr->cache.u.segment.limit_scaled;
|
|
|
|
|
|
|
|
if (! i->as32L()) {
|
|
|
|
// For 16-bit addressing mode, clamp the segment limits to 16bits
|
|
|
|
// so we don't have to worry about computations using si/di
|
|
|
|
// rolling over 16-bit boundaries.
|
|
|
|
if (srcSegLimit > 0xffff)
|
|
|
|
srcSegLimit = 0xffff;
|
|
|
|
if (dstSegLimit > 0xffff)
|
|
|
|
dstSegLimit = 0xffff;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now make sure transfer will fit within the constraints of the
|
|
|
|
// segment boundaries, 0..limit for non expand-down. We know
|
|
|
|
// count >= 1 here.
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
// Counting downward.
|
|
|
|
Bit32u minOffset = (count-1) << 1;
|
|
|
|
if ( srcOff < minOffset )
|
|
|
|
return 0;
|
|
|
|
if ( dstOff < minOffset )
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
// Counting upward.
|
|
|
|
Bit32u srcMaxOffset = (srcSegLimit - (count<<1)) + 1;
|
|
|
|
Bit32u dstMaxOffset = (dstSegLimit - (count<<1)) + 1;
|
|
|
|
if ( srcOff > srcMaxOffset )
|
|
|
|
return 0;
|
|
|
|
if ( dstOff > dstMaxOffset )
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Transfer data directly using host addresses
|
|
|
|
for (unsigned j=0; j<count; j++) {
|
|
|
|
* (Bit16u *) hostAddrDst = * (Bit16u *) hostAddrSrc;
|
|
|
|
hostAddrDst += pointerDelta;
|
|
|
|
hostAddrSrc += pointerDelta;
|
|
|
|
}
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
Bit32u BX_CPU_C::FastRepMOVSD(bxInstruction_c *i, unsigned srcSeg, bx_address srcOff, unsigned dstSeg, bx_address dstOff, Bit32u count)
|
|
|
|
{
|
|
|
|
Bit32u dwordsFitSrc, dwordsFitDst;
|
|
|
|
signed int pointerDelta;
|
|
|
|
bx_address laddrDst, laddrSrc;
|
|
|
|
Bit32u paddrDst, paddrSrc;
|
|
|
|
|
|
|
|
bx_segment_reg_t *srcSegPtr = &BX_CPU_THIS_PTR sregs[srcSeg];
|
|
|
|
bx_segment_reg_t *dstSegPtr = &BX_CPU_THIS_PTR sregs[dstSeg];
|
|
|
|
|
|
|
|
// Do segment checks for the 1st dword. We do not want to
|
|
|
|
// trip an exception beyond this, because the address would
|
|
|
|
// be incorrect. After we know how many bytes we will directly
|
|
|
|
// transfer, we can do the full segment limit check ourselves
|
|
|
|
// without generating an exception.
|
|
|
|
read_virtual_checks(srcSegPtr, srcOff, 4);
|
|
|
|
laddrSrc = BX_CPU_THIS_PTR get_segment_base(srcSeg) + srcOff;
|
|
|
|
if (BX_CPU_THIS_PTR cr0.pg) {
|
|
|
|
paddrSrc = dtranslate_linear(laddrSrc, CPL==3, BX_READ);
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
paddrSrc = laddrSrc;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If we want to write directly into the physical memory array,
|
|
|
|
// we need the A20 address.
|
|
|
|
paddrSrc = A20ADDR(paddrSrc);
|
|
|
|
Bit8u *hostAddrSrc = BX_CPU_THIS_PTR mem->getHostMemAddr(BX_CPU_THIS,
|
|
|
|
paddrSrc, BX_READ);
|
|
|
|
|
|
|
|
if (! hostAddrSrc) return 0;
|
|
|
|
|
|
|
|
write_virtual_checks(dstSegPtr, dstOff, 4);
|
|
|
|
laddrDst = BX_CPU_THIS_PTR get_segment_base(dstSeg) + dstOff;
|
|
|
|
if (BX_CPU_THIS_PTR cr0.pg) {
|
|
|
|
paddrDst = dtranslate_linear(laddrDst, CPL==3, BX_WRITE);
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
paddrDst = laddrDst;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If we want to write directly into the physical memory array,
|
|
|
|
// we need the A20 address.
|
|
|
|
paddrDst = A20ADDR(paddrDst);
|
|
|
|
Bit8u *hostAddrDst = BX_CPU_THIS_PTR mem->getHostMemAddr(BX_CPU_THIS,
|
|
|
|
paddrDst, BX_WRITE);
|
|
|
|
|
|
|
|
if (! hostAddrDst) return 0;
|
|
|
|
|
|
|
|
// See how many dwords can fit in the rest of this page.
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
// Counting downward.
|
|
|
|
// Note: 1st dword must not cross page boundary.
|
|
|
|
if ( ((paddrSrc & 0xfff) > 0xffc) || ((paddrDst & 0xfff) > 0xffc) )
|
|
|
|
return 0;
|
|
|
|
dwordsFitSrc = (4 + (paddrSrc & 0xfff)) >> 2;
|
|
|
|
dwordsFitDst = (4 + (paddrDst & 0xfff)) >> 2;
|
|
|
|
pointerDelta = (signed int) -4;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
// Counting upward.
|
|
|
|
dwordsFitSrc = (0x1000 - (paddrSrc & 0xfff)) >> 2;
|
|
|
|
dwordsFitDst = (0x1000 - (paddrDst & 0xfff)) >> 2;
|
|
|
|
pointerDelta = (signed int) 4;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Restrict dword count to the number that will fit in either
|
|
|
|
// source or dest pages.
|
|
|
|
if (count > dwordsFitSrc)
|
|
|
|
count = dwordsFitSrc;
|
|
|
|
if (count > dwordsFitDst)
|
|
|
|
count = dwordsFitDst;
|
|
|
|
if (count > bx_pc_system.getNumCpuTicksLeftNextEvent())
|
|
|
|
count = bx_pc_system.getNumCpuTicksLeftNextEvent();
|
|
|
|
|
|
|
|
// If after all the restrictions, there is anything left to do...
|
|
|
|
if (count) {
|
|
|
|
// Before we copy memory, we need to make sure that the segments
|
|
|
|
// allow the accesses up to the given source and dest offset. If
|
|
|
|
// the cache.valid bits have SegAccessWOK and ROK, we know that
|
|
|
|
// the cache is valid for those operations, and that the segments
|
|
|
|
// are non expand-down (thus we can make a simple limit check).
|
|
|
|
if ( !(srcSegPtr->cache.valid & SegAccessROK) ||
|
|
|
|
!(dstSegPtr->cache.valid & SegAccessWOK) )
|
|
|
|
{
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2005-08-02 01:40:17 +04:00
|
|
|
if (BX_CPU_THIS_PTR cpu_mode != BX_MODE_LONG_64)
|
2005-07-04 21:44:08 +04:00
|
|
|
{
|
|
|
|
Bit32u srcSegLimit = srcSegPtr->cache.u.segment.limit_scaled;
|
|
|
|
Bit32u dstSegLimit = dstSegPtr->cache.u.segment.limit_scaled;
|
|
|
|
|
|
|
|
if (! i->as32L()) {
|
|
|
|
// For 16-bit addressing mode, clamp the segment limits to 16bits
|
|
|
|
// so we don't have to worry about computations using si/di
|
|
|
|
// rolling over 16-bit boundaries.
|
|
|
|
if (srcSegLimit > 0xffff)
|
|
|
|
srcSegLimit = 0xffff;
|
|
|
|
if (dstSegLimit > 0xffff)
|
|
|
|
dstSegLimit = 0xffff;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now make sure transfer will fit within the constraints of the
|
|
|
|
// segment boundaries, 0..limit for non expand-down. We know
|
|
|
|
// count >= 1 here.
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
// Counting downward.
|
|
|
|
Bit32u minOffset = (count-1) << 2;
|
|
|
|
if ( srcOff < minOffset )
|
|
|
|
return 0;
|
|
|
|
if ( dstOff < minOffset )
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
// Counting upward.
|
|
|
|
Bit32u srcMaxOffset = (srcSegLimit - (count<<2)) + 1;
|
|
|
|
Bit32u dstMaxOffset = (dstSegLimit - (count<<2)) + 1;
|
|
|
|
if ( srcOff > srcMaxOffset )
|
|
|
|
return 0;
|
|
|
|
if ( dstOff > dstMaxOffset )
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Transfer data directly using host addresses
|
|
|
|
for (unsigned j=0; j<count; j++) {
|
|
|
|
* (Bit32u *) hostAddrDst = * (Bit32u *) hostAddrSrc;
|
|
|
|
hostAddrDst += pointerDelta;
|
|
|
|
hostAddrSrc += pointerDelta;
|
|
|
|
}
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
Bit32u BX_CPU_C::FastRepSTOSB(bxInstruction_c *i, unsigned dstSeg, bx_address dstOff, Bit8u val, Bit32u count)
|
|
|
|
{
|
|
|
|
Bit32u bytesFitDst;
|
|
|
|
signed int pointerDelta;
|
|
|
|
bx_address laddrDst;
|
|
|
|
Bit32u paddrDst;
|
|
|
|
|
|
|
|
bx_segment_reg_t *dstSegPtr = &BX_CPU_THIS_PTR sregs[dstSeg];
|
|
|
|
|
|
|
|
write_virtual_checks(dstSegPtr, dstOff, 1);
|
|
|
|
laddrDst = BX_CPU_THIS_PTR get_segment_base(dstSeg) + dstOff;
|
|
|
|
if (BX_CPU_THIS_PTR cr0.pg) {
|
|
|
|
paddrDst = dtranslate_linear(laddrDst, CPL==3, BX_WRITE);
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
paddrDst = laddrDst;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If we want to write directly into the physical memory array,
|
|
|
|
// we need the A20 address.
|
|
|
|
paddrDst = A20ADDR(paddrDst);
|
|
|
|
Bit8u *hostAddrDst = BX_CPU_THIS_PTR mem->getHostMemAddr(BX_CPU_THIS,
|
|
|
|
paddrDst, BX_WRITE);
|
|
|
|
|
|
|
|
if (! hostAddrDst) return 0;
|
|
|
|
|
|
|
|
// See how many bytes can fit in the rest of this page.
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
// Counting downward.
|
|
|
|
bytesFitDst = 1 + (paddrDst & 0xfff);
|
|
|
|
pointerDelta = (signed int) -1;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
// Counting upward.
|
|
|
|
bytesFitDst = (0x1000 - (paddrDst & 0xfff));
|
|
|
|
pointerDelta = (signed int) 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Restrict word count to the number that will fit in either
|
|
|
|
// source or dest pages.
|
|
|
|
if (count > bytesFitDst)
|
|
|
|
count = bytesFitDst;
|
|
|
|
if (count > bx_pc_system.getNumCpuTicksLeftNextEvent())
|
|
|
|
count = bx_pc_system.getNumCpuTicksLeftNextEvent();
|
|
|
|
|
|
|
|
// If after all the restrictions, there is anything left to do...
|
|
|
|
if (count) {
|
|
|
|
// Before we copy memory, we need to make sure that the segments
|
|
|
|
// allow the accesses up to the given source and dest offset. If
|
|
|
|
// the cache.valid bits have SegAccessWOK and ROK, we know that
|
|
|
|
// the cache is valid for those operations, and that the segments
|
|
|
|
// are non expand-down (thus we can make a simple limit check).
|
|
|
|
if ( !(dstSegPtr->cache.valid & SegAccessWOK) ) return 0;
|
|
|
|
|
2005-08-02 01:40:17 +04:00
|
|
|
if (BX_CPU_THIS_PTR cpu_mode != BX_MODE_LONG_64)
|
2005-07-04 21:44:08 +04:00
|
|
|
{
|
|
|
|
Bit32u dstSegLimit = dstSegPtr->cache.u.segment.limit_scaled;
|
|
|
|
|
|
|
|
if (! i->as32L()) {
|
|
|
|
// For 16-bit addressing mode, clamp the segment limits to 16bits
|
|
|
|
// so we don't have to worry about computations using di
|
|
|
|
// rolling over 16-bit boundaries.
|
|
|
|
if (dstSegLimit > 0xffff)
|
|
|
|
dstSegLimit = 0xffff;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now make sure transfer will fit within the constraints of the
|
|
|
|
// segment boundaries, 0..limit for non expand-down. We know
|
|
|
|
// count >= 1 here.
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
Bit32u minOffset = (count-1);
|
|
|
|
if ( dstOff < minOffset )
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
// Counting upward.
|
|
|
|
Bit32u dstMaxOffset = (dstSegLimit - count) + 1;
|
|
|
|
if ( dstOff > dstMaxOffset )
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Transfer data directly using host addresses
|
|
|
|
for (unsigned j=0; j<count; j++) {
|
|
|
|
* (Bit8u *) hostAddrDst = val;
|
|
|
|
hostAddrDst += pointerDelta;
|
|
|
|
}
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
Bit32u BX_CPU_C::FastRepSTOSW(bxInstruction_c *i, unsigned dstSeg, bx_address dstOff, Bit16u val, Bit32u count)
|
|
|
|
{
|
|
|
|
Bit32u wordsFitDst;
|
|
|
|
signed int pointerDelta;
|
|
|
|
bx_address laddrDst;
|
|
|
|
Bit32u paddrDst;
|
|
|
|
|
|
|
|
bx_segment_reg_t *dstSegPtr = &BX_CPU_THIS_PTR sregs[dstSeg];
|
|
|
|
|
|
|
|
write_virtual_checks(dstSegPtr, dstOff, 2);
|
|
|
|
laddrDst = BX_CPU_THIS_PTR get_segment_base(dstSeg) + dstOff;
|
|
|
|
if (BX_CPU_THIS_PTR cr0.pg) {
|
|
|
|
paddrDst = dtranslate_linear(laddrDst, CPL==3, BX_WRITE);
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
paddrDst = laddrDst;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If we want to write directly into the physical memory array,
|
|
|
|
// we need the A20 address.
|
|
|
|
paddrDst = A20ADDR(paddrDst);
|
|
|
|
Bit8u *hostAddrDst = BX_CPU_THIS_PTR mem->getHostMemAddr(BX_CPU_THIS,
|
|
|
|
paddrDst, BX_WRITE);
|
|
|
|
|
|
|
|
if (! hostAddrDst) return 0;
|
|
|
|
|
|
|
|
// See how many words can fit in the rest of this page.
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
// Counting downward.
|
|
|
|
// Note: 1st word must not cross page boundary.
|
|
|
|
if ((paddrDst & 0xfff) > 0xffe) return 0;
|
|
|
|
wordsFitDst = (2 + (paddrDst & 0xfff)) >> 1;
|
|
|
|
pointerDelta = (signed int) -2;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
// Counting upward.
|
|
|
|
wordsFitDst = (0x1000 - (paddrDst & 0xfff)) >> 1;
|
|
|
|
pointerDelta = (signed int) 2;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Restrict word count to the number that will fit in either
|
|
|
|
// source or dest pages.
|
|
|
|
if (count > wordsFitDst)
|
|
|
|
count = wordsFitDst;
|
|
|
|
if (count > bx_pc_system.getNumCpuTicksLeftNextEvent())
|
|
|
|
count = bx_pc_system.getNumCpuTicksLeftNextEvent();
|
|
|
|
|
|
|
|
// If after all the restrictions, there is anything left to do...
|
|
|
|
if (count) {
|
|
|
|
// Before we copy memory, we need to make sure that the segments
|
|
|
|
// allow the accesses up to the given source and dest offset. If
|
|
|
|
// the cache.valid bits have SegAccessWOK and ROK, we know that
|
|
|
|
// the cache is valid for those operations, and that the segments
|
|
|
|
// are non expand-down (thus we can make a simple limit check).
|
|
|
|
if ( !(dstSegPtr->cache.valid & SegAccessWOK) ) return 0;
|
|
|
|
|
2005-08-02 01:40:17 +04:00
|
|
|
if (BX_CPU_THIS_PTR cpu_mode != BX_MODE_LONG_64)
|
2005-07-04 21:44:08 +04:00
|
|
|
{
|
|
|
|
Bit32u dstSegLimit = dstSegPtr->cache.u.segment.limit_scaled;
|
|
|
|
|
|
|
|
if (! i->as32L()) {
|
|
|
|
// For 16-bit addressing mode, clamp the segment limits to 16bits
|
|
|
|
// so we don't have to worry about computations using di
|
|
|
|
// rolling over 16-bit boundaries.
|
|
|
|
if (dstSegLimit > 0xffff)
|
|
|
|
dstSegLimit = 0xffff;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now make sure transfer will fit within the constraints of the
|
|
|
|
// segment boundaries, 0..limit for non expand-down. We know
|
|
|
|
// count >= 1 here.
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
// Counting downward.
|
|
|
|
Bit32u minOffset = (count-1) << 1;
|
|
|
|
if ( dstOff < minOffset )
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
// Counting upward.
|
|
|
|
Bit32u dstMaxOffset = (dstSegLimit - (count<<1)) + 1;
|
|
|
|
if ( dstOff > dstMaxOffset )
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Transfer data directly using host addresses
|
|
|
|
for (unsigned j=0; j<count; j++) {
|
|
|
|
* (Bit16u *) hostAddrDst = val;
|
|
|
|
hostAddrDst += pointerDelta;
|
|
|
|
}
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
Bit32u BX_CPU_C::FastRepSTOSD(bxInstruction_c *i, unsigned dstSeg, bx_address dstOff, Bit32u val, Bit32u count)
|
|
|
|
{
|
|
|
|
Bit32u dwordsFitDst;
|
|
|
|
signed int pointerDelta;
|
|
|
|
bx_address laddrDst;
|
|
|
|
Bit32u paddrDst;
|
|
|
|
|
|
|
|
bx_segment_reg_t *dstSegPtr = &BX_CPU_THIS_PTR sregs[dstSeg];
|
|
|
|
|
|
|
|
write_virtual_checks(dstSegPtr, dstOff, 4);
|
|
|
|
laddrDst = BX_CPU_THIS_PTR get_segment_base(dstSeg) + dstOff;
|
|
|
|
if (BX_CPU_THIS_PTR cr0.pg) {
|
|
|
|
paddrDst = dtranslate_linear(laddrDst, CPL==3, BX_WRITE);
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
paddrDst = laddrDst;
|
|
|
|
}
|
|
|
|
|
|
|
|
// If we want to write directly into the physical memory array,
|
|
|
|
// we need the A20 address.
|
|
|
|
paddrDst = A20ADDR(paddrDst);
|
|
|
|
Bit8u *hostAddrDst = BX_CPU_THIS_PTR mem->getHostMemAddr(BX_CPU_THIS,
|
|
|
|
paddrDst, BX_WRITE);
|
|
|
|
|
|
|
|
if (! hostAddrDst) return 0;
|
|
|
|
|
|
|
|
// See how many dwords can fit in the rest of this page.
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
// Counting downward.
|
|
|
|
// Note: 1st dword must not cross page boundary.
|
|
|
|
if ((paddrDst & 0xfff) > 0xffc) return 0;
|
|
|
|
dwordsFitDst = (4 + (paddrDst & 0xfff)) >> 2;
|
|
|
|
pointerDelta = (signed int) -4;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
// Counting upward.
|
|
|
|
dwordsFitDst = (0x1000 - (paddrDst & 0xfff)) >> 2;
|
|
|
|
pointerDelta = (signed int) 4;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Restrict dword count to the number that will fit in either
|
|
|
|
// source or dest pages.
|
|
|
|
if (count > dwordsFitDst)
|
|
|
|
count = dwordsFitDst;
|
|
|
|
if (count > bx_pc_system.getNumCpuTicksLeftNextEvent())
|
|
|
|
count = bx_pc_system.getNumCpuTicksLeftNextEvent();
|
|
|
|
|
|
|
|
// If after all the restrictions, there is anything left to do...
|
|
|
|
if (count) {
|
|
|
|
// Before we copy memory, we need to make sure that the segments
|
|
|
|
// allow the accesses up to the given source and dest offset. If
|
|
|
|
// the cache.valid bits have SegAccessWOK and ROK, we know that
|
|
|
|
// the cache is valid for those operations, and that the segments
|
|
|
|
// are non expand-down (thus we can make a simple limit check).
|
|
|
|
if ( !(dstSegPtr->cache.valid & SegAccessWOK) ) return 0;
|
|
|
|
|
2005-08-02 01:40:17 +04:00
|
|
|
if (BX_CPU_THIS_PTR cpu_mode != BX_MODE_LONG_64)
|
2005-07-04 21:44:08 +04:00
|
|
|
{
|
|
|
|
Bit32u dstSegLimit = dstSegPtr->cache.u.segment.limit_scaled;
|
|
|
|
|
|
|
|
if (! i->as32L()) {
|
|
|
|
// For 16-bit addressing mode, clamp the segment limits to 16bits
|
|
|
|
// so we don't have to worry about computations using di
|
|
|
|
// rolling over 16-bit boundaries.
|
|
|
|
if (dstSegLimit > 0xffff)
|
|
|
|
dstSegLimit = 0xffff;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Now make sure transfer will fit within the constraints of the
|
|
|
|
// segment boundaries, 0..limit for non expand-down. We know
|
|
|
|
// count >= 1 here.
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
// Counting downward.
|
|
|
|
Bit32u minOffset = (count-1) << 2;
|
|
|
|
if ( dstOff < minOffset )
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
// Counting upward.
|
|
|
|
Bit32u dstMaxOffset = (dstSegLimit - (count<<2)) + 1;
|
|
|
|
if ( dstOff > dstMaxOffset )
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Transfer data directly using host addresses
|
|
|
|
for (unsigned j=0; j<count; j++) {
|
|
|
|
* (Bit32u *) hostAddrDst = val;
|
|
|
|
hostAddrDst += pointerDelta;
|
|
|
|
}
|
|
|
|
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
2001-04-10 05:04:59 +04:00
|
|
|
/* MOVSB ES:[EDI], DS:[ESI] DS may be overridden
|
|
|
|
* mov string from DS:[ESI] into ES:[EDI]
|
|
|
|
*/
|
2004-11-21 02:26:32 +03:00
|
|
|
void BX_CPU_C::MOVSB_XbYb(bxInstruction_c *i)
|
2001-04-10 05:04:59 +04:00
|
|
|
{
|
|
|
|
unsigned seg;
|
|
|
|
Bit8u temp8;
|
|
|
|
|
2002-09-18 09:36:48 +04:00
|
|
|
if (!BX_NULL_SEG_REG(i->seg())) {
|
|
|
|
seg = i->seg();
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else {
|
|
|
|
seg = BX_SEG_REG_DS;
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
#if BX_SUPPORT_X86_64
|
2002-09-18 09:36:48 +04:00
|
|
|
if (i->as64L()) {
|
2002-09-15 09:09:18 +04:00
|
|
|
Bit64u rsi, rdi;
|
|
|
|
|
|
|
|
rsi = RSI;
|
|
|
|
rdi = RDI;
|
|
|
|
|
|
|
|
read_virtual_byte(seg, rsi, &temp8);
|
|
|
|
write_virtual_byte(BX_SEG_REG_ES, rdi, &temp8);
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
/* decrement RSI, RDI */
|
|
|
|
rsi--;
|
|
|
|
rdi--;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
else {
|
|
|
|
/* increment RSI, RDI */
|
|
|
|
rsi++;
|
|
|
|
rdi++;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
|
|
|
|
RSI = rsi;
|
|
|
|
RDI = rdi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
else
|
|
|
|
#endif // #if BX_SUPPORT_X86_64
|
2004-11-21 02:26:32 +03:00
|
|
|
if (i->as32L())
|
|
|
|
{
|
2001-04-10 05:04:59 +04:00
|
|
|
Bit32u esi, edi;
|
|
|
|
|
|
|
|
esi = ESI;
|
|
|
|
edi = EDI;
|
|
|
|
|
|
|
|
read_virtual_byte(seg, esi, &temp8);
|
|
|
|
write_virtual_byte(BX_SEG_REG_ES, edi, &temp8);
|
|
|
|
|
2002-09-12 22:10:46 +04:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
2001-04-10 05:04:59 +04:00
|
|
|
/* decrement ESI, EDI */
|
|
|
|
esi--;
|
|
|
|
edi--;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else {
|
|
|
|
/* increment ESI, EDI */
|
|
|
|
esi++;
|
|
|
|
edi++;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
// zero extension of RSI/RDI
|
|
|
|
RSI = esi;
|
|
|
|
RDI = edi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else
|
2004-11-21 02:26:32 +03:00
|
|
|
{ /* 16 bit address mode */
|
2005-07-04 21:44:08 +04:00
|
|
|
unsigned incr = 1;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
Bit16u si = SI;
|
|
|
|
Bit16u di = DI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-02 22:44:35 +04:00
|
|
|
#if BX_SupportRepeatSpeedups
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
#if (BX_DEBUGGER == 0)
|
|
|
|
/* If conditions are right, we can transfer IO to physical memory
|
2005-07-04 21:44:08 +04:00
|
|
|
* in a batch, rather than one instruction at a time */
|
2004-11-21 02:26:32 +03:00
|
|
|
if (i->repUsedL() && !BX_CPU_THIS_PTR async_event)
|
|
|
|
{
|
2005-07-04 21:44:08 +04:00
|
|
|
Bit32u byteCount = CX;
|
|
|
|
BX_ASSERT(byteCount > 0);
|
|
|
|
byteCount = FastRepMOVSB(i, seg, si, BX_SEG_REG_ES, di, byteCount);
|
|
|
|
if (byteCount)
|
|
|
|
{
|
|
|
|
// Decrement the ticks count by the number of iterations, minus
|
|
|
|
// one, since the main cpu loop will decrement one. Also,
|
|
|
|
// the count is predecremented before examined, so defintely
|
|
|
|
// don't roll it under zero.
|
|
|
|
BX_TICKN(byteCount-1);
|
|
|
|
|
|
|
|
// Decrement eCX. Note, the main loop will decrement 1 also, so
|
|
|
|
// decrement by one less than expected, like the case above.
|
|
|
|
CX -= (byteCount-1);
|
|
|
|
|
|
|
|
incr = byteCount;
|
|
|
|
goto doIncr16;
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-03 23:38:27 +04:00
|
|
|
#endif // (BX_DEBUGGER == 0)
|
|
|
|
#endif // BX_SupportRepeatSpeedups
|
|
|
|
|
2001-04-10 05:04:59 +04:00
|
|
|
read_virtual_byte(seg, si, &temp8);
|
|
|
|
write_virtual_byte(BX_SEG_REG_ES, di, &temp8);
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
|
2002-09-03 23:38:27 +04:00
|
|
|
#if BX_SupportRepeatSpeedups
|
|
|
|
#if (BX_DEBUGGER == 0)
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
doIncr16:
|
2002-09-03 23:38:27 +04:00
|
|
|
#endif // (BX_DEBUGGER == 0)
|
|
|
|
#endif // BX_SupportRepeatSpeedups
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-12 22:10:46 +04:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
2001-04-10 05:04:59 +04:00
|
|
|
/* decrement SI, DI */
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
si -= incr;
|
|
|
|
di -= incr;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else {
|
|
|
|
/* increment SI, DI */
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
si += incr;
|
|
|
|
di += incr;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
|
|
|
SI = si;
|
|
|
|
DI = di;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
}
|
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
/* 16 bit opsize mode */
|
|
|
|
void BX_CPU_C::MOVSW_XwYw(bxInstruction_c *i)
|
2001-04-10 05:04:59 +04:00
|
|
|
{
|
2004-11-21 02:26:32 +03:00
|
|
|
Bit16u temp16;
|
2001-04-10 05:04:59 +04:00
|
|
|
unsigned seg;
|
|
|
|
|
2002-09-18 09:36:48 +04:00
|
|
|
if (!BX_NULL_SEG_REG(i->seg())) {
|
|
|
|
seg = i->seg();
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else {
|
|
|
|
seg = BX_SEG_REG_DS;
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
#if BX_SUPPORT_X86_64
|
2002-09-18 09:36:48 +04:00
|
|
|
if (i->as64L()) {
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
Bit64u rsi = RSI;
|
|
|
|
Bit64u rdi = RDI;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_word(seg, rsi, &temp16);
|
|
|
|
write_virtual_word(BX_SEG_REG_ES, rdi, &temp16);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rsi -= 2;
|
|
|
|
rdi -= 2;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
rsi += 2;
|
|
|
|
rdi += 2;
|
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
|
|
|
|
RSI = rsi;
|
|
|
|
RDI = rdi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
else
|
|
|
|
#endif // #if BX_SUPPORT_X86_64
|
2002-09-18 09:36:48 +04:00
|
|
|
if (i->as32L()) {
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
Bit32u esi = ESI;
|
|
|
|
Bit32u edi = EDI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_word(seg, esi, &temp16);
|
|
|
|
write_virtual_word(BX_SEG_REG_ES, edi, &temp16);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
esi -= 2;
|
|
|
|
edi -= 2;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
esi += 2;
|
|
|
|
edi += 2;
|
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
// zero extension of RSI/RDI
|
|
|
|
RSI = esi;
|
|
|
|
RDI = edi;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{ /* 16bit address mode */
|
2005-07-04 21:44:08 +04:00
|
|
|
unsigned incr = 2;
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
Bit16u si = SI;
|
|
|
|
Bit16u di = DI;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2002-09-02 22:44:35 +04:00
|
|
|
#if BX_SupportRepeatSpeedups
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
#if (BX_DEBUGGER == 0)
|
|
|
|
/* If conditions are right, we can transfer IO to physical memory
|
|
|
|
* in a batch, rather than one instruction at a time.
|
|
|
|
*/
|
2004-11-21 02:26:32 +03:00
|
|
|
if (i->repUsedL() && !BX_CPU_THIS_PTR async_event)
|
|
|
|
{
|
|
|
|
Bit32u wordCount = CX;
|
2005-07-04 21:44:08 +04:00
|
|
|
BX_ASSERT(wordCount > 0);
|
|
|
|
wordCount = FastRepMOVSW(i, seg, si, BX_SEG_REG_ES, di, wordCount);
|
|
|
|
if (wordCount)
|
|
|
|
{
|
|
|
|
// Decrement the ticks count by the number of iterations, minus
|
|
|
|
// one, since the main cpu loop will decrement one. Also,
|
|
|
|
// the count is predecremented before examined, so defintely
|
|
|
|
// don't roll it under zero.
|
|
|
|
BX_TICKN(wordCount-1);
|
|
|
|
|
|
|
|
// Decrement eCX. Note, the main loop will decrement 1 also, so
|
|
|
|
// decrement by one less than expected, like the case above.
|
|
|
|
CX -= (wordCount-1);
|
|
|
|
|
|
|
|
incr = wordCount << 1; // count * 2
|
|
|
|
goto doIncr16;
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-03 23:38:27 +04:00
|
|
|
#endif // (BX_DEBUGGER == 0)
|
|
|
|
#endif // BX_SupportRepeatSpeedups
|
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_word(seg, si, &temp16);
|
|
|
|
write_virtual_word(BX_SEG_REG_ES, di, &temp16);
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
|
2002-09-03 23:38:27 +04:00
|
|
|
#if BX_SupportRepeatSpeedups
|
|
|
|
#if (BX_DEBUGGER == 0)
|
2004-11-21 02:26:32 +03:00
|
|
|
doIncr16:
|
2002-09-03 23:38:27 +04:00
|
|
|
#endif
|
|
|
|
#endif
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
/* decrement SI, DI */
|
|
|
|
si -= incr;
|
|
|
|
di -= incr;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
/* increment SI, DI */
|
|
|
|
si += incr;
|
|
|
|
di += incr;
|
2001-04-10 05:04:59 +04:00
|
|
|
}
|
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SI = si;
|
|
|
|
DI = di;
|
|
|
|
}
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
/* 32 bit opsize mode */
|
|
|
|
void BX_CPU_C::MOVSD_XdYd(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit32u temp32;
|
|
|
|
unsigned seg;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (!BX_NULL_SEG_REG(i->seg())) {
|
|
|
|
seg = i->seg();
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
else {
|
|
|
|
seg = BX_SEG_REG_DS;
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
#if BX_SUPPORT_X86_64
|
|
|
|
if (i->as64L()) {
|
|
|
|
Bit64u rsi = RSI;
|
|
|
|
Bit64u rdi = RDI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_dword(seg, rsi, &temp32);
|
|
|
|
write_virtual_dword(BX_SEG_REG_ES, rdi, &temp32);
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rsi -= 4;
|
|
|
|
rdi -= 4;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
rsi += 4;
|
|
|
|
rdi += 4;
|
|
|
|
}
|
|
|
|
|
|
|
|
RSI = rsi;
|
|
|
|
RDI = rdi;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
#endif // #if BX_SUPPORT_X86_64
|
|
|
|
if (i->as32L())
|
|
|
|
{
|
2005-07-04 21:44:08 +04:00
|
|
|
unsigned incr = 4;
|
2004-11-21 02:26:32 +03:00
|
|
|
|
|
|
|
Bit32u esi = ESI;
|
|
|
|
Bit32u edi = EDI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-02 22:44:35 +04:00
|
|
|
#if BX_SupportRepeatSpeedups
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
#if (BX_DEBUGGER == 0)
|
|
|
|
/* If conditions are right, we can transfer IO to physical memory
|
|
|
|
* in a batch, rather than one instruction at a time.
|
|
|
|
*/
|
2004-11-21 02:26:32 +03:00
|
|
|
if (i->repUsedL() && !BX_CPU_THIS_PTR async_event)
|
|
|
|
{
|
|
|
|
Bit32u dwordCount = ECX;
|
2005-07-04 21:44:08 +04:00
|
|
|
BX_ASSERT(dwordCount > 0);
|
|
|
|
dwordCount = FastRepMOVSD(i, seg, esi, BX_SEG_REG_ES, edi, dwordCount);
|
2004-11-21 02:26:32 +03:00
|
|
|
if (dwordCount)
|
|
|
|
{
|
2005-07-04 21:44:08 +04:00
|
|
|
// Decrement the ticks count by the number of iterations, minus
|
|
|
|
// one, since the main cpu loop will decrement one. Also,
|
|
|
|
// the count is predecremented before examined, so defintely
|
|
|
|
// don't roll it under zero.
|
|
|
|
BX_TICKN(dwordCount-1);
|
|
|
|
|
|
|
|
// Decrement eCX. Note, the main loop will decrement 1 also, so
|
|
|
|
// decrement by one less than expected, like the case above.
|
|
|
|
ECX -= (dwordCount-1);
|
|
|
|
|
|
|
|
incr = dwordCount << 2; // count * 4
|
|
|
|
goto doIncr32;
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-03 23:38:27 +04:00
|
|
|
#endif // (BX_DEBUGGER == 0)
|
|
|
|
#endif // BX_SupportRepeatSpeedups
|
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_dword(seg, esi, &temp32);
|
|
|
|
write_virtual_dword(BX_SEG_REG_ES, edi, &temp32);
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
|
2002-09-03 23:38:27 +04:00
|
|
|
#if BX_SupportRepeatSpeedups
|
|
|
|
#if (BX_DEBUGGER == 0)
|
2004-11-21 02:26:32 +03:00
|
|
|
doIncr32:
|
2002-09-03 23:38:27 +04:00
|
|
|
#endif
|
|
|
|
#endif
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
esi -= incr;
|
|
|
|
edi -= incr;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
esi += incr;
|
|
|
|
edi += incr;
|
|
|
|
}
|
|
|
|
|
|
|
|
// zero extension of RSI/RDI
|
|
|
|
RSI = esi;
|
|
|
|
RDI = edi;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{ /* 16bit address mode */
|
|
|
|
Bit16u si = SI;
|
|
|
|
Bit16u di = DI;
|
|
|
|
|
|
|
|
read_virtual_dword(seg, si, &temp32);
|
|
|
|
write_virtual_dword(BX_SEG_REG_ES, di, &temp32);
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
si -= 4;
|
|
|
|
di -= 4;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
si += 4;
|
|
|
|
di += 4;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
|
|
|
SI = si;
|
|
|
|
DI = di;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#if BX_SUPPORT_X86_64
|
|
|
|
|
|
|
|
/* 64 bit opsize mode */
|
|
|
|
void BX_CPU_C::MOVSQ_XqYq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u temp64;
|
|
|
|
unsigned seg;
|
|
|
|
|
|
|
|
if (!BX_NULL_SEG_REG(i->seg())) {
|
|
|
|
seg = i->seg();
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
else {
|
|
|
|
seg = BX_SEG_REG_DS;
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
|
|
|
|
if (i->as64L()) {
|
|
|
|
Bit64u rsi = RSI;
|
|
|
|
Bit64u rdi = RDI;
|
|
|
|
|
|
|
|
read_virtual_qword(seg, rsi, &temp64);
|
|
|
|
write_virtual_qword(BX_SEG_REG_ES, rdi, &temp64);
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rsi -= 8;
|
|
|
|
rdi -= 8;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
rsi += 8;
|
|
|
|
rdi += 8;
|
|
|
|
}
|
|
|
|
|
|
|
|
RSI = rsi;
|
|
|
|
RDI = rdi;
|
|
|
|
}
|
|
|
|
else /* 32-bit address size mode */
|
|
|
|
{
|
|
|
|
Bit32u esi = ESI;
|
|
|
|
Bit32u edi = EDI;
|
|
|
|
|
|
|
|
read_virtual_qword(seg, esi, &temp64);
|
|
|
|
write_virtual_qword(BX_SEG_REG_ES, edi, &temp64);
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
esi -= 8;
|
|
|
|
edi -= 8;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
esi += 8;
|
|
|
|
edi += 8;
|
|
|
|
}
|
|
|
|
|
|
|
|
// zero extension of RSI/RDI
|
|
|
|
RSI = esi;
|
|
|
|
RDI = edi;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
}
|
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
#endif
|
|
|
|
|
|
|
|
void BX_CPU_C::CMPSB_XbYb(bxInstruction_c *i)
|
2001-04-10 05:04:59 +04:00
|
|
|
{
|
|
|
|
unsigned seg;
|
|
|
|
Bit8u op1_8, op2_8, diff_8;
|
|
|
|
|
2002-09-18 09:36:48 +04:00
|
|
|
if (!BX_NULL_SEG_REG(i->seg())) {
|
|
|
|
seg = i->seg();
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else {
|
|
|
|
seg = BX_SEG_REG_DS;
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
#if BX_SUPPORT_X86_64
|
2002-09-18 09:36:48 +04:00
|
|
|
if (i->as64L()) {
|
2004-11-21 02:26:32 +03:00
|
|
|
Bit64u rsi = RSI;
|
|
|
|
Bit64u rdi = RDI;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
|
|
|
read_virtual_byte(seg, rsi, &op1_8);
|
|
|
|
read_virtual_byte(BX_SEG_REG_ES, rdi, &op2_8);
|
|
|
|
|
|
|
|
diff_8 = op1_8 - op2_8;
|
|
|
|
|
2004-08-17 00:18:01 +04:00
|
|
|
SET_FLAGS_OSZAPC_8(op1_8, op2_8, diff_8, BX_INSTR_COMPARE8);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rsi--;
|
|
|
|
rdi--;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
else {
|
|
|
|
rsi++;
|
|
|
|
rdi++;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
|
|
|
|
RDI = rdi;
|
|
|
|
RSI = rsi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
else
|
|
|
|
#endif // #if BX_SUPPORT_X86_64
|
2002-09-18 09:36:48 +04:00
|
|
|
if (i->as32L()) {
|
2004-11-21 02:26:32 +03:00
|
|
|
Bit32u esi = ESI;
|
|
|
|
Bit32u edi = EDI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
|
|
|
read_virtual_byte(seg, esi, &op1_8);
|
|
|
|
read_virtual_byte(BX_SEG_REG_ES, edi, &op2_8);
|
|
|
|
|
|
|
|
diff_8 = op1_8 - op2_8;
|
|
|
|
|
2004-08-17 00:18:01 +04:00
|
|
|
SET_FLAGS_OSZAPC_8(op1_8, op2_8, diff_8, BX_INSTR_COMPARE8);
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-12 22:10:46 +04:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
2001-04-10 05:04:59 +04:00
|
|
|
esi--;
|
|
|
|
edi--;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else {
|
|
|
|
esi++;
|
|
|
|
edi++;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
// zero extension of RSI/RDI
|
|
|
|
RDI = edi;
|
|
|
|
RSI = esi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else
|
2004-11-21 02:26:32 +03:00
|
|
|
{ /* 16bit address mode */
|
|
|
|
Bit16u si = SI;
|
|
|
|
Bit16u di = DI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
|
|
|
read_virtual_byte(seg, si, &op1_8);
|
|
|
|
read_virtual_byte(BX_SEG_REG_ES, di, &op2_8);
|
|
|
|
|
|
|
|
diff_8 = op1_8 - op2_8;
|
|
|
|
|
2004-08-17 00:18:01 +04:00
|
|
|
SET_FLAGS_OSZAPC_8(op1_8, op2_8, diff_8, BX_INSTR_COMPARE8);
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-12 22:10:46 +04:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
2001-04-10 05:04:59 +04:00
|
|
|
si--;
|
|
|
|
di--;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else {
|
|
|
|
si++;
|
|
|
|
di++;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
|
|
|
DI = di;
|
|
|
|
SI = si;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
}
|
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
/* 16 bit opsize mode */
|
|
|
|
void BX_CPU_C::CMPSW_XwYw(bxInstruction_c *i)
|
2001-04-10 05:04:59 +04:00
|
|
|
{
|
2004-11-21 02:26:32 +03:00
|
|
|
Bit16u op1_16, op2_16, diff_16;
|
2001-04-10 05:04:59 +04:00
|
|
|
unsigned seg;
|
|
|
|
|
2002-09-18 09:36:48 +04:00
|
|
|
if (!BX_NULL_SEG_REG(i->seg())) {
|
|
|
|
seg = i->seg();
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else {
|
|
|
|
seg = BX_SEG_REG_DS;
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
#if BX_SUPPORT_X86_64
|
2002-09-18 09:36:48 +04:00
|
|
|
if (i->as64L()) {
|
2002-09-15 09:09:18 +04:00
|
|
|
Bit64u rsi, rdi;
|
|
|
|
|
|
|
|
rsi = RSI;
|
|
|
|
rdi = RDI;
|
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_word(seg, rsi, &op1_16);
|
|
|
|
read_virtual_word(BX_SEG_REG_ES, rdi, &op2_16);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
diff_16 = op1_16 - op2_16;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SET_FLAGS_OSZAPC_16(op1_16, op2_16, diff_16, BX_INSTR_COMPARE16);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rsi -= 2;
|
|
|
|
rdi -= 2;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
rsi += 2;
|
|
|
|
rdi += 2;
|
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
RDI = rdi;
|
|
|
|
RSI = rsi;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
#endif // #if BX_SUPPORT_X86_64
|
|
|
|
if (i->as32L()) {
|
|
|
|
Bit32u esi = ESI;
|
|
|
|
Bit32u edi = EDI;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_word(seg, esi, &op1_16);
|
|
|
|
read_virtual_word(BX_SEG_REG_ES, edi, &op2_16);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
diff_16 = op1_16 - op2_16;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SET_FLAGS_OSZAPC_16(op1_16, op2_16, diff_16, BX_INSTR_COMPARE16);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
esi -= 2;
|
|
|
|
edi -= 2;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
esi += 2;
|
|
|
|
edi += 2;
|
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
// zero extension of RSI/RDI
|
|
|
|
RDI = edi;
|
|
|
|
RSI = esi;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{ /* 16 bit address mode */
|
|
|
|
Bit16u si = SI;
|
|
|
|
Bit16u di = DI;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_word(seg, si, &op1_16);
|
|
|
|
read_virtual_word(BX_SEG_REG_ES, di, &op2_16);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
diff_16 = op1_16 - op2_16;
|
|
|
|
SET_FLAGS_OSZAPC_16(op1_16, op2_16, diff_16, BX_INSTR_COMPARE16);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
si -= 2;
|
|
|
|
di -= 2;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
si += 2;
|
|
|
|
di += 2;
|
|
|
|
}
|
|
|
|
|
|
|
|
DI = di;
|
|
|
|
SI = si;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* 32 bit opsize mode */
|
|
|
|
void BX_CPU_C::CMPSD_XdYd(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit32u op1_32, op2_32, diff_32;
|
|
|
|
unsigned seg;
|
|
|
|
|
|
|
|
if (!BX_NULL_SEG_REG(i->seg())) {
|
|
|
|
seg = i->seg();
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
else {
|
|
|
|
seg = BX_SEG_REG_DS;
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
|
|
|
|
#if BX_SUPPORT_X86_64
|
|
|
|
if (i->as64L()) {
|
|
|
|
Bit64u rsi = RSI;
|
|
|
|
Bit64u rdi = RDI;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
Bit32u op1_32, op2_32, diff_32;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_dword(seg, rsi, &op1_32);
|
|
|
|
read_virtual_dword(BX_SEG_REG_ES, rdi, &op2_32);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
diff_32 = op1_32 - op2_32;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SET_FLAGS_OSZAPC_32(op1_32, op2_32, diff_32, BX_INSTR_COMPARE32);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rsi -= 4;
|
|
|
|
rdi -= 4;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
rsi += 4;
|
|
|
|
rdi += 4;
|
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
|
|
|
|
RDI = rdi;
|
|
|
|
RSI = rsi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
else
|
|
|
|
#endif // #if BX_SUPPORT_X86_64
|
2002-09-18 09:36:48 +04:00
|
|
|
if (i->as32L()) {
|
2004-11-21 02:26:32 +03:00
|
|
|
Bit32u esi = ESI;
|
|
|
|
Bit32u edi = EDI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_dword(seg, esi, &op1_32);
|
|
|
|
read_virtual_dword(BX_SEG_REG_ES, edi, &op2_32);
|
|
|
|
|
|
|
|
diff_32 = op1_32 - op2_32;
|
|
|
|
|
|
|
|
SET_FLAGS_OSZAPC_32(op1_32, op2_32, diff_32, BX_INSTR_COMPARE32);
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
esi -= 4;
|
|
|
|
edi -= 4;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
esi += 4;
|
|
|
|
edi += 4;
|
|
|
|
}
|
|
|
|
|
|
|
|
// zero extension of RSI/RDI
|
|
|
|
RDI = edi;
|
|
|
|
RSI = esi;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{ /* 16 bit address mode */
|
|
|
|
Bit16u si = SI;
|
|
|
|
Bit16u di = DI;
|
|
|
|
|
|
|
|
read_virtual_dword(seg, si, &op1_32);
|
|
|
|
read_virtual_dword(BX_SEG_REG_ES, di, &op2_32);
|
|
|
|
|
|
|
|
diff_32 = op1_32 - op2_32;
|
|
|
|
|
|
|
|
SET_FLAGS_OSZAPC_32(op1_32, op2_32, diff_32, BX_INSTR_COMPARE32);
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
si -= 4;
|
|
|
|
di -= 4;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
si += 4;
|
|
|
|
di += 4;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
DI = di;
|
|
|
|
SI = si;
|
|
|
|
}
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
#if BX_SUPPORT_X86_64
|
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
/* 64 bit opsize mode */
|
|
|
|
void BX_CPU_C::CMPSQ_XqYq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op1_64, op2_64, diff_64;
|
|
|
|
unsigned seg;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (!BX_NULL_SEG_REG(i->seg())) {
|
|
|
|
seg = i->seg();
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
else {
|
|
|
|
seg = BX_SEG_REG_DS;
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (i->as64L()) {
|
|
|
|
Bit64u rsi = RSI;
|
|
|
|
Bit64u rdi = RDI;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_qword(seg, rsi, &op1_64);
|
|
|
|
read_virtual_qword(BX_SEG_REG_ES, rdi, &op2_64);
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
diff_64 = op1_64 - op2_64;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SET_FLAGS_OSZAPC_64(op1_64, op2_64, diff_64, BX_INSTR_COMPARE64);
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rsi -= 8;
|
|
|
|
rdi -= 8;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
rsi += 8;
|
|
|
|
rdi += 8;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
RDI = rdi;
|
|
|
|
RSI = rsi;
|
|
|
|
}
|
|
|
|
else /* 32 bit address size */
|
|
|
|
{
|
|
|
|
Bit32u esi = ESI;
|
|
|
|
Bit32u edi = EDI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_qword(seg, esi, &op1_64);
|
|
|
|
read_virtual_qword(BX_SEG_REG_ES, edi, &op2_64);
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
diff_64 = op1_64 - op2_64;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SET_FLAGS_OSZAPC_64(op1_64, op2_64, diff_64, BX_INSTR_COMPARE64);
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
esi -= 8;
|
|
|
|
edi -= 8;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
esi += 8;
|
|
|
|
edi += 8;
|
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
// zero extension of RSI/RDI
|
2002-09-15 09:09:18 +04:00
|
|
|
RDI = edi;
|
|
|
|
RSI = esi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
|
|
|
void BX_CPU_C::SCASB_ALXb(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit8u op1_8, op2_8, diff_8;
|
|
|
|
|
|
|
|
op1_8 = AL;
|
|
|
|
|
|
|
|
#if BX_SUPPORT_X86_64
|
|
|
|
if (i->as64L()) {
|
|
|
|
Bit64u rdi = RDI;
|
|
|
|
|
|
|
|
read_virtual_byte(BX_SEG_REG_ES, rdi, &op2_8);
|
|
|
|
|
|
|
|
diff_8 = op1_8 - op2_8;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SET_FLAGS_OSZAPC_8(op1_8, op2_8, diff_8, BX_INSTR_COMPARE8);
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rdi--;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
rdi++;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
RDI = rdi;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
#endif // #if BX_SUPPORT_X86_64
|
|
|
|
if (i->as32L()) {
|
|
|
|
Bit32u edi = EDI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_byte(BX_SEG_REG_ES, edi, &op2_8);
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
diff_8 = op1_8 - op2_8;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SET_FLAGS_OSZAPC_8(op1_8, op2_8, diff_8, BX_INSTR_COMPARE8);
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
edi--;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
edi++;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
// zero extension of RDI
|
|
|
|
RDI = edi;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{ /* 16bit address mode */
|
|
|
|
Bit16u di = DI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_byte(BX_SEG_REG_ES, di, &op2_8);
|
2002-10-08 02:51:58 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
diff_8 = op1_8 - op2_8;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SET_FLAGS_OSZAPC_8(op1_8, op2_8, diff_8, BX_INSTR_COMPARE8);
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
di--;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
di++;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
|
|
|
DI = di;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
}
|
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
/* 16 bit opsize mode */
|
|
|
|
void BX_CPU_C::SCASW_AXXw(bxInstruction_c *i)
|
2001-04-10 05:04:59 +04:00
|
|
|
{
|
2004-11-21 02:26:32 +03:00
|
|
|
Bit16u op1_16, op2_16, diff_16;
|
|
|
|
|
|
|
|
op1_16 = AX;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
#if BX_SUPPORT_X86_64
|
2002-09-18 09:36:48 +04:00
|
|
|
if (i->as64L()) {
|
2004-08-15 00:44:48 +04:00
|
|
|
Bit64u rdi = RDI;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_word(BX_SEG_REG_ES, rdi, &op2_16);
|
|
|
|
diff_16 = op1_16 - op2_16;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SET_FLAGS_OSZAPC_16(op1_16, op2_16, diff_16, BX_INSTR_COMPARE16);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
2004-11-21 02:26:32 +03:00
|
|
|
rdi -= 2;
|
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
else {
|
2004-11-21 02:26:32 +03:00
|
|
|
rdi += 2;
|
2002-09-15 09:09:18 +04:00
|
|
|
}
|
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
RDI = rdi;
|
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
else
|
|
|
|
#endif // #if BX_SUPPORT_X86_64
|
2002-09-18 09:36:48 +04:00
|
|
|
if (i->as32L()) {
|
2004-08-15 00:44:48 +04:00
|
|
|
Bit32u edi = EDI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_word(BX_SEG_REG_ES, edi, &op2_16);
|
|
|
|
diff_16 = op1_16 - op2_16;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SET_FLAGS_OSZAPC_16(op1_16, op2_16, diff_16, BX_INSTR_COMPARE16);
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-12 22:10:46 +04:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
2004-11-21 02:26:32 +03:00
|
|
|
edi -= 2;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else {
|
2004-11-21 02:26:32 +03:00
|
|
|
edi += 2;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
// zero extension of RDI
|
|
|
|
RDI = edi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else
|
2004-11-21 02:26:32 +03:00
|
|
|
{ /* 16bit address mode */
|
2004-08-15 00:44:48 +04:00
|
|
|
Bit16u di = DI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_word(BX_SEG_REG_ES, di, &op2_16);
|
|
|
|
diff_16 = op1_16 - op2_16;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SET_FLAGS_OSZAPC_16(op1_16, op2_16, diff_16, BX_INSTR_COMPARE16);
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-12 22:10:46 +04:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
2004-11-21 02:26:32 +03:00
|
|
|
di -= 2;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else {
|
2004-11-21 02:26:32 +03:00
|
|
|
di += 2;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
|
|
|
DI = di;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
}
|
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
/* 32 bit opsize mode */
|
|
|
|
void BX_CPU_C::SCASD_EAXXd(bxInstruction_c *i)
|
2001-04-10 05:04:59 +04:00
|
|
|
{
|
2004-11-21 02:26:32 +03:00
|
|
|
Bit32u op1_32, op2_32, diff_32;
|
|
|
|
|
|
|
|
op1_32 = EAX;
|
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
#if BX_SUPPORT_X86_64
|
2002-09-18 09:36:48 +04:00
|
|
|
if (i->as64L()) {
|
2004-08-15 00:44:48 +04:00
|
|
|
Bit64u rdi = RDI;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_dword(BX_SEG_REG_ES, rdi, &op2_32);
|
|
|
|
diff_32 = op1_32 - op2_32;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SET_FLAGS_OSZAPC_32(op1_32, op2_32, diff_32, BX_INSTR_COMPARE32);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rdi -= 4;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
rdi += 4;
|
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
|
|
|
|
RDI = rdi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
else
|
|
|
|
#endif // #if BX_SUPPORT_X86_64
|
2002-09-18 09:36:48 +04:00
|
|
|
if (i->as32L()) {
|
2004-08-15 00:44:48 +04:00
|
|
|
Bit32u edi = EDI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_dword(BX_SEG_REG_ES, edi, &op2_32);
|
|
|
|
diff_32 = op1_32 - op2_32;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SET_FLAGS_OSZAPC_32(op1_32, op2_32, diff_32, BX_INSTR_COMPARE32);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
edi -= 4;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
edi += 4;
|
|
|
|
}
|
|
|
|
|
|
|
|
// zero extension of RDI
|
|
|
|
RDI = edi;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{ /* 16bit address mode */
|
|
|
|
Bit16u di = DI;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_dword(BX_SEG_REG_ES, di, &op2_32);
|
|
|
|
diff_32 = op1_32 - op2_32;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SET_FLAGS_OSZAPC_32(op1_32, op2_32, diff_32, BX_INSTR_COMPARE32);
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
di -= 4;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
di += 4;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
DI = di;
|
|
|
|
}
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
#if BX_SUPPORT_X86_64
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
/* 64 bit opsize mode */
|
|
|
|
void BX_CPU_C::SCASQ_RAXXq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u op1_64, op2_64, diff_64;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
op1_64 = RAX;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (i->as64L()) {
|
|
|
|
Bit64u rdi = RDI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_qword(BX_SEG_REG_ES, rdi, &op2_64);
|
|
|
|
diff_64 = op1_64 - op2_64;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SET_FLAGS_OSZAPC_64(op1_64, op2_64, diff_64, BX_INSTR_COMPARE64);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rdi -= 8;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
rdi += 8;
|
2001-04-10 05:04:59 +04:00
|
|
|
}
|
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
RDI = rdi;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{
|
|
|
|
Bit32u edi = EDI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_qword(BX_SEG_REG_ES, edi, &op2_64);
|
|
|
|
diff_64 = op1_64 - op2_64;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SET_FLAGS_OSZAPC_64(op1_64, op2_64, diff_64, BX_INSTR_COMPARE64);
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
edi -= 8;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
edi += 8;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
// zero extension of RDI
|
|
|
|
RDI = edi;
|
|
|
|
}
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-10-03 22:12:40 +04:00
|
|
|
#endif
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
void BX_CPU_C::STOSB_YbAL(bxInstruction_c *i)
|
2001-04-10 05:04:59 +04:00
|
|
|
{
|
2004-11-21 02:26:32 +03:00
|
|
|
Bit8u al = AL;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
|
|
|
#if BX_SUPPORT_X86_64
|
2002-09-18 09:36:48 +04:00
|
|
|
if (i->as64L()) {
|
2004-08-15 00:44:48 +04:00
|
|
|
Bit64u rdi = RDI;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
|
|
|
write_virtual_byte(BX_SEG_REG_ES, rdi, &al);
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rdi--;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
else {
|
|
|
|
rdi++;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
|
|
|
|
RDI = rdi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
else
|
|
|
|
#endif // #if BX_SUPPORT_X86_64
|
|
|
|
{
|
2005-07-04 21:44:08 +04:00
|
|
|
unsigned incr = 1;
|
2004-11-21 02:26:32 +03:00
|
|
|
Bit32u edi;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (i->as32L()) {
|
|
|
|
edi = EDI;
|
2001-04-10 05:04:59 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
else
|
2001-04-10 05:04:59 +04:00
|
|
|
{ /* 16bit address size */
|
2004-11-21 02:26:32 +03:00
|
|
|
edi = DI;
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2002-09-02 22:44:35 +04:00
|
|
|
#if BX_SupportRepeatSpeedups
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
#if (BX_DEBUGGER == 0)
|
|
|
|
/* If conditions are right, we can transfer IO to physical memory
|
|
|
|
* in a batch, rather than one instruction at a time.
|
|
|
|
*/
|
2004-11-21 02:26:32 +03:00
|
|
|
if (i->repUsedL() && !BX_CPU_THIS_PTR async_event)
|
|
|
|
{
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
Bit32u byteCount;
|
|
|
|
|
2002-09-18 09:36:48 +04:00
|
|
|
if (i->as32L())
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
byteCount = ECX;
|
|
|
|
else
|
|
|
|
byteCount = CX;
|
|
|
|
|
2005-07-04 21:44:08 +04:00
|
|
|
BX_ASSERT(byteCount);
|
|
|
|
byteCount = FastRepSTOSB(i, BX_SEG_REG_ES, edi, al, byteCount);
|
|
|
|
if (byteCount)
|
|
|
|
{
|
|
|
|
// Decrement the ticks count by the number of iterations, minus
|
|
|
|
// one, since the main cpu loop will decrement one. Also,
|
|
|
|
// the count is predecremented before examined, so defintely
|
|
|
|
// don't roll it under zero.
|
|
|
|
BX_TICKN(byteCount-1);
|
|
|
|
|
|
|
|
// Decrement eCX. Note, the main loop will decrement 1 also, so
|
|
|
|
// decrement by one less than expected, like the case above.
|
|
|
|
if (i->as32L())
|
|
|
|
ECX -= (byteCount-1);
|
|
|
|
else
|
|
|
|
CX -= (byteCount-1);
|
|
|
|
|
|
|
|
incr = byteCount;
|
|
|
|
goto doIncr16;
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-03 23:38:27 +04:00
|
|
|
#endif // (BX_DEBUGGER == 0)
|
|
|
|
#endif // BX_SupportRepeatSpeedups
|
|
|
|
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
write_virtual_byte(BX_SEG_REG_ES, edi, &al);
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-03 23:38:27 +04:00
|
|
|
#if BX_SupportRepeatSpeedups
|
|
|
|
#if (BX_DEBUGGER == 0)
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
doIncr16:
|
2002-09-03 23:38:27 +04:00
|
|
|
#endif
|
|
|
|
#endif
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
edi -= incr;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
edi += incr;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (i->as32L())
|
|
|
|
// zero extension of RDI
|
|
|
|
RDI = edi;
|
|
|
|
else
|
2005-07-04 21:44:08 +04:00
|
|
|
DI = edi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* 16 bit opsize mode */
|
|
|
|
void BX_CPU_C::STOSW_YwAX(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit16u ax = AX;
|
|
|
|
|
|
|
|
#if BX_SUPPORT_X86_64
|
|
|
|
if (i->as64L()) {
|
|
|
|
Bit64u rdi = RDI;
|
|
|
|
|
|
|
|
write_virtual_word(BX_SEG_REG_ES, rdi, &ax);
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rdi -= 2;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
rdi += 2;
|
|
|
|
}
|
|
|
|
|
|
|
|
RDI = rdi;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
#endif // #if BX_SUPPORT_X86_64
|
|
|
|
if (i->as32L())
|
|
|
|
{
|
|
|
|
Bit32u edi = EDI;
|
|
|
|
|
|
|
|
write_virtual_word(BX_SEG_REG_ES, edi, &ax);
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
edi -= 2;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
edi += 2;
|
|
|
|
}
|
|
|
|
|
|
|
|
// zero extension of RDI
|
|
|
|
RDI = edi;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{ /* 16bit address size */
|
|
|
|
Bit16u di = DI;
|
|
|
|
|
|
|
|
write_virtual_word(BX_SEG_REG_ES, di, &ax);
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
di -= 2;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
di += 2;
|
|
|
|
}
|
|
|
|
|
|
|
|
DI = di;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* 32 bit opsize mode */
|
|
|
|
void BX_CPU_C::STOSD_YdEAX(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit32u eax = EAX;
|
|
|
|
|
|
|
|
#if BX_SUPPORT_X86_64
|
|
|
|
if (i->as64L()) {
|
|
|
|
Bit64u rdi = RDI;
|
|
|
|
|
|
|
|
write_virtual_dword(BX_SEG_REG_ES, rdi, &eax);
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rdi -= 4;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
rdi += 4;
|
|
|
|
}
|
|
|
|
|
|
|
|
RDI = rdi;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
#endif // #if BX_SUPPORT_X86_64
|
|
|
|
if (i->as32L())
|
|
|
|
{
|
|
|
|
Bit32u edi = EDI;
|
|
|
|
|
|
|
|
write_virtual_dword(BX_SEG_REG_ES, edi, &eax);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
edi -= 4;
|
2002-09-15 09:09:18 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
else {
|
|
|
|
edi += 4;
|
2002-09-15 09:09:18 +04:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
// zero extension of RDI
|
|
|
|
RDI = edi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
Integrated patches for:
- Paging code rehash. You must now use --enable-4meg-pages to
use 4Meg pages, with the default of disabled, since we don't well
support 4Meg pages yet. Paging table walks model a real CPU
more closely now, and I fixed some bugs in the old logic.
- Segment check redundancy elimination. After a segment is loaded,
reads and writes are marked when a segment type check succeeds, and
they are skipped thereafter, when possible.
- Repeated IO and memory string copy acceleration. Only some variants
of instructions are available on all platforms, word and dword
variants only on x86 for the moment due to alignment and endian issues.
This is compiled in currently with no option - I should add a configure
option.
- Added a guest linear address to host TLB. Actually, I just stick
the host address (mem.vector[addr] address) in the upper 29 bits
of the field 'combined_access' since they are unused. Convenient
for now. I'm only storing page frame addresses. This was the
simplest for of such a TLB. We can likely enhance this. Also,
I only accelerated the normal read/write routines in access.cc.
Could also modify the read-modify-write versions too. You must
use --enable-guest2host-tlb, to try this out. Currently speeds
up Win95 boot time by about 3.5% for me. More ground to cover...
- Minor mods to CPUI/MOV_CdRd for CMOV.
- Integrated enhancements from Volker to getHostMemAddr() for PCI
being enabled.
2002-09-02 00:12:09 +04:00
|
|
|
else
|
2004-11-21 02:26:32 +03:00
|
|
|
{ /* 16bit address size */
|
|
|
|
Bit16u di = DI;
|
|
|
|
|
|
|
|
write_virtual_dword(BX_SEG_REG_ES, di, &eax);
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
di -= 4;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
di += 4;
|
|
|
|
}
|
|
|
|
|
|
|
|
DI = di;
|
2002-09-15 09:09:18 +04:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
}
|
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
#if BX_SUPPORT_X86_64
|
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
/* 64 bit opsize mode */
|
|
|
|
void BX_CPU_C::STOSQ_YqRAX(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
Bit64u rax = RAX;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (i->as64L()) {
|
|
|
|
Bit64u rdi = RDI;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
write_virtual_qword(BX_SEG_REG_ES, rdi, &rax);
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rdi -= 8;
|
2002-09-15 09:09:18 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
else {
|
|
|
|
rdi += 8;
|
|
|
|
}
|
2005-02-22 21:24:19 +03:00
|
|
|
|
|
|
|
RDI = rdi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
|
|
|
else /* 32 bit address size */
|
|
|
|
{
|
2004-08-15 00:44:48 +04:00
|
|
|
Bit32u edi = EDI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
write_virtual_qword(BX_SEG_REG_ES, edi, &rax);
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
edi -= 8;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
edi += 8;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
// zero extension of RDI
|
|
|
|
RDI = edi;
|
2004-08-15 00:44:48 +04:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
}
|
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
#endif
|
|
|
|
|
|
|
|
void BX_CPU_C::LODSB_ALXb(bxInstruction_c *i)
|
2001-04-10 05:04:59 +04:00
|
|
|
{
|
|
|
|
unsigned seg;
|
|
|
|
Bit8u al;
|
|
|
|
|
2002-09-18 09:36:48 +04:00
|
|
|
if (!BX_NULL_SEG_REG(i->seg())) {
|
|
|
|
seg = i->seg();
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else {
|
|
|
|
seg = BX_SEG_REG_DS;
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
#if BX_SUPPORT_X86_64
|
2002-09-18 09:36:48 +04:00
|
|
|
if (i->as64L()) {
|
2004-08-15 00:44:48 +04:00
|
|
|
Bit64u rsi = RSI;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
|
|
|
read_virtual_byte(seg, rsi, &al);
|
|
|
|
|
|
|
|
AL = al;
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rsi--;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
else {
|
|
|
|
rsi++;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
|
|
|
|
RSI = rsi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
else
|
|
|
|
#endif // #if BX_SUPPORT_X86_64
|
2004-11-21 02:26:32 +03:00
|
|
|
if (i->as32L())
|
|
|
|
{
|
2004-08-15 00:44:48 +04:00
|
|
|
Bit32u esi = ESI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
|
|
|
read_virtual_byte(seg, esi, &al);
|
|
|
|
|
|
|
|
AL = al;
|
2002-09-12 22:10:46 +04:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
2001-04-10 05:04:59 +04:00
|
|
|
esi--;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else {
|
|
|
|
esi++;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
// zero extension of RSI
|
|
|
|
RSI = esi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else
|
2004-11-21 02:26:32 +03:00
|
|
|
{ /* 16bit address mode */
|
2004-08-15 00:44:48 +04:00
|
|
|
Bit16u si = SI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
|
|
|
read_virtual_byte(seg, si, &al);
|
|
|
|
|
|
|
|
AL = al;
|
2002-09-12 22:10:46 +04:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
2001-04-10 05:04:59 +04:00
|
|
|
si--;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else {
|
|
|
|
si++;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
|
|
|
SI = si;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
}
|
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
/* 16 bit opsize mode */
|
|
|
|
void BX_CPU_C::LODSW_AXXw(bxInstruction_c *i)
|
2001-04-10 05:04:59 +04:00
|
|
|
{
|
|
|
|
unsigned seg;
|
2004-11-21 02:26:32 +03:00
|
|
|
Bit16u ax;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-18 09:36:48 +04:00
|
|
|
if (!BX_NULL_SEG_REG(i->seg())) {
|
|
|
|
seg = i->seg();
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else {
|
|
|
|
seg = BX_SEG_REG_DS;
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
#if BX_SUPPORT_X86_64
|
2002-09-18 09:36:48 +04:00
|
|
|
if (i->as64L()) {
|
2004-08-15 00:44:48 +04:00
|
|
|
Bit64u rsi = RSI;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_word(seg, rsi, &ax);
|
|
|
|
AX = ax;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rsi -= 2;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
rsi += 2;
|
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
|
|
|
|
RSI = rsi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2002-09-15 09:09:18 +04:00
|
|
|
else
|
|
|
|
#endif // #if BX_SUPPORT_X86_64
|
2004-11-21 02:26:32 +03:00
|
|
|
if (i->as32L())
|
|
|
|
{
|
2004-08-15 00:44:48 +04:00
|
|
|
Bit32u esi = ESI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_word(seg, esi, &ax);
|
|
|
|
AX = ax;
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
esi -= 2;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
esi += 2;
|
|
|
|
}
|
|
|
|
|
|
|
|
// zero extension of RSI
|
|
|
|
RSI = esi;
|
|
|
|
}
|
|
|
|
else
|
|
|
|
{ /* 16bit address mode */
|
|
|
|
Bit16u si = SI;
|
|
|
|
|
|
|
|
read_virtual_word(seg, si, &ax);
|
|
|
|
AX = ax;
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
si -= 2;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
si += 2;
|
|
|
|
}
|
|
|
|
|
|
|
|
SI = si;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* 32 bit opsize mode */
|
|
|
|
void BX_CPU_C::LODSD_EAXXd(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
unsigned seg;
|
|
|
|
Bit32u eax;
|
|
|
|
|
|
|
|
if (!BX_NULL_SEG_REG(i->seg())) {
|
|
|
|
seg = i->seg();
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
else {
|
|
|
|
seg = BX_SEG_REG_DS;
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
#if BX_SUPPORT_X86_64
|
2004-11-21 02:26:32 +03:00
|
|
|
if (i->as64L()) {
|
|
|
|
Bit64u rsi = RSI;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_dword(seg, rsi, &eax);
|
|
|
|
RAX = eax;
|
2002-09-15 09:09:18 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rsi -= 4;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
rsi += 4;
|
|
|
|
}
|
|
|
|
|
|
|
|
RSI = rsi;
|
|
|
|
}
|
|
|
|
else
|
2002-09-15 09:09:18 +04:00
|
|
|
#endif // #if BX_SUPPORT_X86_64
|
2004-11-21 02:26:32 +03:00
|
|
|
if (i->as32L())
|
|
|
|
{
|
|
|
|
Bit32u esi = ESI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_dword(seg, esi, &eax);
|
|
|
|
RAX = eax;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
esi -= 4;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
esi += 4;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2002-09-15 09:09:18 +04:00
|
|
|
// zero extension of RSI
|
|
|
|
RSI = esi;
|
2004-11-21 02:26:32 +03:00
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
else
|
2004-11-21 02:26:32 +03:00
|
|
|
{ /* 16bit address mode */
|
2004-08-15 00:44:48 +04:00
|
|
|
Bit16u si = SI;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
read_virtual_dword(seg, si, &eax);
|
|
|
|
RAX = eax;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
si -= 4;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
si += 4;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
SI = si;
|
|
|
|
}
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
#if BX_SUPPORT_X86_64
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
/* 64 bit opsize mode */
|
|
|
|
void BX_CPU_C::LODSQ_RAXXq(bxInstruction_c *i)
|
|
|
|
{
|
|
|
|
unsigned seg;
|
|
|
|
Bit64u rax;
|
2001-04-10 05:04:59 +04:00
|
|
|
|
2004-11-21 02:26:32 +03:00
|
|
|
if (!BX_NULL_SEG_REG(i->seg())) {
|
|
|
|
seg = i->seg();
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
else {
|
|
|
|
seg = BX_SEG_REG_DS;
|
2005-06-21 21:01:21 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
|
|
|
|
if (i->as64L()) {
|
|
|
|
Bit64u rsi = RSI;
|
|
|
|
|
|
|
|
read_virtual_qword(seg, rsi, &rax);
|
|
|
|
RAX = rax;
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
rsi -= 8;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
rsi += 8;
|
|
|
|
}
|
|
|
|
|
|
|
|
RSI = rsi;
|
|
|
|
}
|
|
|
|
else /* 32 bit address size */
|
|
|
|
{
|
|
|
|
Bit32u esi = ESI;
|
|
|
|
|
|
|
|
read_virtual_qword(seg, esi, &rax);
|
|
|
|
RAX = rax;
|
|
|
|
|
|
|
|
if (BX_CPU_THIS_PTR get_DF ()) {
|
|
|
|
esi -= 8;
|
|
|
|
}
|
|
|
|
else {
|
|
|
|
esi += 8;
|
2001-04-10 05:04:59 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
|
|
|
|
// zero extension of RSI
|
|
|
|
RSI = esi;
|
|
|
|
}
|
2001-04-10 05:04:59 +04:00
|
|
|
}
|
2004-11-21 02:26:32 +03:00
|
|
|
|
|
|
|
#endif
|