///////////////////////////////////////////////////////////////////////// // $Id: ctrl_xfer32.cc,v 1.41 2005-05-20 20:06:50 sshwarts Exp $ ///////////////////////////////////////////////////////////////////////// // // Copyright (C) 2001 MandrakeSoft S.A. // // MandrakeSoft S.A. // 43, rue d'Aboukir // 75002 Paris - France // http://www.linux-mandrake.com/ // http://www.mandrakesoft.com/ // // This library is free software; you can redistribute it and/or // modify it under the terms of the GNU Lesser General Public // License as published by the Free Software Foundation; either // version 2 of the License, or (at your option) any later version. // // This library is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU // Lesser General Public License for more details. // // You should have received a copy of the GNU Lesser General Public // License along with this library; if not, write to the Free Software // Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA #define NEED_CPU_REG_SHORTCUTS 1 #include "bochs.h" #define LOG_THIS BX_CPU_THIS_PTR #if BX_CPU_LEVEL >= 3 void BX_CPU_C::RETnear32_Iw(bxInstruction_c *i) { Bit32u return_EIP; #if BX_DEBUGGER BX_CPU_THIS_PTR show_flag |= Flag_ret; #endif Bit16u imm16 = i->Iw(); pop_32(&return_EIP); branch_near32(return_EIP); // includes revalidate_prefetch_q() if (BX_CPU_THIS_PTR sregs[BX_SEG_REG_SS].cache.u.segment.d_b) ESP += imm16; else SP += imm16; BX_INSTR_UCNEAR_BRANCH(BX_CPU_ID, BX_INSTR_IS_RET, EIP); } void BX_CPU_C::RETnear32(bxInstruction_c *i) { Bit32u return_EIP; #if BX_DEBUGGER BX_CPU_THIS_PTR show_flag |= Flag_ret; #endif pop_32(&return_EIP); branch_near32(return_EIP); // includes revalidate_prefetch_q() BX_INSTR_UCNEAR_BRANCH(BX_CPU_ID, BX_INSTR_IS_RET, EIP); } void BX_CPU_C::RETfar32_Iw(bxInstruction_c *i) { Bit32u eip, ecs_raw; invalidate_prefetch_q(); #if BX_DEBUGGER BX_CPU_THIS_PTR show_flag |= Flag_ret; #endif Bit16u imm16 = i->Iw(); if (protected_mode()) { BX_CPU_THIS_PTR return_protected(i, imm16); goto done; } pop_32(&eip); pop_32(&ecs_raw); EIP = eip; load_seg_reg(&BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS], (Bit16u) ecs_raw); if (BX_CPU_THIS_PTR sregs[BX_SEG_REG_SS].cache.u.segment.d_b) ESP += imm16; else SP += imm16; done: BX_INSTR_FAR_BRANCH(BX_CPU_ID, BX_INSTR_IS_RET, BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].selector.value, EIP); } void BX_CPU_C::RETfar32(bxInstruction_c *i) { Bit32u eip, ecs_raw; invalidate_prefetch_q(); #if BX_DEBUGGER BX_CPU_THIS_PTR show_flag |= Flag_ret; #endif if ( protected_mode() ) { BX_CPU_THIS_PTR return_protected(i, 0); goto done; } pop_32(&eip); pop_32(&ecs_raw); /* 32bit pop, MSW discarded */ EIP = eip; load_seg_reg(&BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS], (Bit16u) ecs_raw); done: BX_INSTR_FAR_BRANCH(BX_CPU_ID, BX_INSTR_IS_RET, BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].selector.value, EIP); } void BX_CPU_C::CALL_Ad(bxInstruction_c *i) { BailBigRSP("CALL_Ad"); #if BX_DEBUGGER BX_CPU_THIS_PTR show_flag |= Flag_call; #endif Bit32u new_EIP = EIP + i->Id(); if ( new_EIP > BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].cache.u.segment.limit_scaled ) { BX_ERROR(("CALL_Ad: offset outside of CS limits")); exception(BX_GP_EXCEPTION, 0, 0); } /* push 32 bit EA of next instruction */ push_32(EIP); EIP = new_EIP; BX_INSTR_UCNEAR_BRANCH(BX_CPU_ID, BX_INSTR_IS_CALL, EIP); } void BX_CPU_C::CALL32_Ap(bxInstruction_c *i) { Bit16u cs_raw; Bit32u disp32; invalidate_prefetch_q(); #if BX_DEBUGGER BX_CPU_THIS_PTR show_flag |= Flag_call; #endif disp32 = i->Id(); cs_raw = i->Iw2(); if (protected_mode()) { BX_CPU_THIS_PTR call_protected(i, cs_raw, disp32); goto done; } push_32(BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].selector.value); push_32(EIP); EIP = disp32; load_seg_reg(&BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS], cs_raw); done: BX_INSTR_FAR_BRANCH(BX_CPU_ID, BX_INSTR_IS_CALL, BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].selector.value, EIP); } void BX_CPU_C::CALL_Ed(bxInstruction_c *i) { Bit32u op1_32; #if BX_DEBUGGER BX_CPU_THIS_PTR show_flag |= Flag_call; #endif if (i->modC0()) { op1_32 = BX_READ_32BIT_REG(i->rm()); } else { read_virtual_dword(i->seg(), RMAddr(i), &op1_32); } if (op1_32 > BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].cache.u.segment.limit_scaled) { BX_ERROR(("CALL_Ed: EIP out of CS limits!")); exception(BX_GP_EXCEPTION, 0, 0); } push_32(EIP); EIP = op1_32; BX_INSTR_UCNEAR_BRANCH(BX_CPU_ID, BX_INSTR_IS_CALL, EIP); } void BX_CPU_C::CALL32_Ep(bxInstruction_c *i) { Bit16u cs_raw; Bit32u op1_32; invalidate_prefetch_q(); #if BX_DEBUGGER BX_CPU_THIS_PTR show_flag |= Flag_call; #endif /* op1_32 is a register or memory reference */ if (i->modC0()) { BX_INFO(("CALL_Ep: op1 is a register")); exception(BX_UD_EXCEPTION, 0, 0); } /* pointer, segment address pair */ read_virtual_dword(i->seg(), RMAddr(i), &op1_32); read_virtual_word(i->seg(), RMAddr(i)+4, &cs_raw); if ( protected_mode() ) { BX_CPU_THIS_PTR call_protected(i, cs_raw, op1_32); goto done; } push_32(BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].selector.value); push_32(EIP); EIP = op1_32; load_seg_reg(&BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS], cs_raw); done: BX_INSTR_FAR_BRANCH(BX_CPU_ID, BX_INSTR_IS_CALL, BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].selector.value, EIP); } void BX_CPU_C::JMP_Jd(bxInstruction_c *i) { Bit32u new_EIP = EIP + (Bit32s) i->Id(); branch_near32(new_EIP); // includes revalidate_prefetch_q() BX_INSTR_UCNEAR_BRANCH(BX_CPU_ID, BX_INSTR_IS_JMP, new_EIP); } void BX_CPU_C::JCC_Jd(bxInstruction_c *i) { bx_bool condition; switch (i->b1() & 0x0f) { case 0x00: /* JO */ condition = get_OF(); break; case 0x01: /* JNO */ condition = !get_OF(); break; case 0x02: /* JB */ condition = get_CF(); break; case 0x03: /* JNB */ condition = !get_CF(); break; case 0x04: /* JZ */ condition = get_ZF(); break; case 0x05: /* JNZ */ condition = !get_ZF(); break; case 0x06: /* JBE */ condition = get_CF() || get_ZF(); break; case 0x07: /* JNBE */ condition = !get_CF() && !get_ZF(); break; case 0x08: /* JS */ condition = get_SF(); break; case 0x09: /* JNS */ condition = !get_SF(); break; case 0x0A: /* JP */ condition = get_PF(); break; case 0x0B: /* JNP */ condition = !get_PF(); break; case 0x0C: /* JL */ condition = getB_SF() != getB_OF(); break; case 0x0D: /* JNL */ condition = getB_SF() == getB_OF(); break; case 0x0E: /* JLE */ condition = get_ZF() || (getB_SF() != getB_OF()); break; case 0x0F: /* JNLE */ condition = (getB_SF() == getB_OF()) && !get_ZF(); break; default: condition = 0; // For compiler...all targets should set condition. break; } if (condition) { Bit32u new_EIP = EIP + (Bit32s) i->Id(); branch_near32(new_EIP); // includes revalidate_prefetch_q() BX_INSTR_CNEAR_BRANCH_TAKEN(BX_CPU_ID, new_EIP); } #if BX_INSTRUMENTATION else { BX_INSTR_CNEAR_BRANCH_NOT_TAKEN(BX_CPU_ID); } #endif } void BX_CPU_C::JZ_Jd(bxInstruction_c *i) { if (get_ZF()) { Bit32u new_EIP = EIP + (Bit32s) i->Id(); branch_near32(new_EIP); // includes revalidate_prefetch_q() BX_INSTR_CNEAR_BRANCH_TAKEN(BX_CPU_ID, new_EIP); } #if BX_INSTRUMENTATION else { BX_INSTR_CNEAR_BRANCH_NOT_TAKEN(BX_CPU_ID); } #endif } void BX_CPU_C::JNZ_Jd(bxInstruction_c *i) { if (!get_ZF()) { Bit32u new_EIP = EIP + (Bit32s) i->Id(); branch_near32(new_EIP); // includes revalidate_prefetch_q() BX_INSTR_CNEAR_BRANCH_TAKEN(BX_CPU_ID, new_EIP); } #if BX_INSTRUMENTATION else { BX_INSTR_CNEAR_BRANCH_NOT_TAKEN(BX_CPU_ID); } #endif } void BX_CPU_C::JMP_Ap(bxInstruction_c *i) { Bit32u disp32; Bit16u cs_raw; invalidate_prefetch_q(); if (i->os32L()) { disp32 = i->Id(); } else { disp32 = i->Iw(); } cs_raw = i->Iw2(); if (protected_mode()) { BX_CPU_THIS_PTR jump_protected(i, cs_raw, disp32); goto done; } load_seg_reg(&BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS], cs_raw); EIP = disp32; done: BX_INSTR_FAR_BRANCH(BX_CPU_ID, BX_INSTR_IS_JMP, BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].selector.value, EIP); } void BX_CPU_C::JMP_Ed(bxInstruction_c *i) { Bit32u new_EIP; /* op1_32 is a register or memory reference */ if (i->modC0()) { new_EIP = BX_READ_32BIT_REG(i->rm()); } else { /* pointer, segment address pair */ read_virtual_dword(i->seg(), RMAddr(i), &new_EIP); } branch_near32(new_EIP); // includes revalidate_prefetch_q() BX_INSTR_UCNEAR_BRANCH(BX_CPU_ID, BX_INSTR_IS_JMP, new_EIP); } /* Far indirect jump */ void BX_CPU_C::JMP32_Ep(bxInstruction_c *i) { Bit16u cs_raw; Bit32u op1_32; invalidate_prefetch_q(); /* op1_32 is a register or memory reference */ if (i->modC0()) { /* far indirect must specify a memory address */ BX_INFO(("JMP_Ep(): op1 is a register")); exception(BX_UD_EXCEPTION, 0, 0); } /* pointer, segment address pair */ read_virtual_dword(i->seg(), RMAddr(i), &op1_32); read_virtual_word(i->seg(), RMAddr(i)+4, &cs_raw); if ( protected_mode() ) { BX_CPU_THIS_PTR jump_protected(i, cs_raw, op1_32); goto done; } EIP = op1_32; load_seg_reg(&BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS], cs_raw); done: BX_INSTR_FAR_BRANCH(BX_CPU_ID, BX_INSTR_IS_JMP, BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].selector.value, EIP); } void BX_CPU_C::IRET32(bxInstruction_c *i) { invalidate_prefetch_q(); #if BX_DEBUGGER BX_CPU_THIS_PTR show_flag |= Flag_iret; BX_CPU_THIS_PTR show_eip = EIP; #endif if (v8086_mode()) { // IOPL check in stack_return_from_v86() stack_return_from_v86(i); goto done; } if (protected_mode()) { iret_protected(i); goto done; } Bit32u eip, ecs, eflags; if (! can_pop(12)) { BX_PANIC(("IRETD: to 12 bytes of stack not within stack limits")); exception(BX_SS_EXCEPTION, 0, 0); } access_linear(BX_CPU_THIS_PTR get_segment_base(BX_SEG_REG_SS) + ESP, 4, CPL == 3, BX_READ, &eip); // still need to be validated ! if (eip > 0xffff) { BX_PANIC(("IRETD: instruction pointer not within code segment limits")); exception(BX_GP_EXCEPTION, 0, 0); } pop_32(&eip); pop_32(&ecs); pop_32(&eflags); ecs &= 0xffff; eflags = (eflags & 0x257fd5) | (read_eflags() & 0x1a0000); load_seg_reg(&BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS], (Bit16u)ecs); EIP = eip; writeEFlags(eflags, 0xffffffff); done: BX_INSTR_FAR_BRANCH(BX_CPU_ID, BX_INSTR_IS_IRET, BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].selector.value, EIP); } #endif