1110 lines
35 KiB
C++
1110 lines
35 KiB
C++
/////////////////////////////////////////////////////////////////////////
|
||
// $Id$
|
||
/////////////////////////////////////////////////////////////////////////
|
||
//
|
||
// Copyright (C) 2001-2012 The Bochs Project
|
||
//
|
||
// This library is free software; you can redistribute it and/or
|
||
// modify it under the terms of the GNU Lesser General Public
|
||
// License as published by the Free Software Foundation; either
|
||
// version 2 of the License, or (at your option) any later version.
|
||
//
|
||
// This library is distributed in the hope that it will be useful,
|
||
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||
// Lesser General Public License for more details.
|
||
//
|
||
// You should have received a copy of the GNU Lesser General Public
|
||
// License along with this library; if not, write to the Free Software
|
||
// Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
|
||
//
|
||
/////////////////////////////////////////////////////////////////////////
|
||
|
||
#define NEED_CPU_REG_SHORTCUTS 1
|
||
#include "bochs.h"
|
||
#include "cpu.h"
|
||
#define LOG_THIS BX_CPU_THIS_PTR
|
||
|
||
#include "param_names.h"
|
||
|
||
#include <stdlib.h>
|
||
|
||
BX_CPU_C::BX_CPU_C(unsigned id): bx_cpuid(id)
|
||
#if BX_CPU_LEVEL >= 4
|
||
, cpuid(NULL)
|
||
#endif
|
||
#if BX_SUPPORT_APIC
|
||
,lapic (this, id)
|
||
#endif
|
||
{
|
||
// in case of SMF, you cannot reference any member data
|
||
// in the constructor because the only access to it is via
|
||
// global variables which aren't initialized quite yet.
|
||
char name[16], logname[16];
|
||
sprintf(name, "CPU%x", bx_cpuid);
|
||
sprintf(logname, "cpu%x", bx_cpuid);
|
||
put(logname, name);
|
||
|
||
isa_extensions_bitmask = BX_SUPPORT_FPU ? BX_ISA_X87 : 0;
|
||
cpu_extensions_bitmask = 0;
|
||
#if BX_SUPPORT_VMX
|
||
vmx_extensions_bitmask = 0;
|
||
#endif
|
||
#if BX_SUPPORT_SVM
|
||
svm_extensions_bitmask = 0;
|
||
#endif
|
||
|
||
srand(time(NULL)); // initialize random generator for RDRAND/RDSEED
|
||
}
|
||
|
||
#if BX_CPU_LEVEL >= 4
|
||
|
||
#include "generic_cpuid.h"
|
||
|
||
#define bx_define_cpudb(model) \
|
||
extern bx_cpuid_t *create_ ##model##_cpuid(BX_CPU_C *cpu);
|
||
|
||
#include "cpudb.h"
|
||
|
||
#undef bx_define_cpudb
|
||
|
||
static bx_cpuid_t *cpuid_factory(BX_CPU_C *cpu)
|
||
{
|
||
unsigned cpu_model = SIM->get_param_enum(BXPN_CPU_MODEL)->get();
|
||
|
||
#define bx_define_cpudb(model) \
|
||
case bx_cpudb_##model: \
|
||
return create_ ##model##_cpuid(cpu);
|
||
|
||
switch(cpu_model) {
|
||
#include "cpudb.h"
|
||
default:
|
||
return 0;
|
||
}
|
||
#undef bx_define_cpudb
|
||
}
|
||
|
||
#endif
|
||
|
||
// BX_CPU_C constructor
|
||
void BX_CPU_C::initialize(void)
|
||
{
|
||
#if BX_CPU_LEVEL >= 4
|
||
BX_CPU_THIS_PTR cpuid = cpuid_factory(this);
|
||
if (! BX_CPU_THIS_PTR cpuid)
|
||
BX_PANIC(("Failed to create CPUID module !"));
|
||
|
||
BX_CPU_THIS_PTR isa_extensions_bitmask = cpuid->get_isa_extensions_bitmask();
|
||
BX_CPU_THIS_PTR cpu_extensions_bitmask = cpuid->get_cpu_extensions_bitmask();
|
||
#if BX_SUPPORT_VMX
|
||
BX_CPU_THIS_PTR vmx_extensions_bitmask = cpuid->get_vmx_extensions_bitmask();
|
||
#endif
|
||
#if BX_SUPPORT_SVM
|
||
BX_CPU_THIS_PTR svm_extensions_bitmask = cpuid->get_svm_extensions_bitmask();
|
||
#endif
|
||
#endif
|
||
|
||
init_FetchDecodeTables(); // must be called after init_isa_features_bitmask()
|
||
|
||
#if BX_CONFIGURE_MSRS
|
||
for (unsigned n=0; n < BX_MSR_MAX_INDEX; n++) {
|
||
BX_CPU_THIS_PTR msrs[n] = 0;
|
||
}
|
||
const char *msrs_filename = SIM->get_param_string(BXPN_CONFIGURABLE_MSRS_PATH)->getptr();
|
||
load_MSRs(msrs_filename);
|
||
#endif
|
||
|
||
// ignore bad MSRS if user asked for it
|
||
#if BX_CPU_LEVEL >= 5
|
||
BX_CPU_THIS_PTR ignore_bad_msrs = SIM->get_param_bool(BXPN_IGNORE_BAD_MSRS)->get();
|
||
#endif
|
||
|
||
init_SMRAM();
|
||
|
||
#if BX_SUPPORT_VMX
|
||
init_VMCS();
|
||
#endif
|
||
}
|
||
|
||
// save/restore functionality
|
||
void BX_CPU_C::register_state(void)
|
||
{
|
||
unsigned n;
|
||
char name[10];
|
||
|
||
sprintf(name, "cpu%d", BX_CPU_ID);
|
||
|
||
bx_list_c *cpu = new bx_list_c(SIM->get_bochs_root(), name, name);
|
||
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, isa_extensions_bitmask);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, cpu_extensions_bitmask);
|
||
#if BX_SUPPORT_VMX
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, vmx_extensions_bitmask);
|
||
#endif
|
||
#if BX_SUPPORT_SVM
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, svm_extensions_bitmask);
|
||
#endif
|
||
BXRS_DEC_PARAM_SIMPLE(cpu, cpu_mode);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, activity_state);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, inhibit_mask);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, inhibit_icount);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, debug_trap);
|
||
BXRS_DEC_PARAM_SIMPLE(cpu, icount);
|
||
BXRS_DEC_PARAM_SIMPLE(cpu, icount_last_sync);
|
||
#if BX_SUPPORT_X86_64
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, RAX);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, RBX);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, RCX);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, RDX);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, RSP);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, RBP);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, RSI);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, RDI);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, R8);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, R9);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, R10);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, R11);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, R12);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, R13);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, R14);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, R15);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, RIP);
|
||
#else
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, EAX);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, EBX);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, ECX);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, EDX);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, ESP);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, EBP);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, ESI);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, EDI);
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, EIP);
|
||
#endif
|
||
BXRS_PARAM_SPECIAL32(cpu, EFLAGS,
|
||
param_save_handler, param_restore_handler);
|
||
#if BX_CPU_LEVEL >= 3
|
||
BXRS_HEX_PARAM_FIELD(cpu, DR0, dr[0]);
|
||
BXRS_HEX_PARAM_FIELD(cpu, DR1, dr[1]);
|
||
BXRS_HEX_PARAM_FIELD(cpu, DR2, dr[2]);
|
||
BXRS_HEX_PARAM_FIELD(cpu, DR3, dr[3]);
|
||
BXRS_HEX_PARAM_FIELD(cpu, DR6, dr6.val32);
|
||
BXRS_HEX_PARAM_FIELD(cpu, DR7, dr7.val32);
|
||
#endif
|
||
BXRS_HEX_PARAM_FIELD(cpu, CR0, cr0.val32);
|
||
BXRS_HEX_PARAM_FIELD(cpu, CR2, cr2);
|
||
BXRS_HEX_PARAM_FIELD(cpu, CR3, cr3);
|
||
#if BX_CPU_LEVEL >= 5
|
||
BXRS_HEX_PARAM_FIELD(cpu, CR4, cr4.val32);
|
||
#endif
|
||
#if BX_CPU_LEVEL >= 6
|
||
if (BX_CPUID_SUPPORT_ISA_EXTENSION(BX_ISA_XSAVE)) {
|
||
BXRS_HEX_PARAM_FIELD(cpu, XCR0, xcr0.val32);
|
||
}
|
||
#endif
|
||
#if BX_CPU_LEVEL >= 5
|
||
BXRS_HEX_PARAM_FIELD(cpu, tsc_last_reset, tsc_last_reset);
|
||
#if BX_SUPPORT_VMX || BX_SUPPORT_SVM
|
||
BXRS_HEX_PARAM_FIELD(cpu, tsc_offset, tsc_offset);
|
||
#endif
|
||
#endif
|
||
for(n=0; n<6; n++) {
|
||
bx_segment_reg_t *segment = &BX_CPU_THIS_PTR sregs[n];
|
||
bx_list_c *sreg = new bx_list_c(cpu, strseg(segment));
|
||
BXRS_PARAM_SPECIAL16(sreg, selector,
|
||
param_save_handler, param_restore_handler);
|
||
BXRS_HEX_PARAM_FIELD(sreg, valid, segment->cache.valid);
|
||
BXRS_PARAM_BOOL(sreg, p, segment->cache.p);
|
||
BXRS_HEX_PARAM_FIELD(sreg, dpl, segment->cache.dpl);
|
||
BXRS_PARAM_BOOL(sreg, segment, segment->cache.segment);
|
||
BXRS_HEX_PARAM_FIELD(sreg, type, segment->cache.type);
|
||
BXRS_HEX_PARAM_FIELD(sreg, base, segment->cache.u.segment.base);
|
||
BXRS_HEX_PARAM_FIELD(sreg, limit_scaled, segment->cache.u.segment.limit_scaled);
|
||
BXRS_PARAM_BOOL(sreg, granularity, segment->cache.u.segment.g);
|
||
BXRS_PARAM_BOOL(sreg, d_b, segment->cache.u.segment.d_b);
|
||
#if BX_SUPPORT_X86_64
|
||
BXRS_PARAM_BOOL(sreg, l, segment->cache.u.segment.l);
|
||
#endif
|
||
BXRS_PARAM_BOOL(sreg, avl, segment->cache.u.segment.avl);
|
||
}
|
||
|
||
bx_list_c *GDTR = new bx_list_c(cpu, "GDTR");
|
||
BXRS_HEX_PARAM_FIELD(GDTR, base, gdtr.base);
|
||
BXRS_HEX_PARAM_FIELD(GDTR, limit, gdtr.limit);
|
||
|
||
bx_list_c *IDTR = new bx_list_c(cpu, "IDTR");
|
||
BXRS_HEX_PARAM_FIELD(IDTR, base, idtr.base);
|
||
BXRS_HEX_PARAM_FIELD(IDTR, limit, idtr.limit);
|
||
|
||
bx_list_c *LDTR = new bx_list_c(cpu, "LDTR");
|
||
BXRS_PARAM_SPECIAL16(LDTR, selector, param_save_handler, param_restore_handler);
|
||
BXRS_HEX_PARAM_FIELD(LDTR, valid, ldtr.cache.valid);
|
||
BXRS_PARAM_BOOL(LDTR, p, ldtr.cache.p);
|
||
BXRS_HEX_PARAM_FIELD(LDTR, dpl, ldtr.cache.dpl);
|
||
BXRS_PARAM_BOOL(LDTR, segment, ldtr.cache.segment);
|
||
BXRS_HEX_PARAM_FIELD(LDTR, type, ldtr.cache.type);
|
||
BXRS_HEX_PARAM_FIELD(LDTR, base, ldtr.cache.u.segment.base);
|
||
BXRS_HEX_PARAM_FIELD(LDTR, limit_scaled, ldtr.cache.u.segment.limit_scaled);
|
||
BXRS_PARAM_BOOL(LDTR, granularity, ldtr.cache.u.segment.g);
|
||
BXRS_PARAM_BOOL(LDTR, d_b, ldtr.cache.u.segment.d_b);
|
||
BXRS_PARAM_BOOL(LDTR, avl, ldtr.cache.u.segment.avl);
|
||
|
||
bx_list_c *TR = new bx_list_c(cpu, "TR");
|
||
BXRS_PARAM_SPECIAL16(TR, selector, param_save_handler, param_restore_handler);
|
||
BXRS_HEX_PARAM_FIELD(TR, valid, tr.cache.valid);
|
||
BXRS_PARAM_BOOL(TR, p, tr.cache.p);
|
||
BXRS_HEX_PARAM_FIELD(TR, dpl, tr.cache.dpl);
|
||
BXRS_PARAM_BOOL(TR, segment, tr.cache.segment);
|
||
BXRS_HEX_PARAM_FIELD(TR, type, tr.cache.type);
|
||
BXRS_HEX_PARAM_FIELD(TR, base, tr.cache.u.segment.base);
|
||
BXRS_HEX_PARAM_FIELD(TR, limit_scaled, tr.cache.u.segment.limit_scaled);
|
||
BXRS_PARAM_BOOL(TR, granularity, tr.cache.u.segment.g);
|
||
BXRS_PARAM_BOOL(TR, d_b, tr.cache.u.segment.d_b);
|
||
BXRS_PARAM_BOOL(TR, avl, tr.cache.u.segment.avl);
|
||
|
||
BXRS_HEX_PARAM_SIMPLE(cpu, smbase);
|
||
|
||
#if BX_CPU_LEVEL >= 6
|
||
bx_list_c *PDPTRS = new bx_list_c(cpu, "PDPTR_CACHE");
|
||
BXRS_HEX_PARAM_FIELD(PDPTRS, entry0, PDPTR_CACHE.entry[0]);
|
||
BXRS_HEX_PARAM_FIELD(PDPTRS, entry1, PDPTR_CACHE.entry[1]);
|
||
BXRS_HEX_PARAM_FIELD(PDPTRS, entry2, PDPTR_CACHE.entry[2]);
|
||
BXRS_HEX_PARAM_FIELD(PDPTRS, entry3, PDPTR_CACHE.entry[3]);
|
||
#endif
|
||
|
||
#if BX_CPU_LEVEL >= 5
|
||
bx_list_c *MSR = new bx_list_c(cpu, "MSR");
|
||
|
||
#if BX_SUPPORT_APIC
|
||
BXRS_HEX_PARAM_FIELD(MSR, apicbase, msr.apicbase);
|
||
#endif
|
||
BXRS_HEX_PARAM_FIELD(MSR, EFER, efer.val32);
|
||
BXRS_HEX_PARAM_FIELD(MSR, star, msr.star);
|
||
#if BX_SUPPORT_X86_64
|
||
if (BX_CPUID_SUPPORT_CPU_EXTENSION(BX_CPU_LONG_MODE)) {
|
||
BXRS_HEX_PARAM_FIELD(MSR, lstar, msr.lstar);
|
||
BXRS_HEX_PARAM_FIELD(MSR, cstar, msr.cstar);
|
||
BXRS_HEX_PARAM_FIELD(MSR, fmask, msr.fmask);
|
||
BXRS_HEX_PARAM_FIELD(MSR, kernelgsbase, msr.kernelgsbase);
|
||
BXRS_HEX_PARAM_FIELD(MSR, tsc_aux, msr.tsc_aux);
|
||
}
|
||
#endif
|
||
#if BX_CPU_LEVEL >= 6
|
||
BXRS_HEX_PARAM_FIELD(MSR, sysenter_cs_msr, msr.sysenter_cs_msr);
|
||
BXRS_HEX_PARAM_FIELD(MSR, sysenter_esp_msr, msr.sysenter_esp_msr);
|
||
BXRS_HEX_PARAM_FIELD(MSR, sysenter_eip_msr, msr.sysenter_eip_msr);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrphysbase0, msr.mtrrphys[0]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrphysmask0, msr.mtrrphys[1]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrphysbase1, msr.mtrrphys[2]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrphysmask1, msr.mtrrphys[3]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrphysbase2, msr.mtrrphys[4]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrphysmask2, msr.mtrrphys[5]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrphysbase3, msr.mtrrphys[6]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrphysmask3, msr.mtrrphys[7]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrphysbase4, msr.mtrrphys[8]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrphysmask4, msr.mtrrphys[9]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrphysbase5, msr.mtrrphys[10]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrphysmask5, msr.mtrrphys[11]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrphysbase6, msr.mtrrphys[12]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrphysmask6, msr.mtrrphys[13]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrphysbase7, msr.mtrrphys[14]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrphysmask7, msr.mtrrphys[15]);
|
||
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrfix64k_00000, msr.mtrrfix64k_00000);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrfix16k_80000, msr.mtrrfix16k[0]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrfix16k_a0000, msr.mtrrfix16k[1]);
|
||
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrfix4k_c0000, msr.mtrrfix4k[0]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrfix4k_c8000, msr.mtrrfix4k[1]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrfix4k_d0000, msr.mtrrfix4k[2]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrfix4k_d8000, msr.mtrrfix4k[3]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrfix4k_e0000, msr.mtrrfix4k[4]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrfix4k_e8000, msr.mtrrfix4k[5]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrfix4k_f0000, msr.mtrrfix4k[6]);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrrfix4k_f8000, msr.mtrrfix4k[7]);
|
||
|
||
BXRS_HEX_PARAM_FIELD(MSR, pat, msr.pat);
|
||
BXRS_HEX_PARAM_FIELD(MSR, mtrr_deftype, msr.mtrr_deftype);
|
||
#endif
|
||
#if BX_CONFIGURE_MSRS
|
||
bx_list_c *MSRS = new bx_list_c(cpu, "USER_MSR");
|
||
for(n=0; n < BX_MSR_MAX_INDEX; n++) {
|
||
if (! msrs[n]) continue;
|
||
sprintf(name, "msr_0x%03x", n);
|
||
bx_list_c *m = new bx_list_c(MSRS, name);
|
||
BXRS_HEX_PARAM_FIELD(m, index, msrs[n]->index);
|
||
BXRS_DEC_PARAM_FIELD(m, type, msrs[n]->type);
|
||
BXRS_HEX_PARAM_FIELD(m, val64, msrs[n]->val64);
|
||
BXRS_HEX_PARAM_FIELD(m, reset, msrs[n]->reset_value);
|
||
BXRS_HEX_PARAM_FIELD(m, reserved, msrs[n]->reserved);
|
||
BXRS_HEX_PARAM_FIELD(m, ignored, msrs[n]->ignored);
|
||
}
|
||
#endif
|
||
#endif
|
||
|
||
#if BX_SUPPORT_FPU
|
||
bx_list_c *fpu = new bx_list_c(cpu, "FPU");
|
||
BXRS_HEX_PARAM_FIELD(fpu, cwd, the_i387.cwd);
|
||
BXRS_HEX_PARAM_FIELD(fpu, swd, the_i387.swd);
|
||
BXRS_HEX_PARAM_FIELD(fpu, twd, the_i387.twd);
|
||
BXRS_HEX_PARAM_FIELD(fpu, foo, the_i387.foo);
|
||
BXRS_HEX_PARAM_FIELD(fpu, fcs, the_i387.fcs);
|
||
BXRS_HEX_PARAM_FIELD(fpu, fip, the_i387.fip);
|
||
BXRS_HEX_PARAM_FIELD(fpu, fds, the_i387.fds);
|
||
BXRS_HEX_PARAM_FIELD(fpu, fdp, the_i387.fdp);
|
||
for (n=0; n<8; n++) {
|
||
sprintf(name, "st%d", n);
|
||
bx_list_c *STx = new bx_list_c(fpu, name);
|
||
BXRS_HEX_PARAM_FIELD(STx, exp, the_i387.st_space[n].exp);
|
||
BXRS_HEX_PARAM_FIELD(STx, fraction, the_i387.st_space[n].fraction);
|
||
}
|
||
BXRS_DEC_PARAM_FIELD(fpu, tos, the_i387.tos);
|
||
#endif
|
||
|
||
#if BX_CPU_LEVEL >= 6
|
||
if (BX_CPUID_SUPPORT_ISA_EXTENSION(BX_ISA_SSE)) {
|
||
bx_list_c *sse = new bx_list_c(cpu, "SSE");
|
||
BXRS_HEX_PARAM_FIELD(sse, mxcsr, mxcsr.mxcsr);
|
||
for (n=0; n<BX_XMM_REGISTERS; n++) {
|
||
for(unsigned j=0;j < BX_VLMAX*2;j++) {
|
||
sprintf(name, "xmm%02d_%d", n, j);
|
||
#if BX_SUPPORT_AVX
|
||
new bx_shadow_num_c(sse, name, &vmm[n].avx64u(j), BASE_HEX);
|
||
#else
|
||
new bx_shadow_num_c(sse, name, &vmm[n].xmm64u(j), BASE_HEX);
|
||
#endif
|
||
}
|
||
}
|
||
}
|
||
#endif
|
||
|
||
#if BX_SUPPORT_MONITOR_MWAIT
|
||
bx_list_c *monitor_list = new bx_list_c(cpu, "MONITOR");
|
||
BXRS_HEX_PARAM_FIELD(monitor_list, monitor_addr, monitor.monitor_addr);
|
||
BXRS_PARAM_BOOL(monitor_list, armed, monitor.armed);
|
||
#endif
|
||
|
||
#if BX_SUPPORT_APIC
|
||
lapic.register_state(cpu);
|
||
#endif
|
||
|
||
#if BX_SUPPORT_VMX
|
||
register_vmx_state(cpu);
|
||
#endif
|
||
|
||
#if BX_SUPPORT_SVM
|
||
register_svm_state(cpu);
|
||
#endif
|
||
|
||
BXRS_HEX_PARAM_SIMPLE32(cpu, pending_event);
|
||
BXRS_HEX_PARAM_SIMPLE32(cpu, event_mask);
|
||
BXRS_HEX_PARAM_SIMPLE32(cpu, async_event);
|
||
|
||
#if BX_X86_DEBUGGER
|
||
BXRS_PARAM_BOOL(cpu, in_repeat, in_repeat);
|
||
#endif
|
||
|
||
BXRS_PARAM_BOOL(cpu, in_smm, in_smm);
|
||
|
||
#if BX_DEBUGGER
|
||
bx_list_c *tlb = new bx_list_c(cpu, "TLB");
|
||
#if BX_CPU_LEVEL >= 5
|
||
BXRS_PARAM_BOOL(tlb, split_large, TLB.split_large);
|
||
#endif
|
||
for (n=0; n<BX_TLB_SIZE; n++) {
|
||
sprintf(name, "entry%d", n);
|
||
bx_list_c *tlb_entry = new bx_list_c(tlb, name);
|
||
BXRS_HEX_PARAM_FIELD(tlb_entry, lpf, TLB.entry[n].lpf);
|
||
BXRS_HEX_PARAM_FIELD(tlb_entry, lpf_mask, TLB.entry[n].lpf_mask);
|
||
BXRS_HEX_PARAM_FIELD(tlb_entry, ppf, TLB.entry[n].ppf);
|
||
BXRS_HEX_PARAM_FIELD(tlb_entry, accessBits, TLB.entry[n].accessBits);
|
||
}
|
||
#endif
|
||
}
|
||
|
||
Bit64s BX_CPU_C::param_save_handler(void *devptr, bx_param_c *param)
|
||
{
|
||
#if !BX_USE_CPU_SMF
|
||
BX_CPU_C *class_ptr = (BX_CPU_C *) devptr;
|
||
return class_ptr->param_save(param);
|
||
}
|
||
|
||
Bit64s BX_CPU_C::param_save(bx_param_c *param)
|
||
{
|
||
#else
|
||
UNUSED(devptr);
|
||
#endif // !BX_USE_CPU_SMF
|
||
const char *pname, *segname;
|
||
bx_segment_reg_t *segment = NULL;
|
||
Bit64s val = 0;
|
||
|
||
pname = param->get_name();
|
||
if (!strcmp(pname, "EFLAGS")) {
|
||
val = read_eflags();
|
||
} else if (!strcmp(pname, "selector")) {
|
||
segname = param->get_parent()->get_name();
|
||
if (!strcmp(segname, "CS")) {
|
||
segment = &BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS];
|
||
} else if (!strcmp(segname, "DS")) {
|
||
segment = &BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS];
|
||
} else if (!strcmp(segname, "SS")) {
|
||
segment = &BX_CPU_THIS_PTR sregs[BX_SEG_REG_SS];
|
||
} else if (!strcmp(segname, "ES")) {
|
||
segment = &BX_CPU_THIS_PTR sregs[BX_SEG_REG_ES];
|
||
} else if (!strcmp(segname, "FS")) {
|
||
segment = &BX_CPU_THIS_PTR sregs[BX_SEG_REG_FS];
|
||
} else if (!strcmp(segname, "GS")) {
|
||
segment = &BX_CPU_THIS_PTR sregs[BX_SEG_REG_GS];
|
||
} else if (!strcmp(segname, "LDTR")) {
|
||
segment = &BX_CPU_THIS_PTR ldtr;
|
||
} else if (!strcmp(segname, "TR")) {
|
||
segment = &BX_CPU_THIS_PTR tr;
|
||
}
|
||
if (segment != NULL) {
|
||
val = segment->selector.value;
|
||
}
|
||
}
|
||
else {
|
||
BX_PANIC(("Unknown param %s in param_save handler !", pname));
|
||
}
|
||
return val;
|
||
}
|
||
|
||
void BX_CPU_C::param_restore_handler(void *devptr, bx_param_c *param, Bit64s val)
|
||
{
|
||
#if !BX_USE_CPU_SMF
|
||
BX_CPU_C *class_ptr = (BX_CPU_C *) devptr;
|
||
class_ptr->param_restore(param, val);
|
||
}
|
||
|
||
void BX_CPU_C::param_restore(bx_param_c *param, Bit64s val)
|
||
{
|
||
#else
|
||
UNUSED(devptr);
|
||
#endif // !BX_USE_CPU_SMF
|
||
const char *pname, *segname;
|
||
bx_segment_reg_t *segment = NULL;
|
||
|
||
pname = param->get_name();
|
||
if (!strcmp(pname, "EFLAGS")) {
|
||
BX_CPU_THIS_PTR setEFlags((Bit32u)val);
|
||
} else if (!strcmp(pname, "selector")) {
|
||
segname = param->get_parent()->get_name();
|
||
if (!strcmp(segname, "CS")) {
|
||
segment = &BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS];
|
||
} else if (!strcmp(segname, "DS")) {
|
||
segment = &BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS];
|
||
} else if (!strcmp(segname, "SS")) {
|
||
segment = &BX_CPU_THIS_PTR sregs[BX_SEG_REG_SS];
|
||
} else if (!strcmp(segname, "ES")) {
|
||
segment = &BX_CPU_THIS_PTR sregs[BX_SEG_REG_ES];
|
||
} else if (!strcmp(segname, "FS")) {
|
||
segment = &BX_CPU_THIS_PTR sregs[BX_SEG_REG_FS];
|
||
} else if (!strcmp(segname, "GS")) {
|
||
segment = &BX_CPU_THIS_PTR sregs[BX_SEG_REG_GS];
|
||
} else if (!strcmp(segname, "LDTR")) {
|
||
segment = &BX_CPU_THIS_PTR ldtr;
|
||
} else if (!strcmp(segname, "TR")) {
|
||
segment = &BX_CPU_THIS_PTR tr;
|
||
}
|
||
if (segment != NULL) {
|
||
bx_selector_t *selector = &(segment->selector);
|
||
parse_selector((Bit16u)val, selector);
|
||
}
|
||
}
|
||
else {
|
||
BX_PANIC(("Unknown param %s in param_restore handler !", pname));
|
||
}
|
||
}
|
||
|
||
void BX_CPU_C::after_restore_state(void)
|
||
{
|
||
handleCpuContextChange();
|
||
|
||
BX_CPU_THIS_PTR prev_rip = RIP;
|
||
|
||
if (BX_CPU_THIS_PTR cpu_mode == BX_MODE_IA32_REAL) CPL = 0;
|
||
else {
|
||
if (BX_CPU_THIS_PTR cpu_mode == BX_MODE_IA32_V8086) CPL = 3;
|
||
}
|
||
|
||
#if BX_SUPPORT_VMX
|
||
set_VMCSPTR(BX_CPU_THIS_PTR vmcsptr);
|
||
#endif
|
||
|
||
assert_checks();
|
||
debug(RIP);
|
||
}
|
||
// end of save/restore functionality
|
||
|
||
BX_CPU_C::~BX_CPU_C()
|
||
{
|
||
#if BX_CPU_LEVEL >= 4
|
||
delete cpuid;
|
||
#endif
|
||
|
||
BX_INSTR_EXIT(BX_CPU_ID);
|
||
BX_DEBUG(("Exit."));
|
||
}
|
||
|
||
void BX_CPU_C::reset(unsigned source)
|
||
{
|
||
unsigned n;
|
||
|
||
if (source == BX_RESET_HARDWARE)
|
||
BX_INFO(("cpu hardware reset"));
|
||
else if (source == BX_RESET_SOFTWARE)
|
||
BX_INFO(("cpu software reset"));
|
||
else
|
||
BX_INFO(("cpu reset"));
|
||
|
||
for (n=0;n<BX_GENERAL_REGISTERS;n++)
|
||
BX_WRITE_32BIT_REGZ(n, 0);
|
||
|
||
//BX_WRITE_32BIT_REGZ(BX_32BIT_REG_EDX, get_cpu_version_information());
|
||
|
||
// initialize NIL register
|
||
BX_WRITE_32BIT_REGZ(BX_NIL_REGISTER, 0);
|
||
|
||
BX_CPU_THIS_PTR eflags = 0x2; // Bit1 is always set
|
||
// clear lazy flags state to satisfy Valgrind uninitialized variables checker
|
||
memset(&BX_CPU_THIS_PTR oszapc, 0, sizeof(BX_CPU_THIS_PTR oszapc));
|
||
setEFlagsOSZAPC(0); // update lazy flags state
|
||
|
||
if (source == BX_RESET_HARDWARE)
|
||
BX_CPU_THIS_PTR icount = 0;
|
||
BX_CPU_THIS_PTR icount_last_sync = BX_CPU_THIS_PTR icount;
|
||
|
||
BX_CPU_THIS_PTR inhibit_mask = 0;
|
||
BX_CPU_THIS_PTR inhibit_icount = 0;
|
||
|
||
BX_CPU_THIS_PTR activity_state = BX_ACTIVITY_STATE_ACTIVE;
|
||
BX_CPU_THIS_PTR debug_trap = 0;
|
||
|
||
/* instruction pointer */
|
||
#if BX_CPU_LEVEL < 2
|
||
BX_CPU_THIS_PTR prev_rip = EIP = 0x00000000;
|
||
#else /* from 286 up */
|
||
BX_CPU_THIS_PTR prev_rip = RIP = 0x0000FFF0;
|
||
#endif
|
||
|
||
/* CS (Code Segment) and descriptor cache */
|
||
/* Note: on a real cpu, CS initially points to upper memory. After
|
||
* the 1st jump, the descriptor base is zero'd out. Since I'm just
|
||
* going to jump to my BIOS, I don't need to do this.
|
||
* For future reference:
|
||
* processor cs.selector cs.base cs.limit EIP
|
||
* 8086 FFFF FFFF0 FFFF 0000
|
||
* 286 F000 FF0000 FFFF FFF0
|
||
* 386+ F000 FFFF0000 FFFF FFF0
|
||
*/
|
||
parse_selector(0xf000,
|
||
&BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].selector);
|
||
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].cache.valid = SegValidCache | SegAccessROK | SegAccessWOK;
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].cache.p = 1;
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].cache.dpl = 0;
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].cache.segment = 1; /* data/code segment */
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].cache.type = BX_DATA_READ_WRITE_ACCESSED;
|
||
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].cache.u.segment.base = 0xFFFF0000;
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].cache.u.segment.limit_scaled = 0xFFFF;
|
||
|
||
#if BX_CPU_LEVEL >= 3
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].cache.u.segment.g = 0; /* byte granular */
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].cache.u.segment.d_b = 0; /* 16bit default size */
|
||
#if BX_SUPPORT_X86_64
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].cache.u.segment.l = 0; /* 16bit default size */
|
||
#endif
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].cache.u.segment.avl = 0;
|
||
#endif
|
||
|
||
flushICaches();
|
||
|
||
/* DS (Data Segment) and descriptor cache */
|
||
parse_selector(0x0000,
|
||
&BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS].selector);
|
||
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS].cache.valid = SegValidCache | SegAccessROK | SegAccessWOK;
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS].cache.p = 1;
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS].cache.dpl = 0;
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS].cache.segment = 1; /* data/code segment */
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS].cache.type = BX_DATA_READ_WRITE_ACCESSED;
|
||
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS].cache.u.segment.base = 0x00000000;
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS].cache.u.segment.limit_scaled = 0xFFFF;
|
||
#if BX_CPU_LEVEL >= 3
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS].cache.u.segment.avl = 0;
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS].cache.u.segment.g = 0; /* byte granular */
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS].cache.u.segment.d_b = 0; /* 16bit default size */
|
||
#if BX_SUPPORT_X86_64
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS].cache.u.segment.l = 0; /* 16bit default size */
|
||
#endif
|
||
#endif
|
||
|
||
// use DS segment as template for the others
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_SS] = BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS];
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_ES] = BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS];
|
||
#if BX_CPU_LEVEL >= 3
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_FS] = BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS];
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_GS] = BX_CPU_THIS_PTR sregs[BX_SEG_REG_DS];
|
||
#endif
|
||
|
||
/* GDTR (Global Descriptor Table Register) */
|
||
BX_CPU_THIS_PTR gdtr.base = 0x00000000;
|
||
BX_CPU_THIS_PTR gdtr.limit = 0xFFFF;
|
||
|
||
/* IDTR (Interrupt Descriptor Table Register) */
|
||
BX_CPU_THIS_PTR idtr.base = 0x00000000;
|
||
BX_CPU_THIS_PTR idtr.limit = 0xFFFF; /* always byte granular */
|
||
|
||
/* LDTR (Local Descriptor Table Register) */
|
||
BX_CPU_THIS_PTR ldtr.selector.value = 0x0000;
|
||
BX_CPU_THIS_PTR ldtr.selector.index = 0x0000;
|
||
BX_CPU_THIS_PTR ldtr.selector.ti = 0;
|
||
BX_CPU_THIS_PTR ldtr.selector.rpl = 0;
|
||
|
||
BX_CPU_THIS_PTR ldtr.cache.valid = 1; /* valid */
|
||
BX_CPU_THIS_PTR ldtr.cache.p = 1; /* present */
|
||
BX_CPU_THIS_PTR ldtr.cache.dpl = 0; /* field not used */
|
||
BX_CPU_THIS_PTR ldtr.cache.segment = 0; /* system segment */
|
||
BX_CPU_THIS_PTR ldtr.cache.type = BX_SYS_SEGMENT_LDT;
|
||
BX_CPU_THIS_PTR ldtr.cache.u.segment.base = 0x00000000;
|
||
BX_CPU_THIS_PTR ldtr.cache.u.segment.limit_scaled = 0xFFFF;
|
||
BX_CPU_THIS_PTR ldtr.cache.u.segment.avl = 0;
|
||
BX_CPU_THIS_PTR ldtr.cache.u.segment.g = 0; /* byte granular */
|
||
|
||
/* TR (Task Register) */
|
||
BX_CPU_THIS_PTR tr.selector.value = 0x0000;
|
||
BX_CPU_THIS_PTR tr.selector.index = 0x0000; /* undefined */
|
||
BX_CPU_THIS_PTR tr.selector.ti = 0;
|
||
BX_CPU_THIS_PTR tr.selector.rpl = 0;
|
||
|
||
BX_CPU_THIS_PTR tr.cache.valid = 1; /* valid */
|
||
BX_CPU_THIS_PTR tr.cache.p = 1; /* present */
|
||
BX_CPU_THIS_PTR tr.cache.dpl = 0; /* field not used */
|
||
BX_CPU_THIS_PTR tr.cache.segment = 0; /* system segment */
|
||
BX_CPU_THIS_PTR tr.cache.type = BX_SYS_SEGMENT_BUSY_386_TSS;
|
||
BX_CPU_THIS_PTR tr.cache.u.segment.base = 0x00000000;
|
||
BX_CPU_THIS_PTR tr.cache.u.segment.limit_scaled = 0xFFFF;
|
||
BX_CPU_THIS_PTR tr.cache.u.segment.avl = 0;
|
||
BX_CPU_THIS_PTR tr.cache.u.segment.g = 0; /* byte granular */
|
||
|
||
BX_CPU_THIS_PTR cpu_mode = BX_MODE_IA32_REAL;
|
||
|
||
// DR0 - DR7 (Debug Registers)
|
||
#if BX_CPU_LEVEL >= 3
|
||
for (n=0; n<4; n++)
|
||
BX_CPU_THIS_PTR dr[n] = 0;
|
||
#endif
|
||
|
||
#if BX_CPU_LEVEL >= 5
|
||
BX_CPU_THIS_PTR dr6.val32 = 0xFFFF0FF0;
|
||
#else
|
||
BX_CPU_THIS_PTR dr6.val32 = 0xFFFF1FF0;
|
||
#endif
|
||
BX_CPU_THIS_PTR dr7.val32 = 0x00000400;
|
||
|
||
#if BX_X86_DEBUGGER
|
||
BX_CPU_THIS_PTR in_repeat = 0;
|
||
#endif
|
||
BX_CPU_THIS_PTR in_smm = 0;
|
||
|
||
BX_CPU_THIS_PTR pending_event = 0;
|
||
BX_CPU_THIS_PTR event_mask = 0;
|
||
|
||
if (source == BX_RESET_HARDWARE) {
|
||
BX_CPU_THIS_PTR smbase = 0x30000; // do not change SMBASE on INIT
|
||
}
|
||
|
||
BX_CPU_THIS_PTR cr0.set32(0x60000010);
|
||
// handle reserved bits
|
||
#if BX_CPU_LEVEL == 3
|
||
// reserved bits all set to 1 on 386
|
||
BX_CPU_THIS_PTR cr0.val32 |= 0x7ffffff0;
|
||
#endif
|
||
|
||
#if BX_CPU_LEVEL >= 3
|
||
BX_CPU_THIS_PTR cr2 = 0;
|
||
BX_CPU_THIS_PTR cr3 = 0;
|
||
#endif
|
||
|
||
#if BX_CPU_LEVEL >= 5
|
||
BX_CPU_THIS_PTR cr4.set32(0);
|
||
BX_CPU_THIS_PTR cr4_suppmask = get_cr4_allow_mask();
|
||
#endif
|
||
|
||
#if BX_CPU_LEVEL >= 6
|
||
BX_CPU_THIS_PTR xcr0.set32(0x1);
|
||
BX_CPU_THIS_PTR xcr0_suppmask = 0x3;
|
||
#if BX_SUPPORT_AVX
|
||
if (BX_CPUID_SUPPORT_ISA_EXTENSION(BX_ISA_AVX))
|
||
BX_CPU_THIS_PTR xcr0_suppmask |= BX_XCR0_AVX_MASK;
|
||
#endif
|
||
#endif
|
||
|
||
/* initialise MSR registers to defaults */
|
||
#if BX_CPU_LEVEL >= 5
|
||
#if BX_SUPPORT_APIC
|
||
/* APIC Address, APIC enabled and BSP is default, we'll fill in the rest later */
|
||
BX_CPU_THIS_PTR msr.apicbase = BX_LAPIC_BASE_ADDR;
|
||
BX_CPU_THIS_PTR lapic.reset(source);
|
||
BX_CPU_THIS_PTR msr.apicbase |= 0x900;
|
||
BX_CPU_THIS_PTR lapic.set_base(BX_CPU_THIS_PTR msr.apicbase);
|
||
#if BX_CPU_LEVEL >= 6
|
||
if (BX_CPUID_SUPPORT_CPU_EXTENSION(BX_CPU_XAPIC_EXT))
|
||
BX_CPU_THIS_PTR lapic.enable_xapic_extensions();
|
||
#endif
|
||
#endif
|
||
|
||
BX_CPU_THIS_PTR efer.set32(0);
|
||
BX_CPU_THIS_PTR efer_suppmask = 0;
|
||
if (BX_CPUID_SUPPORT_CPU_EXTENSION(BX_CPU_NX))
|
||
BX_CPU_THIS_PTR efer_suppmask |= BX_EFER_NXE_MASK;
|
||
if (BX_CPUID_SUPPORT_ISA_EXTENSION(BX_ISA_SYSCALL_SYSRET_LEGACY))
|
||
BX_CPU_THIS_PTR efer_suppmask |= BX_EFER_SCE_MASK;
|
||
#if BX_SUPPORT_X86_64
|
||
if (BX_CPUID_SUPPORT_CPU_EXTENSION(BX_CPU_LONG_MODE)) {
|
||
BX_CPU_THIS_PTR efer_suppmask |= (BX_EFER_SCE_MASK | BX_EFER_LME_MASK | BX_EFER_LMA_MASK);
|
||
if (BX_CPUID_SUPPORT_CPU_EXTENSION(BX_CPU_FFXSR))
|
||
BX_CPU_THIS_PTR efer_suppmask |= BX_EFER_FFXSR_MASK;
|
||
if (BX_CPUID_SUPPORT_ISA_EXTENSION(BX_ISA_SVM))
|
||
BX_CPU_THIS_PTR efer_suppmask |= BX_EFER_SVME_MASK;
|
||
}
|
||
#endif
|
||
|
||
BX_CPU_THIS_PTR msr.star = 0;
|
||
#if BX_SUPPORT_X86_64
|
||
if (BX_CPUID_SUPPORT_CPU_EXTENSION(BX_CPU_LONG_MODE)) {
|
||
BX_CPU_THIS_PTR msr.lstar = 0;
|
||
BX_CPU_THIS_PTR msr.cstar = 0;
|
||
BX_CPU_THIS_PTR msr.fmask = 0x00020200;
|
||
BX_CPU_THIS_PTR msr.kernelgsbase = 0;
|
||
BX_CPU_THIS_PTR msr.tsc_aux = 0;
|
||
}
|
||
#endif
|
||
|
||
#if BX_SUPPORT_VMX || BX_SUPPORT_SVM
|
||
BX_CPU_THIS_PTR tsc_offset = 0;
|
||
#endif
|
||
if (source == BX_RESET_HARDWARE) {
|
||
BX_CPU_THIS_PTR set_TSC(0); // do not change TSC on INIT
|
||
}
|
||
#endif // BX_CPU_LEVEL >= 5
|
||
|
||
#if BX_CPU_LEVEL >= 6
|
||
BX_CPU_THIS_PTR msr.sysenter_cs_msr = 0;
|
||
BX_CPU_THIS_PTR msr.sysenter_esp_msr = 0;
|
||
BX_CPU_THIS_PTR msr.sysenter_eip_msr = 0;
|
||
#endif
|
||
|
||
// Do not change MTRR on INIT
|
||
#if BX_CPU_LEVEL >= 6
|
||
if (source == BX_RESET_HARDWARE) {
|
||
for (n=0; n<16; n++)
|
||
BX_CPU_THIS_PTR msr.mtrrphys[n] = 0;
|
||
|
||
BX_CPU_THIS_PTR msr.mtrrfix64k_00000 = 0; // all fix range MTRRs undefined according to manual
|
||
BX_CPU_THIS_PTR msr.mtrrfix16k[0] = 0;
|
||
BX_CPU_THIS_PTR msr.mtrrfix16k[1] = 0;
|
||
|
||
for (n=0; n<8; n++)
|
||
BX_CPU_THIS_PTR msr.mtrrfix4k[n] = 0;
|
||
|
||
BX_CPU_THIS_PTR msr.pat = BX_CONST64(0x0007040600070406);
|
||
BX_CPU_THIS_PTR msr.mtrr_deftype = 0;
|
||
}
|
||
#endif
|
||
|
||
// All configurable MSRs do not change on INIT
|
||
#if BX_CONFIGURE_MSRS
|
||
if (source == BX_RESET_HARDWARE) {
|
||
for (n=0; n < BX_MSR_MAX_INDEX; n++) {
|
||
if (BX_CPU_THIS_PTR msrs[n])
|
||
BX_CPU_THIS_PTR msrs[n]->reset();
|
||
}
|
||
}
|
||
#endif
|
||
|
||
BX_CPU_THIS_PTR EXT = 0;
|
||
BX_CPU_THIS_PTR last_exception_type = 0;
|
||
|
||
// invalidate the code prefetch queue
|
||
BX_CPU_THIS_PTR eipPageBias = 0;
|
||
BX_CPU_THIS_PTR eipPageWindowSize = 0;
|
||
BX_CPU_THIS_PTR eipFetchPtr = NULL;
|
||
|
||
// invalidate current stack page
|
||
BX_CPU_THIS_PTR espPageBias = 0;
|
||
BX_CPU_THIS_PTR espPageWindowSize = 0;
|
||
BX_CPU_THIS_PTR espHostPtr = NULL;
|
||
|
||
#if BX_DEBUGGER
|
||
BX_CPU_THIS_PTR stop_reason = STOP_NO_REASON;
|
||
BX_CPU_THIS_PTR magic_break = 0;
|
||
BX_CPU_THIS_PTR trace = 0;
|
||
BX_CPU_THIS_PTR trace_reg = 0;
|
||
BX_CPU_THIS_PTR trace_mem = 0;
|
||
BX_CPU_THIS_PTR mode_break = 0;
|
||
#if BX_SUPPORT_VMX
|
||
BX_CPU_THIS_PTR vmexit_break = 0;
|
||
#endif
|
||
#endif
|
||
|
||
// Reset the Floating Point Unit
|
||
#if BX_SUPPORT_FPU
|
||
if (source == BX_RESET_HARDWARE) {
|
||
BX_CPU_THIS_PTR the_i387.reset();
|
||
}
|
||
#endif
|
||
|
||
#if BX_CPU_LEVEL >= 6
|
||
BX_CPU_THIS_PTR sse_ok = 0;
|
||
#if BX_SUPPORT_AVX
|
||
BX_CPU_THIS_PTR avx_ok = 0;
|
||
#endif
|
||
|
||
// Reset XMM state - unchanged on #INIT
|
||
if (source == BX_RESET_HARDWARE) {
|
||
static BxPackedXmmRegister xmmnil; /* compiler will clear the variable */
|
||
for(n=0; n<BX_XMM_REGISTERS; n++)
|
||
BX_WRITE_XMM_REG_CLEAR_HIGH(n, xmmnil);
|
||
|
||
BX_CPU_THIS_PTR mxcsr.mxcsr = MXCSR_RESET;
|
||
BX_CPU_THIS_PTR mxcsr_mask = 0x0000ffbf;
|
||
if (BX_CPUID_SUPPORT_ISA_EXTENSION(BX_ISA_SSE2))
|
||
BX_CPU_THIS_PTR mxcsr_mask |= MXCSR_DAZ;
|
||
if (BX_CPUID_SUPPORT_CPU_EXTENSION(BX_CPU_MISALIGNED_SSE))
|
||
BX_CPU_THIS_PTR mxcsr_mask |= MXCSR_MISALIGNED_EXCEPTION_MASK;
|
||
}
|
||
#endif
|
||
|
||
#if BX_SUPPORT_VMX
|
||
BX_CPU_THIS_PTR in_vmx = BX_CPU_THIS_PTR in_vmx_guest = 0;
|
||
BX_CPU_THIS_PTR in_smm_vmx = BX_CPU_THIS_PTR in_smm_vmx_guest = 0;
|
||
BX_CPU_THIS_PTR vmcsptr = BX_CPU_THIS_PTR vmxonptr = BX_INVALID_VMCSPTR;
|
||
BX_CPU_THIS_PTR vmcshostptr = 0;
|
||
/* enable VMX, should be done in BIOS instead */
|
||
BX_CPU_THIS_PTR msr.ia32_feature_ctrl =
|
||
/*BX_IA32_FEATURE_CONTROL_LOCK_BIT | */BX_IA32_FEATURE_CONTROL_VMX_ENABLE_BIT;
|
||
#endif
|
||
|
||
#if BX_SUPPORT_SVM
|
||
BX_CPU_THIS_PTR in_svm_guest = 0;
|
||
BX_CPU_THIS_PTR svm_gif = 1;
|
||
BX_CPU_THIS_PTR vmcbptr = 0;
|
||
BX_CPU_THIS_PTR vmcbhostptr = 0;
|
||
#endif
|
||
|
||
#if BX_SUPPORT_VMX || BX_SUPPORT_SVM
|
||
BX_CPU_THIS_PTR in_event = 0;
|
||
#endif
|
||
|
||
#if BX_SUPPORT_VMX
|
||
BX_CPU_THIS_PTR nmi_unblocking_iret = 0;
|
||
#endif
|
||
|
||
#if BX_SUPPORT_SMP
|
||
// notice if I'm the bootstrap processor. If not, do the equivalent of
|
||
// a HALT instruction.
|
||
int apic_id = lapic.get_id();
|
||
if (BX_BOOTSTRAP_PROCESSOR == apic_id) {
|
||
// boot normally
|
||
BX_CPU_THIS_PTR msr.apicbase |= 0x100; /* set bit 8 BSP */
|
||
BX_INFO(("CPU[%d] is the bootstrap processor", apic_id));
|
||
} else {
|
||
// it's an application processor, halt until IPI is heard.
|
||
BX_CPU_THIS_PTR msr.apicbase &= ~0x100; /* clear bit 8 BSP */
|
||
BX_INFO(("CPU[%d] is an application processor. Halting until SIPI.", apic_id));
|
||
enter_sleep_state(BX_ACTIVITY_STATE_WAIT_FOR_SIPI);
|
||
}
|
||
#endif
|
||
|
||
handleCpuContextChange();
|
||
|
||
#if BX_CPU_LEVEL >= 4
|
||
BX_CPU_THIS_PTR cpuid->dump_cpuid();
|
||
#endif
|
||
|
||
BX_INSTR_RESET(BX_CPU_ID, source);
|
||
}
|
||
|
||
void BX_CPU_C::sanity_checks(void)
|
||
{
|
||
Bit32u eax = EAX, ecx = ECX, edx = EDX, ebx = EBX, esp = ESP, ebp = EBP, esi = ESI, edi = EDI;
|
||
|
||
EAX = 0xFFEEDDCC;
|
||
ECX = 0xBBAA9988;
|
||
EDX = 0x77665544;
|
||
EBX = 0x332211FF;
|
||
ESP = 0xEEDDCCBB;
|
||
EBP = 0xAA998877;
|
||
ESI = 0x66554433;
|
||
EDI = 0x2211FFEE;
|
||
|
||
Bit8u al, cl, dl, bl, ah, ch, dh, bh;
|
||
|
||
al = AL;
|
||
cl = CL;
|
||
dl = DL;
|
||
bl = BL;
|
||
ah = AH;
|
||
ch = CH;
|
||
dh = DH;
|
||
bh = BH;
|
||
|
||
if ( al != (EAX & 0xFF) ||
|
||
cl != (ECX & 0xFF) ||
|
||
dl != (EDX & 0xFF) ||
|
||
bl != (EBX & 0xFF) ||
|
||
ah != ((EAX >> 8) & 0xFF) ||
|
||
ch != ((ECX >> 8) & 0xFF) ||
|
||
dh != ((EDX >> 8) & 0xFF) ||
|
||
bh != ((EBX >> 8) & 0xFF) )
|
||
{
|
||
BX_PANIC(("problems using BX_READ_8BIT_REGx()!"));
|
||
}
|
||
|
||
Bit16u ax, cx, dx, bx, sp, bp, si, di;
|
||
|
||
ax = AX;
|
||
cx = CX;
|
||
dx = DX;
|
||
bx = BX;
|
||
sp = SP;
|
||
bp = BP;
|
||
si = SI;
|
||
di = DI;
|
||
|
||
if ( ax != (EAX & 0xFFFF) ||
|
||
cx != (ECX & 0xFFFF) ||
|
||
dx != (EDX & 0xFFFF) ||
|
||
bx != (EBX & 0xFFFF) ||
|
||
sp != (ESP & 0xFFFF) ||
|
||
bp != (EBP & 0xFFFF) ||
|
||
si != (ESI & 0xFFFF) ||
|
||
di != (EDI & 0xFFFF) )
|
||
{
|
||
BX_PANIC(("problems using BX_READ_16BIT_REG()!"));
|
||
}
|
||
|
||
EAX = eax; /* restore registers */
|
||
ECX = ecx;
|
||
EDX = edx;
|
||
EBX = ebx;
|
||
ESP = esp;
|
||
EBP = ebp;
|
||
ESI = esi;
|
||
EDI = edi;
|
||
|
||
if (sizeof(Bit8u) != 1 || sizeof(Bit8s) != 1)
|
||
BX_PANIC(("data type Bit8u or Bit8s is not of length 1 byte!"));
|
||
if (sizeof(Bit16u) != 2 || sizeof(Bit16s) != 2)
|
||
BX_PANIC(("data type Bit16u or Bit16s is not of length 2 bytes!"));
|
||
if (sizeof(Bit32u) != 4 || sizeof(Bit32s) != 4)
|
||
BX_PANIC(("data type Bit32u or Bit32s is not of length 4 bytes!"));
|
||
if (sizeof(Bit64u) != 8 || sizeof(Bit64s) != 8)
|
||
BX_PANIC(("data type Bit64u or Bit64u is not of length 8 bytes!"));
|
||
|
||
BX_DEBUG(("#(%u)all sanity checks passed!", BX_CPU_ID));
|
||
}
|
||
|
||
void BX_CPU_C::assert_checks(void)
|
||
{
|
||
// check CPU mode consistency
|
||
#if BX_SUPPORT_X86_64
|
||
if (BX_CPU_THIS_PTR efer.get_LMA()) {
|
||
if (! BX_CPU_THIS_PTR cr0.get_PE()) {
|
||
BX_PANIC(("assert_checks: EFER.LMA is set when CR0.PE=0 !"));
|
||
}
|
||
if (BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].cache.u.segment.l) {
|
||
if (BX_CPU_THIS_PTR cpu_mode != BX_MODE_LONG_64)
|
||
BX_PANIC(("assert_checks: unconsistent cpu_mode BX_MODE_LONG_64 !"));
|
||
}
|
||
else {
|
||
if (BX_CPU_THIS_PTR cpu_mode != BX_MODE_LONG_COMPAT)
|
||
BX_PANIC(("assert_checks: unconsistent cpu_mode BX_MODE_LONG_COMPAT !"));
|
||
}
|
||
}
|
||
else
|
||
#endif
|
||
{
|
||
if (BX_CPU_THIS_PTR cr0.get_PE()) {
|
||
if (BX_CPU_THIS_PTR get_VM()) {
|
||
if (BX_CPU_THIS_PTR cpu_mode != BX_MODE_IA32_V8086)
|
||
BX_PANIC(("assert_checks: unconsistent cpu_mode BX_MODE_IA32_V8086 !"));
|
||
}
|
||
else {
|
||
if (BX_CPU_THIS_PTR cpu_mode != BX_MODE_IA32_PROTECTED)
|
||
BX_PANIC(("assert_checks: unconsistent cpu_mode BX_MODE_IA32_PROTECTED !"));
|
||
}
|
||
}
|
||
else {
|
||
if (BX_CPU_THIS_PTR cpu_mode != BX_MODE_IA32_REAL)
|
||
BX_PANIC(("assert_checks: unconsistent cpu_mode BX_MODE_IA32_REAL !"));
|
||
}
|
||
}
|
||
|
||
// check CR0 consistency
|
||
if (! check_CR0(BX_CPU_THIS_PTR cr0.val32))
|
||
BX_PANIC(("assert_checks: CR0 consistency checks failed !"));
|
||
|
||
#if BX_CPU_LEVEL >= 5
|
||
// check CR4 consistency
|
||
if (! check_CR4(BX_CPU_THIS_PTR cr4.val32))
|
||
BX_PANIC(("assert_checks: CR4 consistency checks failed !"));
|
||
#endif
|
||
|
||
#if BX_SUPPORT_X86_64
|
||
// VM should be OFF in long mode
|
||
if (long_mode()) {
|
||
if (BX_CPU_THIS_PTR get_VM()) BX_PANIC(("assert_checks: VM is set in long mode !"));
|
||
}
|
||
|
||
// CS.L and CS.D_B are mutualy exclusive
|
||
if (BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].cache.u.segment.l &&
|
||
BX_CPU_THIS_PTR sregs[BX_SEG_REG_CS].cache.u.segment.d_b)
|
||
{
|
||
BX_PANIC(("assert_checks: CS.l and CS.d_b set together !"));
|
||
}
|
||
#endif
|
||
|
||
// check LDTR type
|
||
if (BX_CPU_THIS_PTR ldtr.cache.valid)
|
||
{
|
||
if (BX_CPU_THIS_PTR ldtr.cache.type != BX_SYS_SEGMENT_LDT)
|
||
{
|
||
BX_PANIC(("assert_checks: LDTR is not LDT type !"));
|
||
}
|
||
}
|
||
|
||
// check Task Register type
|
||
if(BX_CPU_THIS_PTR tr.cache.valid)
|
||
{
|
||
switch(BX_CPU_THIS_PTR tr.cache.type)
|
||
{
|
||
case BX_SYS_SEGMENT_BUSY_286_TSS:
|
||
case BX_SYS_SEGMENT_AVAIL_286_TSS:
|
||
#if BX_CPU_LEVEL >= 3
|
||
if (BX_CPU_THIS_PTR tr.cache.u.segment.g != 0)
|
||
BX_PANIC(("assert_checks: tss286.g != 0 !"));
|
||
if (BX_CPU_THIS_PTR tr.cache.u.segment.avl != 0)
|
||
BX_PANIC(("assert_checks: tss286.avl != 0 !"));
|
||
#endif
|
||
break;
|
||
case BX_SYS_SEGMENT_BUSY_386_TSS:
|
||
case BX_SYS_SEGMENT_AVAIL_386_TSS:
|
||
break;
|
||
default:
|
||
BX_PANIC(("assert_checks: TR is not TSS type !"));
|
||
}
|
||
}
|
||
|
||
#if BX_SUPPORT_X86_64 == 0 && BX_CPU_LEVEL >= 5
|
||
if (BX_CPU_THIS_PTR efer_suppmask & (BX_EFER_SCE_MASK |
|
||
BX_EFER_LME_MASK | BX_EFER_LMA_MASK | BX_EFER_FFXSR_MASK))
|
||
{
|
||
BX_PANIC(("assert_checks: EFER supports x86-64 specific bits !"));
|
||
}
|
||
#endif
|
||
}
|