f8d926e9cd
Save and restore all so far neglected KVM-specific CPU states. Handling the TSC stabilizes migration in KVM mode. The interrupt_bitmap and mp_state are currently unused, but will become relevant for in-kernel irqchip support. By including proper saving/restoring already, we avoid having to increment CPU_SAVE_VERSION later on once again. v2: - initialize mp_state runnable (for the boot CPU) Signed-off-by: Jan Kiszka <jan.kiszka@siemens.com> Signed-off-by: Anthony Liguori <aliguori@us.ibm.com>
342 lines
9.5 KiB
C
342 lines
9.5 KiB
C
#include "hw/hw.h"
|
|
#include "hw/boards.h"
|
|
#include "hw/pc.h"
|
|
#include "hw/isa.h"
|
|
|
|
#include "exec-all.h"
|
|
#include "kvm.h"
|
|
|
|
static void cpu_put_seg(QEMUFile *f, SegmentCache *dt)
|
|
{
|
|
qemu_put_be32(f, dt->selector);
|
|
qemu_put_betl(f, dt->base);
|
|
qemu_put_be32(f, dt->limit);
|
|
qemu_put_be32(f, dt->flags);
|
|
}
|
|
|
|
static void cpu_get_seg(QEMUFile *f, SegmentCache *dt)
|
|
{
|
|
dt->selector = qemu_get_be32(f);
|
|
dt->base = qemu_get_betl(f);
|
|
dt->limit = qemu_get_be32(f);
|
|
dt->flags = qemu_get_be32(f);
|
|
}
|
|
|
|
void cpu_save(QEMUFile *f, void *opaque)
|
|
{
|
|
CPUState *env = opaque;
|
|
uint16_t fptag, fpus, fpuc, fpregs_format;
|
|
uint32_t hflags;
|
|
int32_t a20_mask;
|
|
int i;
|
|
|
|
cpu_synchronize_state(env, 0);
|
|
|
|
for(i = 0; i < CPU_NB_REGS; i++)
|
|
qemu_put_betls(f, &env->regs[i]);
|
|
qemu_put_betls(f, &env->eip);
|
|
qemu_put_betls(f, &env->eflags);
|
|
hflags = env->hflags; /* XXX: suppress most of the redundant hflags */
|
|
qemu_put_be32s(f, &hflags);
|
|
|
|
/* FPU */
|
|
fpuc = env->fpuc;
|
|
fpus = (env->fpus & ~0x3800) | (env->fpstt & 0x7) << 11;
|
|
fptag = 0;
|
|
for(i = 0; i < 8; i++) {
|
|
fptag |= ((!env->fptags[i]) << i);
|
|
}
|
|
|
|
qemu_put_be16s(f, &fpuc);
|
|
qemu_put_be16s(f, &fpus);
|
|
qemu_put_be16s(f, &fptag);
|
|
|
|
#ifdef USE_X86LDOUBLE
|
|
fpregs_format = 0;
|
|
#else
|
|
fpregs_format = 1;
|
|
#endif
|
|
qemu_put_be16s(f, &fpregs_format);
|
|
|
|
for(i = 0; i < 8; i++) {
|
|
#ifdef USE_X86LDOUBLE
|
|
{
|
|
uint64_t mant;
|
|
uint16_t exp;
|
|
/* we save the real CPU data (in case of MMX usage only 'mant'
|
|
contains the MMX register */
|
|
cpu_get_fp80(&mant, &exp, env->fpregs[i].d);
|
|
qemu_put_be64(f, mant);
|
|
qemu_put_be16(f, exp);
|
|
}
|
|
#else
|
|
/* if we use doubles for float emulation, we save the doubles to
|
|
avoid losing information in case of MMX usage. It can give
|
|
problems if the image is restored on a CPU where long
|
|
doubles are used instead. */
|
|
qemu_put_be64(f, env->fpregs[i].mmx.MMX_Q(0));
|
|
#endif
|
|
}
|
|
|
|
for(i = 0; i < 6; i++)
|
|
cpu_put_seg(f, &env->segs[i]);
|
|
cpu_put_seg(f, &env->ldt);
|
|
cpu_put_seg(f, &env->tr);
|
|
cpu_put_seg(f, &env->gdt);
|
|
cpu_put_seg(f, &env->idt);
|
|
|
|
qemu_put_be32s(f, &env->sysenter_cs);
|
|
qemu_put_betls(f, &env->sysenter_esp);
|
|
qemu_put_betls(f, &env->sysenter_eip);
|
|
|
|
qemu_put_betls(f, &env->cr[0]);
|
|
qemu_put_betls(f, &env->cr[2]);
|
|
qemu_put_betls(f, &env->cr[3]);
|
|
qemu_put_betls(f, &env->cr[4]);
|
|
|
|
for(i = 0; i < 8; i++)
|
|
qemu_put_betls(f, &env->dr[i]);
|
|
|
|
/* MMU */
|
|
a20_mask = (int32_t) env->a20_mask;
|
|
qemu_put_sbe32s(f, &a20_mask);
|
|
|
|
/* XMM */
|
|
qemu_put_be32s(f, &env->mxcsr);
|
|
for(i = 0; i < CPU_NB_REGS; i++) {
|
|
qemu_put_be64s(f, &env->xmm_regs[i].XMM_Q(0));
|
|
qemu_put_be64s(f, &env->xmm_regs[i].XMM_Q(1));
|
|
}
|
|
|
|
#ifdef TARGET_X86_64
|
|
qemu_put_be64s(f, &env->efer);
|
|
qemu_put_be64s(f, &env->star);
|
|
qemu_put_be64s(f, &env->lstar);
|
|
qemu_put_be64s(f, &env->cstar);
|
|
qemu_put_be64s(f, &env->fmask);
|
|
qemu_put_be64s(f, &env->kernelgsbase);
|
|
#endif
|
|
qemu_put_be32s(f, &env->smbase);
|
|
|
|
qemu_put_be64s(f, &env->pat);
|
|
qemu_put_be32s(f, &env->hflags2);
|
|
|
|
qemu_put_be64s(f, &env->vm_hsave);
|
|
qemu_put_be64s(f, &env->vm_vmcb);
|
|
qemu_put_be64s(f, &env->tsc_offset);
|
|
qemu_put_be64s(f, &env->intercept);
|
|
qemu_put_be16s(f, &env->intercept_cr_read);
|
|
qemu_put_be16s(f, &env->intercept_cr_write);
|
|
qemu_put_be16s(f, &env->intercept_dr_read);
|
|
qemu_put_be16s(f, &env->intercept_dr_write);
|
|
qemu_put_be32s(f, &env->intercept_exceptions);
|
|
qemu_put_8s(f, &env->v_tpr);
|
|
|
|
/* MTRRs */
|
|
for(i = 0; i < 11; i++)
|
|
qemu_put_be64s(f, &env->mtrr_fixed[i]);
|
|
qemu_put_be64s(f, &env->mtrr_deftype);
|
|
for(i = 0; i < 8; i++) {
|
|
qemu_put_be64s(f, &env->mtrr_var[i].base);
|
|
qemu_put_be64s(f, &env->mtrr_var[i].mask);
|
|
}
|
|
|
|
for (i = 0; i < sizeof(env->interrupt_bitmap)/8; i++) {
|
|
qemu_put_be64s(f, &env->interrupt_bitmap[i]);
|
|
}
|
|
qemu_put_be64s(f, &env->tsc);
|
|
qemu_put_be32s(f, &env->mp_state);
|
|
}
|
|
|
|
#ifdef USE_X86LDOUBLE
|
|
/* XXX: add that in a FPU generic layer */
|
|
union x86_longdouble {
|
|
uint64_t mant;
|
|
uint16_t exp;
|
|
};
|
|
|
|
#define MANTD1(fp) (fp & ((1LL << 52) - 1))
|
|
#define EXPBIAS1 1023
|
|
#define EXPD1(fp) ((fp >> 52) & 0x7FF)
|
|
#define SIGND1(fp) ((fp >> 32) & 0x80000000)
|
|
|
|
static void fp64_to_fp80(union x86_longdouble *p, uint64_t temp)
|
|
{
|
|
int e;
|
|
/* mantissa */
|
|
p->mant = (MANTD1(temp) << 11) | (1LL << 63);
|
|
/* exponent + sign */
|
|
e = EXPD1(temp) - EXPBIAS1 + 16383;
|
|
e |= SIGND1(temp) >> 16;
|
|
p->exp = e;
|
|
}
|
|
#endif
|
|
|
|
int cpu_load(QEMUFile *f, void *opaque, int version_id)
|
|
{
|
|
CPUState *env = opaque;
|
|
int i, guess_mmx;
|
|
uint32_t hflags;
|
|
uint16_t fpus, fpuc, fptag, fpregs_format;
|
|
int32_t a20_mask;
|
|
|
|
if (version_id < 3 || version_id > CPU_SAVE_VERSION)
|
|
return -EINVAL;
|
|
for(i = 0; i < CPU_NB_REGS; i++)
|
|
qemu_get_betls(f, &env->regs[i]);
|
|
qemu_get_betls(f, &env->eip);
|
|
qemu_get_betls(f, &env->eflags);
|
|
qemu_get_be32s(f, &hflags);
|
|
|
|
qemu_get_be16s(f, &fpuc);
|
|
qemu_get_be16s(f, &fpus);
|
|
qemu_get_be16s(f, &fptag);
|
|
qemu_get_be16s(f, &fpregs_format);
|
|
|
|
/* NOTE: we cannot always restore the FPU state if the image come
|
|
from a host with a different 'USE_X86LDOUBLE' define. We guess
|
|
if we are in an MMX state to restore correctly in that case. */
|
|
guess_mmx = ((fptag == 0xff) && (fpus & 0x3800) == 0);
|
|
for(i = 0; i < 8; i++) {
|
|
uint64_t mant;
|
|
uint16_t exp;
|
|
|
|
switch(fpregs_format) {
|
|
case 0:
|
|
mant = qemu_get_be64(f);
|
|
exp = qemu_get_be16(f);
|
|
#ifdef USE_X86LDOUBLE
|
|
env->fpregs[i].d = cpu_set_fp80(mant, exp);
|
|
#else
|
|
/* difficult case */
|
|
if (guess_mmx)
|
|
env->fpregs[i].mmx.MMX_Q(0) = mant;
|
|
else
|
|
env->fpregs[i].d = cpu_set_fp80(mant, exp);
|
|
#endif
|
|
break;
|
|
case 1:
|
|
mant = qemu_get_be64(f);
|
|
#ifdef USE_X86LDOUBLE
|
|
{
|
|
union x86_longdouble *p;
|
|
/* difficult case */
|
|
p = (void *)&env->fpregs[i];
|
|
if (guess_mmx) {
|
|
p->mant = mant;
|
|
p->exp = 0xffff;
|
|
} else {
|
|
fp64_to_fp80(p, mant);
|
|
}
|
|
}
|
|
#else
|
|
env->fpregs[i].mmx.MMX_Q(0) = mant;
|
|
#endif
|
|
break;
|
|
default:
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
env->fpuc = fpuc;
|
|
/* XXX: restore FPU round state */
|
|
env->fpstt = (fpus >> 11) & 7;
|
|
env->fpus = fpus & ~0x3800;
|
|
fptag ^= 0xff;
|
|
for(i = 0; i < 8; i++) {
|
|
env->fptags[i] = (fptag >> i) & 1;
|
|
}
|
|
|
|
for(i = 0; i < 6; i++)
|
|
cpu_get_seg(f, &env->segs[i]);
|
|
cpu_get_seg(f, &env->ldt);
|
|
cpu_get_seg(f, &env->tr);
|
|
cpu_get_seg(f, &env->gdt);
|
|
cpu_get_seg(f, &env->idt);
|
|
|
|
qemu_get_be32s(f, &env->sysenter_cs);
|
|
if (version_id >= 7) {
|
|
qemu_get_betls(f, &env->sysenter_esp);
|
|
qemu_get_betls(f, &env->sysenter_eip);
|
|
} else {
|
|
env->sysenter_esp = qemu_get_be32(f);
|
|
env->sysenter_eip = qemu_get_be32(f);
|
|
}
|
|
|
|
qemu_get_betls(f, &env->cr[0]);
|
|
qemu_get_betls(f, &env->cr[2]);
|
|
qemu_get_betls(f, &env->cr[3]);
|
|
qemu_get_betls(f, &env->cr[4]);
|
|
|
|
for(i = 0; i < 8; i++)
|
|
qemu_get_betls(f, &env->dr[i]);
|
|
cpu_breakpoint_remove_all(env, BP_CPU);
|
|
cpu_watchpoint_remove_all(env, BP_CPU);
|
|
for (i = 0; i < 4; i++)
|
|
hw_breakpoint_insert(env, i);
|
|
|
|
/* MMU */
|
|
qemu_get_sbe32s(f, &a20_mask);
|
|
env->a20_mask = a20_mask;
|
|
|
|
qemu_get_be32s(f, &env->mxcsr);
|
|
for(i = 0; i < CPU_NB_REGS; i++) {
|
|
qemu_get_be64s(f, &env->xmm_regs[i].XMM_Q(0));
|
|
qemu_get_be64s(f, &env->xmm_regs[i].XMM_Q(1));
|
|
}
|
|
|
|
#ifdef TARGET_X86_64
|
|
qemu_get_be64s(f, &env->efer);
|
|
qemu_get_be64s(f, &env->star);
|
|
qemu_get_be64s(f, &env->lstar);
|
|
qemu_get_be64s(f, &env->cstar);
|
|
qemu_get_be64s(f, &env->fmask);
|
|
qemu_get_be64s(f, &env->kernelgsbase);
|
|
#endif
|
|
if (version_id >= 4) {
|
|
qemu_get_be32s(f, &env->smbase);
|
|
}
|
|
if (version_id >= 5) {
|
|
qemu_get_be64s(f, &env->pat);
|
|
qemu_get_be32s(f, &env->hflags2);
|
|
if (version_id < 6)
|
|
qemu_get_be32s(f, &env->halted);
|
|
|
|
qemu_get_be64s(f, &env->vm_hsave);
|
|
qemu_get_be64s(f, &env->vm_vmcb);
|
|
qemu_get_be64s(f, &env->tsc_offset);
|
|
qemu_get_be64s(f, &env->intercept);
|
|
qemu_get_be16s(f, &env->intercept_cr_read);
|
|
qemu_get_be16s(f, &env->intercept_cr_write);
|
|
qemu_get_be16s(f, &env->intercept_dr_read);
|
|
qemu_get_be16s(f, &env->intercept_dr_write);
|
|
qemu_get_be32s(f, &env->intercept_exceptions);
|
|
qemu_get_8s(f, &env->v_tpr);
|
|
}
|
|
|
|
if (version_id >= 8) {
|
|
/* MTRRs */
|
|
for(i = 0; i < 11; i++)
|
|
qemu_get_be64s(f, &env->mtrr_fixed[i]);
|
|
qemu_get_be64s(f, &env->mtrr_deftype);
|
|
for(i = 0; i < 8; i++) {
|
|
qemu_get_be64s(f, &env->mtrr_var[i].base);
|
|
qemu_get_be64s(f, &env->mtrr_var[i].mask);
|
|
}
|
|
}
|
|
if (version_id >= 9) {
|
|
for (i = 0; i < sizeof(env->interrupt_bitmap)/8; i++) {
|
|
qemu_get_be64s(f, &env->interrupt_bitmap[i]);
|
|
}
|
|
qemu_get_be64s(f, &env->tsc);
|
|
qemu_get_be32s(f, &env->mp_state);
|
|
}
|
|
|
|
/* XXX: ensure compatiblity for halted bit ? */
|
|
/* XXX: compute redundant hflags bits */
|
|
env->hflags = hflags;
|
|
tlb_flush(env, 1);
|
|
cpu_synchronize_state(env, 1);
|
|
return 0;
|
|
}
|