mirror of
https://git.FreeBSD.org/src.git
synced 2024-12-24 11:29:10 +00:00
becd984900
When the guest is bringing up the APs in the x2APIC mode a write to the ICR register will now trigger a return to userspace with an exitcode of VM_EXITCODE_SPINUP_AP. This gets SMP guests working again with x2APIC. Change the vlapic timer lock to be a spinlock because the vlapic can be accessed from within a critical section (vm run loop) when guest is using x2apic mode. Reviewed by: grehan@
274 lines
6.3 KiB
C
274 lines
6.3 KiB
C
/*-
|
|
* Copyright (c) 2011 NetApp, Inc.
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY NETAPP, INC ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL NETAPP, INC OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* $FreeBSD$
|
|
*/
|
|
|
|
#include <sys/cdefs.h>
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/systm.h>
|
|
#include <sys/smp.h>
|
|
|
|
#include <machine/specialreg.h>
|
|
|
|
#include <machine/vmm.h>
|
|
#include "vmm_lapic.h"
|
|
#include "vmm_msr.h"
|
|
|
|
#define VMM_MSR_F_EMULATE 0x01
|
|
#define VMM_MSR_F_READONLY 0x02
|
|
#define VMM_MSR_F_INVALID 0x04 /* guest_msr_valid() can override this */
|
|
|
|
struct vmm_msr {
|
|
int num;
|
|
int flags;
|
|
uint64_t hostval;
|
|
};
|
|
|
|
static struct vmm_msr vmm_msr[] = {
|
|
{ MSR_LSTAR, 0 },
|
|
{ MSR_CSTAR, 0 },
|
|
{ MSR_STAR, 0 },
|
|
{ MSR_SF_MASK, 0 },
|
|
{ MSR_PAT, VMM_MSR_F_EMULATE | VMM_MSR_F_INVALID },
|
|
{ MSR_BIOS_SIGN,VMM_MSR_F_EMULATE },
|
|
{ MSR_MCG_CAP, VMM_MSR_F_EMULATE | VMM_MSR_F_READONLY },
|
|
{ MSR_IA32_PLATFORM_ID, VMM_MSR_F_EMULATE | VMM_MSR_F_READONLY },
|
|
{ MSR_IA32_MISC_ENABLE, VMM_MSR_F_EMULATE | VMM_MSR_F_READONLY },
|
|
};
|
|
|
|
#define vmm_msr_num (sizeof(vmm_msr) / sizeof(vmm_msr[0]))
|
|
CTASSERT(VMM_MSR_NUM >= vmm_msr_num);
|
|
|
|
#define readonly_msr(idx) \
|
|
((vmm_msr[(idx)].flags & VMM_MSR_F_READONLY) != 0)
|
|
|
|
#define emulated_msr(idx) \
|
|
((vmm_msr[(idx)].flags & VMM_MSR_F_EMULATE) != 0)
|
|
|
|
#define invalid_msr(idx) \
|
|
((vmm_msr[(idx)].flags & VMM_MSR_F_INVALID) != 0)
|
|
|
|
void
|
|
vmm_msr_init(void)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < vmm_msr_num; i++) {
|
|
if (emulated_msr(i))
|
|
continue;
|
|
/*
|
|
* XXX this assumes that the value of the host msr does not
|
|
* change after we have cached it.
|
|
*/
|
|
vmm_msr[i].hostval = rdmsr(vmm_msr[i].num);
|
|
}
|
|
}
|
|
|
|
void
|
|
guest_msrs_init(struct vm *vm, int cpu)
|
|
{
|
|
int i;
|
|
uint64_t *guest_msrs, misc;
|
|
|
|
guest_msrs = vm_guest_msrs(vm, cpu);
|
|
|
|
for (i = 0; i < vmm_msr_num; i++) {
|
|
switch (vmm_msr[i].num) {
|
|
case MSR_LSTAR:
|
|
case MSR_CSTAR:
|
|
case MSR_STAR:
|
|
case MSR_SF_MASK:
|
|
case MSR_BIOS_SIGN:
|
|
case MSR_MCG_CAP:
|
|
guest_msrs[i] = 0;
|
|
break;
|
|
case MSR_PAT:
|
|
guest_msrs[i] = PAT_VALUE(0, PAT_WRITE_BACK) |
|
|
PAT_VALUE(1, PAT_WRITE_THROUGH) |
|
|
PAT_VALUE(2, PAT_UNCACHED) |
|
|
PAT_VALUE(3, PAT_UNCACHEABLE) |
|
|
PAT_VALUE(4, PAT_WRITE_BACK) |
|
|
PAT_VALUE(5, PAT_WRITE_THROUGH) |
|
|
PAT_VALUE(6, PAT_UNCACHED) |
|
|
PAT_VALUE(7, PAT_UNCACHEABLE);
|
|
break;
|
|
case MSR_IA32_MISC_ENABLE:
|
|
misc = rdmsr(MSR_IA32_MISC_ENABLE);
|
|
/*
|
|
* Set mandatory bits
|
|
* 11: branch trace disabled
|
|
* 12: PEBS unavailable
|
|
* Clear unsupported features
|
|
* 16: SpeedStep enable
|
|
* 18: enable MONITOR FSM
|
|
*/
|
|
misc |= (1 << 12) | (1 << 11);
|
|
misc &= ~((1 << 18) | (1 << 16));
|
|
guest_msrs[i] = misc;
|
|
break;
|
|
case MSR_IA32_PLATFORM_ID:
|
|
guest_msrs[i] = 0;
|
|
break;
|
|
default:
|
|
panic("guest_msrs_init: missing initialization for msr "
|
|
"0x%0x", vmm_msr[i].num);
|
|
}
|
|
}
|
|
}
|
|
|
|
static int
|
|
msr_num_to_idx(u_int num)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < vmm_msr_num; i++)
|
|
if (vmm_msr[i].num == num)
|
|
return (i);
|
|
|
|
return (-1);
|
|
}
|
|
|
|
int
|
|
emulate_wrmsr(struct vm *vm, int cpu, u_int num, uint64_t val, bool *retu)
|
|
{
|
|
int idx;
|
|
uint64_t *guest_msrs;
|
|
|
|
if (lapic_msr(num))
|
|
return (lapic_wrmsr(vm, cpu, num, val, retu));
|
|
|
|
idx = msr_num_to_idx(num);
|
|
if (idx < 0 || invalid_msr(idx))
|
|
return (EINVAL);
|
|
|
|
if (!readonly_msr(idx)) {
|
|
guest_msrs = vm_guest_msrs(vm, cpu);
|
|
|
|
/* Stash the value */
|
|
guest_msrs[idx] = val;
|
|
|
|
/* Update processor state for non-emulated MSRs */
|
|
if (!emulated_msr(idx))
|
|
wrmsr(vmm_msr[idx].num, val);
|
|
}
|
|
|
|
return (0);
|
|
}
|
|
|
|
int
|
|
emulate_rdmsr(struct vm *vm, int cpu, u_int num, bool *retu)
|
|
{
|
|
int error, idx;
|
|
uint32_t eax, edx;
|
|
uint64_t result, *guest_msrs;
|
|
|
|
if (lapic_msr(num)) {
|
|
error = lapic_rdmsr(vm, cpu, num, &result, retu);
|
|
goto done;
|
|
}
|
|
|
|
idx = msr_num_to_idx(num);
|
|
if (idx < 0 || invalid_msr(idx)) {
|
|
error = EINVAL;
|
|
goto done;
|
|
}
|
|
|
|
guest_msrs = vm_guest_msrs(vm, cpu);
|
|
result = guest_msrs[idx];
|
|
|
|
/*
|
|
* If this is not an emulated msr register make sure that the processor
|
|
* state matches our cached state.
|
|
*/
|
|
if (!emulated_msr(idx) && (rdmsr(num) != result)) {
|
|
panic("emulate_rdmsr: msr 0x%0x has inconsistent cached "
|
|
"(0x%016lx) and actual (0x%016lx) values", num,
|
|
result, rdmsr(num));
|
|
}
|
|
|
|
error = 0;
|
|
|
|
done:
|
|
if (error == 0) {
|
|
eax = result;
|
|
edx = result >> 32;
|
|
error = vm_set_register(vm, cpu, VM_REG_GUEST_RAX, eax);
|
|
if (error)
|
|
panic("vm_set_register(rax) error %d", error);
|
|
error = vm_set_register(vm, cpu, VM_REG_GUEST_RDX, edx);
|
|
if (error)
|
|
panic("vm_set_register(rdx) error %d", error);
|
|
}
|
|
return (error);
|
|
}
|
|
|
|
void
|
|
restore_guest_msrs(struct vm *vm, int cpu)
|
|
{
|
|
int i;
|
|
uint64_t *guest_msrs;
|
|
|
|
guest_msrs = vm_guest_msrs(vm, cpu);
|
|
|
|
for (i = 0; i < vmm_msr_num; i++) {
|
|
if (emulated_msr(i))
|
|
continue;
|
|
else
|
|
wrmsr(vmm_msr[i].num, guest_msrs[i]);
|
|
}
|
|
}
|
|
|
|
void
|
|
restore_host_msrs(struct vm *vm, int cpu)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < vmm_msr_num; i++) {
|
|
if (emulated_msr(i))
|
|
continue;
|
|
else
|
|
wrmsr(vmm_msr[i].num, vmm_msr[i].hostval);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Must be called by the CPU-specific code before any guests are
|
|
* created
|
|
*/
|
|
void
|
|
guest_msr_valid(int msr)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < vmm_msr_num; i++) {
|
|
if (vmm_msr[i].num == msr && invalid_msr(i)) {
|
|
vmm_msr[i].flags &= ~VMM_MSR_F_INVALID;
|
|
}
|
|
}
|
|
}
|