kvm: x86: Drop KVM_CAP build dependencies

No longer needed with accompanied kernel headers.

Signed-off-by: Jan Kiszka <jan.kiszka@siemens.com>
Signed-off-by: Marcelo Tosatti <mtosatti@redhat.com>
This commit is contained in:
Jan Kiszka 2011-06-08 16:11:02 +02:00 committed by Marcelo Tosatti
parent ba4047cf84
commit e56ff191b6
1 changed files with 3 additions and 64 deletions

View File

@ -94,9 +94,7 @@ struct kvm_para_features {
{ KVM_CAP_CLOCKSOURCE, KVM_FEATURE_CLOCKSOURCE }, { KVM_CAP_CLOCKSOURCE, KVM_FEATURE_CLOCKSOURCE },
{ KVM_CAP_NOP_IO_DELAY, KVM_FEATURE_NOP_IO_DELAY }, { KVM_CAP_NOP_IO_DELAY, KVM_FEATURE_NOP_IO_DELAY },
{ KVM_CAP_PV_MMU, KVM_FEATURE_MMU_OP }, { KVM_CAP_PV_MMU, KVM_FEATURE_MMU_OP },
#ifdef KVM_CAP_ASYNC_PF
{ KVM_CAP_ASYNC_PF, KVM_FEATURE_ASYNC_PF }, { KVM_CAP_ASYNC_PF, KVM_FEATURE_ASYNC_PF },
#endif
{ -1, -1 } { -1, -1 }
}; };
@ -193,7 +191,6 @@ static void kvm_unpoison_all(void *param)
} }
} }
#ifdef KVM_CAP_MCE
static void kvm_hwpoison_page_add(ram_addr_t ram_addr) static void kvm_hwpoison_page_add(ram_addr_t ram_addr)
{ {
HWPoisonPage *page; HWPoisonPage *page;
@ -239,7 +236,6 @@ static void kvm_mce_inject(CPUState *env, target_phys_addr_t paddr, int code)
cpu_x86_support_mca_broadcast(env) ? cpu_x86_support_mca_broadcast(env) ?
MCE_INJECT_BROADCAST : 0); MCE_INJECT_BROADCAST : 0);
} }
#endif /* KVM_CAP_MCE */
static void hardware_memory_error(void) static void hardware_memory_error(void)
{ {
@ -249,7 +245,6 @@ static void hardware_memory_error(void)
int kvm_arch_on_sigbus_vcpu(CPUState *env, int code, void *addr) int kvm_arch_on_sigbus_vcpu(CPUState *env, int code, void *addr)
{ {
#ifdef KVM_CAP_MCE
ram_addr_t ram_addr; ram_addr_t ram_addr;
target_phys_addr_t paddr; target_phys_addr_t paddr;
@ -269,9 +264,7 @@ int kvm_arch_on_sigbus_vcpu(CPUState *env, int code, void *addr)
} }
kvm_hwpoison_page_add(ram_addr); kvm_hwpoison_page_add(ram_addr);
kvm_mce_inject(env, paddr, code); kvm_mce_inject(env, paddr, code);
} else } else {
#endif /* KVM_CAP_MCE */
{
if (code == BUS_MCEERR_AO) { if (code == BUS_MCEERR_AO) {
return 0; return 0;
} else if (code == BUS_MCEERR_AR) { } else if (code == BUS_MCEERR_AR) {
@ -285,7 +278,6 @@ int kvm_arch_on_sigbus_vcpu(CPUState *env, int code, void *addr)
int kvm_arch_on_sigbus(int code, void *addr) int kvm_arch_on_sigbus(int code, void *addr)
{ {
#ifdef KVM_CAP_MCE
if ((first_cpu->mcg_cap & MCG_SER_P) && addr && code == BUS_MCEERR_AO) { if ((first_cpu->mcg_cap & MCG_SER_P) && addr && code == BUS_MCEERR_AO) {
ram_addr_t ram_addr; ram_addr_t ram_addr;
target_phys_addr_t paddr; target_phys_addr_t paddr;
@ -300,9 +292,7 @@ int kvm_arch_on_sigbus(int code, void *addr)
} }
kvm_hwpoison_page_add(ram_addr); kvm_hwpoison_page_add(ram_addr);
kvm_mce_inject(first_cpu, paddr, code); kvm_mce_inject(first_cpu, paddr, code);
} else } else {
#endif /* KVM_CAP_MCE */
{
if (code == BUS_MCEERR_AO) { if (code == BUS_MCEERR_AO) {
return 0; return 0;
} else if (code == BUS_MCEERR_AR) { } else if (code == BUS_MCEERR_AR) {
@ -316,7 +306,6 @@ int kvm_arch_on_sigbus(int code, void *addr)
static int kvm_inject_mce_oldstyle(CPUState *env) static int kvm_inject_mce_oldstyle(CPUState *env)
{ {
#ifdef KVM_CAP_MCE
if (!kvm_has_vcpu_events() && env->exception_injected == EXCP12_MCHK) { if (!kvm_has_vcpu_events() && env->exception_injected == EXCP12_MCHK) {
unsigned int bank, bank_num = env->mcg_cap & 0xff; unsigned int bank, bank_num = env->mcg_cap & 0xff;
struct kvm_x86_mce mce; struct kvm_x86_mce mce;
@ -342,7 +331,6 @@ static int kvm_inject_mce_oldstyle(CPUState *env)
return kvm_vcpu_ioctl(env, KVM_X86_SET_MCE, &mce); return kvm_vcpu_ioctl(env, KVM_X86_SET_MCE, &mce);
} }
#endif /* KVM_CAP_MCE */
return 0; return 0;
} }
@ -398,9 +386,7 @@ int kvm_arch_init_vcpu(CPUState *env)
c->eax = env->cpuid_kvm_features & kvm_arch_get_supported_cpuid(env, c->eax = env->cpuid_kvm_features & kvm_arch_get_supported_cpuid(env,
KVM_CPUID_FEATURES, 0, R_EAX); KVM_CPUID_FEATURES, 0, R_EAX);
#ifdef KVM_CAP_ASYNC_PF
has_msr_async_pf_en = c->eax & (1 << KVM_FEATURE_ASYNC_PF); has_msr_async_pf_en = c->eax & (1 << KVM_FEATURE_ASYNC_PF);
#endif
cpu_x86_cpuid(env, 0, 0, &limit, &unused, &unused, &unused); cpu_x86_cpuid(env, 0, 0, &limit, &unused, &unused, &unused);
@ -481,7 +467,6 @@ int kvm_arch_init_vcpu(CPUState *env)
cpuid_data.cpuid.nent = cpuid_i; cpuid_data.cpuid.nent = cpuid_i;
#ifdef KVM_CAP_MCE
if (((env->cpuid_version >> 8)&0xF) >= 6 if (((env->cpuid_version >> 8)&0xF) >= 6
&& (env->cpuid_features&(CPUID_MCE|CPUID_MCA)) == (CPUID_MCE|CPUID_MCA) && (env->cpuid_features&(CPUID_MCE|CPUID_MCA)) == (CPUID_MCE|CPUID_MCA)
&& kvm_check_extension(env->kvm_state, KVM_CAP_MCE) > 0) { && kvm_check_extension(env->kvm_state, KVM_CAP_MCE) > 0) {
@ -508,7 +493,6 @@ int kvm_arch_init_vcpu(CPUState *env)
env->mcg_cap = mcg_cap; env->mcg_cap = mcg_cap;
} }
#endif
qemu_add_vm_change_state_handler(cpu_update_state, env); qemu_add_vm_change_state_handler(cpu_update_state, env);
@ -600,7 +584,6 @@ int kvm_arch_init(KVMState *s)
* that case we need to stick with the default, i.e. a 256K maximum BIOS * that case we need to stick with the default, i.e. a 256K maximum BIOS
* size. * size.
*/ */
#ifdef KVM_CAP_SET_IDENTITY_MAP_ADDR
if (kvm_check_extension(s, KVM_CAP_SET_IDENTITY_MAP_ADDR)) { if (kvm_check_extension(s, KVM_CAP_SET_IDENTITY_MAP_ADDR)) {
/* Allows up to 16M BIOSes. */ /* Allows up to 16M BIOSes. */
identity_base = 0xfeffc000; identity_base = 0xfeffc000;
@ -610,7 +593,7 @@ int kvm_arch_init(KVMState *s)
return ret; return ret;
} }
} }
#endif
/* Set TSS base one page after EPT identity map. */ /* Set TSS base one page after EPT identity map. */
ret = kvm_vm_ioctl(s, KVM_SET_TSS_ADDR, identity_base + 0x1000); ret = kvm_vm_ioctl(s, KVM_SET_TSS_ADDR, identity_base + 0x1000);
if (ret < 0) { if (ret < 0) {
@ -748,7 +731,6 @@ static int kvm_put_fpu(CPUState *env)
return kvm_vcpu_ioctl(env, KVM_SET_FPU, &fpu); return kvm_vcpu_ioctl(env, KVM_SET_FPU, &fpu);
} }
#ifdef KVM_CAP_XSAVE
#define XSAVE_CWD_RIP 2 #define XSAVE_CWD_RIP 2
#define XSAVE_CWD_RDP 4 #define XSAVE_CWD_RDP 4
#define XSAVE_MXCSR 6 #define XSAVE_MXCSR 6
@ -756,11 +738,9 @@ static int kvm_put_fpu(CPUState *env)
#define XSAVE_XMM_SPACE 40 #define XSAVE_XMM_SPACE 40
#define XSAVE_XSTATE_BV 128 #define XSAVE_XSTATE_BV 128
#define XSAVE_YMMH_SPACE 144 #define XSAVE_YMMH_SPACE 144
#endif
static int kvm_put_xsave(CPUState *env) static int kvm_put_xsave(CPUState *env)
{ {
#ifdef KVM_CAP_XSAVE
int i, r; int i, r;
struct kvm_xsave* xsave; struct kvm_xsave* xsave;
uint16_t cwd, swd, twd; uint16_t cwd, swd, twd;
@ -793,14 +773,10 @@ static int kvm_put_xsave(CPUState *env)
r = kvm_vcpu_ioctl(env, KVM_SET_XSAVE, xsave); r = kvm_vcpu_ioctl(env, KVM_SET_XSAVE, xsave);
qemu_free(xsave); qemu_free(xsave);
return r; return r;
#else
return kvm_put_fpu(env);
#endif
} }
static int kvm_put_xcrs(CPUState *env) static int kvm_put_xcrs(CPUState *env)
{ {
#ifdef KVM_CAP_XCRS
struct kvm_xcrs xcrs; struct kvm_xcrs xcrs;
if (!kvm_has_xcrs()) { if (!kvm_has_xcrs()) {
@ -812,9 +788,6 @@ static int kvm_put_xcrs(CPUState *env)
xcrs.xcrs[0].xcr = 0; xcrs.xcrs[0].xcr = 0;
xcrs.xcrs[0].value = env->xcr0; xcrs.xcrs[0].value = env->xcr0;
return kvm_vcpu_ioctl(env, KVM_SET_XCRS, &xcrs); return kvm_vcpu_ioctl(env, KVM_SET_XCRS, &xcrs);
#else
return 0;
#endif
} }
static int kvm_put_sregs(CPUState *env) static int kvm_put_sregs(CPUState *env)
@ -923,7 +896,6 @@ static int kvm_put_msrs(CPUState *env, int level)
env->async_pf_en_msr); env->async_pf_en_msr);
} }
} }
#ifdef KVM_CAP_MCE
if (env->mcg_cap) { if (env->mcg_cap) {
int i; int i;
@ -933,7 +905,6 @@ static int kvm_put_msrs(CPUState *env, int level)
kvm_msr_entry_set(&msrs[n++], MSR_MC0_CTL + i, env->mce_banks[i]); kvm_msr_entry_set(&msrs[n++], MSR_MC0_CTL + i, env->mce_banks[i]);
} }
} }
#endif
msr_data.info.nmsrs = n; msr_data.info.nmsrs = n;
@ -970,7 +941,6 @@ static int kvm_get_fpu(CPUState *env)
static int kvm_get_xsave(CPUState *env) static int kvm_get_xsave(CPUState *env)
{ {
#ifdef KVM_CAP_XSAVE
struct kvm_xsave* xsave; struct kvm_xsave* xsave;
int ret, i; int ret, i;
uint16_t cwd, swd, twd; uint16_t cwd, swd, twd;
@ -1008,14 +978,10 @@ static int kvm_get_xsave(CPUState *env)
sizeof env->ymmh_regs); sizeof env->ymmh_regs);
qemu_free(xsave); qemu_free(xsave);
return 0; return 0;
#else
return kvm_get_fpu(env);
#endif
} }
static int kvm_get_xcrs(CPUState *env) static int kvm_get_xcrs(CPUState *env)
{ {
#ifdef KVM_CAP_XCRS
int i, ret; int i, ret;
struct kvm_xcrs xcrs; struct kvm_xcrs xcrs;
@ -1036,9 +1002,6 @@ static int kvm_get_xcrs(CPUState *env)
} }
} }
return 0; return 0;
#else
return 0;
#endif
} }
static int kvm_get_sregs(CPUState *env) static int kvm_get_sregs(CPUState *env)
@ -1166,7 +1129,6 @@ static int kvm_get_msrs(CPUState *env)
msrs[n++].index = MSR_KVM_ASYNC_PF_EN; msrs[n++].index = MSR_KVM_ASYNC_PF_EN;
} }
#ifdef KVM_CAP_MCE
if (env->mcg_cap) { if (env->mcg_cap) {
msrs[n++].index = MSR_MCG_STATUS; msrs[n++].index = MSR_MCG_STATUS;
msrs[n++].index = MSR_MCG_CTL; msrs[n++].index = MSR_MCG_CTL;
@ -1174,7 +1136,6 @@ static int kvm_get_msrs(CPUState *env)
msrs[n++].index = MSR_MC0_CTL + i; msrs[n++].index = MSR_MC0_CTL + i;
} }
} }
#endif
msr_data.info.nmsrs = n; msr_data.info.nmsrs = n;
ret = kvm_vcpu_ioctl(env, KVM_GET_MSRS, &msr_data); ret = kvm_vcpu_ioctl(env, KVM_GET_MSRS, &msr_data);
@ -1225,21 +1186,17 @@ static int kvm_get_msrs(CPUState *env)
case MSR_KVM_WALL_CLOCK: case MSR_KVM_WALL_CLOCK:
env->wall_clock_msr = msrs[i].data; env->wall_clock_msr = msrs[i].data;
break; break;
#ifdef KVM_CAP_MCE
case MSR_MCG_STATUS: case MSR_MCG_STATUS:
env->mcg_status = msrs[i].data; env->mcg_status = msrs[i].data;
break; break;
case MSR_MCG_CTL: case MSR_MCG_CTL:
env->mcg_ctl = msrs[i].data; env->mcg_ctl = msrs[i].data;
break; break;
#endif
default: default:
#ifdef KVM_CAP_MCE
if (msrs[i].index >= MSR_MC0_CTL && if (msrs[i].index >= MSR_MC0_CTL &&
msrs[i].index < MSR_MC0_CTL + (env->mcg_cap & 0xff) * 4) { msrs[i].index < MSR_MC0_CTL + (env->mcg_cap & 0xff) * 4) {
env->mce_banks[msrs[i].index - MSR_MC0_CTL] = msrs[i].data; env->mce_banks[msrs[i].index - MSR_MC0_CTL] = msrs[i].data;
} }
#endif
break; break;
case MSR_KVM_ASYNC_PF_EN: case MSR_KVM_ASYNC_PF_EN:
env->async_pf_en_msr = msrs[i].data; env->async_pf_en_msr = msrs[i].data;
@ -1275,7 +1232,6 @@ static int kvm_get_mp_state(CPUState *env)
static int kvm_put_vcpu_events(CPUState *env, int level) static int kvm_put_vcpu_events(CPUState *env, int level)
{ {
#ifdef KVM_CAP_VCPU_EVENTS
struct kvm_vcpu_events events; struct kvm_vcpu_events events;
if (!kvm_has_vcpu_events()) { if (!kvm_has_vcpu_events()) {
@ -1304,14 +1260,10 @@ static int kvm_put_vcpu_events(CPUState *env, int level)
} }
return kvm_vcpu_ioctl(env, KVM_SET_VCPU_EVENTS, &events); return kvm_vcpu_ioctl(env, KVM_SET_VCPU_EVENTS, &events);
#else
return 0;
#endif
} }
static int kvm_get_vcpu_events(CPUState *env) static int kvm_get_vcpu_events(CPUState *env)
{ {
#ifdef KVM_CAP_VCPU_EVENTS
struct kvm_vcpu_events events; struct kvm_vcpu_events events;
int ret; int ret;
@ -1341,7 +1293,6 @@ static int kvm_get_vcpu_events(CPUState *env)
} }
env->sipi_vector = events.sipi_vector; env->sipi_vector = events.sipi_vector;
#endif
return 0; return 0;
} }
@ -1349,7 +1300,6 @@ static int kvm_get_vcpu_events(CPUState *env)
static int kvm_guest_debug_workarounds(CPUState *env) static int kvm_guest_debug_workarounds(CPUState *env)
{ {
int ret = 0; int ret = 0;
#ifdef KVM_CAP_SET_GUEST_DEBUG
unsigned long reinject_trap = 0; unsigned long reinject_trap = 0;
if (!kvm_has_vcpu_events()) { if (!kvm_has_vcpu_events()) {
@ -1373,13 +1323,11 @@ static int kvm_guest_debug_workarounds(CPUState *env)
(!kvm_has_robust_singlestep() && env->singlestep_enabled)) { (!kvm_has_robust_singlestep() && env->singlestep_enabled)) {
ret = kvm_update_guest_debug(env, reinject_trap); ret = kvm_update_guest_debug(env, reinject_trap);
} }
#endif /* KVM_CAP_SET_GUEST_DEBUG */
return ret; return ret;
} }
static int kvm_put_debugregs(CPUState *env) static int kvm_put_debugregs(CPUState *env)
{ {
#ifdef KVM_CAP_DEBUGREGS
struct kvm_debugregs dbgregs; struct kvm_debugregs dbgregs;
int i; int i;
@ -1395,14 +1343,10 @@ static int kvm_put_debugregs(CPUState *env)
dbgregs.flags = 0; dbgregs.flags = 0;
return kvm_vcpu_ioctl(env, KVM_SET_DEBUGREGS, &dbgregs); return kvm_vcpu_ioctl(env, KVM_SET_DEBUGREGS, &dbgregs);
#else
return 0;
#endif
} }
static int kvm_get_debugregs(CPUState *env) static int kvm_get_debugregs(CPUState *env)
{ {
#ifdef KVM_CAP_DEBUGREGS
struct kvm_debugregs dbgregs; struct kvm_debugregs dbgregs;
int i, ret; int i, ret;
@ -1419,7 +1363,6 @@ static int kvm_get_debugregs(CPUState *env)
} }
env->dr[4] = env->dr[6] = dbgregs.dr6; env->dr[4] = env->dr[6] = dbgregs.dr6;
env->dr[5] = env->dr[7] = dbgregs.dr7; env->dr[5] = env->dr[7] = dbgregs.dr7;
#endif
return 0; return 0;
} }
@ -1645,7 +1588,6 @@ static int kvm_handle_halt(CPUState *env)
return 0; return 0;
} }
#ifdef KVM_CAP_SET_GUEST_DEBUG
int kvm_arch_insert_sw_breakpoint(CPUState *env, struct kvm_sw_breakpoint *bp) int kvm_arch_insert_sw_breakpoint(CPUState *env, struct kvm_sw_breakpoint *bp)
{ {
static const uint8_t int3 = 0xcc; static const uint8_t int3 = 0xcc;
@ -1826,7 +1768,6 @@ void kvm_arch_update_guest_debug(CPUState *env, struct kvm_guest_debug *dbg)
} }
} }
} }
#endif /* KVM_CAP_SET_GUEST_DEBUG */
static bool host_supports_vmx(void) static bool host_supports_vmx(void)
{ {
@ -1873,12 +1814,10 @@ int kvm_arch_handle_exit(CPUState *env, struct kvm_run *run)
run->ex.exception, run->ex.error_code); run->ex.exception, run->ex.error_code);
ret = -1; ret = -1;
break; break;
#ifdef KVM_CAP_SET_GUEST_DEBUG
case KVM_EXIT_DEBUG: case KVM_EXIT_DEBUG:
DPRINTF("kvm_exit_debug\n"); DPRINTF("kvm_exit_debug\n");
ret = kvm_handle_debug(&run->debug.arch); ret = kvm_handle_debug(&run->debug.arch);
break; break;
#endif /* KVM_CAP_SET_GUEST_DEBUG */
default: default:
fprintf(stderr, "KVM: unknown exit reason %d\n", run->exit_reason); fprintf(stderr, "KVM: unknown exit reason %d\n", run->exit_reason);
ret = -1; ret = -1;