Merge branch 'x86-platform-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip
Pull x86 platform updates from Ingo Molnar: "The main changes in this cycle were: - a refactoring of the early virt init code by merging 'struct x86_hyper' into 'struct x86_platform' and 'struct x86_init', which allows simplifications and also the addition of a new ->guest_late_init() callback. (Juergen Gross) - timer_setup() conversion of the UV code (Kees Cook)" * 'x86-platform-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip: x86/virt/xen: Use guest_late_init to detect Xen PVH guest x86/virt, x86/platform: Add ->guest_late_init() callback to hypervisor_x86 structure x86/virt, x86/acpi: Add test for ACPI_FADT_NO_VGA x86/virt: Add enum for hypervisors to replace x86_hyper x86/virt, x86/platform: Merge 'struct x86_hyper' into 'struct x86_platform' and 'struct x86_init' x86/platform/UV: Convert timers to use timer_setup()
This commit is contained in:
commit
43ff2f4db9
|
@ -113,7 +113,7 @@ void hyperv_init(void)
|
|||
u64 guest_id;
|
||||
union hv_x64_msr_hypercall_contents hypercall_msr;
|
||||
|
||||
if (x86_hyper != &x86_hyper_ms_hyperv)
|
||||
if (x86_hyper_type != X86_HYPER_MS_HYPERV)
|
||||
return;
|
||||
|
||||
/* Allocate percpu VP index */
|
||||
|
|
|
@ -23,11 +23,22 @@
|
|||
#ifdef CONFIG_HYPERVISOR_GUEST
|
||||
|
||||
#include <asm/kvm_para.h>
|
||||
#include <asm/x86_init.h>
|
||||
#include <asm/xen/hypervisor.h>
|
||||
|
||||
/*
|
||||
* x86 hypervisor information
|
||||
*/
|
||||
|
||||
enum x86_hypervisor_type {
|
||||
X86_HYPER_NATIVE = 0,
|
||||
X86_HYPER_VMWARE,
|
||||
X86_HYPER_MS_HYPERV,
|
||||
X86_HYPER_XEN_PV,
|
||||
X86_HYPER_XEN_HVM,
|
||||
X86_HYPER_KVM,
|
||||
};
|
||||
|
||||
struct hypervisor_x86 {
|
||||
/* Hypervisor name */
|
||||
const char *name;
|
||||
|
@ -35,40 +46,19 @@ struct hypervisor_x86 {
|
|||
/* Detection routine */
|
||||
uint32_t (*detect)(void);
|
||||
|
||||
/* Platform setup (run once per boot) */
|
||||
void (*init_platform)(void);
|
||||
/* Hypervisor type */
|
||||
enum x86_hypervisor_type type;
|
||||
|
||||
/* X2APIC detection (run once per boot) */
|
||||
bool (*x2apic_available)(void);
|
||||
/* init time callbacks */
|
||||
struct x86_hyper_init init;
|
||||
|
||||
/* pin current vcpu to specified physical cpu (run rarely) */
|
||||
void (*pin_vcpu)(int);
|
||||
|
||||
/* called during init_mem_mapping() to setup early mappings. */
|
||||
void (*init_mem_mapping)(void);
|
||||
/* runtime callbacks */
|
||||
struct x86_hyper_runtime runtime;
|
||||
};
|
||||
|
||||
extern const struct hypervisor_x86 *x86_hyper;
|
||||
|
||||
/* Recognized hypervisors */
|
||||
extern const struct hypervisor_x86 x86_hyper_vmware;
|
||||
extern const struct hypervisor_x86 x86_hyper_ms_hyperv;
|
||||
extern const struct hypervisor_x86 x86_hyper_xen_pv;
|
||||
extern const struct hypervisor_x86 x86_hyper_xen_hvm;
|
||||
extern const struct hypervisor_x86 x86_hyper_kvm;
|
||||
|
||||
extern enum x86_hypervisor_type x86_hyper_type;
|
||||
extern void init_hypervisor_platform(void);
|
||||
extern bool hypervisor_x2apic_available(void);
|
||||
extern void hypervisor_pin_vcpu(int cpu);
|
||||
|
||||
static inline void hypervisor_init_mem_mapping(void)
|
||||
{
|
||||
if (x86_hyper && x86_hyper->init_mem_mapping)
|
||||
x86_hyper->init_mem_mapping();
|
||||
}
|
||||
#else
|
||||
static inline void init_hypervisor_platform(void) { }
|
||||
static inline bool hypervisor_x2apic_available(void) { return false; }
|
||||
static inline void hypervisor_init_mem_mapping(void) { }
|
||||
#endif /* CONFIG_HYPERVISOR_GUEST */
|
||||
#endif /* _ASM_X86_HYPERVISOR_H */
|
||||
|
|
|
@ -88,7 +88,6 @@ static inline long kvm_hypercall4(unsigned int nr, unsigned long p1,
|
|||
#ifdef CONFIG_KVM_GUEST
|
||||
bool kvm_para_available(void);
|
||||
unsigned int kvm_arch_para_features(void);
|
||||
void __init kvm_guest_init(void);
|
||||
void kvm_async_pf_task_wait(u32 token, int interrupt_kernel);
|
||||
void kvm_async_pf_task_wake(u32 token);
|
||||
u32 kvm_read_and_reset_pf_reason(void);
|
||||
|
@ -103,7 +102,6 @@ static inline void kvm_spinlock_init(void)
|
|||
#endif /* CONFIG_PARAVIRT_SPINLOCKS */
|
||||
|
||||
#else /* CONFIG_KVM_GUEST */
|
||||
#define kvm_guest_init() do {} while (0)
|
||||
#define kvm_async_pf_task_wait(T, I) do {} while(0)
|
||||
#define kvm_async_pf_task_wake(T) do {} while(0)
|
||||
|
||||
|
|
|
@ -114,6 +114,20 @@ struct x86_init_pci {
|
|||
void (*fixup_irqs)(void);
|
||||
};
|
||||
|
||||
/**
|
||||
* struct x86_hyper_init - x86 hypervisor init functions
|
||||
* @init_platform: platform setup
|
||||
* @guest_late_init: guest late init
|
||||
* @x2apic_available: X2APIC detection
|
||||
* @init_mem_mapping: setup early mappings during init_mem_mapping()
|
||||
*/
|
||||
struct x86_hyper_init {
|
||||
void (*init_platform)(void);
|
||||
void (*guest_late_init)(void);
|
||||
bool (*x2apic_available)(void);
|
||||
void (*init_mem_mapping)(void);
|
||||
};
|
||||
|
||||
/**
|
||||
* struct x86_init_ops - functions for platform specific setup
|
||||
*
|
||||
|
@ -127,6 +141,7 @@ struct x86_init_ops {
|
|||
struct x86_init_timers timers;
|
||||
struct x86_init_iommu iommu;
|
||||
struct x86_init_pci pci;
|
||||
struct x86_hyper_init hyper;
|
||||
};
|
||||
|
||||
/**
|
||||
|
@ -195,10 +210,20 @@ enum x86_legacy_i8042_state {
|
|||
struct x86_legacy_features {
|
||||
enum x86_legacy_i8042_state i8042;
|
||||
int rtc;
|
||||
int no_vga;
|
||||
int reserve_bios_regions;
|
||||
struct x86_legacy_devices devices;
|
||||
};
|
||||
|
||||
/**
|
||||
* struct x86_hyper_runtime - x86 hypervisor specific runtime callbacks
|
||||
*
|
||||
* @pin_vcpu: pin current vcpu to specified physical cpu (run rarely)
|
||||
*/
|
||||
struct x86_hyper_runtime {
|
||||
void (*pin_vcpu)(int cpu);
|
||||
};
|
||||
|
||||
/**
|
||||
* struct x86_platform_ops - platform specific runtime functions
|
||||
* @calibrate_cpu: calibrate CPU
|
||||
|
@ -218,6 +243,7 @@ struct x86_legacy_features {
|
|||
* possible in x86_early_init_platform_quirks() by
|
||||
* only using the current x86_hardware_subarch
|
||||
* semantics.
|
||||
* @hyper: x86 hypervisor specific runtime callbacks
|
||||
*/
|
||||
struct x86_platform_ops {
|
||||
unsigned long (*calibrate_cpu)(void);
|
||||
|
@ -233,6 +259,7 @@ struct x86_platform_ops {
|
|||
void (*apic_post_init)(void);
|
||||
struct x86_legacy_features legacy;
|
||||
void (*set_legacy_features)(void);
|
||||
struct x86_hyper_runtime hyper;
|
||||
};
|
||||
|
||||
struct pci_dev;
|
||||
|
|
|
@ -961,6 +961,11 @@ static int __init acpi_parse_fadt(struct acpi_table_header *table)
|
|||
x86_platform.legacy.rtc = 0;
|
||||
}
|
||||
|
||||
if (acpi_gbl_FADT.boot_flags & ACPI_FADT_NO_VGA) {
|
||||
pr_debug("ACPI: probing for VGA not safe\n");
|
||||
x86_platform.legacy.no_vga = 1;
|
||||
}
|
||||
|
||||
#ifdef CONFIG_X86_PM_TIMER
|
||||
/* detect the location of the ACPI PM Timer */
|
||||
if (acpi_gbl_FADT.header.revision >= FADT2_REVISION_ID) {
|
||||
|
|
|
@ -1645,7 +1645,7 @@ static __init void try_to_enable_x2apic(int remap_mode)
|
|||
* under KVM
|
||||
*/
|
||||
if (max_physical_apicid > 255 ||
|
||||
!hypervisor_x2apic_available()) {
|
||||
!x86_init.hyper.x2apic_available()) {
|
||||
pr_info("x2apic: IRQ remapping doesn't support X2APIC mode\n");
|
||||
x2apic_disable();
|
||||
return;
|
||||
|
|
|
@ -920,9 +920,8 @@ static __init void uv_rtc_init(void)
|
|||
/*
|
||||
* percpu heartbeat timer
|
||||
*/
|
||||
static void uv_heartbeat(unsigned long ignored)
|
||||
static void uv_heartbeat(struct timer_list *timer)
|
||||
{
|
||||
struct timer_list *timer = &uv_scir_info->timer;
|
||||
unsigned char bits = uv_scir_info->state;
|
||||
|
||||
/* Flip heartbeat bit: */
|
||||
|
@ -947,7 +946,7 @@ static int uv_heartbeat_enable(unsigned int cpu)
|
|||
struct timer_list *timer = &uv_cpu_scir_info(cpu)->timer;
|
||||
|
||||
uv_set_cpu_scir_bits(cpu, SCIR_CPU_HEARTBEAT|SCIR_CPU_ACTIVITY);
|
||||
setup_pinned_timer(timer, uv_heartbeat, cpu);
|
||||
timer_setup(timer, uv_heartbeat, TIMER_PINNED);
|
||||
timer->expires = jiffies + SCIR_CPU_HB_INTERVAL;
|
||||
add_timer_on(timer, cpu);
|
||||
uv_cpu_scir_info(cpu)->enabled = 1;
|
||||
|
|
|
@ -26,6 +26,12 @@
|
|||
#include <asm/processor.h>
|
||||
#include <asm/hypervisor.h>
|
||||
|
||||
extern const struct hypervisor_x86 x86_hyper_vmware;
|
||||
extern const struct hypervisor_x86 x86_hyper_ms_hyperv;
|
||||
extern const struct hypervisor_x86 x86_hyper_xen_pv;
|
||||
extern const struct hypervisor_x86 x86_hyper_xen_hvm;
|
||||
extern const struct hypervisor_x86 x86_hyper_kvm;
|
||||
|
||||
static const __initconst struct hypervisor_x86 * const hypervisors[] =
|
||||
{
|
||||
#ifdef CONFIG_XEN_PV
|
||||
|
@ -41,54 +47,52 @@ static const __initconst struct hypervisor_x86 * const hypervisors[] =
|
|||
#endif
|
||||
};
|
||||
|
||||
const struct hypervisor_x86 *x86_hyper;
|
||||
EXPORT_SYMBOL(x86_hyper);
|
||||
enum x86_hypervisor_type x86_hyper_type;
|
||||
EXPORT_SYMBOL(x86_hyper_type);
|
||||
|
||||
static inline void __init
|
||||
static inline const struct hypervisor_x86 * __init
|
||||
detect_hypervisor_vendor(void)
|
||||
{
|
||||
const struct hypervisor_x86 *h, * const *p;
|
||||
const struct hypervisor_x86 *h = NULL, * const *p;
|
||||
uint32_t pri, max_pri = 0;
|
||||
|
||||
for (p = hypervisors; p < hypervisors + ARRAY_SIZE(hypervisors); p++) {
|
||||
h = *p;
|
||||
pri = h->detect();
|
||||
if (pri != 0 && pri > max_pri) {
|
||||
pri = (*p)->detect();
|
||||
if (pri > max_pri) {
|
||||
max_pri = pri;
|
||||
x86_hyper = h;
|
||||
h = *p;
|
||||
}
|
||||
}
|
||||
|
||||
if (max_pri)
|
||||
pr_info("Hypervisor detected: %s\n", x86_hyper->name);
|
||||
if (h)
|
||||
pr_info("Hypervisor detected: %s\n", h->name);
|
||||
|
||||
return h;
|
||||
}
|
||||
|
||||
static void __init copy_array(const void *src, void *target, unsigned int size)
|
||||
{
|
||||
unsigned int i, n = size / sizeof(void *);
|
||||
const void * const *from = (const void * const *)src;
|
||||
const void **to = (const void **)target;
|
||||
|
||||
for (i = 0; i < n; i++)
|
||||
if (from[i])
|
||||
to[i] = from[i];
|
||||
}
|
||||
|
||||
void __init init_hypervisor_platform(void)
|
||||
{
|
||||
const struct hypervisor_x86 *h;
|
||||
|
||||
detect_hypervisor_vendor();
|
||||
h = detect_hypervisor_vendor();
|
||||
|
||||
if (!x86_hyper)
|
||||
if (!h)
|
||||
return;
|
||||
|
||||
if (x86_hyper->init_platform)
|
||||
x86_hyper->init_platform();
|
||||
}
|
||||
copy_array(&h->init, &x86_init.hyper, sizeof(h->init));
|
||||
copy_array(&h->runtime, &x86_platform.hyper, sizeof(h->runtime));
|
||||
|
||||
bool __init hypervisor_x2apic_available(void)
|
||||
{
|
||||
return x86_hyper &&
|
||||
x86_hyper->x2apic_available &&
|
||||
x86_hyper->x2apic_available();
|
||||
}
|
||||
|
||||
void hypervisor_pin_vcpu(int cpu)
|
||||
{
|
||||
if (!x86_hyper)
|
||||
return;
|
||||
|
||||
if (x86_hyper->pin_vcpu)
|
||||
x86_hyper->pin_vcpu(cpu);
|
||||
else
|
||||
WARN_ONCE(1, "vcpu pinning requested but not supported!\n");
|
||||
x86_hyper_type = h->type;
|
||||
x86_init.hyper.init_platform();
|
||||
}
|
||||
|
|
|
@ -254,9 +254,9 @@ static void __init ms_hyperv_init_platform(void)
|
|||
#endif
|
||||
}
|
||||
|
||||
const __refconst struct hypervisor_x86 x86_hyper_ms_hyperv = {
|
||||
const __initconst struct hypervisor_x86 x86_hyper_ms_hyperv = {
|
||||
.name = "Microsoft Hyper-V",
|
||||
.detect = ms_hyperv_platform,
|
||||
.init_platform = ms_hyperv_init_platform,
|
||||
.type = X86_HYPER_MS_HYPERV,
|
||||
.init.init_platform = ms_hyperv_init_platform,
|
||||
};
|
||||
EXPORT_SYMBOL(x86_hyper_ms_hyperv);
|
||||
|
|
|
@ -205,10 +205,10 @@ static bool __init vmware_legacy_x2apic_available(void)
|
|||
(eax & (1 << VMWARE_PORT_CMD_LEGACY_X2APIC)) != 0;
|
||||
}
|
||||
|
||||
const __refconst struct hypervisor_x86 x86_hyper_vmware = {
|
||||
const __initconst struct hypervisor_x86 x86_hyper_vmware = {
|
||||
.name = "VMware",
|
||||
.detect = vmware_platform,
|
||||
.init_platform = vmware_platform_setup,
|
||||
.x2apic_available = vmware_legacy_x2apic_available,
|
||||
.type = X86_HYPER_VMWARE,
|
||||
.init.init_platform = vmware_platform_setup,
|
||||
.init.x2apic_available = vmware_legacy_x2apic_available,
|
||||
};
|
||||
EXPORT_SYMBOL(x86_hyper_vmware);
|
||||
|
|
|
@ -498,7 +498,7 @@ static void __init kvm_apf_trap_init(void)
|
|||
update_intr_gate(X86_TRAP_PF, async_page_fault);
|
||||
}
|
||||
|
||||
void __init kvm_guest_init(void)
|
||||
static void __init kvm_guest_init(void)
|
||||
{
|
||||
int i;
|
||||
|
||||
|
@ -578,12 +578,13 @@ static uint32_t __init kvm_detect(void)
|
|||
return kvm_cpuid_base();
|
||||
}
|
||||
|
||||
const struct hypervisor_x86 x86_hyper_kvm __refconst = {
|
||||
const __initconst struct hypervisor_x86 x86_hyper_kvm = {
|
||||
.name = "KVM",
|
||||
.detect = kvm_detect,
|
||||
.x2apic_available = kvm_para_available,
|
||||
.type = X86_HYPER_KVM,
|
||||
.init.guest_late_init = kvm_guest_init,
|
||||
.init.x2apic_available = kvm_para_available,
|
||||
};
|
||||
EXPORT_SYMBOL_GPL(x86_hyper_kvm);
|
||||
|
||||
static __init int activate_jump_labels(void)
|
||||
{
|
||||
|
|
|
@ -1296,7 +1296,7 @@ void __init setup_arch(char **cmdline_p)
|
|||
|
||||
io_apic_init_mappings();
|
||||
|
||||
kvm_guest_init();
|
||||
x86_init.hyper.guest_late_init();
|
||||
|
||||
e820__reserve_resources();
|
||||
e820__register_nosave_regions(max_low_pfn);
|
||||
|
|
|
@ -28,6 +28,8 @@ void x86_init_noop(void) { }
|
|||
void __init x86_init_uint_noop(unsigned int unused) { }
|
||||
int __init iommu_init_noop(void) { return 0; }
|
||||
void iommu_shutdown_noop(void) { }
|
||||
bool __init bool_x86_init_noop(void) { return false; }
|
||||
void x86_op_int_noop(int cpu) { }
|
||||
|
||||
/*
|
||||
* The platform setup functions are preset with the default functions
|
||||
|
@ -81,6 +83,13 @@ struct x86_init_ops x86_init __initdata = {
|
|||
.init_irq = x86_default_pci_init_irq,
|
||||
.fixup_irqs = x86_default_pci_fixup_irqs,
|
||||
},
|
||||
|
||||
.hyper = {
|
||||
.init_platform = x86_init_noop,
|
||||
.guest_late_init = x86_init_noop,
|
||||
.x2apic_available = bool_x86_init_noop,
|
||||
.init_mem_mapping = x86_init_noop,
|
||||
},
|
||||
};
|
||||
|
||||
struct x86_cpuinit_ops x86_cpuinit = {
|
||||
|
@ -101,6 +110,7 @@ struct x86_platform_ops x86_platform __ro_after_init = {
|
|||
.get_nmi_reason = default_get_nmi_reason,
|
||||
.save_sched_clock_state = tsc_save_sched_clock_state,
|
||||
.restore_sched_clock_state = tsc_restore_sched_clock_state,
|
||||
.hyper.pin_vcpu = x86_op_int_noop,
|
||||
};
|
||||
|
||||
EXPORT_SYMBOL_GPL(x86_platform);
|
||||
|
|
|
@ -671,7 +671,7 @@ void __init init_mem_mapping(void)
|
|||
load_cr3(swapper_pg_dir);
|
||||
__flush_tlb_all();
|
||||
|
||||
hypervisor_init_mem_mapping();
|
||||
x86_init.hyper.init_mem_mapping();
|
||||
|
||||
early_memtest(0, max_pfn_mapped << PAGE_SHIFT);
|
||||
}
|
||||
|
|
|
@ -1,3 +1,4 @@
|
|||
#include <linux/acpi.h>
|
||||
#include <linux/cpu.h>
|
||||
#include <linux/kexec.h>
|
||||
#include <linux/memblock.h>
|
||||
|
@ -188,8 +189,6 @@ static void __init xen_hvm_guest_init(void)
|
|||
xen_hvm_init_time_ops();
|
||||
xen_hvm_init_mmu_ops();
|
||||
|
||||
if (xen_pvh_domain())
|
||||
machine_ops.emergency_restart = xen_emergency_restart;
|
||||
#ifdef CONFIG_KEXEC_CORE
|
||||
machine_ops.shutdown = xen_hvm_shutdown;
|
||||
machine_ops.crash_shutdown = xen_hvm_crash_shutdown;
|
||||
|
@ -226,12 +225,33 @@ static uint32_t __init xen_platform_hvm(void)
|
|||
return xen_cpuid_base();
|
||||
}
|
||||
|
||||
const struct hypervisor_x86 x86_hyper_xen_hvm = {
|
||||
static __init void xen_hvm_guest_late_init(void)
|
||||
{
|
||||
#ifdef CONFIG_XEN_PVH
|
||||
/* Test for PVH domain (PVH boot path taken overrides ACPI flags). */
|
||||
if (!xen_pvh &&
|
||||
(x86_platform.legacy.rtc || !x86_platform.legacy.no_vga))
|
||||
return;
|
||||
|
||||
/* PVH detected. */
|
||||
xen_pvh = true;
|
||||
|
||||
/* Make sure we don't fall back to (default) ACPI_IRQ_MODEL_PIC. */
|
||||
if (!nr_ioapics && acpi_irq_model == ACPI_IRQ_MODEL_PIC)
|
||||
acpi_irq_model = ACPI_IRQ_MODEL_PLATFORM;
|
||||
|
||||
machine_ops.emergency_restart = xen_emergency_restart;
|
||||
pv_info.name = "Xen PVH";
|
||||
#endif
|
||||
}
|
||||
|
||||
const __initconst struct hypervisor_x86 x86_hyper_xen_hvm = {
|
||||
.name = "Xen HVM",
|
||||
.detect = xen_platform_hvm,
|
||||
.init_platform = xen_hvm_guest_init,
|
||||
.pin_vcpu = xen_pin_vcpu,
|
||||
.x2apic_available = xen_x2apic_para_available,
|
||||
.init_mem_mapping = xen_hvm_init_mem_mapping,
|
||||
.type = X86_HYPER_XEN_HVM,
|
||||
.init.init_platform = xen_hvm_guest_init,
|
||||
.init.x2apic_available = xen_x2apic_para_available,
|
||||
.init.init_mem_mapping = xen_hvm_init_mem_mapping,
|
||||
.init.guest_late_init = xen_hvm_guest_late_init,
|
||||
.runtime.pin_vcpu = xen_pin_vcpu,
|
||||
};
|
||||
EXPORT_SYMBOL(x86_hyper_xen_hvm);
|
||||
|
|
|
@ -1459,9 +1459,9 @@ static uint32_t __init xen_platform_pv(void)
|
|||
return 0;
|
||||
}
|
||||
|
||||
const struct hypervisor_x86 x86_hyper_xen_pv = {
|
||||
const __initconst struct hypervisor_x86 x86_hyper_xen_pv = {
|
||||
.name = "Xen PV",
|
||||
.detect = xen_platform_pv,
|
||||
.pin_vcpu = xen_pin_vcpu,
|
||||
.type = X86_HYPER_XEN_PV,
|
||||
.runtime.pin_vcpu = xen_pin_vcpu,
|
||||
};
|
||||
EXPORT_SYMBOL(x86_hyper_xen_pv);
|
||||
|
|
|
@ -25,13 +25,6 @@ struct boot_params pvh_bootparams __attribute__((section(".data")));
|
|||
struct hvm_start_info pvh_start_info;
|
||||
unsigned int pvh_start_info_sz = sizeof(pvh_start_info);
|
||||
|
||||
static void xen_pvh_arch_setup(void)
|
||||
{
|
||||
/* Make sure we don't fall back to (default) ACPI_IRQ_MODEL_PIC. */
|
||||
if (nr_ioapics == 0)
|
||||
acpi_irq_model = ACPI_IRQ_MODEL_PLATFORM;
|
||||
}
|
||||
|
||||
static void __init init_pvh_bootparams(void)
|
||||
{
|
||||
struct xen_memory_map memmap;
|
||||
|
@ -102,6 +95,4 @@ void __init xen_prepare_pvh(void)
|
|||
wrmsr_safe(msr, (u32)pfn, (u32)(pfn >> 32));
|
||||
|
||||
init_pvh_bootparams();
|
||||
|
||||
x86_init.oem.arch_setup = xen_pvh_arch_setup;
|
||||
}
|
||||
|
|
|
@ -1534,7 +1534,7 @@ static int __init hv_acpi_init(void)
|
|||
{
|
||||
int ret, t;
|
||||
|
||||
if (x86_hyper != &x86_hyper_ms_hyperv)
|
||||
if (x86_hyper_type != X86_HYPER_MS_HYPERV)
|
||||
return -ENODEV;
|
||||
|
||||
init_completion(&probe_event);
|
||||
|
|
|
@ -316,11 +316,9 @@ static int vmmouse_enable(struct psmouse *psmouse)
|
|||
/*
|
||||
* Array of supported hypervisors.
|
||||
*/
|
||||
static const struct hypervisor_x86 *vmmouse_supported_hypervisors[] = {
|
||||
&x86_hyper_vmware,
|
||||
#ifdef CONFIG_KVM_GUEST
|
||||
&x86_hyper_kvm,
|
||||
#endif
|
||||
static enum x86_hypervisor_type vmmouse_supported_hypervisors[] = {
|
||||
X86_HYPER_VMWARE,
|
||||
X86_HYPER_KVM,
|
||||
};
|
||||
|
||||
/**
|
||||
|
@ -331,7 +329,7 @@ static bool vmmouse_check_hypervisor(void)
|
|||
int i;
|
||||
|
||||
for (i = 0; i < ARRAY_SIZE(vmmouse_supported_hypervisors); i++)
|
||||
if (vmmouse_supported_hypervisors[i] == x86_hyper)
|
||||
if (vmmouse_supported_hypervisors[i] == x86_hyper_type)
|
||||
return true;
|
||||
|
||||
return false;
|
||||
|
|
|
@ -1271,7 +1271,7 @@ static int __init vmballoon_init(void)
|
|||
* Check if we are running on VMware's hypervisor and bail out
|
||||
* if we are not.
|
||||
*/
|
||||
if (x86_hyper != &x86_hyper_vmware)
|
||||
if (x86_hyper_type != X86_HYPER_VMWARE)
|
||||
return -ENODEV;
|
||||
|
||||
for (is_2m_pages = 0; is_2m_pages < VMW_BALLOON_NUM_PAGE_SIZES;
|
||||
|
|
|
@ -7,8 +7,12 @@
|
|||
* Juergen Gross <jgross@suse.com>
|
||||
*/
|
||||
|
||||
#ifdef CONFIG_HYPERVISOR_GUEST
|
||||
#include <asm/hypervisor.h>
|
||||
#ifdef CONFIG_X86
|
||||
#include <asm/x86_init.h>
|
||||
static inline void hypervisor_pin_vcpu(int cpu)
|
||||
{
|
||||
x86_platform.hyper.pin_vcpu(cpu);
|
||||
}
|
||||
#else
|
||||
static inline void hypervisor_pin_vcpu(int cpu)
|
||||
{
|
||||
|
|
Loading…
Reference in New Issue