KVM SVM changes for 6.18
- Require a minimum GHCB version of 2 when starting SEV-SNP guests via
KVM_SEV_INIT2 so that invalid GHCB versions result in immediate errors
instead of latent guest failures.
- Add support for Secure TSC for SEV-SNP guests, which prevents the untrusted
host from tampering with the guest's TSC frequency, while still allowing the
the VMM to configure the guest's TSC frequency prior to launch.
- Mitigate the potential for TOCTOU bugs when accessing GHCB fields by
wrapping all accesses via READ_ONCE().
- Validate the XCR0 provided by the guest (via the GHCB) to avoid tracking a
bogous XCR0 value in KVM's software model.
- Save an SEV guest's policy if and only if LAUNCH_START fully succeeds to
avoid leaving behind stale state (thankfully not consumed in KVM).
- Explicitly reject non-positive effective lengths during SNP's LAUNCH_UPDATE
instead of subtly relying on guest_memfd to do the "heavy" lifting.
- Reload the pre-VMRUN TSC_AUX on #VMEXIT for SEV-ES guests, not the host's
desired TSC_AUX, to fix a bug where KVM could clobber a different vCPU's
TSC_AUX due to hardware not matching the value cached in the user-return MSR
infrastructure.
- Enable AVIC by default for Zen4+ if x2AVIC (and other prereqs) is supported,
and clean up the AVIC initialization code along the way.
-----BEGIN PGP SIGNATURE-----
iQIzBAABCgAdFiEEKTobbabEP7vbhhN9OlYIJqCjN/0FAmjXH54ACgkQOlYIJqCj
N/0OCw//e+0o6jov6/PO8ljq6sXJySsXKxEFYnvQlWYzjqtlVs05Y2SY0GBTnMu3
g0ie2c4V3VD7cY5bGAWETWvrOMLqGXM3E7v9dVOuE4xU3xx0HkCAlXc/woOLUXoT
jo/komNXnpeiZ1QRO9FlGooHTJ6Y+jg6/mM7asStS2Pk3Mm//wYgQej9mSJDrypo
NB4+BCS9cyt8rndNtCUkyedFYMboVQ8AEvXh/jeydhw4rdbBh0/Ci2IKGcVI5DP1
be8GD/FsNTIUDtieHRYCR+LCKCMFj/hYzlg2nQ6UjxHZbvlDyQuh2Ld2LtZiGSef
ejNr9e+ro6vxWBgX6wplWtKRLxBYEnQ1h/rQ9A3g50TuhrtFJbxBxY7DPQ16hlBJ
EB/E1JFvVgkGVrYN0oPQCvvfhFtpkx43qnEBw4q0pbdAS79XOnG2GJFvI0hpZAP6
qwy19lbsJ5g3qLTlDPChxQJC08gThn3CbarCmZNNzBpPDQoLDUfYBfyN4prRPuiN
UByfaaEC0Fi6JSgmHsO0LsUB9K++k2ucWiIIW4YQhVgPUtCjTNLe9omgGJ1UYe0X
YITqgklewe3QtBJ46JE0APkPaHio7r6zd7QvO+RhRFkjwZfY6dlsrSImykKrpK3O
rPaZnW+UpAnA1XIqroMl1RVoczFCfGcP1Cat9JwScBVVxjJ1DlI=
=zd53
-----END PGP SIGNATURE-----
Merge tag 'kvm-x86-svm-6.18' of https://github.com/kvm-x86/linux into HEAD
KVM SVM changes for 6.18
- Require a minimum GHCB version of 2 when starting SEV-SNP guests via
KVM_SEV_INIT2 so that invalid GHCB versions result in immediate errors
instead of latent guest failures.
- Add support for Secure TSC for SEV-SNP guests, which prevents the untrusted
host from tampering with the guest's TSC frequency, while still allowing the
the VMM to configure the guest's TSC frequency prior to launch.
- Mitigate the potential for TOCTOU bugs when accessing GHCB fields by
wrapping all accesses via READ_ONCE().
- Validate the XCR0 provided by the guest (via the GHCB) to avoid tracking a
bogous XCR0 value in KVM's software model.
- Save an SEV guest's policy if and only if LAUNCH_START fully succeeds to
avoid leaving behind stale state (thankfully not consumed in KVM).
- Explicitly reject non-positive effective lengths during SNP's LAUNCH_UPDATE
instead of subtly relying on guest_memfd to do the "heavy" lifting.
- Reload the pre-VMRUN TSC_AUX on #VMEXIT for SEV-ES guests, not the host's
desired TSC_AUX, to fix a bug where KVM could clobber a different vCPU's
TSC_AUX due to hardware not matching the value cached in the user-return MSR
infrastructure.
- Enable AVIC by default for Zen4+ if x2AVIC (and other prereqs) is supported,
and clean up the AVIC initialization code along the way.
pull/1354/merge
commit
a104e0a305
|
|
@ -444,6 +444,7 @@
|
|||
#define X86_FEATURE_VM_PAGE_FLUSH (19*32+ 2) /* VM Page Flush MSR is supported */
|
||||
#define X86_FEATURE_SEV_ES (19*32+ 3) /* "sev_es" Secure Encrypted Virtualization - Encrypted State */
|
||||
#define X86_FEATURE_SEV_SNP (19*32+ 4) /* "sev_snp" Secure Encrypted Virtualization - Secure Nested Paging */
|
||||
#define X86_FEATURE_SNP_SECURE_TSC (19*32+ 8) /* SEV-SNP Secure TSC */
|
||||
#define X86_FEATURE_V_TSC_AUX (19*32+ 9) /* Virtual TSC_AUX */
|
||||
#define X86_FEATURE_SME_COHERENT (19*32+10) /* hardware-enforced cache coherency */
|
||||
#define X86_FEATURE_DEBUG_SWAP (19*32+14) /* "debug_swap" SEV-ES full debug state swap support */
|
||||
|
|
|
|||
|
|
@ -2200,6 +2200,7 @@ int kvm_set_dr(struct kvm_vcpu *vcpu, int dr, unsigned long val);
|
|||
unsigned long kvm_get_dr(struct kvm_vcpu *vcpu, int dr);
|
||||
unsigned long kvm_get_cr8(struct kvm_vcpu *vcpu);
|
||||
void kvm_lmsw(struct kvm_vcpu *vcpu, unsigned long msw);
|
||||
int __kvm_set_xcr(struct kvm_vcpu *vcpu, u32 index, u64 xcr);
|
||||
int kvm_emulate_xsetbv(struct kvm_vcpu *vcpu);
|
||||
|
||||
int kvm_get_msr_common(struct kvm_vcpu *vcpu, struct msr_data *msr);
|
||||
|
|
@ -2367,6 +2368,7 @@ int kvm_add_user_return_msr(u32 msr);
|
|||
int kvm_find_user_return_msr(u32 msr);
|
||||
int kvm_set_user_return_msr(unsigned index, u64 val, u64 mask);
|
||||
void kvm_user_return_msr_update_cache(unsigned int index, u64 val);
|
||||
u64 kvm_get_user_return_msr(unsigned int slot);
|
||||
|
||||
static inline bool kvm_is_supported_user_return_msr(u32 msr)
|
||||
{
|
||||
|
|
|
|||
|
|
@ -299,6 +299,7 @@ static_assert((X2AVIC_MAX_PHYSICAL_ID & AVIC_PHYSICAL_MAX_INDEX_MASK) == X2AVIC_
|
|||
#define SVM_SEV_FEAT_RESTRICTED_INJECTION BIT(3)
|
||||
#define SVM_SEV_FEAT_ALTERNATE_INJECTION BIT(4)
|
||||
#define SVM_SEV_FEAT_DEBUG_SWAP BIT(5)
|
||||
#define SVM_SEV_FEAT_SECURE_TSC BIT(9)
|
||||
|
||||
#define VMCB_ALLOWED_SEV_FEATURES_VALID BIT_ULL(63)
|
||||
|
||||
|
|
|
|||
|
|
@ -64,6 +64,34 @@
|
|||
|
||||
static_assert(__AVIC_GATAG(AVIC_VM_ID_MASK, AVIC_VCPU_IDX_MASK) == -1u);
|
||||
|
||||
#define AVIC_AUTO_MODE -1
|
||||
|
||||
static int avic_param_set(const char *val, const struct kernel_param *kp)
|
||||
{
|
||||
if (val && sysfs_streq(val, "auto")) {
|
||||
*(int *)kp->arg = AVIC_AUTO_MODE;
|
||||
return 0;
|
||||
}
|
||||
|
||||
return param_set_bint(val, kp);
|
||||
}
|
||||
|
||||
static const struct kernel_param_ops avic_ops = {
|
||||
.flags = KERNEL_PARAM_OPS_FL_NOARG,
|
||||
.set = avic_param_set,
|
||||
.get = param_get_bool,
|
||||
};
|
||||
|
||||
/*
|
||||
* Enable / disable AVIC. In "auto" mode (default behavior), AVIC is enabled
|
||||
* for Zen4+ CPUs with x2AVIC (and all other criteria for enablement are met).
|
||||
*/
|
||||
static int avic = AVIC_AUTO_MODE;
|
||||
module_param_cb(avic, &avic_ops, &avic, 0444);
|
||||
__MODULE_PARM_TYPE(avic, "bool");
|
||||
|
||||
module_param(enable_ipiv, bool, 0444);
|
||||
|
||||
static bool force_avic;
|
||||
module_param_unsafe(force_avic, bool, 0444);
|
||||
|
||||
|
|
@ -77,7 +105,58 @@ static DEFINE_HASHTABLE(svm_vm_data_hash, SVM_VM_DATA_HASH_BITS);
|
|||
static u32 next_vm_id = 0;
|
||||
static bool next_vm_id_wrapped = 0;
|
||||
static DEFINE_SPINLOCK(svm_vm_data_hash_lock);
|
||||
bool x2avic_enabled;
|
||||
static bool x2avic_enabled;
|
||||
|
||||
|
||||
static void avic_set_x2apic_msr_interception(struct vcpu_svm *svm,
|
||||
bool intercept)
|
||||
{
|
||||
static const u32 x2avic_passthrough_msrs[] = {
|
||||
X2APIC_MSR(APIC_ID),
|
||||
X2APIC_MSR(APIC_LVR),
|
||||
X2APIC_MSR(APIC_TASKPRI),
|
||||
X2APIC_MSR(APIC_ARBPRI),
|
||||
X2APIC_MSR(APIC_PROCPRI),
|
||||
X2APIC_MSR(APIC_EOI),
|
||||
X2APIC_MSR(APIC_RRR),
|
||||
X2APIC_MSR(APIC_LDR),
|
||||
X2APIC_MSR(APIC_DFR),
|
||||
X2APIC_MSR(APIC_SPIV),
|
||||
X2APIC_MSR(APIC_ISR),
|
||||
X2APIC_MSR(APIC_TMR),
|
||||
X2APIC_MSR(APIC_IRR),
|
||||
X2APIC_MSR(APIC_ESR),
|
||||
X2APIC_MSR(APIC_ICR),
|
||||
X2APIC_MSR(APIC_ICR2),
|
||||
|
||||
/*
|
||||
* Note! Always intercept LVTT, as TSC-deadline timer mode
|
||||
* isn't virtualized by hardware, and the CPU will generate a
|
||||
* #GP instead of a #VMEXIT.
|
||||
*/
|
||||
X2APIC_MSR(APIC_LVTTHMR),
|
||||
X2APIC_MSR(APIC_LVTPC),
|
||||
X2APIC_MSR(APIC_LVT0),
|
||||
X2APIC_MSR(APIC_LVT1),
|
||||
X2APIC_MSR(APIC_LVTERR),
|
||||
X2APIC_MSR(APIC_TMICT),
|
||||
X2APIC_MSR(APIC_TMCCT),
|
||||
X2APIC_MSR(APIC_TDCR),
|
||||
};
|
||||
int i;
|
||||
|
||||
if (intercept == svm->x2avic_msrs_intercepted)
|
||||
return;
|
||||
|
||||
if (!x2avic_enabled)
|
||||
return;
|
||||
|
||||
for (i = 0; i < ARRAY_SIZE(x2avic_passthrough_msrs); i++)
|
||||
svm_set_intercept_for_msr(&svm->vcpu, x2avic_passthrough_msrs[i],
|
||||
MSR_TYPE_RW, intercept);
|
||||
|
||||
svm->x2avic_msrs_intercepted = intercept;
|
||||
}
|
||||
|
||||
static void avic_activate_vmcb(struct vcpu_svm *svm)
|
||||
{
|
||||
|
|
@ -99,7 +178,7 @@ static void avic_activate_vmcb(struct vcpu_svm *svm)
|
|||
vmcb->control.int_ctl |= X2APIC_MODE_MASK;
|
||||
vmcb->control.avic_physical_id |= X2AVIC_MAX_PHYSICAL_ID;
|
||||
/* Disabling MSR intercept for x2APIC registers */
|
||||
svm_set_x2apic_msr_interception(svm, false);
|
||||
avic_set_x2apic_msr_interception(svm, false);
|
||||
} else {
|
||||
/*
|
||||
* Flush the TLB, the guest may have inserted a non-APIC
|
||||
|
|
@ -110,7 +189,7 @@ static void avic_activate_vmcb(struct vcpu_svm *svm)
|
|||
/* For xAVIC and hybrid-xAVIC modes */
|
||||
vmcb->control.avic_physical_id |= AVIC_MAX_PHYSICAL_ID;
|
||||
/* Enabling MSR intercept for x2APIC registers */
|
||||
svm_set_x2apic_msr_interception(svm, true);
|
||||
avic_set_x2apic_msr_interception(svm, true);
|
||||
}
|
||||
}
|
||||
|
||||
|
|
@ -130,7 +209,7 @@ static void avic_deactivate_vmcb(struct vcpu_svm *svm)
|
|||
return;
|
||||
|
||||
/* Enabling MSR intercept for x2APIC registers */
|
||||
svm_set_x2apic_msr_interception(svm, true);
|
||||
avic_set_x2apic_msr_interception(svm, true);
|
||||
}
|
||||
|
||||
/* Note:
|
||||
|
|
@ -1090,23 +1169,27 @@ void avic_vcpu_unblocking(struct kvm_vcpu *vcpu)
|
|||
avic_vcpu_load(vcpu, vcpu->cpu);
|
||||
}
|
||||
|
||||
/*
|
||||
* Note:
|
||||
* - The module param avic enable both xAPIC and x2APIC mode.
|
||||
* - Hypervisor can support both xAVIC and x2AVIC in the same guest.
|
||||
* - The mode can be switched at run-time.
|
||||
*/
|
||||
bool avic_hardware_setup(void)
|
||||
static bool __init avic_want_avic_enabled(void)
|
||||
{
|
||||
if (!npt_enabled)
|
||||
/*
|
||||
* In "auto" mode, enable AVIC by default for Zen4+ if x2AVIC is
|
||||
* supported (to avoid enabling partial support by default, and because
|
||||
* x2AVIC should be supported by all Zen4+ CPUs). Explicitly check for
|
||||
* family 0x19 and later (Zen5+), as the kernel's synthetic ZenX flags
|
||||
* aren't inclusive of previous generations, i.e. the kernel will set
|
||||
* at most one ZenX feature flag.
|
||||
*/
|
||||
if (avic == AVIC_AUTO_MODE)
|
||||
avic = boot_cpu_has(X86_FEATURE_X2AVIC) &&
|
||||
(boot_cpu_data.x86 > 0x19 || cpu_feature_enabled(X86_FEATURE_ZEN4));
|
||||
|
||||
if (!avic || !npt_enabled)
|
||||
return false;
|
||||
|
||||
/* AVIC is a prerequisite for x2AVIC. */
|
||||
if (!boot_cpu_has(X86_FEATURE_AVIC) && !force_avic) {
|
||||
if (boot_cpu_has(X86_FEATURE_X2AVIC)) {
|
||||
pr_warn(FW_BUG "Cannot support x2AVIC due to AVIC is disabled");
|
||||
pr_warn(FW_BUG "Try enable AVIC using force_avic option");
|
||||
}
|
||||
if (boot_cpu_has(X86_FEATURE_X2AVIC))
|
||||
pr_warn(FW_BUG "Cannot enable x2AVIC, AVIC is unsupported\n");
|
||||
return false;
|
||||
}
|
||||
|
||||
|
|
@ -1116,21 +1199,37 @@ bool avic_hardware_setup(void)
|
|||
return false;
|
||||
}
|
||||
|
||||
if (boot_cpu_has(X86_FEATURE_AVIC)) {
|
||||
pr_info("AVIC enabled\n");
|
||||
} else if (force_avic) {
|
||||
/*
|
||||
* Some older systems does not advertise AVIC support.
|
||||
* See Revision Guide for specific AMD processor for more detail.
|
||||
*/
|
||||
pr_warn("AVIC is not supported in CPUID but force enabled");
|
||||
pr_warn("Your system might crash and burn");
|
||||
}
|
||||
/*
|
||||
* Print a scary message if AVIC is force enabled to make it abundantly
|
||||
* clear that ignoring CPUID could have repercussions. See Revision
|
||||
* Guide for specific AMD processor for more details.
|
||||
*/
|
||||
if (!boot_cpu_has(X86_FEATURE_AVIC))
|
||||
pr_warn("AVIC unsupported in CPUID but force enabled, your system might crash and burn\n");
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
/*
|
||||
* Note:
|
||||
* - The module param avic enable both xAPIC and x2APIC mode.
|
||||
* - Hypervisor can support both xAVIC and x2AVIC in the same guest.
|
||||
* - The mode can be switched at run-time.
|
||||
*/
|
||||
bool __init avic_hardware_setup(void)
|
||||
{
|
||||
avic = avic_want_avic_enabled();
|
||||
if (!avic)
|
||||
return false;
|
||||
|
||||
pr_info("AVIC enabled\n");
|
||||
|
||||
/* AVIC is a prerequisite for x2AVIC. */
|
||||
x2avic_enabled = boot_cpu_has(X86_FEATURE_X2AVIC);
|
||||
if (x2avic_enabled)
|
||||
pr_info("x2AVIC enabled\n");
|
||||
else
|
||||
svm_x86_ops.allow_apicv_in_x2apic_without_x2apic_virtualization = true;
|
||||
|
||||
/*
|
||||
* Disable IPI virtualization for AMD Family 17h CPUs (Zen1 and Zen2)
|
||||
|
|
|
|||
|
|
@ -1798,17 +1798,15 @@ static int svm_set_nested_state(struct kvm_vcpu *vcpu,
|
|||
if (kvm_state->size < sizeof(*kvm_state) + KVM_STATE_NESTED_SVM_VMCB_SIZE)
|
||||
return -EINVAL;
|
||||
|
||||
ret = -ENOMEM;
|
||||
ctl = kzalloc(sizeof(*ctl), GFP_KERNEL);
|
||||
save = kzalloc(sizeof(*save), GFP_KERNEL);
|
||||
if (!ctl || !save)
|
||||
goto out_free;
|
||||
ctl = memdup_user(&user_vmcb->control, sizeof(*ctl));
|
||||
if (IS_ERR(ctl))
|
||||
return PTR_ERR(ctl);
|
||||
|
||||
ret = -EFAULT;
|
||||
if (copy_from_user(ctl, &user_vmcb->control, sizeof(*ctl)))
|
||||
goto out_free;
|
||||
if (copy_from_user(save, &user_vmcb->save, sizeof(*save)))
|
||||
goto out_free;
|
||||
save = memdup_user(&user_vmcb->save, sizeof(*save));
|
||||
if (IS_ERR(save)) {
|
||||
kfree(ctl);
|
||||
return PTR_ERR(save);
|
||||
}
|
||||
|
||||
ret = -EINVAL;
|
||||
__nested_copy_vmcb_control_to_cache(vcpu, &ctl_cached, ctl);
|
||||
|
|
|
|||
|
|
@ -37,7 +37,6 @@
|
|||
#include "trace.h"
|
||||
|
||||
#define GHCB_VERSION_MAX 2ULL
|
||||
#define GHCB_VERSION_DEFAULT 2ULL
|
||||
#define GHCB_VERSION_MIN 1ULL
|
||||
|
||||
#define GHCB_HV_FT_SUPPORTED (GHCB_HV_FT_SNP | GHCB_HV_FT_SNP_AP_CREATION)
|
||||
|
|
@ -147,6 +146,14 @@ static bool sev_vcpu_has_debug_swap(struct vcpu_svm *svm)
|
|||
return sev->vmsa_features & SVM_SEV_FEAT_DEBUG_SWAP;
|
||||
}
|
||||
|
||||
static bool snp_is_secure_tsc_enabled(struct kvm *kvm)
|
||||
{
|
||||
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
||||
|
||||
return (sev->vmsa_features & SVM_SEV_FEAT_SECURE_TSC) &&
|
||||
!WARN_ON_ONCE(!sev_snp_guest(kvm));
|
||||
}
|
||||
|
||||
/* Must be called with the sev_bitmap_lock held */
|
||||
static bool __sev_recycle_asids(unsigned int min_asid, unsigned int max_asid)
|
||||
{
|
||||
|
|
@ -406,6 +413,7 @@ static int __sev_guest_init(struct kvm *kvm, struct kvm_sev_cmd *argp,
|
|||
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
||||
struct sev_platform_init_args init_args = {0};
|
||||
bool es_active = vm_type != KVM_X86_SEV_VM;
|
||||
bool snp_active = vm_type == KVM_X86_SNP_VM;
|
||||
u64 valid_vmsa_features = es_active ? sev_supported_vmsa_features : 0;
|
||||
int ret;
|
||||
|
||||
|
|
@ -415,12 +423,26 @@ static int __sev_guest_init(struct kvm *kvm, struct kvm_sev_cmd *argp,
|
|||
if (data->flags)
|
||||
return -EINVAL;
|
||||
|
||||
if (!snp_active)
|
||||
valid_vmsa_features &= ~SVM_SEV_FEAT_SECURE_TSC;
|
||||
|
||||
if (data->vmsa_features & ~valid_vmsa_features)
|
||||
return -EINVAL;
|
||||
|
||||
if (data->ghcb_version > GHCB_VERSION_MAX || (!es_active && data->ghcb_version))
|
||||
return -EINVAL;
|
||||
|
||||
/*
|
||||
* KVM supports the full range of mandatory features defined by version
|
||||
* 2 of the GHCB protocol, so default to that for SEV-ES guests created
|
||||
* via KVM_SEV_INIT2 (KVM_SEV_INIT forces version 1).
|
||||
*/
|
||||
if (es_active && !data->ghcb_version)
|
||||
data->ghcb_version = 2;
|
||||
|
||||
if (snp_active && data->ghcb_version < 2)
|
||||
return -EINVAL;
|
||||
|
||||
if (unlikely(sev->active))
|
||||
return -EINVAL;
|
||||
|
||||
|
|
@ -429,15 +451,7 @@ static int __sev_guest_init(struct kvm *kvm, struct kvm_sev_cmd *argp,
|
|||
sev->vmsa_features = data->vmsa_features;
|
||||
sev->ghcb_version = data->ghcb_version;
|
||||
|
||||
/*
|
||||
* Currently KVM supports the full range of mandatory features defined
|
||||
* by version 2 of the GHCB protocol, so default to that for SEV-ES
|
||||
* guests created via KVM_SEV_INIT2.
|
||||
*/
|
||||
if (sev->es_active && !sev->ghcb_version)
|
||||
sev->ghcb_version = GHCB_VERSION_DEFAULT;
|
||||
|
||||
if (vm_type == KVM_X86_SNP_VM)
|
||||
if (snp_active)
|
||||
sev->vmsa_features |= SVM_SEV_FEAT_SNP_ACTIVE;
|
||||
|
||||
ret = sev_asid_new(sev);
|
||||
|
|
@ -455,7 +469,7 @@ static int __sev_guest_init(struct kvm *kvm, struct kvm_sev_cmd *argp,
|
|||
}
|
||||
|
||||
/* This needs to happen after SEV/SNP firmware initialization. */
|
||||
if (vm_type == KVM_X86_SNP_VM) {
|
||||
if (snp_active) {
|
||||
ret = snp_guest_req_init(kvm);
|
||||
if (ret)
|
||||
goto e_free;
|
||||
|
|
@ -569,8 +583,6 @@ static int sev_launch_start(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|||
if (copy_from_user(¶ms, u64_to_user_ptr(argp->data), sizeof(params)))
|
||||
return -EFAULT;
|
||||
|
||||
sev->policy = params.policy;
|
||||
|
||||
memset(&start, 0, sizeof(start));
|
||||
|
||||
dh_blob = NULL;
|
||||
|
|
@ -618,6 +630,7 @@ static int sev_launch_start(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|||
goto e_free_session;
|
||||
}
|
||||
|
||||
sev->policy = params.policy;
|
||||
sev->handle = start.handle;
|
||||
sev->fd = argp->sev_fd;
|
||||
|
||||
|
|
@ -1968,7 +1981,7 @@ static void sev_migrate_from(struct kvm *dst_kvm, struct kvm *src_kvm)
|
|||
kvm_for_each_vcpu(i, dst_vcpu, dst_kvm) {
|
||||
dst_svm = to_svm(dst_vcpu);
|
||||
|
||||
sev_init_vmcb(dst_svm);
|
||||
sev_init_vmcb(dst_svm, false);
|
||||
|
||||
if (!dst->es_active)
|
||||
continue;
|
||||
|
|
@ -2180,7 +2193,12 @@ static int snp_launch_start(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|||
if (!(params.policy & SNP_POLICY_MASK_RSVD_MBO))
|
||||
return -EINVAL;
|
||||
|
||||
sev->policy = params.policy;
|
||||
if (snp_is_secure_tsc_enabled(kvm)) {
|
||||
if (WARN_ON_ONCE(!kvm->arch.default_tsc_khz))
|
||||
return -EINVAL;
|
||||
|
||||
start.desired_tsc_khz = kvm->arch.default_tsc_khz;
|
||||
}
|
||||
|
||||
sev->snp_context = snp_context_create(kvm, argp);
|
||||
if (!sev->snp_context)
|
||||
|
|
@ -2188,6 +2206,7 @@ static int snp_launch_start(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|||
|
||||
start.gctx_paddr = __psp_pa(sev->snp_context);
|
||||
start.policy = params.policy;
|
||||
|
||||
memcpy(start.gosvw, params.gosvw, sizeof(params.gosvw));
|
||||
rc = __sev_issue_cmd(argp->sev_fd, SEV_CMD_SNP_LAUNCH_START, &start, &argp->error);
|
||||
if (rc) {
|
||||
|
|
@ -2196,6 +2215,7 @@ static int snp_launch_start(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|||
goto e_free_context;
|
||||
}
|
||||
|
||||
sev->policy = params.policy;
|
||||
sev->fd = argp->sev_fd;
|
||||
rc = snp_bind_asid(kvm, &argp->error);
|
||||
if (rc) {
|
||||
|
|
@ -2329,7 +2349,7 @@ static int snp_launch_update(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|||
pr_debug("%s: GFN start 0x%llx length 0x%llx type %d flags %d\n", __func__,
|
||||
params.gfn_start, params.len, params.type, params.flags);
|
||||
|
||||
if (!PAGE_ALIGNED(params.len) || params.flags ||
|
||||
if (!params.len || !PAGE_ALIGNED(params.len) || params.flags ||
|
||||
(params.type != KVM_SEV_SNP_PAGE_TYPE_NORMAL &&
|
||||
params.type != KVM_SEV_SNP_PAGE_TYPE_ZERO &&
|
||||
params.type != KVM_SEV_SNP_PAGE_TYPE_UNMEASURED &&
|
||||
|
|
@ -3078,6 +3098,9 @@ out:
|
|||
sev_supported_vmsa_features = 0;
|
||||
if (sev_es_debug_swap_enabled)
|
||||
sev_supported_vmsa_features |= SVM_SEV_FEAT_DEBUG_SWAP;
|
||||
|
||||
if (sev_snp_enabled && tsc_khz && cpu_feature_enabled(X86_FEATURE_SNP_SECURE_TSC))
|
||||
sev_supported_vmsa_features |= SVM_SEV_FEAT_SECURE_TSC;
|
||||
}
|
||||
|
||||
void sev_hardware_unsetup(void)
|
||||
|
|
@ -3193,7 +3216,7 @@ skip_vmsa_free:
|
|||
kvfree(svm->sev_es.ghcb_sa);
|
||||
}
|
||||
|
||||
static u64 kvm_ghcb_get_sw_exit_code(struct vmcb_control_area *control)
|
||||
static u64 kvm_get_cached_sw_exit_code(struct vmcb_control_area *control)
|
||||
{
|
||||
return (((u64)control->exit_code_hi) << 32) | control->exit_code;
|
||||
}
|
||||
|
|
@ -3219,7 +3242,7 @@ static void dump_ghcb(struct vcpu_svm *svm)
|
|||
*/
|
||||
pr_err("GHCB (GPA=%016llx) snapshot:\n", svm->vmcb->control.ghcb_gpa);
|
||||
pr_err("%-20s%016llx is_valid: %u\n", "sw_exit_code",
|
||||
kvm_ghcb_get_sw_exit_code(control), kvm_ghcb_sw_exit_code_is_valid(svm));
|
||||
kvm_get_cached_sw_exit_code(control), kvm_ghcb_sw_exit_code_is_valid(svm));
|
||||
pr_err("%-20s%016llx is_valid: %u\n", "sw_exit_info_1",
|
||||
control->exit_info_1, kvm_ghcb_sw_exit_info_1_is_valid(svm));
|
||||
pr_err("%-20s%016llx is_valid: %u\n", "sw_exit_info_2",
|
||||
|
|
@ -3272,26 +3295,24 @@ static void sev_es_sync_from_ghcb(struct vcpu_svm *svm)
|
|||
BUILD_BUG_ON(sizeof(svm->sev_es.valid_bitmap) != sizeof(ghcb->save.valid_bitmap));
|
||||
memcpy(&svm->sev_es.valid_bitmap, &ghcb->save.valid_bitmap, sizeof(ghcb->save.valid_bitmap));
|
||||
|
||||
vcpu->arch.regs[VCPU_REGS_RAX] = kvm_ghcb_get_rax_if_valid(svm, ghcb);
|
||||
vcpu->arch.regs[VCPU_REGS_RBX] = kvm_ghcb_get_rbx_if_valid(svm, ghcb);
|
||||
vcpu->arch.regs[VCPU_REGS_RCX] = kvm_ghcb_get_rcx_if_valid(svm, ghcb);
|
||||
vcpu->arch.regs[VCPU_REGS_RDX] = kvm_ghcb_get_rdx_if_valid(svm, ghcb);
|
||||
vcpu->arch.regs[VCPU_REGS_RSI] = kvm_ghcb_get_rsi_if_valid(svm, ghcb);
|
||||
vcpu->arch.regs[VCPU_REGS_RAX] = kvm_ghcb_get_rax_if_valid(svm);
|
||||
vcpu->arch.regs[VCPU_REGS_RBX] = kvm_ghcb_get_rbx_if_valid(svm);
|
||||
vcpu->arch.regs[VCPU_REGS_RCX] = kvm_ghcb_get_rcx_if_valid(svm);
|
||||
vcpu->arch.regs[VCPU_REGS_RDX] = kvm_ghcb_get_rdx_if_valid(svm);
|
||||
vcpu->arch.regs[VCPU_REGS_RSI] = kvm_ghcb_get_rsi_if_valid(svm);
|
||||
|
||||
svm->vmcb->save.cpl = kvm_ghcb_get_cpl_if_valid(svm, ghcb);
|
||||
svm->vmcb->save.cpl = kvm_ghcb_get_cpl_if_valid(svm);
|
||||
|
||||
if (kvm_ghcb_xcr0_is_valid(svm)) {
|
||||
vcpu->arch.xcr0 = ghcb_get_xcr0(ghcb);
|
||||
vcpu->arch.cpuid_dynamic_bits_dirty = true;
|
||||
}
|
||||
if (kvm_ghcb_xcr0_is_valid(svm))
|
||||
__kvm_set_xcr(vcpu, 0, kvm_ghcb_get_xcr0(svm));
|
||||
|
||||
/* Copy the GHCB exit information into the VMCB fields */
|
||||
exit_code = ghcb_get_sw_exit_code(ghcb);
|
||||
exit_code = kvm_ghcb_get_sw_exit_code(svm);
|
||||
control->exit_code = lower_32_bits(exit_code);
|
||||
control->exit_code_hi = upper_32_bits(exit_code);
|
||||
control->exit_info_1 = ghcb_get_sw_exit_info_1(ghcb);
|
||||
control->exit_info_2 = ghcb_get_sw_exit_info_2(ghcb);
|
||||
svm->sev_es.sw_scratch = kvm_ghcb_get_sw_scratch_if_valid(svm, ghcb);
|
||||
control->exit_info_1 = kvm_ghcb_get_sw_exit_info_1(svm);
|
||||
control->exit_info_2 = kvm_ghcb_get_sw_exit_info_2(svm);
|
||||
svm->sev_es.sw_scratch = kvm_ghcb_get_sw_scratch_if_valid(svm);
|
||||
|
||||
/* Clear the valid entries fields */
|
||||
memset(ghcb->save.valid_bitmap, 0, sizeof(ghcb->save.valid_bitmap));
|
||||
|
|
@ -3308,7 +3329,7 @@ static int sev_es_validate_vmgexit(struct vcpu_svm *svm)
|
|||
* Retrieve the exit code now even though it may not be marked valid
|
||||
* as it could help with debugging.
|
||||
*/
|
||||
exit_code = kvm_ghcb_get_sw_exit_code(control);
|
||||
exit_code = kvm_get_cached_sw_exit_code(control);
|
||||
|
||||
/* Only GHCB Usage code 0 is supported */
|
||||
if (svm->sev_es.ghcb->ghcb_usage) {
|
||||
|
|
@ -3880,7 +3901,7 @@ next_range:
|
|||
/*
|
||||
* Invoked as part of svm_vcpu_reset() processing of an init event.
|
||||
*/
|
||||
void sev_snp_init_protected_guest_state(struct kvm_vcpu *vcpu)
|
||||
static void sev_snp_init_protected_guest_state(struct kvm_vcpu *vcpu)
|
||||
{
|
||||
struct vcpu_svm *svm = to_svm(vcpu);
|
||||
struct kvm_memory_slot *slot;
|
||||
|
|
@ -3888,9 +3909,6 @@ void sev_snp_init_protected_guest_state(struct kvm_vcpu *vcpu)
|
|||
kvm_pfn_t pfn;
|
||||
gfn_t gfn;
|
||||
|
||||
if (!sev_snp_guest(vcpu->kvm))
|
||||
return;
|
||||
|
||||
guard(mutex)(&svm->sev_es.snp_vmsa_mutex);
|
||||
|
||||
if (!svm->sev_es.snp_ap_waiting_for_reset)
|
||||
|
|
@ -4316,7 +4334,7 @@ int sev_handle_vmgexit(struct kvm_vcpu *vcpu)
|
|||
|
||||
svm_vmgexit_success(svm, 0);
|
||||
|
||||
exit_code = kvm_ghcb_get_sw_exit_code(control);
|
||||
exit_code = kvm_get_cached_sw_exit_code(control);
|
||||
switch (exit_code) {
|
||||
case SVM_VMGEXIT_MMIO_READ:
|
||||
ret = setup_vmgexit_scratch(svm, true, control->exit_info_2);
|
||||
|
|
@ -4448,6 +4466,9 @@ void sev_es_recalc_msr_intercepts(struct kvm_vcpu *vcpu)
|
|||
!guest_cpu_cap_has(vcpu, X86_FEATURE_RDTSCP) &&
|
||||
!guest_cpu_cap_has(vcpu, X86_FEATURE_RDPID));
|
||||
|
||||
svm_set_intercept_for_msr(vcpu, MSR_AMD64_GUEST_TSC_FREQ, MSR_TYPE_R,
|
||||
!snp_is_secure_tsc_enabled(vcpu->kvm));
|
||||
|
||||
/*
|
||||
* For SEV-ES, accesses to MSR_IA32_XSS should not be intercepted if
|
||||
* the host/guest supports its use.
|
||||
|
|
@ -4476,7 +4497,7 @@ void sev_vcpu_after_set_cpuid(struct vcpu_svm *svm)
|
|||
vcpu->arch.reserved_gpa_bits &= ~(1UL << (best->ebx & 0x3f));
|
||||
}
|
||||
|
||||
static void sev_es_init_vmcb(struct vcpu_svm *svm)
|
||||
static void sev_es_init_vmcb(struct vcpu_svm *svm, bool init_event)
|
||||
{
|
||||
struct kvm_sev_info *sev = to_kvm_sev_info(svm->vcpu.kvm);
|
||||
struct vmcb *vmcb = svm->vmcb01.ptr;
|
||||
|
|
@ -4537,10 +4558,21 @@ static void sev_es_init_vmcb(struct vcpu_svm *svm)
|
|||
|
||||
/* Can't intercept XSETBV, HV can't modify XCR0 directly */
|
||||
svm_clr_intercept(svm, INTERCEPT_XSETBV);
|
||||
|
||||
/*
|
||||
* Set the GHCB MSR value as per the GHCB specification when emulating
|
||||
* vCPU RESET for an SEV-ES guest.
|
||||
*/
|
||||
if (!init_event)
|
||||
set_ghcb_msr(svm, GHCB_MSR_SEV_INFO((__u64)sev->ghcb_version,
|
||||
GHCB_VERSION_MIN,
|
||||
sev_enc_bit));
|
||||
}
|
||||
|
||||
void sev_init_vmcb(struct vcpu_svm *svm)
|
||||
void sev_init_vmcb(struct vcpu_svm *svm, bool init_event)
|
||||
{
|
||||
struct kvm_vcpu *vcpu = &svm->vcpu;
|
||||
|
||||
svm->vmcb->control.nested_ctl |= SVM_NESTED_CTL_SEV_ENABLE;
|
||||
clr_exception_intercept(svm, UD_VECTOR);
|
||||
|
||||
|
|
@ -4550,24 +4582,36 @@ void sev_init_vmcb(struct vcpu_svm *svm)
|
|||
*/
|
||||
clr_exception_intercept(svm, GP_VECTOR);
|
||||
|
||||
if (sev_es_guest(svm->vcpu.kvm))
|
||||
sev_es_init_vmcb(svm);
|
||||
if (init_event && sev_snp_guest(vcpu->kvm))
|
||||
sev_snp_init_protected_guest_state(vcpu);
|
||||
|
||||
if (sev_es_guest(vcpu->kvm))
|
||||
sev_es_init_vmcb(svm, init_event);
|
||||
}
|
||||
|
||||
void sev_es_vcpu_reset(struct vcpu_svm *svm)
|
||||
int sev_vcpu_create(struct kvm_vcpu *vcpu)
|
||||
{
|
||||
struct kvm_vcpu *vcpu = &svm->vcpu;
|
||||
struct kvm_sev_info *sev = to_kvm_sev_info(vcpu->kvm);
|
||||
|
||||
/*
|
||||
* Set the GHCB MSR value as per the GHCB specification when emulating
|
||||
* vCPU RESET for an SEV-ES guest.
|
||||
*/
|
||||
set_ghcb_msr(svm, GHCB_MSR_SEV_INFO((__u64)sev->ghcb_version,
|
||||
GHCB_VERSION_MIN,
|
||||
sev_enc_bit));
|
||||
struct vcpu_svm *svm = to_svm(vcpu);
|
||||
struct page *vmsa_page;
|
||||
|
||||
mutex_init(&svm->sev_es.snp_vmsa_mutex);
|
||||
|
||||
if (!sev_es_guest(vcpu->kvm))
|
||||
return 0;
|
||||
|
||||
/*
|
||||
* SEV-ES guests require a separate (from the VMCB) VMSA page used to
|
||||
* contain the encrypted register state of the guest.
|
||||
*/
|
||||
vmsa_page = snp_safe_alloc_page();
|
||||
if (!vmsa_page)
|
||||
return -ENOMEM;
|
||||
|
||||
svm->sev_es.vmsa = page_address(vmsa_page);
|
||||
|
||||
vcpu->arch.guest_tsc_protected = snp_is_secure_tsc_enabled(vcpu->kvm);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
void sev_es_prepare_switch_to_guest(struct vcpu_svm *svm, struct sev_es_save_area *hostsa)
|
||||
|
|
@ -4618,6 +4662,16 @@ void sev_es_prepare_switch_to_guest(struct vcpu_svm *svm, struct sev_es_save_are
|
|||
hostsa->dr2_addr_mask = amd_get_dr_addr_mask(2);
|
||||
hostsa->dr3_addr_mask = amd_get_dr_addr_mask(3);
|
||||
}
|
||||
|
||||
/*
|
||||
* TSC_AUX is always virtualized for SEV-ES guests when the feature is
|
||||
* available, i.e. TSC_AUX is loaded on #VMEXIT from the host save area.
|
||||
* Set the save area to the current hardware value, i.e. the current
|
||||
* user return value, so that the correct value is restored on #VMEXIT.
|
||||
*/
|
||||
if (cpu_feature_enabled(X86_FEATURE_V_TSC_AUX) &&
|
||||
!WARN_ON_ONCE(tsc_aux_uret_slot < 0))
|
||||
hostsa->tsc_aux = kvm_get_user_return_msr(tsc_aux_uret_slot);
|
||||
}
|
||||
|
||||
void sev_vcpu_deliver_sipi_vector(struct kvm_vcpu *vcpu, u8 vector)
|
||||
|
|
|
|||
|
|
@ -158,14 +158,6 @@ module_param(lbrv, int, 0444);
|
|||
static int tsc_scaling = true;
|
||||
module_param(tsc_scaling, int, 0444);
|
||||
|
||||
/*
|
||||
* enable / disable AVIC. Because the defaults differ for APICv
|
||||
* support between VMX and SVM we cannot use module_param_named.
|
||||
*/
|
||||
static bool avic;
|
||||
module_param(avic, bool, 0444);
|
||||
module_param(enable_ipiv, bool, 0444);
|
||||
|
||||
module_param(enable_device_posted_irqs, bool, 0444);
|
||||
|
||||
bool __read_mostly dump_invalid_vmcb;
|
||||
|
|
@ -195,7 +187,7 @@ static DEFINE_MUTEX(vmcb_dump_mutex);
|
|||
* RDTSCP and RDPID are not used in the kernel, specifically to allow KVM to
|
||||
* defer the restoration of TSC_AUX until the CPU returns to userspace.
|
||||
*/
|
||||
static int tsc_aux_uret_slot __read_mostly = -1;
|
||||
int tsc_aux_uret_slot __ro_after_init = -1;
|
||||
|
||||
static int get_npt_level(void)
|
||||
{
|
||||
|
|
@ -577,18 +569,6 @@ static int svm_enable_virtualization_cpu(void)
|
|||
|
||||
amd_pmu_enable_virt();
|
||||
|
||||
/*
|
||||
* If TSC_AUX virtualization is supported, TSC_AUX becomes a swap type
|
||||
* "B" field (see sev_es_prepare_switch_to_guest()) for SEV-ES guests.
|
||||
* Since Linux does not change the value of TSC_AUX once set, prime the
|
||||
* TSC_AUX field now to avoid a RDMSR on every vCPU run.
|
||||
*/
|
||||
if (boot_cpu_has(X86_FEATURE_V_TSC_AUX)) {
|
||||
u32 __maybe_unused msr_hi;
|
||||
|
||||
rdmsr(MSR_TSC_AUX, sev_es_host_save_area(sd)->tsc_aux, msr_hi);
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
|
@ -736,55 +716,6 @@ static void svm_recalc_lbr_msr_intercepts(struct kvm_vcpu *vcpu)
|
|||
svm_set_intercept_for_msr(vcpu, MSR_IA32_DEBUGCTLMSR, MSR_TYPE_RW, intercept);
|
||||
}
|
||||
|
||||
void svm_set_x2apic_msr_interception(struct vcpu_svm *svm, bool intercept)
|
||||
{
|
||||
static const u32 x2avic_passthrough_msrs[] = {
|
||||
X2APIC_MSR(APIC_ID),
|
||||
X2APIC_MSR(APIC_LVR),
|
||||
X2APIC_MSR(APIC_TASKPRI),
|
||||
X2APIC_MSR(APIC_ARBPRI),
|
||||
X2APIC_MSR(APIC_PROCPRI),
|
||||
X2APIC_MSR(APIC_EOI),
|
||||
X2APIC_MSR(APIC_RRR),
|
||||
X2APIC_MSR(APIC_LDR),
|
||||
X2APIC_MSR(APIC_DFR),
|
||||
X2APIC_MSR(APIC_SPIV),
|
||||
X2APIC_MSR(APIC_ISR),
|
||||
X2APIC_MSR(APIC_TMR),
|
||||
X2APIC_MSR(APIC_IRR),
|
||||
X2APIC_MSR(APIC_ESR),
|
||||
X2APIC_MSR(APIC_ICR),
|
||||
X2APIC_MSR(APIC_ICR2),
|
||||
|
||||
/*
|
||||
* Note! Always intercept LVTT, as TSC-deadline timer mode
|
||||
* isn't virtualized by hardware, and the CPU will generate a
|
||||
* #GP instead of a #VMEXIT.
|
||||
*/
|
||||
X2APIC_MSR(APIC_LVTTHMR),
|
||||
X2APIC_MSR(APIC_LVTPC),
|
||||
X2APIC_MSR(APIC_LVT0),
|
||||
X2APIC_MSR(APIC_LVT1),
|
||||
X2APIC_MSR(APIC_LVTERR),
|
||||
X2APIC_MSR(APIC_TMICT),
|
||||
X2APIC_MSR(APIC_TMCCT),
|
||||
X2APIC_MSR(APIC_TDCR),
|
||||
};
|
||||
int i;
|
||||
|
||||
if (intercept == svm->x2avic_msrs_intercepted)
|
||||
return;
|
||||
|
||||
if (!x2avic_enabled)
|
||||
return;
|
||||
|
||||
for (i = 0; i < ARRAY_SIZE(x2avic_passthrough_msrs); i++)
|
||||
svm_set_intercept_for_msr(&svm->vcpu, x2avic_passthrough_msrs[i],
|
||||
MSR_TYPE_RW, intercept);
|
||||
|
||||
svm->x2avic_msrs_intercepted = intercept;
|
||||
}
|
||||
|
||||
void svm_vcpu_free_msrpm(void *msrpm)
|
||||
{
|
||||
__free_pages(virt_to_page(msrpm), get_order(MSRPM_SIZE));
|
||||
|
|
@ -1083,7 +1014,7 @@ static void svm_recalc_intercepts_after_set_cpuid(struct kvm_vcpu *vcpu)
|
|||
svm_recalc_msr_intercepts(vcpu);
|
||||
}
|
||||
|
||||
static void init_vmcb(struct kvm_vcpu *vcpu)
|
||||
static void init_vmcb(struct kvm_vcpu *vcpu, bool init_event)
|
||||
{
|
||||
struct vcpu_svm *svm = to_svm(vcpu);
|
||||
struct vmcb *vmcb = svm->vmcb01.ptr;
|
||||
|
|
@ -1221,7 +1152,7 @@ static void init_vmcb(struct kvm_vcpu *vcpu)
|
|||
svm_set_intercept(svm, INTERCEPT_BUSLOCK);
|
||||
|
||||
if (sev_guest(vcpu->kvm))
|
||||
sev_init_vmcb(svm);
|
||||
sev_init_vmcb(svm, init_event);
|
||||
|
||||
svm_hv_init_vmcb(vmcb);
|
||||
|
||||
|
|
@ -1244,9 +1175,6 @@ static void __svm_vcpu_reset(struct kvm_vcpu *vcpu)
|
|||
|
||||
svm->nmi_masked = false;
|
||||
svm->awaiting_iret_completion = false;
|
||||
|
||||
if (sev_es_guest(vcpu->kvm))
|
||||
sev_es_vcpu_reset(svm);
|
||||
}
|
||||
|
||||
static void svm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event)
|
||||
|
|
@ -1256,10 +1184,7 @@ static void svm_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event)
|
|||
svm->spec_ctrl = 0;
|
||||
svm->virt_spec_ctrl = 0;
|
||||
|
||||
if (init_event)
|
||||
sev_snp_init_protected_guest_state(vcpu);
|
||||
|
||||
init_vmcb(vcpu);
|
||||
init_vmcb(vcpu, init_event);
|
||||
|
||||
if (!init_event)
|
||||
__svm_vcpu_reset(vcpu);
|
||||
|
|
@ -1275,7 +1200,6 @@ static int svm_vcpu_create(struct kvm_vcpu *vcpu)
|
|||
{
|
||||
struct vcpu_svm *svm;
|
||||
struct page *vmcb01_page;
|
||||
struct page *vmsa_page = NULL;
|
||||
int err;
|
||||
|
||||
BUILD_BUG_ON(offsetof(struct vcpu_svm, vcpu) != 0);
|
||||
|
|
@ -1286,24 +1210,18 @@ static int svm_vcpu_create(struct kvm_vcpu *vcpu)
|
|||
if (!vmcb01_page)
|
||||
goto out;
|
||||
|
||||
if (sev_es_guest(vcpu->kvm)) {
|
||||
/*
|
||||
* SEV-ES guests require a separate VMSA page used to contain
|
||||
* the encrypted register state of the guest.
|
||||
*/
|
||||
vmsa_page = snp_safe_alloc_page();
|
||||
if (!vmsa_page)
|
||||
goto error_free_vmcb_page;
|
||||
}
|
||||
err = sev_vcpu_create(vcpu);
|
||||
if (err)
|
||||
goto error_free_vmcb_page;
|
||||
|
||||
err = avic_init_vcpu(svm);
|
||||
if (err)
|
||||
goto error_free_vmsa_page;
|
||||
goto error_free_sev;
|
||||
|
||||
svm->msrpm = svm_vcpu_alloc_msrpm();
|
||||
if (!svm->msrpm) {
|
||||
err = -ENOMEM;
|
||||
goto error_free_vmsa_page;
|
||||
goto error_free_sev;
|
||||
}
|
||||
|
||||
svm->x2avic_msrs_intercepted = true;
|
||||
|
|
@ -1312,16 +1230,12 @@ static int svm_vcpu_create(struct kvm_vcpu *vcpu)
|
|||
svm->vmcb01.pa = __sme_set(page_to_pfn(vmcb01_page) << PAGE_SHIFT);
|
||||
svm_switch_vmcb(svm, &svm->vmcb01);
|
||||
|
||||
if (vmsa_page)
|
||||
svm->sev_es.vmsa = page_address(vmsa_page);
|
||||
|
||||
svm->guest_state_loaded = false;
|
||||
|
||||
return 0;
|
||||
|
||||
error_free_vmsa_page:
|
||||
if (vmsa_page)
|
||||
__free_page(vmsa_page);
|
||||
error_free_sev:
|
||||
sev_free_vcpu(vcpu);
|
||||
error_free_vmcb_page:
|
||||
__free_page(vmcb01_page);
|
||||
out:
|
||||
|
|
@ -1423,10 +1337,10 @@ static void svm_prepare_switch_to_guest(struct kvm_vcpu *vcpu)
|
|||
__svm_write_tsc_multiplier(vcpu->arch.tsc_scaling_ratio);
|
||||
|
||||
/*
|
||||
* TSC_AUX is always virtualized for SEV-ES guests when the feature is
|
||||
* available. The user return MSR support is not required in this case
|
||||
* because TSC_AUX is restored on #VMEXIT from the host save area
|
||||
* (which has been initialized in svm_enable_virtualization_cpu()).
|
||||
* TSC_AUX is always virtualized (context switched by hardware) for
|
||||
* SEV-ES guests when the feature is available. For non-SEV-ES guests,
|
||||
* context switch TSC_AUX via the user_return MSR infrastructure (not
|
||||
* all CPUs support TSC_AUX virtualization).
|
||||
*/
|
||||
if (likely(tsc_aux_uret_slot >= 0) &&
|
||||
(!boot_cpu_has(X86_FEATURE_V_TSC_AUX) || !sev_es_guest(vcpu->kvm)))
|
||||
|
|
@ -3021,8 +2935,7 @@ static int svm_set_msr(struct kvm_vcpu *vcpu, struct msr_data *msr)
|
|||
* TSC_AUX is always virtualized for SEV-ES guests when the
|
||||
* feature is available. The user return MSR support is not
|
||||
* required in this case because TSC_AUX is restored on #VMEXIT
|
||||
* from the host save area (which has been initialized in
|
||||
* svm_enable_virtualization_cpu()).
|
||||
* from the host save area.
|
||||
*/
|
||||
if (boot_cpu_has(X86_FEATURE_V_TSC_AUX) && sev_es_guest(vcpu->kvm))
|
||||
break;
|
||||
|
|
@ -5041,7 +4954,7 @@ static void *svm_alloc_apic_backing_page(struct kvm_vcpu *vcpu)
|
|||
return page_address(page);
|
||||
}
|
||||
|
||||
static struct kvm_x86_ops svm_x86_ops __initdata = {
|
||||
struct kvm_x86_ops svm_x86_ops __initdata = {
|
||||
.name = KBUILD_MODNAME,
|
||||
|
||||
.check_processor_compatibility = svm_check_processor_compat,
|
||||
|
|
@ -5408,15 +5321,12 @@ static __init int svm_hardware_setup(void)
|
|||
goto err;
|
||||
}
|
||||
|
||||
enable_apicv = avic = avic && avic_hardware_setup();
|
||||
|
||||
enable_apicv = avic_hardware_setup();
|
||||
if (!enable_apicv) {
|
||||
enable_ipiv = false;
|
||||
svm_x86_ops.vcpu_blocking = NULL;
|
||||
svm_x86_ops.vcpu_unblocking = NULL;
|
||||
svm_x86_ops.vcpu_get_apicv_inhibit_reasons = NULL;
|
||||
} else if (!x2avic_enabled) {
|
||||
svm_x86_ops.allow_apicv_in_x2apic_without_x2apic_virtualization = true;
|
||||
}
|
||||
|
||||
if (vls) {
|
||||
|
|
|
|||
|
|
@ -48,10 +48,13 @@ extern bool npt_enabled;
|
|||
extern int nrips;
|
||||
extern int vgif;
|
||||
extern bool intercept_smi;
|
||||
extern bool x2avic_enabled;
|
||||
extern bool vnmi;
|
||||
extern int lbrv;
|
||||
|
||||
extern int tsc_aux_uret_slot __ro_after_init;
|
||||
|
||||
extern struct kvm_x86_ops svm_x86_ops __initdata;
|
||||
|
||||
/*
|
||||
* Clean bits in VMCB.
|
||||
* VMCB_ALL_CLEAN_MASK might also need to
|
||||
|
|
@ -699,7 +702,6 @@ void svm_set_gif(struct vcpu_svm *svm, bool value);
|
|||
int svm_invoke_exit_handler(struct kvm_vcpu *vcpu, u64 exit_code);
|
||||
void set_msr_interception(struct kvm_vcpu *vcpu, u32 *msrpm, u32 msr,
|
||||
int read, int write);
|
||||
void svm_set_x2apic_msr_interception(struct vcpu_svm *svm, bool disable);
|
||||
void svm_complete_interrupt_delivery(struct kvm_vcpu *vcpu, int delivery_mode,
|
||||
int trig_mode, int vec);
|
||||
|
||||
|
|
@ -801,7 +803,7 @@ extern struct kvm_x86_nested_ops svm_nested_ops;
|
|||
BIT(APICV_INHIBIT_REASON_PHYSICAL_ID_TOO_BIG) \
|
||||
)
|
||||
|
||||
bool avic_hardware_setup(void);
|
||||
bool __init avic_hardware_setup(void);
|
||||
int avic_ga_log_notifier(u32 ga_tag);
|
||||
void avic_vm_destroy(struct kvm *kvm);
|
||||
int avic_vm_init(struct kvm *kvm);
|
||||
|
|
@ -826,10 +828,9 @@ void avic_refresh_virtual_apic_mode(struct kvm_vcpu *vcpu);
|
|||
/* sev.c */
|
||||
|
||||
int pre_sev_run(struct vcpu_svm *svm, int cpu);
|
||||
void sev_init_vmcb(struct vcpu_svm *svm);
|
||||
void sev_init_vmcb(struct vcpu_svm *svm, bool init_event);
|
||||
void sev_vcpu_after_set_cpuid(struct vcpu_svm *svm);
|
||||
int sev_es_string_io(struct vcpu_svm *svm, int size, unsigned int port, int in);
|
||||
void sev_es_vcpu_reset(struct vcpu_svm *svm);
|
||||
void sev_es_recalc_msr_intercepts(struct kvm_vcpu *vcpu);
|
||||
void sev_vcpu_deliver_sipi_vector(struct kvm_vcpu *vcpu, u8 vector);
|
||||
void sev_es_prepare_switch_to_guest(struct vcpu_svm *svm, struct sev_es_save_area *hostsa);
|
||||
|
|
@ -854,6 +855,7 @@ static inline struct page *snp_safe_alloc_page(void)
|
|||
return snp_safe_alloc_page_node(numa_node_id(), GFP_KERNEL_ACCOUNT);
|
||||
}
|
||||
|
||||
int sev_vcpu_create(struct kvm_vcpu *vcpu);
|
||||
void sev_free_vcpu(struct kvm_vcpu *vcpu);
|
||||
void sev_vm_destroy(struct kvm *kvm);
|
||||
void __init sev_set_cpu_caps(void);
|
||||
|
|
@ -863,7 +865,6 @@ int sev_cpu_init(struct svm_cpu_data *sd);
|
|||
int sev_dev_get_attr(u32 group, u64 attr, u64 *val);
|
||||
extern unsigned int max_sev_asid;
|
||||
void sev_handle_rmp_fault(struct kvm_vcpu *vcpu, gpa_t gpa, u64 error_code);
|
||||
void sev_snp_init_protected_guest_state(struct kvm_vcpu *vcpu);
|
||||
int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order);
|
||||
void sev_gmem_invalidate(kvm_pfn_t start, kvm_pfn_t end);
|
||||
int sev_gmem_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn, bool is_private);
|
||||
|
|
@ -880,6 +881,7 @@ static inline struct page *snp_safe_alloc_page(void)
|
|||
return snp_safe_alloc_page_node(numa_node_id(), GFP_KERNEL_ACCOUNT);
|
||||
}
|
||||
|
||||
static inline int sev_vcpu_create(struct kvm_vcpu *vcpu) { return 0; }
|
||||
static inline void sev_free_vcpu(struct kvm_vcpu *vcpu) {}
|
||||
static inline void sev_vm_destroy(struct kvm *kvm) {}
|
||||
static inline void __init sev_set_cpu_caps(void) {}
|
||||
|
|
@ -889,7 +891,6 @@ static inline int sev_cpu_init(struct svm_cpu_data *sd) { return 0; }
|
|||
static inline int sev_dev_get_attr(u32 group, u64 attr, u64 *val) { return -ENXIO; }
|
||||
#define max_sev_asid 0
|
||||
static inline void sev_handle_rmp_fault(struct kvm_vcpu *vcpu, gpa_t gpa, u64 error_code) {}
|
||||
static inline void sev_snp_init_protected_guest_state(struct kvm_vcpu *vcpu) {}
|
||||
static inline int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order)
|
||||
{
|
||||
return 0;
|
||||
|
|
@ -914,16 +915,21 @@ void __svm_sev_es_vcpu_run(struct vcpu_svm *svm, bool spec_ctrl_intercepted,
|
|||
void __svm_vcpu_run(struct vcpu_svm *svm, bool spec_ctrl_intercepted);
|
||||
|
||||
#define DEFINE_KVM_GHCB_ACCESSORS(field) \
|
||||
static __always_inline bool kvm_ghcb_##field##_is_valid(const struct vcpu_svm *svm) \
|
||||
{ \
|
||||
return test_bit(GHCB_BITMAP_IDX(field), \
|
||||
(unsigned long *)&svm->sev_es.valid_bitmap); \
|
||||
} \
|
||||
\
|
||||
static __always_inline u64 kvm_ghcb_get_##field##_if_valid(struct vcpu_svm *svm, struct ghcb *ghcb) \
|
||||
{ \
|
||||
return kvm_ghcb_##field##_is_valid(svm) ? ghcb->save.field : 0; \
|
||||
} \
|
||||
static __always_inline u64 kvm_ghcb_get_##field(struct vcpu_svm *svm) \
|
||||
{ \
|
||||
return READ_ONCE(svm->sev_es.ghcb->save.field); \
|
||||
} \
|
||||
\
|
||||
static __always_inline bool kvm_ghcb_##field##_is_valid(const struct vcpu_svm *svm) \
|
||||
{ \
|
||||
return test_bit(GHCB_BITMAP_IDX(field), \
|
||||
(unsigned long *)&svm->sev_es.valid_bitmap); \
|
||||
} \
|
||||
\
|
||||
static __always_inline u64 kvm_ghcb_get_##field##_if_valid(struct vcpu_svm *svm) \
|
||||
{ \
|
||||
return kvm_ghcb_##field##_is_valid(svm) ? kvm_ghcb_get_##field(svm) : 0; \
|
||||
}
|
||||
|
||||
DEFINE_KVM_GHCB_ACCESSORS(cpl)
|
||||
DEFINE_KVM_GHCB_ACCESSORS(rax)
|
||||
|
|
|
|||
|
|
@ -15,7 +15,7 @@
|
|||
#include "kvm_onhyperv.h"
|
||||
#include "svm_onhyperv.h"
|
||||
|
||||
int svm_hv_enable_l2_tlb_flush(struct kvm_vcpu *vcpu)
|
||||
static int svm_hv_enable_l2_tlb_flush(struct kvm_vcpu *vcpu)
|
||||
{
|
||||
struct hv_vmcb_enlightenments *hve;
|
||||
hpa_t partition_assist_page = hv_get_partition_assist_page(vcpu);
|
||||
|
|
@ -35,3 +35,29 @@ int svm_hv_enable_l2_tlb_flush(struct kvm_vcpu *vcpu)
|
|||
return 0;
|
||||
}
|
||||
|
||||
__init void svm_hv_hardware_setup(void)
|
||||
{
|
||||
if (npt_enabled &&
|
||||
ms_hyperv.nested_features & HV_X64_NESTED_ENLIGHTENED_TLB) {
|
||||
pr_info(KBUILD_MODNAME ": Hyper-V enlightened NPT TLB flush enabled\n");
|
||||
svm_x86_ops.flush_remote_tlbs = hv_flush_remote_tlbs;
|
||||
svm_x86_ops.flush_remote_tlbs_range = hv_flush_remote_tlbs_range;
|
||||
}
|
||||
|
||||
if (ms_hyperv.nested_features & HV_X64_NESTED_DIRECT_FLUSH) {
|
||||
int cpu;
|
||||
|
||||
pr_info(KBUILD_MODNAME ": Hyper-V Direct TLB Flush enabled\n");
|
||||
for_each_online_cpu(cpu) {
|
||||
struct hv_vp_assist_page *vp_ap =
|
||||
hv_get_vp_assist_page(cpu);
|
||||
|
||||
if (!vp_ap)
|
||||
continue;
|
||||
|
||||
vp_ap->nested_control.features.directhypercall = 1;
|
||||
}
|
||||
svm_x86_ops.enable_l2_tlb_flush =
|
||||
svm_hv_enable_l2_tlb_flush;
|
||||
}
|
||||
}
|
||||
|
|
|
|||
|
|
@ -13,9 +13,7 @@
|
|||
#include "kvm_onhyperv.h"
|
||||
#include "svm/hyperv.h"
|
||||
|
||||
static struct kvm_x86_ops svm_x86_ops;
|
||||
|
||||
int svm_hv_enable_l2_tlb_flush(struct kvm_vcpu *vcpu);
|
||||
__init void svm_hv_hardware_setup(void);
|
||||
|
||||
static inline bool svm_hv_is_enlightened_tlb_enabled(struct kvm_vcpu *vcpu)
|
||||
{
|
||||
|
|
@ -40,33 +38,6 @@ static inline void svm_hv_init_vmcb(struct vmcb *vmcb)
|
|||
hve->hv_enlightenments_control.msr_bitmap = 1;
|
||||
}
|
||||
|
||||
static inline __init void svm_hv_hardware_setup(void)
|
||||
{
|
||||
if (npt_enabled &&
|
||||
ms_hyperv.nested_features & HV_X64_NESTED_ENLIGHTENED_TLB) {
|
||||
pr_info(KBUILD_MODNAME ": Hyper-V enlightened NPT TLB flush enabled\n");
|
||||
svm_x86_ops.flush_remote_tlbs = hv_flush_remote_tlbs;
|
||||
svm_x86_ops.flush_remote_tlbs_range = hv_flush_remote_tlbs_range;
|
||||
}
|
||||
|
||||
if (ms_hyperv.nested_features & HV_X64_NESTED_DIRECT_FLUSH) {
|
||||
int cpu;
|
||||
|
||||
pr_info(KBUILD_MODNAME ": Hyper-V Direct TLB Flush enabled\n");
|
||||
for_each_online_cpu(cpu) {
|
||||
struct hv_vp_assist_page *vp_ap =
|
||||
hv_get_vp_assist_page(cpu);
|
||||
|
||||
if (!vp_ap)
|
||||
continue;
|
||||
|
||||
vp_ap->nested_control.features.directhypercall = 1;
|
||||
}
|
||||
svm_x86_ops.enable_l2_tlb_flush =
|
||||
svm_hv_enable_l2_tlb_flush;
|
||||
}
|
||||
}
|
||||
|
||||
static inline void svm_hv_vmcb_dirty_nested_enlightenments(
|
||||
struct kvm_vcpu *vcpu)
|
||||
{
|
||||
|
|
|
|||
|
|
@ -677,6 +677,12 @@ void kvm_user_return_msr_update_cache(unsigned int slot, u64 value)
|
|||
}
|
||||
EXPORT_SYMBOL_GPL(kvm_user_return_msr_update_cache);
|
||||
|
||||
u64 kvm_get_user_return_msr(unsigned int slot)
|
||||
{
|
||||
return this_cpu_ptr(user_return_msrs)->values[slot].curr;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(kvm_get_user_return_msr);
|
||||
|
||||
static void drop_user_return_notifiers(void)
|
||||
{
|
||||
struct kvm_user_return_msrs *msrs = this_cpu_ptr(user_return_msrs);
|
||||
|
|
@ -1237,7 +1243,7 @@ static inline u64 kvm_guest_supported_xfd(struct kvm_vcpu *vcpu)
|
|||
}
|
||||
#endif
|
||||
|
||||
static int __kvm_set_xcr(struct kvm_vcpu *vcpu, u32 index, u64 xcr)
|
||||
int __kvm_set_xcr(struct kvm_vcpu *vcpu, u32 index, u64 xcr)
|
||||
{
|
||||
u64 xcr0 = xcr;
|
||||
u64 old_xcr0 = vcpu->arch.xcr0;
|
||||
|
|
@ -1281,6 +1287,7 @@ static int __kvm_set_xcr(struct kvm_vcpu *vcpu, u32 index, u64 xcr)
|
|||
vcpu->arch.cpuid_dynamic_bits_dirty = true;
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(__kvm_set_xcr);
|
||||
|
||||
int kvm_emulate_xsetbv(struct kvm_vcpu *vcpu)
|
||||
{
|
||||
|
|
|
|||
|
|
@ -716,7 +716,8 @@ long kvm_gmem_populate(struct kvm *kvm, gfn_t start_gfn, void __user *src, long
|
|||
long i;
|
||||
|
||||
lockdep_assert_held(&kvm->slots_lock);
|
||||
if (npages < 0)
|
||||
|
||||
if (WARN_ON_ONCE(npages <= 0))
|
||||
return -EINVAL;
|
||||
|
||||
slot = gfn_to_memslot(kvm, start_gfn);
|
||||
|
|
|
|||
Loading…
Reference in New Issue