|
|
|
|
@@ -140,7 +140,7 @@ static inline bool is_mirroring_enc_context(struct kvm *kvm)
|
|
|
|
|
static bool sev_vcpu_has_debug_swap(struct vcpu_svm *svm)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_vcpu *vcpu = &svm->vcpu;
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(vcpu->kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(vcpu->kvm);
|
|
|
|
|
|
|
|
|
|
return sev->vmsa_features & SVM_SEV_FEAT_DEBUG_SWAP;
|
|
|
|
|
}
|
|
|
|
|
@@ -226,9 +226,7 @@ static int sev_asid_new(struct kvm_sev_info *sev)
|
|
|
|
|
|
|
|
|
|
static unsigned int sev_get_asid(struct kvm *kvm)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
|
|
|
|
|
return sev->asid;
|
|
|
|
|
return to_kvm_sev_info(kvm)->asid;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void sev_asid_free(struct kvm_sev_info *sev)
|
|
|
|
|
@@ -403,7 +401,7 @@ static int __sev_guest_init(struct kvm *kvm, struct kvm_sev_cmd *argp,
|
|
|
|
|
struct kvm_sev_init *data,
|
|
|
|
|
unsigned long vm_type)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
|
|
|
|
struct sev_platform_init_args init_args = {0};
|
|
|
|
|
bool es_active = vm_type != KVM_X86_SEV_VM;
|
|
|
|
|
u64 valid_vmsa_features = es_active ? sev_supported_vmsa_features : 0;
|
|
|
|
|
@@ -500,10 +498,9 @@ static int sev_guest_init(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
|
|
|
|
|
static int sev_guest_init2(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_init data;
|
|
|
|
|
|
|
|
|
|
if (!sev->need_init)
|
|
|
|
|
if (!to_kvm_sev_info(kvm)->need_init)
|
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
|
|
if (kvm->arch.vm_type != KVM_X86_SEV_VM &&
|
|
|
|
|
@@ -543,14 +540,14 @@ static int __sev_issue_cmd(int fd, int id, void *data, int *error)
|
|
|
|
|
|
|
|
|
|
static int sev_issue_cmd(struct kvm *kvm, int id, void *data, int *error)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
|
|
|
|
|
|
|
|
|
return __sev_issue_cmd(sev->fd, id, data, error);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static int sev_launch_start(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
|
|
|
|
struct sev_data_launch_start start;
|
|
|
|
|
struct kvm_sev_launch_start params;
|
|
|
|
|
void *dh_blob, *session_blob;
|
|
|
|
|
@@ -622,9 +619,9 @@ static int sev_launch_start(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
|
|
|
|
|
static struct page **sev_pin_memory(struct kvm *kvm, unsigned long uaddr,
|
|
|
|
|
unsigned long ulen, unsigned long *n,
|
|
|
|
|
int write)
|
|
|
|
|
unsigned int flags)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
|
|
|
|
unsigned long npages, size;
|
|
|
|
|
int npinned;
|
|
|
|
|
unsigned long locked, lock_limit;
|
|
|
|
|
@@ -663,7 +660,7 @@ static struct page **sev_pin_memory(struct kvm *kvm, unsigned long uaddr,
|
|
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
|
|
|
|
|
|
/* Pin the user virtual address. */
|
|
|
|
|
npinned = pin_user_pages_fast(uaddr, npages, write ? FOLL_WRITE : 0, pages);
|
|
|
|
|
npinned = pin_user_pages_fast(uaddr, npages, flags, pages);
|
|
|
|
|
if (npinned != npages) {
|
|
|
|
|
pr_err("SEV: Failure locking %lu pages.\n", npages);
|
|
|
|
|
ret = -ENOMEM;
|
|
|
|
|
@@ -686,11 +683,9 @@ static struct page **sev_pin_memory(struct kvm *kvm, unsigned long uaddr,
|
|
|
|
|
static void sev_unpin_memory(struct kvm *kvm, struct page **pages,
|
|
|
|
|
unsigned long npages)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
|
|
|
|
|
unpin_user_pages(pages, npages);
|
|
|
|
|
kvfree(pages);
|
|
|
|
|
sev->pages_locked -= npages;
|
|
|
|
|
to_kvm_sev_info(kvm)->pages_locked -= npages;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void sev_clflush_pages(struct page *pages[], unsigned long npages)
|
|
|
|
|
@@ -734,7 +729,6 @@ static unsigned long get_num_contig_pages(unsigned long idx,
|
|
|
|
|
static int sev_launch_update_data(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
unsigned long vaddr, vaddr_end, next_vaddr, npages, pages, size, i;
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_launch_update_data params;
|
|
|
|
|
struct sev_data_launch_update_data data;
|
|
|
|
|
struct page **inpages;
|
|
|
|
|
@@ -751,7 +745,7 @@ static int sev_launch_update_data(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
vaddr_end = vaddr + size;
|
|
|
|
|
|
|
|
|
|
/* Lock the user memory. */
|
|
|
|
|
inpages = sev_pin_memory(kvm, vaddr, size, &npages, 1);
|
|
|
|
|
inpages = sev_pin_memory(kvm, vaddr, size, &npages, FOLL_WRITE);
|
|
|
|
|
if (IS_ERR(inpages))
|
|
|
|
|
return PTR_ERR(inpages);
|
|
|
|
|
|
|
|
|
|
@@ -762,7 +756,7 @@ static int sev_launch_update_data(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
sev_clflush_pages(inpages, npages);
|
|
|
|
|
|
|
|
|
|
data.reserved = 0;
|
|
|
|
|
data.handle = sev->handle;
|
|
|
|
|
data.handle = to_kvm_sev_info(kvm)->handle;
|
|
|
|
|
|
|
|
|
|
for (i = 0; vaddr < vaddr_end; vaddr = next_vaddr, i += pages) {
|
|
|
|
|
int offset, len;
|
|
|
|
|
@@ -802,7 +796,7 @@ static int sev_launch_update_data(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
static int sev_es_sync_vmsa(struct vcpu_svm *svm)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_vcpu *vcpu = &svm->vcpu;
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(vcpu->kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(vcpu->kvm);
|
|
|
|
|
struct sev_es_save_area *save = svm->sev_es.vmsa;
|
|
|
|
|
struct xregs_state *xsave;
|
|
|
|
|
const u8 *s;
|
|
|
|
|
@@ -972,7 +966,6 @@ static int sev_launch_update_vmsa(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
static int sev_launch_measure(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
void __user *measure = u64_to_user_ptr(argp->data);
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct sev_data_launch_measure data;
|
|
|
|
|
struct kvm_sev_launch_measure params;
|
|
|
|
|
void __user *p = NULL;
|
|
|
|
|
@@ -1005,7 +998,7 @@ static int sev_launch_measure(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
cmd:
|
|
|
|
|
data.handle = sev->handle;
|
|
|
|
|
data.handle = to_kvm_sev_info(kvm)->handle;
|
|
|
|
|
ret = sev_issue_cmd(kvm, SEV_CMD_LAUNCH_MEASURE, &data, &argp->error);
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
@@ -1033,19 +1026,17 @@ static int sev_launch_measure(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
|
|
|
|
|
static int sev_launch_finish(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct sev_data_launch_finish data;
|
|
|
|
|
|
|
|
|
|
if (!sev_guest(kvm))
|
|
|
|
|
return -ENOTTY;
|
|
|
|
|
|
|
|
|
|
data.handle = sev->handle;
|
|
|
|
|
data.handle = to_kvm_sev_info(kvm)->handle;
|
|
|
|
|
return sev_issue_cmd(kvm, SEV_CMD_LAUNCH_FINISH, &data, &argp->error);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static int sev_guest_status(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_guest_status params;
|
|
|
|
|
struct sev_data_guest_status data;
|
|
|
|
|
int ret;
|
|
|
|
|
@@ -1055,7 +1046,7 @@ static int sev_guest_status(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
|
|
|
|
|
memset(&data, 0, sizeof(data));
|
|
|
|
|
|
|
|
|
|
data.handle = sev->handle;
|
|
|
|
|
data.handle = to_kvm_sev_info(kvm)->handle;
|
|
|
|
|
ret = sev_issue_cmd(kvm, SEV_CMD_GUEST_STATUS, &data, &argp->error);
|
|
|
|
|
if (ret)
|
|
|
|
|
return ret;
|
|
|
|
|
@@ -1074,11 +1065,10 @@ static int __sev_issue_dbg_cmd(struct kvm *kvm, unsigned long src,
|
|
|
|
|
unsigned long dst, int size,
|
|
|
|
|
int *error, bool enc)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct sev_data_dbg data;
|
|
|
|
|
|
|
|
|
|
data.reserved = 0;
|
|
|
|
|
data.handle = sev->handle;
|
|
|
|
|
data.handle = to_kvm_sev_info(kvm)->handle;
|
|
|
|
|
data.dst_addr = dst;
|
|
|
|
|
data.src_addr = src;
|
|
|
|
|
data.len = size;
|
|
|
|
|
@@ -1250,7 +1240,7 @@ static int sev_dbg_crypt(struct kvm *kvm, struct kvm_sev_cmd *argp, bool dec)
|
|
|
|
|
if (IS_ERR(src_p))
|
|
|
|
|
return PTR_ERR(src_p);
|
|
|
|
|
|
|
|
|
|
dst_p = sev_pin_memory(kvm, dst_vaddr & PAGE_MASK, PAGE_SIZE, &n, 1);
|
|
|
|
|
dst_p = sev_pin_memory(kvm, dst_vaddr & PAGE_MASK, PAGE_SIZE, &n, FOLL_WRITE);
|
|
|
|
|
if (IS_ERR(dst_p)) {
|
|
|
|
|
sev_unpin_memory(kvm, src_p, n);
|
|
|
|
|
return PTR_ERR(dst_p);
|
|
|
|
|
@@ -1302,7 +1292,6 @@ static int sev_dbg_crypt(struct kvm *kvm, struct kvm_sev_cmd *argp, bool dec)
|
|
|
|
|
|
|
|
|
|
static int sev_launch_secret(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct sev_data_launch_secret data;
|
|
|
|
|
struct kvm_sev_launch_secret params;
|
|
|
|
|
struct page **pages;
|
|
|
|
|
@@ -1316,7 +1305,7 @@ static int sev_launch_secret(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
if (copy_from_user(¶ms, u64_to_user_ptr(argp->data), sizeof(params)))
|
|
|
|
|
return -EFAULT;
|
|
|
|
|
|
|
|
|
|
pages = sev_pin_memory(kvm, params.guest_uaddr, params.guest_len, &n, 1);
|
|
|
|
|
pages = sev_pin_memory(kvm, params.guest_uaddr, params.guest_len, &n, FOLL_WRITE);
|
|
|
|
|
if (IS_ERR(pages))
|
|
|
|
|
return PTR_ERR(pages);
|
|
|
|
|
|
|
|
|
|
@@ -1358,7 +1347,7 @@ static int sev_launch_secret(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
data.hdr_address = __psp_pa(hdr);
|
|
|
|
|
data.hdr_len = params.hdr_len;
|
|
|
|
|
|
|
|
|
|
data.handle = sev->handle;
|
|
|
|
|
data.handle = to_kvm_sev_info(kvm)->handle;
|
|
|
|
|
ret = sev_issue_cmd(kvm, SEV_CMD_LAUNCH_UPDATE_SECRET, &data, &argp->error);
|
|
|
|
|
|
|
|
|
|
kfree(hdr);
|
|
|
|
|
@@ -1378,7 +1367,6 @@ static int sev_launch_secret(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
static int sev_get_attestation_report(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
void __user *report = u64_to_user_ptr(argp->data);
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct sev_data_attestation_report data;
|
|
|
|
|
struct kvm_sev_attestation_report params;
|
|
|
|
|
void __user *p;
|
|
|
|
|
@@ -1411,7 +1399,7 @@ static int sev_get_attestation_report(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
memcpy(data.mnonce, params.mnonce, sizeof(params.mnonce));
|
|
|
|
|
}
|
|
|
|
|
cmd:
|
|
|
|
|
data.handle = sev->handle;
|
|
|
|
|
data.handle = to_kvm_sev_info(kvm)->handle;
|
|
|
|
|
ret = sev_issue_cmd(kvm, SEV_CMD_ATTESTATION_REPORT, &data, &argp->error);
|
|
|
|
|
/*
|
|
|
|
|
* If we query the session length, FW responded with expected data.
|
|
|
|
|
@@ -1441,12 +1429,11 @@ static int
|
|
|
|
|
__sev_send_start_query_session_length(struct kvm *kvm, struct kvm_sev_cmd *argp,
|
|
|
|
|
struct kvm_sev_send_start *params)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct sev_data_send_start data;
|
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
|
|
memset(&data, 0, sizeof(data));
|
|
|
|
|
data.handle = sev->handle;
|
|
|
|
|
data.handle = to_kvm_sev_info(kvm)->handle;
|
|
|
|
|
ret = sev_issue_cmd(kvm, SEV_CMD_SEND_START, &data, &argp->error);
|
|
|
|
|
|
|
|
|
|
params->session_len = data.session_len;
|
|
|
|
|
@@ -1459,7 +1446,6 @@ __sev_send_start_query_session_length(struct kvm *kvm, struct kvm_sev_cmd *argp,
|
|
|
|
|
|
|
|
|
|
static int sev_send_start(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct sev_data_send_start data;
|
|
|
|
|
struct kvm_sev_send_start params;
|
|
|
|
|
void *amd_certs, *session_data;
|
|
|
|
|
@@ -1520,7 +1506,7 @@ static int sev_send_start(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
data.amd_certs_len = params.amd_certs_len;
|
|
|
|
|
data.session_address = __psp_pa(session_data);
|
|
|
|
|
data.session_len = params.session_len;
|
|
|
|
|
data.handle = sev->handle;
|
|
|
|
|
data.handle = to_kvm_sev_info(kvm)->handle;
|
|
|
|
|
|
|
|
|
|
ret = sev_issue_cmd(kvm, SEV_CMD_SEND_START, &data, &argp->error);
|
|
|
|
|
|
|
|
|
|
@@ -1552,12 +1538,11 @@ static int
|
|
|
|
|
__sev_send_update_data_query_lengths(struct kvm *kvm, struct kvm_sev_cmd *argp,
|
|
|
|
|
struct kvm_sev_send_update_data *params)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct sev_data_send_update_data data;
|
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
|
|
memset(&data, 0, sizeof(data));
|
|
|
|
|
data.handle = sev->handle;
|
|
|
|
|
data.handle = to_kvm_sev_info(kvm)->handle;
|
|
|
|
|
ret = sev_issue_cmd(kvm, SEV_CMD_SEND_UPDATE_DATA, &data, &argp->error);
|
|
|
|
|
|
|
|
|
|
params->hdr_len = data.hdr_len;
|
|
|
|
|
@@ -1572,7 +1557,6 @@ __sev_send_update_data_query_lengths(struct kvm *kvm, struct kvm_sev_cmd *argp,
|
|
|
|
|
|
|
|
|
|
static int sev_send_update_data(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct sev_data_send_update_data data;
|
|
|
|
|
struct kvm_sev_send_update_data params;
|
|
|
|
|
void *hdr, *trans_data;
|
|
|
|
|
@@ -1626,7 +1610,7 @@ static int sev_send_update_data(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
data.guest_address = (page_to_pfn(guest_page[0]) << PAGE_SHIFT) + offset;
|
|
|
|
|
data.guest_address |= sev_me_mask;
|
|
|
|
|
data.guest_len = params.guest_len;
|
|
|
|
|
data.handle = sev->handle;
|
|
|
|
|
data.handle = to_kvm_sev_info(kvm)->handle;
|
|
|
|
|
|
|
|
|
|
ret = sev_issue_cmd(kvm, SEV_CMD_SEND_UPDATE_DATA, &data, &argp->error);
|
|
|
|
|
|
|
|
|
|
@@ -1657,31 +1641,29 @@ static int sev_send_update_data(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
|
|
|
|
|
static int sev_send_finish(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct sev_data_send_finish data;
|
|
|
|
|
|
|
|
|
|
if (!sev_guest(kvm))
|
|
|
|
|
return -ENOTTY;
|
|
|
|
|
|
|
|
|
|
data.handle = sev->handle;
|
|
|
|
|
data.handle = to_kvm_sev_info(kvm)->handle;
|
|
|
|
|
return sev_issue_cmd(kvm, SEV_CMD_SEND_FINISH, &data, &argp->error);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static int sev_send_cancel(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct sev_data_send_cancel data;
|
|
|
|
|
|
|
|
|
|
if (!sev_guest(kvm))
|
|
|
|
|
return -ENOTTY;
|
|
|
|
|
|
|
|
|
|
data.handle = sev->handle;
|
|
|
|
|
data.handle = to_kvm_sev_info(kvm)->handle;
|
|
|
|
|
return sev_issue_cmd(kvm, SEV_CMD_SEND_CANCEL, &data, &argp->error);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static int sev_receive_start(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
|
|
|
|
struct sev_data_receive_start start;
|
|
|
|
|
struct kvm_sev_receive_start params;
|
|
|
|
|
int *error = &argp->error;
|
|
|
|
|
@@ -1755,7 +1737,6 @@ static int sev_receive_start(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
|
|
|
|
|
static int sev_receive_update_data(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_receive_update_data params;
|
|
|
|
|
struct sev_data_receive_update_data data;
|
|
|
|
|
void *hdr = NULL, *trans = NULL;
|
|
|
|
|
@@ -1798,7 +1779,7 @@ static int sev_receive_update_data(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
|
|
|
|
|
/* Pin guest memory */
|
|
|
|
|
guest_page = sev_pin_memory(kvm, params.guest_uaddr & PAGE_MASK,
|
|
|
|
|
PAGE_SIZE, &n, 1);
|
|
|
|
|
PAGE_SIZE, &n, FOLL_WRITE);
|
|
|
|
|
if (IS_ERR(guest_page)) {
|
|
|
|
|
ret = PTR_ERR(guest_page);
|
|
|
|
|
goto e_free_trans;
|
|
|
|
|
@@ -1815,7 +1796,7 @@ static int sev_receive_update_data(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
data.guest_address = (page_to_pfn(guest_page[0]) << PAGE_SHIFT) + offset;
|
|
|
|
|
data.guest_address |= sev_me_mask;
|
|
|
|
|
data.guest_len = params.guest_len;
|
|
|
|
|
data.handle = sev->handle;
|
|
|
|
|
data.handle = to_kvm_sev_info(kvm)->handle;
|
|
|
|
|
|
|
|
|
|
ret = sev_issue_cmd(kvm, SEV_CMD_RECEIVE_UPDATE_DATA, &data,
|
|
|
|
|
&argp->error);
|
|
|
|
|
@@ -1832,13 +1813,12 @@ static int sev_receive_update_data(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
|
|
|
|
|
static int sev_receive_finish(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct sev_data_receive_finish data;
|
|
|
|
|
|
|
|
|
|
if (!sev_guest(kvm))
|
|
|
|
|
return -ENOTTY;
|
|
|
|
|
|
|
|
|
|
data.handle = sev->handle;
|
|
|
|
|
data.handle = to_kvm_sev_info(kvm)->handle;
|
|
|
|
|
return sev_issue_cmd(kvm, SEV_CMD_RECEIVE_FINISH, &data, &argp->error);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
@@ -1858,8 +1838,8 @@ static bool is_cmd_allowed_from_mirror(u32 cmd_id)
|
|
|
|
|
|
|
|
|
|
static int sev_lock_two_vms(struct kvm *dst_kvm, struct kvm *src_kvm)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *dst_sev = &to_kvm_svm(dst_kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *src_sev = &to_kvm_svm(src_kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *dst_sev = to_kvm_sev_info(dst_kvm);
|
|
|
|
|
struct kvm_sev_info *src_sev = to_kvm_sev_info(src_kvm);
|
|
|
|
|
int r = -EBUSY;
|
|
|
|
|
|
|
|
|
|
if (dst_kvm == src_kvm)
|
|
|
|
|
@@ -1893,8 +1873,8 @@ static int sev_lock_two_vms(struct kvm *dst_kvm, struct kvm *src_kvm)
|
|
|
|
|
|
|
|
|
|
static void sev_unlock_two_vms(struct kvm *dst_kvm, struct kvm *src_kvm)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *dst_sev = &to_kvm_svm(dst_kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *src_sev = &to_kvm_svm(src_kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *dst_sev = to_kvm_sev_info(dst_kvm);
|
|
|
|
|
struct kvm_sev_info *src_sev = to_kvm_sev_info(src_kvm);
|
|
|
|
|
|
|
|
|
|
mutex_unlock(&dst_kvm->lock);
|
|
|
|
|
mutex_unlock(&src_kvm->lock);
|
|
|
|
|
@@ -1968,8 +1948,8 @@ static void sev_unlock_vcpus_for_migration(struct kvm *kvm)
|
|
|
|
|
|
|
|
|
|
static void sev_migrate_from(struct kvm *dst_kvm, struct kvm *src_kvm)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *dst = &to_kvm_svm(dst_kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *src = &to_kvm_svm(src_kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *dst = to_kvm_sev_info(dst_kvm);
|
|
|
|
|
struct kvm_sev_info *src = to_kvm_sev_info(src_kvm);
|
|
|
|
|
struct kvm_vcpu *dst_vcpu, *src_vcpu;
|
|
|
|
|
struct vcpu_svm *dst_svm, *src_svm;
|
|
|
|
|
struct kvm_sev_info *mirror;
|
|
|
|
|
@@ -2009,8 +1989,7 @@ static void sev_migrate_from(struct kvm *dst_kvm, struct kvm *src_kvm)
|
|
|
|
|
* and add the new mirror to the list.
|
|
|
|
|
*/
|
|
|
|
|
if (is_mirroring_enc_context(dst_kvm)) {
|
|
|
|
|
struct kvm_sev_info *owner_sev_info =
|
|
|
|
|
&to_kvm_svm(dst->enc_context_owner)->sev_info;
|
|
|
|
|
struct kvm_sev_info *owner_sev_info = to_kvm_sev_info(dst->enc_context_owner);
|
|
|
|
|
|
|
|
|
|
list_del(&src->mirror_entry);
|
|
|
|
|
list_add_tail(&dst->mirror_entry, &owner_sev_info->mirror_vms);
|
|
|
|
|
@@ -2069,7 +2048,7 @@ static int sev_check_source_vcpus(struct kvm *dst, struct kvm *src)
|
|
|
|
|
|
|
|
|
|
int sev_vm_move_enc_context_from(struct kvm *kvm, unsigned int source_fd)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *dst_sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *dst_sev = to_kvm_sev_info(kvm);
|
|
|
|
|
struct kvm_sev_info *src_sev, *cg_cleanup_sev;
|
|
|
|
|
CLASS(fd, f)(source_fd);
|
|
|
|
|
struct kvm *source_kvm;
|
|
|
|
|
@@ -2093,7 +2072,7 @@ int sev_vm_move_enc_context_from(struct kvm *kvm, unsigned int source_fd)
|
|
|
|
|
goto out_unlock;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
src_sev = &to_kvm_svm(source_kvm)->sev_info;
|
|
|
|
|
src_sev = to_kvm_sev_info(source_kvm);
|
|
|
|
|
|
|
|
|
|
dst_sev->misc_cg = get_current_misc_cg();
|
|
|
|
|
cg_cleanup_sev = dst_sev;
|
|
|
|
|
@@ -2181,7 +2160,7 @@ static void *snp_context_create(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
|
|
|
|
|
static int snp_bind_asid(struct kvm *kvm, int *error)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
|
|
|
|
struct sev_data_snp_activate data = {0};
|
|
|
|
|
|
|
|
|
|
data.gctx_paddr = __psp_pa(sev->snp_context);
|
|
|
|
|
@@ -2191,7 +2170,7 @@ static int snp_bind_asid(struct kvm *kvm, int *error)
|
|
|
|
|
|
|
|
|
|
static int snp_launch_start(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
|
|
|
|
struct sev_data_snp_launch_start start = {0};
|
|
|
|
|
struct kvm_sev_snp_launch_start params;
|
|
|
|
|
int rc;
|
|
|
|
|
@@ -2260,7 +2239,7 @@ static int sev_gmem_post_populate(struct kvm *kvm, gfn_t gfn_start, kvm_pfn_t pf
|
|
|
|
|
void __user *src, int order, void *opaque)
|
|
|
|
|
{
|
|
|
|
|
struct sev_gmem_populate_args *sev_populate_args = opaque;
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
|
|
|
|
int n_private = 0, ret, i;
|
|
|
|
|
int npages = (1 << order);
|
|
|
|
|
gfn_t gfn;
|
|
|
|
|
@@ -2350,7 +2329,7 @@ static int sev_gmem_post_populate(struct kvm *kvm, gfn_t gfn_start, kvm_pfn_t pf
|
|
|
|
|
|
|
|
|
|
static int snp_launch_update(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
|
|
|
|
struct sev_gmem_populate_args sev_populate_args = {0};
|
|
|
|
|
struct kvm_sev_snp_launch_update params;
|
|
|
|
|
struct kvm_memory_slot *memslot;
|
|
|
|
|
@@ -2434,7 +2413,7 @@ static int snp_launch_update(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
|
|
|
|
|
static int snp_launch_update_vmsa(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
|
|
|
|
struct sev_data_snp_launch_update data = {};
|
|
|
|
|
struct kvm_vcpu *vcpu;
|
|
|
|
|
unsigned long i;
|
|
|
|
|
@@ -2482,7 +2461,7 @@ static int snp_launch_update_vmsa(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
|
|
|
|
|
static int snp_launch_finish(struct kvm *kvm, struct kvm_sev_cmd *argp)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
|
|
|
|
struct kvm_sev_snp_launch_finish params;
|
|
|
|
|
struct sev_data_snp_launch_finish *data;
|
|
|
|
|
void *id_block = NULL, *id_auth = NULL;
|
|
|
|
|
@@ -2677,7 +2656,7 @@ int sev_mem_enc_ioctl(struct kvm *kvm, void __user *argp)
|
|
|
|
|
int sev_mem_enc_register_region(struct kvm *kvm,
|
|
|
|
|
struct kvm_enc_region *range)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
|
|
|
|
struct enc_region *region;
|
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
|
|
@@ -2696,7 +2675,8 @@ int sev_mem_enc_register_region(struct kvm *kvm,
|
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
|
|
mutex_lock(&kvm->lock);
|
|
|
|
|
region->pages = sev_pin_memory(kvm, range->addr, range->size, ®ion->npages, 1);
|
|
|
|
|
region->pages = sev_pin_memory(kvm, range->addr, range->size, ®ion->npages,
|
|
|
|
|
FOLL_WRITE | FOLL_LONGTERM);
|
|
|
|
|
if (IS_ERR(region->pages)) {
|
|
|
|
|
ret = PTR_ERR(region->pages);
|
|
|
|
|
mutex_unlock(&kvm->lock);
|
|
|
|
|
@@ -2729,7 +2709,7 @@ int sev_mem_enc_register_region(struct kvm *kvm,
|
|
|
|
|
static struct enc_region *
|
|
|
|
|
find_enc_region(struct kvm *kvm, struct kvm_enc_region *range)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
|
|
|
|
struct list_head *head = &sev->regions_list;
|
|
|
|
|
struct enc_region *i;
|
|
|
|
|
|
|
|
|
|
@@ -2824,9 +2804,9 @@ int sev_vm_copy_enc_context_from(struct kvm *kvm, unsigned int source_fd)
|
|
|
|
|
* The mirror kvm holds an enc_context_owner ref so its asid can't
|
|
|
|
|
* disappear until we're done with it
|
|
|
|
|
*/
|
|
|
|
|
source_sev = &to_kvm_svm(source_kvm)->sev_info;
|
|
|
|
|
source_sev = to_kvm_sev_info(source_kvm);
|
|
|
|
|
kvm_get_kvm(source_kvm);
|
|
|
|
|
mirror_sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
mirror_sev = to_kvm_sev_info(kvm);
|
|
|
|
|
list_add_tail(&mirror_sev->mirror_entry, &source_sev->mirror_vms);
|
|
|
|
|
|
|
|
|
|
/* Set enc_context_owner and copy its encryption context over */
|
|
|
|
|
@@ -2854,7 +2834,7 @@ int sev_vm_copy_enc_context_from(struct kvm *kvm, unsigned int source_fd)
|
|
|
|
|
|
|
|
|
|
static int snp_decommission_context(struct kvm *kvm)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
|
|
|
|
struct sev_data_snp_addr data = {};
|
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
|
|
@@ -2879,7 +2859,7 @@ static int snp_decommission_context(struct kvm *kvm)
|
|
|
|
|
|
|
|
|
|
void sev_vm_destroy(struct kvm *kvm)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
|
|
|
|
struct list_head *head = &sev->regions_list;
|
|
|
|
|
struct list_head *pos, *q;
|
|
|
|
|
|
|
|
|
|
@@ -3430,8 +3410,7 @@ static int sev_es_validate_vmgexit(struct vcpu_svm *svm)
|
|
|
|
|
dump_ghcb(svm);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
ghcb_set_sw_exit_info_1(svm->sev_es.ghcb, 2);
|
|
|
|
|
ghcb_set_sw_exit_info_2(svm->sev_es.ghcb, reason);
|
|
|
|
|
svm_vmgexit_bad_input(svm, reason);
|
|
|
|
|
|
|
|
|
|
/* Resume the guest to "return" the error code. */
|
|
|
|
|
return 1;
|
|
|
|
|
@@ -3472,10 +3451,19 @@ void sev_es_unmap_ghcb(struct vcpu_svm *svm)
|
|
|
|
|
svm->sev_es.ghcb = NULL;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
void pre_sev_run(struct vcpu_svm *svm, int cpu)
|
|
|
|
|
int pre_sev_run(struct vcpu_svm *svm, int cpu)
|
|
|
|
|
{
|
|
|
|
|
struct svm_cpu_data *sd = per_cpu_ptr(&svm_data, cpu);
|
|
|
|
|
unsigned int asid = sev_get_asid(svm->vcpu.kvm);
|
|
|
|
|
struct kvm *kvm = svm->vcpu.kvm;
|
|
|
|
|
unsigned int asid = sev_get_asid(kvm);
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Reject KVM_RUN if userspace attempts to run the vCPU with an invalid
|
|
|
|
|
* VMSA, e.g. if userspace forces the vCPU to be RUNNABLE after an SNP
|
|
|
|
|
* AP Destroy event.
|
|
|
|
|
*/
|
|
|
|
|
if (sev_es_guest(kvm) && !VALID_PAGE(svm->vmcb->control.vmsa_pa))
|
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
|
|
/* Assign the asid allocated with this SEV guest */
|
|
|
|
|
svm->asid = asid;
|
|
|
|
|
@@ -3488,11 +3476,12 @@ void pre_sev_run(struct vcpu_svm *svm, int cpu)
|
|
|
|
|
*/
|
|
|
|
|
if (sd->sev_vmcbs[asid] == svm->vmcb &&
|
|
|
|
|
svm->vcpu.arch.last_vmentry_cpu == cpu)
|
|
|
|
|
return;
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
sd->sev_vmcbs[asid] = svm->vmcb;
|
|
|
|
|
svm->vmcb->control.tlb_ctl = TLB_CONTROL_FLUSH_ASID;
|
|
|
|
|
vmcb_mark_dirty(svm->vmcb, VMCB_ASID);
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
#define GHCB_SCRATCH_AREA_LIMIT (16ULL * PAGE_SIZE)
|
|
|
|
|
@@ -3574,8 +3563,7 @@ static int setup_vmgexit_scratch(struct vcpu_svm *svm, bool sync, u64 len)
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
|
|
e_scratch:
|
|
|
|
|
ghcb_set_sw_exit_info_1(svm->sev_es.ghcb, 2);
|
|
|
|
|
ghcb_set_sw_exit_info_2(svm->sev_es.ghcb, GHCB_ERR_INVALID_SCRATCH_AREA);
|
|
|
|
|
svm_vmgexit_bad_input(svm, GHCB_ERR_INVALID_SCRATCH_AREA);
|
|
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
|
}
|
|
|
|
|
@@ -3675,7 +3663,14 @@ static void snp_complete_psc(struct vcpu_svm *svm, u64 psc_ret)
|
|
|
|
|
svm->sev_es.psc_inflight = 0;
|
|
|
|
|
svm->sev_es.psc_idx = 0;
|
|
|
|
|
svm->sev_es.psc_2m = false;
|
|
|
|
|
ghcb_set_sw_exit_info_2(svm->sev_es.ghcb, psc_ret);
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* PSC requests always get a "no action" response in SW_EXITINFO1, with
|
|
|
|
|
* a PSC-specific return code in SW_EXITINFO2 that provides the "real"
|
|
|
|
|
* return code. E.g. if the PSC request was interrupted, the need to
|
|
|
|
|
* retry is communicated via SW_EXITINFO2, not SW_EXITINFO1.
|
|
|
|
|
*/
|
|
|
|
|
svm_vmgexit_no_action(svm, psc_ret);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static void __snp_complete_one_psc(struct vcpu_svm *svm)
|
|
|
|
|
@@ -3847,11 +3842,26 @@ static int snp_begin_psc(struct vcpu_svm *svm, struct psc_buffer *psc)
|
|
|
|
|
BUG();
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static int __sev_snp_update_protected_guest_state(struct kvm_vcpu *vcpu)
|
|
|
|
|
/*
|
|
|
|
|
* Invoked as part of svm_vcpu_reset() processing of an init event.
|
|
|
|
|
*/
|
|
|
|
|
void sev_snp_init_protected_guest_state(struct kvm_vcpu *vcpu)
|
|
|
|
|
{
|
|
|
|
|
struct vcpu_svm *svm = to_svm(vcpu);
|
|
|
|
|
struct kvm_memory_slot *slot;
|
|
|
|
|
struct page *page;
|
|
|
|
|
kvm_pfn_t pfn;
|
|
|
|
|
gfn_t gfn;
|
|
|
|
|
|
|
|
|
|
WARN_ON(!mutex_is_locked(&svm->sev_es.snp_vmsa_mutex));
|
|
|
|
|
if (!sev_snp_guest(vcpu->kvm))
|
|
|
|
|
return;
|
|
|
|
|
|
|
|
|
|
guard(mutex)(&svm->sev_es.snp_vmsa_mutex);
|
|
|
|
|
|
|
|
|
|
if (!svm->sev_es.snp_ap_waiting_for_reset)
|
|
|
|
|
return;
|
|
|
|
|
|
|
|
|
|
svm->sev_es.snp_ap_waiting_for_reset = false;
|
|
|
|
|
|
|
|
|
|
/* Mark the vCPU as offline and not runnable */
|
|
|
|
|
vcpu->arch.pv.pv_unhalted = false;
|
|
|
|
|
@@ -3860,96 +3870,62 @@ static int __sev_snp_update_protected_guest_state(struct kvm_vcpu *vcpu)
|
|
|
|
|
/* Clear use of the VMSA */
|
|
|
|
|
svm->vmcb->control.vmsa_pa = INVALID_PAGE;
|
|
|
|
|
|
|
|
|
|
if (VALID_PAGE(svm->sev_es.snp_vmsa_gpa)) {
|
|
|
|
|
gfn_t gfn = gpa_to_gfn(svm->sev_es.snp_vmsa_gpa);
|
|
|
|
|
struct kvm_memory_slot *slot;
|
|
|
|
|
struct page *page;
|
|
|
|
|
kvm_pfn_t pfn;
|
|
|
|
|
|
|
|
|
|
slot = gfn_to_memslot(vcpu->kvm, gfn);
|
|
|
|
|
if (!slot)
|
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* The new VMSA will be private memory guest memory, so
|
|
|
|
|
* retrieve the PFN from the gmem backend.
|
|
|
|
|
*/
|
|
|
|
|
if (kvm_gmem_get_pfn(vcpu->kvm, slot, gfn, &pfn, &page, NULL))
|
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* From this point forward, the VMSA will always be a
|
|
|
|
|
* guest-mapped page rather than the initial one allocated
|
|
|
|
|
* by KVM in svm->sev_es.vmsa. In theory, svm->sev_es.vmsa
|
|
|
|
|
* could be free'd and cleaned up here, but that involves
|
|
|
|
|
* cleanups like wbinvd_on_all_cpus() which would ideally
|
|
|
|
|
* be handled during teardown rather than guest boot.
|
|
|
|
|
* Deferring that also allows the existing logic for SEV-ES
|
|
|
|
|
* VMSAs to be re-used with minimal SNP-specific changes.
|
|
|
|
|
*/
|
|
|
|
|
svm->sev_es.snp_has_guest_vmsa = true;
|
|
|
|
|
|
|
|
|
|
/* Use the new VMSA */
|
|
|
|
|
svm->vmcb->control.vmsa_pa = pfn_to_hpa(pfn);
|
|
|
|
|
|
|
|
|
|
/* Mark the vCPU as runnable */
|
|
|
|
|
kvm_set_mp_state(vcpu, KVM_MP_STATE_RUNNABLE);
|
|
|
|
|
|
|
|
|
|
svm->sev_es.snp_vmsa_gpa = INVALID_PAGE;
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* gmem pages aren't currently migratable, but if this ever
|
|
|
|
|
* changes then care should be taken to ensure
|
|
|
|
|
* svm->sev_es.vmsa is pinned through some other means.
|
|
|
|
|
*/
|
|
|
|
|
kvm_release_page_clean(page);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* When replacing the VMSA during SEV-SNP AP creation,
|
|
|
|
|
* mark the VMCB dirty so that full state is always reloaded.
|
|
|
|
|
*/
|
|
|
|
|
vmcb_mark_all_dirty(svm->vmcb);
|
|
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Invoked as part of svm_vcpu_reset() processing of an init event.
|
|
|
|
|
*/
|
|
|
|
|
void sev_snp_init_protected_guest_state(struct kvm_vcpu *vcpu)
|
|
|
|
|
{
|
|
|
|
|
struct vcpu_svm *svm = to_svm(vcpu);
|
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
|
|
if (!sev_snp_guest(vcpu->kvm))
|
|
|
|
|
if (!VALID_PAGE(svm->sev_es.snp_vmsa_gpa))
|
|
|
|
|
return;
|
|
|
|
|
|
|
|
|
|
mutex_lock(&svm->sev_es.snp_vmsa_mutex);
|
|
|
|
|
gfn = gpa_to_gfn(svm->sev_es.snp_vmsa_gpa);
|
|
|
|
|
svm->sev_es.snp_vmsa_gpa = INVALID_PAGE;
|
|
|
|
|
|
|
|
|
|
if (!svm->sev_es.snp_ap_waiting_for_reset)
|
|
|
|
|
goto unlock;
|
|
|
|
|
slot = gfn_to_memslot(vcpu->kvm, gfn);
|
|
|
|
|
if (!slot)
|
|
|
|
|
return;
|
|
|
|
|
|
|
|
|
|
svm->sev_es.snp_ap_waiting_for_reset = false;
|
|
|
|
|
/*
|
|
|
|
|
* The new VMSA will be private memory guest memory, so retrieve the
|
|
|
|
|
* PFN from the gmem backend.
|
|
|
|
|
*/
|
|
|
|
|
if (kvm_gmem_get_pfn(vcpu->kvm, slot, gfn, &pfn, &page, NULL))
|
|
|
|
|
return;
|
|
|
|
|
|
|
|
|
|
ret = __sev_snp_update_protected_guest_state(vcpu);
|
|
|
|
|
if (ret)
|
|
|
|
|
vcpu_unimpl(vcpu, "snp: AP state update on init failed\n");
|
|
|
|
|
/*
|
|
|
|
|
* From this point forward, the VMSA will always be a guest-mapped page
|
|
|
|
|
* rather than the initial one allocated by KVM in svm->sev_es.vmsa. In
|
|
|
|
|
* theory, svm->sev_es.vmsa could be free'd and cleaned up here, but
|
|
|
|
|
* that involves cleanups like wbinvd_on_all_cpus() which would ideally
|
|
|
|
|
* be handled during teardown rather than guest boot. Deferring that
|
|
|
|
|
* also allows the existing logic for SEV-ES VMSAs to be re-used with
|
|
|
|
|
* minimal SNP-specific changes.
|
|
|
|
|
*/
|
|
|
|
|
svm->sev_es.snp_has_guest_vmsa = true;
|
|
|
|
|
|
|
|
|
|
unlock:
|
|
|
|
|
mutex_unlock(&svm->sev_es.snp_vmsa_mutex);
|
|
|
|
|
/* Use the new VMSA */
|
|
|
|
|
svm->vmcb->control.vmsa_pa = pfn_to_hpa(pfn);
|
|
|
|
|
|
|
|
|
|
/* Mark the vCPU as runnable */
|
|
|
|
|
kvm_set_mp_state(vcpu, KVM_MP_STATE_RUNNABLE);
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* gmem pages aren't currently migratable, but if this ever changes
|
|
|
|
|
* then care should be taken to ensure svm->sev_es.vmsa is pinned
|
|
|
|
|
* through some other means.
|
|
|
|
|
*/
|
|
|
|
|
kvm_release_page_clean(page);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static int sev_snp_ap_creation(struct vcpu_svm *svm)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(svm->vcpu.kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(svm->vcpu.kvm);
|
|
|
|
|
struct kvm_vcpu *vcpu = &svm->vcpu;
|
|
|
|
|
struct kvm_vcpu *target_vcpu;
|
|
|
|
|
struct vcpu_svm *target_svm;
|
|
|
|
|
unsigned int request;
|
|
|
|
|
unsigned int apic_id;
|
|
|
|
|
bool kick;
|
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
|
|
request = lower_32_bits(svm->vmcb->control.exit_info_1);
|
|
|
|
|
apic_id = upper_32_bits(svm->vmcb->control.exit_info_1);
|
|
|
|
|
@@ -3962,47 +3938,23 @@ static int sev_snp_ap_creation(struct vcpu_svm *svm)
|
|
|
|
|
return -EINVAL;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
ret = 0;
|
|
|
|
|
|
|
|
|
|
target_svm = to_svm(target_vcpu);
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* The target vCPU is valid, so the vCPU will be kicked unless the
|
|
|
|
|
* request is for CREATE_ON_INIT. For any errors at this stage, the
|
|
|
|
|
* kick will place the vCPU in an non-runnable state.
|
|
|
|
|
*/
|
|
|
|
|
kick = true;
|
|
|
|
|
|
|
|
|
|
mutex_lock(&target_svm->sev_es.snp_vmsa_mutex);
|
|
|
|
|
|
|
|
|
|
target_svm->sev_es.snp_vmsa_gpa = INVALID_PAGE;
|
|
|
|
|
target_svm->sev_es.snp_ap_waiting_for_reset = true;
|
|
|
|
|
|
|
|
|
|
/* Interrupt injection mode shouldn't change for AP creation */
|
|
|
|
|
if (request < SVM_VMGEXIT_AP_DESTROY) {
|
|
|
|
|
u64 sev_features;
|
|
|
|
|
|
|
|
|
|
sev_features = vcpu->arch.regs[VCPU_REGS_RAX];
|
|
|
|
|
sev_features ^= sev->vmsa_features;
|
|
|
|
|
|
|
|
|
|
if (sev_features & SVM_SEV_FEAT_INT_INJ_MODES) {
|
|
|
|
|
vcpu_unimpl(vcpu, "vmgexit: invalid AP injection mode [%#lx] from guest\n",
|
|
|
|
|
vcpu->arch.regs[VCPU_REGS_RAX]);
|
|
|
|
|
ret = -EINVAL;
|
|
|
|
|
goto out;
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
guard(mutex)(&target_svm->sev_es.snp_vmsa_mutex);
|
|
|
|
|
|
|
|
|
|
switch (request) {
|
|
|
|
|
case SVM_VMGEXIT_AP_CREATE_ON_INIT:
|
|
|
|
|
kick = false;
|
|
|
|
|
fallthrough;
|
|
|
|
|
case SVM_VMGEXIT_AP_CREATE:
|
|
|
|
|
if (vcpu->arch.regs[VCPU_REGS_RAX] != sev->vmsa_features) {
|
|
|
|
|
vcpu_unimpl(vcpu, "vmgexit: mismatched AP sev_features [%#lx] != [%#llx] from guest\n",
|
|
|
|
|
vcpu->arch.regs[VCPU_REGS_RAX], sev->vmsa_features);
|
|
|
|
|
return -EINVAL;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if (!page_address_valid(vcpu, svm->vmcb->control.exit_info_2)) {
|
|
|
|
|
vcpu_unimpl(vcpu, "vmgexit: invalid AP VMSA address [%#llx] from guest\n",
|
|
|
|
|
svm->vmcb->control.exit_info_2);
|
|
|
|
|
ret = -EINVAL;
|
|
|
|
|
goto out;
|
|
|
|
|
return -EINVAL;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
@@ -4016,30 +3968,32 @@ static int sev_snp_ap_creation(struct vcpu_svm *svm)
|
|
|
|
|
vcpu_unimpl(vcpu,
|
|
|
|
|
"vmgexit: AP VMSA address [%llx] from guest is unsafe as it is 2M aligned\n",
|
|
|
|
|
svm->vmcb->control.exit_info_2);
|
|
|
|
|
ret = -EINVAL;
|
|
|
|
|
goto out;
|
|
|
|
|
return -EINVAL;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
target_svm->sev_es.snp_vmsa_gpa = svm->vmcb->control.exit_info_2;
|
|
|
|
|
break;
|
|
|
|
|
case SVM_VMGEXIT_AP_DESTROY:
|
|
|
|
|
target_svm->sev_es.snp_vmsa_gpa = INVALID_PAGE;
|
|
|
|
|
break;
|
|
|
|
|
default:
|
|
|
|
|
vcpu_unimpl(vcpu, "vmgexit: invalid AP creation request [%#x] from guest\n",
|
|
|
|
|
request);
|
|
|
|
|
ret = -EINVAL;
|
|
|
|
|
break;
|
|
|
|
|
return -EINVAL;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
out:
|
|
|
|
|
if (kick) {
|
|
|
|
|
target_svm->sev_es.snp_ap_waiting_for_reset = true;
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Unless Creation is deferred until INIT, signal the vCPU to update
|
|
|
|
|
* its state.
|
|
|
|
|
*/
|
|
|
|
|
if (request != SVM_VMGEXIT_AP_CREATE_ON_INIT) {
|
|
|
|
|
kvm_make_request(KVM_REQ_UPDATE_PROTECTED_GUEST_STATE, target_vcpu);
|
|
|
|
|
kvm_vcpu_kick(target_vcpu);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
mutex_unlock(&target_svm->sev_es.snp_vmsa_mutex);
|
|
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
|
return 0;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
static int snp_handle_guest_req(struct vcpu_svm *svm, gpa_t req_gpa, gpa_t resp_gpa)
|
|
|
|
|
@@ -4078,7 +4032,8 @@ static int snp_handle_guest_req(struct vcpu_svm *svm, gpa_t req_gpa, gpa_t resp_
|
|
|
|
|
goto out_unlock;
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
ghcb_set_sw_exit_info_2(svm->sev_es.ghcb, SNP_GUEST_ERR(0, fw_err));
|
|
|
|
|
/* No action is requested *from KVM* if there was a firmware error. */
|
|
|
|
|
svm_vmgexit_no_action(svm, SNP_GUEST_ERR(0, fw_err));
|
|
|
|
|
|
|
|
|
|
ret = 1; /* resume guest */
|
|
|
|
|
|
|
|
|
|
@@ -4134,8 +4089,7 @@ static int snp_handle_ext_guest_req(struct vcpu_svm *svm, gpa_t req_gpa, gpa_t r
|
|
|
|
|
return snp_handle_guest_req(svm, req_gpa, resp_gpa);
|
|
|
|
|
|
|
|
|
|
request_invalid:
|
|
|
|
|
ghcb_set_sw_exit_info_1(svm->sev_es.ghcb, 2);
|
|
|
|
|
ghcb_set_sw_exit_info_2(svm->sev_es.ghcb, GHCB_ERR_INVALID_INPUT);
|
|
|
|
|
svm_vmgexit_bad_input(svm, GHCB_ERR_INVALID_INPUT);
|
|
|
|
|
return 1; /* resume guest */
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
@@ -4143,7 +4097,7 @@ static int sev_handle_vmgexit_msr_protocol(struct vcpu_svm *svm)
|
|
|
|
|
{
|
|
|
|
|
struct vmcb_control_area *control = &svm->vmcb->control;
|
|
|
|
|
struct kvm_vcpu *vcpu = &svm->vcpu;
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(vcpu->kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(vcpu->kvm);
|
|
|
|
|
u64 ghcb_info;
|
|
|
|
|
int ret = 1;
|
|
|
|
|
|
|
|
|
|
@@ -4327,8 +4281,7 @@ int sev_handle_vmgexit(struct kvm_vcpu *vcpu)
|
|
|
|
|
if (ret)
|
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
|
|
ghcb_set_sw_exit_info_1(svm->sev_es.ghcb, 0);
|
|
|
|
|
ghcb_set_sw_exit_info_2(svm->sev_es.ghcb, 0);
|
|
|
|
|
svm_vmgexit_success(svm, 0);
|
|
|
|
|
|
|
|
|
|
exit_code = kvm_ghcb_get_sw_exit_code(control);
|
|
|
|
|
switch (exit_code) {
|
|
|
|
|
@@ -4363,7 +4316,7 @@ int sev_handle_vmgexit(struct kvm_vcpu *vcpu)
|
|
|
|
|
ret = kvm_emulate_ap_reset_hold(vcpu);
|
|
|
|
|
break;
|
|
|
|
|
case SVM_VMGEXIT_AP_JUMP_TABLE: {
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(vcpu->kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(vcpu->kvm);
|
|
|
|
|
|
|
|
|
|
switch (control->exit_info_1) {
|
|
|
|
|
case 0:
|
|
|
|
|
@@ -4372,21 +4325,19 @@ int sev_handle_vmgexit(struct kvm_vcpu *vcpu)
|
|
|
|
|
break;
|
|
|
|
|
case 1:
|
|
|
|
|
/* Get AP jump table address */
|
|
|
|
|
ghcb_set_sw_exit_info_2(svm->sev_es.ghcb, sev->ap_jump_table);
|
|
|
|
|
svm_vmgexit_success(svm, sev->ap_jump_table);
|
|
|
|
|
break;
|
|
|
|
|
default:
|
|
|
|
|
pr_err("svm: vmgexit: unsupported AP jump table request - exit_info_1=%#llx\n",
|
|
|
|
|
control->exit_info_1);
|
|
|
|
|
ghcb_set_sw_exit_info_1(svm->sev_es.ghcb, 2);
|
|
|
|
|
ghcb_set_sw_exit_info_2(svm->sev_es.ghcb, GHCB_ERR_INVALID_INPUT);
|
|
|
|
|
svm_vmgexit_bad_input(svm, GHCB_ERR_INVALID_INPUT);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
ret = 1;
|
|
|
|
|
break;
|
|
|
|
|
}
|
|
|
|
|
case SVM_VMGEXIT_HV_FEATURES:
|
|
|
|
|
ghcb_set_sw_exit_info_2(svm->sev_es.ghcb, GHCB_HV_FT_SUPPORTED);
|
|
|
|
|
|
|
|
|
|
svm_vmgexit_success(svm, GHCB_HV_FT_SUPPORTED);
|
|
|
|
|
ret = 1;
|
|
|
|
|
break;
|
|
|
|
|
case SVM_VMGEXIT_TERM_REQUEST:
|
|
|
|
|
@@ -4407,8 +4358,7 @@ int sev_handle_vmgexit(struct kvm_vcpu *vcpu)
|
|
|
|
|
case SVM_VMGEXIT_AP_CREATION:
|
|
|
|
|
ret = sev_snp_ap_creation(svm);
|
|
|
|
|
if (ret) {
|
|
|
|
|
ghcb_set_sw_exit_info_1(svm->sev_es.ghcb, 2);
|
|
|
|
|
ghcb_set_sw_exit_info_2(svm->sev_es.ghcb, GHCB_ERR_INVALID_INPUT);
|
|
|
|
|
svm_vmgexit_bad_input(svm, GHCB_ERR_INVALID_INPUT);
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
ret = 1;
|
|
|
|
|
@@ -4574,7 +4524,7 @@ void sev_init_vmcb(struct vcpu_svm *svm)
|
|
|
|
|
void sev_es_vcpu_reset(struct vcpu_svm *svm)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_vcpu *vcpu = &svm->vcpu;
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(vcpu->kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(vcpu->kvm);
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
|
* Set the GHCB MSR value as per the GHCB specification when emulating
|
|
|
|
|
@@ -4654,7 +4604,7 @@ void sev_vcpu_deliver_sipi_vector(struct kvm_vcpu *vcpu, u8 vector)
|
|
|
|
|
* Return from an AP Reset Hold VMGEXIT, where the guest will
|
|
|
|
|
* set the CS and RIP. Set SW_EXIT_INFO_2 to a non-zero value.
|
|
|
|
|
*/
|
|
|
|
|
ghcb_set_sw_exit_info_2(svm->sev_es.ghcb, 1);
|
|
|
|
|
svm_vmgexit_success(svm, 1);
|
|
|
|
|
break;
|
|
|
|
|
case AP_RESET_HOLD_MSR_PROTO:
|
|
|
|
|
/*
|
|
|
|
|
@@ -4852,7 +4802,7 @@ static bool is_large_rmp_possible(struct kvm *kvm, kvm_pfn_t pfn, int order)
|
|
|
|
|
|
|
|
|
|
int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order)
|
|
|
|
|
{
|
|
|
|
|
struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
|
|
|
|
|
struct kvm_sev_info *sev = to_kvm_sev_info(kvm);
|
|
|
|
|
kvm_pfn_t pfn_aligned;
|
|
|
|
|
gfn_t gfn_aligned;
|
|
|
|
|
int level, rc;
|
|
|
|
|
|