[RFC 30/33] KVM: x86: hyper-v: Introduce KVM_REQ_HV_INJECT_INTERCEPT request

From: Nicolas Saenz Julienne
Date: Wed Nov 08 2023 - 06:25:12 EST


Introduce a new request type, KVM_REQ_HV_INJECT_INTERCEPT which allows
injecting out-of-band Hyper-V secure intercepts. For now only memory
access intercepts are supported. These are triggered when access a GPA
protected by a higher VTL. The memory intercept metadata is filled based
on the GPA provided through struct kvm_vcpu_hv_intercept_info, and
injected into the guest through SynIC message.

Signed-off-by: Nicolas Saenz Julienne <nsaenz@xxxxxxxxxx>
---
arch/x86/include/asm/kvm_host.h | 10 +++
arch/x86/kvm/hyperv.c | 114 ++++++++++++++++++++++++++++++++
arch/x86/kvm/hyperv.h | 2 +
arch/x86/kvm/x86.c | 3 +
4 files changed, 129 insertions(+)

diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h
index 1a854776d91e..39671e075555 100644
--- a/arch/x86/include/asm/kvm_host.h
+++ b/arch/x86/include/asm/kvm_host.h
@@ -113,6 +113,7 @@
KVM_ARCH_REQ_FLAGS(31, KVM_REQUEST_WAIT | KVM_REQUEST_NO_WAKEUP)
#define KVM_REQ_HV_TLB_FLUSH \
KVM_ARCH_REQ_FLAGS(32, KVM_REQUEST_WAIT | KVM_REQUEST_NO_WAKEUP)
+#define KVM_REQ_HV_INJECT_INTERCEPT KVM_ARCH_REQ(33)

#define CR0_RESERVED_BITS \
(~(unsigned long)(X86_CR0_PE | X86_CR0_MP | X86_CR0_EM | X86_CR0_TS \
@@ -639,6 +640,13 @@ struct kvm_vcpu_hv_tlb_flush_fifo {
DECLARE_KFIFO(entries, u64, KVM_HV_TLB_FLUSH_FIFO_SIZE);
};

+struct kvm_vcpu_hv_intercept_info {
+ struct kvm_vcpu *vcpu;
+ int type;
+ u64 gpa;
+ u8 access;
+};
+
/* Hyper-V per vcpu emulation context */
struct kvm_vcpu_hv {
struct kvm_vcpu *vcpu;
@@ -673,6 +681,8 @@ struct kvm_vcpu_hv {
u64 vm_id;
u32 vp_id;
} nested;
+
+ struct kvm_vcpu_hv_intercept_info intercept_info;
};

struct kvm_hypervisor_cpuid {
diff --git a/arch/x86/kvm/hyperv.c b/arch/x86/kvm/hyperv.c
index eb6a4848e306..38ee3abdef9c 100644
--- a/arch/x86/kvm/hyperv.c
+++ b/arch/x86/kvm/hyperv.c
@@ -2789,6 +2789,120 @@ int kvm_hv_hypercall(struct kvm_vcpu *vcpu)
return 0;
}

+static void store_kvm_segment(const struct kvm_segment *kvmseg,
+ struct hv_x64_segment_register *reg)
+{
+ reg->base = kvmseg->base;
+ reg->limit = kvmseg->limit;
+ reg->selector = kvmseg->selector;
+ reg->segment_type = kvmseg->type;
+ reg->present = kvmseg->present;
+ reg->descriptor_privilege_level = kvmseg->dpl;
+ reg->_default = kvmseg->db;
+ reg->non_system_segment = kvmseg->s;
+ reg->_long = kvmseg->l;
+ reg->granularity = kvmseg->g;
+ reg->available = kvmseg->avl;
+}
+
+static void deliver_gpa_intercept(struct kvm_vcpu *target_vcpu,
+ struct kvm_vcpu *intercepted_vcpu, u64 gpa,
+ u64 gva, u8 access_type_mask)
+{
+ ulong cr0;
+ struct hv_message msg = { 0 };
+ struct hv_memory_intercept_message *intercept = (struct hv_memory_intercept_message *)msg.u.payload;
+ struct kvm_vcpu_hv *hv_vcpu = to_hv_vcpu(target_vcpu);
+ struct x86_exception e;
+ struct kvm_segment kvmseg;
+
+ msg.header.message_type = HVMSG_GPA_INTERCEPT;
+ msg.header.payload_size = sizeof(*intercept);
+
+ intercept->header.vp_index = to_hv_vcpu(intercepted_vcpu)->vp_index;
+ intercept->header.instruction_length = intercepted_vcpu->arch.exit_instruction_len;
+ intercept->header.access_type_mask = access_type_mask;
+ kvm_x86_ops.get_segment(intercepted_vcpu, &kvmseg, VCPU_SREG_CS);
+ store_kvm_segment(&kvmseg, &intercept->header.cs);
+
+ cr0 = kvm_read_cr0(intercepted_vcpu);
+ intercept->header.exec_state.cr0_pe = (cr0 & X86_CR0_PE);
+ intercept->header.exec_state.cr0_am = (cr0 & X86_CR0_AM);
+ intercept->header.exec_state.cpl = kvm_x86_ops.get_cpl(intercepted_vcpu);
+ intercept->header.exec_state.efer_lma = is_long_mode(intercepted_vcpu);
+ intercept->header.exec_state.debug_active = 0;
+ intercept->header.exec_state.interruption_pending = 0;
+ intercept->header.rip = kvm_rip_read(intercepted_vcpu);
+ intercept->header.rflags = kvm_get_rflags(intercepted_vcpu);
+
+ /*
+ * For exec violations we don't have a way to decode an instruction that issued a fetch
+ * to a non-X page because CPU points RIP and GPA to the fetch destination in the faulted page.
+ * Instruction length though is the length of the fetch source.
+ * Seems like Hyper-V is aware of that and is not trying to access those fields.
+ */
+ if (access_type_mask == HV_INTERCEPT_ACCESS_EXECUTE) {
+ intercept->instruction_byte_count = 0;
+ } else {
+ intercept->instruction_byte_count = intercepted_vcpu->arch.exit_instruction_len;
+ if (intercept->instruction_byte_count > sizeof(intercept->instruction_bytes))
+ intercept->instruction_byte_count = sizeof(intercept->instruction_bytes);
+ if (kvm_read_guest_virt(intercepted_vcpu,
+ kvm_rip_read(intercepted_vcpu),
+ intercept->instruction_bytes,
+ intercept->instruction_byte_count, &e))
+ goto inject_ud;
+ }
+
+ intercept->memory_access_info.gva_valid = (gva != 0);
+ intercept->gva = gva;
+ intercept->gpa = gpa;
+ intercept->cache_type = HV_X64_CACHE_TYPE_WRITEBACK;
+ kvm_x86_ops.get_segment(intercepted_vcpu, &kvmseg, VCPU_SREG_DS);
+ store_kvm_segment(&kvmseg, &intercept->ds);
+ kvm_x86_ops.get_segment(intercepted_vcpu, &kvmseg, VCPU_SREG_SS);
+ store_kvm_segment(&kvmseg, &intercept->ss);
+ intercept->rax = kvm_rax_read(intercepted_vcpu);
+ intercept->rcx = kvm_rcx_read(intercepted_vcpu);
+ intercept->rdx = kvm_rdx_read(intercepted_vcpu);
+ intercept->rbx = kvm_rbx_read(intercepted_vcpu);
+ intercept->rsp = kvm_rsp_read(intercepted_vcpu);
+ intercept->rbp = kvm_rbp_read(intercepted_vcpu);
+ intercept->rsi = kvm_rsi_read(intercepted_vcpu);
+ intercept->rdi = kvm_rdi_read(intercepted_vcpu);
+ intercept->r8 = kvm_r8_read(intercepted_vcpu);
+ intercept->r9 = kvm_r9_read(intercepted_vcpu);
+ intercept->r10 = kvm_r10_read(intercepted_vcpu);
+ intercept->r11 = kvm_r11_read(intercepted_vcpu);
+ intercept->r12 = kvm_r12_read(intercepted_vcpu);
+ intercept->r13 = kvm_r13_read(intercepted_vcpu);
+ intercept->r14 = kvm_r14_read(intercepted_vcpu);
+ intercept->r15 = kvm_r15_read(intercepted_vcpu);
+
+ if (synic_deliver_msg(&hv_vcpu->synic, 0, &msg, true))
+ goto inject_ud;
+
+ return;
+
+inject_ud:
+ kvm_queue_exception(target_vcpu, UD_VECTOR);
+}
+
+void kvm_hv_deliver_intercept(struct kvm_vcpu *vcpu)
+{
+ struct kvm_vcpu_hv_intercept_info *info = &to_hv_vcpu(vcpu)->intercept_info;
+
+ switch (info->type) {
+ case HVMSG_GPA_INTERCEPT:
+ deliver_gpa_intercept(vcpu, info->vcpu, info->gpa, 0,
+ info->access);
+ break;
+ default:
+ pr_warn("Unknown exception\n");
+ }
+}
+EXPORT_SYMBOL_GPL(kvm_hv_deliver_intercept);
+
void kvm_hv_init_vm(struct kvm *kvm)
{
struct kvm_hv *hv = to_kvm_hv(kvm);
diff --git a/arch/x86/kvm/hyperv.h b/arch/x86/kvm/hyperv.h
index ae781b4d4669..8efc4916e0cb 100644
--- a/arch/x86/kvm/hyperv.h
+++ b/arch/x86/kvm/hyperv.h
@@ -273,4 +273,6 @@ int kvm_hv_vtl_dev_register(void);
void kvm_hv_vtl_dev_unregister(void);
int kvm_hv_faultin_pfn(struct kvm_vcpu *vcpu, struct kvm_page_fault *fault);

+void kvm_hv_deliver_intercept(struct kvm_vcpu *vcpu);
+
#endif
diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
index 82d3b86d9c93..f2581eec39a9 100644
--- a/arch/x86/kvm/x86.c
+++ b/arch/x86/kvm/x86.c
@@ -10707,6 +10707,9 @@ static int vcpu_enter_guest(struct kvm_vcpu *vcpu)

if (kvm_check_request(KVM_REQ_UPDATE_CPU_DIRTY_LOGGING, vcpu))
static_call(kvm_x86_update_cpu_dirty_logging)(vcpu);
+
+ if (kvm_check_request(KVM_REQ_HV_INJECT_INTERCEPT, vcpu))
+ kvm_hv_deliver_intercept(vcpu);
}

if (kvm_check_request(KVM_REQ_EVENT, vcpu) || req_int_win ||
--
2.40.1