Re: [PATCH v12 24/31] LoongArch: KVM: Implement handle mmio exception

From: bibo, mao
Date: Mon Jun 05 2023 - 21:41:14 EST


Reviewed-by: Bibo, Mao <maobibo@xxxxxxxxxxx>

在 2023/5/30 09:52, Tianrui Zhao 写道:
> Implement handle mmio exception, setting the mmio info into vcpu_run and
> return to user space to handle it.
>
> Signed-off-by: Tianrui Zhao <zhaotianrui@xxxxxxxxxxx>
> ---
> arch/loongarch/kvm/exit.c | 308 ++++++++++++++++++++++++++++++++++++++
> 1 file changed, 308 insertions(+)
>
> diff --git a/arch/loongarch/kvm/exit.c b/arch/loongarch/kvm/exit.c
> index 26283a9d3c6d..1237ceb06d0c 100644
> --- a/arch/loongarch/kvm/exit.c
> +++ b/arch/loongarch/kvm/exit.c
> @@ -209,6 +209,265 @@ int _kvm_emu_idle(struct kvm_vcpu *vcpu)
> return EMULATE_DONE;
> }
>
> +int _kvm_emu_mmio_write(struct kvm_vcpu *vcpu, larch_inst inst)
> +{
> + struct kvm_run *run = vcpu->run;
> + unsigned int rd, op8, opcode;
> + unsigned long rd_val = 0;
> + void *data = run->mmio.data;
> + unsigned long curr_pc;
> + int ret;
> +
> + /*
> + * Update PC and hold onto current PC in case there is
> + * an error and we want to rollback the PC
> + */
> + curr_pc = vcpu->arch.pc;
> + update_pc(&vcpu->arch);
> +
> + op8 = (inst.word >> 24) & 0xff;
> + run->mmio.phys_addr = vcpu->arch.badv;
> + ret = EMULATE_DO_MMIO;
> + if (op8 < 0x28) {
> + /* stptrw/d process */
> + rd = inst.reg2i14_format.rd;
> + opcode = inst.reg2i14_format.opcode;
> +
> + switch (opcode) {
> + case stptrd_op:
> + run->mmio.len = 8;
> + *(unsigned long *)data = vcpu->arch.gprs[rd];
> + break;
> + case stptrw_op:
> + run->mmio.len = 4;
> + *(unsigned int *)data = vcpu->arch.gprs[rd];
> + break;
> + default:
> + ret = EMULATE_FAIL;
> + break;
> + }
> + } else if (op8 < 0x30) {
> + /* st.b/h/w/d process */
> + rd = inst.reg2i12_format.rd;
> + opcode = inst.reg2i12_format.opcode;
> + rd_val = vcpu->arch.gprs[rd];
> +
> + switch (opcode) {
> + case std_op:
> + run->mmio.len = 8;
> + *(unsigned long *)data = rd_val;
> + break;
> + case stw_op:
> + run->mmio.len = 4;
> + *(unsigned int *)data = rd_val;
> + break;
> + case sth_op:
> + run->mmio.len = 2;
> + *(unsigned short *)data = rd_val;
> + break;
> + case stb_op:
> + run->mmio.len = 1;
> + *(unsigned char *)data = rd_val;
> + break;
> + default:
> + ret = EMULATE_FAIL;
> + break;
> + }
> + } else if (op8 == 0x38) {
> + /* stxb/h/w/d process */
> + rd = inst.reg3_format.rd;
> + opcode = inst.reg3_format.opcode;
> +
> + switch (opcode) {
> + case stxb_op:
> + run->mmio.len = 1;
> + *(unsigned char *)data = vcpu->arch.gprs[rd];
> + break;
> + case stxh_op:
> + run->mmio.len = 2;
> + *(unsigned short *)data = vcpu->arch.gprs[rd];
> + break;
> + case stxw_op:
> + run->mmio.len = 4;
> + *(unsigned int *)data = vcpu->arch.gprs[rd];
> + break;
> + case stxd_op:
> + run->mmio.len = 8;
> + *(unsigned long *)data = vcpu->arch.gprs[rd];
> + break;
> + default:
> + ret = EMULATE_FAIL;
> + break;
> + }
> + } else
> + ret = EMULATE_FAIL;
> +
> + if (ret == EMULATE_DO_MMIO) {
> + run->mmio.is_write = 1;
> + vcpu->mmio_needed = 1;
> + vcpu->mmio_is_write = 1;
> + } else {
> + vcpu->arch.pc = curr_pc;
> + kvm_err("Write not supporded inst=0x%08x @%lx BadVaddr:%#lx\n",
> + inst.word, vcpu->arch.pc, vcpu->arch.badv);
> + kvm_arch_vcpu_dump_regs(vcpu);
> + /* Rollback PC if emulation was unsuccessful */
> + }
> +
> + return ret;
> +}
> +
> +int _kvm_emu_mmio_read(struct kvm_vcpu *vcpu, larch_inst inst)
> +{
> + unsigned int op8, opcode, rd;
> + struct kvm_run *run = vcpu->run;
> + int ret;
> +
> + run->mmio.phys_addr = vcpu->arch.badv;
> + vcpu->mmio_needed = 2; /* signed */
> + op8 = (inst.word >> 24) & 0xff;
> + ret = EMULATE_DO_MMIO;
> +
> + if (op8 < 0x28) {
> + /* ldptr.w/d process */
> + rd = inst.reg2i14_format.rd;
> + opcode = inst.reg2i14_format.opcode;
> +
> + switch (opcode) {
> + case ldptrd_op:
> + run->mmio.len = 8;
> + break;
> + case ldptrw_op:
> + run->mmio.len = 4;
> + break;
> + default:
> + break;
> + }
> + } else if (op8 < 0x2f) {
> + /* ld.b/h/w/d, ld.bu/hu/wu process */
> + rd = inst.reg2i12_format.rd;
> + opcode = inst.reg2i12_format.opcode;
> +
> + switch (opcode) {
> + case ldd_op:
> + run->mmio.len = 8;
> + break;
> + case ldwu_op:
> + vcpu->mmio_needed = 1; /* unsigned */
> + run->mmio.len = 4;
> + break;
> + case ldw_op:
> + run->mmio.len = 4;
> + break;
> + case ldhu_op:
> + vcpu->mmio_needed = 1; /* unsigned */
> + run->mmio.len = 2;
> + break;
> + case ldh_op:
> + run->mmio.len = 2;
> + break;
> + case ldbu_op:
> + vcpu->mmio_needed = 1; /* unsigned */
> + run->mmio.len = 1;
> + break;
> + case ldb_op:
> + run->mmio.len = 1;
> + break;
> + default:
> + ret = EMULATE_FAIL;
> + break;
> + }
> + } else if (op8 == 0x38) {
> + /* ldxb/h/w/d, ldxb/h/wu, ldgtb/h/w/d, ldleb/h/w/d process */
> + rd = inst.reg3_format.rd;
> + opcode = inst.reg3_format.opcode;
> +
> + switch (opcode) {
> + case ldxb_op:
> + run->mmio.len = 1;
> + break;
> + case ldxbu_op:
> + run->mmio.len = 1;
> + vcpu->mmio_needed = 1; /* unsigned */
> + break;
> + case ldxh_op:
> + run->mmio.len = 2;
> + break;
> + case ldxhu_op:
> + run->mmio.len = 2;
> + vcpu->mmio_needed = 1; /* unsigned */
> + break;
> + case ldxw_op:
> + run->mmio.len = 4;
> + break;
> + case ldxwu_op:
> + run->mmio.len = 4;
> + vcpu->mmio_needed = 1; /* unsigned */
> + break;
> + case ldxd_op:
> + run->mmio.len = 8;
> + break;
> + default:
> + ret = EMULATE_FAIL;
> + break;
> + }
> + } else
> + ret = EMULATE_FAIL;
> +
> + if (ret == EMULATE_DO_MMIO) {
> + /* Set for _kvm_complete_mmio_read use */
> + vcpu->arch.io_gpr = rd;
> + run->mmio.is_write = 0;
> + vcpu->mmio_is_write = 0;
> + } else {
> + kvm_err("Load not supporded inst=0x%08x @%lx BadVaddr:%#lx\n",
> + inst.word, vcpu->arch.pc, vcpu->arch.badv);
> + kvm_arch_vcpu_dump_regs(vcpu);
> + vcpu->mmio_needed = 0;
> + }
> + return ret;
> +}
> +
> +int _kvm_complete_mmio_read(struct kvm_vcpu *vcpu, struct kvm_run *run)
> +{
> + unsigned long *gpr = &vcpu->arch.gprs[vcpu->arch.io_gpr];
> + enum emulation_result er = EMULATE_DONE;
> +
> + /* update with new PC */
> + update_pc(&vcpu->arch);
> + switch (run->mmio.len) {
> + case 8:
> + *gpr = *(s64 *)run->mmio.data;
> + break;
> + case 4:
> + if (vcpu->mmio_needed == 2)
> + *gpr = *(int *)run->mmio.data;
> + else
> + *gpr = *(unsigned int *)run->mmio.data;
> + break;
> + case 2:
> + if (vcpu->mmio_needed == 2)
> + *gpr = *(short *) run->mmio.data;
> + else
> + *gpr = *(unsigned short *)run->mmio.data;
> +
> + break;
> + case 1:
> + if (vcpu->mmio_needed == 2)
> + *gpr = *(char *) run->mmio.data;
> + else
> + *gpr = *(unsigned char *) run->mmio.data;
> + break;
> + default:
> + kvm_err("Bad MMIO length: %d,addr is 0x%lx",
> + run->mmio.len, vcpu->arch.badv);
> + er = EMULATE_FAIL;
> + break;
> + }
> +
> + return er;
> +}
> +
> static int _kvm_trap_handle_gspr(struct kvm_vcpu *vcpu)
> {
> enum emulation_result er = EMULATE_DONE;
> @@ -323,3 +582,52 @@ static int _kvm_handle_gspr(struct kvm_vcpu *vcpu)
> }
> return ret;
> }
> +
> +static int _kvm_handle_mmu_fault(struct kvm_vcpu *vcpu, bool write)
> +{
> + struct kvm_run *run = vcpu->run;
> + unsigned long badv = vcpu->arch.badv;
> + larch_inst inst;
> + enum emulation_result er = EMULATE_DONE;
> + int ret;
> +
> + ret = kvm_handle_mm_fault(vcpu, badv, write);
> + if (ret) {
> + /* Treat as MMIO */
> + inst.word = vcpu->arch.badi;
> + if (write) {
> + er = _kvm_emu_mmio_write(vcpu, inst);
> + } else {
> + /* A code fetch fault doesn't count as an MMIO */
> + if (kvm_is_ifetch_fault(&vcpu->arch)) {
> + kvm_err("%s ifetch error addr:%lx\n", __func__, badv);
> + run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
> + return RESUME_HOST;
> + }
> +
> + er = _kvm_emu_mmio_read(vcpu, inst);
> + }
> + }
> +
> + if (er == EMULATE_DONE) {
> + ret = RESUME_GUEST;
> + } else if (er == EMULATE_DO_MMIO) {
> + run->exit_reason = KVM_EXIT_MMIO;
> + ret = RESUME_HOST;
> + } else {
> + run->exit_reason = KVM_EXIT_INTERNAL_ERROR;
> + ret = RESUME_HOST;
> + }
> +
> + return ret;
> +}
> +
> +static int _kvm_handle_write_fault(struct kvm_vcpu *vcpu)
> +{
> + return _kvm_handle_mmu_fault(vcpu, true);
> +}
> +
> +static int _kvm_handle_read_fault(struct kvm_vcpu *vcpu)
> +{
> + return _kvm_handle_mmu_fault(vcpu, false);
> +}