This patch implements kvm_cpu__<xyz> Guest/VM VCPU arch functions. These functions mostly deal with: 1. VCPU allocation and initialization 2. VCPU reset 3. VCPU show/dump code 4. VCPU show/dump registers We also save RISC-V ISA, XLEN, and TIMEBASE frequency for each VCPU so that it can be later used for generating Guest/VM FDT. Signed-off-by: Atish Patra <atish.patra@xxxxxxx> Signed-off-by: Anup Patel <anup.patel@xxxxxxx> --- riscv/include/kvm/kvm-cpu-arch.h | 4 + riscv/kvm-cpu.c | 393 ++++++++++++++++++++++++++++++- 2 files changed, 390 insertions(+), 7 deletions(-) diff --git a/riscv/include/kvm/kvm-cpu-arch.h b/riscv/include/kvm/kvm-cpu-arch.h index ae6ae0a..78fcd01 100644 --- a/riscv/include/kvm/kvm-cpu-arch.h +++ b/riscv/include/kvm/kvm-cpu-arch.h @@ -12,6 +12,10 @@ struct kvm_cpu { unsigned long cpu_id; + unsigned long riscv_xlen; + unsigned long riscv_isa; + unsigned long riscv_timebase; + struct kvm *kvm; int vcpu_fd; struct kvm_run *kvm_run; diff --git a/riscv/kvm-cpu.c b/riscv/kvm-cpu.c index e4b8fa5..8adaddd 100644 --- a/riscv/kvm-cpu.c +++ b/riscv/kvm-cpu.c @@ -17,10 +17,88 @@ int kvm_cpu__get_debug_fd(void) return debug_fd; } +static __u64 __kvm_reg_id(__u64 type, __u64 idx, __u64 size) +{ + return KVM_REG_RISCV | type | idx | size; +} + +#if __riscv_xlen == 64 +#define KVM_REG_SIZE_ULONG KVM_REG_SIZE_U64 +#else +#define KVM_REG_SIZE_ULONG KVM_REG_SIZE_U32 +#endif + +#define RISCV_CONFIG_REG(name) __kvm_reg_id(KVM_REG_RISCV_CONFIG, \ + KVM_REG_RISCV_CONFIG_REG(name), \ + KVM_REG_SIZE_ULONG) + +#define RISCV_CORE_REG(name) __kvm_reg_id(KVM_REG_RISCV_CORE, \ + KVM_REG_RISCV_CORE_REG(name), \ + KVM_REG_SIZE_ULONG) + +#define RISCV_CSR_REG(name) __kvm_reg_id(KVM_REG_RISCV_CSR, \ + KVM_REG_RISCV_CSR_REG(name), \ + KVM_REG_SIZE_ULONG) + +#define RISCV_TIMER_REG(name) __kvm_reg_id(KVM_REG_RISCV_TIMER, \ + KVM_REG_RISCV_TIMER_REG(name), \ + KVM_REG_SIZE_U64) + struct kvm_cpu *kvm_cpu__arch_init(struct kvm *kvm, unsigned long cpu_id) { - /* TODO: */ - return NULL; + struct kvm_cpu *vcpu; + u64 timebase = 0; + unsigned long isa = 0; + int coalesced_offset, mmap_size; + struct kvm_one_reg reg; + + vcpu = calloc(1, sizeof(struct kvm_cpu)); + if (!vcpu) + return NULL; + + vcpu->vcpu_fd = ioctl(kvm->vm_fd, KVM_CREATE_VCPU, cpu_id); + if (vcpu->vcpu_fd < 0) + die_perror("KVM_CREATE_VCPU ioctl"); + + reg.id = RISCV_CONFIG_REG(isa); + reg.addr = (unsigned long)&isa; + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (config.isa)"); + + reg.id = RISCV_TIMER_REG(frequency); + reg.addr = (unsigned long)&timebase; + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (timer.frequency)"); + + mmap_size = ioctl(kvm->sys_fd, KVM_GET_VCPU_MMAP_SIZE, 0); + if (mmap_size < 0) + die_perror("KVM_GET_VCPU_MMAP_SIZE ioctl"); + + vcpu->kvm_run = mmap(NULL, mmap_size, PROT_RW, MAP_SHARED, + vcpu->vcpu_fd, 0); + if (vcpu->kvm_run == MAP_FAILED) + die("unable to mmap vcpu fd"); + + coalesced_offset = ioctl(kvm->sys_fd, KVM_CHECK_EXTENSION, + KVM_CAP_COALESCED_MMIO); + if (coalesced_offset) + vcpu->ring = (void *)vcpu->kvm_run + + (coalesced_offset * PAGE_SIZE); + + reg.id = RISCV_CONFIG_REG(isa); + reg.addr = (unsigned long)&isa; + if (ioctl(vcpu->vcpu_fd, KVM_SET_ONE_REG, ®) < 0) + die("KVM_SET_ONE_REG failed (config.isa)"); + + /* Populate the vcpu structure. */ + vcpu->kvm = kvm; + vcpu->cpu_id = cpu_id; + vcpu->riscv_isa = isa; + vcpu->riscv_xlen = __riscv_xlen; + vcpu->riscv_timebase = timebase; + vcpu->is_running = true; + + return vcpu; } void kvm_cpu__arch_nmi(struct kvm_cpu *cpu) @@ -29,7 +107,7 @@ void kvm_cpu__arch_nmi(struct kvm_cpu *cpu) void kvm_cpu__delete(struct kvm_cpu *vcpu) { - /* TODO: */ + free(vcpu); } bool kvm_cpu__handle_exit(struct kvm_cpu *vcpu) @@ -40,12 +118,43 @@ bool kvm_cpu__handle_exit(struct kvm_cpu *vcpu) void kvm_cpu__show_page_tables(struct kvm_cpu *vcpu) { - /* TODO: */ } void kvm_cpu__reset_vcpu(struct kvm_cpu *vcpu) { - /* TODO: */ + struct kvm *kvm = vcpu->kvm; + struct kvm_mp_state mp_state; + struct kvm_one_reg reg; + unsigned long data; + + if (ioctl(vcpu->vcpu_fd, KVM_GET_MP_STATE, &mp_state) < 0) + die_perror("KVM_GET_MP_STATE failed"); + + /* + * If MP state is stopped then it means Linux KVM RISC-V emulates + * SBI v0.2 (or higher) with HART power managment and give VCPU + * will power-up at boot-time by boot VCPU. For such VCPU, we + * don't update PC, A0 and A1 here. + */ + if (mp_state.mp_state == KVM_MP_STATE_STOPPED) + return; + + reg.addr = (unsigned long)&data; + + data = kvm->arch.kern_guest_start; + reg.id = RISCV_CORE_REG(regs.pc); + if (ioctl(vcpu->vcpu_fd, KVM_SET_ONE_REG, ®) < 0) + die_perror("KVM_SET_ONE_REG failed (pc)"); + + data = vcpu->cpu_id; + reg.id = RISCV_CORE_REG(regs.a0); + if (ioctl(vcpu->vcpu_fd, KVM_SET_ONE_REG, ®) < 0) + die_perror("KVM_SET_ONE_REG failed (a0)"); + + data = kvm->arch.dtb_guest_start; + reg.id = RISCV_CORE_REG(regs.a1); + if (ioctl(vcpu->vcpu_fd, KVM_SET_ONE_REG, ®) < 0) + die_perror("KVM_SET_ONE_REG failed (a1)"); } int kvm_cpu__get_endianness(struct kvm_cpu *vcpu) @@ -55,10 +164,280 @@ int kvm_cpu__get_endianness(struct kvm_cpu *vcpu) void kvm_cpu__show_code(struct kvm_cpu *vcpu) { - /* TODO: */ + struct kvm_one_reg reg; + unsigned long data; + int debug_fd = kvm_cpu__get_debug_fd(); + + reg.addr = (unsigned long)&data; + + dprintf(debug_fd, "\n*PC:\n"); + reg.id = RISCV_CORE_REG(regs.pc); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (show_code @ PC)"); + + kvm__dump_mem(vcpu->kvm, data, 32, debug_fd); + + dprintf(debug_fd, "\n*RA:\n"); + reg.id = RISCV_CORE_REG(regs.ra); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (show_code @ RA)"); + + kvm__dump_mem(vcpu->kvm, data, 32, debug_fd); +} + +static void kvm_cpu__show_csrs(struct kvm_cpu *vcpu) +{ + struct kvm_one_reg reg; + struct kvm_riscv_csr csr; + unsigned long data; + int debug_fd = kvm_cpu__get_debug_fd(); + + reg.addr = (unsigned long)&data; + dprintf(debug_fd, "\n Control Status Registers:\n"); + dprintf(debug_fd, " ------------------------\n"); + + reg.id = RISCV_CSR_REG(sstatus); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (sstatus)"); + csr.sstatus = data; + + reg.id = RISCV_CSR_REG(sie); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (sie)"); + csr.sie = data; + + reg.id = RISCV_CSR_REG(stvec); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (stvec)"); + csr.stvec = data; + + reg.id = RISCV_CSR_REG(sip); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (sip)"); + csr.sip = data; + + reg.id = RISCV_CSR_REG(satp); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (satp)"); + csr.satp = data; + + reg.id = RISCV_CSR_REG(stval); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (stval)"); + csr.stval = data; + + reg.id = RISCV_CSR_REG(scause); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (SCAUSE)"); + csr.scause = data; + + reg.id = RISCV_CSR_REG(sscratch); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (sscartch)"); + csr.sscratch = data; + dprintf(debug_fd, " SSTATUS: 0x%016lx\n", csr.sstatus); + dprintf(debug_fd, " SIE: 0x%016lx\n", csr.sie); + dprintf(debug_fd, " STVEC: 0x%016lx\n", csr.stvec); + dprintf(debug_fd, " SIP: 0x%016lx\n", csr.sip); + dprintf(debug_fd, " SATP: 0x%016lx\n", csr.satp); + dprintf(debug_fd, " STVAL: 0x%016lx\n", csr.stval); + dprintf(debug_fd, " SCAUSE: 0x%016lx\n", csr.scause); + dprintf(debug_fd, " SSCRATCH: 0x%016lx\n", csr.sscratch); } void kvm_cpu__show_registers(struct kvm_cpu *vcpu) { - /* TODO: */ + struct kvm_one_reg reg; + unsigned long data; + int debug_fd = kvm_cpu__get_debug_fd(); + struct kvm_riscv_core core; + + reg.addr = (unsigned long)&data; + + reg.id = RISCV_CORE_REG(mode); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (mode)"); + core.mode = data; + + reg.id = RISCV_CORE_REG(regs.pc); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (pc)"); + core.regs.pc = data; + + reg.id = RISCV_CORE_REG(regs.ra); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (ra)"); + core.regs.ra = data; + + reg.id = RISCV_CORE_REG(regs.sp); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (sp)"); + core.regs.sp = data; + + reg.id = RISCV_CORE_REG(regs.gp); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (gp)"); + core.regs.gp = data; + + reg.id = RISCV_CORE_REG(regs.tp); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (tp)"); + core.regs.tp = data; + + reg.id = RISCV_CORE_REG(regs.t0); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (t0)"); + core.regs.t0 = data; + + reg.id = RISCV_CORE_REG(regs.t1); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (t1)"); + core.regs.t1 = data; + + reg.id = RISCV_CORE_REG(regs.t2); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (t2)"); + core.regs.t2 = data; + + reg.id = RISCV_CORE_REG(regs.s0); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (s0)"); + core.regs.s0 = data; + + reg.id = RISCV_CORE_REG(regs.s1); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (s1)"); + core.regs.s1 = data; + + reg.id = RISCV_CORE_REG(regs.a0); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (a0)"); + core.regs.a0 = data; + + reg.id = RISCV_CORE_REG(regs.a1); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (a1)"); + core.regs.a1 = data; + + reg.id = RISCV_CORE_REG(regs.a2); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (a2)"); + core.regs.a2 = data; + + reg.id = RISCV_CORE_REG(regs.a3); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (a3)"); + core.regs.a3 = data; + + reg.id = RISCV_CORE_REG(regs.a4); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (a4)"); + core.regs.a4 = data; + + reg.id = RISCV_CORE_REG(regs.a5); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (a5)"); + core.regs.a5 = data; + + reg.id = RISCV_CORE_REG(regs.a6); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (a6)"); + core.regs.a6 = data; + + reg.id = RISCV_CORE_REG(regs.a7); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (a7)"); + core.regs.a7 = data; + + reg.id = RISCV_CORE_REG(regs.s2); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (s2)"); + core.regs.s2 = data; + + reg.id = RISCV_CORE_REG(regs.s3); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (s3)"); + core.regs.s3 = data; + + reg.id = RISCV_CORE_REG(regs.s4); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (s4)"); + core.regs.s4 = data; + + reg.id = RISCV_CORE_REG(regs.s5); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (s5)"); + core.regs.s5 = data; + + reg.id = RISCV_CORE_REG(regs.s6); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (s6)"); + core.regs.s6 = data; + + reg.id = RISCV_CORE_REG(regs.s7); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (s7)"); + core.regs.s7 = data; + + reg.id = RISCV_CORE_REG(regs.s8); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (s8)"); + core.regs.s8 = data; + + reg.id = RISCV_CORE_REG(regs.s9); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (s9)"); + core.regs.s9 = data; + + reg.id = RISCV_CORE_REG(regs.s10); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (s10)"); + core.regs.s10 = data; + + reg.id = RISCV_CORE_REG(regs.s11); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (s11)"); + core.regs.s11 = data; + + reg.id = RISCV_CORE_REG(regs.t3); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (t3)"); + core.regs.t3 = data; + + reg.id = RISCV_CORE_REG(regs.t4); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (t4)"); + core.regs.t4 = data; + + reg.id = RISCV_CORE_REG(regs.t5); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (t5)"); + core.regs.t5 = data; + + reg.id = RISCV_CORE_REG(regs.t6); + if (ioctl(vcpu->vcpu_fd, KVM_GET_ONE_REG, ®) < 0) + die("KVM_GET_ONE_REG failed (t6)"); + core.regs.t6 = data; + + dprintf(debug_fd, "\n General Purpose Registers:\n"); + dprintf(debug_fd, " -------------------------\n"); + dprintf(debug_fd, " MODE: 0x%lx\n", data); + dprintf(debug_fd, " PC: 0x%016lx RA: 0x%016lx SP: 0x%016lx GP: 0x%016lx\n", + core.regs.pc, core.regs.ra, core.regs.sp, core.regs.gp); + dprintf(debug_fd, " TP: 0x%016lx T0: 0x%016lx T1: 0x%016lx T2: 0x%016lx\n", + core.regs.tp, core.regs.t0, core.regs.t1, core.regs.t2); + dprintf(debug_fd, " S0: 0x%016lx S1: 0x%016lx A0: 0x%016lx A1: 0x%016lx\n", + core.regs.s0, core.regs.s1, core.regs.a0, core.regs.a1); + dprintf(debug_fd, " A2: 0x%016lx A3: 0x%016lx A4: 0x%016lx A5: 0x%016lx\n", + core.regs.a2, core.regs.a3, core.regs.a4, core.regs.a5); + dprintf(debug_fd, " A6: 0x%016lx A7: 0x%016lx S2: 0x%016lx S3: 0x%016lx\n", + core.regs.a6, core.regs.a7, core.regs.s2, core.regs.s3); + dprintf(debug_fd, " S4: 0x%016lx S5: 0x%016lx S6: 0x%016lx S7: 0x%016lx\n", + core.regs.s4, core.regs.s5, core.regs.s6, core.regs.s7); + dprintf(debug_fd, " S8: 0x%016lx S9: 0x%016lx S10: 0x%016lx S11: 0x%016lx\n", + core.regs.s8, core.regs.s9, core.regs.s10, core.regs.s11); + dprintf(debug_fd, " T3: 0x%016lx T4: 0x%016lx T5: 0x%016lx T6: 0x%016lx\n", + core.regs.t3, core.regs.t4, core.regs.t5, core.regs.t6); + + kvm_cpu__show_csrs(vcpu); } -- 2.25.1