On Thu, Jan 11, 2018 at 09:36:58AM +1100, Paul Mackerras wrote: > On Mon, Jan 08, 2018 at 04:29:31PM -0200, Jose Ricardo Ziviani wrote: > > This patch provides the MMIO load/store vector indexed > > X-Form emulation. > > > > Instructions implemented: lvx, stvx > > > > Signed-off-by: Jose Ricardo Ziviani <joserz@xxxxxxxxxxxxxxxxxx> > > What testing has been done of this patch? In particular, what > combinations of endianness of host and guest have been tested? For a > cross-endian situation (endianness of guest different from host) you > would need to load the two halves of the VMX register image in > reversed order (note that lvx/stvx are different from lxvd2x/stxvd2x > and lxvw4x/stxvw4x in this respect), and I don't see that being done. Hello! Thank you for reviewing it. Yes, I tested it with both in little endian mode. You're absolutely right, I'll review the cross-endian scenario that's not covered indeed. > > Also, lvx/stvx mask off the bottom 4 bits of the EA, and I don't see > that being done. > > [snip] > > > diff --git a/arch/powerpc/kvm/powerpc.c b/arch/powerpc/kvm/powerpc.c > > index 1915e86cef6f..7d0f60359ee0 100644 > > --- a/arch/powerpc/kvm/powerpc.c > > +++ b/arch/powerpc/kvm/powerpc.c > > @@ -832,23 +832,7 @@ void kvm_arch_irq_bypass_del_producer(struct irq_bypass_consumer *cons, > > kvm->arch.kvm_ops->irq_bypass_del_producer(cons, prod); > > } > > > > -#ifdef CONFIG_VSX > > -static inline int kvmppc_get_vsr_dword_offset(int index) > > -{ > > - int offset; > > - > > - if ((index != 0) && (index != 1)) > > - return -1; > > - > > -#ifdef __BIG_ENDIAN > > - offset = index; > > -#else > > - offset = 1 - index; > > -#endif > > - > > - return offset; > > -} > > - > > Why does this function need to be moved down in the file? The main reason is that I need it too but I need it available with ALTIVEC(1) support, so I changed the guard to ALTIVEC and moved it down closer to other related functions under the same guard. But I can move it back if you prefer. (1)AFAIK it's possible to have ALTIVEC enabled but VSX disabled, not the other way around. > > > +#ifdef CONFIG_ALTIVEC > > static inline int kvmppc_get_vsr_word_offset(int index) > > { > > int offset; > > @@ -864,6 +848,40 @@ static inline int kvmppc_get_vsr_word_offset(int index) > > return offset; > > } > > > > +static inline void kvmppc_set_vmx_dword(struct kvm_vcpu *vcpu, > > + u64 gpr) > > +{ > > + int index = vcpu->arch.io_gpr & KVM_MMIO_REG_MASK; > > + u64 hi = gpr >> 32; > > + u64 lo = gpr & 0xffffffff; > > + > > + if (vcpu->arch.mmio_vmx_copy_nums == 1) { > > + VCPU_VSX_VR(vcpu, index).u[kvmppc_get_vsr_word_offset(2)] = hi; > > + VCPU_VSX_VR(vcpu, index).u[kvmppc_get_vsr_word_offset(3)] = lo; > > + } else if (vcpu->arch.mmio_vmx_copy_nums == 2) { > > + VCPU_VSX_VR(vcpu, index).u[kvmppc_get_vsr_word_offset(0)] = hi; > > + VCPU_VSX_VR(vcpu, index).u[kvmppc_get_vsr_word_offset(1)] = lo; > > + } > > This splitting of the value into hi/lo words looks to me like you're > assuming a big-endian byte ordering. It looks like it will end up > swapping the words in each half of the register on a little-endian > host (regardless of the endianness of the guest). Yep! I focused on VCPU_VSX_VR code to fill the vector correctly but missed the 64-bit value splitting. I'll fix it and improve the testing, considering other endian scenario. > > > +} > > +#endif /* CONFIG_ALTIVEC */ > > + > > +#ifdef CONFIG_VSX > > +static inline int kvmppc_get_vsr_dword_offset(int index) > > +{ > > + int offset; > > + > > + if ((index != 0) && (index != 1)) > > + return -1; > > + > > +#ifdef __BIG_ENDIAN > > + offset = index; > > +#else > > + offset = 1 - index; > > +#endif > > + > > + return offset; > > +} > > + > > static inline void kvmppc_set_vsr_dword(struct kvm_vcpu *vcpu, > > u64 gpr) > > { > > @@ -1027,6 +1045,11 @@ static void kvmppc_complete_mmio_load(struct kvm_vcpu *vcpu, > > KVMPPC_VSX_COPY_DWORD_LOAD_DUMP) > > kvmppc_set_vsr_dword_dump(vcpu, gpr); > > break; > > +#endif > > +#ifdef CONFIG_ALTIVEC > > + case KVM_MMIO_REG_VMX: > > + kvmppc_set_vmx_dword(vcpu, gpr); > > + break; > > #endif > > default: > > BUG(); > > @@ -1307,6 +1330,99 @@ static int kvmppc_emulate_mmio_vsx_loadstore(struct kvm_vcpu *vcpu, > > } > > #endif /* CONFIG_VSX */ > > > > +#ifdef CONFIG_ALTIVEC > > +/* handle quadword load access in two halves */ > > +int kvmppc_handle_load128_by2x64(struct kvm_run *run, struct kvm_vcpu *vcpu, > > + unsigned int rt, int is_default_endian) > > +{ > > + enum emulation_result emulated = EMULATE_DONE; > > + > > + while (vcpu->arch.mmio_vmx_copy_nums) { > > + emulated = __kvmppc_handle_load(run, vcpu, rt, 8, > > + is_default_endian, 0); > > + if (emulated != EMULATE_DONE) > > + break; > > + > > + vcpu->arch.mmio_vmx_copy_nums--; > > + } > > + return emulated; > > +} > > + > > +static inline int kvmppc_get_vmx_data(struct kvm_vcpu *vcpu, int rs, u64 *val) > > +{ > > + > > + if (vcpu->arch.mmio_vmx_copy_nums == 1) { > > + *val = VCPU_VSX_VR(vcpu, rs).u[kvmppc_get_vsr_word_offset(2)]; > > + *val = (*val << 32) | VCPU_VSX_VR(vcpu, rs). > > + u[kvmppc_get_vsr_word_offset(3)]; > > + return 0; > > + } else if (vcpu->arch.mmio_vmx_copy_nums == 2) { > > + *val = VCPU_VSX_VR(vcpu, rs).u[kvmppc_get_vsr_word_offset(0)]; > > + *val = (*val << 32) | VCPU_VSX_VR(vcpu, rs). > > + u[kvmppc_get_vsr_word_offset(1)]; > > + return 0; > > + } > > + return -1; > > +} > > Once again, the way this puts two word values together to get a > doubleword looks wrong for a little-endian host. Got it. It's clear the mixed-endianess confusion. I'll fix it, improve the test cases and send a V2. Thank you Paul! > > Paul. > -- > To unsubscribe from this list: send the line "unsubscribe kvm-ppc" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html > -- To unsubscribe from this list: send the line "unsubscribe kvm-ppc" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html