On Tue, Jan 14, 2014 at 11:36:30PM +0530, Anup Patel wrote: > On 14 January 2014 23:21, Christoffer Dall <christoffer.dall@xxxxxxxxxx> wrote: > > Present: > > -------- > > - Christoffer > > - Maria > > - Antonios > > - Stuart Yoder > > - Way (Samsung) > > - Victor Kamensky > > - Andrew Jones > > - Marc > > - Janner from Freescale > > - Peter > > +Anup > Sorry, missed that. (also sorry for people who's name I may have misspelled or misheard) > > > > Minutes: > > -------- > > - We should have some shared document for minutes in the future. > > - PSCI 0.2 support in KVM > > * Add the PSCI 0.2 function IDs to KVM and SYSTEM ops as part of that > > * Anup would add the PSCI 0.2 support and a capability accordingly > > * QEMU and kvmtools need to adjust the device tree for guests to add > > psci 0.2 nodes on supported systems > > * For this to work we need the guest bindings in place. > > - kvm-unit-tests status and discussion for future directions. > > * Drew will post a v3 soon that supports DT directly in our 'test > > guests' instead of the host-tools approach. > > * We need to move on to apply Christoffer's work > > * virtio-testdev needs to be reviewed and merged on the QEMU side, > > no reviews yet. > > * Next on Drew's side is aarch64 support. > > - Cache coherency on guest startup (guest runs with MMU disabled). > > This is a common problem that has also been observed on X-Gene and > > with Xen. It is prevalant with kvm-unit-tests because of the small > > payload. Suggested solutions include user space flushing, enabling > > the DC bit, flushing at fault time, ... I think we need to come up > > with some plan here. > > > > Possible solutions (Marc, please improve this): > > * DC enable bit will make all accesses cacheables, and disable the > > MMU on Stage-1, but if you have directly assigned devices and a > > bootloader that does uncached incoherent DMA accesses, this will > > not work. > > * At page fault time, if you're doing non-cacheable accesses, do > > cache clean before mapping in pages. But this breaks for code > > patching when turning the MMU on in the guest. This can be solved > > by having the TVM bit set when the MMU is off, and invalidate when > > MMU is enabled. This still breaks when the guest disables the MMU > > due to potential speculative loads. The final alternative is a > > paravirtualized MMU off operation. > > * Cleaning caches from user space only deals with the initial > > payload. It is easy to do this on v8, but not necessarily on v7, > > needs to be investigated. This may prevent the need to flush on > > non-cacheable accesses on a per-page basis. > > * Someone needs to look into whether user space caching is 'coherent > > enough' when done from user space in v7 and v8. > > * Marc is going to send out a patch seris that addresses the initial > > payload issue soon. > > * Investigate if I-cache is coherent when restoring a VM with MMU > > enabled/disabled. Verify it. > > > > - Dummy hypercall for kvm unit tests > > - Reading perf counters in kvm unit tests > > > > - Big Endian host support > > * Victor has posted v7 BE host support patches. > > * Victor is working on the v8 BE host support patches. > > * We discussed the vcpu_data_to_host and vice versa patch. Marc and > > Christoffer to take another look at the code and patches. > > _______________________________________________ > > kvmarm mailing list > > kvmarm@xxxxxxxxxxxxxxxxxxxxx > > https://lists.cs.columbia.edu/cucslists/listinfo/kvmarm -- Christoffer _______________________________________________ kvmarm mailing list kvmarm@xxxxxxxxxxxxxxxxxxxxx https://lists.cs.columbia.edu/cucslists/listinfo/kvmarm