On Fri, Jun 08, 2018 at 03:46:30PM +1000, Alexey Kardashevskiy wrote: > We are going to reuse multilevel TCE code for the userspace copy of > the TCE table and since it is big endian, let's make the copy big endian > too. > > Signed-off-by: Alexey Kardashevskiy <aik@xxxxxxxxx> Reviewed-by: David Gibson <david@xxxxxxxxxxxxxxxxxxxxx> > --- > arch/powerpc/include/asm/iommu.h | 2 +- > arch/powerpc/kvm/book3s_64_vio.c | 11 ++++++----- > arch/powerpc/kvm/book3s_64_vio_hv.c | 10 +++++----- > drivers/vfio/vfio_iommu_spapr_tce.c | 19 +++++++++---------- > 4 files changed, 21 insertions(+), 21 deletions(-) > > diff --git a/arch/powerpc/include/asm/iommu.h b/arch/powerpc/include/asm/iommu.h > index 20febe0..803ac70 100644 > --- a/arch/powerpc/include/asm/iommu.h > +++ b/arch/powerpc/include/asm/iommu.h > @@ -117,7 +117,7 @@ struct iommu_table { > unsigned long *it_map; /* A simple allocation bitmap for now */ > unsigned long it_page_shift;/* table iommu page size */ > struct list_head it_group_list;/* List of iommu_table_group_link */ > - unsigned long *it_userspace; /* userspace view of the table */ > + __be64 *it_userspace; /* userspace view of the table */ > struct iommu_table_ops *it_ops; > struct kref it_kref; > }; > diff --git a/arch/powerpc/kvm/book3s_64_vio.c b/arch/powerpc/kvm/book3s_64_vio.c > index 80ead38..1dbca4b 100644 > --- a/arch/powerpc/kvm/book3s_64_vio.c > +++ b/arch/powerpc/kvm/book3s_64_vio.c > @@ -378,19 +378,19 @@ static long kvmppc_tce_iommu_mapped_dec(struct kvm *kvm, > { > struct mm_iommu_table_group_mem_t *mem = NULL; > const unsigned long pgsize = 1ULL << tbl->it_page_shift; > - unsigned long *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry); > + __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry); > > if (!pua) > /* it_userspace allocation might be delayed */ > return H_TOO_HARD; > > - mem = mm_iommu_lookup(kvm->mm, *pua, pgsize); > + mem = mm_iommu_lookup(kvm->mm, be64_to_cpu(*pua), pgsize); > if (!mem) > return H_TOO_HARD; > > mm_iommu_mapped_dec(mem); > > - *pua = 0; > + *pua = cpu_to_be64(0); > > return H_SUCCESS; > } > @@ -437,7 +437,8 @@ long kvmppc_tce_iommu_do_map(struct kvm *kvm, struct iommu_table *tbl, > enum dma_data_direction dir) > { > long ret; > - unsigned long hpa, *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry); > + unsigned long hpa; > + __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry); > struct mm_iommu_table_group_mem_t *mem; > > if (!pua) > @@ -464,7 +465,7 @@ long kvmppc_tce_iommu_do_map(struct kvm *kvm, struct iommu_table *tbl, > if (dir != DMA_NONE) > kvmppc_tce_iommu_mapped_dec(kvm, tbl, entry); > > - *pua = ua; > + *pua = cpu_to_be64(ua); > > return 0; > } > diff --git a/arch/powerpc/kvm/book3s_64_vio_hv.c b/arch/powerpc/kvm/book3s_64_vio_hv.c > index 635f3ca..18109f3 100644 > --- a/arch/powerpc/kvm/book3s_64_vio_hv.c > +++ b/arch/powerpc/kvm/book3s_64_vio_hv.c > @@ -200,7 +200,7 @@ static long kvmppc_rm_tce_iommu_mapped_dec(struct kvm *kvm, > { > struct mm_iommu_table_group_mem_t *mem = NULL; > const unsigned long pgsize = 1ULL << tbl->it_page_shift; > - unsigned long *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry); > + __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry); > > if (!pua) > /* it_userspace allocation might be delayed */ > @@ -210,13 +210,13 @@ static long kvmppc_rm_tce_iommu_mapped_dec(struct kvm *kvm, > if (WARN_ON_ONCE_RM(!pua)) > return H_HARDWARE; > > - mem = mm_iommu_lookup_rm(kvm->mm, *pua, pgsize); > + mem = mm_iommu_lookup_rm(kvm->mm, be64_to_cpu(*pua), pgsize); > if (!mem) > return H_TOO_HARD; > > mm_iommu_mapped_dec(mem); > > - *pua = 0; > + *pua = cpu_to_be64(0); > > return H_SUCCESS; > } > @@ -268,7 +268,7 @@ static long kvmppc_rm_tce_iommu_do_map(struct kvm *kvm, struct iommu_table *tbl, > { > long ret; > unsigned long hpa = 0; > - unsigned long *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry); > + __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry); > struct mm_iommu_table_group_mem_t *mem; > > if (!pua) > @@ -302,7 +302,7 @@ static long kvmppc_rm_tce_iommu_do_map(struct kvm *kvm, struct iommu_table *tbl, > if (dir != DMA_NONE) > kvmppc_rm_tce_iommu_mapped_dec(kvm, tbl, entry); > > - *pua = ua; > + *pua = cpu_to_be64(ua); > > return 0; > } > diff --git a/drivers/vfio/vfio_iommu_spapr_tce.c b/drivers/vfio/vfio_iommu_spapr_tce.c > index 47071f3..81f48114 100644 > --- a/drivers/vfio/vfio_iommu_spapr_tce.c > +++ b/drivers/vfio/vfio_iommu_spapr_tce.c > @@ -231,7 +231,7 @@ static long tce_iommu_userspace_view_alloc(struct iommu_table *tbl, > decrement_locked_vm(mm, cb >> PAGE_SHIFT); > return -ENOMEM; > } > - tbl->it_userspace = uas; > + tbl->it_userspace = (__be64 *) uas; > > return 0; > } > @@ -494,20 +494,20 @@ static void tce_iommu_unuse_page_v2(struct tce_container *container, > struct mm_iommu_table_group_mem_t *mem = NULL; > int ret; > unsigned long hpa = 0; > - unsigned long *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry); > + __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry); > > if (!pua) > return; > > - ret = tce_iommu_prereg_ua_to_hpa(container, *pua, IOMMU_PAGE_SIZE(tbl), > - &hpa, &mem); > + ret = tce_iommu_prereg_ua_to_hpa(container, be64_to_cpu(*pua), > + IOMMU_PAGE_SIZE(tbl), &hpa, &mem); > if (ret) > - pr_debug("%s: tce %lx at #%lx was not cached, ret=%d\n", > - __func__, *pua, entry, ret); > + pr_debug("%s: tce %llx at #%lx was not cached, ret=%d\n", > + __func__, be64_to_cpu(*pua), entry, ret); > if (mem) > mm_iommu_mapped_dec(mem); > > - *pua = 0; > + *pua = cpu_to_be64(0); > } > > static int tce_iommu_clear(struct tce_container *container, > @@ -616,8 +616,7 @@ static long tce_iommu_build_v2(struct tce_container *container, > > for (i = 0; i < pages; ++i) { > struct mm_iommu_table_group_mem_t *mem = NULL; > - unsigned long *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, > - entry + i); > + __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry + i); > > ret = tce_iommu_prereg_ua_to_hpa(container, > tce, IOMMU_PAGE_SIZE(tbl), &hpa, &mem); > @@ -650,7 +649,7 @@ static long tce_iommu_build_v2(struct tce_container *container, > if (dirtmp != DMA_NONE) > tce_iommu_unuse_page_v2(container, tbl, entry + i); > > - *pua = tce; > + *pua = cpu_to_be64(tce); > > tce += IOMMU_PAGE_SIZE(tbl); > } -- David Gibson | I'll have my music baroque, and my code david AT gibson.dropbear.id.au | minimalist, thank you. NOT _the_ _other_ | _way_ _around_! http://www.ozlabs.org/~dgibson
Attachment:
signature.asc
Description: PGP signature