On 07.01.20 21:59, Logan Gunthorpe wrote: > In prepartion to support a pgprot_t argument for arch_add_memory(). > > Cc: Heiko Carstens <heiko.carstens@xxxxxxxxxx> > Cc: Vasily Gorbik <gor@xxxxxxxxxxxxx> > Cc: Christian Borntraeger <borntraeger@xxxxxxxxxx> > Signed-off-by: Logan Gunthorpe <logang@xxxxxxxxxxxx> > --- > arch/s390/include/asm/pgtable.h | 3 ++- > arch/s390/mm/extmem.c | 3 ++- > arch/s390/mm/init.c | 2 +- > arch/s390/mm/vmem.c | 10 +++++----- > 4 files changed, 10 insertions(+), 8 deletions(-) > > diff --git a/arch/s390/include/asm/pgtable.h b/arch/s390/include/asm/pgtable.h > index 7b03037a8475..e667a1a96879 100644 > --- a/arch/s390/include/asm/pgtable.h > +++ b/arch/s390/include/asm/pgtable.h > @@ -1640,7 +1640,8 @@ static inline swp_entry_t __swp_entry(unsigned long type, unsigned long offset) > > #define kern_addr_valid(addr) (1) > > -extern int vmem_add_mapping(unsigned long start, unsigned long size); > +extern int vmem_add_mapping(unsigned long start, unsigned long size, > + pgprot_t prot); > extern int vmem_remove_mapping(unsigned long start, unsigned long size); > extern int s390_enable_sie(void); > extern int s390_enable_skey(void); > diff --git a/arch/s390/mm/extmem.c b/arch/s390/mm/extmem.c > index fd0dae9d10f4..6cf7029a7b35 100644 > --- a/arch/s390/mm/extmem.c > +++ b/arch/s390/mm/extmem.c > @@ -313,7 +313,8 @@ __segment_load (char *name, int do_nonshared, unsigned long *addr, unsigned long > goto out_free; > } > > - rc = vmem_add_mapping(seg->start_addr, seg->end - seg->start_addr + 1); > + rc = vmem_add_mapping(seg->start_addr, seg->end - seg->start_addr + 1, > + PAGE_KERNEL); > > if (rc) > goto out_free; > diff --git a/arch/s390/mm/init.c b/arch/s390/mm/init.c > index a0c88c1c9ad0..ef19522ddad2 100644 > --- a/arch/s390/mm/init.c > +++ b/arch/s390/mm/init.c > @@ -277,7 +277,7 @@ int arch_add_memory(int nid, u64 start, u64 size, > if (WARN_ON_ONCE(modifiers->altmap)) > return -EINVAL; > > - rc = vmem_add_mapping(start, size); > + rc = vmem_add_mapping(start, size, PAGE_KERNEL); > if (rc) > return rc; > > diff --git a/arch/s390/mm/vmem.c b/arch/s390/mm/vmem.c > index b403fa14847d..8a5e95f184a2 100644 > --- a/arch/s390/mm/vmem.c > +++ b/arch/s390/mm/vmem.c > @@ -66,7 +66,7 @@ pte_t __ref *vmem_pte_alloc(void) > /* > * Add a physical memory range to the 1:1 mapping. > */ > -static int vmem_add_mem(unsigned long start, unsigned long size) > +static int vmem_add_mem(unsigned long start, unsigned long size, pgprot_t prot) > { > unsigned long pgt_prot, sgt_prot, r3_prot; > unsigned long pages4k, pages1m, pages2g; > @@ -79,7 +79,7 @@ static int vmem_add_mem(unsigned long start, unsigned long size) > pte_t *pt_dir; > int ret = -ENOMEM; > > - pgt_prot = pgprot_val(PAGE_KERNEL); > + pgt_prot = pgprot_val(prot); > sgt_prot = pgprot_val(SEGMENT_KERNEL); > r3_prot = pgprot_val(REGION3_KERNEL); So, if we map as huge/gigantic pages, the protection would be discarded? That looks wrong. s390x does not support ZONE_DEVICE yet. Maybe simply bail out for s390x as you do for sh to make your life easier? [...] -- Thanks, David / dhildenb