mm: move ioremap_page_range to vmalloc.c
Patch series "small ioremap cleanups". The first patch moves a little code around the vmalloc/ioremap boundary following a bigger move by Nick earlier. The second enforces non-executable mapping on ioremap just like we do for vmap. No driver currently uses executable mappings anyway, as they should. This patch (of 2): This keeps it together with the implementation, and to remove the vmap_range wrapper. Link: https://lkml.kernel.org/r/20210824091259.1324527-1-hch@lst.de Link: https://lkml.kernel.org/r/20210824091259.1324527-2-hch@lst.de Signed-off-by: Christoph Hellwig <hch@lst.de> Reviewed-by: Nicholas Piggin <npiggin@gmail.com> Cc: Peter Zijlstra <peterz@infradead.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
8350229ffc
commit
82a70ce042
|
@ -225,9 +225,6 @@ static inline bool is_vm_area_hugepages(const void *addr)
|
|||
}
|
||||
|
||||
#ifdef CONFIG_MMU
|
||||
int vmap_range(unsigned long addr, unsigned long end,
|
||||
phys_addr_t phys_addr, pgprot_t prot,
|
||||
unsigned int max_page_shift);
|
||||
void vunmap_range(unsigned long addr, unsigned long end);
|
||||
static inline void set_vm_flush_reset_perms(void *addr)
|
||||
{
|
||||
|
|
|
@ -38,7 +38,7 @@ mmu-y := nommu.o
|
|||
mmu-$(CONFIG_MMU) := highmem.o memory.o mincore.o \
|
||||
mlock.o mmap.o mmu_gather.o mprotect.o mremap.o \
|
||||
msync.o page_vma_mapped.o pagewalk.o \
|
||||
pgtable-generic.o rmap.o vmalloc.o ioremap.o
|
||||
pgtable-generic.o rmap.o vmalloc.o
|
||||
|
||||
|
||||
ifdef CONFIG_CROSS_MEMORY_ATTACH
|
||||
|
@ -128,3 +128,4 @@ obj-$(CONFIG_PTDUMP_CORE) += ptdump.o
|
|||
obj-$(CONFIG_PAGE_REPORTING) += page_reporting.o
|
||||
obj-$(CONFIG_IO_MAPPING) += io-mapping.o
|
||||
obj-$(CONFIG_HAVE_BOOTMEM_INFO_NODE) += bootmem_info.o
|
||||
obj-$(CONFIG_GENERIC_IOREMAP) += ioremap.o
|
||||
|
|
25
mm/ioremap.c
25
mm/ioremap.c
|
@ -8,33 +8,9 @@
|
|||
*/
|
||||
#include <linux/vmalloc.h>
|
||||
#include <linux/mm.h>
|
||||
#include <linux/sched.h>
|
||||
#include <linux/io.h>
|
||||
#include <linux/export.h>
|
||||
#include <asm/cacheflush.h>
|
||||
|
||||
#include "pgalloc-track.h"
|
||||
|
||||
#ifdef CONFIG_HAVE_ARCH_HUGE_VMAP
|
||||
static unsigned int __ro_after_init iomap_max_page_shift = BITS_PER_LONG - 1;
|
||||
|
||||
static int __init set_nohugeiomap(char *str)
|
||||
{
|
||||
iomap_max_page_shift = PAGE_SHIFT;
|
||||
return 0;
|
||||
}
|
||||
early_param("nohugeiomap", set_nohugeiomap);
|
||||
#else /* CONFIG_HAVE_ARCH_HUGE_VMAP */
|
||||
static const unsigned int iomap_max_page_shift = PAGE_SHIFT;
|
||||
#endif /* CONFIG_HAVE_ARCH_HUGE_VMAP */
|
||||
|
||||
int ioremap_page_range(unsigned long addr,
|
||||
unsigned long end, phys_addr_t phys_addr, pgprot_t prot)
|
||||
{
|
||||
return vmap_range(addr, end, phys_addr, prot, iomap_max_page_shift);
|
||||
}
|
||||
|
||||
#ifdef CONFIG_GENERIC_IOREMAP
|
||||
void __iomem *ioremap_prot(phys_addr_t addr, size_t size, unsigned long prot)
|
||||
{
|
||||
unsigned long offset, vaddr;
|
||||
|
@ -71,4 +47,3 @@ void iounmap(volatile void __iomem *addr)
|
|||
vunmap((void *)((unsigned long)addr & PAGE_MASK));
|
||||
}
|
||||
EXPORT_SYMBOL(iounmap);
|
||||
#endif /* CONFIG_GENERIC_IOREMAP */
|
||||
|
|
22
mm/vmalloc.c
22
mm/vmalloc.c
|
@ -44,6 +44,19 @@
|
|||
#include "internal.h"
|
||||
#include "pgalloc-track.h"
|
||||
|
||||
#ifdef CONFIG_HAVE_ARCH_HUGE_VMAP
|
||||
static unsigned int __ro_after_init ioremap_max_page_shift = BITS_PER_LONG - 1;
|
||||
|
||||
static int __init set_nohugeiomap(char *str)
|
||||
{
|
||||
ioremap_max_page_shift = PAGE_SHIFT;
|
||||
return 0;
|
||||
}
|
||||
early_param("nohugeiomap", set_nohugeiomap);
|
||||
#else /* CONFIG_HAVE_ARCH_HUGE_VMAP */
|
||||
static const unsigned int ioremap_max_page_shift = PAGE_SHIFT;
|
||||
#endif /* CONFIG_HAVE_ARCH_HUGE_VMAP */
|
||||
|
||||
#ifdef CONFIG_HAVE_ARCH_HUGE_VMALLOC
|
||||
static bool __ro_after_init vmap_allow_huge = true;
|
||||
|
||||
|
@ -298,15 +311,14 @@ static int vmap_range_noflush(unsigned long addr, unsigned long end,
|
|||
return err;
|
||||
}
|
||||
|
||||
int vmap_range(unsigned long addr, unsigned long end,
|
||||
phys_addr_t phys_addr, pgprot_t prot,
|
||||
unsigned int max_page_shift)
|
||||
int ioremap_page_range(unsigned long addr, unsigned long end,
|
||||
phys_addr_t phys_addr, pgprot_t prot)
|
||||
{
|
||||
int err;
|
||||
|
||||
err = vmap_range_noflush(addr, end, phys_addr, prot, max_page_shift);
|
||||
err = vmap_range_noflush(addr, end, phys_addr, prot,
|
||||
ioremap_max_page_shift);
|
||||
flush_cache_vmap(addr, end);
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
|
|
Loading…
Reference in New Issue