2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
|
|
* for more details.
|
|
|
|
*
|
|
|
|
* Copyright (C) 1995 Linus Torvalds
|
|
|
|
* Copyright (C) 1995 Waldorf Electronics
|
|
|
|
* Copyright (C) 1994, 95, 96, 97, 98, 99, 2000, 01, 02, 03 Ralf Baechle
|
|
|
|
* Copyright (C) 1996 Stoned Elipot
|
|
|
|
* Copyright (C) 1999 Silicon Graphics, Inc.
|
2013-01-22 19:59:30 +08:00
|
|
|
* Copyright (C) 2000, 2001, 2002, 2007 Maciej W. Rozycki
|
2005-04-17 06:20:36 +08:00
|
|
|
*/
|
|
|
|
#include <linux/init.h>
|
|
|
|
#include <linux/ioport.h>
|
2011-07-24 04:30:40 +08:00
|
|
|
#include <linux/export.h>
|
2006-07-10 19:44:13 +08:00
|
|
|
#include <linux/screen_info.h>
|
2011-12-09 02:22:09 +08:00
|
|
|
#include <linux/memblock.h>
|
2005-04-17 06:20:36 +08:00
|
|
|
#include <linux/initrd.h>
|
|
|
|
#include <linux/root_dev.h>
|
|
|
|
#include <linux/highmem.h>
|
|
|
|
#include <linux/console.h>
|
2006-03-27 17:16:04 +08:00
|
|
|
#include <linux/pfn.h>
|
2007-06-29 23:55:48 +08:00
|
|
|
#include <linux/debugfs.h>
|
2012-10-12 00:14:58 +08:00
|
|
|
#include <linux/kexec.h>
|
2013-04-13 19:15:47 +08:00
|
|
|
#include <linux/sizes.h>
|
2014-07-16 23:51:32 +08:00
|
|
|
#include <linux/device.h>
|
|
|
|
#include <linux/dma-contiguous.h>
|
2016-05-11 06:50:03 +08:00
|
|
|
#include <linux/decompress/generic.h>
|
2016-11-23 21:43:46 +08:00
|
|
|
#include <linux/of_fdt.h>
|
2019-05-04 01:50:41 +08:00
|
|
|
#include <linux/of_reserved_mem.h>
|
2020-02-05 12:08:33 +08:00
|
|
|
#include <linux/dmi.h>
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
#include <asm/addrspace.h>
|
|
|
|
#include <asm/bootinfo.h>
|
2007-10-23 19:43:11 +08:00
|
|
|
#include <asm/bugs.h>
|
2005-07-13 19:48:45 +08:00
|
|
|
#include <asm/cache.h>
|
2015-01-29 19:14:13 +08:00
|
|
|
#include <asm/cdmm.h>
|
2005-04-17 06:20:36 +08:00
|
|
|
#include <asm/cpu.h>
|
2015-09-23 01:10:55 +08:00
|
|
|
#include <asm/debug.h>
|
2018-06-15 19:08:45 +08:00
|
|
|
#include <asm/dma-coherence.h>
|
2005-04-17 06:20:36 +08:00
|
|
|
#include <asm/sections.h>
|
|
|
|
#include <asm/setup.h>
|
2007-11-19 20:23:51 +08:00
|
|
|
#include <asm/smp-ops.h>
|
2010-10-13 14:52:46 +08:00
|
|
|
#include <asm/prom.h>
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2015-09-11 22:46:14 +08:00
|
|
|
#ifdef CONFIG_MIPS_ELF_APPENDED_DTB
|
|
|
|
const char __section(.appended_dtb) __appended_dtb[0x100000];
|
|
|
|
#endif /* CONFIG_MIPS_ELF_APPENDED_DTB */
|
|
|
|
|
2005-07-13 19:48:45 +08:00
|
|
|
struct cpuinfo_mips cpu_data[NR_CPUS] __read_mostly;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
EXPORT_SYMBOL(cpu_data);
|
|
|
|
|
|
|
|
#ifdef CONFIG_VT
|
|
|
|
struct screen_info screen_info;
|
|
|
|
#endif
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Setup information
|
|
|
|
*
|
|
|
|
* These are initialized so they are in the .data section
|
|
|
|
*/
|
2005-07-13 19:48:45 +08:00
|
|
|
unsigned long mips_machtype __read_mostly = MACH_UNKNOWN;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
EXPORT_SYMBOL(mips_machtype);
|
|
|
|
|
2009-11-22 04:34:41 +08:00
|
|
|
static char __initdata command_line[COMMAND_LINE_SIZE];
|
|
|
|
char __initdata arcs_cmdline[COMMAND_LINE_SIZE];
|
|
|
|
|
|
|
|
#ifdef CONFIG_CMDLINE_BOOL
|
2019-10-13 04:43:38 +08:00
|
|
|
static const char builtin_cmdline[] __initconst = CONFIG_CMDLINE;
|
2019-10-13 04:43:36 +08:00
|
|
|
#else
|
|
|
|
static const char builtin_cmdline[] __initconst = "";
|
2009-11-22 04:34:41 +08:00
|
|
|
#endif
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* mips_io_port_base is the begin of the address space to which x86 style
|
|
|
|
* I/O ports are mapped.
|
|
|
|
*/
|
2019-07-30 05:10:12 +08:00
|
|
|
unsigned long mips_io_port_base = -1;
|
2005-04-17 06:20:36 +08:00
|
|
|
EXPORT_SYMBOL(mips_io_port_base);
|
|
|
|
|
|
|
|
static struct resource code_resource = { .name = "Kernel code", };
|
|
|
|
static struct resource data_resource = { .name = "Kernel data", };
|
2017-10-13 03:50:34 +08:00
|
|
|
static struct resource bss_resource = { .name = "Kernel bss", };
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2013-04-13 19:15:47 +08:00
|
|
|
static void *detect_magic __initdata = detect_memory_region;
|
|
|
|
|
2018-07-28 09:23:20 +08:00
|
|
|
#ifdef CONFIG_MIPS_AUTO_PFN_OFFSET
|
|
|
|
unsigned long ARCH_PFN_OFFSET;
|
|
|
|
EXPORT_SYMBOL(ARCH_PFN_OFFSET);
|
|
|
|
#endif
|
|
|
|
|
2014-11-22 07:22:09 +08:00
|
|
|
void __init add_memory_region(phys_addr_t start, phys_addr_t size, long type)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2019-08-19 22:23:13 +08:00
|
|
|
/*
|
|
|
|
* Note: This function only exists for historical reason,
|
|
|
|
* new code should use memblock_add or memblock_add_node instead.
|
|
|
|
*/
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2016-08-09 20:21:48 +08:00
|
|
|
/*
|
|
|
|
* If the region reaches the top of the physical address space, adjust
|
|
|
|
* the size slightly so that (start + size) doesn't overflow
|
|
|
|
*/
|
2018-06-15 06:28:02 +08:00
|
|
|
if (start + size - 1 == PHYS_ADDR_MAX)
|
2016-08-09 20:21:48 +08:00
|
|
|
--size;
|
|
|
|
|
2006-08-11 23:51:48 +08:00
|
|
|
/* Sanity check */
|
|
|
|
if (start + size < start) {
|
2014-10-05 00:50:42 +08:00
|
|
|
pr_warn("Trying to add an invalid memory region, skipped\n");
|
2006-08-11 23:51:48 +08:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2019-09-24 23:20:51 +08:00
|
|
|
if (start < PHYS_OFFSET)
|
|
|
|
return;
|
|
|
|
|
2019-08-19 22:23:13 +08:00
|
|
|
memblock_add(start, size);
|
|
|
|
/* Reserve any memory except the ordinary RAM ranges. */
|
|
|
|
switch (type) {
|
|
|
|
case BOOT_MEM_RAM:
|
|
|
|
break;
|
2012-11-15 19:53:59 +08:00
|
|
|
|
2019-08-19 22:23:13 +08:00
|
|
|
case BOOT_MEM_NOMAP: /* Discard the range from the system. */
|
|
|
|
memblock_remove(start, size);
|
|
|
|
break;
|
2012-11-15 19:53:59 +08:00
|
|
|
|
2019-08-19 22:23:13 +08:00
|
|
|
default: /* Reserve the rest of the memory types at boot time */
|
|
|
|
memblock_reserve(start, size);
|
|
|
|
break;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-11-22 07:22:09 +08:00
|
|
|
void __init detect_memory_region(phys_addr_t start, phys_addr_t sz_min, phys_addr_t sz_max)
|
2013-04-13 19:15:47 +08:00
|
|
|
{
|
|
|
|
void *dm = &detect_magic;
|
2014-11-22 07:22:09 +08:00
|
|
|
phys_addr_t size;
|
2013-04-13 19:15:47 +08:00
|
|
|
|
|
|
|
for (size = sz_min; size < sz_max; size <<= 1) {
|
|
|
|
if (!memcmp(dm, dm + size, sizeof(detect_magic)))
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
pr_debug("Memory: %lluMB of RAM detected at 0x%llx (min: %lluMB, max: %lluMB)\n",
|
|
|
|
((unsigned long long) size) / SZ_1M,
|
|
|
|
(unsigned long long) start,
|
|
|
|
((unsigned long long) sz_min) / SZ_1M,
|
|
|
|
((unsigned long long) sz_max) / SZ_1M);
|
|
|
|
|
|
|
|
add_memory_region(start, size, BOOT_MEM_RAM);
|
|
|
|
}
|
|
|
|
|
2006-08-11 23:51:49 +08:00
|
|
|
/*
|
|
|
|
* Manage initrd
|
|
|
|
*/
|
|
|
|
#ifdef CONFIG_BLK_DEV_INITRD
|
|
|
|
|
2006-08-11 23:51:53 +08:00
|
|
|
static int __init rd_start_early(char *p)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2006-08-11 23:51:53 +08:00
|
|
|
unsigned long start = memparse(p, &p);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2005-09-04 06:56:16 +08:00
|
|
|
#ifdef CONFIG_64BIT
|
2006-10-19 19:20:04 +08:00
|
|
|
/* Guess if the sign extension was forgotten by bootloader */
|
|
|
|
if (start < XKPHYS)
|
|
|
|
start = (int)start;
|
2005-04-17 06:20:36 +08:00
|
|
|
#endif
|
2006-08-11 23:51:53 +08:00
|
|
|
initrd_start = start;
|
|
|
|
initrd_end += start;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
early_param("rd_start", rd_start_early);
|
|
|
|
|
|
|
|
static int __init rd_size_early(char *p)
|
|
|
|
{
|
|
|
|
initrd_end += memparse(p, &p);
|
2005-04-17 06:20:36 +08:00
|
|
|
return 0;
|
|
|
|
}
|
2006-08-11 23:51:53 +08:00
|
|
|
early_param("rd_size", rd_size_early);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2006-10-19 19:20:04 +08:00
|
|
|
/* it returns the next free pfn after initrd */
|
2006-08-11 23:51:49 +08:00
|
|
|
static unsigned long __init init_initrd(void)
|
|
|
|
{
|
2006-10-19 19:20:04 +08:00
|
|
|
unsigned long end;
|
2006-08-11 23:51:49 +08:00
|
|
|
|
|
|
|
/*
|
2006-08-11 23:51:53 +08:00
|
|
|
* Board specific code or command line parser should have
|
|
|
|
* already set up initrd_start and initrd_end. In these cases
|
|
|
|
* perfom sanity checks and use them if all looks good.
|
2006-08-11 23:51:49 +08:00
|
|
|
*/
|
2009-12-17 09:57:07 +08:00
|
|
|
if (!initrd_start || initrd_end <= initrd_start)
|
2006-10-19 19:20:04 +08:00
|
|
|
goto disable;
|
|
|
|
|
|
|
|
if (initrd_start & ~PAGE_MASK) {
|
2008-07-28 20:12:52 +08:00
|
|
|
pr_err("initrd start must be page aligned\n");
|
2006-10-19 19:20:04 +08:00
|
|
|
goto disable;
|
2006-08-11 23:51:49 +08:00
|
|
|
}
|
2006-10-19 19:20:04 +08:00
|
|
|
if (initrd_start < PAGE_OFFSET) {
|
2008-07-28 20:12:52 +08:00
|
|
|
pr_err("initrd start < PAGE_OFFSET\n");
|
2006-10-19 19:20:04 +08:00
|
|
|
goto disable;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Sanitize initrd addresses. For example firmware
|
|
|
|
* can't guess if they need to pass them through
|
|
|
|
* 64-bits values if the kernel has been built in pure
|
|
|
|
* 32-bit. We need also to switch from KSEG0 to XKPHYS
|
|
|
|
* addresses now, so the code can now safely use __pa().
|
|
|
|
*/
|
|
|
|
end = __pa(initrd_end);
|
|
|
|
initrd_end = (unsigned long)__va(end);
|
|
|
|
initrd_start = (unsigned long)__va(__pa(initrd_start));
|
|
|
|
|
|
|
|
ROOT_DEV = Root_RAM0;
|
|
|
|
return PFN_UP(end);
|
|
|
|
disable:
|
|
|
|
initrd_start = 0;
|
|
|
|
initrd_end = 0;
|
|
|
|
return 0;
|
2006-08-11 23:51:49 +08:00
|
|
|
}
|
|
|
|
|
2016-05-11 06:50:03 +08:00
|
|
|
/* In some conditions (e.g. big endian bootloader with a little endian
|
|
|
|
kernel), the initrd might appear byte swapped. Try to detect this and
|
|
|
|
byte swap it if needed. */
|
|
|
|
static void __init maybe_bswap_initrd(void)
|
|
|
|
{
|
|
|
|
#if defined(CONFIG_CPU_CAVIUM_OCTEON)
|
|
|
|
u64 buf;
|
|
|
|
|
|
|
|
/* Check for CPIO signature */
|
|
|
|
if (!memcmp((void *)initrd_start, "070701", 6))
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* Check for compressed initrd */
|
|
|
|
if (decompress_method((unsigned char *)initrd_start, 8, NULL))
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* Try again with a byte swapped header */
|
|
|
|
buf = swab64p((u64 *)initrd_start);
|
|
|
|
if (!memcmp(&buf, "070701", 6) ||
|
|
|
|
decompress_method((unsigned char *)(&buf), 8, NULL)) {
|
|
|
|
unsigned long i;
|
|
|
|
|
|
|
|
pr_info("Byteswapped initrd detected\n");
|
|
|
|
for (i = initrd_start; i < ALIGN(initrd_end, 8); i += 8)
|
|
|
|
swab64s((u64 *)i);
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2006-08-11 23:51:49 +08:00
|
|
|
static void __init finalize_initrd(void)
|
|
|
|
{
|
|
|
|
unsigned long size = initrd_end - initrd_start;
|
|
|
|
|
|
|
|
if (size == 0) {
|
|
|
|
printk(KERN_INFO "Initrd not found or empty");
|
|
|
|
goto disable;
|
|
|
|
}
|
2006-10-19 19:20:01 +08:00
|
|
|
if (__pa(initrd_end) > PFN_PHYS(max_low_pfn)) {
|
2008-07-28 20:12:52 +08:00
|
|
|
printk(KERN_ERR "Initrd extends beyond end of memory");
|
2006-08-11 23:51:49 +08:00
|
|
|
goto disable;
|
|
|
|
}
|
|
|
|
|
2016-05-11 06:50:03 +08:00
|
|
|
maybe_bswap_initrd();
|
|
|
|
|
2018-09-10 17:23:18 +08:00
|
|
|
memblock_reserve(__pa(initrd_start), size);
|
2006-08-11 23:51:49 +08:00
|
|
|
initrd_below_start_ok = 1;
|
|
|
|
|
2008-07-28 20:12:52 +08:00
|
|
|
pr_info("Initial ramdisk at: 0x%lx (%lu bytes)\n",
|
|
|
|
initrd_start, size);
|
2006-08-11 23:51:49 +08:00
|
|
|
return;
|
|
|
|
disable:
|
2008-07-28 20:12:52 +08:00
|
|
|
printk(KERN_CONT " - disabling initrd\n");
|
2006-08-11 23:51:49 +08:00
|
|
|
initrd_start = 0;
|
|
|
|
initrd_end = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
#else /* !CONFIG_BLK_DEV_INITRD */
|
|
|
|
|
2006-10-13 18:22:52 +08:00
|
|
|
static unsigned long __init init_initrd(void)
|
|
|
|
{
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2006-08-11 23:51:49 +08:00
|
|
|
#define finalize_initrd() do {} while (0)
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
2006-08-11 23:51:48 +08:00
|
|
|
/*
|
|
|
|
* Initialize the bootmem allocator. It also setup initrd related data
|
|
|
|
* if needed.
|
|
|
|
*/
|
2019-10-20 22:43:13 +08:00
|
|
|
#if defined(CONFIG_SGI_IP27) || (defined(CONFIG_CPU_LOONGSON64) && defined(CONFIG_NUMA))
|
2006-08-11 23:51:49 +08:00
|
|
|
|
2006-08-11 23:51:48 +08:00
|
|
|
static void __init bootmem_init(void)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2006-08-11 23:51:49 +08:00
|
|
|
init_initrd();
|
|
|
|
finalize_initrd();
|
|
|
|
}
|
|
|
|
|
|
|
|
#else /* !CONFIG_SGI_IP27 */
|
|
|
|
|
|
|
|
static void __init bootmem_init(void)
|
|
|
|
{
|
2019-08-19 22:23:13 +08:00
|
|
|
struct memblock_region *mem;
|
|
|
|
phys_addr_t ramstart, ramend;
|
|
|
|
|
|
|
|
ramstart = memblock_start_of_DRAM();
|
|
|
|
ramend = memblock_end_of_DRAM();
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
/*
|
2010-09-08 13:50:43 +08:00
|
|
|
* Sanity check any INITRD first. We don't take it into account
|
|
|
|
* for bootmem setup initially, rely on the end-of-kernel-code
|
|
|
|
* as our memory range starting point. Once bootmem is inited we
|
|
|
|
* will reserve the area used for the initrd.
|
2005-04-17 06:20:36 +08:00
|
|
|
*/
|
2010-09-08 13:50:43 +08:00
|
|
|
init_initrd();
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2019-04-24 06:47:40 +08:00
|
|
|
/* Reserve memory occupied by kernel. */
|
|
|
|
memblock_reserve(__pa_symbol(&_text),
|
|
|
|
__pa_symbol(&_end) - __pa_symbol(&_text));
|
2018-09-10 17:23:18 +08:00
|
|
|
|
2019-08-19 22:23:13 +08:00
|
|
|
/* max_low_pfn is not a number of pages but the end pfn of low mem */
|
|
|
|
|
|
|
|
#ifdef CONFIG_MIPS_AUTO_PFN_OFFSET
|
|
|
|
ARCH_PFN_OFFSET = PFN_UP(ramstart);
|
|
|
|
#else
|
2007-01-10 16:44:04 +08:00
|
|
|
/*
|
2019-08-19 22:23:13 +08:00
|
|
|
* Reserve any memory between the start of RAM and PHYS_OFFSET
|
2007-01-10 16:44:04 +08:00
|
|
|
*/
|
2019-08-19 22:23:13 +08:00
|
|
|
if (ramstart > PHYS_OFFSET)
|
2019-09-24 23:19:56 +08:00
|
|
|
memblock_reserve(PHYS_OFFSET, ramstart - PHYS_OFFSET);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2019-08-19 22:23:13 +08:00
|
|
|
if (PFN_UP(ramstart) > ARCH_PFN_OFFSET) {
|
|
|
|
pr_info("Wasting %lu bytes for tracking %lu unused pages\n",
|
|
|
|
(unsigned long)((PFN_UP(ramstart) - ARCH_PFN_OFFSET) * sizeof(struct page)),
|
|
|
|
(unsigned long)(PFN_UP(ramstart) - ARCH_PFN_OFFSET));
|
|
|
|
}
|
|
|
|
#endif
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2019-08-19 22:23:13 +08:00
|
|
|
min_low_pfn = ARCH_PFN_OFFSET;
|
|
|
|
max_pfn = PFN_DOWN(ramend);
|
|
|
|
for_each_memblock(memory, mem) {
|
|
|
|
unsigned long start = memblock_region_memory_base_pfn(mem);
|
|
|
|
unsigned long end = memblock_region_memory_end_pfn(mem);
|
2018-02-01 19:37:21 +08:00
|
|
|
|
2016-11-01 21:59:09 +08:00
|
|
|
/*
|
|
|
|
* Skip highmem here so we get an accurate max_low_pfn if low
|
|
|
|
* memory stops short of high memory.
|
|
|
|
* If the region overlaps HIGHMEM_START, end is clipped so
|
|
|
|
* max_pfn excludes the highmem portion.
|
|
|
|
*/
|
2019-08-19 22:23:13 +08:00
|
|
|
if (memblock_is_nomap(mem))
|
|
|
|
continue;
|
2016-11-01 21:59:09 +08:00
|
|
|
if (start >= PFN_DOWN(HIGHMEM_START))
|
|
|
|
continue;
|
|
|
|
if (end > PFN_DOWN(HIGHMEM_START))
|
|
|
|
end = PFN_DOWN(HIGHMEM_START);
|
2007-01-10 16:44:04 +08:00
|
|
|
if (end > max_low_pfn)
|
|
|
|
max_low_pfn = end;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2018-07-28 09:23:20 +08:00
|
|
|
if (min_low_pfn >= max_low_pfn)
|
|
|
|
panic("Incorrect memory mapping !!!");
|
|
|
|
|
2019-08-19 22:23:13 +08:00
|
|
|
if (max_pfn > PFN_DOWN(HIGHMEM_START)) {
|
2006-08-11 23:51:48 +08:00
|
|
|
#ifdef CONFIG_HIGHMEM
|
|
|
|
highstart_pfn = PFN_DOWN(HIGHMEM_START);
|
2019-08-19 22:23:13 +08:00
|
|
|
highend_pfn = max_pfn;
|
|
|
|
#else
|
2007-01-10 16:44:04 +08:00
|
|
|
max_low_pfn = PFN_DOWN(HIGHMEM_START);
|
2019-08-19 22:23:13 +08:00
|
|
|
max_pfn = max_low_pfn;
|
2019-04-24 06:47:38 +08:00
|
|
|
#endif
|
2019-08-19 22:23:13 +08:00
|
|
|
}
|
2011-11-22 22:38:03 +08:00
|
|
|
|
2006-08-11 23:51:49 +08:00
|
|
|
/*
|
|
|
|
* Reserve initrd memory if needed.
|
|
|
|
*/
|
|
|
|
finalize_initrd();
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2006-08-11 23:51:49 +08:00
|
|
|
#endif /* CONFIG_SGI_IP27 */
|
|
|
|
|
2009-09-17 08:25:07 +08:00
|
|
|
static int usermem __initdata;
|
2006-08-11 23:51:53 +08:00
|
|
|
|
|
|
|
static int __init early_parse_mem(char *p)
|
|
|
|
{
|
2014-12-13 02:51:15 +08:00
|
|
|
phys_addr_t start, size;
|
2006-08-11 23:51:53 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* If a user specifies memory size, we
|
|
|
|
* blow away any automatically generated
|
|
|
|
* size.
|
|
|
|
*/
|
|
|
|
if (usermem == 0) {
|
|
|
|
usermem = 1;
|
2019-08-19 22:23:13 +08:00
|
|
|
memblock_remove(memblock_start_of_DRAM(),
|
|
|
|
memblock_end_of_DRAM() - memblock_start_of_DRAM());
|
2013-01-22 19:59:30 +08:00
|
|
|
}
|
2006-08-11 23:51:53 +08:00
|
|
|
start = 0;
|
|
|
|
size = memparse(p, &p);
|
|
|
|
if (*p == '@')
|
|
|
|
start = memparse(p + 1, &p);
|
|
|
|
|
|
|
|
add_memory_region(start, size, BOOT_MEM_RAM);
|
2016-11-23 21:43:49 +08:00
|
|
|
|
2006-08-11 23:51:53 +08:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
early_param("mem", early_parse_mem);
|
2006-06-18 08:32:22 +08:00
|
|
|
|
2017-06-19 23:50:08 +08:00
|
|
|
static int __init early_parse_memmap(char *p)
|
|
|
|
{
|
|
|
|
char *oldp;
|
|
|
|
u64 start_at, mem_size;
|
|
|
|
|
|
|
|
if (!p)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (!strncmp(p, "exactmap", 8)) {
|
|
|
|
pr_err("\"memmap=exactmap\" invalid on MIPS\n");
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
oldp = p;
|
|
|
|
mem_size = memparse(p, &p);
|
|
|
|
if (p == oldp)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (*p == '@') {
|
|
|
|
start_at = memparse(p+1, &p);
|
|
|
|
add_memory_region(start_at, mem_size, BOOT_MEM_RAM);
|
|
|
|
} else if (*p == '#') {
|
|
|
|
pr_err("\"memmap=nn#ss\" (force ACPI data) invalid on MIPS\n");
|
|
|
|
return -EINVAL;
|
|
|
|
} else if (*p == '$') {
|
|
|
|
start_at = memparse(p+1, &p);
|
|
|
|
add_memory_region(start_at, mem_size, BOOT_MEM_RESERVED);
|
|
|
|
} else {
|
|
|
|
pr_err("\"memmap\" invalid format!\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (*p == '\0') {
|
|
|
|
usermem = 1;
|
|
|
|
return 0;
|
|
|
|
} else
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
early_param("memmap", early_parse_memmap);
|
|
|
|
|
2013-02-13 03:41:48 +08:00
|
|
|
#ifdef CONFIG_PROC_VMCORE
|
|
|
|
unsigned long setup_elfcorehdr, setup_elfcorehdr_size;
|
|
|
|
static int __init early_parse_elfcorehdr(char *p)
|
|
|
|
{
|
2019-08-19 22:23:13 +08:00
|
|
|
struct memblock_region *mem;
|
2013-02-13 03:41:48 +08:00
|
|
|
|
|
|
|
setup_elfcorehdr = memparse(p, &p);
|
|
|
|
|
2019-08-19 22:23:13 +08:00
|
|
|
for_each_memblock(memory, mem) {
|
|
|
|
unsigned long start = mem->base;
|
2019-08-24 01:00:38 +08:00
|
|
|
unsigned long end = start + mem->size;
|
2013-02-13 03:41:48 +08:00
|
|
|
if (setup_elfcorehdr >= start && setup_elfcorehdr < end) {
|
|
|
|
/*
|
|
|
|
* Reserve from the elf core header to the end of
|
|
|
|
* the memory segment, that should all be kdump
|
|
|
|
* reserved memory.
|
|
|
|
*/
|
|
|
|
setup_elfcorehdr_size = end - setup_elfcorehdr;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If we don't find it in the memory map, then we shouldn't
|
|
|
|
* have to worry about it, as the new kernel won't use it.
|
|
|
|
*/
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
early_param("elfcorehdr", early_parse_elfcorehdr);
|
|
|
|
#endif
|
|
|
|
|
2013-09-05 01:56:24 +08:00
|
|
|
#ifdef CONFIG_KEXEC
|
|
|
|
static void __init mips_parse_crashkernel(void)
|
|
|
|
{
|
|
|
|
unsigned long long total_mem;
|
|
|
|
unsigned long long crash_size, crash_base;
|
|
|
|
int ret;
|
|
|
|
|
2019-08-19 22:23:13 +08:00
|
|
|
total_mem = memblock_phys_mem_size();
|
2013-09-05 01:56:24 +08:00
|
|
|
ret = parse_crashkernel(boot_command_line, total_mem,
|
|
|
|
&crash_size, &crash_base);
|
|
|
|
if (ret != 0 || crash_size <= 0)
|
|
|
|
return;
|
|
|
|
|
2020-07-25 13:56:38 +08:00
|
|
|
if (!memblock_find_in_range(crash_base, crash_base + crash_size, crash_size, 1)) {
|
2016-11-23 21:43:50 +08:00
|
|
|
pr_warn("Invalid memory region reserved for crash kernel\n");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2013-09-05 01:56:24 +08:00
|
|
|
crashk_res.start = crash_base;
|
|
|
|
crashk_res.end = crash_base + crash_size - 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void __init request_crashkernel(struct resource *res)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
2016-11-23 21:43:43 +08:00
|
|
|
if (crashk_res.start == crashk_res.end)
|
|
|
|
return;
|
|
|
|
|
2013-09-05 01:56:24 +08:00
|
|
|
ret = request_resource(res, &crashk_res);
|
|
|
|
if (!ret)
|
|
|
|
pr_info("Reserving %ldMB of memory at %ldMB for crashkernel\n",
|
2020-01-02 01:49:48 +08:00
|
|
|
(unsigned long)(resource_size(&crashk_res) >> 20),
|
2013-09-05 01:56:24 +08:00
|
|
|
(unsigned long)(crashk_res.start >> 20));
|
|
|
|
}
|
|
|
|
#else /* !defined(CONFIG_KEXEC) */
|
|
|
|
static void __init mips_parse_crashkernel(void)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
|
|
|
static void __init request_crashkernel(struct resource *res)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
#endif /* !defined(CONFIG_KEXEC) */
|
|
|
|
|
2019-08-19 22:23:13 +08:00
|
|
|
static void __init check_kernel_sections_mem(void)
|
|
|
|
{
|
|
|
|
phys_addr_t start = PFN_PHYS(PFN_DOWN(__pa_symbol(&_text)));
|
|
|
|
phys_addr_t size = PFN_PHYS(PFN_UP(__pa_symbol(&_end))) - start;
|
|
|
|
|
|
|
|
if (!memblock_is_region_memory(start, size)) {
|
|
|
|
pr_info("Kernel sections are not in the memory maps\n");
|
|
|
|
memblock_add(start, size);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-10-10 07:09:45 +08:00
|
|
|
static void __init bootcmdline_append(const char *s, size_t max)
|
|
|
|
{
|
|
|
|
if (!s[0] || !max)
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (boot_command_line[0])
|
|
|
|
strlcat(boot_command_line, " ", COMMAND_LINE_SIZE);
|
|
|
|
|
|
|
|
strlcat(boot_command_line, s, max);
|
|
|
|
}
|
|
|
|
|
2019-10-13 04:43:37 +08:00
|
|
|
#ifdef CONFIG_OF_EARLY_FLATTREE
|
|
|
|
|
2019-10-10 07:09:45 +08:00
|
|
|
static int __init bootcmdline_scan_chosen(unsigned long node, const char *uname,
|
|
|
|
int depth, void *data)
|
|
|
|
{
|
|
|
|
bool *dt_bootargs = data;
|
|
|
|
const char *p;
|
|
|
|
int l;
|
|
|
|
|
|
|
|
if (depth != 1 || !data ||
|
|
|
|
(strcmp(uname, "chosen") != 0 && strcmp(uname, "chosen@0") != 0))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
p = of_get_flat_dt_prop(node, "bootargs", &l);
|
|
|
|
if (p != NULL && l > 0) {
|
|
|
|
bootcmdline_append(p, min(l, COMMAND_LINE_SIZE));
|
|
|
|
*dt_bootargs = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2019-10-13 04:43:37 +08:00
|
|
|
#endif /* CONFIG_OF_EARLY_FLATTREE */
|
|
|
|
|
2020-05-18 15:08:08 +08:00
|
|
|
static void __init bootcmdline_init(void)
|
2019-10-10 07:09:45 +08:00
|
|
|
{
|
|
|
|
bool dt_bootargs = false;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If CMDLINE_OVERRIDE is enabled then initializing the command line is
|
|
|
|
* trivial - we simply use the built-in command line unconditionally &
|
|
|
|
* unmodified.
|
|
|
|
*/
|
|
|
|
if (IS_ENABLED(CONFIG_CMDLINE_OVERRIDE)) {
|
|
|
|
strlcpy(boot_command_line, builtin_cmdline, COMMAND_LINE_SIZE);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If the user specified a built-in command line &
|
|
|
|
* MIPS_CMDLINE_BUILTIN_EXTEND, then the built-in command line is
|
|
|
|
* prepended to arguments from the bootloader or DT so we'll copy them
|
|
|
|
* to the start of boot_command_line here. Otherwise, empty
|
|
|
|
* boot_command_line to undo anything early_init_dt_scan_chosen() did.
|
|
|
|
*/
|
|
|
|
if (IS_ENABLED(CONFIG_MIPS_CMDLINE_BUILTIN_EXTEND))
|
|
|
|
strlcpy(boot_command_line, builtin_cmdline, COMMAND_LINE_SIZE);
|
|
|
|
else
|
|
|
|
boot_command_line[0] = 0;
|
|
|
|
|
2019-10-13 04:43:37 +08:00
|
|
|
#ifdef CONFIG_OF_EARLY_FLATTREE
|
2019-10-10 07:09:45 +08:00
|
|
|
/*
|
|
|
|
* If we're configured to take boot arguments from DT, look for those
|
|
|
|
* now.
|
|
|
|
*/
|
2020-02-25 23:28:09 +08:00
|
|
|
if (IS_ENABLED(CONFIG_MIPS_CMDLINE_FROM_DTB) ||
|
|
|
|
IS_ENABLED(CONFIG_MIPS_CMDLINE_DTB_EXTEND))
|
2019-10-10 07:09:45 +08:00
|
|
|
of_scan_flat_dt(bootcmdline_scan_chosen, &dt_bootargs);
|
2019-10-13 04:43:37 +08:00
|
|
|
#endif
|
2019-10-10 07:09:45 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* If we didn't get any arguments from DT (regardless of whether that's
|
|
|
|
* because we weren't configured to look for them, or because we looked
|
|
|
|
* & found none) then we'll take arguments from the bootloader.
|
|
|
|
* plat_mem_setup() should have filled arcs_cmdline with arguments from
|
|
|
|
* the bootloader.
|
|
|
|
*/
|
|
|
|
if (IS_ENABLED(CONFIG_MIPS_CMDLINE_DTB_EXTEND) || !dt_bootargs)
|
|
|
|
bootcmdline_append(arcs_cmdline, COMMAND_LINE_SIZE);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If the user specified a built-in command line & we didn't already
|
|
|
|
* prepend it, we append it to boot_command_line here.
|
|
|
|
*/
|
|
|
|
if (IS_ENABLED(CONFIG_CMDLINE_BOOL) &&
|
|
|
|
!IS_ENABLED(CONFIG_MIPS_CMDLINE_BUILTIN_EXTEND))
|
|
|
|
bootcmdline_append(builtin_cmdline, COMMAND_LINE_SIZE);
|
|
|
|
}
|
2015-10-12 19:13:02 +08:00
|
|
|
|
2018-09-01 06:28:57 +08:00
|
|
|
/*
|
|
|
|
* arch_mem_init - initialize memory management subsystem
|
|
|
|
*
|
|
|
|
* o plat_mem_setup() detects the memory configuration and will record detected
|
|
|
|
* memory areas using add_memory_region.
|
|
|
|
*
|
|
|
|
* At this stage the memory configuration of the system is known to the
|
|
|
|
* kernel but generic memory management system is still entirely uninitialized.
|
|
|
|
*
|
|
|
|
* o bootmem_init()
|
|
|
|
* o sparse_init()
|
|
|
|
* o paging_init()
|
|
|
|
* o dma_contiguous_reserve()
|
|
|
|
*
|
|
|
|
* At this stage the bootmem allocator is ready to use.
|
|
|
|
*
|
|
|
|
* NOTE: historically plat_mem_setup did the entire platform initialization.
|
|
|
|
* This was rather impractical because it meant plat_mem_setup had to
|
|
|
|
* get away without any kind of memory allocator. To keep old code from
|
|
|
|
* breaking plat_setup was just renamed to plat_mem_setup and a second platform
|
|
|
|
* initialization hook for anything else was introduced.
|
|
|
|
*/
|
2013-02-13 03:41:47 +08:00
|
|
|
static void __init arch_mem_init(char **cmdline_p)
|
|
|
|
{
|
2018-09-28 06:59:18 +08:00
|
|
|
/* call board setup routine */
|
|
|
|
plat_mem_setup();
|
2018-11-10 11:50:14 +08:00
|
|
|
memblock_set_bottom_up(true);
|
2018-09-28 06:59:18 +08:00
|
|
|
|
2020-05-18 15:08:08 +08:00
|
|
|
bootcmdline_init();
|
2009-11-22 04:34:41 +08:00
|
|
|
strlcpy(command_line, boot_command_line, COMMAND_LINE_SIZE);
|
2006-06-18 08:32:22 +08:00
|
|
|
*cmdline_p = command_line;
|
|
|
|
|
2006-08-11 23:51:53 +08:00
|
|
|
parse_early_param();
|
|
|
|
|
2019-08-19 22:23:13 +08:00
|
|
|
if (usermem)
|
|
|
|
pr_info("User-defined physical RAM map overwrite\n");
|
|
|
|
|
|
|
|
check_kernel_sections_mem();
|
2006-08-11 23:51:53 +08:00
|
|
|
|
2016-11-23 21:43:46 +08:00
|
|
|
early_init_fdt_reserve_self();
|
|
|
|
early_init_fdt_scan_reserved_mem();
|
|
|
|
|
2019-08-19 22:23:13 +08:00
|
|
|
#ifndef CONFIG_NUMA
|
|
|
|
memblock_set_node(0, PHYS_ADDR_MAX, &memblock.memory, 0);
|
|
|
|
#endif
|
2006-06-18 08:32:22 +08:00
|
|
|
bootmem_init();
|
2018-09-10 17:23:18 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Prevent memblock from allocating high memory.
|
|
|
|
* This cannot be done before max_low_pfn is detected, so up
|
|
|
|
* to this point is possible to only reserve physical memory
|
2018-10-31 06:08:04 +08:00
|
|
|
* with memblock_reserve; memblock_alloc* can be used
|
2018-09-10 17:23:18 +08:00
|
|
|
* only after this point
|
|
|
|
*/
|
|
|
|
memblock_set_current_limit(PFN_PHYS(max_low_pfn));
|
|
|
|
|
2013-02-13 03:41:48 +08:00
|
|
|
#ifdef CONFIG_PROC_VMCORE
|
|
|
|
if (setup_elfcorehdr && setup_elfcorehdr_size) {
|
|
|
|
printk(KERN_INFO "kdump reserved memory at %lx-%lx\n",
|
|
|
|
setup_elfcorehdr, setup_elfcorehdr_size);
|
2018-09-10 17:23:18 +08:00
|
|
|
memblock_reserve(setup_elfcorehdr, setup_elfcorehdr_size);
|
2013-02-13 03:41:48 +08:00
|
|
|
}
|
|
|
|
#endif
|
2013-09-05 01:56:24 +08:00
|
|
|
|
|
|
|
mips_parse_crashkernel();
|
2012-10-12 00:14:58 +08:00
|
|
|
#ifdef CONFIG_KEXEC
|
|
|
|
if (crashk_res.start != crashk_res.end)
|
2020-01-02 01:49:48 +08:00
|
|
|
memblock_reserve(crashk_res.start, resource_size(&crashk_res));
|
2012-10-12 00:14:58 +08:00
|
|
|
#endif
|
2010-10-13 14:52:46 +08:00
|
|
|
device_tree_init();
|
MIPS: Make sparse_init() using top-down allocation
In the current code, if CONFIG_SWIOTLB is set, when failed to get IO TLB
memory from the low pages by plat_swiotlb_setup(), it may lead to the boot
process failed with kernel panic.
(1) On the Loongson and SiByte platform
arch/mips/loongson64/dma.c
arch/mips/sibyte/common/dma.c
void __init plat_swiotlb_setup(void)
{
swiotlb_init(1);
}
kernel/dma/swiotlb.c
void __init
swiotlb_init(int verbose)
{
...
vstart = memblock_alloc_low(PAGE_ALIGN(bytes), PAGE_SIZE);
if (vstart && !swiotlb_init_with_tbl(vstart, io_tlb_nslabs, verbose))
return;
...
pr_warn("Cannot allocate buffer");
no_iotlb_memory = true;
}
phys_addr_t swiotlb_tbl_map_single()
{
...
if (no_iotlb_memory)
panic("Can not allocate SWIOTLB buffer earlier ...");
...
}
(2) On the Cavium OCTEON platform
arch/mips/cavium-octeon/dma-octeon.c
void __init plat_swiotlb_setup(void)
{
...
octeon_swiotlb = memblock_alloc_low(swiotlbsize, PAGE_SIZE);
if (!octeon_swiotlb)
panic("%s: Failed to allocate %zu bytes align=%lx\n",
__func__, swiotlbsize, PAGE_SIZE);
...
}
Because IO_TLB_DEFAULT_SIZE is 64M, if the rest size of low memory is less
than 64M when call plat_swiotlb_setup(), we can easily reproduce the panic
case.
In order to reduce the possibility of kernel panic when failed to get IO
TLB memory under CONFIG_SWIOTLB, it is better to allocate low memory as
small as possible before plat_swiotlb_setup(), so make sparse_init() using
top-down allocation.
Reported-by: Juxin Gao <gaojuxin@loongson.cn>
Co-developed-by: Juxin Gao <gaojuxin@loongson.cn>
Signed-off-by: Juxin Gao <gaojuxin@loongson.cn>
Signed-off-by: Tiezhu Yang <yangtiezhu@loongson.cn>
Signed-off-by: Thomas Bogendoerfer <tsbogend@alpha.franken.de>
2020-04-21 19:59:46 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* In order to reduce the possibility of kernel panic when failed to
|
|
|
|
* get IO TLB memory under CONFIG_SWIOTLB, it is better to allocate
|
|
|
|
* low memory as small as possible before plat_swiotlb_setup(), so
|
|
|
|
* make sparse_init() using top-down allocation.
|
|
|
|
*/
|
|
|
|
memblock_set_bottom_up(false);
|
2006-06-18 08:32:22 +08:00
|
|
|
sparse_init();
|
MIPS: Make sparse_init() using top-down allocation
In the current code, if CONFIG_SWIOTLB is set, when failed to get IO TLB
memory from the low pages by plat_swiotlb_setup(), it may lead to the boot
process failed with kernel panic.
(1) On the Loongson and SiByte platform
arch/mips/loongson64/dma.c
arch/mips/sibyte/common/dma.c
void __init plat_swiotlb_setup(void)
{
swiotlb_init(1);
}
kernel/dma/swiotlb.c
void __init
swiotlb_init(int verbose)
{
...
vstart = memblock_alloc_low(PAGE_ALIGN(bytes), PAGE_SIZE);
if (vstart && !swiotlb_init_with_tbl(vstart, io_tlb_nslabs, verbose))
return;
...
pr_warn("Cannot allocate buffer");
no_iotlb_memory = true;
}
phys_addr_t swiotlb_tbl_map_single()
{
...
if (no_iotlb_memory)
panic("Can not allocate SWIOTLB buffer earlier ...");
...
}
(2) On the Cavium OCTEON platform
arch/mips/cavium-octeon/dma-octeon.c
void __init plat_swiotlb_setup(void)
{
...
octeon_swiotlb = memblock_alloc_low(swiotlbsize, PAGE_SIZE);
if (!octeon_swiotlb)
panic("%s: Failed to allocate %zu bytes align=%lx\n",
__func__, swiotlbsize, PAGE_SIZE);
...
}
Because IO_TLB_DEFAULT_SIZE is 64M, if the rest size of low memory is less
than 64M when call plat_swiotlb_setup(), we can easily reproduce the panic
case.
In order to reduce the possibility of kernel panic when failed to get IO
TLB memory under CONFIG_SWIOTLB, it is better to allocate low memory as
small as possible before plat_swiotlb_setup(), so make sparse_init() using
top-down allocation.
Reported-by: Juxin Gao <gaojuxin@loongson.cn>
Co-developed-by: Juxin Gao <gaojuxin@loongson.cn>
Signed-off-by: Juxin Gao <gaojuxin@loongson.cn>
Signed-off-by: Tiezhu Yang <yangtiezhu@loongson.cn>
Signed-off-by: Thomas Bogendoerfer <tsbogend@alpha.franken.de>
2020-04-21 19:59:46 +08:00
|
|
|
memblock_set_bottom_up(true);
|
|
|
|
|
2010-10-02 04:27:33 +08:00
|
|
|
plat_swiotlb_setup();
|
2014-07-16 23:51:32 +08:00
|
|
|
|
|
|
|
dma_contiguous_reserve(PFN_PHYS(max_low_pfn));
|
2016-03-17 20:37:10 +08:00
|
|
|
|
2019-04-24 06:47:42 +08:00
|
|
|
/* Reserve for hibernation. */
|
|
|
|
memblock_reserve(__pa_symbol(&__nosave_begin),
|
|
|
|
__pa_symbol(&__nosave_end) - __pa_symbol(&__nosave_begin));
|
2019-05-04 01:50:37 +08:00
|
|
|
|
2019-05-04 01:50:41 +08:00
|
|
|
fdt_init_reserved_mem();
|
|
|
|
|
2019-05-04 01:50:37 +08:00
|
|
|
memblock_dump_all();
|
2019-05-04 01:50:38 +08:00
|
|
|
|
2019-08-19 22:23:13 +08:00
|
|
|
early_memtest(PFN_PHYS(ARCH_PFN_OFFSET), PFN_PHYS(max_low_pfn));
|
2006-06-18 08:32:22 +08:00
|
|
|
}
|
|
|
|
|
2006-08-11 23:51:51 +08:00
|
|
|
static void __init resource_init(void)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2019-08-19 22:23:13 +08:00
|
|
|
struct memblock_region *region;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2006-06-20 19:47:53 +08:00
|
|
|
if (UNCAC_BASE != IO_BASE)
|
|
|
|
return;
|
|
|
|
|
2006-10-19 19:20:03 +08:00
|
|
|
code_resource.start = __pa_symbol(&_text);
|
|
|
|
code_resource.end = __pa_symbol(&_etext) - 1;
|
|
|
|
data_resource.start = __pa_symbol(&_etext);
|
|
|
|
data_resource.end = __pa_symbol(&_edata) - 1;
|
2017-10-13 03:50:34 +08:00
|
|
|
bss_resource.start = __pa_symbol(&__bss_start);
|
|
|
|
bss_resource.end = __pa_symbol(&__bss_stop) - 1;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2019-08-19 22:23:13 +08:00
|
|
|
for_each_memblock(memory, region) {
|
|
|
|
phys_addr_t start = PFN_PHYS(memblock_region_memory_base_pfn(region));
|
|
|
|
phys_addr_t end = PFN_PHYS(memblock_region_memory_end_pfn(region)) - 1;
|
2005-04-17 06:20:36 +08:00
|
|
|
struct resource *res;
|
|
|
|
|
memblock: stop using implicit alignment to SMP_CACHE_BYTES
When a memblock allocation APIs are called with align = 0, the alignment
is implicitly set to SMP_CACHE_BYTES.
Implicit alignment is done deep in the memblock allocator and it can
come as a surprise. Not that such an alignment would be wrong even
when used incorrectly but it is better to be explicit for the sake of
clarity and the prinicple of the least surprise.
Replace all such uses of memblock APIs with the 'align' parameter
explicitly set to SMP_CACHE_BYTES and stop implicit alignment assignment
in the memblock internal allocation functions.
For the case when memblock APIs are used via helper functions, e.g. like
iommu_arena_new_node() in Alpha, the helper functions were detected with
Coccinelle's help and then manually examined and updated where
appropriate.
The direct memblock APIs users were updated using the semantic patch below:
@@
expression size, min_addr, max_addr, nid;
@@
(
|
- memblock_alloc_try_nid_raw(size, 0, min_addr, max_addr, nid)
+ memblock_alloc_try_nid_raw(size, SMP_CACHE_BYTES, min_addr, max_addr,
nid)
|
- memblock_alloc_try_nid_nopanic(size, 0, min_addr, max_addr, nid)
+ memblock_alloc_try_nid_nopanic(size, SMP_CACHE_BYTES, min_addr, max_addr,
nid)
|
- memblock_alloc_try_nid(size, 0, min_addr, max_addr, nid)
+ memblock_alloc_try_nid(size, SMP_CACHE_BYTES, min_addr, max_addr, nid)
|
- memblock_alloc(size, 0)
+ memblock_alloc(size, SMP_CACHE_BYTES)
|
- memblock_alloc_raw(size, 0)
+ memblock_alloc_raw(size, SMP_CACHE_BYTES)
|
- memblock_alloc_from(size, 0, min_addr)
+ memblock_alloc_from(size, SMP_CACHE_BYTES, min_addr)
|
- memblock_alloc_nopanic(size, 0)
+ memblock_alloc_nopanic(size, SMP_CACHE_BYTES)
|
- memblock_alloc_low(size, 0)
+ memblock_alloc_low(size, SMP_CACHE_BYTES)
|
- memblock_alloc_low_nopanic(size, 0)
+ memblock_alloc_low_nopanic(size, SMP_CACHE_BYTES)
|
- memblock_alloc_from_nopanic(size, 0, min_addr)
+ memblock_alloc_from_nopanic(size, SMP_CACHE_BYTES, min_addr)
|
- memblock_alloc_node(size, 0, nid)
+ memblock_alloc_node(size, SMP_CACHE_BYTES, nid)
)
[mhocko@suse.com: changelog update]
[akpm@linux-foundation.org: coding-style fixes]
[rppt@linux.ibm.com: fix missed uses of implicit alignment]
Link: http://lkml.kernel.org/r/20181016133656.GA10925@rapoport-lnx
Link: http://lkml.kernel.org/r/1538687224-17535-1-git-send-email-rppt@linux.vnet.ibm.com
Signed-off-by: Mike Rapoport <rppt@linux.vnet.ibm.com>
Suggested-by: Michal Hocko <mhocko@suse.com>
Acked-by: Paul Burton <paul.burton@mips.com> [MIPS]
Acked-by: Michael Ellerman <mpe@ellerman.id.au> [powerpc]
Acked-by: Michal Hocko <mhocko@suse.com>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: Chris Zankel <chris@zankel.net>
Cc: Geert Uytterhoeven <geert@linux-m68k.org>
Cc: Guan Xuetao <gxt@pku.edu.cn>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Matt Turner <mattst88@gmail.com>
Cc: Michal Simek <monstr@monstr.eu>
Cc: Richard Weinberger <richard@nod.at>
Cc: Russell King <linux@armlinux.org.uk>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Tony Luck <tony.luck@intel.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2018-10-31 06:09:57 +08:00
|
|
|
res = memblock_alloc(sizeof(struct resource), SMP_CACHE_BYTES);
|
2019-03-12 14:30:31 +08:00
|
|
|
if (!res)
|
|
|
|
panic("%s: Failed to allocate %zu bytes\n", __func__,
|
|
|
|
sizeof(struct resource));
|
2016-01-27 04:57:22 +08:00
|
|
|
|
|
|
|
res->start = start;
|
|
|
|
res->end = end;
|
2019-08-19 22:23:13 +08:00
|
|
|
res->flags = IORESOURCE_SYSTEM_RAM | IORESOURCE_BUSY;
|
|
|
|
res->name = "System RAM";
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
request_resource(&iomem_resource, res);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We don't know which RAM region contains kernel data,
|
|
|
|
* so we try it repeatedly and let the resource manager
|
|
|
|
* test it.
|
|
|
|
*/
|
|
|
|
request_resource(res, &code_resource);
|
|
|
|
request_resource(res, &data_resource);
|
2017-10-13 03:50:34 +08:00
|
|
|
request_resource(res, &bss_resource);
|
2012-10-12 00:14:58 +08:00
|
|
|
request_crashkernel(res);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-06-26 11:41:25 +08:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
static void __init prefill_possible_map(void)
|
|
|
|
{
|
|
|
|
int i, possible = num_possible_cpus();
|
|
|
|
|
|
|
|
if (possible > nr_cpu_ids)
|
|
|
|
possible = nr_cpu_ids;
|
|
|
|
|
|
|
|
for (i = 0; i < possible; i++)
|
|
|
|
set_cpu_possible(i, true);
|
|
|
|
for (; i < NR_CPUS; i++)
|
|
|
|
set_cpu_possible(i, false);
|
|
|
|
|
|
|
|
nr_cpu_ids = possible;
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
static inline void prefill_possible_map(void) {}
|
|
|
|
#endif
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
void __init setup_arch(char **cmdline_p)
|
|
|
|
{
|
|
|
|
cpu_probe();
|
2016-02-09 01:46:31 +08:00
|
|
|
mips_cm_probe();
|
2005-04-17 06:20:36 +08:00
|
|
|
prom_init();
|
2007-03-01 19:56:43 +08:00
|
|
|
|
2015-01-29 19:14:13 +08:00
|
|
|
setup_early_fdc_console();
|
2007-03-01 19:56:43 +08:00
|
|
|
#ifdef CONFIG_EARLY_PRINTK
|
2008-05-29 22:57:08 +08:00
|
|
|
setup_early_printk();
|
2007-03-01 19:56:43 +08:00
|
|
|
#endif
|
2005-04-17 06:20:36 +08:00
|
|
|
cpu_report();
|
2007-10-23 19:43:11 +08:00
|
|
|
check_bugs_early();
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
#if defined(CONFIG_VT)
|
|
|
|
#if defined(CONFIG_VGA_CONSOLE)
|
2007-02-05 08:10:11 +08:00
|
|
|
conswitchp = &vga_con;
|
2005-04-17 06:20:36 +08:00
|
|
|
#endif
|
|
|
|
#endif
|
|
|
|
|
2006-06-18 08:32:22 +08:00
|
|
|
arch_mem_init(cmdline_p);
|
2020-02-05 12:08:33 +08:00
|
|
|
dmi_setup();
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
resource_init();
|
2006-02-23 20:23:27 +08:00
|
|
|
plat_smp_setup();
|
2014-06-26 11:41:25 +08:00
|
|
|
prefill_possible_map();
|
2012-05-15 15:04:50 +08:00
|
|
|
|
|
|
|
cpu_cache_init();
|
2016-09-02 22:17:31 +08:00
|
|
|
paging_init();
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2007-01-24 00:21:05 +08:00
|
|
|
unsigned long kernelsp[NR_CPUS];
|
|
|
|
unsigned long fw_arg0, fw_arg1, fw_arg2, fw_arg3;
|
2007-06-29 23:55:48 +08:00
|
|
|
|
2016-06-20 17:27:37 +08:00
|
|
|
#ifdef CONFIG_USE_OF
|
|
|
|
unsigned long fw_passed_dtb;
|
|
|
|
#endif
|
|
|
|
|
2007-06-29 23:55:48 +08:00
|
|
|
#ifdef CONFIG_DEBUG_FS
|
|
|
|
struct dentry *mips_debugfs_dir;
|
|
|
|
static int __init debugfs_mips(void)
|
|
|
|
{
|
2019-01-22 22:57:42 +08:00
|
|
|
mips_debugfs_dir = debugfs_create_dir("mips", NULL);
|
2007-06-29 23:55:48 +08:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
arch_initcall(debugfs_mips);
|
|
|
|
#endif
|
2018-06-15 19:08:45 +08:00
|
|
|
|
2018-08-16 21:47:53 +08:00
|
|
|
#ifdef CONFIG_DMA_MAYBE_COHERENT
|
2018-06-15 19:08:45 +08:00
|
|
|
/* User defined DMA coherency from command line. */
|
|
|
|
enum coherent_io_user_state coherentio = IO_COHERENCE_DEFAULT;
|
|
|
|
EXPORT_SYMBOL_GPL(coherentio);
|
2020-04-21 19:59:44 +08:00
|
|
|
int hw_coherentio; /* Actual hardware supported DMA coherency setting. */
|
2018-06-15 19:08:45 +08:00
|
|
|
|
|
|
|
static int __init setcoherentio(char *str)
|
|
|
|
{
|
|
|
|
coherentio = IO_COHERENCE_ENABLED;
|
|
|
|
pr_info("Hardware DMA cache coherency (command line)\n");
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
early_param("coherentio", setcoherentio);
|
|
|
|
|
|
|
|
static int __init setnocoherentio(char *str)
|
|
|
|
{
|
|
|
|
coherentio = IO_COHERENCE_DISABLED;
|
|
|
|
pr_info("Software DMA cache coherency (command line)\n");
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
early_param("nocoherentio", setnocoherentio);
|
|
|
|
#endif
|