2019-06-04 16:11:33 +08:00
|
|
|
// SPDX-License-Identifier: GPL-2.0-only
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* linux/arch/arm/kernel/module.c
|
|
|
|
*
|
|
|
|
* Copyright (C) 2002 Russell King.
|
2006-09-26 16:37:07 +08:00
|
|
|
* Modified for nommu by Hyok S. Choi
|
2005-04-17 06:20:36 +08:00
|
|
|
*
|
|
|
|
* Module allocation method suggested by Andi Kleen.
|
|
|
|
*/
|
|
|
|
#include <linux/module.h>
|
2005-10-28 21:29:43 +08:00
|
|
|
#include <linux/moduleloader.h>
|
2005-04-17 06:20:36 +08:00
|
|
|
#include <linux/kernel.h>
|
2008-07-24 12:28:13 +08:00
|
|
|
#include <linux/mm.h>
|
2005-04-17 06:20:36 +08:00
|
|
|
#include <linux/elf.h>
|
|
|
|
#include <linux/vmalloc.h>
|
|
|
|
#include <linux/fs.h>
|
|
|
|
#include <linux/string.h>
|
include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit slab.h inclusion from percpu.h
percpu.h is included by sched.h and module.h and thus ends up being
included when building most .c files. percpu.h includes slab.h which
in turn includes gfp.h making everything defined by the two files
universally available and complicating inclusion dependencies.
percpu.h -> slab.h dependency is about to be removed. Prepare for
this change by updating users of gfp and slab facilities include those
headers directly instead of assuming availability. As this conversion
needs to touch large number of source files, the following script is
used as the basis of conversion.
http://userweb.kernel.org/~tj/misc/slabh-sweep.py
The script does the followings.
* Scan files for gfp and slab usages and update includes such that
only the necessary includes are there. ie. if only gfp is used,
gfp.h, if slab is used, slab.h.
* When the script inserts a new include, it looks at the include
blocks and try to put the new include such that its order conforms
to its surrounding. It's put in the include block which contains
core kernel includes, in the same order that the rest are ordered -
alphabetical, Christmas tree, rev-Xmas-tree or at the end if there
doesn't seem to be any matching order.
* If the script can't find a place to put a new include (mostly
because the file doesn't have fitting include block), it prints out
an error message indicating which .h file needs to be added to the
file.
The conversion was done in the following steps.
1. The initial automatic conversion of all .c files updated slightly
over 4000 files, deleting around 700 includes and adding ~480 gfp.h
and ~3000 slab.h inclusions. The script emitted errors for ~400
files.
2. Each error was manually checked. Some didn't need the inclusion,
some needed manual addition while adding it to implementation .h or
embedding .c file was more appropriate for others. This step added
inclusions to around 150 files.
3. The script was run again and the output was compared to the edits
from #2 to make sure no file was left behind.
4. Several build tests were done and a couple of problems were fixed.
e.g. lib/decompress_*.c used malloc/free() wrappers around slab
APIs requiring slab.h to be added manually.
5. The script was run on all .h files but without automatically
editing them as sprinkling gfp.h and slab.h inclusions around .h
files could easily lead to inclusion dependency hell. Most gfp.h
inclusion directives were ignored as stuff from gfp.h was usually
wildly available and often used in preprocessor macros. Each
slab.h inclusion directive was examined and added manually as
necessary.
6. percpu.h was updated not to include slab.h.
7. Build test were done on the following configurations and failures
were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my
distributed build env didn't work with gcov compiles) and a few
more options had to be turned off depending on archs to make things
build (like ipr on powerpc/64 which failed due to missing writeq).
* x86 and x86_64 UP and SMP allmodconfig and a custom test config.
* powerpc and powerpc64 SMP allmodconfig
* sparc and sparc64 SMP allmodconfig
* ia64 SMP allmodconfig
* s390 SMP allmodconfig
* alpha SMP allmodconfig
* um on x86_64 SMP allmodconfig
8. percpu.h modifications were reverted so that it could be applied as
a separate patch and serve as bisection point.
Given the fact that I had only a couple of failures from tests on step
6, I'm fairly confident about the coverage of this conversion patch.
If there is a breakage, it's likely to be something in one of the arch
headers which should be easily discoverable easily on most builds of
the specific arch.
Signed-off-by: Tejun Heo <tj@kernel.org>
Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
2010-03-24 16:04:11 +08:00
|
|
|
#include <linux/gfp.h>
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2008-12-01 19:53:07 +08:00
|
|
|
#include <asm/sections.h>
|
2011-02-10 23:25:18 +08:00
|
|
|
#include <asm/smp_plat.h>
|
2009-02-11 20:09:54 +08:00
|
|
|
#include <asm/unwind.h>
|
2013-07-20 01:27:23 +08:00
|
|
|
#include <asm/opcodes.h>
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
#ifdef CONFIG_XIP_KERNEL
|
|
|
|
/*
|
|
|
|
* The XIP kernel text is mapped in the module area for modules and
|
|
|
|
* some other stuff to work without any indirect relocations.
|
2008-11-07 01:11:07 +08:00
|
|
|
* MODULES_VADDR is redefined here and not in asm/memory.h to avoid
|
2005-04-17 06:20:36 +08:00
|
|
|
* recompiling the whole kernel when CONFIG_XIP_KERNEL is turned on/off.
|
|
|
|
*/
|
2008-11-07 01:11:07 +08:00
|
|
|
#undef MODULES_VADDR
|
2016-02-10 02:34:43 +08:00
|
|
|
#define MODULES_VADDR (((unsigned long)_exiprom + ~PMD_MASK) & PMD_MASK)
|
2005-04-17 06:20:36 +08:00
|
|
|
#endif
|
|
|
|
|
2006-09-26 16:37:07 +08:00
|
|
|
#ifdef CONFIG_MMU
|
2005-04-17 06:20:36 +08:00
|
|
|
void *module_alloc(unsigned long size)
|
|
|
|
{
|
2017-04-28 02:19:01 +08:00
|
|
|
gfp_t gfp_mask = GFP_KERNEL;
|
|
|
|
void *p;
|
|
|
|
|
|
|
|
/* Silence the initial allocation */
|
|
|
|
if (IS_ENABLED(CONFIG_ARM_MODULE_PLTS))
|
|
|
|
gfp_mask |= __GFP_NOWARN;
|
|
|
|
|
|
|
|
p = __vmalloc_node_range(size, 1, MODULES_VADDR, MODULES_END,
|
|
|
|
gfp_mask, PAGE_KERNEL_EXEC, 0, NUMA_NO_NODE,
|
2014-11-24 23:54:35 +08:00
|
|
|
__builtin_return_address(0));
|
|
|
|
if (!IS_ENABLED(CONFIG_ARM_MODULE_PLTS) || p)
|
|
|
|
return p;
|
|
|
|
return __vmalloc_node_range(size, 1, VMALLOC_START, VMALLOC_END,
|
2015-02-14 06:40:07 +08:00
|
|
|
GFP_KERNEL, PAGE_KERNEL_EXEC, 0, NUMA_NO_NODE,
|
2011-01-14 07:46:02 +08:00
|
|
|
__builtin_return_address(0));
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2011-07-01 03:22:12 +08:00
|
|
|
#endif
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2020-05-14 18:36:42 +08:00
|
|
|
bool module_init_section(const char *name)
|
|
|
|
{
|
|
|
|
return strstarts(name, ".init") ||
|
|
|
|
strstarts(name, ".ARM.extab.init") ||
|
|
|
|
strstarts(name, ".ARM.exidx.init");
|
|
|
|
}
|
|
|
|
|
2019-06-07 18:49:12 +08:00
|
|
|
bool module_exit_section(const char *name)
|
|
|
|
{
|
|
|
|
return strstarts(name, ".exit") ||
|
|
|
|
strstarts(name, ".ARM.extab.exit") ||
|
|
|
|
strstarts(name, ".ARM.exidx.exit");
|
|
|
|
}
|
|
|
|
|
2021-11-25 01:56:22 +08:00
|
|
|
/*
|
|
|
|
* This implements the partitioning algorithm for group relocations as
|
|
|
|
* documented in the ARM AArch32 ELF psABI (IHI 0044).
|
|
|
|
*
|
|
|
|
* A single PC-relative symbol reference is divided in up to 3 add or subtract
|
|
|
|
* operations, where the final one could be incorporated into a load/store
|
|
|
|
* instruction with immediate offset. E.g.,
|
|
|
|
*
|
|
|
|
* ADD Rd, PC, #... or ADD Rd, PC, #...
|
|
|
|
* ADD Rd, Rd, #... ADD Rd, Rd, #...
|
|
|
|
* LDR Rd, [Rd, #...] ADD Rd, Rd, #...
|
|
|
|
*
|
|
|
|
* The latter has a guaranteed range of only 16 MiB (3x8 == 24 bits), so it is
|
|
|
|
* of limited use in the kernel. However, the ADD/ADD/LDR combo has a range of
|
|
|
|
* -/+ 256 MiB, (2x8 + 12 == 28 bits), which means it has sufficient range for
|
|
|
|
* any in-kernel symbol reference (unless module PLTs are being used).
|
|
|
|
*
|
|
|
|
* The main advantage of this approach over the typical pattern using a literal
|
|
|
|
* load is that literal loads may miss in the D-cache, and generally lead to
|
|
|
|
* lower cache efficiency for variables that are referenced often from many
|
|
|
|
* different places in the code.
|
|
|
|
*/
|
|
|
|
static u32 get_group_rem(u32 group, u32 *offset)
|
|
|
|
{
|
|
|
|
u32 val = *offset;
|
|
|
|
u32 shift;
|
|
|
|
do {
|
|
|
|
shift = val ? (31 - __fls(val)) & ~1 : 32;
|
|
|
|
*offset = val;
|
|
|
|
if (!val)
|
|
|
|
break;
|
|
|
|
val &= 0xffffff >> shift;
|
|
|
|
} while (group--);
|
|
|
|
return shift;
|
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
int
|
|
|
|
apply_relocate(Elf32_Shdr *sechdrs, const char *strtab, unsigned int symindex,
|
|
|
|
unsigned int relindex, struct module *module)
|
|
|
|
{
|
|
|
|
Elf32_Shdr *symsec = sechdrs + symindex;
|
|
|
|
Elf32_Shdr *relsec = sechdrs + relindex;
|
|
|
|
Elf32_Shdr *dstsec = sechdrs + relsec->sh_info;
|
|
|
|
Elf32_Rel *rel = (void *)relsec->sh_addr;
|
|
|
|
unsigned int i;
|
|
|
|
|
|
|
|
for (i = 0; i < relsec->sh_size / sizeof(Elf32_Rel); i++, rel++) {
|
|
|
|
unsigned long loc;
|
|
|
|
Elf32_Sym *sym;
|
2011-01-12 03:57:14 +08:00
|
|
|
const char *symname;
|
2021-11-25 01:56:22 +08:00
|
|
|
u32 shift, group = 1;
|
2005-04-17 06:20:36 +08:00
|
|
|
s32 offset;
|
2013-07-20 01:27:23 +08:00
|
|
|
u32 tmp;
|
2010-06-21 22:11:38 +08:00
|
|
|
#ifdef CONFIG_THUMB2_KERNEL
|
2009-07-24 19:32:59 +08:00
|
|
|
u32 upper, lower, sign, j1, j2;
|
2010-06-21 22:11:38 +08:00
|
|
|
#endif
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
offset = ELF32_R_SYM(rel->r_info);
|
|
|
|
if (offset < 0 || offset > (symsec->sh_size / sizeof(Elf32_Sym))) {
|
2011-01-12 03:57:14 +08:00
|
|
|
pr_err("%s: section %u reloc %u: bad relocation sym offset\n",
|
2005-04-17 06:20:36 +08:00
|
|
|
module->name, relindex, i);
|
|
|
|
return -ENOEXEC;
|
|
|
|
}
|
|
|
|
|
|
|
|
sym = ((Elf32_Sym *)symsec->sh_addr) + offset;
|
2011-01-12 03:57:14 +08:00
|
|
|
symname = strtab + sym->st_name;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
if (rel->r_offset < 0 || rel->r_offset > dstsec->sh_size - sizeof(u32)) {
|
2011-01-12 03:57:14 +08:00
|
|
|
pr_err("%s: section %u reloc %u sym '%s': out of bounds relocation, offset %d size %u\n",
|
|
|
|
module->name, relindex, i, symname,
|
|
|
|
rel->r_offset, dstsec->sh_size);
|
2005-04-17 06:20:36 +08:00
|
|
|
return -ENOEXEC;
|
|
|
|
}
|
|
|
|
|
|
|
|
loc = dstsec->sh_addr + rel->r_offset;
|
|
|
|
|
|
|
|
switch (ELF32_R_TYPE(rel->r_info)) {
|
2009-02-11 20:09:54 +08:00
|
|
|
case R_ARM_NONE:
|
|
|
|
/* ignore */
|
|
|
|
break;
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
case R_ARM_ABS32:
|
2014-08-08 21:12:17 +08:00
|
|
|
case R_ARM_TARGET1:
|
2005-04-17 06:20:36 +08:00
|
|
|
*(u32 *)loc += sym->st_value;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case R_ARM_PC24:
|
2005-12-15 06:04:22 +08:00
|
|
|
case R_ARM_CALL:
|
|
|
|
case R_ARM_JUMP24:
|
2014-11-24 23:45:12 +08:00
|
|
|
if (sym->st_value & 3) {
|
|
|
|
pr_err("%s: section %u reloc %u sym '%s': unsupported interworking call (ARM -> Thumb)\n",
|
|
|
|
module->name, relindex, i, symname);
|
|
|
|
return -ENOEXEC;
|
|
|
|
}
|
|
|
|
|
2013-07-20 01:27:23 +08:00
|
|
|
offset = __mem_to_opcode_arm(*(u32 *)loc);
|
|
|
|
offset = (offset & 0x00ffffff) << 2;
|
2005-04-17 06:20:36 +08:00
|
|
|
if (offset & 0x02000000)
|
|
|
|
offset -= 0x04000000;
|
|
|
|
|
|
|
|
offset += sym->st_value - loc;
|
2014-11-24 23:54:35 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Route through a PLT entry if 'offset' exceeds the
|
|
|
|
* supported range. Note that 'offset + loc + 8'
|
|
|
|
* contains the absolute jump target, i.e.,
|
|
|
|
* @sym + addend, corrected for the +8 PC bias.
|
|
|
|
*/
|
|
|
|
if (IS_ENABLED(CONFIG_ARM_MODULE_PLTS) &&
|
|
|
|
(offset <= (s32)0xfe000000 ||
|
|
|
|
offset >= (s32)0x02000000))
|
|
|
|
offset = get_module_plt(module, loc,
|
|
|
|
offset + loc + 8)
|
|
|
|
- loc - 8;
|
|
|
|
|
2014-11-24 23:45:12 +08:00
|
|
|
if (offset <= (s32)0xfe000000 ||
|
2007-05-09 05:05:25 +08:00
|
|
|
offset >= (s32)0x02000000) {
|
2011-01-12 03:57:14 +08:00
|
|
|
pr_err("%s: section %u reloc %u sym '%s': relocation %u out of range (%#lx -> %#x)\n",
|
|
|
|
module->name, relindex, i, symname,
|
|
|
|
ELF32_R_TYPE(rel->r_info), loc,
|
|
|
|
sym->st_value);
|
2005-04-17 06:20:36 +08:00
|
|
|
return -ENOEXEC;
|
|
|
|
}
|
|
|
|
|
|
|
|
offset >>= 2;
|
2013-07-20 01:27:23 +08:00
|
|
|
offset &= 0x00ffffff;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2013-07-20 01:27:23 +08:00
|
|
|
*(u32 *)loc &= __opcode_to_mem_arm(0xff000000);
|
|
|
|
*(u32 *)loc |= __opcode_to_mem_arm(offset);
|
2005-04-17 06:20:36 +08:00
|
|
|
break;
|
|
|
|
|
2009-03-20 18:11:43 +08:00
|
|
|
case R_ARM_V4BX:
|
|
|
|
/* Preserve Rm and the condition code. Alter
|
|
|
|
* other bits to re-code instruction as
|
|
|
|
* MOV PC,Rm.
|
|
|
|
*/
|
2013-07-20 01:27:23 +08:00
|
|
|
*(u32 *)loc &= __opcode_to_mem_arm(0xf000000f);
|
|
|
|
*(u32 *)loc |= __opcode_to_mem_arm(0x01a0f000);
|
2009-03-20 18:11:43 +08:00
|
|
|
break;
|
|
|
|
|
2009-02-11 20:09:54 +08:00
|
|
|
case R_ARM_PREL31:
|
2017-01-31 01:29:28 +08:00
|
|
|
offset = (*(s32 *)loc << 1) >> 1; /* sign extend */
|
|
|
|
offset += sym->st_value - loc;
|
|
|
|
if (offset >= 0x40000000 || offset < -0x40000000) {
|
|
|
|
pr_err("%s: section %u reloc %u sym '%s': relocation %u out of range (%#lx -> %#x)\n",
|
|
|
|
module->name, relindex, i, symname,
|
|
|
|
ELF32_R_TYPE(rel->r_info), loc,
|
|
|
|
sym->st_value);
|
|
|
|
return -ENOEXEC;
|
|
|
|
}
|
|
|
|
*(u32 *)loc &= 0x80000000;
|
|
|
|
*(u32 *)loc |= offset & 0x7fffffff;
|
2009-02-11 20:09:54 +08:00
|
|
|
break;
|
|
|
|
|
2020-09-14 16:24:37 +08:00
|
|
|
case R_ARM_REL32:
|
|
|
|
*(u32 *)loc += sym->st_value - loc;
|
|
|
|
break;
|
|
|
|
|
2009-05-07 23:18:40 +08:00
|
|
|
case R_ARM_MOVW_ABS_NC:
|
|
|
|
case R_ARM_MOVT_ABS:
|
2020-09-14 16:24:37 +08:00
|
|
|
case R_ARM_MOVW_PREL_NC:
|
|
|
|
case R_ARM_MOVT_PREL:
|
2013-07-20 01:27:23 +08:00
|
|
|
offset = tmp = __mem_to_opcode_arm(*(u32 *)loc);
|
2009-05-07 23:18:40 +08:00
|
|
|
offset = ((offset & 0xf0000) >> 4) | (offset & 0xfff);
|
|
|
|
offset = (offset ^ 0x8000) - 0x8000;
|
|
|
|
|
|
|
|
offset += sym->st_value;
|
2020-09-14 16:24:37 +08:00
|
|
|
if (ELF32_R_TYPE(rel->r_info) == R_ARM_MOVT_PREL ||
|
|
|
|
ELF32_R_TYPE(rel->r_info) == R_ARM_MOVW_PREL_NC)
|
|
|
|
offset -= loc;
|
|
|
|
if (ELF32_R_TYPE(rel->r_info) == R_ARM_MOVT_ABS ||
|
|
|
|
ELF32_R_TYPE(rel->r_info) == R_ARM_MOVT_PREL)
|
2009-05-07 23:18:40 +08:00
|
|
|
offset >>= 16;
|
|
|
|
|
2013-07-20 01:27:23 +08:00
|
|
|
tmp &= 0xfff0f000;
|
|
|
|
tmp |= ((offset & 0xf000) << 4) |
|
|
|
|
(offset & 0x0fff);
|
|
|
|
|
|
|
|
*(u32 *)loc = __opcode_to_mem_arm(tmp);
|
2009-05-07 23:18:40 +08:00
|
|
|
break;
|
|
|
|
|
2021-11-25 01:56:22 +08:00
|
|
|
case R_ARM_ALU_PC_G0_NC:
|
|
|
|
group = 0;
|
|
|
|
fallthrough;
|
|
|
|
case R_ARM_ALU_PC_G1_NC:
|
|
|
|
tmp = __mem_to_opcode_arm(*(u32 *)loc);
|
|
|
|
offset = ror32(tmp & 0xff, (tmp & 0xf00) >> 7);
|
|
|
|
if (tmp & BIT(22))
|
|
|
|
offset = -offset;
|
|
|
|
offset += sym->st_value - loc;
|
|
|
|
if (offset < 0) {
|
|
|
|
offset = -offset;
|
|
|
|
tmp = (tmp & ~BIT(23)) | BIT(22); // SUB opcode
|
|
|
|
} else {
|
|
|
|
tmp = (tmp & ~BIT(22)) | BIT(23); // ADD opcode
|
|
|
|
}
|
|
|
|
|
|
|
|
shift = get_group_rem(group, &offset);
|
|
|
|
if (shift < 24) {
|
|
|
|
offset >>= 24 - shift;
|
|
|
|
offset |= (shift + 8) << 7;
|
|
|
|
}
|
|
|
|
*(u32 *)loc = __opcode_to_mem_arm((tmp & ~0xfff) | offset);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case R_ARM_LDR_PC_G2:
|
|
|
|
tmp = __mem_to_opcode_arm(*(u32 *)loc);
|
|
|
|
offset = tmp & 0xfff;
|
|
|
|
if (~tmp & BIT(23)) // U bit cleared?
|
|
|
|
offset = -offset;
|
|
|
|
offset += sym->st_value - loc;
|
|
|
|
if (offset < 0) {
|
|
|
|
offset = -offset;
|
|
|
|
tmp &= ~BIT(23); // clear U bit
|
|
|
|
} else {
|
|
|
|
tmp |= BIT(23); // set U bit
|
|
|
|
}
|
|
|
|
get_group_rem(2, &offset);
|
|
|
|
|
|
|
|
if (offset > 0xfff) {
|
|
|
|
pr_err("%s: section %u reloc %u sym '%s': relocation %u out of range (%#lx -> %#x)\n",
|
|
|
|
module->name, relindex, i, symname,
|
|
|
|
ELF32_R_TYPE(rel->r_info), loc,
|
|
|
|
sym->st_value);
|
|
|
|
return -ENOEXEC;
|
|
|
|
}
|
|
|
|
*(u32 *)loc = __opcode_to_mem_arm((tmp & ~0xfff) | offset);
|
|
|
|
break;
|
|
|
|
|
2010-06-21 22:11:38 +08:00
|
|
|
#ifdef CONFIG_THUMB2_KERNEL
|
2009-07-24 19:32:59 +08:00
|
|
|
case R_ARM_THM_CALL:
|
|
|
|
case R_ARM_THM_JUMP24:
|
2014-11-24 23:45:12 +08:00
|
|
|
/*
|
|
|
|
* For function symbols, only Thumb addresses are
|
|
|
|
* allowed (no interworking).
|
|
|
|
*
|
|
|
|
* For non-function symbols, the destination
|
|
|
|
* has no specific ARM/Thumb disposition, so
|
|
|
|
* the branch is resolved under the assumption
|
|
|
|
* that interworking is not required.
|
|
|
|
*/
|
|
|
|
if (ELF32_ST_TYPE(sym->st_info) == STT_FUNC &&
|
|
|
|
!(sym->st_value & 1)) {
|
|
|
|
pr_err("%s: section %u reloc %u sym '%s': unsupported interworking call (Thumb -> ARM)\n",
|
|
|
|
module->name, relindex, i, symname);
|
|
|
|
return -ENOEXEC;
|
|
|
|
}
|
|
|
|
|
2013-07-20 01:27:23 +08:00
|
|
|
upper = __mem_to_opcode_thumb16(*(u16 *)loc);
|
|
|
|
lower = __mem_to_opcode_thumb16(*(u16 *)(loc + 2));
|
2009-07-24 19:32:59 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* 25 bit signed address range (Thumb-2 BL and B.W
|
|
|
|
* instructions):
|
|
|
|
* S:I1:I2:imm10:imm11:0
|
|
|
|
* where:
|
|
|
|
* S = upper[10] = offset[24]
|
|
|
|
* I1 = ~(J1 ^ S) = offset[23]
|
|
|
|
* I2 = ~(J2 ^ S) = offset[22]
|
|
|
|
* imm10 = upper[9:0] = offset[21:12]
|
|
|
|
* imm11 = lower[10:0] = offset[11:1]
|
|
|
|
* J1 = lower[13]
|
|
|
|
* J2 = lower[11]
|
|
|
|
*/
|
|
|
|
sign = (upper >> 10) & 1;
|
|
|
|
j1 = (lower >> 13) & 1;
|
|
|
|
j2 = (lower >> 11) & 1;
|
|
|
|
offset = (sign << 24) | ((~(j1 ^ sign) & 1) << 23) |
|
|
|
|
((~(j2 ^ sign) & 1) << 22) |
|
|
|
|
((upper & 0x03ff) << 12) |
|
|
|
|
((lower & 0x07ff) << 1);
|
|
|
|
if (offset & 0x01000000)
|
|
|
|
offset -= 0x02000000;
|
|
|
|
offset += sym->st_value - loc;
|
|
|
|
|
2014-11-24 23:54:35 +08:00
|
|
|
/*
|
|
|
|
* Route through a PLT entry if 'offset' exceeds the
|
|
|
|
* supported range.
|
|
|
|
*/
|
|
|
|
if (IS_ENABLED(CONFIG_ARM_MODULE_PLTS) &&
|
|
|
|
(offset <= (s32)0xff000000 ||
|
|
|
|
offset >= (s32)0x01000000))
|
|
|
|
offset = get_module_plt(module, loc,
|
|
|
|
offset + loc + 4)
|
|
|
|
- loc - 4;
|
|
|
|
|
2014-11-24 23:45:12 +08:00
|
|
|
if (offset <= (s32)0xff000000 ||
|
2009-07-24 19:32:59 +08:00
|
|
|
offset >= (s32)0x01000000) {
|
2011-01-12 03:57:14 +08:00
|
|
|
pr_err("%s: section %u reloc %u sym '%s': relocation %u out of range (%#lx -> %#x)\n",
|
|
|
|
module->name, relindex, i, symname,
|
|
|
|
ELF32_R_TYPE(rel->r_info), loc,
|
|
|
|
sym->st_value);
|
2009-07-24 19:32:59 +08:00
|
|
|
return -ENOEXEC;
|
|
|
|
}
|
|
|
|
|
|
|
|
sign = (offset >> 24) & 1;
|
|
|
|
j1 = sign ^ (~(offset >> 23) & 1);
|
|
|
|
j2 = sign ^ (~(offset >> 22) & 1);
|
2013-07-20 01:27:23 +08:00
|
|
|
upper = (u16)((upper & 0xf800) | (sign << 10) |
|
2009-07-24 19:32:59 +08:00
|
|
|
((offset >> 12) & 0x03ff));
|
2013-07-20 01:27:23 +08:00
|
|
|
lower = (u16)((lower & 0xd000) |
|
|
|
|
(j1 << 13) | (j2 << 11) |
|
|
|
|
((offset >> 1) & 0x07ff));
|
|
|
|
|
|
|
|
*(u16 *)loc = __opcode_to_mem_thumb16(upper);
|
|
|
|
*(u16 *)(loc + 2) = __opcode_to_mem_thumb16(lower);
|
2009-07-24 19:32:59 +08:00
|
|
|
break;
|
|
|
|
|
2010-06-21 22:10:37 +08:00
|
|
|
case R_ARM_THM_MOVW_ABS_NC:
|
|
|
|
case R_ARM_THM_MOVT_ABS:
|
2020-09-14 16:24:37 +08:00
|
|
|
case R_ARM_THM_MOVW_PREL_NC:
|
|
|
|
case R_ARM_THM_MOVT_PREL:
|
2013-07-20 01:27:23 +08:00
|
|
|
upper = __mem_to_opcode_thumb16(*(u16 *)loc);
|
|
|
|
lower = __mem_to_opcode_thumb16(*(u16 *)(loc + 2));
|
2010-06-21 22:10:37 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* MOVT/MOVW instructions encoding in Thumb-2:
|
|
|
|
*
|
|
|
|
* i = upper[10]
|
|
|
|
* imm4 = upper[3:0]
|
|
|
|
* imm3 = lower[14:12]
|
|
|
|
* imm8 = lower[7:0]
|
|
|
|
*
|
|
|
|
* imm16 = imm4:i:imm3:imm8
|
|
|
|
*/
|
|
|
|
offset = ((upper & 0x000f) << 12) |
|
|
|
|
((upper & 0x0400) << 1) |
|
|
|
|
((lower & 0x7000) >> 4) | (lower & 0x00ff);
|
|
|
|
offset = (offset ^ 0x8000) - 0x8000;
|
|
|
|
offset += sym->st_value;
|
|
|
|
|
2020-09-14 16:24:37 +08:00
|
|
|
if (ELF32_R_TYPE(rel->r_info) == R_ARM_THM_MOVT_PREL ||
|
|
|
|
ELF32_R_TYPE(rel->r_info) == R_ARM_THM_MOVW_PREL_NC)
|
|
|
|
offset -= loc;
|
|
|
|
if (ELF32_R_TYPE(rel->r_info) == R_ARM_THM_MOVT_ABS ||
|
|
|
|
ELF32_R_TYPE(rel->r_info) == R_ARM_THM_MOVT_PREL)
|
2010-06-21 22:10:37 +08:00
|
|
|
offset >>= 16;
|
|
|
|
|
2013-07-20 01:27:23 +08:00
|
|
|
upper = (u16)((upper & 0xfbf0) |
|
|
|
|
((offset & 0xf000) >> 12) |
|
|
|
|
((offset & 0x0800) >> 1));
|
|
|
|
lower = (u16)((lower & 0x8f00) |
|
|
|
|
((offset & 0x0700) << 4) |
|
|
|
|
(offset & 0x00ff));
|
|
|
|
*(u16 *)loc = __opcode_to_mem_thumb16(upper);
|
|
|
|
*(u16 *)(loc + 2) = __opcode_to_mem_thumb16(lower);
|
2010-06-21 22:10:37 +08:00
|
|
|
break;
|
2010-06-21 22:11:38 +08:00
|
|
|
#endif
|
2010-06-21 22:10:37 +08:00
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
default:
|
2014-10-28 19:26:42 +08:00
|
|
|
pr_err("%s: unknown relocation: %u\n",
|
2005-04-17 06:20:36 +08:00
|
|
|
module->name, ELF32_R_TYPE(rel->r_info));
|
|
|
|
return -ENOEXEC;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2010-11-12 21:02:46 +08:00
|
|
|
struct mod_unwind_map {
|
|
|
|
const Elf_Shdr *unw_sec;
|
|
|
|
const Elf_Shdr *txt_sec;
|
|
|
|
};
|
|
|
|
|
2011-02-10 23:25:18 +08:00
|
|
|
static const Elf_Shdr *find_mod_section(const Elf32_Ehdr *hdr,
|
|
|
|
const Elf_Shdr *sechdrs, const char *name)
|
|
|
|
{
|
|
|
|
const Elf_Shdr *s, *se;
|
|
|
|
const char *secstrs = (void *)hdr + sechdrs[hdr->e_shstrndx].sh_offset;
|
|
|
|
|
|
|
|
for (s = sechdrs, se = sechdrs + hdr->e_shnum; s < se; s++)
|
|
|
|
if (strcmp(name, secstrs + s->sh_name) == 0)
|
|
|
|
return s;
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
ARM: P2V: introduce phys_to_virt/virt_to_phys runtime patching
This idea came from Nicolas, Eric Miao produced an initial version,
which was then rewritten into this.
Patch the physical to virtual translations at runtime. As we modify
the code, this makes it incompatible with XIP kernels, but allows us
to achieve this with minimal loss of performance.
As many translations are of the form:
physical = virtual + (PHYS_OFFSET - PAGE_OFFSET)
virtual = physical - (PHYS_OFFSET - PAGE_OFFSET)
we generate an 'add' instruction for __virt_to_phys(), and a 'sub'
instruction for __phys_to_virt(). We calculate at run time (PHYS_OFFSET
- PAGE_OFFSET) by comparing the address prior to MMU initialization with
where it should be once the MMU has been initialized, and place this
constant into the above add/sub instructions.
Once we have (PHYS_OFFSET - PAGE_OFFSET), we can calculate the real
PHYS_OFFSET as PAGE_OFFSET is a build-time constant, and save this for
the C-mode PHYS_OFFSET variable definition to use.
At present, we are unable to support Realview with Sparsemem enabled
as this uses a complex mapping function, and MSM as this requires a
constant which will not fit in our math instruction.
Add a module version magic string for this feature to prevent
incompatible modules being loaded.
Tested-by: Tony Lindgren <tony@atomide.com>
Reviewed-by: Nicolas Pitre <nicolas.pitre@linaro.org>
Tested-by: Nicolas Pitre <nicolas.pitre@linaro.org>
Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
2011-01-05 03:09:43 +08:00
|
|
|
extern void fixup_pv_table(const void *, unsigned long);
|
2011-02-10 23:25:18 +08:00
|
|
|
extern void fixup_smp(const void *, unsigned long);
|
|
|
|
|
2010-11-12 21:02:46 +08:00
|
|
|
int module_finalize(const Elf32_Ehdr *hdr, const Elf_Shdr *sechdrs,
|
|
|
|
struct module *mod)
|
2009-02-11 20:09:54 +08:00
|
|
|
{
|
ARM: P2V: introduce phys_to_virt/virt_to_phys runtime patching
This idea came from Nicolas, Eric Miao produced an initial version,
which was then rewritten into this.
Patch the physical to virtual translations at runtime. As we modify
the code, this makes it incompatible with XIP kernels, but allows us
to achieve this with minimal loss of performance.
As many translations are of the form:
physical = virtual + (PHYS_OFFSET - PAGE_OFFSET)
virtual = physical - (PHYS_OFFSET - PAGE_OFFSET)
we generate an 'add' instruction for __virt_to_phys(), and a 'sub'
instruction for __phys_to_virt(). We calculate at run time (PHYS_OFFSET
- PAGE_OFFSET) by comparing the address prior to MMU initialization with
where it should be once the MMU has been initialized, and place this
constant into the above add/sub instructions.
Once we have (PHYS_OFFSET - PAGE_OFFSET), we can calculate the real
PHYS_OFFSET as PAGE_OFFSET is a build-time constant, and save this for
the C-mode PHYS_OFFSET variable definition to use.
At present, we are unable to support Realview with Sparsemem enabled
as this uses a complex mapping function, and MSM as this requires a
constant which will not fit in our math instruction.
Add a module version magic string for this feature to prevent
incompatible modules being loaded.
Tested-by: Tony Lindgren <tony@atomide.com>
Reviewed-by: Nicolas Pitre <nicolas.pitre@linaro.org>
Tested-by: Nicolas Pitre <nicolas.pitre@linaro.org>
Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
2011-01-05 03:09:43 +08:00
|
|
|
const Elf_Shdr *s = NULL;
|
2010-11-12 21:02:46 +08:00
|
|
|
#ifdef CONFIG_ARM_UNWIND
|
|
|
|
const char *secstrs = (void *)hdr + sechdrs[hdr->e_shstrndx].sh_offset;
|
2011-02-10 23:25:18 +08:00
|
|
|
const Elf_Shdr *sechdrs_end = sechdrs + hdr->e_shnum;
|
2010-11-12 21:02:46 +08:00
|
|
|
struct mod_unwind_map maps[ARM_SEC_MAX];
|
2010-08-19 22:16:37 +08:00
|
|
|
int i;
|
2010-11-12 21:02:46 +08:00
|
|
|
|
|
|
|
memset(maps, 0, sizeof(maps));
|
|
|
|
|
|
|
|
for (s = sechdrs; s < sechdrs_end; s++) {
|
|
|
|
const char *secname = secstrs + s->sh_name;
|
|
|
|
|
2010-11-12 21:04:16 +08:00
|
|
|
if (!(s->sh_flags & SHF_ALLOC))
|
|
|
|
continue;
|
|
|
|
|
2010-11-12 21:02:46 +08:00
|
|
|
if (strcmp(".ARM.exidx.init.text", secname) == 0)
|
|
|
|
maps[ARM_SEC_INIT].unw_sec = s;
|
|
|
|
else if (strcmp(".ARM.exidx", secname) == 0)
|
|
|
|
maps[ARM_SEC_CORE].unw_sec = s;
|
|
|
|
else if (strcmp(".ARM.exidx.exit.text", secname) == 0)
|
|
|
|
maps[ARM_SEC_EXIT].unw_sec = s;
|
2013-08-29 07:08:01 +08:00
|
|
|
else if (strcmp(".ARM.exidx.text.unlikely", secname) == 0)
|
|
|
|
maps[ARM_SEC_UNLIKELY].unw_sec = s;
|
|
|
|
else if (strcmp(".ARM.exidx.text.hot", secname) == 0)
|
|
|
|
maps[ARM_SEC_HOT].unw_sec = s;
|
2010-11-12 21:02:46 +08:00
|
|
|
else if (strcmp(".init.text", secname) == 0)
|
|
|
|
maps[ARM_SEC_INIT].txt_sec = s;
|
|
|
|
else if (strcmp(".text", secname) == 0)
|
|
|
|
maps[ARM_SEC_CORE].txt_sec = s;
|
|
|
|
else if (strcmp(".exit.text", secname) == 0)
|
|
|
|
maps[ARM_SEC_EXIT].txt_sec = s;
|
2013-08-29 07:08:01 +08:00
|
|
|
else if (strcmp(".text.unlikely", secname) == 0)
|
|
|
|
maps[ARM_SEC_UNLIKELY].txt_sec = s;
|
|
|
|
else if (strcmp(".text.hot", secname) == 0)
|
|
|
|
maps[ARM_SEC_HOT].txt_sec = s;
|
2010-08-19 22:16:37 +08:00
|
|
|
}
|
2009-02-11 20:09:54 +08:00
|
|
|
|
2010-11-12 21:02:46 +08:00
|
|
|
for (i = 0; i < ARM_SEC_MAX; i++)
|
|
|
|
if (maps[i].unw_sec && maps[i].txt_sec)
|
|
|
|
mod->arch.unwind[i] =
|
|
|
|
unwind_table_add(maps[i].unw_sec->sh_addr,
|
|
|
|
maps[i].unw_sec->sh_size,
|
|
|
|
maps[i].txt_sec->sh_addr,
|
|
|
|
maps[i].txt_sec->sh_size);
|
ARM: P2V: introduce phys_to_virt/virt_to_phys runtime patching
This idea came from Nicolas, Eric Miao produced an initial version,
which was then rewritten into this.
Patch the physical to virtual translations at runtime. As we modify
the code, this makes it incompatible with XIP kernels, but allows us
to achieve this with minimal loss of performance.
As many translations are of the form:
physical = virtual + (PHYS_OFFSET - PAGE_OFFSET)
virtual = physical - (PHYS_OFFSET - PAGE_OFFSET)
we generate an 'add' instruction for __virt_to_phys(), and a 'sub'
instruction for __phys_to_virt(). We calculate at run time (PHYS_OFFSET
- PAGE_OFFSET) by comparing the address prior to MMU initialization with
where it should be once the MMU has been initialized, and place this
constant into the above add/sub instructions.
Once we have (PHYS_OFFSET - PAGE_OFFSET), we can calculate the real
PHYS_OFFSET as PAGE_OFFSET is a build-time constant, and save this for
the C-mode PHYS_OFFSET variable definition to use.
At present, we are unable to support Realview with Sparsemem enabled
as this uses a complex mapping function, and MSM as this requires a
constant which will not fit in our math instruction.
Add a module version magic string for this feature to prevent
incompatible modules being loaded.
Tested-by: Tony Lindgren <tony@atomide.com>
Reviewed-by: Nicolas Pitre <nicolas.pitre@linaro.org>
Tested-by: Nicolas Pitre <nicolas.pitre@linaro.org>
Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
2011-01-05 03:09:43 +08:00
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_ARM_PATCH_PHYS_VIRT
|
|
|
|
s = find_mod_section(hdr, sechdrs, ".pv_table");
|
|
|
|
if (s)
|
|
|
|
fixup_pv_table((void *)s->sh_addr, s->sh_size);
|
2009-02-11 20:09:54 +08:00
|
|
|
#endif
|
2011-02-10 23:25:18 +08:00
|
|
|
s = find_mod_section(hdr, sechdrs, ".alt.smp.init");
|
|
|
|
if (s && !is_smp())
|
2011-08-04 07:01:00 +08:00
|
|
|
#ifdef CONFIG_SMP_ON_UP
|
2011-02-10 23:25:18 +08:00
|
|
|
fixup_smp((void *)s->sh_addr, s->sh_size);
|
2011-08-04 07:01:00 +08:00
|
|
|
#else
|
|
|
|
return -EINVAL;
|
|
|
|
#endif
|
2005-04-17 06:20:36 +08:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
module_arch_cleanup(struct module *mod)
|
|
|
|
{
|
2010-11-12 21:02:46 +08:00
|
|
|
#ifdef CONFIG_ARM_UNWIND
|
|
|
|
int i;
|
|
|
|
|
2020-05-14 18:36:42 +08:00
|
|
|
for (i = 0; i < ARM_SEC_MAX; i++) {
|
|
|
|
unwind_table_del(mod->arch.unwind[i]);
|
|
|
|
mod->arch.unwind[i] = NULL;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
void __weak module_arch_freeing_init(struct module *mod)
|
|
|
|
{
|
|
|
|
#ifdef CONFIG_ARM_UNWIND
|
|
|
|
unwind_table_del(mod->arch.unwind[ARM_SEC_INIT]);
|
|
|
|
mod->arch.unwind[ARM_SEC_INIT] = NULL;
|
2010-11-12 21:02:46 +08:00
|
|
|
#endif
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|