2009-06-08 02:46:37 +08:00
|
|
|
/*
|
|
|
|
* Helper macros to support writing architecture specific
|
|
|
|
* linker scripts.
|
|
|
|
*
|
|
|
|
* A minimal linker scripts has following content:
|
|
|
|
* [This is a sample, architectures may have special requiriements]
|
|
|
|
*
|
|
|
|
* OUTPUT_FORMAT(...)
|
|
|
|
* OUTPUT_ARCH(...)
|
|
|
|
* ENTRY(...)
|
|
|
|
* SECTIONS
|
|
|
|
* {
|
|
|
|
* . = START;
|
|
|
|
* __init_begin = .;
|
2009-06-15 04:10:41 +08:00
|
|
|
* HEAD_TEXT_SECTION
|
2009-06-08 02:46:37 +08:00
|
|
|
* INIT_TEXT_SECTION(PAGE_SIZE)
|
|
|
|
* INIT_DATA_SECTION(...)
|
2011-03-25 01:50:09 +08:00
|
|
|
* PERCPU_SECTION(CACHELINE_SIZE)
|
2009-06-08 02:46:37 +08:00
|
|
|
* __init_end = .;
|
|
|
|
*
|
|
|
|
* _stext = .;
|
|
|
|
* TEXT_SECTION = 0
|
|
|
|
* _etext = .;
|
|
|
|
*
|
|
|
|
* _sdata = .;
|
2019-10-30 05:13:34 +08:00
|
|
|
* RO_DATA(PAGE_SIZE)
|
2009-06-08 02:46:37 +08:00
|
|
|
* RW_DATA_SECTION(...)
|
|
|
|
* _edata = .;
|
|
|
|
*
|
|
|
|
* EXCEPTION_TABLE(...)
|
|
|
|
*
|
2009-07-13 06:23:33 +08:00
|
|
|
* BSS_SECTION(0, 0, 0)
|
2009-06-08 02:46:37 +08:00
|
|
|
* _end = .;
|
|
|
|
*
|
|
|
|
* STABS_DEBUG
|
|
|
|
* DWARF_DEBUG
|
linker script: unify usage of discard definition
Discarded sections in different archs share some commonality but have
considerable differences. This led to linker script for each arch
implementing its own /DISCARD/ definition, which makes maintaining
tedious and adding new entries error-prone.
This patch makes all linker scripts to move discard definitions to the
end of the linker script and use the common DISCARDS macro. As ld
uses the first matching section definition, archs can include default
discarded sections by including them earlier in the linker script.
ia64 is notable because it first throws away some ia64 specific
subsections and then include the rest of the sections into the final
image, so those sections must be discarded before the inclusion.
defconfig compile tested for x86, x86-64, powerpc, powerpc64, ia64,
alpha, sparc, sparc64 and s390. Michal Simek tested microblaze.
Signed-off-by: Tejun Heo <tj@kernel.org>
Acked-by: Paul Mundt <lethal@linux-sh.org>
Acked-by: Mike Frysinger <vapier@gentoo.org>
Tested-by: Michal Simek <monstr@monstr.eu>
Cc: linux-arch@vger.kernel.org
Cc: Michal Simek <monstr@monstr.eu>
Cc: microblaze-uclinux@itee.uq.edu.au
Cc: Sam Ravnborg <sam@ravnborg.org>
Cc: Tony Luck <tony.luck@intel.com>
2009-07-09 10:27:40 +08:00
|
|
|
*
|
|
|
|
* DISCARDS // must be the last
|
2009-06-08 02:46:37 +08:00
|
|
|
* }
|
|
|
|
*
|
|
|
|
* [__init_begin, __init_end] is the init section that may be freed after init
|
2014-09-26 10:30:59 +08:00
|
|
|
* // __init_begin and __init_end should be page aligned, so that we can
|
|
|
|
* // free the whole .init memory
|
2009-06-08 02:46:37 +08:00
|
|
|
* [_stext, _etext] is the text section
|
|
|
|
* [_sdata, _edata] is the data section
|
|
|
|
*
|
|
|
|
* Some of the included output section have their own set of constants.
|
|
|
|
* Examples are: [__initramfs_start, __initramfs_end] for initramfs and
|
|
|
|
* [__nosave_begin, __nosave_end] for the nosave data
|
|
|
|
*/
|
2009-04-26 10:10:56 +08:00
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
#ifndef LOAD_OFFSET
|
|
|
|
#define LOAD_OFFSET 0
|
|
|
|
#endif
|
|
|
|
|
2019-10-30 05:13:30 +08:00
|
|
|
/*
|
|
|
|
* Only some architectures want to have the .notes segment visible in
|
2019-10-30 05:13:31 +08:00
|
|
|
* a separate PT_NOTE ELF Program Header. When this happens, it needs
|
|
|
|
* to be visible in both the kernel text's PT_LOAD and the PT_NOTE
|
|
|
|
* Program Headers. In this case, though, the PT_LOAD needs to be made
|
|
|
|
* the default again so that all the following sections don't also end
|
|
|
|
* up in the PT_NOTE Program Header.
|
2019-10-30 05:13:30 +08:00
|
|
|
*/
|
|
|
|
#ifdef EMITS_PT_NOTE
|
|
|
|
#define NOTES_HEADERS :text :note
|
2019-10-30 05:13:31 +08:00
|
|
|
#define NOTES_HEADERS_RESTORE __restore_ph : { *(.__restore_ph) } :text
|
|
|
|
#else
|
|
|
|
#define NOTES_HEADERS
|
|
|
|
#define NOTES_HEADERS_RESTORE
|
2019-10-30 05:13:30 +08:00
|
|
|
#endif
|
|
|
|
|
2005-07-15 04:15:44 +08:00
|
|
|
/* Align . to a 8 byte boundary equals to maximum function alignment. */
|
|
|
|
#define ALIGN_FUNCTION() . = ALIGN(8)
|
|
|
|
|
2017-07-26 20:46:27 +08:00
|
|
|
/*
|
|
|
|
* LD_DEAD_CODE_DATA_ELIMINATION option enables -fdata-sections, which
|
|
|
|
* generates .data.identifier sections, which need to be pulled in with
|
|
|
|
* .data. We don't want to pull in .data..other sections, which Linux
|
|
|
|
* has defined. Same for text and bss.
|
2018-05-09 20:59:58 +08:00
|
|
|
*
|
|
|
|
* RODATA_MAIN is not used because existing code already defines .rodata.x
|
|
|
|
* sections to be brought in with rodata.
|
2017-07-26 20:46:27 +08:00
|
|
|
*/
|
|
|
|
#ifdef CONFIG_LD_DEAD_CODE_DATA_ELIMINATION
|
|
|
|
#define TEXT_MAIN .text .text.[0-9a-zA-Z_]*
|
2018-09-13 19:00:00 +08:00
|
|
|
#define DATA_MAIN .data .data.[0-9a-zA-Z_]* .data..LPBX*
|
2018-05-09 20:59:58 +08:00
|
|
|
#define SDATA_MAIN .sdata .sdata.[0-9a-zA-Z_]*
|
|
|
|
#define RODATA_MAIN .rodata .rodata.[0-9a-zA-Z_]*
|
2017-07-26 20:46:27 +08:00
|
|
|
#define BSS_MAIN .bss .bss.[0-9a-zA-Z_]*
|
2018-05-09 20:59:58 +08:00
|
|
|
#define SBSS_MAIN .sbss .sbss.[0-9a-zA-Z_]*
|
2017-07-26 20:46:27 +08:00
|
|
|
#else
|
|
|
|
#define TEXT_MAIN .text
|
|
|
|
#define DATA_MAIN .data
|
2018-05-09 20:59:58 +08:00
|
|
|
#define SDATA_MAIN .sdata
|
|
|
|
#define RODATA_MAIN .rodata
|
2017-07-26 20:46:27 +08:00
|
|
|
#define BSS_MAIN .bss
|
2018-05-09 20:59:58 +08:00
|
|
|
#define SBSS_MAIN .sbss
|
2017-07-26 20:46:27 +08:00
|
|
|
#endif
|
|
|
|
|
2010-07-10 14:35:00 +08:00
|
|
|
/*
|
|
|
|
* Align to a 32 byte boundary equal to the
|
|
|
|
* alignment gcc 4.5 uses for a struct
|
|
|
|
*/
|
2010-12-23 03:57:26 +08:00
|
|
|
#define STRUCT_ALIGNMENT 32
|
|
|
|
#define STRUCT_ALIGN() . = ALIGN(STRUCT_ALIGNMENT)
|
2010-07-10 14:35:00 +08:00
|
|
|
|
2008-01-21 03:07:28 +08:00
|
|
|
/* The actual configuration determine if the init/exit sections
|
|
|
|
* are handled as text/data or they can be discarded (which
|
|
|
|
* often happens at runtime)
|
|
|
|
*/
|
|
|
|
#ifdef CONFIG_HOTPLUG_CPU
|
|
|
|
#define CPU_KEEP(sec) *(.cpu##sec)
|
|
|
|
#define CPU_DISCARD(sec)
|
|
|
|
#else
|
|
|
|
#define CPU_KEEP(sec)
|
|
|
|
#define CPU_DISCARD(sec) *(.cpu##sec)
|
|
|
|
#endif
|
|
|
|
|
2008-01-25 05:20:18 +08:00
|
|
|
#if defined(CONFIG_MEMORY_HOTPLUG)
|
2008-01-21 03:07:28 +08:00
|
|
|
#define MEM_KEEP(sec) *(.mem##sec)
|
|
|
|
#define MEM_DISCARD(sec)
|
|
|
|
#else
|
|
|
|
#define MEM_KEEP(sec)
|
|
|
|
#define MEM_DISCARD(sec) *(.mem##sec)
|
|
|
|
#endif
|
|
|
|
|
ftrace: create __mcount_loc section
This patch creates a section in the kernel called "__mcount_loc".
This will hold a list of pointers to the mcount relocation for
each call site of mcount.
For example:
objdump -dr init/main.o
[...]
Disassembly of section .text:
0000000000000000 <do_one_initcall>:
0: 55 push %rbp
[...]
000000000000017b <init_post>:
17b: 55 push %rbp
17c: 48 89 e5 mov %rsp,%rbp
17f: 53 push %rbx
180: 48 83 ec 08 sub $0x8,%rsp
184: e8 00 00 00 00 callq 189 <init_post+0xe>
185: R_X86_64_PC32 mcount+0xfffffffffffffffc
[...]
We will add a section to point to each function call.
.section __mcount_loc,"a",@progbits
[...]
.quad .text + 0x185
[...]
The offset to of the mcount call site in init_post is an offset from
the start of the section, and not the start of the function init_post.
The mcount relocation is at the call site 0x185 from the start of the
.text section.
.text + 0x185 == init_post + 0xa
We need a way to add this __mcount_loc section in a way that we do not
lose the relocations after final link. The .text section here will
be attached to all other .text sections after final link and the
offsets will be meaningless. We need to keep track of where these
.text sections are.
To do this, we use the start of the first function in the section.
do_one_initcall. We can make a tmp.s file with this function as a reference
to the start of the .text section.
.section __mcount_loc,"a",@progbits
[...]
.quad do_one_initcall + 0x185
[...]
Then we can compile the tmp.s into a tmp.o
gcc -c tmp.s -o tmp.o
And link it into back into main.o.
ld -r main.o tmp.o -o tmp_main.o
mv tmp_main.o main.o
But we have a problem. What happens if the first function in a section
is not exported, and is a static function. The linker will not let
the tmp.o use it. This case exists in main.o as well.
Disassembly of section .init.text:
0000000000000000 <set_reset_devices>:
0: 55 push %rbp
1: 48 89 e5 mov %rsp,%rbp
4: e8 00 00 00 00 callq 9 <set_reset_devices+0x9>
5: R_X86_64_PC32 mcount+0xfffffffffffffffc
The first function in .init.text is a static function.
00000000000000a8 t __setup_set_reset_devices
000000000000105f t __setup_str_set_reset_devices
0000000000000000 t set_reset_devices
The lowercase 't' means that set_reset_devices is local and is not exported.
If we simply try to link the tmp.o with the set_reset_devices we end
up with two symbols: one local and one global.
.section __mcount_loc,"a",@progbits
.quad set_reset_devices + 0x10
00000000000000a8 t __setup_set_reset_devices
000000000000105f t __setup_str_set_reset_devices
0000000000000000 t set_reset_devices
U set_reset_devices
We still have an undefined reference to set_reset_devices, and if we try
to compile the kernel, we will end up with an undefined reference to
set_reset_devices, or even worst, it could be exported someplace else,
and then we will have a reference to the wrong location.
To handle this case, we make an intermediate step using objcopy.
We convert set_reset_devices into a global exported symbol before linking
it with tmp.o and set it back afterwards.
00000000000000a8 t __setup_set_reset_devices
000000000000105f t __setup_str_set_reset_devices
0000000000000000 T set_reset_devices
00000000000000a8 t __setup_set_reset_devices
000000000000105f t __setup_str_set_reset_devices
0000000000000000 T set_reset_devices
00000000000000a8 t __setup_set_reset_devices
000000000000105f t __setup_str_set_reset_devices
0000000000000000 t set_reset_devices
Now we have a section in main.o called __mcount_loc that we can place
somewhere in the kernel using vmlinux.ld.S and access it to convert
all these locations that call mcount into nops before starting SMP
and thus, eliminating the need to do this with kstop_machine.
Note, A well documented perl script (scripts/recordmcount.pl) is used
to do all this in one location.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-08-15 03:45:07 +08:00
|
|
|
#ifdef CONFIG_FTRACE_MCOUNT_RECORD
|
parisc: add dynamic ftrace
This patch implements dynamic ftrace for PA-RISC. The required mcount
call sequences can get pretty long, so instead of patching the
whole call sequence out of the functions, we are using
-fpatchable-function-entry from gcc. This puts a configurable amount of
NOPS before/at the start of the function. Taking do_sys_open() as example,
which would look like this when the call is patched out:
1036b248: 08 00 02 40 nop
1036b24c: 08 00 02 40 nop
1036b250: 08 00 02 40 nop
1036b254: 08 00 02 40 nop
1036b258 <do_sys_open>:
1036b258: 08 00 02 40 nop
1036b25c: 08 03 02 41 copy r3,r1
1036b260: 6b c2 3f d9 stw rp,-14(sp)
1036b264: 08 1e 02 43 copy sp,r3
1036b268: 6f c1 01 00 stw,ma r1,80(sp)
When ftrace gets enabled for this function the kernel will patch these
NOPs to:
1036b248: 10 19 57 20 <address of ftrace>
1036b24c: 6f c1 00 80 stw,ma r1,40(sp)
1036b250: 48 21 3f d1 ldw -18(r1),r1
1036b254: e8 20 c0 02 bv,n r0(r1)
1036b258 <do_sys_open>:
1036b258: e8 3f 1f df b,l,n .-c,r1
1036b25c: 08 03 02 41 copy r3,r1
1036b260: 6b c2 3f d9 stw rp,-14(sp)
1036b264: 08 1e 02 43 copy sp,r3
1036b268: 6f c1 01 00 stw,ma r1,80(sp)
So the first NOP in do_sys_open() will be patched to jump backwards into
some minimal trampoline code which pushes a stackframe, saves r1 which
holds the return address, loads the address of the real ftrace function,
and branches to that location. For 64 Bit things are getting a bit more
complicated (and longer) because we must make sure that the address of
ftrace location is 8 byte aligned, and the offset passed to ldd for
fetching the address is 8 byte aligned as well.
Note that gcc has a bug which misplaces the function label, and needs a
patch to make dynamic ftrace work. See
https://gcc.gnu.org/bugzilla/show_bug.cgi?id=90751 for details.
Signed-off-by: Sven Schnelle <svens@stackframe.org>
Signed-off-by: Helge Deller <deller@gmx.de>
2019-06-06 04:32:22 +08:00
|
|
|
#ifdef CC_USING_PATCHABLE_FUNCTION_ENTRY
|
|
|
|
#define MCOUNT_REC() . = ALIGN(8); \
|
|
|
|
__start_mcount_loc = .; \
|
|
|
|
KEEP(*(__patchable_function_entries)) \
|
|
|
|
__stop_mcount_loc = .;
|
|
|
|
#else
|
2009-07-28 02:23:50 +08:00
|
|
|
#define MCOUNT_REC() . = ALIGN(8); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start_mcount_loc = .; \
|
2018-05-09 20:59:58 +08:00
|
|
|
KEEP(*(__mcount_loc)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop_mcount_loc = .;
|
parisc: add dynamic ftrace
This patch implements dynamic ftrace for PA-RISC. The required mcount
call sequences can get pretty long, so instead of patching the
whole call sequence out of the functions, we are using
-fpatchable-function-entry from gcc. This puts a configurable amount of
NOPS before/at the start of the function. Taking do_sys_open() as example,
which would look like this when the call is patched out:
1036b248: 08 00 02 40 nop
1036b24c: 08 00 02 40 nop
1036b250: 08 00 02 40 nop
1036b254: 08 00 02 40 nop
1036b258 <do_sys_open>:
1036b258: 08 00 02 40 nop
1036b25c: 08 03 02 41 copy r3,r1
1036b260: 6b c2 3f d9 stw rp,-14(sp)
1036b264: 08 1e 02 43 copy sp,r3
1036b268: 6f c1 01 00 stw,ma r1,80(sp)
When ftrace gets enabled for this function the kernel will patch these
NOPs to:
1036b248: 10 19 57 20 <address of ftrace>
1036b24c: 6f c1 00 80 stw,ma r1,40(sp)
1036b250: 48 21 3f d1 ldw -18(r1),r1
1036b254: e8 20 c0 02 bv,n r0(r1)
1036b258 <do_sys_open>:
1036b258: e8 3f 1f df b,l,n .-c,r1
1036b25c: 08 03 02 41 copy r3,r1
1036b260: 6b c2 3f d9 stw rp,-14(sp)
1036b264: 08 1e 02 43 copy sp,r3
1036b268: 6f c1 01 00 stw,ma r1,80(sp)
So the first NOP in do_sys_open() will be patched to jump backwards into
some minimal trampoline code which pushes a stackframe, saves r1 which
holds the return address, loads the address of the real ftrace function,
and branches to that location. For 64 Bit things are getting a bit more
complicated (and longer) because we must make sure that the address of
ftrace location is 8 byte aligned, and the offset passed to ldd for
fetching the address is 8 byte aligned as well.
Note that gcc has a bug which misplaces the function label, and needs a
patch to make dynamic ftrace work. See
https://gcc.gnu.org/bugzilla/show_bug.cgi?id=90751 for details.
Signed-off-by: Sven Schnelle <svens@stackframe.org>
Signed-off-by: Helge Deller <deller@gmx.de>
2019-06-06 04:32:22 +08:00
|
|
|
#endif
|
ftrace: create __mcount_loc section
This patch creates a section in the kernel called "__mcount_loc".
This will hold a list of pointers to the mcount relocation for
each call site of mcount.
For example:
objdump -dr init/main.o
[...]
Disassembly of section .text:
0000000000000000 <do_one_initcall>:
0: 55 push %rbp
[...]
000000000000017b <init_post>:
17b: 55 push %rbp
17c: 48 89 e5 mov %rsp,%rbp
17f: 53 push %rbx
180: 48 83 ec 08 sub $0x8,%rsp
184: e8 00 00 00 00 callq 189 <init_post+0xe>
185: R_X86_64_PC32 mcount+0xfffffffffffffffc
[...]
We will add a section to point to each function call.
.section __mcount_loc,"a",@progbits
[...]
.quad .text + 0x185
[...]
The offset to of the mcount call site in init_post is an offset from
the start of the section, and not the start of the function init_post.
The mcount relocation is at the call site 0x185 from the start of the
.text section.
.text + 0x185 == init_post + 0xa
We need a way to add this __mcount_loc section in a way that we do not
lose the relocations after final link. The .text section here will
be attached to all other .text sections after final link and the
offsets will be meaningless. We need to keep track of where these
.text sections are.
To do this, we use the start of the first function in the section.
do_one_initcall. We can make a tmp.s file with this function as a reference
to the start of the .text section.
.section __mcount_loc,"a",@progbits
[...]
.quad do_one_initcall + 0x185
[...]
Then we can compile the tmp.s into a tmp.o
gcc -c tmp.s -o tmp.o
And link it into back into main.o.
ld -r main.o tmp.o -o tmp_main.o
mv tmp_main.o main.o
But we have a problem. What happens if the first function in a section
is not exported, and is a static function. The linker will not let
the tmp.o use it. This case exists in main.o as well.
Disassembly of section .init.text:
0000000000000000 <set_reset_devices>:
0: 55 push %rbp
1: 48 89 e5 mov %rsp,%rbp
4: e8 00 00 00 00 callq 9 <set_reset_devices+0x9>
5: R_X86_64_PC32 mcount+0xfffffffffffffffc
The first function in .init.text is a static function.
00000000000000a8 t __setup_set_reset_devices
000000000000105f t __setup_str_set_reset_devices
0000000000000000 t set_reset_devices
The lowercase 't' means that set_reset_devices is local and is not exported.
If we simply try to link the tmp.o with the set_reset_devices we end
up with two symbols: one local and one global.
.section __mcount_loc,"a",@progbits
.quad set_reset_devices + 0x10
00000000000000a8 t __setup_set_reset_devices
000000000000105f t __setup_str_set_reset_devices
0000000000000000 t set_reset_devices
U set_reset_devices
We still have an undefined reference to set_reset_devices, and if we try
to compile the kernel, we will end up with an undefined reference to
set_reset_devices, or even worst, it could be exported someplace else,
and then we will have a reference to the wrong location.
To handle this case, we make an intermediate step using objcopy.
We convert set_reset_devices into a global exported symbol before linking
it with tmp.o and set it back afterwards.
00000000000000a8 t __setup_set_reset_devices
000000000000105f t __setup_str_set_reset_devices
0000000000000000 T set_reset_devices
00000000000000a8 t __setup_set_reset_devices
000000000000105f t __setup_str_set_reset_devices
0000000000000000 T set_reset_devices
00000000000000a8 t __setup_set_reset_devices
000000000000105f t __setup_str_set_reset_devices
0000000000000000 t set_reset_devices
Now we have a section in main.o called __mcount_loc that we can place
somewhere in the kernel using vmlinux.ld.S and access it to convert
all these locations that call mcount into nops before starting SMP
and thus, eliminating the need to do this with kstop_machine.
Note, A well documented perl script (scripts/recordmcount.pl) is used
to do all this in one location.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-08-15 03:45:07 +08:00
|
|
|
#else
|
|
|
|
#define MCOUNT_REC()
|
|
|
|
#endif
|
2008-01-21 03:07:28 +08:00
|
|
|
|
2008-11-13 04:24:24 +08:00
|
|
|
#ifdef CONFIG_TRACE_BRANCH_PROFILING
|
2018-05-09 15:23:51 +08:00
|
|
|
#define LIKELY_PROFILE() __start_annotated_branch_profile = .; \
|
2018-05-09 20:59:58 +08:00
|
|
|
KEEP(*(_ftrace_annotated_branch)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop_annotated_branch_profile = .;
|
2008-11-12 13:14:39 +08:00
|
|
|
#else
|
|
|
|
#define LIKELY_PROFILE()
|
|
|
|
#endif
|
|
|
|
|
2008-11-21 14:30:54 +08:00
|
|
|
#ifdef CONFIG_PROFILE_ALL_BRANCHES
|
2018-05-09 15:23:51 +08:00
|
|
|
#define BRANCH_PROFILE() __start_branch_profile = .; \
|
2018-05-09 20:59:58 +08:00
|
|
|
KEEP(*(_ftrace_branch)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop_branch_profile = .;
|
2008-11-21 14:30:54 +08:00
|
|
|
#else
|
|
|
|
#define BRANCH_PROFILE()
|
|
|
|
#endif
|
|
|
|
|
2014-04-17 16:17:05 +08:00
|
|
|
#ifdef CONFIG_KPROBES
|
2014-05-01 13:26:44 +08:00
|
|
|
#define KPROBE_BLACKLIST() . = ALIGN(8); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start_kprobe_blacklist = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(_kprobe_blacklist)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop_kprobe_blacklist = .;
|
2014-04-17 16:17:05 +08:00
|
|
|
#else
|
|
|
|
#define KPROBE_BLACKLIST()
|
|
|
|
#endif
|
|
|
|
|
2018-01-13 01:55:03 +08:00
|
|
|
#ifdef CONFIG_FUNCTION_ERROR_INJECTION
|
2018-01-13 01:55:33 +08:00
|
|
|
#define ERROR_INJECT_WHITELIST() STRUCT_ALIGN(); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start_error_injection_whitelist = .; \
|
2018-01-13 01:55:03 +08:00
|
|
|
KEEP(*(_error_injection_whitelist)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop_error_injection_whitelist = .;
|
2017-12-12 00:36:46 +08:00
|
|
|
#else
|
2018-01-13 01:55:03 +08:00
|
|
|
#define ERROR_INJECT_WHITELIST()
|
2017-12-12 00:36:46 +08:00
|
|
|
#endif
|
|
|
|
|
2009-04-08 16:14:01 +08:00
|
|
|
#ifdef CONFIG_EVENT_TRACING
|
tracing: Replace trace_event struct array with pointer array
Currently the trace_event structures are placed in the _ftrace_events
section, and at link time, the linker makes one large array of all
the trace_event structures. On boot up, this array is read (much like
the initcall sections) and the events are processed.
The problem is that there is no guarantee that gcc will place complex
structures nicely together in an array format. Two structures in the
same file may be placed awkwardly, because gcc has no clue that they
are suppose to be in an array.
A hack was used previous to force the alignment to 4, to pack the
structures together. But this caused alignment issues with other
architectures (sparc).
Instead of packing the structures into an array, the structures' addresses
are now put into the _ftrace_event section. As pointers are always the
natural alignment, gcc should always pack them tightly together
(otherwise initcall, extable, etc would also fail).
By having the pointers to the structures in the section, we can still
iterate the trace_events without causing unnecessary alignment problems
with other architectures, or depending on the current behaviour of
gcc that will likely change in the future just to tick us kernel developers
off a little more.
The _ftrace_event section is also moved into the .init.data section
as it is now only needed at boot up.
Suggested-by: David Miller <davem@davemloft.net>
Cc: Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
Acked-by: David S. Miller <davem@davemloft.net>
Signed-off-by: Steven Rostedt <rostedt@goodmis.org>
2011-01-27 22:15:30 +08:00
|
|
|
#define FTRACE_EVENTS() . = ALIGN(8); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start_ftrace_events = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(_ftrace_events)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop_ftrace_events = .; \
|
|
|
|
__start_ftrace_eval_maps = .; \
|
2017-06-01 05:56:42 +08:00
|
|
|
KEEP(*(_ftrace_eval_map)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop_ftrace_eval_maps = .;
|
2009-02-24 23:21:36 +08:00
|
|
|
#else
|
|
|
|
#define FTRACE_EVENTS()
|
|
|
|
#endif
|
|
|
|
|
2009-03-07 00:21:48 +08:00
|
|
|
#ifdef CONFIG_TRACING
|
2018-05-09 15:23:51 +08:00
|
|
|
#define TRACE_PRINTKS() __start___trace_bprintk_fmt = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(__trace_printk_fmt)) /* Trace_printk fmt' pointer */ \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___trace_bprintk_fmt = .;
|
|
|
|
#define TRACEPOINT_STR() __start___tracepoint_str = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(__tracepoint_str)) /* Trace_printk fmt' pointer */ \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___tracepoint_str = .;
|
2009-03-07 00:21:48 +08:00
|
|
|
#else
|
|
|
|
#define TRACE_PRINTKS()
|
2013-07-13 05:07:27 +08:00
|
|
|
#define TRACEPOINT_STR()
|
2009-03-07 00:21:48 +08:00
|
|
|
#endif
|
|
|
|
|
2009-03-13 22:42:11 +08:00
|
|
|
#ifdef CONFIG_FTRACE_SYSCALLS
|
tracing: Replace syscall_meta_data struct array with pointer array
Currently the syscall_meta structures for the syscall tracepoints are
placed in the __syscall_metadata section, and at link time, the linker
makes one large array of all these syscall metadata structures. On boot
up, this array is read (much like the initcall sections) and the syscall
data is processed.
The problem is that there is no guarantee that gcc will place complex
structures nicely together in an array format. Two structures in the
same file may be placed awkwardly, because gcc has no clue that they
are suppose to be in an array.
A hack was used previous to force the alignment to 4, to pack the
structures together. But this caused alignment issues with other
architectures (sparc).
Instead of packing the structures into an array, the structures' addresses
are now put into the __syscall_metadata section. As pointers are always the
natural alignment, gcc should always pack them tightly together
(otherwise initcall, extable, etc would also fail).
By having the pointers to the structures in the section, we can still
iterate the trace_events without causing unnecessary alignment problems
with other architectures, or depending on the current behaviour of
gcc that will likely change in the future just to tick us kernel developers
off a little more.
The __syscall_metadata section is also moved into the .init.data section
as it is now only needed at boot up.
Suggested-by: David Miller <davem@davemloft.net>
Acked-by: David S. Miller <davem@davemloft.net>
Cc: Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
Signed-off-by: Steven Rostedt <rostedt@goodmis.org>
2011-02-03 06:06:09 +08:00
|
|
|
#define TRACE_SYSCALLS() . = ALIGN(8); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start_syscalls_metadata = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(__syscalls_metadata)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop_syscalls_metadata = .;
|
2009-03-13 22:42:11 +08:00
|
|
|
#else
|
|
|
|
#define TRACE_SYSCALLS()
|
|
|
|
#endif
|
|
|
|
|
2018-03-29 03:05:37 +08:00
|
|
|
#ifdef CONFIG_BPF_EVENTS
|
|
|
|
#define BPF_RAW_TP() STRUCT_ALIGN(); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start__bpf_raw_tp = .; \
|
2018-03-29 03:05:37 +08:00
|
|
|
KEEP(*(__bpf_raw_tp_map)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop__bpf_raw_tp = .;
|
2018-03-29 03:05:37 +08:00
|
|
|
#else
|
|
|
|
#define BPF_RAW_TP()
|
|
|
|
#endif
|
|
|
|
|
2015-03-10 04:27:21 +08:00
|
|
|
#ifdef CONFIG_SERIAL_EARLYCON
|
2018-04-07 07:21:53 +08:00
|
|
|
#define EARLYCON_TABLE() . = ALIGN(8); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__earlycon_table = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(__earlycon_table)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__earlycon_table_end = .;
|
2015-03-10 04:27:21 +08:00
|
|
|
#else
|
|
|
|
#define EARLYCON_TABLE()
|
|
|
|
#endif
|
2010-12-23 03:57:26 +08:00
|
|
|
|
2018-10-11 08:18:22 +08:00
|
|
|
#ifdef CONFIG_SECURITY
|
|
|
|
#define LSM_TABLE() . = ALIGN(8); \
|
|
|
|
__start_lsm_info = .; \
|
|
|
|
KEEP(*(.lsm_info.init)) \
|
|
|
|
__end_lsm_info = .;
|
2019-08-20 08:17:37 +08:00
|
|
|
#define EARLY_LSM_TABLE() . = ALIGN(8); \
|
|
|
|
__start_early_lsm_info = .; \
|
|
|
|
KEEP(*(.early_lsm_info.init)) \
|
|
|
|
__end_early_lsm_info = .;
|
2018-10-11 08:18:22 +08:00
|
|
|
#else
|
|
|
|
#define LSM_TABLE()
|
2019-08-20 08:17:37 +08:00
|
|
|
#define EARLY_LSM_TABLE()
|
2018-10-11 08:18:22 +08:00
|
|
|
#endif
|
|
|
|
|
2014-03-25 05:59:20 +08:00
|
|
|
#define ___OF_TABLE(cfg, name) _OF_TABLE_##cfg(name)
|
|
|
|
#define __OF_TABLE(cfg, name) ___OF_TABLE(cfg, name)
|
2016-06-14 13:58:58 +08:00
|
|
|
#define OF_TABLE(cfg, name) __OF_TABLE(IS_ENABLED(cfg), name)
|
2014-03-25 05:59:20 +08:00
|
|
|
#define _OF_TABLE_0(name)
|
|
|
|
#define _OF_TABLE_1(name) \
|
irqchip: add basic infrastructure
With the recent creation of the drivers/irqchip/ directory, it is
desirable to move irq controller drivers here. At the moment, the only
driver here is irq-bcm2835, the driver for the irq controller found in
the ARM BCM2835 SoC, present in Rasberry Pi systems. This irq
controller driver was exporting its initialization function and its
irq handling function through a header file in
<linux/irqchip/bcm2835.h>.
When proposing to also move another irq controller driver in
drivers/irqchip, Rob Herring raised the very valid point that moving
things to drivers/irqchip was good in order to remove more stuff from
arch/arm, but if it means adding gazillions of headers files in
include/linux/irqchip/, it would not be very nice.
So, upon the suggestion of Rob Herring and Arnd Bergmann, this commit
introduces a small infrastructure that defines a central
irqchip_init() function in drivers/irqchip/irqchip.c, which is meant
to be called as the ->init_irq() callback of ARM platforms. This
function calls of_irq_init() with an array of match strings and init
functions generated from a special linker section.
Note that the irq controller driver initialization function is
responsible for setting the global handle_arch_irq() variable, so that
ARM platforms no longer have to define the ->handle_irq field in their
DT_MACHINE structure.
A global header, <linux/irqchip.h> is also added to expose the single
irqchip_init() function to the reset of the kernel.
A further commit moves the BCM2835 irq controller driver to this new
small infrastructure, therefore removing the include/linux/irqchip/
directory.
Signed-off-by: Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
Reviewed-by: Stephen Warren <swarren@wwwdotorg.org>
Reviewed-by: Rob Herring <rob.herring@calxeda.com>
Acked-by: Arnd Bergmann <arnd@arndb.de>
[rob.herring: reword commit message to reflect use of linker sections.]
Signed-off-by: Rob Herring <rob.herring@calxeda.com>
2012-11-21 06:00:52 +08:00
|
|
|
. = ALIGN(8); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__##name##_of_table = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(__##name##_of_table)) \
|
|
|
|
KEEP(*(__##name##_of_table_end))
|
2014-03-25 05:59:20 +08:00
|
|
|
|
2017-05-27 01:34:11 +08:00
|
|
|
#define TIMER_OF_TABLES() OF_TABLE(CONFIG_TIMER_OF, timer)
|
2014-03-25 05:59:20 +08:00
|
|
|
#define IRQCHIP_OF_MATCH_TABLE() OF_TABLE(CONFIG_IRQCHIP, irqchip)
|
|
|
|
#define CLK_OF_TABLES() OF_TABLE(CONFIG_COMMON_CLK, clk)
|
|
|
|
#define RESERVEDMEM_OF_TABLES() OF_TABLE(CONFIG_OF_RESERVED_MEM, reservedmem)
|
|
|
|
#define CPU_METHOD_OF_TABLES() OF_TABLE(CONFIG_SMP, cpu_method)
|
2015-02-02 23:32:45 +08:00
|
|
|
#define CPUIDLE_METHOD_OF_TABLES() OF_TABLE(CONFIG_CPU_IDLE, cpuidle_method)
|
2013-10-31 09:21:09 +08:00
|
|
|
|
2015-09-28 22:49:12 +08:00
|
|
|
#ifdef CONFIG_ACPI
|
|
|
|
#define ACPI_PROBE_TABLE(name) \
|
|
|
|
. = ALIGN(8); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__##name##_acpi_probe_table = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(__##name##_acpi_probe_table)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__##name##_acpi_probe_table_end = .;
|
2015-09-28 22:49:12 +08:00
|
|
|
#else
|
|
|
|
#define ACPI_PROBE_TABLE(name)
|
|
|
|
#endif
|
|
|
|
|
2019-06-13 04:13:24 +08:00
|
|
|
#ifdef CONFIG_THERMAL
|
|
|
|
#define THERMAL_TABLE(name) \
|
|
|
|
. = ALIGN(8); \
|
|
|
|
__##name##_thermal_table = .; \
|
|
|
|
KEEP(*(__##name##_thermal_table)) \
|
|
|
|
__##name##_thermal_table_end = .;
|
|
|
|
#else
|
|
|
|
#define THERMAL_TABLE(name)
|
|
|
|
#endif
|
|
|
|
|
2010-12-23 03:57:26 +08:00
|
|
|
#define KERNEL_DTB() \
|
|
|
|
STRUCT_ALIGN(); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__dtb_start = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(.dtb.init.rodata)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__dtb_end = .;
|
2010-12-23 03:57:26 +08:00
|
|
|
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
/*
|
|
|
|
* .data section
|
|
|
|
*/
|
2007-05-17 19:38:44 +08:00
|
|
|
#define DATA_DATA \
|
mtd: only use __xipram annotation when XIP_KERNEL is set
When XIP_KERNEL is enabled, some functions are defined in the .data
ELF section because we require them to be in RAM whenever we communicate
with the flash chip. However this causes problems when FTRACE is
enabled and gcc emits calls to __gnu_mcount_nc in the function
prolog:
drivers/built-in.o: In function `cfi_chip_setup':
:(.data+0x272fc): relocation truncated to fit: R_ARM_CALL against symbol `__gnu_mcount_nc' defined in .text section in arch/arm/kernel/built-in.o
drivers/built-in.o: In function `cfi_probe_chip':
:(.data+0x27de8): relocation truncated to fit: R_ARM_CALL against symbol `__gnu_mcount_nc' defined in .text section in arch/arm/kernel/built-in.o
/tmp/ccY172rP.s: Assembler messages:
/tmp/ccY172rP.s:70: Warning: ignoring changed section attributes for .data
/tmp/ccY172rP.s: Error: 1 warning, treating warnings as errors
make[5]: *** [drivers/mtd/chips/cfi_probe.o] Error 1
/tmp/ccK4rjeO.s: Assembler messages:
/tmp/ccK4rjeO.s:421: Warning: ignoring changed section attributes for .data
/tmp/ccK4rjeO.s: Error: 1 warning, treating warnings as errors
make[5]: *** [drivers/mtd/chips/cfi_util.o] Error 1
/tmp/ccUvhCYR.s: Assembler messages:
/tmp/ccUvhCYR.s:1895: Warning: ignoring changed section attributes for .data
/tmp/ccUvhCYR.s: Error: 1 warning, treating warnings as errors
Specifically, this does not work because the .data section is not
marked executable, which leads LD to not generate trampolines for
long calls.
This moves the __xipram functions into their own .xiptext section instead.
The section is still placed next to .data and located in RAM but is marked
executable, which avoids the build errors.
Also, we only need to place the XIP functions into a separate section
if both CONFIG_XIP_KERNEL and CONFIG_MTD_XIP are set: When only MTD_XIP
is used, the whole kernel is still in RAM and we do not need to worry
about pulling out the rug under it. When only XIP_KERNEL but not MTD_XIP
is set, the kernel is in some form of ROM, but we never write to it.
Note that MTD_XIP has been broken on ARM since around 2011 or 2012. I
have sent another patch[2] to fix compilation, which I plan to merge
through arm-soc unless there are objections. The obvious alternative
to that would be to completely rip out the MTD_XIP support from the
kernel, since obviously nobody has been using it in a long while.
Link: [1] https://patchwork.kernel.org/patch/8109771/
Link: [2] https://patchwork.kernel.org/patch/9855225/
Signed-off-by: Arnd Bergmann <arnd@arndb.de>
Signed-off-by: Boris Brezillon <boris.brezillon@free-electrons.com>
2017-07-22 04:26:25 +08:00
|
|
|
*(.xiptext) \
|
2017-07-26 20:46:27 +08:00
|
|
|
*(DATA_MAIN) \
|
2008-01-29 03:21:15 +08:00
|
|
|
*(.ref.data) \
|
2010-10-27 05:22:29 +08:00
|
|
|
*(.data..shared_aligned) /* percpu related */ \
|
2018-05-09 20:59:58 +08:00
|
|
|
MEM_KEEP(init.data*) \
|
|
|
|
MEM_KEEP(exit.data*) \
|
2012-03-24 06:01:52 +08:00
|
|
|
*(.data.unlikely) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start_once = .; \
|
2017-11-18 07:27:03 +08:00
|
|
|
*(.data.once) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__end_once = .; \
|
2011-01-27 06:26:22 +08:00
|
|
|
STRUCT_ALIGN(); \
|
tracing: Kernel Tracepoints
Implementation of kernel tracepoints. Inspired from the Linux Kernel
Markers. Allows complete typing verification by declaring both tracing
statement inline functions and probe registration/unregistration static
inline functions within the same macro "DEFINE_TRACE". No format string
is required. See the tracepoint Documentation and Samples patches for
usage examples.
Taken from the documentation patch :
"A tracepoint placed in code provides a hook to call a function (probe)
that you can provide at runtime. A tracepoint can be "on" (a probe is
connected to it) or "off" (no probe is attached). When a tracepoint is
"off" it has no effect, except for adding a tiny time penalty (checking
a condition for a branch) and space penalty (adding a few bytes for the
function call at the end of the instrumented function and adds a data
structure in a separate section). When a tracepoint is "on", the
function you provide is called each time the tracepoint is executed, in
the execution context of the caller. When the function provided ends its
execution, it returns to the caller (continuing from the tracepoint
site).
You can put tracepoints at important locations in the code. They are
lightweight hooks that can pass an arbitrary number of parameters, which
prototypes are described in a tracepoint declaration placed in a header
file."
Addition and removal of tracepoints is synchronized by RCU using the
scheduler (and preempt_disable) as guarantees to find a quiescent state
(this is really RCU "classic"). The update side uses rcu_barrier_sched()
with call_rcu_sched() and the read/execute side uses
"preempt_disable()/preempt_enable()".
We make sure the previous array containing probes, which has been
scheduled for deletion by the rcu callback, is indeed freed before we
proceed to the next update. It therefore limits the rate of modification
of a single tracepoint to one update per RCU period. The objective here
is to permit fast batch add/removal of probes on _different_
tracepoints.
Changelog :
- Use #name ":" #proto as string to identify the tracepoint in the
tracepoint table. This will make sure not type mismatch happens due to
connexion of a probe with the wrong type to a tracepoint declared with
the same name in a different header.
- Add tracepoint_entry_free_old.
- Change __TO_TRACE to get rid of the 'i' iterator.
Masami Hiramatsu <mhiramat@redhat.com> :
Tested on x86-64.
Performance impact of a tracepoint : same as markers, except that it
adds about 70 bytes of instructions in an unlikely branch of each
instrumented function (the for loop, the stack setup and the function
call). It currently adds a memory read, a test and a conditional branch
at the instrumentation site (in the hot path). Immediate values will
eventually change this into a load immediate, test and branch, which
removes the memory read which will make the i-cache impact smaller
(changing the memory read for a load immediate removes 3-4 bytes per
site on x86_32 (depending on mov prefixes), or 7-8 bytes on x86_64, it
also saves the d-cache hit).
About the performance impact of tracepoints (which is comparable to
markers), even without immediate values optimizations, tests done by
Hideo Aoki on ia64 show no regression. His test case was using hackbench
on a kernel where scheduler instrumentation (about 5 events in code
scheduler code) was added.
Quoting Hideo Aoki about Markers :
I evaluated overhead of kernel marker using linux-2.6-sched-fixes git
tree, which includes several markers for LTTng, using an ia64 server.
While the immediate trace mark feature isn't implemented on ia64, there
is no major performance regression. So, I think that we don't have any
issues to propose merging marker point patches into Linus's tree from
the viewpoint of performance impact.
I prepared two kernels to evaluate. The first one was compiled without
CONFIG_MARKERS. The second one was enabled CONFIG_MARKERS.
I downloaded the original hackbench from the following URL:
http://devresources.linux-foundation.org/craiger/hackbench/src/hackbench.c
I ran hackbench 5 times in each condition and calculated the average and
difference between the kernels.
The parameter of hackbench: every 50 from 50 to 800
The number of CPUs of the server: 2, 4, and 8
Below is the results. As you can see, major performance regression
wasn't found in any case. Even if number of processes increases,
differences between marker-enabled kernel and marker- disabled kernel
doesn't increase. Moreover, if number of CPUs increases, the differences
doesn't increase either.
Curiously, marker-enabled kernel is better than marker-disabled kernel
in more than half cases, although I guess it comes from the difference
of memory access pattern.
* 2 CPUs
Number of | without | with | diff | diff |
processes | Marker [Sec] | Marker [Sec] | [Sec] | [%] |
--------------------------------------------------------------
50 | 4.811 | 4.872 | +0.061 | +1.27 |
100 | 9.854 | 10.309 | +0.454 | +4.61 |
150 | 15.602 | 15.040 | -0.562 | -3.6 |
200 | 20.489 | 20.380 | -0.109 | -0.53 |
250 | 25.798 | 25.652 | -0.146 | -0.56 |
300 | 31.260 | 30.797 | -0.463 | -1.48 |
350 | 36.121 | 35.770 | -0.351 | -0.97 |
400 | 42.288 | 42.102 | -0.186 | -0.44 |
450 | 47.778 | 47.253 | -0.526 | -1.1 |
500 | 51.953 | 52.278 | +0.325 | +0.63 |
550 | 58.401 | 57.700 | -0.701 | -1.2 |
600 | 63.334 | 63.222 | -0.112 | -0.18 |
650 | 68.816 | 68.511 | -0.306 | -0.44 |
700 | 74.667 | 74.088 | -0.579 | -0.78 |
750 | 78.612 | 79.582 | +0.970 | +1.23 |
800 | 85.431 | 85.263 | -0.168 | -0.2 |
--------------------------------------------------------------
* 4 CPUs
Number of | without | with | diff | diff |
processes | Marker [Sec] | Marker [Sec] | [Sec] | [%] |
--------------------------------------------------------------
50 | 2.586 | 2.584 | -0.003 | -0.1 |
100 | 5.254 | 5.283 | +0.030 | +0.56 |
150 | 8.012 | 8.074 | +0.061 | +0.76 |
200 | 11.172 | 11.000 | -0.172 | -1.54 |
250 | 13.917 | 14.036 | +0.119 | +0.86 |
300 | 16.905 | 16.543 | -0.362 | -2.14 |
350 | 19.901 | 20.036 | +0.135 | +0.68 |
400 | 22.908 | 23.094 | +0.186 | +0.81 |
450 | 26.273 | 26.101 | -0.172 | -0.66 |
500 | 29.554 | 29.092 | -0.461 | -1.56 |
550 | 32.377 | 32.274 | -0.103 | -0.32 |
600 | 35.855 | 35.322 | -0.533 | -1.49 |
650 | 39.192 | 38.388 | -0.804 | -2.05 |
700 | 41.744 | 41.719 | -0.025 | -0.06 |
750 | 45.016 | 44.496 | -0.520 | -1.16 |
800 | 48.212 | 47.603 | -0.609 | -1.26 |
--------------------------------------------------------------
* 8 CPUs
Number of | without | with | diff | diff |
processes | Marker [Sec] | Marker [Sec] | [Sec] | [%] |
--------------------------------------------------------------
50 | 2.094 | 2.072 | -0.022 | -1.07 |
100 | 4.162 | 4.273 | +0.111 | +2.66 |
150 | 6.485 | 6.540 | +0.055 | +0.84 |
200 | 8.556 | 8.478 | -0.078 | -0.91 |
250 | 10.458 | 10.258 | -0.200 | -1.91 |
300 | 12.425 | 12.750 | +0.325 | +2.62 |
350 | 14.807 | 14.839 | +0.032 | +0.22 |
400 | 16.801 | 16.959 | +0.158 | +0.94 |
450 | 19.478 | 19.009 | -0.470 | -2.41 |
500 | 21.296 | 21.504 | +0.208 | +0.98 |
550 | 23.842 | 23.979 | +0.137 | +0.57 |
600 | 26.309 | 26.111 | -0.198 | -0.75 |
650 | 28.705 | 28.446 | -0.259 | -0.9 |
700 | 31.233 | 31.394 | +0.161 | +0.52 |
750 | 34.064 | 33.720 | -0.344 | -1.01 |
800 | 36.320 | 36.114 | -0.206 | -0.57 |
--------------------------------------------------------------
Signed-off-by: Mathieu Desnoyers <mathieu.desnoyers@polymtl.ca>
Acked-by: Masami Hiramatsu <mhiramat@redhat.com>
Acked-by: 'Peter Zijlstra' <peterz@infradead.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-07-19 00:16:16 +08:00
|
|
|
*(__tracepoints) \
|
2009-02-06 00:51:38 +08:00
|
|
|
/* implement dynamic printk debug */ \
|
|
|
|
. = ALIGN(8); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start___verbose = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(__verbose)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___verbose = .; \
|
2008-11-21 14:30:54 +08:00
|
|
|
LIKELY_PROFILE() \
|
2009-02-24 23:21:36 +08:00
|
|
|
BRANCH_PROFILE() \
|
2013-07-13 05:07:27 +08:00
|
|
|
TRACE_PRINTKS() \
|
2018-03-29 03:05:37 +08:00
|
|
|
BPF_RAW_TP() \
|
2013-07-13 05:07:27 +08:00
|
|
|
TRACEPOINT_STR()
|
2007-05-17 19:38:44 +08:00
|
|
|
|
2009-06-08 02:46:37 +08:00
|
|
|
/*
|
|
|
|
* Data section helpers
|
|
|
|
*/
|
|
|
|
#define NOSAVE_DATA \
|
|
|
|
. = ALIGN(PAGE_SIZE); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__nosave_begin = .; \
|
2010-02-20 08:03:52 +08:00
|
|
|
*(.data..nosave) \
|
2009-06-08 02:46:37 +08:00
|
|
|
. = ALIGN(PAGE_SIZE); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__nosave_end = .;
|
2009-06-08 02:46:37 +08:00
|
|
|
|
|
|
|
#define PAGE_ALIGNED_DATA(page_align) \
|
|
|
|
. = ALIGN(page_align); \
|
2010-02-20 08:03:37 +08:00
|
|
|
*(.data..page_aligned)
|
2009-06-08 02:46:37 +08:00
|
|
|
|
|
|
|
#define READ_MOSTLY_DATA(align) \
|
|
|
|
. = ALIGN(align); \
|
2011-01-13 08:59:38 +08:00
|
|
|
*(.data..read_mostly) \
|
|
|
|
. = ALIGN(align);
|
2009-06-08 02:46:37 +08:00
|
|
|
|
|
|
|
#define CACHELINE_ALIGNED_DATA(align) \
|
|
|
|
. = ALIGN(align); \
|
2010-02-20 08:03:34 +08:00
|
|
|
*(.data..cacheline_aligned)
|
2009-06-08 02:46:37 +08:00
|
|
|
|
2009-06-24 06:53:15 +08:00
|
|
|
#define INIT_TASK_DATA(align) \
|
2009-06-08 02:46:37 +08:00
|
|
|
. = ALIGN(align); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start_init_task = .; \
|
|
|
|
init_thread_union = .; \
|
|
|
|
init_stack = .; \
|
2018-05-09 20:59:58 +08:00
|
|
|
KEEP(*(.data..init_task)) \
|
|
|
|
KEEP(*(.data..init_thread_info)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
. = __start_init_task + THREAD_SIZE; \
|
|
|
|
__end_init_task = .;
|
2009-06-08 02:46:37 +08:00
|
|
|
|
2018-09-19 14:51:43 +08:00
|
|
|
#define JUMP_TABLE_DATA \
|
|
|
|
. = ALIGN(8); \
|
|
|
|
__start___jump_table = .; \
|
|
|
|
KEEP(*(__jump_table)) \
|
|
|
|
__stop___jump_table = .;
|
|
|
|
|
2016-06-07 18:20:51 +08:00
|
|
|
/*
|
|
|
|
* Allow architectures to handle ro_after_init data on their
|
|
|
|
* own by defining an empty RO_AFTER_INIT_DATA.
|
|
|
|
*/
|
|
|
|
#ifndef RO_AFTER_INIT_DATA
|
2016-11-11 02:46:44 +08:00
|
|
|
#define RO_AFTER_INIT_DATA \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start_ro_after_init = .; \
|
2016-11-11 02:46:44 +08:00
|
|
|
*(.data..ro_after_init) \
|
2018-09-19 14:51:43 +08:00
|
|
|
JUMP_TABLE_DATA \
|
2018-05-09 15:23:51 +08:00
|
|
|
__end_ro_after_init = .;
|
2016-06-07 18:20:51 +08:00
|
|
|
#endif
|
|
|
|
|
2009-06-08 02:46:37 +08:00
|
|
|
/*
|
|
|
|
* Read only Data
|
|
|
|
*/
|
2019-10-30 05:13:34 +08:00
|
|
|
#define RO_DATA(align) \
|
2007-05-30 03:29:00 +08:00
|
|
|
. = ALIGN((align)); \
|
2005-04-17 06:20:36 +08:00
|
|
|
.rodata : AT(ADDR(.rodata) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start_rodata = .; \
|
2005-04-17 06:20:36 +08:00
|
|
|
*(.rodata) *(.rodata.*) \
|
2016-06-07 18:20:51 +08:00
|
|
|
RO_AFTER_INIT_DATA /* Read only after init */ \
|
2011-01-27 06:26:22 +08:00
|
|
|
. = ALIGN(8); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start___tracepoints_ptrs = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(__tracepoints_ptrs)) /* Tracepoints: pointer array */ \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___tracepoints_ptrs = .; \
|
tracing: Kernel Tracepoints
Implementation of kernel tracepoints. Inspired from the Linux Kernel
Markers. Allows complete typing verification by declaring both tracing
statement inline functions and probe registration/unregistration static
inline functions within the same macro "DEFINE_TRACE". No format string
is required. See the tracepoint Documentation and Samples patches for
usage examples.
Taken from the documentation patch :
"A tracepoint placed in code provides a hook to call a function (probe)
that you can provide at runtime. A tracepoint can be "on" (a probe is
connected to it) or "off" (no probe is attached). When a tracepoint is
"off" it has no effect, except for adding a tiny time penalty (checking
a condition for a branch) and space penalty (adding a few bytes for the
function call at the end of the instrumented function and adds a data
structure in a separate section). When a tracepoint is "on", the
function you provide is called each time the tracepoint is executed, in
the execution context of the caller. When the function provided ends its
execution, it returns to the caller (continuing from the tracepoint
site).
You can put tracepoints at important locations in the code. They are
lightweight hooks that can pass an arbitrary number of parameters, which
prototypes are described in a tracepoint declaration placed in a header
file."
Addition and removal of tracepoints is synchronized by RCU using the
scheduler (and preempt_disable) as guarantees to find a quiescent state
(this is really RCU "classic"). The update side uses rcu_barrier_sched()
with call_rcu_sched() and the read/execute side uses
"preempt_disable()/preempt_enable()".
We make sure the previous array containing probes, which has been
scheduled for deletion by the rcu callback, is indeed freed before we
proceed to the next update. It therefore limits the rate of modification
of a single tracepoint to one update per RCU period. The objective here
is to permit fast batch add/removal of probes on _different_
tracepoints.
Changelog :
- Use #name ":" #proto as string to identify the tracepoint in the
tracepoint table. This will make sure not type mismatch happens due to
connexion of a probe with the wrong type to a tracepoint declared with
the same name in a different header.
- Add tracepoint_entry_free_old.
- Change __TO_TRACE to get rid of the 'i' iterator.
Masami Hiramatsu <mhiramat@redhat.com> :
Tested on x86-64.
Performance impact of a tracepoint : same as markers, except that it
adds about 70 bytes of instructions in an unlikely branch of each
instrumented function (the for loop, the stack setup and the function
call). It currently adds a memory read, a test and a conditional branch
at the instrumentation site (in the hot path). Immediate values will
eventually change this into a load immediate, test and branch, which
removes the memory read which will make the i-cache impact smaller
(changing the memory read for a load immediate removes 3-4 bytes per
site on x86_32 (depending on mov prefixes), or 7-8 bytes on x86_64, it
also saves the d-cache hit).
About the performance impact of tracepoints (which is comparable to
markers), even without immediate values optimizations, tests done by
Hideo Aoki on ia64 show no regression. His test case was using hackbench
on a kernel where scheduler instrumentation (about 5 events in code
scheduler code) was added.
Quoting Hideo Aoki about Markers :
I evaluated overhead of kernel marker using linux-2.6-sched-fixes git
tree, which includes several markers for LTTng, using an ia64 server.
While the immediate trace mark feature isn't implemented on ia64, there
is no major performance regression. So, I think that we don't have any
issues to propose merging marker point patches into Linus's tree from
the viewpoint of performance impact.
I prepared two kernels to evaluate. The first one was compiled without
CONFIG_MARKERS. The second one was enabled CONFIG_MARKERS.
I downloaded the original hackbench from the following URL:
http://devresources.linux-foundation.org/craiger/hackbench/src/hackbench.c
I ran hackbench 5 times in each condition and calculated the average and
difference between the kernels.
The parameter of hackbench: every 50 from 50 to 800
The number of CPUs of the server: 2, 4, and 8
Below is the results. As you can see, major performance regression
wasn't found in any case. Even if number of processes increases,
differences between marker-enabled kernel and marker- disabled kernel
doesn't increase. Moreover, if number of CPUs increases, the differences
doesn't increase either.
Curiously, marker-enabled kernel is better than marker-disabled kernel
in more than half cases, although I guess it comes from the difference
of memory access pattern.
* 2 CPUs
Number of | without | with | diff | diff |
processes | Marker [Sec] | Marker [Sec] | [Sec] | [%] |
--------------------------------------------------------------
50 | 4.811 | 4.872 | +0.061 | +1.27 |
100 | 9.854 | 10.309 | +0.454 | +4.61 |
150 | 15.602 | 15.040 | -0.562 | -3.6 |
200 | 20.489 | 20.380 | -0.109 | -0.53 |
250 | 25.798 | 25.652 | -0.146 | -0.56 |
300 | 31.260 | 30.797 | -0.463 | -1.48 |
350 | 36.121 | 35.770 | -0.351 | -0.97 |
400 | 42.288 | 42.102 | -0.186 | -0.44 |
450 | 47.778 | 47.253 | -0.526 | -1.1 |
500 | 51.953 | 52.278 | +0.325 | +0.63 |
550 | 58.401 | 57.700 | -0.701 | -1.2 |
600 | 63.334 | 63.222 | -0.112 | -0.18 |
650 | 68.816 | 68.511 | -0.306 | -0.44 |
700 | 74.667 | 74.088 | -0.579 | -0.78 |
750 | 78.612 | 79.582 | +0.970 | +1.23 |
800 | 85.431 | 85.263 | -0.168 | -0.2 |
--------------------------------------------------------------
* 4 CPUs
Number of | without | with | diff | diff |
processes | Marker [Sec] | Marker [Sec] | [Sec] | [%] |
--------------------------------------------------------------
50 | 2.586 | 2.584 | -0.003 | -0.1 |
100 | 5.254 | 5.283 | +0.030 | +0.56 |
150 | 8.012 | 8.074 | +0.061 | +0.76 |
200 | 11.172 | 11.000 | -0.172 | -1.54 |
250 | 13.917 | 14.036 | +0.119 | +0.86 |
300 | 16.905 | 16.543 | -0.362 | -2.14 |
350 | 19.901 | 20.036 | +0.135 | +0.68 |
400 | 22.908 | 23.094 | +0.186 | +0.81 |
450 | 26.273 | 26.101 | -0.172 | -0.66 |
500 | 29.554 | 29.092 | -0.461 | -1.56 |
550 | 32.377 | 32.274 | -0.103 | -0.32 |
600 | 35.855 | 35.322 | -0.533 | -1.49 |
650 | 39.192 | 38.388 | -0.804 | -2.05 |
700 | 41.744 | 41.719 | -0.025 | -0.06 |
750 | 45.016 | 44.496 | -0.520 | -1.16 |
800 | 48.212 | 47.603 | -0.609 | -1.26 |
--------------------------------------------------------------
* 8 CPUs
Number of | without | with | diff | diff |
processes | Marker [Sec] | Marker [Sec] | [Sec] | [%] |
--------------------------------------------------------------
50 | 2.094 | 2.072 | -0.022 | -1.07 |
100 | 4.162 | 4.273 | +0.111 | +2.66 |
150 | 6.485 | 6.540 | +0.055 | +0.84 |
200 | 8.556 | 8.478 | -0.078 | -0.91 |
250 | 10.458 | 10.258 | -0.200 | -1.91 |
300 | 12.425 | 12.750 | +0.325 | +2.62 |
350 | 14.807 | 14.839 | +0.032 | +0.22 |
400 | 16.801 | 16.959 | +0.158 | +0.94 |
450 | 19.478 | 19.009 | -0.470 | -2.41 |
500 | 21.296 | 21.504 | +0.208 | +0.98 |
550 | 23.842 | 23.979 | +0.137 | +0.57 |
600 | 26.309 | 26.111 | -0.198 | -0.75 |
650 | 28.705 | 28.446 | -0.259 | -0.9 |
700 | 31.233 | 31.394 | +0.161 | +0.52 |
750 | 34.064 | 33.720 | -0.344 | -1.01 |
800 | 36.320 | 36.114 | -0.206 | -0.57 |
--------------------------------------------------------------
Signed-off-by: Mathieu Desnoyers <mathieu.desnoyers@polymtl.ca>
Acked-by: Masami Hiramatsu <mhiramat@redhat.com>
Acked-by: 'Peter Zijlstra' <peterz@infradead.org>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
2008-07-19 00:16:16 +08:00
|
|
|
*(__tracepoints_strings)/* Tracepoints: strings */ \
|
2005-04-17 06:20:36 +08:00
|
|
|
} \
|
|
|
|
\
|
|
|
|
.rodata1 : AT(ADDR(.rodata1) - LOAD_OFFSET) { \
|
|
|
|
*(.rodata1) \
|
|
|
|
} \
|
|
|
|
\
|
|
|
|
/* PCI quirks */ \
|
|
|
|
.pci_fixup : AT(ADDR(.pci_fixup) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start_pci_fixups_early = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(.pci_fixup_early)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__end_pci_fixups_early = .; \
|
|
|
|
__start_pci_fixups_header = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(.pci_fixup_header)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__end_pci_fixups_header = .; \
|
|
|
|
__start_pci_fixups_final = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(.pci_fixup_final)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__end_pci_fixups_final = .; \
|
|
|
|
__start_pci_fixups_enable = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(.pci_fixup_enable)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__end_pci_fixups_enable = .; \
|
|
|
|
__start_pci_fixups_resume = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(.pci_fixup_resume)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__end_pci_fixups_resume = .; \
|
|
|
|
__start_pci_fixups_resume_early = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(.pci_fixup_resume_early)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__end_pci_fixups_resume_early = .; \
|
|
|
|
__start_pci_fixups_suspend = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(.pci_fixup_suspend)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__end_pci_fixups_suspend = .; \
|
|
|
|
__start_pci_fixups_suspend_late = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(.pci_fixup_suspend_late)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__end_pci_fixups_suspend_late = .; \
|
2005-04-17 06:20:36 +08:00
|
|
|
} \
|
|
|
|
\
|
2008-05-23 20:52:42 +08:00
|
|
|
/* Built-in firmware blobs */ \
|
|
|
|
.builtin_fw : AT(ADDR(.builtin_fw) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start_builtin_fw = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(.builtin_fw)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__end_builtin_fw = .; \
|
2008-05-23 20:52:42 +08:00
|
|
|
} \
|
|
|
|
\
|
2008-05-12 21:44:41 +08:00
|
|
|
TRACEDATA \
|
|
|
|
\
|
2005-04-17 06:20:36 +08:00
|
|
|
/* Kernel symbol table: Normal symbols */ \
|
|
|
|
__ksymtab : AT(ADDR(__ksymtab) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start___ksymtab = .; \
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
KEEP(*(SORT(___ksymtab+*))) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___ksymtab = .; \
|
2005-04-17 06:20:36 +08:00
|
|
|
} \
|
|
|
|
\
|
|
|
|
/* Kernel symbol table: GPL-only symbols */ \
|
|
|
|
__ksymtab_gpl : AT(ADDR(__ksymtab_gpl) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start___ksymtab_gpl = .; \
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
KEEP(*(SORT(___ksymtab_gpl+*))) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___ksymtab_gpl = .; \
|
2005-04-17 06:20:36 +08:00
|
|
|
} \
|
|
|
|
\
|
2006-06-28 19:26:45 +08:00
|
|
|
/* Kernel symbol table: Normal unused symbols */ \
|
|
|
|
__ksymtab_unused : AT(ADDR(__ksymtab_unused) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start___ksymtab_unused = .; \
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
KEEP(*(SORT(___ksymtab_unused+*))) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___ksymtab_unused = .; \
|
2006-06-28 19:26:45 +08:00
|
|
|
} \
|
|
|
|
\
|
|
|
|
/* Kernel symbol table: GPL-only unused symbols */ \
|
|
|
|
__ksymtab_unused_gpl : AT(ADDR(__ksymtab_unused_gpl) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start___ksymtab_unused_gpl = .; \
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
KEEP(*(SORT(___ksymtab_unused_gpl+*))) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___ksymtab_unused_gpl = .; \
|
2006-06-28 19:26:45 +08:00
|
|
|
} \
|
|
|
|
\
|
2006-03-21 05:17:13 +08:00
|
|
|
/* Kernel symbol table: GPL-future-only symbols */ \
|
|
|
|
__ksymtab_gpl_future : AT(ADDR(__ksymtab_gpl_future) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start___ksymtab_gpl_future = .; \
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
KEEP(*(SORT(___ksymtab_gpl_future+*))) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___ksymtab_gpl_future = .; \
|
2006-03-21 05:17:13 +08:00
|
|
|
} \
|
|
|
|
\
|
2005-04-17 06:20:36 +08:00
|
|
|
/* Kernel symbol table: Normal symbols */ \
|
|
|
|
__kcrctab : AT(ADDR(__kcrctab) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start___kcrctab = .; \
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
KEEP(*(SORT(___kcrctab+*))) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___kcrctab = .; \
|
2005-04-17 06:20:36 +08:00
|
|
|
} \
|
|
|
|
\
|
|
|
|
/* Kernel symbol table: GPL-only symbols */ \
|
|
|
|
__kcrctab_gpl : AT(ADDR(__kcrctab_gpl) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start___kcrctab_gpl = .; \
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
KEEP(*(SORT(___kcrctab_gpl+*))) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___kcrctab_gpl = .; \
|
2005-04-17 06:20:36 +08:00
|
|
|
} \
|
|
|
|
\
|
2006-06-28 19:26:45 +08:00
|
|
|
/* Kernel symbol table: Normal unused symbols */ \
|
|
|
|
__kcrctab_unused : AT(ADDR(__kcrctab_unused) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start___kcrctab_unused = .; \
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
KEEP(*(SORT(___kcrctab_unused+*))) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___kcrctab_unused = .; \
|
2006-06-28 19:26:45 +08:00
|
|
|
} \
|
|
|
|
\
|
|
|
|
/* Kernel symbol table: GPL-only unused symbols */ \
|
|
|
|
__kcrctab_unused_gpl : AT(ADDR(__kcrctab_unused_gpl) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start___kcrctab_unused_gpl = .; \
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
KEEP(*(SORT(___kcrctab_unused_gpl+*))) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___kcrctab_unused_gpl = .; \
|
2006-06-28 19:26:45 +08:00
|
|
|
} \
|
|
|
|
\
|
2006-03-21 05:17:13 +08:00
|
|
|
/* Kernel symbol table: GPL-future-only symbols */ \
|
|
|
|
__kcrctab_gpl_future : AT(ADDR(__kcrctab_gpl_future) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start___kcrctab_gpl_future = .; \
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
KEEP(*(SORT(___kcrctab_gpl_future+*))) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___kcrctab_gpl_future = .; \
|
2006-03-21 05:17:13 +08:00
|
|
|
} \
|
|
|
|
\
|
2005-04-17 06:20:36 +08:00
|
|
|
/* Kernel symbol table: strings */ \
|
|
|
|
__ksymtab_strings : AT(ADDR(__ksymtab_strings) - LOAD_OFFSET) { \
|
2016-11-24 00:41:41 +08:00
|
|
|
*(__ksymtab_strings) \
|
2005-04-17 06:20:36 +08:00
|
|
|
} \
|
|
|
|
\
|
2008-01-21 03:07:28 +08:00
|
|
|
/* __*init sections */ \
|
|
|
|
__init_rodata : AT(ADDR(__init_rodata) - LOAD_OFFSET) { \
|
2008-01-29 03:21:15 +08:00
|
|
|
*(.ref.rodata) \
|
2008-01-21 03:07:28 +08:00
|
|
|
MEM_KEEP(init.rodata) \
|
|
|
|
MEM_KEEP(exit.rodata) \
|
|
|
|
} \
|
|
|
|
\
|
2005-04-17 06:20:36 +08:00
|
|
|
/* Built-in module parameters. */ \
|
|
|
|
__param : AT(ADDR(__param) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start___param = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(__param)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___param = .; \
|
2010-12-16 06:00:19 +08:00
|
|
|
} \
|
|
|
|
\
|
|
|
|
/* Built-in module versions. */ \
|
|
|
|
__modver : AT(ADDR(__modver) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start___modver = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(__modver)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___modver = .; \
|
2006-09-27 16:51:02 +08:00
|
|
|
} \
|
2019-10-30 05:13:32 +08:00
|
|
|
\
|
|
|
|
NOTES \
|
|
|
|
\
|
|
|
|
. = ALIGN((align)); \
|
|
|
|
__end_rodata = .;
|
2007-05-30 03:29:00 +08:00
|
|
|
|
2017-07-26 20:46:27 +08:00
|
|
|
/*
|
|
|
|
* .text section. Map to function alignment to avoid address changes
|
2016-09-14 10:24:03 +08:00
|
|
|
* during second ld run in second ld pass when generating System.map
|
2017-07-26 20:46:27 +08:00
|
|
|
*
|
|
|
|
* TEXT_MAIN here will match .text.fixup and .text.unlikely if dead
|
|
|
|
* code elimination is enabled, so these sections should be converted
|
|
|
|
* to use ".." first.
|
|
|
|
*/
|
2007-05-13 06:31:33 +08:00
|
|
|
#define TEXT_TEXT \
|
|
|
|
ALIGN_FUNCTION(); \
|
2017-07-26 20:46:27 +08:00
|
|
|
*(.text.hot TEXT_MAIN .text.fixup .text.unlikely) \
|
locking/refcounts, x86/asm: Use unique .text section for refcount exceptions
Using .text.unlikely for refcount exceptions isn't safe because gcc may
move entire functions into .text.unlikely (e.g. in6_dev_dev()), which
would cause any uses of a protected refcount_t function to stay inline
with the function, triggering the protection unconditionally:
.section .text.unlikely,"ax",@progbits
.type in6_dev_get, @function
in6_dev_getx:
.LFB4673:
.loc 2 4128 0
.cfi_startproc
...
lock; incl 480(%rbx)
js 111f
.pushsection .text.unlikely
111: lea 480(%rbx), %rcx
112: .byte 0x0f, 0xff
.popsection
113:
This creates a unique .text..refcount section and adds an additional
test to the exception handler to WARN in the case of having none of OF,
SF, nor ZF set so we can see things like this more easily in the future.
The double dot for the section name keeps it out of the TEXT_MAIN macro
namespace, to avoid collisions and so it can be put at the end with
text.unlikely to keep the cold code together.
See commit:
cb87481ee89db ("kbuild: linker script do not match C names unless LD_DEAD_CODE_DATA_ELIMINATION is configured")
... which matches C names: [a-zA-Z0-9_] but not ".".
Reported-by: Mike Galbraith <efault@gmx.de>
Signed-off-by: Kees Cook <keescook@chromium.org>
Cc: Ard Biesheuvel <ard.biesheuvel@linaro.org>
Cc: Elena <elena.reshetova@intel.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: linux-arch <linux-arch@vger.kernel.org>
Fixes: 7a46ec0e2f48 ("locking/refcounts, x86/asm: Implement fast refcount overflow protection")
Link: http://lkml.kernel.org/r/1504382986-49301-2-git-send-email-keescook@chromium.org
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2017-09-03 04:09:45 +08:00
|
|
|
*(.text..refcount) \
|
2008-01-29 03:21:15 +08:00
|
|
|
*(.ref.text) \
|
2018-05-09 20:59:58 +08:00
|
|
|
MEM_KEEP(init.text*) \
|
|
|
|
MEM_KEEP(exit.text*) \
|
2008-01-21 03:07:28 +08:00
|
|
|
|
2007-05-13 06:31:33 +08:00
|
|
|
|
2005-07-15 04:15:44 +08:00
|
|
|
/* sched.text is aling to function alignment to secure we have same
|
|
|
|
* address even at second ld pass when generating System.map */
|
2005-04-17 06:20:36 +08:00
|
|
|
#define SCHED_TEXT \
|
2005-07-15 04:15:44 +08:00
|
|
|
ALIGN_FUNCTION(); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__sched_text_start = .; \
|
2005-04-17 06:20:36 +08:00
|
|
|
*(.sched.text) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__sched_text_end = .;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2005-07-15 04:15:44 +08:00
|
|
|
/* spinlock.text is aling to function alignment to secure we have same
|
|
|
|
* address even at second ld pass when generating System.map */
|
2005-04-17 06:20:36 +08:00
|
|
|
#define LOCK_TEXT \
|
2005-07-15 04:15:44 +08:00
|
|
|
ALIGN_FUNCTION(); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__lock_text_start = .; \
|
2005-04-17 06:20:36 +08:00
|
|
|
*(.spinlock.text) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__lock_text_end = .;
|
2005-09-07 06:19:26 +08:00
|
|
|
|
2016-10-08 08:02:55 +08:00
|
|
|
#define CPUIDLE_TEXT \
|
|
|
|
ALIGN_FUNCTION(); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__cpuidle_text_start = .; \
|
2016-10-08 08:02:55 +08:00
|
|
|
*(.cpuidle.text) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__cpuidle_text_end = .;
|
2016-10-08 08:02:55 +08:00
|
|
|
|
2005-09-07 06:19:26 +08:00
|
|
|
#define KPROBES_TEXT \
|
|
|
|
ALIGN_FUNCTION(); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__kprobes_text_start = .; \
|
2005-09-07 06:19:26 +08:00
|
|
|
*(.kprobes.text) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__kprobes_text_end = .;
|
2005-09-11 01:44:54 +08:00
|
|
|
|
2011-03-08 02:10:39 +08:00
|
|
|
#define ENTRY_TEXT \
|
|
|
|
ALIGN_FUNCTION(); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__entry_text_start = .; \
|
2011-03-08 02:10:39 +08:00
|
|
|
*(.entry.text) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__entry_text_end = .;
|
2011-03-08 02:10:39 +08:00
|
|
|
|
2008-12-10 06:53:16 +08:00
|
|
|
#define IRQENTRY_TEXT \
|
|
|
|
ALIGN_FUNCTION(); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__irqentry_text_start = .; \
|
2008-12-10 06:53:16 +08:00
|
|
|
*(.irqentry.text) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__irqentry_text_end = .;
|
2008-12-10 06:53:16 +08:00
|
|
|
|
2016-03-26 05:22:05 +08:00
|
|
|
#define SOFTIRQENTRY_TEXT \
|
|
|
|
ALIGN_FUNCTION(); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__softirqentry_text_start = .; \
|
2016-03-26 05:22:05 +08:00
|
|
|
*(.softirqentry.text) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__softirqentry_text_end = .;
|
2016-03-26 05:22:05 +08:00
|
|
|
|
2008-02-20 04:00:18 +08:00
|
|
|
/* Section used for early init (in .S files) */
|
2018-05-09 20:59:58 +08:00
|
|
|
#define HEAD_TEXT KEEP(*(.head.text))
|
2008-02-20 04:00:18 +08:00
|
|
|
|
2009-06-15 04:10:41 +08:00
|
|
|
#define HEAD_TEXT_SECTION \
|
2009-06-08 02:46:37 +08:00
|
|
|
.head.text : AT(ADDR(.head.text) - LOAD_OFFSET) { \
|
|
|
|
HEAD_TEXT \
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Exception table
|
|
|
|
*/
|
|
|
|
#define EXCEPTION_TABLE(align) \
|
|
|
|
. = ALIGN(align); \
|
|
|
|
__ex_table : AT(ADDR(__ex_table) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start___ex_table = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(__ex_table)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___ex_table = .; \
|
2009-06-08 02:46:37 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Init task
|
|
|
|
*/
|
2009-06-24 06:53:15 +08:00
|
|
|
#define INIT_TASK_DATA_SECTION(align) \
|
2009-06-08 02:46:37 +08:00
|
|
|
. = ALIGN(align); \
|
2010-07-13 17:39:42 +08:00
|
|
|
.data..init_task : AT(ADDR(.data..init_task) - LOAD_OFFSET) { \
|
2009-06-24 06:53:15 +08:00
|
|
|
INIT_TASK_DATA(align) \
|
2009-06-08 02:46:37 +08:00
|
|
|
}
|
2008-02-20 04:00:18 +08:00
|
|
|
|
2009-06-18 07:28:03 +08:00
|
|
|
#ifdef CONFIG_CONSTRUCTORS
|
2009-07-01 02:41:13 +08:00
|
|
|
#define KERNEL_CTORS() . = ALIGN(8); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__ctors_start = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(.ctors)) \
|
|
|
|
KEEP(*(SORT(.init_array.*))) \
|
|
|
|
KEEP(*(.init_array)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__ctors_end = .;
|
2009-06-18 07:28:03 +08:00
|
|
|
#else
|
|
|
|
#define KERNEL_CTORS()
|
|
|
|
#endif
|
|
|
|
|
2008-01-20 21:15:03 +08:00
|
|
|
/* init and exit section handling */
|
2008-01-21 03:07:28 +08:00
|
|
|
#define INIT_DATA \
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
KEEP(*(SORT(___kentry+*))) \
|
2018-05-09 20:59:58 +08:00
|
|
|
*(.init.data init.data.*) \
|
|
|
|
MEM_DISCARD(init.data*) \
|
2009-06-18 07:28:03 +08:00
|
|
|
KERNEL_CTORS() \
|
2009-07-28 02:23:50 +08:00
|
|
|
MCOUNT_REC() \
|
2018-05-09 20:59:58 +08:00
|
|
|
*(.init.rodata .init.rodata.*) \
|
tracing: Replace trace_event struct array with pointer array
Currently the trace_event structures are placed in the _ftrace_events
section, and at link time, the linker makes one large array of all
the trace_event structures. On boot up, this array is read (much like
the initcall sections) and the events are processed.
The problem is that there is no guarantee that gcc will place complex
structures nicely together in an array format. Two structures in the
same file may be placed awkwardly, because gcc has no clue that they
are suppose to be in an array.
A hack was used previous to force the alignment to 4, to pack the
structures together. But this caused alignment issues with other
architectures (sparc).
Instead of packing the structures into an array, the structures' addresses
are now put into the _ftrace_event section. As pointers are always the
natural alignment, gcc should always pack them tightly together
(otherwise initcall, extable, etc would also fail).
By having the pointers to the structures in the section, we can still
iterate the trace_events without causing unnecessary alignment problems
with other architectures, or depending on the current behaviour of
gcc that will likely change in the future just to tick us kernel developers
off a little more.
The _ftrace_event section is also moved into the .init.data section
as it is now only needed at boot up.
Suggested-by: David Miller <davem@davemloft.net>
Cc: Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
Acked-by: David S. Miller <davem@davemloft.net>
Signed-off-by: Steven Rostedt <rostedt@goodmis.org>
2011-01-27 22:15:30 +08:00
|
|
|
FTRACE_EVENTS() \
|
tracing: Replace syscall_meta_data struct array with pointer array
Currently the syscall_meta structures for the syscall tracepoints are
placed in the __syscall_metadata section, and at link time, the linker
makes one large array of all these syscall metadata structures. On boot
up, this array is read (much like the initcall sections) and the syscall
data is processed.
The problem is that there is no guarantee that gcc will place complex
structures nicely together in an array format. Two structures in the
same file may be placed awkwardly, because gcc has no clue that they
are suppose to be in an array.
A hack was used previous to force the alignment to 4, to pack the
structures together. But this caused alignment issues with other
architectures (sparc).
Instead of packing the structures into an array, the structures' addresses
are now put into the __syscall_metadata section. As pointers are always the
natural alignment, gcc should always pack them tightly together
(otherwise initcall, extable, etc would also fail).
By having the pointers to the structures in the section, we can still
iterate the trace_events without causing unnecessary alignment problems
with other architectures, or depending on the current behaviour of
gcc that will likely change in the future just to tick us kernel developers
off a little more.
The __syscall_metadata section is also moved into the .init.data section
as it is now only needed at boot up.
Suggested-by: David Miller <davem@davemloft.net>
Acked-by: David S. Miller <davem@davemloft.net>
Cc: Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
Signed-off-by: Steven Rostedt <rostedt@goodmis.org>
2011-02-03 06:06:09 +08:00
|
|
|
TRACE_SYSCALLS() \
|
2014-04-17 16:17:05 +08:00
|
|
|
KPROBE_BLACKLIST() \
|
2018-01-13 01:55:03 +08:00
|
|
|
ERROR_INJECT_WHITELIST() \
|
2010-12-23 03:57:26 +08:00
|
|
|
MEM_DISCARD(init.rodata) \
|
2013-01-04 15:00:52 +08:00
|
|
|
CLK_OF_TABLES() \
|
2014-02-28 21:42:49 +08:00
|
|
|
RESERVEDMEM_OF_TABLES() \
|
2017-05-27 00:33:27 +08:00
|
|
|
TIMER_OF_TABLES() \
|
2013-10-31 09:21:09 +08:00
|
|
|
CPU_METHOD_OF_TABLES() \
|
2015-02-02 23:32:45 +08:00
|
|
|
CPUIDLE_METHOD_OF_TABLES() \
|
irqchip: add basic infrastructure
With the recent creation of the drivers/irqchip/ directory, it is
desirable to move irq controller drivers here. At the moment, the only
driver here is irq-bcm2835, the driver for the irq controller found in
the ARM BCM2835 SoC, present in Rasberry Pi systems. This irq
controller driver was exporting its initialization function and its
irq handling function through a header file in
<linux/irqchip/bcm2835.h>.
When proposing to also move another irq controller driver in
drivers/irqchip, Rob Herring raised the very valid point that moving
things to drivers/irqchip was good in order to remove more stuff from
arch/arm, but if it means adding gazillions of headers files in
include/linux/irqchip/, it would not be very nice.
So, upon the suggestion of Rob Herring and Arnd Bergmann, this commit
introduces a small infrastructure that defines a central
irqchip_init() function in drivers/irqchip/irqchip.c, which is meant
to be called as the ->init_irq() callback of ARM platforms. This
function calls of_irq_init() with an array of match strings and init
functions generated from a special linker section.
Note that the irq controller driver initialization function is
responsible for setting the global handle_arch_irq() variable, so that
ARM platforms no longer have to define the ->handle_irq field in their
DT_MACHINE structure.
A global header, <linux/irqchip.h> is also added to expose the single
irqchip_init() function to the reset of the kernel.
A further commit moves the BCM2835 irq controller driver to this new
small infrastructure, therefore removing the include/linux/irqchip/
directory.
Signed-off-by: Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
Reviewed-by: Stephen Warren <swarren@wwwdotorg.org>
Reviewed-by: Rob Herring <rob.herring@calxeda.com>
Acked-by: Arnd Bergmann <arnd@arndb.de>
[rob.herring: reword commit message to reflect use of linker sections.]
Signed-off-by: Rob Herring <rob.herring@calxeda.com>
2012-11-21 06:00:52 +08:00
|
|
|
KERNEL_DTB() \
|
2014-03-27 21:06:16 +08:00
|
|
|
IRQCHIP_OF_MATCH_TABLE() \
|
2015-09-28 22:49:13 +08:00
|
|
|
ACPI_PROBE_TABLE(irqchip) \
|
2017-05-27 00:33:27 +08:00
|
|
|
ACPI_PROBE_TABLE(timer) \
|
2019-06-13 04:13:24 +08:00
|
|
|
THERMAL_TABLE(governor) \
|
2018-10-11 08:18:22 +08:00
|
|
|
EARLYCON_TABLE() \
|
2019-08-20 08:17:37 +08:00
|
|
|
LSM_TABLE() \
|
|
|
|
EARLY_LSM_TABLE()
|
2008-01-21 03:07:28 +08:00
|
|
|
|
|
|
|
#define INIT_TEXT \
|
2018-05-09 20:59:58 +08:00
|
|
|
*(.init.text .init.text.*) \
|
2016-07-15 03:07:29 +08:00
|
|
|
*(.text.startup) \
|
2018-05-09 20:59:58 +08:00
|
|
|
MEM_DISCARD(init.text*)
|
2008-01-21 03:07:28 +08:00
|
|
|
|
|
|
|
#define EXIT_DATA \
|
2018-05-09 20:59:58 +08:00
|
|
|
*(.exit.data .exit.data.*) \
|
2018-09-13 18:59:59 +08:00
|
|
|
*(.fini_array .fini_array.*) \
|
|
|
|
*(.dtors .dtors.*) \
|
2018-05-09 20:59:58 +08:00
|
|
|
MEM_DISCARD(exit.data*) \
|
|
|
|
MEM_DISCARD(exit.rodata*)
|
2008-01-20 21:15:03 +08:00
|
|
|
|
2008-01-21 03:07:28 +08:00
|
|
|
#define EXIT_TEXT \
|
|
|
|
*(.exit.text) \
|
2016-07-15 03:07:29 +08:00
|
|
|
*(.text.exit) \
|
2008-01-21 03:07:28 +08:00
|
|
|
MEM_DISCARD(exit.text)
|
2008-01-20 21:15:03 +08:00
|
|
|
|
2009-06-15 04:10:41 +08:00
|
|
|
#define EXIT_CALL \
|
|
|
|
*(.exitcall.exit)
|
|
|
|
|
2009-06-08 02:46:37 +08:00
|
|
|
/*
|
|
|
|
* bss (Block Started by Symbol) - uninitialized data
|
|
|
|
* zeroed during startup
|
|
|
|
*/
|
2009-07-13 06:23:33 +08:00
|
|
|
#define SBSS(sbss_align) \
|
|
|
|
. = ALIGN(sbss_align); \
|
2009-06-08 02:46:37 +08:00
|
|
|
.sbss : AT(ADDR(.sbss) - LOAD_OFFSET) { \
|
2017-05-12 01:40:40 +08:00
|
|
|
*(.dynsbss) \
|
2018-05-09 20:59:58 +08:00
|
|
|
*(SBSS_MAIN) \
|
2009-06-08 02:46:37 +08:00
|
|
|
*(.scommon) \
|
|
|
|
}
|
|
|
|
|
2012-08-15 02:08:00 +08:00
|
|
|
/*
|
|
|
|
* Allow archectures to redefine BSS_FIRST_SECTIONS to add extra
|
|
|
|
* sections to the front of bss.
|
|
|
|
*/
|
|
|
|
#ifndef BSS_FIRST_SECTIONS
|
|
|
|
#define BSS_FIRST_SECTIONS
|
|
|
|
#endif
|
|
|
|
|
2009-06-08 02:46:37 +08:00
|
|
|
#define BSS(bss_align) \
|
|
|
|
. = ALIGN(bss_align); \
|
|
|
|
.bss : AT(ADDR(.bss) - LOAD_OFFSET) { \
|
2012-08-15 02:08:00 +08:00
|
|
|
BSS_FIRST_SECTIONS \
|
2010-02-20 08:03:38 +08:00
|
|
|
*(.bss..page_aligned) \
|
2009-06-08 02:46:37 +08:00
|
|
|
*(.dynbss) \
|
2017-07-26 20:46:27 +08:00
|
|
|
*(BSS_MAIN) \
|
2009-06-08 02:46:37 +08:00
|
|
|
*(COMMON) \
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* DWARF debug sections.
|
|
|
|
* Symbols in the DWARF debugging sections are relative to
|
|
|
|
* the beginning of the section so we begin them at 0.
|
|
|
|
*/
|
2005-09-11 01:44:54 +08:00
|
|
|
#define DWARF_DEBUG \
|
|
|
|
/* DWARF 1 */ \
|
|
|
|
.debug 0 : { *(.debug) } \
|
|
|
|
.line 0 : { *(.line) } \
|
|
|
|
/* GNU DWARF 1 extensions */ \
|
|
|
|
.debug_srcinfo 0 : { *(.debug_srcinfo) } \
|
|
|
|
.debug_sfnames 0 : { *(.debug_sfnames) } \
|
|
|
|
/* DWARF 1.1 and DWARF 2 */ \
|
|
|
|
.debug_aranges 0 : { *(.debug_aranges) } \
|
|
|
|
.debug_pubnames 0 : { *(.debug_pubnames) } \
|
|
|
|
/* DWARF 2 */ \
|
|
|
|
.debug_info 0 : { *(.debug_info \
|
|
|
|
.gnu.linkonce.wi.*) } \
|
|
|
|
.debug_abbrev 0 : { *(.debug_abbrev) } \
|
|
|
|
.debug_line 0 : { *(.debug_line) } \
|
|
|
|
.debug_frame 0 : { *(.debug_frame) } \
|
|
|
|
.debug_str 0 : { *(.debug_str) } \
|
|
|
|
.debug_loc 0 : { *(.debug_loc) } \
|
|
|
|
.debug_macinfo 0 : { *(.debug_macinfo) } \
|
2017-05-12 01:40:40 +08:00
|
|
|
.debug_pubtypes 0 : { *(.debug_pubtypes) } \
|
|
|
|
/* DWARF 3 */ \
|
|
|
|
.debug_ranges 0 : { *(.debug_ranges) } \
|
2005-09-11 01:44:54 +08:00
|
|
|
/* SGI/MIPS DWARF 2 extensions */ \
|
|
|
|
.debug_weaknames 0 : { *(.debug_weaknames) } \
|
|
|
|
.debug_funcnames 0 : { *(.debug_funcnames) } \
|
|
|
|
.debug_typenames 0 : { *(.debug_typenames) } \
|
|
|
|
.debug_varnames 0 : { *(.debug_varnames) } \
|
2017-05-12 01:40:40 +08:00
|
|
|
/* GNU DWARF 2 extensions */ \
|
|
|
|
.debug_gnu_pubnames 0 : { *(.debug_gnu_pubnames) } \
|
|
|
|
.debug_gnu_pubtypes 0 : { *(.debug_gnu_pubtypes) } \
|
|
|
|
/* DWARF 4 */ \
|
|
|
|
.debug_types 0 : { *(.debug_types) } \
|
|
|
|
/* DWARF 5 */ \
|
|
|
|
.debug_macro 0 : { *(.debug_macro) } \
|
|
|
|
.debug_addr 0 : { *(.debug_addr) }
|
2005-09-11 01:44:54 +08:00
|
|
|
|
|
|
|
/* Stabs debugging sections. */
|
|
|
|
#define STABS_DEBUG \
|
|
|
|
.stab 0 : { *(.stab) } \
|
|
|
|
.stabstr 0 : { *(.stabstr) } \
|
|
|
|
.stab.excl 0 : { *(.stab.excl) } \
|
|
|
|
.stab.exclstr 0 : { *(.stab.exclstr) } \
|
|
|
|
.stab.index 0 : { *(.stab.index) } \
|
|
|
|
.stab.indexstr 0 : { *(.stab.indexstr) } \
|
|
|
|
.comment 0 : { *(.comment) }
|
2006-09-26 14:32:26 +08:00
|
|
|
|
2008-05-12 21:44:41 +08:00
|
|
|
#ifdef CONFIG_GENERIC_BUG
|
[PATCH] Generic BUG implementation
This patch adds common handling for kernel BUGs, for use by architectures as
they wish. The code is derived from arch/powerpc.
The advantages of having common BUG handling are:
- consistent BUG reporting across architectures
- shared implementation of out-of-line file/line data
- implement CONFIG_DEBUG_BUGVERBOSE consistently
This means that in inline impact of BUG is just the illegal instruction
itself, which is an improvement for i386 and x86-64.
A BUG is represented in the instruction stream as an illegal instruction,
which has file/line information associated with it. This extra information is
stored in the __bug_table section in the ELF file.
When the kernel gets an illegal instruction, it first confirms it might
possibly be from a BUG (ie, in kernel mode, the right illegal instruction).
It then calls report_bug(). This searches __bug_table for a matching
instruction pointer, and if found, prints the corresponding file/line
information. If report_bug() determines that it wasn't a BUG which caused the
trap, it returns BUG_TRAP_TYPE_NONE.
Some architectures (powerpc) implement WARN using the same mechanism; if the
illegal instruction was the result of a WARN, then report_bug(Q) returns
CONFIG_DEBUG_BUGVERBOSE; otherwise it returns BUG_TRAP_TYPE_BUG.
lib/bug.c keeps a list of loaded modules which can be searched for __bug_table
entries. The architecture must call
module_bug_finalize()/module_bug_cleanup() from its corresponding
module_finalize/cleanup functions.
Unsetting CONFIG_DEBUG_BUGVERBOSE will reduce the kernel size by some amount.
At the very least, filename and line information will not be recorded for each
but, but architectures may decide to store no extra information per BUG at
all.
Unfortunately, gcc doesn't have a general way to mark an asm() as noreturn, so
architectures will generally have to include an infinite loop (or similar) in
the BUG code, so that gcc knows execution won't continue beyond that point.
gcc does have a __builtin_trap() operator which may be useful to achieve the
same effect, unfortunately it cannot be used to actually implement the BUG
itself, because there's no way to get the instruction's address for use in
generating the __bug_table entry.
[randy.dunlap@oracle.com: Handle BUG=n, GENERIC_BUG=n to prevent build errors]
[bunk@stusta.de: include/linux/bug.h must always #include <linux/module.h]
Signed-off-by: Jeremy Fitzhardinge <jeremy@goop.org>
Cc: Andi Kleen <ak@muc.de>
Cc: Hugh Dickens <hugh@veritas.com>
Cc: Michael Ellerman <michael@ellerman.id.au>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Rusty Russell <rusty@rustcorp.com.au>
Signed-off-by: Adrian Bunk <bunk@stusta.de>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-12-08 18:36:19 +08:00
|
|
|
#define BUG_TABLE \
|
|
|
|
. = ALIGN(8); \
|
|
|
|
__bug_table : AT(ADDR(__bug_table) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start___bug_table = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(__bug_table)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop___bug_table = .; \
|
[PATCH] Generic BUG implementation
This patch adds common handling for kernel BUGs, for use by architectures as
they wish. The code is derived from arch/powerpc.
The advantages of having common BUG handling are:
- consistent BUG reporting across architectures
- shared implementation of out-of-line file/line data
- implement CONFIG_DEBUG_BUGVERBOSE consistently
This means that in inline impact of BUG is just the illegal instruction
itself, which is an improvement for i386 and x86-64.
A BUG is represented in the instruction stream as an illegal instruction,
which has file/line information associated with it. This extra information is
stored in the __bug_table section in the ELF file.
When the kernel gets an illegal instruction, it first confirms it might
possibly be from a BUG (ie, in kernel mode, the right illegal instruction).
It then calls report_bug(). This searches __bug_table for a matching
instruction pointer, and if found, prints the corresponding file/line
information. If report_bug() determines that it wasn't a BUG which caused the
trap, it returns BUG_TRAP_TYPE_NONE.
Some architectures (powerpc) implement WARN using the same mechanism; if the
illegal instruction was the result of a WARN, then report_bug(Q) returns
CONFIG_DEBUG_BUGVERBOSE; otherwise it returns BUG_TRAP_TYPE_BUG.
lib/bug.c keeps a list of loaded modules which can be searched for __bug_table
entries. The architecture must call
module_bug_finalize()/module_bug_cleanup() from its corresponding
module_finalize/cleanup functions.
Unsetting CONFIG_DEBUG_BUGVERBOSE will reduce the kernel size by some amount.
At the very least, filename and line information will not be recorded for each
but, but architectures may decide to store no extra information per BUG at
all.
Unfortunately, gcc doesn't have a general way to mark an asm() as noreturn, so
architectures will generally have to include an infinite loop (or similar) in
the BUG code, so that gcc knows execution won't continue beyond that point.
gcc does have a __builtin_trap() operator which may be useful to achieve the
same effect, unfortunately it cannot be used to actually implement the BUG
itself, because there's no way to get the instruction's address for use in
generating the __bug_table entry.
[randy.dunlap@oracle.com: Handle BUG=n, GENERIC_BUG=n to prevent build errors]
[bunk@stusta.de: include/linux/bug.h must always #include <linux/module.h]
Signed-off-by: Jeremy Fitzhardinge <jeremy@goop.org>
Cc: Andi Kleen <ak@muc.de>
Cc: Hugh Dickens <hugh@veritas.com>
Cc: Michael Ellerman <michael@ellerman.id.au>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Rusty Russell <rusty@rustcorp.com.au>
Signed-off-by: Adrian Bunk <bunk@stusta.de>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-12-08 18:36:19 +08:00
|
|
|
}
|
2008-05-12 21:44:41 +08:00
|
|
|
#else
|
|
|
|
#define BUG_TABLE
|
|
|
|
#endif
|
[PATCH] Generic BUG implementation
This patch adds common handling for kernel BUGs, for use by architectures as
they wish. The code is derived from arch/powerpc.
The advantages of having common BUG handling are:
- consistent BUG reporting across architectures
- shared implementation of out-of-line file/line data
- implement CONFIG_DEBUG_BUGVERBOSE consistently
This means that in inline impact of BUG is just the illegal instruction
itself, which is an improvement for i386 and x86-64.
A BUG is represented in the instruction stream as an illegal instruction,
which has file/line information associated with it. This extra information is
stored in the __bug_table section in the ELF file.
When the kernel gets an illegal instruction, it first confirms it might
possibly be from a BUG (ie, in kernel mode, the right illegal instruction).
It then calls report_bug(). This searches __bug_table for a matching
instruction pointer, and if found, prints the corresponding file/line
information. If report_bug() determines that it wasn't a BUG which caused the
trap, it returns BUG_TRAP_TYPE_NONE.
Some architectures (powerpc) implement WARN using the same mechanism; if the
illegal instruction was the result of a WARN, then report_bug(Q) returns
CONFIG_DEBUG_BUGVERBOSE; otherwise it returns BUG_TRAP_TYPE_BUG.
lib/bug.c keeps a list of loaded modules which can be searched for __bug_table
entries. The architecture must call
module_bug_finalize()/module_bug_cleanup() from its corresponding
module_finalize/cleanup functions.
Unsetting CONFIG_DEBUG_BUGVERBOSE will reduce the kernel size by some amount.
At the very least, filename and line information will not be recorded for each
but, but architectures may decide to store no extra information per BUG at
all.
Unfortunately, gcc doesn't have a general way to mark an asm() as noreturn, so
architectures will generally have to include an infinite loop (or similar) in
the BUG code, so that gcc knows execution won't continue beyond that point.
gcc does have a __builtin_trap() operator which may be useful to achieve the
same effect, unfortunately it cannot be used to actually implement the BUG
itself, because there's no way to get the instruction's address for use in
generating the __bug_table entry.
[randy.dunlap@oracle.com: Handle BUG=n, GENERIC_BUG=n to prevent build errors]
[bunk@stusta.de: include/linux/bug.h must always #include <linux/module.h]
Signed-off-by: Jeremy Fitzhardinge <jeremy@goop.org>
Cc: Andi Kleen <ak@muc.de>
Cc: Hugh Dickens <hugh@veritas.com>
Cc: Michael Ellerman <michael@ellerman.id.au>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Rusty Russell <rusty@rustcorp.com.au>
Signed-off-by: Adrian Bunk <bunk@stusta.de>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
2006-12-08 18:36:19 +08:00
|
|
|
|
2017-10-14 04:02:00 +08:00
|
|
|
#ifdef CONFIG_UNWINDER_ORC
|
2017-07-25 07:36:57 +08:00
|
|
|
#define ORC_UNWIND_TABLE \
|
|
|
|
. = ALIGN(4); \
|
|
|
|
.orc_unwind_ip : AT(ADDR(.orc_unwind_ip) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start_orc_unwind_ip = .; \
|
2017-07-25 07:36:57 +08:00
|
|
|
KEEP(*(.orc_unwind_ip)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop_orc_unwind_ip = .; \
|
2017-07-25 07:36:57 +08:00
|
|
|
} \
|
2019-03-07 01:07:24 +08:00
|
|
|
. = ALIGN(2); \
|
2017-07-25 07:36:57 +08:00
|
|
|
.orc_unwind : AT(ADDR(.orc_unwind) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start_orc_unwind = .; \
|
2017-07-25 07:36:57 +08:00
|
|
|
KEEP(*(.orc_unwind)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop_orc_unwind = .; \
|
2017-07-25 07:36:57 +08:00
|
|
|
} \
|
|
|
|
. = ALIGN(4); \
|
|
|
|
.orc_lookup : AT(ADDR(.orc_lookup) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
orc_lookup = .; \
|
2017-07-25 07:36:57 +08:00
|
|
|
. += (((SIZEOF(.text) + LOOKUP_BLOCK_SIZE - 1) / \
|
|
|
|
LOOKUP_BLOCK_SIZE) + 1) * 4; \
|
2018-05-09 15:23:51 +08:00
|
|
|
orc_lookup_end = .; \
|
2017-07-25 07:36:57 +08:00
|
|
|
}
|
|
|
|
#else
|
|
|
|
#define ORC_UNWIND_TABLE
|
|
|
|
#endif
|
|
|
|
|
2008-05-12 21:44:41 +08:00
|
|
|
#ifdef CONFIG_PM_TRACE
|
|
|
|
#define TRACEDATA \
|
|
|
|
. = ALIGN(4); \
|
|
|
|
.tracedata : AT(ADDR(.tracedata) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__tracedata_start = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(.tracedata)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__tracedata_end = .; \
|
2008-05-12 21:44:41 +08:00
|
|
|
}
|
|
|
|
#else
|
|
|
|
#define TRACEDATA
|
|
|
|
#endif
|
|
|
|
|
2006-09-26 14:32:26 +08:00
|
|
|
#define NOTES \
|
2007-07-19 16:48:36 +08:00
|
|
|
.notes : AT(ADDR(.notes) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__start_notes = .; \
|
2018-05-09 20:59:58 +08:00
|
|
|
KEEP(*(.note.*)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__stop_notes = .; \
|
2019-10-30 05:13:31 +08:00
|
|
|
} NOTES_HEADERS \
|
|
|
|
NOTES_HEADERS_RESTORE
|
2006-10-28 02:41:44 +08:00
|
|
|
|
2009-06-08 02:46:37 +08:00
|
|
|
#define INIT_SETUP(initsetup_align) \
|
|
|
|
. = ALIGN(initsetup_align); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__setup_start = .; \
|
2016-11-24 00:41:41 +08:00
|
|
|
KEEP(*(.init.setup)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__setup_end = .;
|
2009-06-08 02:46:37 +08:00
|
|
|
|
2012-03-26 10:20:51 +08:00
|
|
|
#define INIT_CALLS_LEVEL(level) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__initcall##level##_start = .; \
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
KEEP(*(.initcall##level##.init)) \
|
|
|
|
KEEP(*(.initcall##level##s.init)) \
|
2006-10-28 02:41:44 +08:00
|
|
|
|
2009-06-08 02:46:37 +08:00
|
|
|
#define INIT_CALLS \
|
2018-05-09 15:23:51 +08:00
|
|
|
__initcall_start = .; \
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
KEEP(*(.initcallearly.init)) \
|
2012-03-26 10:20:51 +08:00
|
|
|
INIT_CALLS_LEVEL(0) \
|
|
|
|
INIT_CALLS_LEVEL(1) \
|
|
|
|
INIT_CALLS_LEVEL(2) \
|
|
|
|
INIT_CALLS_LEVEL(3) \
|
|
|
|
INIT_CALLS_LEVEL(4) \
|
|
|
|
INIT_CALLS_LEVEL(5) \
|
|
|
|
INIT_CALLS_LEVEL(rootfs) \
|
|
|
|
INIT_CALLS_LEVEL(6) \
|
|
|
|
INIT_CALLS_LEVEL(7) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__initcall_end = .;
|
2009-06-08 02:46:37 +08:00
|
|
|
|
|
|
|
#define CON_INITCALL \
|
2018-05-09 15:23:51 +08:00
|
|
|
__con_initcall_start = .; \
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
KEEP(*(.con_initcall.init)) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__con_initcall_end = .;
|
2009-06-08 02:46:37 +08:00
|
|
|
|
|
|
|
#ifdef CONFIG_BLK_DEV_INITRD
|
|
|
|
#define INIT_RAM_FS \
|
2010-10-27 05:22:30 +08:00
|
|
|
. = ALIGN(4); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__initramfs_start = .; \
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
KEEP(*(.init.ramfs)) \
|
initramfs: fix initramfs size calculation
The size of a built-in initramfs is calculated in init/initramfs.c by
"__initramfs_end - __initramfs_start". Those symbols are defined in the
linker script include/asm-generic/vmlinux.lds.h:
#define INIT_RAM_FS \
. = ALIGN(PAGE_SIZE); \
VMLINUX_SYMBOL(__initramfs_start) = .; \
*(.init.ramfs) \
VMLINUX_SYMBOL(__initramfs_end) = .;
If the initramfs file has an odd number of bytes, the "__initramfs_end"
symbol points to an odd address, for example, the symbols in the
System.map might look like:
0000000000572000 T __initramfs_start
00000000005bcd05 T __initramfs_end <-- odd address
At least on s390 this causes a problem:
Certain s390 instructions, especially instructions for loading addresses
(larl) or branch addresses must be on even addresses. The compiler loads
the symbol addresses with the "larl" instruction. This instruction sets
the last bit to 0 and, therefore, for odd size files, the calculated size
is one byte less than it should be:
0000000000540a9c <populate_rootfs>:
540a9c: eb cf f0 78 00 24 stmg %r12,%r15,120(%r15),
540aa2: c0 10 00 01 8a af larl %r1,572000 <__initramfs_start>
540aa8: c0 c0 00 03 e1 2e larl %r12,5bcd04 <initramfs_end>
(Instead of 5bcd05)
...
540abe: 1b c1 sr %r12,%r1
To fix the problem, this patch introduces the global variable
__initramfs_size, which is calculated in the "usr/initramfs_data.S" file.
The populate_rootfs() function can then use the start marker of the
.init.ramfs section and the value of __initramfs_size for loading the
initramfs. Because the start marker and size is sufficient, the
__initramfs_end symbol is no longer needed and is removed.
Signed-off-by: Michael Holzheu <holzheu@linux.vnet.ibm.com>
Signed-off-by: Hendrik Brueckner <brueckner@linux.vnet.ibm.com>
Reviewed-by: WANG Cong <xiyou.wangcong@gmail.com>
Acked-by: Michal Marek <mmarek@suse.cz>
Acked-by: "H. Peter Anvin" <hpa@zytor.com>
Cc: Heiko Carstens <heiko.carstens@de.ibm.com>
Cc: Martin Schwidefsky <schwidefsky@de.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Michal Marek <mmarek@suse.cz>
2010-09-18 06:24:11 +08:00
|
|
|
. = ALIGN(8); \
|
kbuild: allow archs to select link dead code/data elimination
Introduce LD_DEAD_CODE_DATA_ELIMINATION option for architectures to
select to build with -ffunction-sections, -fdata-sections, and link
with --gc-sections. It requires some work (documented) to ensure all
unreferenced entrypoints are live, and requires toolchain and build
verification, so it is made a per-arch option for now.
On a random powerpc64le build, this yelds a significant size saving,
it boots and runs fine, but there is a lot I haven't tested as yet, so
these savings may be reduced if there are bugs in the link.
text data bss dec filename
11169741 1180744 1923176 14273661 vmlinux
10445269 1004127 1919707 13369103 vmlinux.dce
~700K text, ~170K data, 6% removed from kernel image size.
Signed-off-by: Nicholas Piggin <npiggin@gmail.com>
Signed-off-by: Michal Marek <mmarek@suse.com>
2016-08-24 20:29:20 +08:00
|
|
|
KEEP(*(.init.ramfs.info))
|
2009-06-08 02:46:37 +08:00
|
|
|
#else
|
2009-06-22 22:32:31 +08:00
|
|
|
#define INIT_RAM_FS
|
2009-06-08 02:46:37 +08:00
|
|
|
#endif
|
|
|
|
|
2017-10-20 22:30:57 +08:00
|
|
|
/*
|
|
|
|
* Memory encryption operates on a page basis. Since we need to clear
|
|
|
|
* the memory encryption mask for this section, it needs to be aligned
|
|
|
|
* on a page boundary and be a page-size multiple in length.
|
|
|
|
*
|
|
|
|
* Note: We use a separate section so that only this section gets
|
|
|
|
* decrypted to avoid exposing more than we wish.
|
|
|
|
*/
|
|
|
|
#ifdef CONFIG_AMD_MEM_ENCRYPT
|
|
|
|
#define PERCPU_DECRYPTED_SECTION \
|
|
|
|
. = ALIGN(PAGE_SIZE); \
|
|
|
|
*(.data..percpu..decrypted) \
|
|
|
|
. = ALIGN(PAGE_SIZE);
|
|
|
|
#else
|
|
|
|
#define PERCPU_DECRYPTED_SECTION
|
|
|
|
#endif
|
|
|
|
|
|
|
|
|
linker script: unify usage of discard definition
Discarded sections in different archs share some commonality but have
considerable differences. This led to linker script for each arch
implementing its own /DISCARD/ definition, which makes maintaining
tedious and adding new entries error-prone.
This patch makes all linker scripts to move discard definitions to the
end of the linker script and use the common DISCARDS macro. As ld
uses the first matching section definition, archs can include default
discarded sections by including them earlier in the linker script.
ia64 is notable because it first throws away some ia64 specific
subsections and then include the rest of the sections into the final
image, so those sections must be discarded before the inclusion.
defconfig compile tested for x86, x86-64, powerpc, powerpc64, ia64,
alpha, sparc, sparc64 and s390. Michal Simek tested microblaze.
Signed-off-by: Tejun Heo <tj@kernel.org>
Acked-by: Paul Mundt <lethal@linux-sh.org>
Acked-by: Mike Frysinger <vapier@gentoo.org>
Tested-by: Michal Simek <monstr@monstr.eu>
Cc: linux-arch@vger.kernel.org
Cc: Michal Simek <monstr@monstr.eu>
Cc: microblaze-uclinux@itee.uq.edu.au
Cc: Sam Ravnborg <sam@ravnborg.org>
Cc: Tony Luck <tony.luck@intel.com>
2009-07-09 10:27:40 +08:00
|
|
|
/*
|
|
|
|
* Default discarded sections.
|
|
|
|
*
|
|
|
|
* Some archs want to discard exit text/data at runtime rather than
|
|
|
|
* link time due to cross-section references such as alt instructions,
|
|
|
|
* bug table, eh_frame, etc. DISCARDS must be the last of output
|
|
|
|
* section definitions so that such archs put those in earlier section
|
|
|
|
* definitions.
|
|
|
|
*/
|
2009-06-24 14:13:38 +08:00
|
|
|
#define DISCARDS \
|
|
|
|
/DISCARD/ : { \
|
|
|
|
EXIT_TEXT \
|
|
|
|
EXIT_DATA \
|
linker script: unify usage of discard definition
Discarded sections in different archs share some commonality but have
considerable differences. This led to linker script for each arch
implementing its own /DISCARD/ definition, which makes maintaining
tedious and adding new entries error-prone.
This patch makes all linker scripts to move discard definitions to the
end of the linker script and use the common DISCARDS macro. As ld
uses the first matching section definition, archs can include default
discarded sections by including them earlier in the linker script.
ia64 is notable because it first throws away some ia64 specific
subsections and then include the rest of the sections into the final
image, so those sections must be discarded before the inclusion.
defconfig compile tested for x86, x86-64, powerpc, powerpc64, ia64,
alpha, sparc, sparc64 and s390. Michal Simek tested microblaze.
Signed-off-by: Tejun Heo <tj@kernel.org>
Acked-by: Paul Mundt <lethal@linux-sh.org>
Acked-by: Mike Frysinger <vapier@gentoo.org>
Tested-by: Michal Simek <monstr@monstr.eu>
Cc: linux-arch@vger.kernel.org
Cc: Michal Simek <monstr@monstr.eu>
Cc: microblaze-uclinux@itee.uq.edu.au
Cc: Sam Ravnborg <sam@ravnborg.org>
Cc: Tony Luck <tony.luck@intel.com>
2009-07-09 10:27:40 +08:00
|
|
|
EXIT_CALL \
|
2009-06-24 14:13:38 +08:00
|
|
|
*(.discard) \
|
2010-07-23 13:58:01 +08:00
|
|
|
*(.discard.*) \
|
2019-04-30 00:11:14 +08:00
|
|
|
*(.modinfo) \
|
2009-06-24 14:13:38 +08:00
|
|
|
}
|
|
|
|
|
2011-04-04 07:41:32 +08:00
|
|
|
/**
|
|
|
|
* PERCPU_INPUT - the percpu input sections
|
|
|
|
* @cacheline: cacheline size
|
|
|
|
*
|
|
|
|
* The core percpu section names and core symbols which do not rely
|
|
|
|
* directly upon load addresses.
|
|
|
|
*
|
|
|
|
* @cacheline is used to align subsections to avoid false cacheline
|
|
|
|
* sharing between subsections for different purposes.
|
|
|
|
*/
|
|
|
|
#define PERCPU_INPUT(cacheline) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__per_cpu_start = .; \
|
2011-04-04 07:41:32 +08:00
|
|
|
*(.data..percpu..first) \
|
|
|
|
. = ALIGN(PAGE_SIZE); \
|
|
|
|
*(.data..percpu..page_aligned) \
|
|
|
|
. = ALIGN(cacheline); \
|
2014-07-02 03:11:47 +08:00
|
|
|
*(.data..percpu..read_mostly) \
|
2011-04-04 07:41:32 +08:00
|
|
|
. = ALIGN(cacheline); \
|
|
|
|
*(.data..percpu) \
|
|
|
|
*(.data..percpu..shared_aligned) \
|
2017-10-20 22:30:57 +08:00
|
|
|
PERCPU_DECRYPTED_SECTION \
|
2018-05-09 15:23:51 +08:00
|
|
|
__per_cpu_end = .;
|
2011-04-04 07:41:32 +08:00
|
|
|
|
2009-01-13 19:41:35 +08:00
|
|
|
/**
|
2009-01-19 11:21:28 +08:00
|
|
|
* PERCPU_VADDR - define output section for percpu area
|
2011-01-25 21:26:50 +08:00
|
|
|
* @cacheline: cacheline size
|
2009-01-13 19:41:35 +08:00
|
|
|
* @vaddr: explicit base address (optional)
|
|
|
|
* @phdr: destination PHDR (optional)
|
|
|
|
*
|
2011-01-25 21:26:50 +08:00
|
|
|
* Macro which expands to output section for percpu area.
|
|
|
|
*
|
|
|
|
* @cacheline is used to align subsections to avoid false cacheline
|
|
|
|
* sharing between subsections for different purposes.
|
|
|
|
*
|
|
|
|
* If @vaddr is not blank, it specifies explicit base address and all
|
|
|
|
* percpu symbols will be offset from the given address. If blank,
|
|
|
|
* @vaddr always equals @laddr + LOAD_OFFSET.
|
2009-01-13 19:41:35 +08:00
|
|
|
*
|
|
|
|
* @phdr defines the output PHDR to use if not blank. Be warned that
|
|
|
|
* output PHDR is sticky. If @phdr is specified, the next output
|
|
|
|
* section in the linker script will go there too. @phdr should have
|
|
|
|
* a leading colon.
|
|
|
|
*
|
2009-01-30 15:32:22 +08:00
|
|
|
* Note that this macros defines __per_cpu_load as an absolute symbol.
|
|
|
|
* If there is no need to put the percpu section at a predetermined
|
2011-03-25 01:50:09 +08:00
|
|
|
* address, use PERCPU_SECTION.
|
2009-01-13 19:41:35 +08:00
|
|
|
*/
|
2011-01-25 21:26:50 +08:00
|
|
|
#define PERCPU_VADDR(cacheline, vaddr, phdr) \
|
2018-05-09 15:23:51 +08:00
|
|
|
__per_cpu_load = .; \
|
|
|
|
.data..percpu vaddr : AT(__per_cpu_load - LOAD_OFFSET) { \
|
2011-04-04 07:41:32 +08:00
|
|
|
PERCPU_INPUT(cacheline) \
|
2009-01-19 11:21:28 +08:00
|
|
|
} phdr \
|
2018-05-09 15:23:51 +08:00
|
|
|
. = __per_cpu_load + SIZEOF(.data..percpu);
|
2009-01-13 19:41:35 +08:00
|
|
|
|
|
|
|
/**
|
2011-03-25 01:50:09 +08:00
|
|
|
* PERCPU_SECTION - define output section for percpu area, simple version
|
2011-01-25 21:26:50 +08:00
|
|
|
* @cacheline: cacheline size
|
2009-01-13 19:41:35 +08:00
|
|
|
*
|
2011-03-25 01:50:09 +08:00
|
|
|
* Align to PAGE_SIZE and outputs output section for percpu area. This
|
|
|
|
* macro doesn't manipulate @vaddr or @phdr and __per_cpu_load and
|
2009-01-13 19:41:35 +08:00
|
|
|
* __per_cpu_start will be identical.
|
2009-01-30 15:32:22 +08:00
|
|
|
*
|
2011-03-25 01:50:09 +08:00
|
|
|
* This macro is equivalent to ALIGN(PAGE_SIZE); PERCPU_VADDR(@cacheline,,)
|
2011-01-25 21:26:50 +08:00
|
|
|
* except that __per_cpu_load is defined as a relative symbol against
|
|
|
|
* .data..percpu which is required for relocatable x86_32 configuration.
|
2009-01-13 19:41:35 +08:00
|
|
|
*/
|
2011-03-25 01:50:09 +08:00
|
|
|
#define PERCPU_SECTION(cacheline) \
|
|
|
|
. = ALIGN(PAGE_SIZE); \
|
2010-02-20 08:03:43 +08:00
|
|
|
.data..percpu : AT(ADDR(.data..percpu) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
__per_cpu_load = .; \
|
2011-04-04 07:41:32 +08:00
|
|
|
PERCPU_INPUT(cacheline) \
|
2009-01-30 15:32:22 +08:00
|
|
|
}
|
2009-06-08 02:46:37 +08:00
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Definition of the high level *_SECTION macros
|
|
|
|
* They will fit only a subset of the architectures
|
|
|
|
*/
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Writeable data.
|
|
|
|
* All sections are combined in a single .data section.
|
|
|
|
* The sections following CONSTRUCTORS are arranged so their
|
|
|
|
* typical alignment matches.
|
|
|
|
* A cacheline is typical/always less than a PAGE_SIZE so
|
|
|
|
* the sections that has this restriction (or similar)
|
|
|
|
* is located before the ones requiring PAGE_SIZE alignment.
|
|
|
|
* NOSAVE_DATA starts and ends with a PAGE_SIZE alignment which
|
2011-03-31 09:57:33 +08:00
|
|
|
* matches the requirement of PAGE_ALIGNED_DATA.
|
2009-06-08 02:46:37 +08:00
|
|
|
*
|
2009-06-15 04:10:41 +08:00
|
|
|
* use 0 as page_align if page_aligned data is not used */
|
2009-06-24 00:04:36 +08:00
|
|
|
#define RW_DATA_SECTION(cacheline, pagealigned, inittask) \
|
2009-06-08 02:46:37 +08:00
|
|
|
. = ALIGN(PAGE_SIZE); \
|
|
|
|
.data : AT(ADDR(.data) - LOAD_OFFSET) { \
|
2009-06-24 06:53:15 +08:00
|
|
|
INIT_TASK_DATA(inittask) \
|
2009-09-24 22:36:16 +08:00
|
|
|
NOSAVE_DATA \
|
|
|
|
PAGE_ALIGNED_DATA(pagealigned) \
|
2009-06-08 02:46:37 +08:00
|
|
|
CACHELINE_ALIGNED_DATA(cacheline) \
|
|
|
|
READ_MOSTLY_DATA(cacheline) \
|
|
|
|
DATA_DATA \
|
|
|
|
CONSTRUCTORS \
|
2017-02-25 15:56:53 +08:00
|
|
|
} \
|
2017-07-25 07:36:57 +08:00
|
|
|
BUG_TABLE \
|
2009-06-08 02:46:37 +08:00
|
|
|
|
|
|
|
#define INIT_TEXT_SECTION(inittext_align) \
|
|
|
|
. = ALIGN(inittext_align); \
|
|
|
|
.init.text : AT(ADDR(.init.text) - LOAD_OFFSET) { \
|
2018-05-09 15:23:51 +08:00
|
|
|
_sinittext = .; \
|
2009-06-08 02:46:37 +08:00
|
|
|
INIT_TEXT \
|
2018-05-09 15:23:51 +08:00
|
|
|
_einittext = .; \
|
2009-06-08 02:46:37 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
#define INIT_DATA_SECTION(initsetup_align) \
|
|
|
|
.init.data : AT(ADDR(.init.data) - LOAD_OFFSET) { \
|
|
|
|
INIT_DATA \
|
|
|
|
INIT_SETUP(initsetup_align) \
|
|
|
|
INIT_CALLS \
|
|
|
|
CON_INITCALL \
|
|
|
|
INIT_RAM_FS \
|
|
|
|
}
|
|
|
|
|
2009-07-13 06:23:33 +08:00
|
|
|
#define BSS_SECTION(sbss_align, bss_align, stop_align) \
|
|
|
|
. = ALIGN(sbss_align); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__bss_start = .; \
|
2009-07-13 06:23:33 +08:00
|
|
|
SBSS(sbss_align) \
|
2009-06-08 02:46:37 +08:00
|
|
|
BSS(bss_align) \
|
2009-07-13 06:23:33 +08:00
|
|
|
. = ALIGN(stop_align); \
|
2018-05-09 15:23:51 +08:00
|
|
|
__bss_stop = .;
|