2019-06-01 16:08:42 +08:00
|
|
|
// SPDX-License-Identifier: GPL-2.0-only
|
2013-02-08 23:37:06 +08:00
|
|
|
/*
|
|
|
|
* efi.c - EFI subsystem
|
|
|
|
*
|
|
|
|
* Copyright (C) 2001,2003,2004 Dell <Matt_Domsch@dell.com>
|
|
|
|
* Copyright (C) 2004 Intel Corporation <matthew.e.tolentino@intel.com>
|
|
|
|
* Copyright (C) 2013 Tom Gundersen <teg@jklm.no>
|
|
|
|
*
|
|
|
|
* This code registers /sys/firmware/efi{,/efivars} when EFI is supported,
|
|
|
|
* allowing the efivarfs to be mounted or the efivars module to be loaded.
|
|
|
|
* The existance of /sys/firmware/efi may also be used by userspace to
|
|
|
|
* determine that the system supports EFI.
|
|
|
|
*/
|
|
|
|
|
2013-09-05 18:34:54 +08:00
|
|
|
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
|
|
|
|
|
2013-02-08 23:37:06 +08:00
|
|
|
#include <linux/kobject.h>
|
|
|
|
#include <linux/module.h>
|
|
|
|
#include <linux/init.h>
|
2020-01-16 00:35:45 +08:00
|
|
|
#include <linux/debugfs.h>
|
2013-02-08 23:37:06 +08:00
|
|
|
#include <linux/device.h>
|
|
|
|
#include <linux/efi.h>
|
2013-12-31 01:12:12 +08:00
|
|
|
#include <linux/of.h>
|
2013-09-05 18:34:54 +08:00
|
|
|
#include <linux/io.h>
|
2016-11-13 05:32:31 +08:00
|
|
|
#include <linux/kexec.h>
|
2014-07-09 18:39:29 +08:00
|
|
|
#include <linux/platform_device.h>
|
2016-11-13 05:32:31 +08:00
|
|
|
#include <linux/random.h>
|
|
|
|
#include <linux/reboot.h>
|
2016-07-09 00:13:12 +08:00
|
|
|
#include <linux/slab.h>
|
|
|
|
#include <linux/acpi.h>
|
|
|
|
#include <linux/ucs2_string.h>
|
2016-03-01 05:22:52 +08:00
|
|
|
#include <linux/memblock.h>
|
2019-08-20 08:18:04 +08:00
|
|
|
#include <linux/security.h>
|
2013-09-05 18:34:54 +08:00
|
|
|
|
2016-01-12 21:22:46 +08:00
|
|
|
#include <asm/early_ioremap.h>
|
2015-11-30 20:28:19 +08:00
|
|
|
|
2013-09-05 18:34:54 +08:00
|
|
|
struct efi __read_mostly efi = {
|
2020-01-21 18:17:47 +08:00
|
|
|
.runtime_supported_mask = EFI_RT_SUPPORTED_ALL,
|
2015-09-09 16:08:15 +08:00
|
|
|
.acpi = EFI_INVALID_TABLE_ADDR,
|
|
|
|
.acpi20 = EFI_INVALID_TABLE_ADDR,
|
|
|
|
.smbios = EFI_INVALID_TABLE_ADDR,
|
|
|
|
.smbios3 = EFI_INVALID_TABLE_ADDR,
|
|
|
|
.esrt = EFI_INVALID_TABLE_ADDR,
|
2018-09-22 00:32:44 +08:00
|
|
|
.tpm_log = EFI_INVALID_TABLE_ADDR,
|
2019-05-21 04:54:59 +08:00
|
|
|
.tpm_final_log = EFI_INVALID_TABLE_ADDR,
|
2020-09-05 09:31:05 +08:00
|
|
|
#ifdef CONFIG_LOAD_UEFI_KEYS
|
|
|
|
.mokvar_table = EFI_INVALID_TABLE_ADDR,
|
|
|
|
#endif
|
2013-09-05 18:34:54 +08:00
|
|
|
};
|
|
|
|
EXPORT_SYMBOL(efi);
|
2013-02-08 23:37:06 +08:00
|
|
|
|
2020-02-28 20:14:04 +08:00
|
|
|
unsigned long __ro_after_init efi_rng_seed = EFI_INVALID_TABLE_ADDR;
|
2020-01-22 22:06:54 +08:00
|
|
|
static unsigned long __initdata mem_reserve = EFI_INVALID_TABLE_ADDR;
|
2020-01-23 20:10:25 +08:00
|
|
|
static unsigned long __initdata rt_prop = EFI_INVALID_TABLE_ADDR;
|
2020-01-22 21:58:15 +08:00
|
|
|
|
2018-03-12 16:44:56 +08:00
|
|
|
struct mm_struct efi_mm = {
|
|
|
|
.mm_rb = RB_ROOT,
|
|
|
|
.mm_users = ATOMIC_INIT(2),
|
|
|
|
.mm_count = ATOMIC_INIT(1),
|
2020-12-15 11:05:44 +08:00
|
|
|
.write_protect_seq = SEQCNT_ZERO(efi_mm.write_protect_seq),
|
2020-06-09 12:33:40 +08:00
|
|
|
MMAP_LOCK_INITIALIZER(efi_mm)
|
2018-03-12 16:44:56 +08:00
|
|
|
.page_table_lock = __SPIN_LOCK_UNLOCKED(efi_mm.page_table_lock),
|
|
|
|
.mmlist = LIST_HEAD_INIT(efi_mm.mmlist),
|
2018-07-17 03:03:31 +08:00
|
|
|
.cpu_bitmap = { [BITS_TO_LONGS(NR_CPUS)] = 0},
|
2018-03-12 16:44:56 +08:00
|
|
|
};
|
|
|
|
|
efi: Use a work queue to invoke EFI Runtime Services
Presently, when a user process requests the kernel to execute any
UEFI runtime service, the kernel temporarily switches to a separate
set of page tables that describe the virtual mapping of the UEFI
runtime services regions in memory. Since UEFI runtime services are
typically invoked with interrupts enabled, any code that may be called
during this time, will have an incorrect view of the process's address
space. Although it is unusual for code running in interrupt context to
make assumptions about the process context it runs in, there are cases
(such as the perf subsystem taking samples) where this causes problems.
So let's set up a work queue for calling UEFI runtime services, so that
the actual calls are made when the work queue items are dispatched by a
work queue worker running in a separate kernel thread. Such threads are
not expected to have userland mappings in the first place, and so the
additional mappings created for the UEFI runtime services can never
clash with any.
The ResetSystem() runtime service is not covered by the work queue
handling, since it is not expected to return, and may be called at a
time when the kernel is torn down to the point where we cannot expect
work queues to still be operational.
The non-blocking variants of SetVariable() and QueryVariableInfo()
are also excluded: these are intended to be used from atomic context,
which obviously rules out waiting for a completion to be signalled by
another thread. Note that these variants are currently only used for
UEFI runtime services calls that occur very early in the boot, and
for ones that occur in critical conditions, e.g., to flush kernel logs
to UEFI variables via efi-pstore.
Suggested-by: Andy Lutomirski <luto@kernel.org>
Signed-off-by: Sai Praneeth Prakhya <sai.praneeth.prakhya@intel.com>
[ardb: exclude ResetSystem() from the workqueue treatment
merge from 2 separate patches and rewrite commit log]
Signed-off-by: Ard Biesheuvel <ard.biesheuvel@linaro.org>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: linux-efi@vger.kernel.org
Link: http://lkml.kernel.org/r/20180711094040.12506-4-ard.biesheuvel@linaro.org
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2018-07-11 17:40:35 +08:00
|
|
|
struct workqueue_struct *efi_rts_wq;
|
|
|
|
|
2021-09-24 21:49:18 +08:00
|
|
|
static bool disable_runtime = IS_ENABLED(CONFIG_PREEMPT_RT);
|
2014-08-14 17:15:26 +08:00
|
|
|
static int __init setup_noefi(char *arg)
|
|
|
|
{
|
|
|
|
disable_runtime = true;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
early_param("noefi", setup_noefi);
|
|
|
|
|
|
|
|
bool efi_runtime_disabled(void)
|
|
|
|
{
|
|
|
|
return disable_runtime;
|
|
|
|
}
|
|
|
|
|
2019-11-07 09:43:11 +08:00
|
|
|
bool __pure __efi_soft_reserve_enabled(void)
|
|
|
|
{
|
|
|
|
return !efi_enabled(EFI_MEM_NO_SOFT_RESERVE);
|
|
|
|
}
|
|
|
|
|
2014-08-14 17:15:28 +08:00
|
|
|
static int __init parse_efi_cmdline(char *str)
|
|
|
|
{
|
2015-07-16 10:36:03 +08:00
|
|
|
if (!str) {
|
|
|
|
pr_warn("need at least one option\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
2015-08-26 21:24:56 +08:00
|
|
|
if (parse_option_str(str, "debug"))
|
|
|
|
set_bit(EFI_DBG, &efi.flags);
|
|
|
|
|
2014-08-14 17:15:28 +08:00
|
|
|
if (parse_option_str(str, "noruntime"))
|
|
|
|
disable_runtime = true;
|
|
|
|
|
2021-09-24 21:49:19 +08:00
|
|
|
if (parse_option_str(str, "runtime"))
|
|
|
|
disable_runtime = false;
|
|
|
|
|
2019-11-07 09:43:11 +08:00
|
|
|
if (parse_option_str(str, "nosoftreserve"))
|
|
|
|
set_bit(EFI_MEM_NO_SOFT_RESERVE, &efi.flags);
|
2014-08-14 17:15:28 +08:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
early_param("efi", parse_efi_cmdline);
|
|
|
|
|
2015-04-29 06:44:31 +08:00
|
|
|
struct kobject *efi_kobj;
|
2013-02-08 23:37:06 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Let's not leave out systab information that snuck into
|
|
|
|
* the efivars driver
|
2017-12-06 17:50:10 +08:00
|
|
|
* Note, do not add more fields in systab sysfs file as it breaks sysfs
|
|
|
|
* one value per file rule!
|
2013-02-08 23:37:06 +08:00
|
|
|
*/
|
|
|
|
static ssize_t systab_show(struct kobject *kobj,
|
|
|
|
struct kobj_attribute *attr, char *buf)
|
|
|
|
{
|
|
|
|
char *str = buf;
|
|
|
|
|
|
|
|
if (!kobj || !buf)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
if (efi.acpi20 != EFI_INVALID_TABLE_ADDR)
|
|
|
|
str += sprintf(str, "ACPI20=0x%lx\n", efi.acpi20);
|
|
|
|
if (efi.acpi != EFI_INVALID_TABLE_ADDR)
|
|
|
|
str += sprintf(str, "ACPI=0x%lx\n", efi.acpi);
|
2015-04-30 21:23:05 +08:00
|
|
|
/*
|
|
|
|
* If both SMBIOS and SMBIOS3 entry points are implemented, the
|
|
|
|
* SMBIOS3 entry point shall be preferred, so we list it first to
|
|
|
|
* let applications stop parsing after the first match.
|
|
|
|
*/
|
2014-10-14 22:34:47 +08:00
|
|
|
if (efi.smbios3 != EFI_INVALID_TABLE_ADDR)
|
|
|
|
str += sprintf(str, "SMBIOS3=0x%lx\n", efi.smbios3);
|
2015-04-30 21:23:05 +08:00
|
|
|
if (efi.smbios != EFI_INVALID_TABLE_ADDR)
|
|
|
|
str += sprintf(str, "SMBIOS=0x%lx\n", efi.smbios);
|
2013-02-08 23:37:06 +08:00
|
|
|
|
2020-05-16 21:26:47 +08:00
|
|
|
if (IS_ENABLED(CONFIG_IA64) || IS_ENABLED(CONFIG_X86))
|
2020-01-19 22:43:53 +08:00
|
|
|
str = efi_systab_show_arch(str);
|
|
|
|
|
2013-02-08 23:37:06 +08:00
|
|
|
return str - buf;
|
|
|
|
}
|
|
|
|
|
2017-12-06 17:50:08 +08:00
|
|
|
static struct kobj_attribute efi_attr_systab = __ATTR_RO_MODE(systab, 0400);
|
2013-02-08 23:37:06 +08:00
|
|
|
|
2015-01-09 23:29:53 +08:00
|
|
|
static ssize_t fw_platform_size_show(struct kobject *kobj,
|
|
|
|
struct kobj_attribute *attr, char *buf)
|
|
|
|
{
|
|
|
|
return sprintf(buf, "%d\n", efi_enabled(EFI_64BIT) ? 64 : 32);
|
|
|
|
}
|
|
|
|
|
2020-01-21 00:23:21 +08:00
|
|
|
extern __weak struct kobj_attribute efi_attr_fw_vendor;
|
|
|
|
extern __weak struct kobj_attribute efi_attr_runtime;
|
|
|
|
extern __weak struct kobj_attribute efi_attr_config_table;
|
2015-01-09 23:29:53 +08:00
|
|
|
static struct kobj_attribute efi_attr_fw_platform_size =
|
|
|
|
__ATTR_RO(fw_platform_size);
|
2013-12-20 18:02:17 +08:00
|
|
|
|
2013-02-08 23:37:06 +08:00
|
|
|
static struct attribute *efi_subsys_attrs[] = {
|
|
|
|
&efi_attr_systab.attr,
|
2020-01-21 00:23:21 +08:00
|
|
|
&efi_attr_fw_platform_size.attr,
|
2013-12-20 18:02:17 +08:00
|
|
|
&efi_attr_fw_vendor.attr,
|
|
|
|
&efi_attr_runtime.attr,
|
|
|
|
&efi_attr_config_table.attr,
|
|
|
|
NULL,
|
2013-02-08 23:37:06 +08:00
|
|
|
};
|
|
|
|
|
2020-01-21 00:23:21 +08:00
|
|
|
umode_t __weak efi_attr_is_visible(struct kobject *kobj, struct attribute *attr,
|
|
|
|
int n)
|
2013-12-20 18:02:17 +08:00
|
|
|
{
|
2014-07-01 01:52:58 +08:00
|
|
|
return attr->mode;
|
2013-12-20 18:02:17 +08:00
|
|
|
}
|
|
|
|
|
2017-08-19 03:49:46 +08:00
|
|
|
static const struct attribute_group efi_subsys_attr_group = {
|
2013-02-08 23:37:06 +08:00
|
|
|
.attrs = efi_subsys_attrs,
|
2013-12-20 18:02:17 +08:00
|
|
|
.is_visible = efi_attr_is_visible,
|
2013-02-08 23:37:06 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
static struct efivars generic_efivars;
|
|
|
|
static struct efivar_operations generic_ops;
|
|
|
|
|
|
|
|
static int generic_ops_register(void)
|
|
|
|
{
|
|
|
|
generic_ops.get_variable = efi.get_variable;
|
|
|
|
generic_ops.get_next_variable = efi.get_next_variable;
|
2013-04-30 18:30:24 +08:00
|
|
|
generic_ops.query_variable_store = efi_query_variable_store;
|
2013-02-08 23:37:06 +08:00
|
|
|
|
2020-07-08 18:01:57 +08:00
|
|
|
if (efi_rt_services_supported(EFI_RT_SUPPORTED_SET_VARIABLE)) {
|
|
|
|
generic_ops.set_variable = efi.set_variable;
|
|
|
|
generic_ops.set_variable_nonblocking = efi.set_variable_nonblocking;
|
|
|
|
}
|
2013-02-08 23:37:06 +08:00
|
|
|
return efivars_register(&generic_efivars, &generic_ops, efi_kobj);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void generic_ops_unregister(void)
|
|
|
|
{
|
|
|
|
efivars_unregister(&generic_efivars);
|
|
|
|
}
|
|
|
|
|
2020-06-16 04:24:08 +08:00
|
|
|
#ifdef CONFIG_EFI_CUSTOM_SSDT_OVERLAYS
|
2016-07-09 00:13:12 +08:00
|
|
|
#define EFIVAR_SSDT_NAME_MAX 16
|
|
|
|
static char efivar_ssdt[EFIVAR_SSDT_NAME_MAX] __initdata;
|
|
|
|
static int __init efivar_ssdt_setup(char *str)
|
|
|
|
{
|
2019-08-20 08:18:04 +08:00
|
|
|
int ret = security_locked_down(LOCKDOWN_ACPI_TABLES);
|
|
|
|
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
2016-07-09 00:13:12 +08:00
|
|
|
if (strlen(str) < sizeof(efivar_ssdt))
|
|
|
|
memcpy(efivar_ssdt, str, strlen(str));
|
|
|
|
else
|
|
|
|
pr_warn("efivar_ssdt: name too long: %s\n", str);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
__setup("efivar_ssdt=", efivar_ssdt_setup);
|
|
|
|
|
|
|
|
static __init int efivar_ssdt_iter(efi_char16_t *name, efi_guid_t vendor,
|
|
|
|
unsigned long name_size, void *data)
|
|
|
|
{
|
|
|
|
struct efivar_entry *entry;
|
|
|
|
struct list_head *list = data;
|
|
|
|
char utf8_name[EFIVAR_SSDT_NAME_MAX];
|
|
|
|
int limit = min_t(unsigned long, EFIVAR_SSDT_NAME_MAX, name_size);
|
|
|
|
|
|
|
|
ucs2_as_utf8(utf8_name, name, limit - 1);
|
|
|
|
if (strncmp(utf8_name, efivar_ssdt, limit) != 0)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
entry = kmalloc(sizeof(*entry), GFP_KERNEL);
|
|
|
|
if (!entry)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
memcpy(entry->var.VariableName, name, name_size);
|
|
|
|
memcpy(&entry->var.VendorGuid, &vendor, sizeof(efi_guid_t));
|
|
|
|
|
|
|
|
efivar_entry_add(entry, list);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static __init int efivar_ssdt_load(void)
|
|
|
|
{
|
|
|
|
LIST_HEAD(entries);
|
|
|
|
struct efivar_entry *entry, *aux;
|
|
|
|
unsigned long size;
|
|
|
|
void *data;
|
|
|
|
int ret;
|
|
|
|
|
2019-10-03 00:58:59 +08:00
|
|
|
if (!efivar_ssdt[0])
|
|
|
|
return 0;
|
|
|
|
|
2016-07-09 00:13:12 +08:00
|
|
|
ret = efivar_init(efivar_ssdt_iter, &entries, true, &entries);
|
|
|
|
|
|
|
|
list_for_each_entry_safe(entry, aux, &entries, list) {
|
|
|
|
pr_info("loading SSDT from variable %s-%pUl\n", efivar_ssdt,
|
|
|
|
&entry->var.VendorGuid);
|
|
|
|
|
|
|
|
list_del(&entry->list);
|
|
|
|
|
|
|
|
ret = efivar_entry_size(entry, &size);
|
|
|
|
if (ret) {
|
|
|
|
pr_err("failed to get var size\n");
|
|
|
|
goto free_entry;
|
|
|
|
}
|
|
|
|
|
|
|
|
data = kmalloc(size, GFP_KERNEL);
|
2016-10-18 22:33:18 +08:00
|
|
|
if (!data) {
|
|
|
|
ret = -ENOMEM;
|
2016-07-09 00:13:12 +08:00
|
|
|
goto free_entry;
|
2016-10-18 22:33:18 +08:00
|
|
|
}
|
2016-07-09 00:13:12 +08:00
|
|
|
|
|
|
|
ret = efivar_entry_get(entry, NULL, &size, data);
|
|
|
|
if (ret) {
|
|
|
|
pr_err("failed to get var data\n");
|
|
|
|
goto free_data;
|
|
|
|
}
|
|
|
|
|
2019-10-26 05:36:53 +08:00
|
|
|
ret = acpi_load_table(data, NULL);
|
2016-07-09 00:13:12 +08:00
|
|
|
if (ret) {
|
|
|
|
pr_err("failed to load table: %d\n", ret);
|
|
|
|
goto free_data;
|
|
|
|
}
|
|
|
|
|
|
|
|
goto free_entry;
|
|
|
|
|
|
|
|
free_data:
|
|
|
|
kfree(data);
|
|
|
|
|
|
|
|
free_entry:
|
|
|
|
kfree(entry);
|
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
static inline int efivar_ssdt_load(void) { return 0; }
|
|
|
|
#endif
|
|
|
|
|
2020-01-16 00:35:45 +08:00
|
|
|
#ifdef CONFIG_DEBUG_FS
|
|
|
|
|
|
|
|
#define EFI_DEBUGFS_MAX_BLOBS 32
|
|
|
|
|
|
|
|
static struct debugfs_blob_wrapper debugfs_blob[EFI_DEBUGFS_MAX_BLOBS];
|
|
|
|
|
|
|
|
static void __init efi_debugfs_init(void)
|
|
|
|
{
|
|
|
|
struct dentry *efi_debugfs;
|
|
|
|
efi_memory_desc_t *md;
|
|
|
|
char name[32];
|
|
|
|
int type_count[EFI_BOOT_SERVICES_DATA + 1] = {};
|
|
|
|
int i = 0;
|
|
|
|
|
|
|
|
efi_debugfs = debugfs_create_dir("efi", NULL);
|
|
|
|
if (IS_ERR_OR_NULL(efi_debugfs))
|
|
|
|
return;
|
|
|
|
|
|
|
|
for_each_efi_memory_desc(md) {
|
|
|
|
switch (md->type) {
|
|
|
|
case EFI_BOOT_SERVICES_CODE:
|
|
|
|
snprintf(name, sizeof(name), "boot_services_code%d",
|
|
|
|
type_count[md->type]++);
|
|
|
|
break;
|
|
|
|
case EFI_BOOT_SERVICES_DATA:
|
|
|
|
snprintf(name, sizeof(name), "boot_services_data%d",
|
|
|
|
type_count[md->type]++);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (i >= EFI_DEBUGFS_MAX_BLOBS) {
|
|
|
|
pr_warn("More then %d EFI boot service segments, only showing first %d in debugfs\n",
|
|
|
|
EFI_DEBUGFS_MAX_BLOBS, EFI_DEBUGFS_MAX_BLOBS);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
debugfs_blob[i].size = md->num_pages << EFI_PAGE_SHIFT;
|
|
|
|
debugfs_blob[i].data = memremap(md->phys_addr,
|
|
|
|
debugfs_blob[i].size,
|
|
|
|
MEMREMAP_WB);
|
|
|
|
if (!debugfs_blob[i].data)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
debugfs_create_blob(name, 0400, efi_debugfs, &debugfs_blob[i]);
|
|
|
|
i++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
static inline void efi_debugfs_init(void) {}
|
|
|
|
#endif
|
|
|
|
|
2013-02-08 23:37:06 +08:00
|
|
|
/*
|
|
|
|
* We register the efi subsystem with the firmware subsystem and the
|
|
|
|
* efivars subsystem with the efi subsystem, if the system was booted with
|
|
|
|
* EFI.
|
|
|
|
*/
|
|
|
|
static int __init efisubsys_init(void)
|
|
|
|
{
|
|
|
|
int error;
|
|
|
|
|
2020-01-21 18:17:47 +08:00
|
|
|
if (!efi_enabled(EFI_RUNTIME_SERVICES))
|
|
|
|
efi.runtime_supported_mask = 0;
|
|
|
|
|
2020-02-28 20:14:08 +08:00
|
|
|
if (!efi_enabled(EFI_BOOT))
|
|
|
|
return 0;
|
|
|
|
|
2020-01-21 18:17:47 +08:00
|
|
|
if (efi.runtime_supported_mask) {
|
|
|
|
/*
|
|
|
|
* Since we process only one efi_runtime_service() at a time, an
|
|
|
|
* ordered workqueue (which creates only one execution context)
|
|
|
|
* should suffice for all our needs.
|
|
|
|
*/
|
|
|
|
efi_rts_wq = alloc_ordered_workqueue("efi_rts_wq", 0);
|
|
|
|
if (!efi_rts_wq) {
|
|
|
|
pr_err("Creating efi_rts_wq failed, EFI runtime services disabled.\n");
|
|
|
|
clear_bit(EFI_RUNTIME_SERVICES, &efi.flags);
|
|
|
|
efi.runtime_supported_mask = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
efi: Use a work queue to invoke EFI Runtime Services
Presently, when a user process requests the kernel to execute any
UEFI runtime service, the kernel temporarily switches to a separate
set of page tables that describe the virtual mapping of the UEFI
runtime services regions in memory. Since UEFI runtime services are
typically invoked with interrupts enabled, any code that may be called
during this time, will have an incorrect view of the process's address
space. Although it is unusual for code running in interrupt context to
make assumptions about the process context it runs in, there are cases
(such as the perf subsystem taking samples) where this causes problems.
So let's set up a work queue for calling UEFI runtime services, so that
the actual calls are made when the work queue items are dispatched by a
work queue worker running in a separate kernel thread. Such threads are
not expected to have userland mappings in the first place, and so the
additional mappings created for the UEFI runtime services can never
clash with any.
The ResetSystem() runtime service is not covered by the work queue
handling, since it is not expected to return, and may be called at a
time when the kernel is torn down to the point where we cannot expect
work queues to still be operational.
The non-blocking variants of SetVariable() and QueryVariableInfo()
are also excluded: these are intended to be used from atomic context,
which obviously rules out waiting for a completion to be signalled by
another thread. Note that these variants are currently only used for
UEFI runtime services calls that occur very early in the boot, and
for ones that occur in critical conditions, e.g., to flush kernel logs
to UEFI variables via efi-pstore.
Suggested-by: Andy Lutomirski <luto@kernel.org>
Signed-off-by: Sai Praneeth Prakhya <sai.praneeth.prakhya@intel.com>
[ardb: exclude ResetSystem() from the workqueue treatment
merge from 2 separate patches and rewrite commit log]
Signed-off-by: Ard Biesheuvel <ard.biesheuvel@linaro.org>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: linux-efi@vger.kernel.org
Link: http://lkml.kernel.org/r/20180711094040.12506-4-ard.biesheuvel@linaro.org
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2018-07-11 17:40:35 +08:00
|
|
|
}
|
|
|
|
|
2020-01-23 16:14:09 +08:00
|
|
|
if (efi_rt_services_supported(EFI_RT_SUPPORTED_TIME_SERVICES))
|
|
|
|
platform_device_register_simple("rtc-efi", 0, NULL, 0);
|
|
|
|
|
2013-02-08 23:37:06 +08:00
|
|
|
/* We register the efi directory at /sys/firmware/efi */
|
|
|
|
efi_kobj = kobject_create_and_add("efi", firmware_kobj);
|
|
|
|
if (!efi_kobj) {
|
|
|
|
pr_err("efi: Firmware registration failed.\n");
|
2020-07-20 15:22:18 +08:00
|
|
|
destroy_workqueue(efi_rts_wq);
|
2013-02-08 23:37:06 +08:00
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
2020-07-08 18:01:57 +08:00
|
|
|
if (efi_rt_services_supported(EFI_RT_SUPPORTED_GET_VARIABLE |
|
|
|
|
EFI_RT_SUPPORTED_GET_NEXT_VARIABLE_NAME)) {
|
2020-01-23 16:12:00 +08:00
|
|
|
error = generic_ops_register();
|
|
|
|
if (error)
|
|
|
|
goto err_put;
|
2020-11-24 01:28:17 +08:00
|
|
|
efivar_ssdt_load();
|
2020-01-23 16:12:00 +08:00
|
|
|
platform_device_register_simple("efivars", 0, NULL, 0);
|
|
|
|
}
|
2016-07-09 00:13:12 +08:00
|
|
|
|
2013-02-08 23:37:06 +08:00
|
|
|
error = sysfs_create_group(efi_kobj, &efi_subsys_attr_group);
|
|
|
|
if (error) {
|
|
|
|
pr_err("efi: Sysfs attribute export failed with error %d.\n",
|
|
|
|
error);
|
|
|
|
goto err_unregister;
|
|
|
|
}
|
|
|
|
|
2013-12-20 18:02:18 +08:00
|
|
|
error = efi_runtime_map_init(efi_kobj);
|
|
|
|
if (error)
|
|
|
|
goto err_remove_group;
|
|
|
|
|
2013-02-08 23:37:06 +08:00
|
|
|
/* and the standard mountpoint for efivarfs */
|
2015-05-14 06:35:41 +08:00
|
|
|
error = sysfs_create_mount_point(efi_kobj, "efivars");
|
|
|
|
if (error) {
|
2013-02-08 23:37:06 +08:00
|
|
|
pr_err("efivars: Subsystem registration failed.\n");
|
|
|
|
goto err_remove_group;
|
|
|
|
}
|
|
|
|
|
2020-01-16 00:35:45 +08:00
|
|
|
if (efi_enabled(EFI_DBG) && efi_enabled(EFI_PRESERVE_BS_REGIONS))
|
|
|
|
efi_debugfs_init();
|
|
|
|
|
2013-02-08 23:37:06 +08:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
err_remove_group:
|
|
|
|
sysfs_remove_group(efi_kobj, &efi_subsys_attr_group);
|
|
|
|
err_unregister:
|
2020-07-08 18:01:57 +08:00
|
|
|
if (efi_rt_services_supported(EFI_RT_SUPPORTED_GET_VARIABLE |
|
|
|
|
EFI_RT_SUPPORTED_GET_NEXT_VARIABLE_NAME))
|
2020-01-23 16:12:00 +08:00
|
|
|
generic_ops_unregister();
|
2013-02-08 23:37:06 +08:00
|
|
|
err_put:
|
|
|
|
kobject_put(efi_kobj);
|
2020-07-20 15:22:18 +08:00
|
|
|
destroy_workqueue(efi_rts_wq);
|
2013-02-08 23:37:06 +08:00
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
|
|
|
subsys_initcall(efisubsys_init);
|
2013-09-05 18:34:54 +08:00
|
|
|
|
2015-04-29 06:44:31 +08:00
|
|
|
/*
|
|
|
|
* Find the efi memory descriptor for a given physical address. Given a
|
2016-02-27 23:52:50 +08:00
|
|
|
* physical address, determine if it exists within an EFI Memory Map entry,
|
2015-04-29 06:44:31 +08:00
|
|
|
* and if so, populate the supplied memory descriptor with the appropriate
|
|
|
|
* data.
|
|
|
|
*/
|
efi: Drop type and attribute checks in efi_mem_desc_lookup()
The current implementation of efi_mem_desc_lookup() includes the
following check on the memory descriptor it returns:
if (!(md->attribute & EFI_MEMORY_RUNTIME) &&
md->type != EFI_BOOT_SERVICES_DATA &&
md->type != EFI_RUNTIME_SERVICES_DATA) {
continue;
}
This means that only EfiBootServicesData or EfiRuntimeServicesData
regions are considered, or any other region type provided that it
has the EFI_MEMORY_RUNTIME attribute set.
Given what the name of the function implies, and the fact that any
physical address can be described in the UEFI memory map only a single
time, it does not make sense to impose this condition in the body of the
loop, but instead, should be imposed by the caller depending on the value
that is returned to it.
Two such callers exist at the moment:
- The BGRT code when running on x86, via efi_mem_reserve() and
efi_arch_mem_reserve(). In this case, the region is already known to
be EfiBootServicesData, and so the check is redundant.
- The ESRT handling code which introduced this function, which calls it
both directly from efi_esrt_init() and again via efi_mem_reserve() and
efi_arch_mem_reserve() [on x86].
So let's move this check into the callers instead. This preserves the
current behavior both for BGRT and ESRT handling, and allows the lookup
routine to be reused by other [upcoming] users that don't have this
limitation.
In the ESRT case, keep the entire condition, so that platforms that
deviate from the UEFI spec and use something other than
EfiBootServicesData for the ESRT table will keep working as before.
For x86's efi_arch_mem_reserve() implementation, limit the type to
EfiBootServicesData, since it is the only type the reservation code
expects to operate on in the first place.
While we're at it, drop the __init annotation so that drivers can use it
as well.
Tested-by: Laszlo Ersek <lersek@redhat.com>
Signed-off-by: Ard Biesheuvel <ard.biesheuvel@linaro.org>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Peter Jones <pjones@redhat.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: linux-efi@vger.kernel.org
Link: http://lkml.kernel.org/r/20180711094040.12506-8-ard.biesheuvel@linaro.org
Signed-off-by: Ingo Molnar <mingo@kernel.org>
2018-07-11 17:40:39 +08:00
|
|
|
int efi_mem_desc_lookup(u64 phys_addr, efi_memory_desc_t *out_md)
|
2015-04-29 06:44:31 +08:00
|
|
|
{
|
2016-02-27 23:52:50 +08:00
|
|
|
efi_memory_desc_t *md;
|
2015-04-29 06:44:31 +08:00
|
|
|
|
|
|
|
if (!efi_enabled(EFI_MEMMAP)) {
|
|
|
|
pr_err_once("EFI_MEMMAP is not enabled.\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!out_md) {
|
|
|
|
pr_err_once("out_md is null.\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
2016-02-27 23:52:50 +08:00
|
|
|
for_each_efi_memory_desc(md) {
|
2015-04-29 06:44:31 +08:00
|
|
|
u64 size;
|
|
|
|
u64 end;
|
|
|
|
|
|
|
|
size = md->num_pages << EFI_PAGE_SHIFT;
|
|
|
|
end = md->phys_addr + size;
|
|
|
|
if (phys_addr >= md->phys_addr && phys_addr < end) {
|
|
|
|
memcpy(out_md, md, sizeof(*out_md));
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return -ENOENT;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Calculate the highest address of an efi memory descriptor.
|
|
|
|
*/
|
|
|
|
u64 __init efi_mem_desc_end(efi_memory_desc_t *md)
|
|
|
|
{
|
|
|
|
u64 size = md->num_pages << EFI_PAGE_SHIFT;
|
|
|
|
u64 end = md->phys_addr + size;
|
|
|
|
return end;
|
|
|
|
}
|
2013-09-05 18:34:54 +08:00
|
|
|
|
2016-03-01 05:22:52 +08:00
|
|
|
void __init __weak efi_arch_mem_reserve(phys_addr_t addr, u64 size) {}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* efi_mem_reserve - Reserve an EFI memory region
|
|
|
|
* @addr: Physical address to reserve
|
|
|
|
* @size: Size of reservation
|
|
|
|
*
|
|
|
|
* Mark a region as reserved from general kernel allocation and
|
|
|
|
* prevent it being released by efi_free_boot_services().
|
|
|
|
*
|
|
|
|
* This function should be called drivers once they've parsed EFI
|
|
|
|
* configuration tables to figure out where their data lives, e.g.
|
|
|
|
* efi_esrt_init().
|
|
|
|
*/
|
|
|
|
void __init efi_mem_reserve(phys_addr_t addr, u64 size)
|
|
|
|
{
|
|
|
|
if (!memblock_is_region_reserved(addr, size))
|
|
|
|
memblock_reserve(addr, size);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Some architectures (x86) reserve all boot services ranges
|
|
|
|
* until efi_free_boot_services() because of buggy firmware
|
|
|
|
* implementations. This means the above memblock_reserve() is
|
|
|
|
* superfluous on x86 and instead what it needs to do is
|
|
|
|
* ensure the @start, @size is not freed.
|
|
|
|
*/
|
|
|
|
efi_arch_mem_reserve(addr, size);
|
|
|
|
}
|
|
|
|
|
2020-01-22 21:40:57 +08:00
|
|
|
static const efi_config_table_type_t common_tables[] __initconst = {
|
2020-03-26 16:24:14 +08:00
|
|
|
{ACPI_20_TABLE_GUID, &efi.acpi20, "ACPI 2.0" },
|
|
|
|
{ACPI_TABLE_GUID, &efi.acpi, "ACPI" },
|
|
|
|
{SMBIOS_TABLE_GUID, &efi.smbios, "SMBIOS" },
|
|
|
|
{SMBIOS3_TABLE_GUID, &efi.smbios3, "SMBIOS 3.0" },
|
|
|
|
{EFI_SYSTEM_RESOURCE_TABLE_GUID, &efi.esrt, "ESRT" },
|
|
|
|
{EFI_MEMORY_ATTRIBUTES_TABLE_GUID, &efi_mem_attr_table, "MEMATTR" },
|
|
|
|
{LINUX_EFI_RANDOM_SEED_TABLE_GUID, &efi_rng_seed, "RNG" },
|
|
|
|
{LINUX_EFI_TPM_EVENT_LOG_GUID, &efi.tpm_log, "TPMEventLog" },
|
|
|
|
{LINUX_EFI_TPM_FINAL_LOG_GUID, &efi.tpm_final_log, "TPMFinalLog" },
|
|
|
|
{LINUX_EFI_MEMRESERVE_TABLE_GUID, &mem_reserve, "MEMRESERVE" },
|
|
|
|
{EFI_RT_PROPERTIES_TABLE_GUID, &rt_prop, "RTPROP" },
|
2019-07-11 02:59:15 +08:00
|
|
|
#ifdef CONFIG_EFI_RCI2_TABLE
|
2020-03-26 16:24:14 +08:00
|
|
|
{DELLEMC_EFI_RCI2_TABLE_GUID, &rci2_table_phys },
|
2020-09-05 09:31:05 +08:00
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_LOAD_UEFI_KEYS
|
|
|
|
{LINUX_EFI_MOK_VARIABLE_TABLE_GUID, &efi.mokvar_table, "MOKvar" },
|
2019-07-11 02:59:15 +08:00
|
|
|
#endif
|
2020-03-26 16:24:14 +08:00
|
|
|
{},
|
2013-09-05 18:34:54 +08:00
|
|
|
};
|
|
|
|
|
2020-01-22 21:40:57 +08:00
|
|
|
static __init int match_config_table(const efi_guid_t *guid,
|
2013-09-05 18:34:54 +08:00
|
|
|
unsigned long table,
|
2020-01-22 21:40:57 +08:00
|
|
|
const efi_config_table_type_t *table_types)
|
2013-09-05 18:34:54 +08:00
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
2020-03-26 16:34:35 +08:00
|
|
|
for (i = 0; efi_guidcmp(table_types[i].guid, NULL_GUID); i++) {
|
|
|
|
if (!efi_guidcmp(*guid, table_types[i].guid)) {
|
|
|
|
*(table_types[i].ptr) = table;
|
|
|
|
if (table_types[i].name[0])
|
|
|
|
pr_cont("%s=0x%lx ",
|
|
|
|
table_types[i].name, table);
|
|
|
|
return 1;
|
2013-09-05 18:34:54 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2020-01-22 21:40:57 +08:00
|
|
|
int __init efi_config_parse_tables(const efi_config_table_t *config_tables,
|
|
|
|
int count,
|
|
|
|
const efi_config_table_type_t *arch_tables)
|
2013-09-05 18:34:54 +08:00
|
|
|
{
|
2020-01-22 21:40:57 +08:00
|
|
|
const efi_config_table_64_t *tbl64 = (void *)config_tables;
|
|
|
|
const efi_config_table_32_t *tbl32 = (void *)config_tables;
|
|
|
|
const efi_guid_t *guid;
|
|
|
|
unsigned long table;
|
2014-10-18 21:04:15 +08:00
|
|
|
int i;
|
2013-09-05 18:34:54 +08:00
|
|
|
|
|
|
|
pr_info("");
|
2014-10-18 21:04:15 +08:00
|
|
|
for (i = 0; i < count; i++) {
|
2020-01-22 21:40:57 +08:00
|
|
|
if (!IS_ENABLED(CONFIG_X86)) {
|
|
|
|
guid = &config_tables[i].guid;
|
|
|
|
table = (unsigned long)config_tables[i].table;
|
|
|
|
} else if (efi_enabled(EFI_64BIT)) {
|
|
|
|
guid = &tbl64[i].guid;
|
|
|
|
table = tbl64[i].table;
|
|
|
|
|
|
|
|
if (IS_ENABLED(CONFIG_X86_32) &&
|
|
|
|
tbl64[i].table > U32_MAX) {
|
2013-09-05 18:34:54 +08:00
|
|
|
pr_cont("\n");
|
|
|
|
pr_err("Table located above 4GB, disabling EFI.\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
} else {
|
2020-01-22 21:40:57 +08:00
|
|
|
guid = &tbl32[i].guid;
|
|
|
|
table = tbl32[i].table;
|
2013-09-05 18:34:54 +08:00
|
|
|
}
|
|
|
|
|
2020-03-26 16:34:35 +08:00
|
|
|
if (!match_config_table(guid, table, common_tables) && arch_tables)
|
2020-01-22 21:40:57 +08:00
|
|
|
match_config_table(guid, table, arch_tables);
|
2013-09-05 18:34:54 +08:00
|
|
|
}
|
|
|
|
pr_cont("\n");
|
2014-01-15 21:36:33 +08:00
|
|
|
set_bit(EFI_CONFIG_TABLES, &efi.flags);
|
2015-09-23 22:29:34 +08:00
|
|
|
|
2020-02-28 20:14:04 +08:00
|
|
|
if (efi_rng_seed != EFI_INVALID_TABLE_ADDR) {
|
2016-11-13 05:32:31 +08:00
|
|
|
struct linux_efi_random_seed *seed;
|
|
|
|
u32 size = 0;
|
|
|
|
|
2020-02-28 20:14:04 +08:00
|
|
|
seed = early_memremap(efi_rng_seed, sizeof(*seed));
|
2016-11-13 05:32:31 +08:00
|
|
|
if (seed != NULL) {
|
2020-02-21 16:48:49 +08:00
|
|
|
size = READ_ONCE(seed->size);
|
2016-11-13 05:32:31 +08:00
|
|
|
early_memunmap(seed, sizeof(*seed));
|
|
|
|
} else {
|
|
|
|
pr_err("Could not map UEFI random seed!\n");
|
|
|
|
}
|
|
|
|
if (size > 0) {
|
2020-02-28 20:14:04 +08:00
|
|
|
seed = early_memremap(efi_rng_seed,
|
|
|
|
sizeof(*seed) + size);
|
2016-11-13 05:32:31 +08:00
|
|
|
if (seed != NULL) {
|
2018-03-08 16:00:18 +08:00
|
|
|
pr_notice("seeding entropy pool\n");
|
2020-02-21 16:48:49 +08:00
|
|
|
add_bootloader_randomness(seed->bits, size);
|
2016-11-13 05:32:31 +08:00
|
|
|
early_memunmap(seed, sizeof(*seed) + size);
|
|
|
|
} else {
|
|
|
|
pr_err("Could not map UEFI random seed!\n");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-03-08 16:08:51 +08:00
|
|
|
if (!IS_ENABLED(CONFIG_X86_32) && efi_enabled(EFI_MEMMAP))
|
2017-06-22 18:51:36 +08:00
|
|
|
efi_memattr_init();
|
2017-01-31 21:21:35 +08:00
|
|
|
|
2017-09-20 16:13:39 +08:00
|
|
|
efi_tpm_eventlog_init();
|
|
|
|
|
2020-01-22 22:06:54 +08:00
|
|
|
if (mem_reserve != EFI_INVALID_TABLE_ADDR) {
|
|
|
|
unsigned long prsv = mem_reserve;
|
2018-09-22 00:32:44 +08:00
|
|
|
|
|
|
|
while (prsv) {
|
|
|
|
struct linux_efi_memreserve *rsv;
|
2018-11-30 01:12:28 +08:00
|
|
|
u8 *p;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Just map a full page: that is what we will get
|
|
|
|
* anyway, and it permits us to map the entire entry
|
|
|
|
* before knowing its size.
|
|
|
|
*/
|
|
|
|
p = early_memremap(ALIGN_DOWN(prsv, PAGE_SIZE),
|
|
|
|
PAGE_SIZE);
|
|
|
|
if (p == NULL) {
|
2018-09-22 00:32:44 +08:00
|
|
|
pr_err("Could not map UEFI memreserve entry!\n");
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
2018-11-30 01:12:28 +08:00
|
|
|
rsv = (void *)(p + prsv % PAGE_SIZE);
|
|
|
|
|
|
|
|
/* reserve the entry itself */
|
efi: Replace zero-length array and use struct_size() helper
The current codebase makes use of the zero-length array language
extension to the C90 standard, but the preferred mechanism to declare
variable-length types such as these ones is a flexible array member[1][2],
introduced in C99:
struct foo {
int stuff;
struct boo array[];
};
By making use of the mechanism above, we will get a compiler warning
in case the flexible array does not occur last in the structure, which
will help us prevent some kind of undefined behavior bugs from being
inadvertently introduced[3] to the codebase from now on.
Also, notice that, dynamic memory allocations won't be affected by
this change:
"Flexible array members have incomplete type, and so the sizeof operator
may not be applied. As a quirk of the original implementation of
zero-length arrays, sizeof evaluates to zero."[1]
sizeof(flexible-array-member) triggers a warning because flexible array
members have incomplete type[1]. There are some instances of code in
which the sizeof operator is being incorrectly/erroneously applied to
zero-length arrays and the result is zero. Such instances may be hiding
some bugs. So, this work (flexible-array member conversions) will also
help to get completely rid of those sorts of issues.
Lastly, make use of the sizeof_field() helper instead of an open-coded
version.
This issue was found with the help of Coccinelle and audited _manually_.
[1] https://gcc.gnu.org/onlinedocs/gcc/Zero-Length.html
[2] https://github.com/KSPP/linux/issues/21
[3] commit 76497732932f ("cxgb3/l2t: Fix undefined behaviour")
Signed-off-by: Gustavo A. R. Silva <gustavoars@kernel.org>
Reviewed-by: Kees Cook <keescook@chromium.org>
Link: https://lore.kernel.org/r/20200527171425.GA4053@embeddedor
Signed-off-by: Ard Biesheuvel <ardb@kernel.org>
2020-05-28 01:14:25 +08:00
|
|
|
memblock_reserve(prsv,
|
|
|
|
struct_size(rsv, entry, rsv->size));
|
2018-11-30 01:12:28 +08:00
|
|
|
|
|
|
|
for (i = 0; i < atomic_read(&rsv->count); i++) {
|
|
|
|
memblock_reserve(rsv->entry[i].base,
|
|
|
|
rsv->entry[i].size);
|
|
|
|
}
|
2018-09-22 00:32:44 +08:00
|
|
|
|
|
|
|
prsv = rsv->next;
|
2018-11-30 01:12:28 +08:00
|
|
|
early_memunmap(p, PAGE_SIZE);
|
2018-09-22 00:32:44 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-01-23 20:10:25 +08:00
|
|
|
if (rt_prop != EFI_INVALID_TABLE_ADDR) {
|
|
|
|
efi_rt_properties_table_t *tbl;
|
|
|
|
|
|
|
|
tbl = early_memremap(rt_prop, sizeof(*tbl));
|
|
|
|
if (tbl) {
|
|
|
|
efi.runtime_supported_mask &= tbl->runtime_services_supported;
|
|
|
|
early_memunmap(tbl, sizeof(*tbl));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-09-05 18:34:54 +08:00
|
|
|
return 0;
|
|
|
|
}
|
2013-12-31 01:12:12 +08:00
|
|
|
|
2020-01-20 17:49:11 +08:00
|
|
|
int __init efi_systab_check_header(const efi_table_hdr_t *systab_hdr,
|
|
|
|
int min_major_version)
|
|
|
|
{
|
|
|
|
if (systab_hdr->signature != EFI_SYSTEM_TABLE_SIGNATURE) {
|
|
|
|
pr_err("System table signature incorrect!\n");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((systab_hdr->revision >> 16) < min_major_version)
|
|
|
|
pr_err("Warning: System table version %d.%02d, expected %d.00 or greater!\n",
|
|
|
|
systab_hdr->revision >> 16,
|
|
|
|
systab_hdr->revision & 0xffff,
|
|
|
|
min_major_version);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifndef CONFIG_IA64
|
|
|
|
static const efi_char16_t *__init map_fw_vendor(unsigned long fw_vendor,
|
|
|
|
size_t size)
|
|
|
|
{
|
|
|
|
const efi_char16_t *ret;
|
|
|
|
|
|
|
|
ret = early_memremap_ro(fw_vendor, size);
|
|
|
|
if (!ret)
|
|
|
|
pr_err("Could not map the firmware vendor!\n");
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void __init unmap_fw_vendor(const void *fw_vendor, size_t size)
|
|
|
|
{
|
|
|
|
early_memunmap((void *)fw_vendor, size);
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
#define map_fw_vendor(p, s) __va(p)
|
|
|
|
#define unmap_fw_vendor(v, s)
|
|
|
|
#endif
|
|
|
|
|
|
|
|
void __init efi_systab_report_header(const efi_table_hdr_t *systab_hdr,
|
|
|
|
unsigned long fw_vendor)
|
|
|
|
{
|
|
|
|
char vendor[100] = "unknown";
|
|
|
|
const efi_char16_t *c16;
|
|
|
|
size_t i;
|
|
|
|
|
|
|
|
c16 = map_fw_vendor(fw_vendor, sizeof(vendor) * sizeof(efi_char16_t));
|
|
|
|
if (c16) {
|
|
|
|
for (i = 0; i < sizeof(vendor) - 1 && c16[i]; ++i)
|
|
|
|
vendor[i] = c16[i];
|
|
|
|
vendor[i] = '\0';
|
|
|
|
|
|
|
|
unmap_fw_vendor(c16, sizeof(vendor) * sizeof(efi_char16_t));
|
|
|
|
}
|
|
|
|
|
|
|
|
pr_info("EFI v%u.%.02u by %s\n",
|
|
|
|
systab_hdr->revision >> 16,
|
|
|
|
systab_hdr->revision & 0xffff,
|
|
|
|
vendor);
|
efi: runtime: avoid EFIv2 runtime services on Apple x86 machines
Aditya reports [0] that his recent MacbookPro crashes in the firmware
when using the variable services at runtime. The culprit appears to be a
call to QueryVariableInfo(), which we did not use to call on Apple x86
machines in the past as they only upgraded from EFI v1.10 to EFI v2.40
firmware fairly recently, and QueryVariableInfo() (along with
UpdateCapsule() et al) was added in EFI v2.00.
The only runtime service introduced in EFI v2.00 that we actually use in
Linux is QueryVariableInfo(), as the capsule based ones are optional,
generally not used at runtime (all the LVFS/fwupd firmware update
infrastructure uses helper EFI programs that invoke capsule update at
boot time, not runtime), and not implemented by Apple machines in the
first place. QueryVariableInfo() is used to 'safely' set variables,
i.e., only when there is enough space. This prevents machines with buggy
firmwares from corrupting their NVRAMs when they run out of space.
Given that Apple machines have been using EFI v1.10 services only for
the longest time (the EFI v2.0 spec was released in 2006, and Linux
support for the newly introduced runtime services was added in 2011, but
the MacbookPro12,1 released in 2015 still claims to be EFI v1.10 only),
let's avoid the EFI v2.0 ones on all Apple x86 machines.
[0] https://lore.kernel.org/all/6D757C75-65B1-468B-842D-10410081A8E4@live.com/
Cc: <stable@vger.kernel.org>
Cc: Jeremy Kerr <jk@ozlabs.org>
Cc: Matthew Garrett <mjg59@srcf.ucam.org>
Reported-by: Aditya Garg <gargaditya08@live.com>
Tested-by: Orlando Chamberlain <redecorating@protonmail.com>
Signed-off-by: Ard Biesheuvel <ardb@kernel.org>
Tested-by: Aditya Garg <gargaditya08@live.com>
Link: https://bugzilla.kernel.org/show_bug.cgi?id=215277
2022-01-12 18:14:13 +08:00
|
|
|
|
|
|
|
if (IS_ENABLED(CONFIG_X86_64) &&
|
|
|
|
systab_hdr->revision > EFI_1_10_SYSTEM_TABLE_REVISION &&
|
|
|
|
!strcmp(vendor, "Apple")) {
|
|
|
|
pr_info("Apple Mac detected, using EFI v1.10 runtime services only\n");
|
|
|
|
efi.runtime_version = EFI_1_10_SYSTEM_TABLE_REVISION;
|
|
|
|
}
|
2020-01-20 17:49:11 +08:00
|
|
|
}
|
|
|
|
|
2020-09-24 19:52:24 +08:00
|
|
|
static __initdata char memory_type_name[][13] = {
|
2014-09-03 19:32:20 +08:00
|
|
|
"Reserved",
|
|
|
|
"Loader Code",
|
|
|
|
"Loader Data",
|
|
|
|
"Boot Code",
|
|
|
|
"Boot Data",
|
|
|
|
"Runtime Code",
|
|
|
|
"Runtime Data",
|
2020-09-24 19:52:24 +08:00
|
|
|
"Conventional",
|
|
|
|
"Unusable",
|
|
|
|
"ACPI Reclaim",
|
|
|
|
"ACPI Mem NVS",
|
|
|
|
"MMIO",
|
|
|
|
"MMIO Port",
|
2016-02-02 06:07:07 +08:00
|
|
|
"PAL Code",
|
2020-09-24 19:52:24 +08:00
|
|
|
"Persistent",
|
2014-09-03 19:32:20 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
char * __init efi_md_typeattr_format(char *buf, size_t size,
|
|
|
|
const efi_memory_desc_t *md)
|
|
|
|
{
|
|
|
|
char *pos;
|
|
|
|
int type_len;
|
|
|
|
u64 attr;
|
|
|
|
|
|
|
|
pos = buf;
|
|
|
|
if (md->type >= ARRAY_SIZE(memory_type_name))
|
|
|
|
type_len = snprintf(pos, size, "[type=%u", md->type);
|
|
|
|
else
|
|
|
|
type_len = snprintf(pos, size, "[%-*s",
|
|
|
|
(int)(sizeof(memory_type_name[0]) - 1),
|
|
|
|
memory_type_name[md->type]);
|
|
|
|
if (type_len >= size)
|
|
|
|
return buf;
|
|
|
|
|
|
|
|
pos += type_len;
|
|
|
|
size -= type_len;
|
|
|
|
|
|
|
|
attr = md->attribute;
|
|
|
|
if (attr & ~(EFI_MEMORY_UC | EFI_MEMORY_WC | EFI_MEMORY_WT |
|
2015-08-07 16:36:54 +08:00
|
|
|
EFI_MEMORY_WB | EFI_MEMORY_UCE | EFI_MEMORY_RO |
|
|
|
|
EFI_MEMORY_WP | EFI_MEMORY_RP | EFI_MEMORY_XP |
|
2020-09-24 19:52:24 +08:00
|
|
|
EFI_MEMORY_NV | EFI_MEMORY_SP | EFI_MEMORY_CPU_CRYPTO |
|
2015-08-27 01:11:19 +08:00
|
|
|
EFI_MEMORY_RUNTIME | EFI_MEMORY_MORE_RELIABLE))
|
2014-09-03 19:32:20 +08:00
|
|
|
snprintf(pos, size, "|attr=0x%016llx]",
|
|
|
|
(unsigned long long)attr);
|
|
|
|
else
|
2016-02-02 06:07:06 +08:00
|
|
|
snprintf(pos, size,
|
2020-09-24 19:52:24 +08:00
|
|
|
"|%3s|%2s|%2s|%2s|%2s|%2s|%2s|%2s|%2s|%3s|%2s|%2s|%2s|%2s]",
|
|
|
|
attr & EFI_MEMORY_RUNTIME ? "RUN" : "",
|
|
|
|
attr & EFI_MEMORY_MORE_RELIABLE ? "MR" : "",
|
|
|
|
attr & EFI_MEMORY_CPU_CRYPTO ? "CC" : "",
|
|
|
|
attr & EFI_MEMORY_SP ? "SP" : "",
|
|
|
|
attr & EFI_MEMORY_NV ? "NV" : "",
|
|
|
|
attr & EFI_MEMORY_XP ? "XP" : "",
|
|
|
|
attr & EFI_MEMORY_RP ? "RP" : "",
|
|
|
|
attr & EFI_MEMORY_WP ? "WP" : "",
|
|
|
|
attr & EFI_MEMORY_RO ? "RO" : "",
|
|
|
|
attr & EFI_MEMORY_UCE ? "UCE" : "",
|
|
|
|
attr & EFI_MEMORY_WB ? "WB" : "",
|
|
|
|
attr & EFI_MEMORY_WT ? "WT" : "",
|
|
|
|
attr & EFI_MEMORY_WC ? "WC" : "",
|
|
|
|
attr & EFI_MEMORY_UC ? "UC" : "");
|
2014-09-03 19:32:20 +08:00
|
|
|
return buf;
|
|
|
|
}
|
2015-08-07 16:36:57 +08:00
|
|
|
|
2017-08-25 23:50:18 +08:00
|
|
|
/*
|
|
|
|
* IA64 has a funky EFI memory map that doesn't work the same way as
|
|
|
|
* other architectures.
|
|
|
|
*/
|
|
|
|
#ifndef CONFIG_IA64
|
2015-08-07 16:36:57 +08:00
|
|
|
/*
|
|
|
|
* efi_mem_attributes - lookup memmap attributes for physical address
|
|
|
|
* @phys_addr: the physical address to lookup
|
|
|
|
*
|
|
|
|
* Search in the EFI memory map for the region covering
|
|
|
|
* @phys_addr. Returns the EFI memory attributes if the region
|
|
|
|
* was found in the memory map, 0 otherwise.
|
|
|
|
*/
|
2017-08-25 23:50:18 +08:00
|
|
|
u64 efi_mem_attributes(unsigned long phys_addr)
|
2015-08-07 16:36:57 +08:00
|
|
|
{
|
|
|
|
efi_memory_desc_t *md;
|
|
|
|
|
|
|
|
if (!efi_enabled(EFI_MEMMAP))
|
|
|
|
return 0;
|
|
|
|
|
2016-04-26 04:06:38 +08:00
|
|
|
for_each_efi_memory_desc(md) {
|
2015-08-07 16:36:57 +08:00
|
|
|
if ((md->phys_addr <= phys_addr) &&
|
|
|
|
(phys_addr < (md->phys_addr +
|
|
|
|
(md->num_pages << EFI_PAGE_SHIFT))))
|
|
|
|
return md->attribute;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
2016-04-26 04:06:58 +08:00
|
|
|
|
2017-08-25 23:50:18 +08:00
|
|
|
/*
|
|
|
|
* efi_mem_type - lookup memmap type for physical address
|
|
|
|
* @phys_addr: the physical address to lookup
|
|
|
|
*
|
|
|
|
* Search in the EFI memory map for the region covering @phys_addr.
|
|
|
|
* Returns the EFI memory type if the region was found in the memory
|
2020-01-14 01:22:41 +08:00
|
|
|
* map, -EINVAL otherwise.
|
2017-08-25 23:50:18 +08:00
|
|
|
*/
|
|
|
|
int efi_mem_type(unsigned long phys_addr)
|
|
|
|
{
|
|
|
|
const efi_memory_desc_t *md;
|
|
|
|
|
|
|
|
if (!efi_enabled(EFI_MEMMAP))
|
|
|
|
return -ENOTSUPP;
|
|
|
|
|
|
|
|
for_each_efi_memory_desc(md) {
|
|
|
|
if ((md->phys_addr <= phys_addr) &&
|
|
|
|
(phys_addr < (md->phys_addr +
|
|
|
|
(md->num_pages << EFI_PAGE_SHIFT))))
|
|
|
|
return md->type;
|
|
|
|
}
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2016-04-26 04:06:58 +08:00
|
|
|
int efi_status_to_err(efi_status_t status)
|
|
|
|
{
|
|
|
|
int err;
|
|
|
|
|
|
|
|
switch (status) {
|
|
|
|
case EFI_SUCCESS:
|
|
|
|
err = 0;
|
|
|
|
break;
|
|
|
|
case EFI_INVALID_PARAMETER:
|
|
|
|
err = -EINVAL;
|
|
|
|
break;
|
|
|
|
case EFI_OUT_OF_RESOURCES:
|
|
|
|
err = -ENOSPC;
|
|
|
|
break;
|
|
|
|
case EFI_DEVICE_ERROR:
|
|
|
|
err = -EIO;
|
|
|
|
break;
|
|
|
|
case EFI_WRITE_PROTECTED:
|
|
|
|
err = -EROFS;
|
|
|
|
break;
|
|
|
|
case EFI_SECURITY_VIOLATION:
|
|
|
|
err = -EACCES;
|
|
|
|
break;
|
|
|
|
case EFI_NOT_FOUND:
|
|
|
|
err = -ENOENT;
|
|
|
|
break;
|
2016-07-16 03:36:31 +08:00
|
|
|
case EFI_ABORTED:
|
|
|
|
err = -EINTR;
|
|
|
|
break;
|
2016-04-26 04:06:58 +08:00
|
|
|
default:
|
|
|
|
err = -EINVAL;
|
|
|
|
}
|
|
|
|
|
|
|
|
return err;
|
2017-07-18 05:10:13 +08:00
|
|
|
}
|
|
|
|
|
2018-09-22 00:32:46 +08:00
|
|
|
static DEFINE_SPINLOCK(efi_mem_reserve_persistent_lock);
|
2018-11-15 01:55:44 +08:00
|
|
|
static struct linux_efi_memreserve *efi_memreserve_root __ro_after_init;
|
2018-09-22 00:32:46 +08:00
|
|
|
|
2018-11-24 05:51:32 +08:00
|
|
|
static int __init efi_memreserve_map_root(void)
|
|
|
|
{
|
2020-01-22 22:06:54 +08:00
|
|
|
if (mem_reserve == EFI_INVALID_TABLE_ADDR)
|
2018-11-24 05:51:32 +08:00
|
|
|
return -ENODEV;
|
|
|
|
|
2020-01-22 22:06:54 +08:00
|
|
|
efi_memreserve_root = memremap(mem_reserve,
|
2018-11-24 05:51:32 +08:00
|
|
|
sizeof(*efi_memreserve_root),
|
|
|
|
MEMREMAP_WB);
|
|
|
|
if (WARN_ON_ONCE(!efi_memreserve_root))
|
|
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2019-12-07 00:55:37 +08:00
|
|
|
static int efi_mem_reserve_iomem(phys_addr_t addr, u64 size)
|
|
|
|
{
|
|
|
|
struct resource *res, *parent;
|
2021-07-14 02:43:26 +08:00
|
|
|
int ret;
|
2019-12-07 00:55:37 +08:00
|
|
|
|
|
|
|
res = kzalloc(sizeof(struct resource), GFP_ATOMIC);
|
|
|
|
if (!res)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
res->name = "reserved";
|
|
|
|
res->flags = IORESOURCE_MEM;
|
|
|
|
res->start = addr;
|
|
|
|
res->end = addr + size - 1;
|
|
|
|
|
|
|
|
/* we expect a conflict with a 'System RAM' region */
|
|
|
|
parent = request_resource_conflict(&iomem_resource, res);
|
2021-07-14 02:43:26 +08:00
|
|
|
ret = parent ? request_resource(parent, res) : 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Given that efi_mem_reserve_iomem() can be called at any
|
|
|
|
* time, only call memblock_reserve() if the architecture
|
|
|
|
* keeps the infrastructure around.
|
|
|
|
*/
|
|
|
|
if (IS_ENABLED(CONFIG_ARCH_KEEP_MEMBLOCK) && !ret)
|
|
|
|
memblock_reserve(addr, size);
|
|
|
|
|
|
|
|
return ret;
|
2019-12-07 00:55:37 +08:00
|
|
|
}
|
|
|
|
|
2018-11-24 05:51:32 +08:00
|
|
|
int __ref efi_mem_reserve_persistent(phys_addr_t addr, u64 size)
|
2018-09-22 00:32:46 +08:00
|
|
|
{
|
2018-11-15 01:55:44 +08:00
|
|
|
struct linux_efi_memreserve *rsv;
|
2018-11-30 01:12:29 +08:00
|
|
|
unsigned long prsv;
|
|
|
|
int rc, index;
|
2018-09-22 00:32:46 +08:00
|
|
|
|
2018-11-24 05:51:32 +08:00
|
|
|
if (efi_memreserve_root == (void *)ULONG_MAX)
|
2018-09-22 00:32:46 +08:00
|
|
|
return -ENODEV;
|
|
|
|
|
2018-11-24 05:51:32 +08:00
|
|
|
if (!efi_memreserve_root) {
|
|
|
|
rc = efi_memreserve_map_root();
|
|
|
|
if (rc)
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
2018-11-30 01:12:29 +08:00
|
|
|
/* first try to find a slot in an existing linked list entry */
|
2021-03-10 16:31:27 +08:00
|
|
|
for (prsv = efi_memreserve_root->next; prsv; ) {
|
2019-06-10 02:17:44 +08:00
|
|
|
rsv = memremap(prsv, sizeof(*rsv), MEMREMAP_WB);
|
2018-11-30 01:12:29 +08:00
|
|
|
index = atomic_fetch_add_unless(&rsv->count, 1, rsv->size);
|
|
|
|
if (index < rsv->size) {
|
|
|
|
rsv->entry[index].base = addr;
|
|
|
|
rsv->entry[index].size = size;
|
|
|
|
|
2019-06-10 02:17:44 +08:00
|
|
|
memunmap(rsv);
|
2019-12-07 00:55:37 +08:00
|
|
|
return efi_mem_reserve_iomem(addr, size);
|
2018-11-30 01:12:29 +08:00
|
|
|
}
|
2021-03-10 16:31:27 +08:00
|
|
|
prsv = rsv->next;
|
2019-06-10 02:17:44 +08:00
|
|
|
memunmap(rsv);
|
2018-11-30 01:12:29 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/* no slot found - allocate a new linked list entry */
|
|
|
|
rsv = (struct linux_efi_memreserve *)__get_free_page(GFP_ATOMIC);
|
2018-09-22 00:32:46 +08:00
|
|
|
if (!rsv)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
2019-12-07 00:55:37 +08:00
|
|
|
rc = efi_mem_reserve_iomem(__pa(rsv), SZ_4K);
|
|
|
|
if (rc) {
|
|
|
|
free_page((unsigned long)rsv);
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
2019-06-10 02:17:44 +08:00
|
|
|
/*
|
|
|
|
* The memremap() call above assumes that a linux_efi_memreserve entry
|
|
|
|
* never crosses a page boundary, so let's ensure that this remains true
|
|
|
|
* even when kexec'ing a 4k pages kernel from a >4k pages kernel, by
|
|
|
|
* using SZ_4K explicitly in the size calculation below.
|
|
|
|
*/
|
|
|
|
rsv->size = EFI_MEMRESERVE_COUNT(SZ_4K);
|
2018-11-30 01:12:28 +08:00
|
|
|
atomic_set(&rsv->count, 1);
|
|
|
|
rsv->entry[0].base = addr;
|
|
|
|
rsv->entry[0].size = size;
|
2018-09-22 00:32:46 +08:00
|
|
|
|
|
|
|
spin_lock(&efi_mem_reserve_persistent_lock);
|
2018-11-15 01:55:44 +08:00
|
|
|
rsv->next = efi_memreserve_root->next;
|
|
|
|
efi_memreserve_root->next = __pa(rsv);
|
2018-09-22 00:32:46 +08:00
|
|
|
spin_unlock(&efi_mem_reserve_persistent_lock);
|
|
|
|
|
2019-12-07 00:55:37 +08:00
|
|
|
return efi_mem_reserve_iomem(addr, size);
|
2018-11-15 01:55:44 +08:00
|
|
|
}
|
2018-09-22 00:32:46 +08:00
|
|
|
|
2018-11-15 01:55:44 +08:00
|
|
|
static int __init efi_memreserve_root_init(void)
|
|
|
|
{
|
2018-11-24 05:51:32 +08:00
|
|
|
if (efi_memreserve_root)
|
|
|
|
return 0;
|
|
|
|
if (efi_memreserve_map_root())
|
|
|
|
efi_memreserve_root = (void *)ULONG_MAX;
|
2018-09-22 00:32:46 +08:00
|
|
|
return 0;
|
|
|
|
}
|
2018-11-15 01:55:44 +08:00
|
|
|
early_initcall(efi_memreserve_root_init);
|
2018-09-22 00:32:46 +08:00
|
|
|
|
2016-11-13 05:32:31 +08:00
|
|
|
#ifdef CONFIG_KEXEC
|
|
|
|
static int update_efi_random_seed(struct notifier_block *nb,
|
|
|
|
unsigned long code, void *unused)
|
|
|
|
{
|
|
|
|
struct linux_efi_random_seed *seed;
|
|
|
|
u32 size = 0;
|
|
|
|
|
|
|
|
if (!kexec_in_progress)
|
|
|
|
return NOTIFY_DONE;
|
|
|
|
|
2020-02-28 20:14:04 +08:00
|
|
|
seed = memremap(efi_rng_seed, sizeof(*seed), MEMREMAP_WB);
|
2016-11-13 05:32:31 +08:00
|
|
|
if (seed != NULL) {
|
2017-08-25 23:50:16 +08:00
|
|
|
size = min(seed->size, EFI_RANDOM_SEED_SIZE);
|
2016-11-13 05:32:31 +08:00
|
|
|
memunmap(seed);
|
|
|
|
} else {
|
|
|
|
pr_err("Could not map UEFI random seed!\n");
|
|
|
|
}
|
|
|
|
if (size > 0) {
|
2020-02-28 20:14:04 +08:00
|
|
|
seed = memremap(efi_rng_seed, sizeof(*seed) + size,
|
|
|
|
MEMREMAP_WB);
|
2016-11-13 05:32:31 +08:00
|
|
|
if (seed != NULL) {
|
|
|
|
seed->size = size;
|
|
|
|
get_random_bytes(seed->bits, seed->size);
|
|
|
|
memunmap(seed);
|
|
|
|
} else {
|
|
|
|
pr_err("Could not map UEFI random seed!\n");
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return NOTIFY_DONE;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct notifier_block efi_random_seed_nb = {
|
|
|
|
.notifier_call = update_efi_random_seed,
|
|
|
|
};
|
|
|
|
|
2020-01-22 21:58:15 +08:00
|
|
|
static int __init register_update_efi_random_seed(void)
|
2016-11-13 05:32:31 +08:00
|
|
|
{
|
2020-02-28 20:14:04 +08:00
|
|
|
if (efi_rng_seed == EFI_INVALID_TABLE_ADDR)
|
2016-11-13 05:32:31 +08:00
|
|
|
return 0;
|
|
|
|
return register_reboot_notifier(&efi_random_seed_nb);
|
|
|
|
}
|
|
|
|
late_initcall(register_update_efi_random_seed);
|
|
|
|
#endif
|