]> git.proxmox.com Git - mirror_ubuntu-zesty-kernel.git/blob - arch/arm64/kernel/efi.c
Merge branch 'for-linus-4.10' of git://git.kernel.org/pub/scm/linux/kernel/git/mason...
[mirror_ubuntu-zesty-kernel.git] / arch / arm64 / kernel / efi.c
1 /*
2 * Extensible Firmware Interface
3 *
4 * Based on Extensible Firmware Interface Specification version 2.4
5 *
6 * Copyright (C) 2013, 2014 Linaro Ltd.
7 *
8 * This program is free software; you can redistribute it and/or modify
9 * it under the terms of the GNU General Public License version 2 as
10 * published by the Free Software Foundation.
11 *
12 */
13
14 #include <linux/dmi.h>
15 #include <linux/efi.h>
16 #include <linux/init.h>
17
18 #include <asm/efi.h>
19
20 /*
21 * Only regions of type EFI_RUNTIME_SERVICES_CODE need to be
22 * executable, everything else can be mapped with the XN bits
23 * set. Also take the new (optional) RO/XP bits into account.
24 */
25 static __init pteval_t create_mapping_protection(efi_memory_desc_t *md)
26 {
27 u64 attr = md->attribute;
28 u32 type = md->type;
29
30 if (type == EFI_MEMORY_MAPPED_IO)
31 return PROT_DEVICE_nGnRE;
32
33 if (WARN_ONCE(!PAGE_ALIGNED(md->phys_addr),
34 "UEFI Runtime regions are not aligned to 64 KB -- buggy firmware?"))
35 /*
36 * If the region is not aligned to the page size of the OS, we
37 * can not use strict permissions, since that would also affect
38 * the mapping attributes of the adjacent regions.
39 */
40 return pgprot_val(PAGE_KERNEL_EXEC);
41
42 /* R-- */
43 if ((attr & (EFI_MEMORY_XP | EFI_MEMORY_RO)) ==
44 (EFI_MEMORY_XP | EFI_MEMORY_RO))
45 return pgprot_val(PAGE_KERNEL_RO);
46
47 /* R-X */
48 if (attr & EFI_MEMORY_RO)
49 return pgprot_val(PAGE_KERNEL_ROX);
50
51 /* RW- */
52 if (attr & EFI_MEMORY_XP || type != EFI_RUNTIME_SERVICES_CODE)
53 return pgprot_val(PAGE_KERNEL);
54
55 /* RWX */
56 return pgprot_val(PAGE_KERNEL_EXEC);
57 }
58
59 /* we will fill this structure from the stub, so don't put it in .bss */
60 struct screen_info screen_info __section(.data);
61
62 int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md)
63 {
64 pteval_t prot_val = create_mapping_protection(md);
65 bool page_mappings_only = (md->type == EFI_RUNTIME_SERVICES_CODE ||
66 md->type == EFI_RUNTIME_SERVICES_DATA);
67
68 if (!PAGE_ALIGNED(md->phys_addr) ||
69 !PAGE_ALIGNED(md->num_pages << EFI_PAGE_SHIFT)) {
70 /*
71 * If the end address of this region is not aligned to page
72 * size, the mapping is rounded up, and may end up sharing a
73 * page frame with the next UEFI memory region. If we create
74 * a block entry now, we may need to split it again when mapping
75 * the next region, and support for that is going to be removed
76 * from the MMU routines. So avoid block mappings altogether in
77 * that case.
78 */
79 page_mappings_only = true;
80 }
81
82 create_pgd_mapping(mm, md->phys_addr, md->virt_addr,
83 md->num_pages << EFI_PAGE_SHIFT,
84 __pgprot(prot_val | PTE_NG), page_mappings_only);
85 return 0;
86 }
87
88 static int __init set_permissions(pte_t *ptep, pgtable_t token,
89 unsigned long addr, void *data)
90 {
91 efi_memory_desc_t *md = data;
92 pte_t pte = *ptep;
93
94 if (md->attribute & EFI_MEMORY_RO)
95 pte = set_pte_bit(pte, __pgprot(PTE_RDONLY));
96 if (md->attribute & EFI_MEMORY_XP)
97 pte = set_pte_bit(pte, __pgprot(PTE_PXN));
98 set_pte(ptep, pte);
99 return 0;
100 }
101
102 int __init efi_set_mapping_permissions(struct mm_struct *mm,
103 efi_memory_desc_t *md)
104 {
105 BUG_ON(md->type != EFI_RUNTIME_SERVICES_CODE &&
106 md->type != EFI_RUNTIME_SERVICES_DATA);
107
108 /*
109 * Calling apply_to_page_range() is only safe on regions that are
110 * guaranteed to be mapped down to pages. Since we are only called
111 * for regions that have been mapped using efi_create_mapping() above
112 * (and this is checked by the generic Memory Attributes table parsing
113 * routines), there is no need to check that again here.
114 */
115 return apply_to_page_range(mm, md->virt_addr,
116 md->num_pages << EFI_PAGE_SHIFT,
117 set_permissions, md);
118 }
119
120 static int __init arm64_dmi_init(void)
121 {
122 /*
123 * On arm64, DMI depends on UEFI, and dmi_scan_machine() needs to
124 * be called early because dmi_id_init(), which is an arch_initcall
125 * itself, depends on dmi_scan_machine() having been called already.
126 */
127 dmi_scan_machine();
128 if (dmi_available)
129 dmi_set_dump_stack_arch_desc();
130 return 0;
131 }
132 core_initcall(arm64_dmi_init);
133
134 /*
135 * UpdateCapsule() depends on the system being shutdown via
136 * ResetSystem().
137 */
138 bool efi_poweroff_required(void)
139 {
140 return efi_enabled(EFI_RUNTIME_SERVICES);
141 }