WIP FPC-III support
[linux/fpc-iii.git] / drivers / firmware / efi / capsule.c
blob768430293669fc678b15b35397ea1dbe582e8067
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * EFI capsule support.
5 * Copyright 2013 Intel Corporation; author Matt Fleming
6 */
8 #define pr_fmt(fmt) "efi: " fmt
10 #include <linux/slab.h>
11 #include <linux/mutex.h>
12 #include <linux/highmem.h>
13 #include <linux/efi.h>
14 #include <linux/vmalloc.h>
15 #include <asm/efi.h>
16 #include <asm/io.h>
18 typedef struct {
19 u64 length;
20 u64 data;
21 } efi_capsule_block_desc_t;
23 static bool capsule_pending;
24 static bool stop_capsules;
25 static int efi_reset_type = -1;
28 * capsule_mutex serialises access to both capsule_pending and
29 * efi_reset_type and stop_capsules.
31 static DEFINE_MUTEX(capsule_mutex);
33 /**
34 * efi_capsule_pending - has a capsule been passed to the firmware?
35 * @reset_type: store the type of EFI reset if capsule is pending
37 * To ensure that the registered capsule is processed correctly by the
38 * firmware we need to perform a specific type of reset. If a capsule is
39 * pending return the reset type in @reset_type.
41 * This function will race with callers of efi_capsule_update(), for
42 * example, calling this function while somebody else is in
43 * efi_capsule_update() but hasn't reached efi_capsue_update_locked()
44 * will miss the updates to capsule_pending and efi_reset_type after
45 * efi_capsule_update_locked() completes.
47 * A non-racy use is from platform reboot code because we use
48 * system_state to ensure no capsules can be sent to the firmware once
49 * we're at SYSTEM_RESTART. See efi_capsule_update_locked().
51 bool efi_capsule_pending(int *reset_type)
53 if (!capsule_pending)
54 return false;
56 if (reset_type)
57 *reset_type = efi_reset_type;
59 return true;
63 * Whitelist of EFI capsule flags that we support.
65 * We do not handle EFI_CAPSULE_INITIATE_RESET because that would
66 * require us to prepare the kernel for reboot. Refuse to load any
67 * capsules with that flag and any other flags that we do not know how
68 * to handle.
70 #define EFI_CAPSULE_SUPPORTED_FLAG_MASK \
71 (EFI_CAPSULE_PERSIST_ACROSS_RESET | EFI_CAPSULE_POPULATE_SYSTEM_TABLE)
73 /**
74 * efi_capsule_supported - does the firmware support the capsule?
75 * @guid: vendor guid of capsule
76 * @flags: capsule flags
77 * @size: size of capsule data
78 * @reset: the reset type required for this capsule
80 * Check whether a capsule with @flags is supported by the firmware
81 * and that @size doesn't exceed the maximum size for a capsule.
83 * No attempt is made to check @reset against the reset type required
84 * by any pending capsules because of the races involved.
86 int efi_capsule_supported(efi_guid_t guid, u32 flags, size_t size, int *reset)
88 efi_capsule_header_t capsule;
89 efi_capsule_header_t *cap_list[] = { &capsule };
90 efi_status_t status;
91 u64 max_size;
93 if (flags & ~EFI_CAPSULE_SUPPORTED_FLAG_MASK)
94 return -EINVAL;
96 capsule.headersize = capsule.imagesize = sizeof(capsule);
97 memcpy(&capsule.guid, &guid, sizeof(efi_guid_t));
98 capsule.flags = flags;
100 status = efi.query_capsule_caps(cap_list, 1, &max_size, reset);
101 if (status != EFI_SUCCESS)
102 return efi_status_to_err(status);
104 if (size > max_size)
105 return -ENOSPC;
107 return 0;
109 EXPORT_SYMBOL_GPL(efi_capsule_supported);
112 * Every scatter gather list (block descriptor) page must end with a
113 * continuation pointer. The last continuation pointer of the last
114 * page must be zero to mark the end of the chain.
116 #define SGLIST_PER_PAGE ((PAGE_SIZE / sizeof(efi_capsule_block_desc_t)) - 1)
119 * How many scatter gather list (block descriptor) pages do we need
120 * to map @count pages?
122 static inline unsigned int sg_pages_num(unsigned int count)
124 return DIV_ROUND_UP(count, SGLIST_PER_PAGE);
128 * efi_capsule_update_locked - pass a single capsule to the firmware
129 * @capsule: capsule to send to the firmware
130 * @sg_pages: array of scatter gather (block descriptor) pages
131 * @reset: the reset type required for @capsule
133 * Since this function must be called under capsule_mutex check
134 * whether efi_reset_type will conflict with @reset, and atomically
135 * set it and capsule_pending if a capsule was successfully sent to
136 * the firmware.
138 * We also check to see if the system is about to restart, and if so,
139 * abort. This avoids races between efi_capsule_update() and
140 * efi_capsule_pending().
142 static int
143 efi_capsule_update_locked(efi_capsule_header_t *capsule,
144 struct page **sg_pages, int reset)
146 efi_physical_addr_t sglist_phys;
147 efi_status_t status;
149 lockdep_assert_held(&capsule_mutex);
152 * If someone has already registered a capsule that requires a
153 * different reset type, we're out of luck and must abort.
155 if (efi_reset_type >= 0 && efi_reset_type != reset) {
156 pr_err("Conflicting capsule reset type %d (%d).\n",
157 reset, efi_reset_type);
158 return -EINVAL;
162 * If the system is getting ready to restart it may have
163 * called efi_capsule_pending() to make decisions (such as
164 * whether to force an EFI reboot), and we're racing against
165 * that call. Abort in that case.
167 if (unlikely(stop_capsules)) {
168 pr_warn("Capsule update raced with reboot, aborting.\n");
169 return -EINVAL;
172 sglist_phys = page_to_phys(sg_pages[0]);
174 status = efi.update_capsule(&capsule, 1, sglist_phys);
175 if (status == EFI_SUCCESS) {
176 capsule_pending = true;
177 efi_reset_type = reset;
180 return efi_status_to_err(status);
184 * efi_capsule_update - send a capsule to the firmware
185 * @capsule: capsule to send to firmware
186 * @pages: an array of capsule data pages
188 * Build a scatter gather list with EFI capsule block descriptors to
189 * map the capsule described by @capsule with its data in @pages and
190 * send it to the firmware via the UpdateCapsule() runtime service.
192 * @capsule must be a virtual mapping of the complete capsule update in the
193 * kernel address space, as the capsule can be consumed immediately.
194 * A capsule_header_t that describes the entire contents of the capsule
195 * must be at the start of the first data page.
197 * Even though this function will validate that the firmware supports
198 * the capsule guid, users will likely want to check that
199 * efi_capsule_supported() returns true before calling this function
200 * because it makes it easier to print helpful error messages.
202 * If the capsule is successfully submitted to the firmware, any
203 * subsequent calls to efi_capsule_pending() will return true. @pages
204 * must not be released or modified if this function returns
205 * successfully.
207 * Callers must be prepared for this function to fail, which can
208 * happen if we raced with system reboot or if there is already a
209 * pending capsule that has a reset type that conflicts with the one
210 * required by @capsule. Do NOT use efi_capsule_pending() to detect
211 * this conflict since that would be racy. Instead, submit the capsule
212 * to efi_capsule_update() and check the return value.
214 * Return 0 on success, a converted EFI status code on failure.
216 int efi_capsule_update(efi_capsule_header_t *capsule, phys_addr_t *pages)
218 u32 imagesize = capsule->imagesize;
219 efi_guid_t guid = capsule->guid;
220 unsigned int count, sg_count;
221 u32 flags = capsule->flags;
222 struct page **sg_pages;
223 int rv, reset_type;
224 int i, j;
226 rv = efi_capsule_supported(guid, flags, imagesize, &reset_type);
227 if (rv)
228 return rv;
230 count = DIV_ROUND_UP(imagesize, PAGE_SIZE);
231 sg_count = sg_pages_num(count);
233 sg_pages = kcalloc(sg_count, sizeof(*sg_pages), GFP_KERNEL);
234 if (!sg_pages)
235 return -ENOMEM;
237 for (i = 0; i < sg_count; i++) {
238 sg_pages[i] = alloc_page(GFP_KERNEL);
239 if (!sg_pages[i]) {
240 rv = -ENOMEM;
241 goto out;
245 for (i = 0; i < sg_count; i++) {
246 efi_capsule_block_desc_t *sglist;
248 sglist = kmap_atomic(sg_pages[i]);
250 for (j = 0; j < SGLIST_PER_PAGE && count > 0; j++) {
251 u64 sz = min_t(u64, imagesize,
252 PAGE_SIZE - (u64)*pages % PAGE_SIZE);
254 sglist[j].length = sz;
255 sglist[j].data = *pages++;
257 imagesize -= sz;
258 count--;
261 /* Continuation pointer */
262 sglist[j].length = 0;
264 if (i + 1 == sg_count)
265 sglist[j].data = 0;
266 else
267 sglist[j].data = page_to_phys(sg_pages[i + 1]);
269 #if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
271 * At runtime, the firmware has no way to find out where the
272 * sglist elements are mapped, if they are mapped in the first
273 * place. Therefore, on architectures that can only perform
274 * cache maintenance by virtual address, the firmware is unable
275 * to perform this maintenance, and so it is up to the OS to do
276 * it instead.
278 efi_capsule_flush_cache_range(sglist, PAGE_SIZE);
279 #endif
280 kunmap_atomic(sglist);
283 mutex_lock(&capsule_mutex);
284 rv = efi_capsule_update_locked(capsule, sg_pages, reset_type);
285 mutex_unlock(&capsule_mutex);
287 out:
288 for (i = 0; rv && i < sg_count; i++) {
289 if (sg_pages[i])
290 __free_page(sg_pages[i]);
293 kfree(sg_pages);
294 return rv;
296 EXPORT_SYMBOL_GPL(efi_capsule_update);
298 static int capsule_reboot_notify(struct notifier_block *nb, unsigned long event, void *cmd)
300 mutex_lock(&capsule_mutex);
301 stop_capsules = true;
302 mutex_unlock(&capsule_mutex);
304 return NOTIFY_DONE;
307 static struct notifier_block capsule_reboot_nb = {
308 .notifier_call = capsule_reboot_notify,
311 static int __init capsule_reboot_register(void)
313 return register_reboot_notifier(&capsule_reboot_nb);
315 core_initcall(capsule_reboot_register);