]> git.proxmox.com Git - mirror_ubuntu-jammy-kernel.git/blob - include/linux/jump_label.h
Merge tag 'spi-fix-v5.14-rc2' of git://git.kernel.org/pub/scm/linux/kernel/git/brooni...
[mirror_ubuntu-jammy-kernel.git] / include / linux / jump_label.h
1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef _LINUX_JUMP_LABEL_H
3 #define _LINUX_JUMP_LABEL_H
4
5 /*
6 * Jump label support
7 *
8 * Copyright (C) 2009-2012 Jason Baron <jbaron@redhat.com>
9 * Copyright (C) 2011-2012 Red Hat, Inc., Peter Zijlstra
10 *
11 * DEPRECATED API:
12 *
13 * The use of 'struct static_key' directly, is now DEPRECATED. In addition
14 * static_key_{true,false}() is also DEPRECATED. IE DO NOT use the following:
15 *
16 * struct static_key false = STATIC_KEY_INIT_FALSE;
17 * struct static_key true = STATIC_KEY_INIT_TRUE;
18 * static_key_true()
19 * static_key_false()
20 *
21 * The updated API replacements are:
22 *
23 * DEFINE_STATIC_KEY_TRUE(key);
24 * DEFINE_STATIC_KEY_FALSE(key);
25 * DEFINE_STATIC_KEY_ARRAY_TRUE(keys, count);
26 * DEFINE_STATIC_KEY_ARRAY_FALSE(keys, count);
27 * static_branch_likely()
28 * static_branch_unlikely()
29 *
30 * Jump labels provide an interface to generate dynamic branches using
31 * self-modifying code. Assuming toolchain and architecture support, if we
32 * define a "key" that is initially false via "DEFINE_STATIC_KEY_FALSE(key)",
33 * an "if (static_branch_unlikely(&key))" statement is an unconditional branch
34 * (which defaults to false - and the true block is placed out of line).
35 * Similarly, we can define an initially true key via
36 * "DEFINE_STATIC_KEY_TRUE(key)", and use it in the same
37 * "if (static_branch_unlikely(&key))", in which case we will generate an
38 * unconditional branch to the out-of-line true branch. Keys that are
39 * initially true or false can be using in both static_branch_unlikely()
40 * and static_branch_likely() statements.
41 *
42 * At runtime we can change the branch target by setting the key
43 * to true via a call to static_branch_enable(), or false using
44 * static_branch_disable(). If the direction of the branch is switched by
45 * these calls then we run-time modify the branch target via a
46 * no-op -> jump or jump -> no-op conversion. For example, for an
47 * initially false key that is used in an "if (static_branch_unlikely(&key))"
48 * statement, setting the key to true requires us to patch in a jump
49 * to the out-of-line of true branch.
50 *
51 * In addition to static_branch_{enable,disable}, we can also reference count
52 * the key or branch direction via static_branch_{inc,dec}. Thus,
53 * static_branch_inc() can be thought of as a 'make more true' and
54 * static_branch_dec() as a 'make more false'.
55 *
56 * Since this relies on modifying code, the branch modifying functions
57 * must be considered absolute slow paths (machine wide synchronization etc.).
58 * OTOH, since the affected branches are unconditional, their runtime overhead
59 * will be absolutely minimal, esp. in the default (off) case where the total
60 * effect is a single NOP of appropriate size. The on case will patch in a jump
61 * to the out-of-line block.
62 *
63 * When the control is directly exposed to userspace, it is prudent to delay the
64 * decrement to avoid high frequency code modifications which can (and do)
65 * cause significant performance degradation. Struct static_key_deferred and
66 * static_key_slow_dec_deferred() provide for this.
67 *
68 * Lacking toolchain and or architecture support, static keys fall back to a
69 * simple conditional branch.
70 *
71 * Additional babbling in: Documentation/staging/static-keys.rst
72 */
73
74 #ifndef __ASSEMBLY__
75
76 #include <linux/types.h>
77 #include <linux/compiler.h>
78
79 extern bool static_key_initialized;
80
81 #define STATIC_KEY_CHECK_USE(key) WARN(!static_key_initialized, \
82 "%s(): static key '%pS' used before call to jump_label_init()", \
83 __func__, (key))
84
85 #ifdef CONFIG_JUMP_LABEL
86
87 struct static_key {
88 atomic_t enabled;
89 /*
90 * Note:
91 * To make anonymous unions work with old compilers, the static
92 * initialization of them requires brackets. This creates a dependency
93 * on the order of the struct with the initializers. If any fields
94 * are added, STATIC_KEY_INIT_TRUE and STATIC_KEY_INIT_FALSE may need
95 * to be modified.
96 *
97 * bit 0 => 1 if key is initially true
98 * 0 if initially false
99 * bit 1 => 1 if points to struct static_key_mod
100 * 0 if points to struct jump_entry
101 */
102 union {
103 unsigned long type;
104 struct jump_entry *entries;
105 struct static_key_mod *next;
106 };
107 };
108
109 #else
110 struct static_key {
111 atomic_t enabled;
112 };
113 #endif /* CONFIG_JUMP_LABEL */
114 #endif /* __ASSEMBLY__ */
115
116 #ifdef CONFIG_JUMP_LABEL
117 #include <asm/jump_label.h>
118
119 #ifndef __ASSEMBLY__
120 #ifdef CONFIG_HAVE_ARCH_JUMP_LABEL_RELATIVE
121
122 struct jump_entry {
123 s32 code;
124 s32 target;
125 long key; // key may be far away from the core kernel under KASLR
126 };
127
128 static inline unsigned long jump_entry_code(const struct jump_entry *entry)
129 {
130 return (unsigned long)&entry->code + entry->code;
131 }
132
133 static inline unsigned long jump_entry_target(const struct jump_entry *entry)
134 {
135 return (unsigned long)&entry->target + entry->target;
136 }
137
138 static inline struct static_key *jump_entry_key(const struct jump_entry *entry)
139 {
140 long offset = entry->key & ~3L;
141
142 return (struct static_key *)((unsigned long)&entry->key + offset);
143 }
144
145 #else
146
147 static inline unsigned long jump_entry_code(const struct jump_entry *entry)
148 {
149 return entry->code;
150 }
151
152 static inline unsigned long jump_entry_target(const struct jump_entry *entry)
153 {
154 return entry->target;
155 }
156
157 static inline struct static_key *jump_entry_key(const struct jump_entry *entry)
158 {
159 return (struct static_key *)((unsigned long)entry->key & ~3UL);
160 }
161
162 #endif
163
164 static inline bool jump_entry_is_branch(const struct jump_entry *entry)
165 {
166 return (unsigned long)entry->key & 1UL;
167 }
168
169 static inline bool jump_entry_is_init(const struct jump_entry *entry)
170 {
171 return (unsigned long)entry->key & 2UL;
172 }
173
174 static inline void jump_entry_set_init(struct jump_entry *entry, bool set)
175 {
176 if (set)
177 entry->key |= 2;
178 else
179 entry->key &= ~2;
180 }
181
182 static inline int jump_entry_size(struct jump_entry *entry)
183 {
184 #ifdef JUMP_LABEL_NOP_SIZE
185 return JUMP_LABEL_NOP_SIZE;
186 #else
187 return arch_jump_entry_size(entry);
188 #endif
189 }
190
191 #endif
192 #endif
193
194 #ifndef __ASSEMBLY__
195
196 enum jump_label_type {
197 JUMP_LABEL_NOP = 0,
198 JUMP_LABEL_JMP,
199 };
200
201 struct module;
202
203 #ifdef CONFIG_JUMP_LABEL
204
205 #define JUMP_TYPE_FALSE 0UL
206 #define JUMP_TYPE_TRUE 1UL
207 #define JUMP_TYPE_LINKED 2UL
208 #define JUMP_TYPE_MASK 3UL
209
210 static __always_inline bool static_key_false(struct static_key *key)
211 {
212 return arch_static_branch(key, false);
213 }
214
215 static __always_inline bool static_key_true(struct static_key *key)
216 {
217 return !arch_static_branch(key, true);
218 }
219
220 extern struct jump_entry __start___jump_table[];
221 extern struct jump_entry __stop___jump_table[];
222
223 extern void jump_label_init(void);
224 extern void jump_label_lock(void);
225 extern void jump_label_unlock(void);
226 extern void arch_jump_label_transform(struct jump_entry *entry,
227 enum jump_label_type type);
228 extern void arch_jump_label_transform_static(struct jump_entry *entry,
229 enum jump_label_type type);
230 extern bool arch_jump_label_transform_queue(struct jump_entry *entry,
231 enum jump_label_type type);
232 extern void arch_jump_label_transform_apply(void);
233 extern int jump_label_text_reserved(void *start, void *end);
234 extern void static_key_slow_inc(struct static_key *key);
235 extern void static_key_slow_dec(struct static_key *key);
236 extern void static_key_slow_inc_cpuslocked(struct static_key *key);
237 extern void static_key_slow_dec_cpuslocked(struct static_key *key);
238 extern void jump_label_apply_nops(struct module *mod);
239 extern int static_key_count(struct static_key *key);
240 extern void static_key_enable(struct static_key *key);
241 extern void static_key_disable(struct static_key *key);
242 extern void static_key_enable_cpuslocked(struct static_key *key);
243 extern void static_key_disable_cpuslocked(struct static_key *key);
244
245 /*
246 * We should be using ATOMIC_INIT() for initializing .enabled, but
247 * the inclusion of atomic.h is problematic for inclusion of jump_label.h
248 * in 'low-level' headers. Thus, we are initializing .enabled with a
249 * raw value, but have added a BUILD_BUG_ON() to catch any issues in
250 * jump_label_init() see: kernel/jump_label.c.
251 */
252 #define STATIC_KEY_INIT_TRUE \
253 { .enabled = { 1 }, \
254 { .entries = (void *)JUMP_TYPE_TRUE } }
255 #define STATIC_KEY_INIT_FALSE \
256 { .enabled = { 0 }, \
257 { .entries = (void *)JUMP_TYPE_FALSE } }
258
259 #else /* !CONFIG_JUMP_LABEL */
260
261 #include <linux/atomic.h>
262 #include <linux/bug.h>
263
264 static inline int static_key_count(struct static_key *key)
265 {
266 return atomic_read(&key->enabled);
267 }
268
269 static __always_inline void jump_label_init(void)
270 {
271 static_key_initialized = true;
272 }
273
274 static __always_inline bool static_key_false(struct static_key *key)
275 {
276 if (unlikely_notrace(static_key_count(key) > 0))
277 return true;
278 return false;
279 }
280
281 static __always_inline bool static_key_true(struct static_key *key)
282 {
283 if (likely_notrace(static_key_count(key) > 0))
284 return true;
285 return false;
286 }
287
288 static inline void static_key_slow_inc(struct static_key *key)
289 {
290 STATIC_KEY_CHECK_USE(key);
291 atomic_inc(&key->enabled);
292 }
293
294 static inline void static_key_slow_dec(struct static_key *key)
295 {
296 STATIC_KEY_CHECK_USE(key);
297 atomic_dec(&key->enabled);
298 }
299
300 #define static_key_slow_inc_cpuslocked(key) static_key_slow_inc(key)
301 #define static_key_slow_dec_cpuslocked(key) static_key_slow_dec(key)
302
303 static inline int jump_label_text_reserved(void *start, void *end)
304 {
305 return 0;
306 }
307
308 static inline void jump_label_lock(void) {}
309 static inline void jump_label_unlock(void) {}
310
311 static inline int jump_label_apply_nops(struct module *mod)
312 {
313 return 0;
314 }
315
316 static inline void static_key_enable(struct static_key *key)
317 {
318 STATIC_KEY_CHECK_USE(key);
319
320 if (atomic_read(&key->enabled) != 0) {
321 WARN_ON_ONCE(atomic_read(&key->enabled) != 1);
322 return;
323 }
324 atomic_set(&key->enabled, 1);
325 }
326
327 static inline void static_key_disable(struct static_key *key)
328 {
329 STATIC_KEY_CHECK_USE(key);
330
331 if (atomic_read(&key->enabled) != 1) {
332 WARN_ON_ONCE(atomic_read(&key->enabled) != 0);
333 return;
334 }
335 atomic_set(&key->enabled, 0);
336 }
337
338 #define static_key_enable_cpuslocked(k) static_key_enable((k))
339 #define static_key_disable_cpuslocked(k) static_key_disable((k))
340
341 #define STATIC_KEY_INIT_TRUE { .enabled = ATOMIC_INIT(1) }
342 #define STATIC_KEY_INIT_FALSE { .enabled = ATOMIC_INIT(0) }
343
344 #endif /* CONFIG_JUMP_LABEL */
345
346 #define STATIC_KEY_INIT STATIC_KEY_INIT_FALSE
347 #define jump_label_enabled static_key_enabled
348
349 /* -------------------------------------------------------------------------- */
350
351 /*
352 * Two type wrappers around static_key, such that we can use compile time
353 * type differentiation to emit the right code.
354 *
355 * All the below code is macros in order to play type games.
356 */
357
358 struct static_key_true {
359 struct static_key key;
360 };
361
362 struct static_key_false {
363 struct static_key key;
364 };
365
366 #define STATIC_KEY_TRUE_INIT (struct static_key_true) { .key = STATIC_KEY_INIT_TRUE, }
367 #define STATIC_KEY_FALSE_INIT (struct static_key_false){ .key = STATIC_KEY_INIT_FALSE, }
368
369 #define DEFINE_STATIC_KEY_TRUE(name) \
370 struct static_key_true name = STATIC_KEY_TRUE_INIT
371
372 #define DEFINE_STATIC_KEY_TRUE_RO(name) \
373 struct static_key_true name __ro_after_init = STATIC_KEY_TRUE_INIT
374
375 #define DECLARE_STATIC_KEY_TRUE(name) \
376 extern struct static_key_true name
377
378 #define DEFINE_STATIC_KEY_FALSE(name) \
379 struct static_key_false name = STATIC_KEY_FALSE_INIT
380
381 #define DEFINE_STATIC_KEY_FALSE_RO(name) \
382 struct static_key_false name __ro_after_init = STATIC_KEY_FALSE_INIT
383
384 #define DECLARE_STATIC_KEY_FALSE(name) \
385 extern struct static_key_false name
386
387 #define DEFINE_STATIC_KEY_ARRAY_TRUE(name, count) \
388 struct static_key_true name[count] = { \
389 [0 ... (count) - 1] = STATIC_KEY_TRUE_INIT, \
390 }
391
392 #define DEFINE_STATIC_KEY_ARRAY_FALSE(name, count) \
393 struct static_key_false name[count] = { \
394 [0 ... (count) - 1] = STATIC_KEY_FALSE_INIT, \
395 }
396
397 #define _DEFINE_STATIC_KEY_1(name) DEFINE_STATIC_KEY_TRUE(name)
398 #define _DEFINE_STATIC_KEY_0(name) DEFINE_STATIC_KEY_FALSE(name)
399 #define DEFINE_STATIC_KEY_MAYBE(cfg, name) \
400 __PASTE(_DEFINE_STATIC_KEY_, IS_ENABLED(cfg))(name)
401
402 #define _DEFINE_STATIC_KEY_RO_1(name) DEFINE_STATIC_KEY_TRUE_RO(name)
403 #define _DEFINE_STATIC_KEY_RO_0(name) DEFINE_STATIC_KEY_FALSE_RO(name)
404 #define DEFINE_STATIC_KEY_MAYBE_RO(cfg, name) \
405 __PASTE(_DEFINE_STATIC_KEY_RO_, IS_ENABLED(cfg))(name)
406
407 #define _DECLARE_STATIC_KEY_1(name) DECLARE_STATIC_KEY_TRUE(name)
408 #define _DECLARE_STATIC_KEY_0(name) DECLARE_STATIC_KEY_FALSE(name)
409 #define DECLARE_STATIC_KEY_MAYBE(cfg, name) \
410 __PASTE(_DECLARE_STATIC_KEY_, IS_ENABLED(cfg))(name)
411
412 extern bool ____wrong_branch_error(void);
413
414 #define static_key_enabled(x) \
415 ({ \
416 if (!__builtin_types_compatible_p(typeof(*x), struct static_key) && \
417 !__builtin_types_compatible_p(typeof(*x), struct static_key_true) &&\
418 !__builtin_types_compatible_p(typeof(*x), struct static_key_false)) \
419 ____wrong_branch_error(); \
420 static_key_count((struct static_key *)x) > 0; \
421 })
422
423 #ifdef CONFIG_JUMP_LABEL
424
425 /*
426 * Combine the right initial value (type) with the right branch order
427 * to generate the desired result.
428 *
429 *
430 * type\branch| likely (1) | unlikely (0)
431 * -----------+-----------------------+------------------
432 * | |
433 * true (1) | ... | ...
434 * | NOP | JMP L
435 * | <br-stmts> | 1: ...
436 * | L: ... |
437 * | |
438 * | | L: <br-stmts>
439 * | | jmp 1b
440 * | |
441 * -----------+-----------------------+------------------
442 * | |
443 * false (0) | ... | ...
444 * | JMP L | NOP
445 * | <br-stmts> | 1: ...
446 * | L: ... |
447 * | |
448 * | | L: <br-stmts>
449 * | | jmp 1b
450 * | |
451 * -----------+-----------------------+------------------
452 *
453 * The initial value is encoded in the LSB of static_key::entries,
454 * type: 0 = false, 1 = true.
455 *
456 * The branch type is encoded in the LSB of jump_entry::key,
457 * branch: 0 = unlikely, 1 = likely.
458 *
459 * This gives the following logic table:
460 *
461 * enabled type branch instuction
462 * -----------------------------+-----------
463 * 0 0 0 | NOP
464 * 0 0 1 | JMP
465 * 0 1 0 | NOP
466 * 0 1 1 | JMP
467 *
468 * 1 0 0 | JMP
469 * 1 0 1 | NOP
470 * 1 1 0 | JMP
471 * 1 1 1 | NOP
472 *
473 * Which gives the following functions:
474 *
475 * dynamic: instruction = enabled ^ branch
476 * static: instruction = type ^ branch
477 *
478 * See jump_label_type() / jump_label_init_type().
479 */
480
481 #define static_branch_likely(x) \
482 ({ \
483 bool branch; \
484 if (__builtin_types_compatible_p(typeof(*x), struct static_key_true)) \
485 branch = !arch_static_branch(&(x)->key, true); \
486 else if (__builtin_types_compatible_p(typeof(*x), struct static_key_false)) \
487 branch = !arch_static_branch_jump(&(x)->key, true); \
488 else \
489 branch = ____wrong_branch_error(); \
490 likely_notrace(branch); \
491 })
492
493 #define static_branch_unlikely(x) \
494 ({ \
495 bool branch; \
496 if (__builtin_types_compatible_p(typeof(*x), struct static_key_true)) \
497 branch = arch_static_branch_jump(&(x)->key, false); \
498 else if (__builtin_types_compatible_p(typeof(*x), struct static_key_false)) \
499 branch = arch_static_branch(&(x)->key, false); \
500 else \
501 branch = ____wrong_branch_error(); \
502 unlikely_notrace(branch); \
503 })
504
505 #else /* !CONFIG_JUMP_LABEL */
506
507 #define static_branch_likely(x) likely_notrace(static_key_enabled(&(x)->key))
508 #define static_branch_unlikely(x) unlikely_notrace(static_key_enabled(&(x)->key))
509
510 #endif /* CONFIG_JUMP_LABEL */
511
512 #define static_branch_maybe(config, x) \
513 (IS_ENABLED(config) ? static_branch_likely(x) \
514 : static_branch_unlikely(x))
515
516 /*
517 * Advanced usage; refcount, branch is enabled when: count != 0
518 */
519
520 #define static_branch_inc(x) static_key_slow_inc(&(x)->key)
521 #define static_branch_dec(x) static_key_slow_dec(&(x)->key)
522 #define static_branch_inc_cpuslocked(x) static_key_slow_inc_cpuslocked(&(x)->key)
523 #define static_branch_dec_cpuslocked(x) static_key_slow_dec_cpuslocked(&(x)->key)
524
525 /*
526 * Normal usage; boolean enable/disable.
527 */
528
529 #define static_branch_enable(x) static_key_enable(&(x)->key)
530 #define static_branch_disable(x) static_key_disable(&(x)->key)
531 #define static_branch_enable_cpuslocked(x) static_key_enable_cpuslocked(&(x)->key)
532 #define static_branch_disable_cpuslocked(x) static_key_disable_cpuslocked(&(x)->key)
533
534 #endif /* __ASSEMBLY__ */
535
536 #endif /* _LINUX_JUMP_LABEL_H */