]>
Commit | Line | Data |
---|---|---|
64db4cff PM |
1 | /* |
2 | * Read-Copy Update mechanism for mutual exclusion (tree-based version) | |
3 | * | |
4 | * This program is free software; you can redistribute it and/or modify | |
5 | * it under the terms of the GNU General Public License as published by | |
6 | * the Free Software Foundation; either version 2 of the License, or | |
7 | * (at your option) any later version. | |
8 | * | |
9 | * This program is distributed in the hope that it will be useful, | |
10 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
11 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
12 | * GNU General Public License for more details. | |
13 | * | |
14 | * You should have received a copy of the GNU General Public License | |
87de1cfd PM |
15 | * along with this program; if not, you can access it online at |
16 | * http://www.gnu.org/licenses/gpl-2.0.html. | |
64db4cff PM |
17 | * |
18 | * Copyright IBM Corporation, 2008 | |
19 | * | |
20 | * Author: Dipankar Sarma <dipankar@in.ibm.com> | |
21 | * Paul E. McKenney <paulmck@linux.vnet.ibm.com> Hierarchical algorithm | |
22 | * | |
23 | * Based on the original work by Paul McKenney <paulmck@us.ibm.com> | |
24 | * and inputs from Rusty Russell, Andrea Arcangeli and Andi Kleen. | |
25 | * | |
26 | * For detailed explanation of Read-Copy Update mechanism see - | |
a71fca58 | 27 | * Documentation/RCU |
64db4cff PM |
28 | */ |
29 | ||
30 | #ifndef __LINUX_RCUTREE_H | |
31 | #define __LINUX_RCUTREE_H | |
32 | ||
38200cf2 | 33 | void rcu_note_context_switch(void); |
c1ad348b | 34 | int rcu_needs_cpu(u64 basem, u64 *nextevt); |
584dc4ce | 35 | void rcu_cpu_stall_reset(void); |
64db4cff | 36 | |
29ce8310 GN |
37 | /* |
38 | * Note a virtualization-based context switch. This is simply a | |
39 | * wrapper around rcu_note_context_switch(), which allows TINY_RCU | |
40 | * to save a few bytes. | |
41 | */ | |
42 | static inline void rcu_virt_note_context_switch(int cpu) | |
43 | { | |
38200cf2 | 44 | rcu_note_context_switch(); |
29ce8310 GN |
45 | } |
46 | ||
584dc4ce TB |
47 | void synchronize_rcu_bh(void); |
48 | void synchronize_sched_expedited(void); | |
49 | void synchronize_rcu_expedited(void); | |
03b042bf | 50 | |
b6a4ae76 | 51 | void kfree_call_rcu(struct rcu_head *head, rcu_callback_t func); |
486e2593 | 52 | |
236fefaf PM |
53 | /** |
54 | * synchronize_rcu_bh_expedited - Brute-force RCU-bh grace period | |
55 | * | |
56 | * Wait for an RCU-bh grace period to elapse, but use a "big hammer" | |
57 | * approach to force the grace period to end quickly. This consumes | |
58 | * significant time on all CPUs and is unfriendly to real-time workloads, | |
59 | * so is thus not recommended for any sort of common-case code. In fact, | |
60 | * if you are using synchronize_rcu_bh_expedited() in a loop, please | |
61 | * restructure your code to batch your updates, and then use a single | |
62 | * synchronize_rcu_bh() instead. | |
63 | * | |
64 | * Note that it is illegal to call this function while holding any lock | |
65 | * that is acquired by a CPU-hotplug notifier. And yes, it is also illegal | |
66 | * to call this function from a CPU-hotplug notifier. Failing to observe | |
67 | * these restriction will result in deadlock. | |
68 | */ | |
03b042bf | 69 | static inline void synchronize_rcu_bh_expedited(void) |
64db4cff | 70 | { |
03b042bf | 71 | synchronize_sched_expedited(); |
64db4cff PM |
72 | } |
73 | ||
584dc4ce TB |
74 | void rcu_barrier(void); |
75 | void rcu_barrier_bh(void); | |
76 | void rcu_barrier_sched(void); | |
765a3f4f PM |
77 | unsigned long get_state_synchronize_rcu(void); |
78 | void cond_synchronize_rcu(unsigned long oldstate); | |
24560056 PM |
79 | unsigned long get_state_synchronize_sched(void); |
80 | void cond_synchronize_sched(unsigned long oldstate); | |
a57eb940 | 81 | |
4a298656 PM |
82 | extern unsigned long rcutorture_testseq; |
83 | extern unsigned long rcutorture_vernum; | |
917963d0 PM |
84 | unsigned long rcu_batches_started(void); |
85 | unsigned long rcu_batches_started_bh(void); | |
86 | unsigned long rcu_batches_started_sched(void); | |
9733e4f0 PM |
87 | unsigned long rcu_batches_completed(void); |
88 | unsigned long rcu_batches_completed_bh(void); | |
89 | unsigned long rcu_batches_completed_sched(void); | |
afea227f | 90 | void show_rcu_gp_kthreads(void); |
4a298656 | 91 | |
584dc4ce TB |
92 | void rcu_force_quiescent_state(void); |
93 | void rcu_bh_force_quiescent_state(void); | |
94 | void rcu_sched_force_quiescent_state(void); | |
64db4cff | 95 | |
51952bc6 PM |
96 | void rcu_idle_enter(void); |
97 | void rcu_idle_exit(void); | |
98 | void rcu_irq_enter(void); | |
99 | void rcu_irq_exit(void); | |
100 | ||
584dc4ce | 101 | void exit_rcu(void); |
2439b696 | 102 | |
584dc4ce | 103 | void rcu_scheduler_starting(void); |
bbad9379 PM |
104 | extern int rcu_scheduler_active __read_mostly; |
105 | ||
584dc4ce | 106 | bool rcu_is_watching(void); |
cc6783f7 | 107 | |
5cd37193 PM |
108 | void rcu_all_qs(void); |
109 | ||
64db4cff | 110 | #endif /* __LINUX_RCUTREE_H */ |