]>
Commit | Line | Data |
---|---|---|
ddbcc7e8 PM |
1 | #ifndef _LINUX_CGROUP_H |
2 | #define _LINUX_CGROUP_H | |
3 | /* | |
4 | * cgroup interface | |
5 | * | |
6 | * Copyright (C) 2003 BULL SA | |
7 | * Copyright (C) 2004-2006 Silicon Graphics, Inc. | |
8 | * | |
9 | */ | |
10 | ||
11 | #include <linux/sched.h> | |
ddbcc7e8 PM |
12 | #include <linux/cpumask.h> |
13 | #include <linux/nodemask.h> | |
14 | #include <linux/rcupdate.h> | |
eb6fd504 | 15 | #include <linux/rculist.h> |
846c7bb0 | 16 | #include <linux/cgroupstats.h> |
31a7df01 | 17 | #include <linux/prio_heap.h> |
cc31edce | 18 | #include <linux/rwsem.h> |
38460b48 | 19 | #include <linux/idr.h> |
48ddbe19 | 20 | #include <linux/workqueue.h> |
03b1cde6 | 21 | #include <linux/xattr.h> |
25a7e684 | 22 | #include <linux/fs.h> |
d3daf28d | 23 | #include <linux/percpu-refcount.h> |
ddbcc7e8 PM |
24 | |
25 | #ifdef CONFIG_CGROUPS | |
26 | ||
27 | struct cgroupfs_root; | |
28 | struct cgroup_subsys; | |
29 | struct inode; | |
84eea842 | 30 | struct cgroup; |
38460b48 | 31 | struct css_id; |
a27bb332 | 32 | struct eventfd_ctx; |
ddbcc7e8 PM |
33 | |
34 | extern int cgroup_init_early(void); | |
35 | extern int cgroup_init(void); | |
b4f48b63 | 36 | extern void cgroup_fork(struct task_struct *p); |
817929ec | 37 | extern void cgroup_post_fork(struct task_struct *p); |
b4f48b63 | 38 | extern void cgroup_exit(struct task_struct *p, int run_callbacks); |
846c7bb0 BS |
39 | extern int cgroupstats_build(struct cgroupstats *stats, |
40 | struct dentry *dentry); | |
e6a1105b | 41 | extern int cgroup_load_subsys(struct cgroup_subsys *ss); |
cf5d5941 | 42 | extern void cgroup_unload_subsys(struct cgroup_subsys *ss); |
ddbcc7e8 | 43 | |
8d8b97ba | 44 | extern int proc_cgroup_show(struct seq_file *, void *); |
a424316c | 45 | |
7d8e0bf5 LZ |
46 | /* |
47 | * Define the enumeration of all cgroup subsystems. | |
48 | * | |
49 | * We define ids for builtin subsystems and then modular ones. | |
50 | */ | |
817929ec PM |
51 | #define SUBSYS(_x) _x ## _subsys_id, |
52 | enum cgroup_subsys_id { | |
7d8e0bf5 | 53 | #define IS_SUBSYS_ENABLED(option) IS_BUILTIN(option) |
817929ec | 54 | #include <linux/cgroup_subsys.h> |
7d8e0bf5 LZ |
55 | #undef IS_SUBSYS_ENABLED |
56 | CGROUP_BUILTIN_SUBSYS_COUNT, | |
57 | ||
58 | __CGROUP_SUBSYS_TEMP_PLACEHOLDER = CGROUP_BUILTIN_SUBSYS_COUNT - 1, | |
59 | ||
60 | #define IS_SUBSYS_ENABLED(option) IS_MODULE(option) | |
817929ec | 61 | #include <linux/cgroup_subsys.h> |
7d8e0bf5 | 62 | #undef IS_SUBSYS_ENABLED |
a6f00298 | 63 | CGROUP_SUBSYS_COUNT, |
817929ec PM |
64 | }; |
65 | #undef SUBSYS | |
66 | ||
ddbcc7e8 PM |
67 | /* Per-subsystem/per-cgroup state maintained by the system. */ |
68 | struct cgroup_subsys_state { | |
72c97e54 | 69 | /* the cgroup that this css is attached to */ |
ddbcc7e8 PM |
70 | struct cgroup *cgroup; |
71 | ||
72c97e54 TH |
72 | /* the cgroup subsystem that this css is attached to */ |
73 | struct cgroup_subsys *ss; | |
74 | ||
d3daf28d TH |
75 | /* reference count - access via css_[try]get() and css_put() */ |
76 | struct percpu_ref refcnt; | |
ddbcc7e8 PM |
77 | |
78 | unsigned long flags; | |
38460b48 | 79 | /* ID for this css, if possible */ |
2c392b8c | 80 | struct css_id __rcu *id; |
48ddbe19 TH |
81 | |
82 | /* Used to put @cgroup->dentry on the last css_put() */ | |
83 | struct work_struct dput_work; | |
ddbcc7e8 PM |
84 | }; |
85 | ||
86 | /* bits in struct cgroup_subsys_state flags field */ | |
87 | enum { | |
38b53aba | 88 | CSS_ROOT = (1 << 0), /* this CSS is the root of the subsystem */ |
92fb9748 | 89 | CSS_ONLINE = (1 << 1), /* between ->css_online() and ->css_offline() */ |
ddbcc7e8 PM |
90 | }; |
91 | ||
5de0107e TH |
92 | /** |
93 | * css_get - obtain a reference on the specified css | |
94 | * @css: target css | |
95 | * | |
96 | * The caller must already have a reference. | |
ddbcc7e8 | 97 | */ |
ddbcc7e8 PM |
98 | static inline void css_get(struct cgroup_subsys_state *css) |
99 | { | |
100 | /* We don't need to reference count the root state */ | |
38b53aba | 101 | if (!(css->flags & CSS_ROOT)) |
d3daf28d | 102 | percpu_ref_get(&css->refcnt); |
ddbcc7e8 | 103 | } |
e7c5ec91 | 104 | |
5de0107e TH |
105 | /** |
106 | * css_tryget - try to obtain a reference on the specified css | |
107 | * @css: target css | |
108 | * | |
109 | * Obtain a reference on @css if it's alive. The caller naturally needs to | |
110 | * ensure that @css is accessible but doesn't have to be holding a | |
111 | * reference on it - IOW, RCU protected access is good enough for this | |
112 | * function. Returns %true if a reference count was successfully obtained; | |
113 | * %false otherwise. | |
114 | */ | |
e7c5ec91 PM |
115 | static inline bool css_tryget(struct cgroup_subsys_state *css) |
116 | { | |
38b53aba | 117 | if (css->flags & CSS_ROOT) |
e7c5ec91 | 118 | return true; |
d3daf28d | 119 | return percpu_ref_tryget(&css->refcnt); |
e7c5ec91 PM |
120 | } |
121 | ||
5de0107e TH |
122 | /** |
123 | * css_put - put a css reference | |
124 | * @css: target css | |
125 | * | |
126 | * Put a reference obtained via css_get() and css_tryget(). | |
127 | */ | |
ddbcc7e8 PM |
128 | static inline void css_put(struct cgroup_subsys_state *css) |
129 | { | |
38b53aba | 130 | if (!(css->flags & CSS_ROOT)) |
d3daf28d | 131 | percpu_ref_put(&css->refcnt); |
ddbcc7e8 PM |
132 | } |
133 | ||
3116f0e3 PM |
134 | /* bits in struct cgroup flags field */ |
135 | enum { | |
136 | /* Control Group is dead */ | |
54766d4a | 137 | CGRP_DEAD, |
d20a390a PM |
138 | /* |
139 | * Control Group has previously had a child cgroup or a task, | |
140 | * but no longer (only if CGRP_NOTIFY_ON_RELEASE is set) | |
141 | */ | |
3116f0e3 PM |
142 | CGRP_RELEASABLE, |
143 | /* Control Group requires release notifications to userspace */ | |
144 | CGRP_NOTIFY_ON_RELEASE, | |
97978e6d | 145 | /* |
2260e7fc TH |
146 | * Clone the parent's configuration when creating a new child |
147 | * cpuset cgroup. For historical reasons, this option can be | |
148 | * specified at mount time and thus is implemented here. | |
97978e6d | 149 | */ |
2260e7fc | 150 | CGRP_CPUSET_CLONE_CHILDREN, |
873fe09e TH |
151 | /* see the comment above CGRP_ROOT_SANE_BEHAVIOR for details */ |
152 | CGRP_SANE_BEHAVIOR, | |
3116f0e3 PM |
153 | }; |
154 | ||
65dff759 LZ |
155 | struct cgroup_name { |
156 | struct rcu_head rcu_head; | |
157 | char name[]; | |
3116f0e3 PM |
158 | }; |
159 | ||
ddbcc7e8 PM |
160 | struct cgroup { |
161 | unsigned long flags; /* "unsigned long" so bitops work */ | |
162 | ||
b414dc09 LZ |
163 | /* |
164 | * idr allocated in-hierarchy ID. | |
165 | * | |
166 | * The ID of the root cgroup is always 0, and a new cgroup | |
167 | * will be assigned with a smallest available ID. | |
168 | */ | |
169 | int id; | |
0a950f65 | 170 | |
ddbcc7e8 PM |
171 | /* |
172 | * We link our 'sibling' struct into our parent's 'children'. | |
173 | * Our children link their 'sibling' into our 'children'. | |
174 | */ | |
175 | struct list_head sibling; /* my parent's children */ | |
176 | struct list_head children; /* my children */ | |
05ef1d7c | 177 | struct list_head files; /* my files */ |
ddbcc7e8 | 178 | |
d20a390a | 179 | struct cgroup *parent; /* my parent */ |
febfcef6 | 180 | struct dentry *dentry; /* cgroup fs entry, RCU protected */ |
ddbcc7e8 | 181 | |
53fa5261 TH |
182 | /* |
183 | * Monotonically increasing unique serial number which defines a | |
184 | * uniform order among all cgroups. It's guaranteed that all | |
185 | * ->children lists are in the ascending order of ->serial_nr. | |
186 | * It's used to allow interrupting and resuming iterations. | |
187 | */ | |
188 | u64 serial_nr; | |
189 | ||
65dff759 LZ |
190 | /* |
191 | * This is a copy of dentry->d_name, and it's needed because | |
192 | * we can't use dentry->d_name in cgroup_path(). | |
193 | * | |
194 | * You must acquire rcu_read_lock() to access cgrp->name, and | |
195 | * the only place that can change it is rename(), which is | |
196 | * protected by parent dir's i_mutex. | |
197 | * | |
198 | * Normally you should use cgroup_name() wrapper rather than | |
199 | * access it directly. | |
200 | */ | |
201 | struct cgroup_name __rcu *name; | |
202 | ||
ddbcc7e8 PM |
203 | /* Private pointers for each registered subsystem */ |
204 | struct cgroup_subsys_state *subsys[CGROUP_SUBSYS_COUNT]; | |
205 | ||
206 | struct cgroupfs_root *root; | |
817929ec PM |
207 | |
208 | /* | |
69d0206c TH |
209 | * List of cgrp_cset_links pointing at css_sets with tasks in this |
210 | * cgroup. Protected by css_set_lock. | |
817929ec | 211 | */ |
69d0206c | 212 | struct list_head cset_links; |
81a6a5cd PM |
213 | |
214 | /* | |
215 | * Linked list running through all cgroups that can | |
216 | * potentially be reaped by the release agent. Protected by | |
217 | * release_list_lock | |
218 | */ | |
219 | struct list_head release_list; | |
cc31edce | 220 | |
72a8cb30 BB |
221 | /* |
222 | * list of pidlists, up to two for each namespace (one for procs, one | |
223 | * for tasks); created on demand. | |
224 | */ | |
225 | struct list_head pidlists; | |
226 | struct mutex pidlist_mutex; | |
a47295e6 | 227 | |
67f4c36f TH |
228 | /* dummy css with NULL ->ss, points back to this cgroup */ |
229 | struct cgroup_subsys_state dummy_css; | |
230 | ||
d3daf28d | 231 | /* For css percpu_ref killing and RCU-protected deletion */ |
a47295e6 | 232 | struct rcu_head rcu_head; |
ea15f8cc | 233 | struct work_struct destroy_work; |
d3daf28d | 234 | atomic_t css_kill_cnt; |
0dea1168 | 235 | |
25985edc | 236 | /* List of events which userspace want to receive */ |
0dea1168 KS |
237 | struct list_head event_list; |
238 | spinlock_t event_list_lock; | |
03b1cde6 AR |
239 | |
240 | /* directory xattrs */ | |
241 | struct simple_xattrs xattrs; | |
817929ec PM |
242 | }; |
243 | ||
25a7e684 TH |
244 | #define MAX_CGROUP_ROOT_NAMELEN 64 |
245 | ||
246 | /* cgroupfs_root->flags */ | |
247 | enum { | |
873fe09e TH |
248 | /* |
249 | * Unfortunately, cgroup core and various controllers are riddled | |
250 | * with idiosyncrasies and pointless options. The following flag, | |
251 | * when set, will force sane behavior - some options are forced on, | |
252 | * others are disallowed, and some controllers will change their | |
253 | * hierarchical or other behaviors. | |
254 | * | |
255 | * The set of behaviors affected by this flag are still being | |
256 | * determined and developed and the mount option for this flag is | |
257 | * prefixed with __DEVEL__. The prefix will be dropped once we | |
258 | * reach the point where all behaviors are compatible with the | |
259 | * planned unified hierarchy, which will automatically turn on this | |
260 | * flag. | |
261 | * | |
262 | * The followings are the behaviors currently affected this flag. | |
263 | * | |
264 | * - Mount options "noprefix" and "clone_children" are disallowed. | |
265 | * Also, cgroupfs file cgroup.clone_children is not created. | |
266 | * | |
267 | * - When mounting an existing superblock, mount options should | |
268 | * match. | |
269 | * | |
270 | * - Remount is disallowed. | |
271 | * | |
0b0585c3 LT |
272 | * - rename(2) is disallowed. |
273 | * | |
f63674fd TH |
274 | * - "tasks" is removed. Everything should be at process |
275 | * granularity. Use "cgroup.procs" instead. | |
f00baae7 | 276 | * |
f63674fd TH |
277 | * - "release_agent" and "notify_on_release" are removed. |
278 | * Replacement notification mechanism will be implemented. | |
873fe09e | 279 | * |
5c5cc623 LZ |
280 | * - cpuset: tasks will be kept in empty cpusets when hotplug happens |
281 | * and take masks of ancestors with non-empty cpus/mems, instead of | |
282 | * being moved to an ancestor. | |
283 | * | |
88fa523b LZ |
284 | * - cpuset: a task can be moved into an empty cpuset, and again it |
285 | * takes masks of ancestors. | |
6db8e85c | 286 | * |
f63674fd TH |
287 | * - memcg: use_hierarchy is on by default and the cgroup file for |
288 | * the flag is not created. | |
f00baae7 | 289 | * |
9138125b | 290 | * - blkcg: blk-throttle becomes properly hierarchical. |
873fe09e TH |
291 | */ |
292 | CGRP_ROOT_SANE_BEHAVIOR = (1 << 0), | |
293 | ||
25a7e684 TH |
294 | CGRP_ROOT_NOPREFIX = (1 << 1), /* mounted subsystems have no named prefix */ |
295 | CGRP_ROOT_XATTR = (1 << 2), /* supports extended attributes */ | |
0ce6cba3 TH |
296 | |
297 | /* mount options live below bit 16 */ | |
298 | CGRP_ROOT_OPTION_MASK = (1 << 16) - 1, | |
299 | ||
300 | CGRP_ROOT_SUBSYS_BOUND = (1 << 16), /* subsystems finished binding */ | |
25a7e684 TH |
301 | }; |
302 | ||
303 | /* | |
304 | * A cgroupfs_root represents the root of a cgroup hierarchy, and may be | |
305 | * associated with a superblock to form an active hierarchy. This is | |
306 | * internal to cgroup core. Don't access directly from controllers. | |
307 | */ | |
308 | struct cgroupfs_root { | |
309 | struct super_block *sb; | |
310 | ||
a8a648c4 | 311 | /* The bitmask of subsystems attached to this hierarchy */ |
25a7e684 TH |
312 | unsigned long subsys_mask; |
313 | ||
314 | /* Unique id for this hierarchy. */ | |
315 | int hierarchy_id; | |
316 | ||
25a7e684 TH |
317 | /* A list running through the attached subsystems */ |
318 | struct list_head subsys_list; | |
319 | ||
320 | /* The root cgroup for this hierarchy */ | |
321 | struct cgroup top_cgroup; | |
322 | ||
323 | /* Tracks how many cgroups are currently defined in hierarchy.*/ | |
324 | int number_of_cgroups; | |
325 | ||
326 | /* A list running through the active hierarchies */ | |
327 | struct list_head root_list; | |
328 | ||
25a7e684 TH |
329 | /* Hierarchy-specific flags */ |
330 | unsigned long flags; | |
331 | ||
332 | /* IDs for cgroups in this hierarchy */ | |
4e96ee8e | 333 | struct idr cgroup_idr; |
25a7e684 TH |
334 | |
335 | /* The path to use for release notifications. */ | |
336 | char release_agent_path[PATH_MAX]; | |
337 | ||
338 | /* The name for this hierarchy - may be empty */ | |
339 | char name[MAX_CGROUP_ROOT_NAMELEN]; | |
340 | }; | |
341 | ||
d20a390a PM |
342 | /* |
343 | * A css_set is a structure holding pointers to a set of | |
817929ec PM |
344 | * cgroup_subsys_state objects. This saves space in the task struct |
345 | * object and speeds up fork()/exit(), since a single inc/dec and a | |
d20a390a PM |
346 | * list_add()/del() can bump the reference count on the entire cgroup |
347 | * set for a task. | |
817929ec PM |
348 | */ |
349 | ||
350 | struct css_set { | |
351 | ||
352 | /* Reference count */ | |
146aa1bd | 353 | atomic_t refcount; |
817929ec | 354 | |
472b1053 LZ |
355 | /* |
356 | * List running through all cgroup groups in the same hash | |
357 | * slot. Protected by css_set_lock | |
358 | */ | |
359 | struct hlist_node hlist; | |
360 | ||
817929ec PM |
361 | /* |
362 | * List running through all tasks using this cgroup | |
363 | * group. Protected by css_set_lock | |
364 | */ | |
365 | struct list_head tasks; | |
366 | ||
367 | /* | |
69d0206c TH |
368 | * List of cgrp_cset_links pointing at cgroups referenced from this |
369 | * css_set. Protected by css_set_lock. | |
817929ec | 370 | */ |
69d0206c | 371 | struct list_head cgrp_links; |
817929ec PM |
372 | |
373 | /* | |
374 | * Set of subsystem states, one for each subsystem. This array | |
375 | * is immutable after creation apart from the init_css_set | |
cf5d5941 BB |
376 | * during subsystem registration (at boot time) and modular subsystem |
377 | * loading/unloading. | |
817929ec PM |
378 | */ |
379 | struct cgroup_subsys_state *subsys[CGROUP_SUBSYS_COUNT]; | |
c378369d BB |
380 | |
381 | /* For RCU-protected deletion */ | |
382 | struct rcu_head rcu_head; | |
ddbcc7e8 PM |
383 | }; |
384 | ||
91796569 PM |
385 | /* |
386 | * cgroup_map_cb is an abstract callback API for reporting map-valued | |
387 | * control files | |
388 | */ | |
389 | ||
390 | struct cgroup_map_cb { | |
391 | int (*fill)(struct cgroup_map_cb *cb, const char *key, u64 value); | |
392 | void *state; | |
393 | }; | |
394 | ||
d20a390a PM |
395 | /* |
396 | * struct cftype: handler definitions for cgroup control files | |
ddbcc7e8 PM |
397 | * |
398 | * When reading/writing to a file: | |
a043e3b2 | 399 | * - the cgroup to use is file->f_dentry->d_parent->d_fsdata |
ddbcc7e8 PM |
400 | * - the 'cftype' of the file is file->f_dentry->d_fsdata |
401 | */ | |
402 | ||
8e3f6541 | 403 | /* cftype->flags */ |
02c402d9 | 404 | enum { |
6f4b7e63 LZ |
405 | CFTYPE_ONLY_ON_ROOT = (1 << 0), /* only create on root cgrp */ |
406 | CFTYPE_NOT_ON_ROOT = (1 << 1), /* don't create on root cgrp */ | |
02c402d9 TH |
407 | CFTYPE_INSANE = (1 << 2), /* don't create if sane_behavior */ |
408 | }; | |
8e3f6541 TH |
409 | |
410 | #define MAX_CFTYPE_NAME 64 | |
411 | ||
ddbcc7e8 | 412 | struct cftype { |
d20a390a PM |
413 | /* |
414 | * By convention, the name should begin with the name of the | |
8e3f6541 TH |
415 | * subsystem, followed by a period. Zero length string indicates |
416 | * end of cftype array. | |
d20a390a | 417 | */ |
ddbcc7e8 PM |
418 | char name[MAX_CFTYPE_NAME]; |
419 | int private; | |
099fca32 LZ |
420 | /* |
421 | * If not 0, file mode is set to this value, otherwise it will | |
422 | * be figured out automatically | |
423 | */ | |
a5e7ed32 | 424 | umode_t mode; |
db3b1497 PM |
425 | |
426 | /* | |
427 | * If non-zero, defines the maximum length of string that can | |
428 | * be passed to write_string; defaults to 64 | |
429 | */ | |
430 | size_t max_write_len; | |
431 | ||
8e3f6541 TH |
432 | /* CFTYPE_* flags */ |
433 | unsigned int flags; | |
434 | ||
2bb566cb TH |
435 | /* |
436 | * The subsys this file belongs to. Initialized automatically | |
437 | * during registration. NULL for cgroup core files. | |
438 | */ | |
439 | struct cgroup_subsys *ss; | |
440 | ||
ce16b49d | 441 | int (*open)(struct inode *inode, struct file *file); |
182446d0 | 442 | ssize_t (*read)(struct cgroup_subsys_state *css, struct cftype *cft, |
ce16b49d PM |
443 | struct file *file, |
444 | char __user *buf, size_t nbytes, loff_t *ppos); | |
ddbcc7e8 | 445 | /* |
f4c753b7 | 446 | * read_u64() is a shortcut for the common case of returning a |
ddbcc7e8 PM |
447 | * single integer. Use it in place of read() |
448 | */ | |
182446d0 | 449 | u64 (*read_u64)(struct cgroup_subsys_state *css, struct cftype *cft); |
e73d2c61 PM |
450 | /* |
451 | * read_s64() is a signed version of read_u64() | |
452 | */ | |
182446d0 | 453 | s64 (*read_s64)(struct cgroup_subsys_state *css, struct cftype *cft); |
91796569 PM |
454 | /* |
455 | * read_map() is used for defining a map of key/value | |
456 | * pairs. It should call cb->fill(cb, key, value) for each | |
457 | * entry. The key/value pairs (and their ordering) should not | |
458 | * change between reboots. | |
459 | */ | |
182446d0 | 460 | int (*read_map)(struct cgroup_subsys_state *css, struct cftype *cft, |
ce16b49d | 461 | struct cgroup_map_cb *cb); |
29486df3 SH |
462 | /* |
463 | * read_seq_string() is used for outputting a simple sequence | |
464 | * using seqfile. | |
465 | */ | |
182446d0 TH |
466 | int (*read_seq_string)(struct cgroup_subsys_state *css, |
467 | struct cftype *cft, struct seq_file *m); | |
91796569 | 468 | |
182446d0 | 469 | ssize_t (*write)(struct cgroup_subsys_state *css, struct cftype *cft, |
ce16b49d PM |
470 | struct file *file, |
471 | const char __user *buf, size_t nbytes, loff_t *ppos); | |
355e0c48 PM |
472 | |
473 | /* | |
f4c753b7 | 474 | * write_u64() is a shortcut for the common case of accepting |
355e0c48 PM |
475 | * a single integer (as parsed by simple_strtoull) from |
476 | * userspace. Use in place of write(); return 0 or error. | |
477 | */ | |
182446d0 TH |
478 | int (*write_u64)(struct cgroup_subsys_state *css, struct cftype *cft, |
479 | u64 val); | |
e73d2c61 PM |
480 | /* |
481 | * write_s64() is a signed version of write_u64() | |
482 | */ | |
182446d0 TH |
483 | int (*write_s64)(struct cgroup_subsys_state *css, struct cftype *cft, |
484 | s64 val); | |
355e0c48 | 485 | |
db3b1497 PM |
486 | /* |
487 | * write_string() is passed a nul-terminated kernelspace | |
488 | * buffer of maximum length determined by max_write_len. | |
489 | * Returns 0 or -ve error code. | |
490 | */ | |
182446d0 | 491 | int (*write_string)(struct cgroup_subsys_state *css, struct cftype *cft, |
db3b1497 | 492 | const char *buffer); |
d447ea2f PE |
493 | /* |
494 | * trigger() callback can be used to get some kick from the | |
495 | * userspace, when the actual string written is not important | |
496 | * at all. The private field can be used to determine the | |
497 | * kick type for multiplexing. | |
498 | */ | |
182446d0 | 499 | int (*trigger)(struct cgroup_subsys_state *css, unsigned int event); |
d447ea2f | 500 | |
ce16b49d | 501 | int (*release)(struct inode *inode, struct file *file); |
0dea1168 KS |
502 | |
503 | /* | |
504 | * register_event() callback will be used to add new userspace | |
505 | * waiter for changes related to the cftype. Implement it if | |
506 | * you want to provide this functionality. Use eventfd_signal() | |
507 | * on eventfd to send notification to userspace. | |
508 | */ | |
509 | int (*register_event)(struct cgroup *cgrp, struct cftype *cft, | |
510 | struct eventfd_ctx *eventfd, const char *args); | |
511 | /* | |
512 | * unregister_event() callback will be called when userspace | |
513 | * closes the eventfd or on cgroup removing. | |
514 | * This callback must be implemented, if you want provide | |
515 | * notification functionality. | |
0dea1168 | 516 | */ |
907860ed | 517 | void (*unregister_event)(struct cgroup *cgrp, struct cftype *cft, |
0dea1168 | 518 | struct eventfd_ctx *eventfd); |
ddbcc7e8 PM |
519 | }; |
520 | ||
8e3f6541 TH |
521 | /* |
522 | * cftype_sets describe cftypes belonging to a subsystem and are chained at | |
523 | * cgroup_subsys->cftsets. Each cftset points to an array of cftypes | |
524 | * terminated by zero length name. | |
525 | */ | |
526 | struct cftype_set { | |
527 | struct list_head node; /* chained at subsys->cftsets */ | |
03b1cde6 | 528 | struct cftype *cfts; |
8e3f6541 TH |
529 | }; |
530 | ||
31a7df01 | 531 | struct cgroup_scanner { |
6f4b7e63 | 532 | struct cgroup *cgrp; |
31a7df01 CW |
533 | int (*test_task)(struct task_struct *p, struct cgroup_scanner *scan); |
534 | void (*process_task)(struct task_struct *p, | |
535 | struct cgroup_scanner *scan); | |
536 | struct ptr_heap *heap; | |
bd1a8ab7 | 537 | void *data; |
31a7df01 CW |
538 | }; |
539 | ||
873fe09e TH |
540 | /* |
541 | * See the comment above CGRP_ROOT_SANE_BEHAVIOR for details. This | |
542 | * function can be called as long as @cgrp is accessible. | |
543 | */ | |
544 | static inline bool cgroup_sane_behavior(const struct cgroup *cgrp) | |
545 | { | |
546 | return cgrp->root->flags & CGRP_ROOT_SANE_BEHAVIOR; | |
547 | } | |
548 | ||
65dff759 LZ |
549 | /* Caller should hold rcu_read_lock() */ |
550 | static inline const char *cgroup_name(const struct cgroup *cgrp) | |
551 | { | |
552 | return rcu_dereference(cgrp->name)->name; | |
553 | } | |
554 | ||
03b1cde6 | 555 | int cgroup_add_cftypes(struct cgroup_subsys *ss, struct cftype *cfts); |
2bb566cb | 556 | int cgroup_rm_cftypes(struct cftype *cfts); |
8e3f6541 | 557 | |
78574cf9 | 558 | bool cgroup_is_descendant(struct cgroup *cgrp, struct cgroup *ancestor); |
ddbcc7e8 | 559 | |
ffd2d883 | 560 | int cgroup_path(const struct cgroup *cgrp, char *buf, int buflen); |
913ffdb5 | 561 | int task_cgroup_path(struct task_struct *task, char *buf, size_t buflen); |
ddbcc7e8 | 562 | |
ffd2d883 | 563 | int cgroup_task_count(const struct cgroup *cgrp); |
bbcb81d0 | 564 | |
2f7ee569 TH |
565 | /* |
566 | * Control Group taskset, used to pass around set of tasks to cgroup_subsys | |
567 | * methods. | |
568 | */ | |
569 | struct cgroup_taskset; | |
570 | struct task_struct *cgroup_taskset_first(struct cgroup_taskset *tset); | |
571 | struct task_struct *cgroup_taskset_next(struct cgroup_taskset *tset); | |
572 | struct cgroup *cgroup_taskset_cur_cgroup(struct cgroup_taskset *tset); | |
573 | int cgroup_taskset_size(struct cgroup_taskset *tset); | |
574 | ||
575 | /** | |
576 | * cgroup_taskset_for_each - iterate cgroup_taskset | |
577 | * @task: the loop cursor | |
578 | * @skip_cgrp: skip if task's cgroup matches this, %NULL to iterate through all | |
579 | * @tset: taskset to iterate | |
580 | */ | |
581 | #define cgroup_taskset_for_each(task, skip_cgrp, tset) \ | |
582 | for ((task) = cgroup_taskset_first((tset)); (task); \ | |
583 | (task) = cgroup_taskset_next((tset))) \ | |
584 | if (!(skip_cgrp) || \ | |
585 | cgroup_taskset_cur_cgroup((tset)) != (skip_cgrp)) | |
586 | ||
21acb9ca TLSC |
587 | /* |
588 | * Control Group subsystem type. | |
589 | * See Documentation/cgroups/cgroups.txt for details | |
590 | */ | |
ddbcc7e8 PM |
591 | |
592 | struct cgroup_subsys { | |
eb95419b TH |
593 | struct cgroup_subsys_state *(*css_alloc)(struct cgroup_subsys_state *parent_css); |
594 | int (*css_online)(struct cgroup_subsys_state *css); | |
595 | void (*css_offline)(struct cgroup_subsys_state *css); | |
596 | void (*css_free)(struct cgroup_subsys_state *css); | |
597 | ||
598 | int (*can_attach)(struct cgroup_subsys_state *css, | |
599 | struct cgroup_taskset *tset); | |
600 | void (*cancel_attach)(struct cgroup_subsys_state *css, | |
601 | struct cgroup_taskset *tset); | |
602 | void (*attach)(struct cgroup_subsys_state *css, | |
603 | struct cgroup_taskset *tset); | |
761b3ef5 | 604 | void (*fork)(struct task_struct *task); |
eb95419b TH |
605 | void (*exit)(struct cgroup_subsys_state *css, |
606 | struct cgroup_subsys_state *old_css, | |
761b3ef5 | 607 | struct task_struct *task); |
eb95419b | 608 | void (*bind)(struct cgroup_subsys_state *root_css); |
e5991371 | 609 | |
ddbcc7e8 | 610 | int subsys_id; |
8bab8dde | 611 | int disabled; |
ddbcc7e8 | 612 | int early_init; |
38460b48 KH |
613 | /* |
614 | * True if this subsys uses ID. ID is not available before cgroup_init() | |
615 | * (not available in early_init time.) | |
616 | */ | |
617 | bool use_id; | |
48ddbe19 | 618 | |
8c7f6edb TH |
619 | /* |
620 | * If %false, this subsystem is properly hierarchical - | |
621 | * configuration, resource accounting and restriction on a parent | |
622 | * cgroup cover those of its children. If %true, hierarchy support | |
623 | * is broken in some ways - some subsystems ignore hierarchy | |
624 | * completely while others are only implemented half-way. | |
625 | * | |
626 | * It's now disallowed to create nested cgroups if the subsystem is | |
627 | * broken and cgroup core will emit a warning message on such | |
628 | * cases. Eventually, all subsystems will be made properly | |
629 | * hierarchical and this will go away. | |
630 | */ | |
631 | bool broken_hierarchy; | |
632 | bool warned_broken_hierarchy; | |
633 | ||
ddbcc7e8 PM |
634 | #define MAX_CGROUP_TYPE_NAMELEN 32 |
635 | const char *name; | |
636 | ||
999cd8a4 PM |
637 | /* |
638 | * Link to parent, and list entry in parent's children. | |
6be96a5c | 639 | * Protected by cgroup_lock() |
999cd8a4 PM |
640 | */ |
641 | struct cgroupfs_root *root; | |
ddbcc7e8 | 642 | struct list_head sibling; |
38460b48 KH |
643 | /* used when use_id == true */ |
644 | struct idr idr; | |
42aee6c4 | 645 | spinlock_t id_lock; |
e6a1105b | 646 | |
8e3f6541 TH |
647 | /* list of cftype_sets */ |
648 | struct list_head cftsets; | |
649 | ||
650 | /* base cftypes, automatically [de]registered with subsys itself */ | |
651 | struct cftype *base_cftypes; | |
652 | struct cftype_set base_cftset; | |
653 | ||
e6a1105b BB |
654 | /* should be defined only by modular subsystems */ |
655 | struct module *module; | |
ddbcc7e8 PM |
656 | }; |
657 | ||
658 | #define SUBSYS(_x) extern struct cgroup_subsys _x ## _subsys; | |
5fc0b025 | 659 | #define IS_SUBSYS_ENABLED(option) IS_BUILTIN(option) |
ddbcc7e8 | 660 | #include <linux/cgroup_subsys.h> |
5fc0b025 | 661 | #undef IS_SUBSYS_ENABLED |
ddbcc7e8 PM |
662 | #undef SUBSYS |
663 | ||
63876986 TH |
664 | /** |
665 | * css_parent - find the parent css | |
666 | * @css: the target cgroup_subsys_state | |
667 | * | |
668 | * Return the parent css of @css. This function is guaranteed to return | |
669 | * non-NULL parent as long as @css isn't the root. | |
670 | */ | |
671 | static inline | |
672 | struct cgroup_subsys_state *css_parent(struct cgroup_subsys_state *css) | |
673 | { | |
674 | struct cgroup *parent_cgrp = css->cgroup->parent; | |
675 | ||
67f4c36f TH |
676 | if (!parent_cgrp) |
677 | return NULL; | |
678 | ||
679 | if (css->ss) | |
680 | return parent_cgrp->subsys[css->ss->subsys_id]; | |
681 | else | |
682 | return &parent_cgrp->dummy_css; | |
63876986 TH |
683 | } |
684 | ||
8af01f56 TH |
685 | /** |
686 | * cgroup_css - obtain a cgroup's css for the specified subsystem | |
687 | * @cgrp: the cgroup of interest | |
688 | * @subsys_id: the subsystem of interest | |
689 | * | |
690 | * Return @cgrp's css (cgroup_subsys_state) associated with @subsys_id. | |
691 | */ | |
692 | static inline struct cgroup_subsys_state *cgroup_css(struct cgroup *cgrp, | |
693 | int subsys_id) | |
ddbcc7e8 | 694 | { |
ffd2d883 | 695 | return cgrp->subsys[subsys_id]; |
ddbcc7e8 PM |
696 | } |
697 | ||
14611e51 TH |
698 | /** |
699 | * task_css_set_check - obtain a task's css_set with extra access conditions | |
700 | * @task: the task to obtain css_set for | |
701 | * @__c: extra condition expression to be passed to rcu_dereference_check() | |
702 | * | |
703 | * A task's css_set is RCU protected, initialized and exited while holding | |
704 | * task_lock(), and can only be modified while holding both cgroup_mutex | |
705 | * and task_lock() while the task is alive. This macro verifies that the | |
706 | * caller is inside proper critical section and returns @task's css_set. | |
707 | * | |
708 | * The caller can also specify additional allowed conditions via @__c, such | |
709 | * as locks used during the cgroup_subsys::attach() methods. | |
dc61b1d6 | 710 | */ |
2219449a TH |
711 | #ifdef CONFIG_PROVE_RCU |
712 | extern struct mutex cgroup_mutex; | |
14611e51 TH |
713 | #define task_css_set_check(task, __c) \ |
714 | rcu_dereference_check((task)->cgroups, \ | |
715 | lockdep_is_held(&(task)->alloc_lock) || \ | |
716 | lockdep_is_held(&cgroup_mutex) || (__c)) | |
2219449a | 717 | #else |
14611e51 TH |
718 | #define task_css_set_check(task, __c) \ |
719 | rcu_dereference((task)->cgroups) | |
2219449a | 720 | #endif |
dc61b1d6 | 721 | |
14611e51 | 722 | /** |
8af01f56 | 723 | * task_css_check - obtain css for (task, subsys) w/ extra access conds |
14611e51 TH |
724 | * @task: the target task |
725 | * @subsys_id: the target subsystem ID | |
726 | * @__c: extra condition expression to be passed to rcu_dereference_check() | |
727 | * | |
728 | * Return the cgroup_subsys_state for the (@task, @subsys_id) pair. The | |
729 | * synchronization rules are the same as task_css_set_check(). | |
730 | */ | |
8af01f56 | 731 | #define task_css_check(task, subsys_id, __c) \ |
14611e51 TH |
732 | task_css_set_check((task), (__c))->subsys[(subsys_id)] |
733 | ||
734 | /** | |
735 | * task_css_set - obtain a task's css_set | |
736 | * @task: the task to obtain css_set for | |
737 | * | |
738 | * See task_css_set_check(). | |
739 | */ | |
740 | static inline struct css_set *task_css_set(struct task_struct *task) | |
741 | { | |
742 | return task_css_set_check(task, false); | |
743 | } | |
744 | ||
745 | /** | |
8af01f56 | 746 | * task_css - obtain css for (task, subsys) |
14611e51 TH |
747 | * @task: the target task |
748 | * @subsys_id: the target subsystem ID | |
749 | * | |
8af01f56 | 750 | * See task_css_check(). |
14611e51 | 751 | */ |
8af01f56 TH |
752 | static inline struct cgroup_subsys_state *task_css(struct task_struct *task, |
753 | int subsys_id) | |
ddbcc7e8 | 754 | { |
8af01f56 | 755 | return task_css_check(task, subsys_id, false); |
ddbcc7e8 PM |
756 | } |
757 | ||
8af01f56 TH |
758 | static inline struct cgroup *task_cgroup(struct task_struct *task, |
759 | int subsys_id) | |
ddbcc7e8 | 760 | { |
8af01f56 | 761 | return task_css(task, subsys_id)->cgroup; |
ddbcc7e8 PM |
762 | } |
763 | ||
e14880f7 LZ |
764 | /** |
765 | * cgroup_from_id - lookup cgroup by id | |
766 | * @ss: cgroup subsys to be looked into | |
767 | * @id: the cgroup id | |
768 | * | |
769 | * Returns the cgroup if there's valid one with @id, otherwise returns NULL. | |
770 | * Should be called under rcu_read_lock(). | |
771 | */ | |
772 | static inline struct cgroup *cgroup_from_id(struct cgroup_subsys *ss, int id) | |
773 | { | |
774 | #ifdef CONFIG_PROVE_RCU | |
775 | rcu_lockdep_assert(rcu_read_lock_held() || | |
776 | lockdep_is_held(&cgroup_mutex), | |
777 | "cgroup_from_id() needs proper protection"); | |
778 | #endif | |
779 | return idr_find(&ss->root->cgroup_idr, id); | |
780 | } | |
781 | ||
492eb21b TH |
782 | struct cgroup_subsys_state *css_next_child(struct cgroup_subsys_state *pos, |
783 | struct cgroup_subsys_state *parent); | |
53fa5261 | 784 | |
574bd9f7 | 785 | /** |
492eb21b TH |
786 | * css_for_each_child - iterate through children of a css |
787 | * @pos: the css * to use as the loop cursor | |
788 | * @parent: css whose children to walk | |
574bd9f7 | 789 | * |
492eb21b TH |
790 | * Walk @parent's children. Must be called under rcu_read_lock(). A child |
791 | * css which hasn't finished ->css_online() or already has finished | |
92fb9748 | 792 | * ->css_offline() may show up during traversal and it's each subsystem's |
574bd9f7 TH |
793 | * responsibility to verify that each @pos is alive. |
794 | * | |
92fb9748 | 795 | * If a subsystem synchronizes against the parent in its ->css_online() and |
492eb21b | 796 | * before starting iterating, a css which finished ->css_online() is |
92fb9748 | 797 | * guaranteed to be visible in the future iterations. |
75501a6d TH |
798 | * |
799 | * It is allowed to temporarily drop RCU read lock during iteration. The | |
800 | * caller is responsible for ensuring that @pos remains accessible until | |
801 | * the start of the next iteration by, for example, bumping the css refcnt. | |
574bd9f7 | 802 | */ |
492eb21b TH |
803 | #define css_for_each_child(pos, parent) \ |
804 | for ((pos) = css_next_child(NULL, (parent)); (pos); \ | |
805 | (pos) = css_next_child((pos), (parent))) | |
574bd9f7 | 806 | |
492eb21b TH |
807 | struct cgroup_subsys_state * |
808 | css_next_descendant_pre(struct cgroup_subsys_state *pos, | |
809 | struct cgroup_subsys_state *css); | |
810 | ||
811 | struct cgroup_subsys_state * | |
812 | css_rightmost_descendant(struct cgroup_subsys_state *pos); | |
574bd9f7 TH |
813 | |
814 | /** | |
492eb21b TH |
815 | * css_for_each_descendant_pre - pre-order walk of a css's descendants |
816 | * @pos: the css * to use as the loop cursor | |
817 | * @root: css whose descendants to walk | |
574bd9f7 | 818 | * |
492eb21b TH |
819 | * Walk @root's descendants. Must be called under rcu_read_lock(). A |
820 | * descendant css which hasn't finished ->css_online() or already has | |
92fb9748 | 821 | * finished ->css_offline() may show up during traversal and it's each |
574bd9f7 TH |
822 | * subsystem's responsibility to verify that each @pos is alive. |
823 | * | |
92fb9748 TH |
824 | * If a subsystem synchronizes against the parent in its ->css_online() and |
825 | * before starting iterating, and synchronizes against @pos on each | |
492eb21b | 826 | * iteration, any descendant css which finished ->css_online() is |
574bd9f7 TH |
827 | * guaranteed to be visible in the future iterations. |
828 | * | |
829 | * In other words, the following guarantees that a descendant can't escape | |
830 | * state updates of its ancestors. | |
831 | * | |
492eb21b | 832 | * my_online(@css) |
574bd9f7 | 833 | * { |
492eb21b TH |
834 | * Lock @css's parent and @css; |
835 | * Inherit state from the parent; | |
574bd9f7 TH |
836 | * Unlock both. |
837 | * } | |
838 | * | |
492eb21b | 839 | * my_update_state(@css) |
574bd9f7 | 840 | * { |
492eb21b TH |
841 | * Lock @css; |
842 | * Update @css's state; | |
843 | * Unlock @css; | |
574bd9f7 | 844 | * |
492eb21b | 845 | * css_for_each_descendant_pre(@pos, @css) { |
574bd9f7 | 846 | * Lock @pos; |
492eb21b | 847 | * Verify @pos is alive and inherit state from @pos's parent; |
574bd9f7 TH |
848 | * Unlock @pos; |
849 | * } | |
850 | * } | |
851 | * | |
852 | * As long as the inheriting step, including checking the parent state, is | |
853 | * enclosed inside @pos locking, double-locking the parent isn't necessary | |
854 | * while inheriting. The state update to the parent is guaranteed to be | |
855 | * visible by walking order and, as long as inheriting operations to the | |
856 | * same @pos are atomic to each other, multiple updates racing each other | |
857 | * still result in the correct state. It's guaranateed that at least one | |
492eb21b | 858 | * inheritance happens for any css after the latest update to its parent. |
574bd9f7 TH |
859 | * |
860 | * If checking parent's state requires locking the parent, each inheriting | |
861 | * iteration should lock and unlock both @pos->parent and @pos. | |
862 | * | |
863 | * Alternatively, a subsystem may choose to use a single global lock to | |
92fb9748 | 864 | * synchronize ->css_online() and ->css_offline() against tree-walking |
574bd9f7 | 865 | * operations. |
75501a6d TH |
866 | * |
867 | * It is allowed to temporarily drop RCU read lock during iteration. The | |
868 | * caller is responsible for ensuring that @pos remains accessible until | |
869 | * the start of the next iteration by, for example, bumping the css refcnt. | |
574bd9f7 | 870 | */ |
492eb21b TH |
871 | #define css_for_each_descendant_pre(pos, css) \ |
872 | for ((pos) = css_next_descendant_pre(NULL, (css)); (pos); \ | |
873 | (pos) = css_next_descendant_pre((pos), (css))) | |
574bd9f7 | 874 | |
492eb21b TH |
875 | struct cgroup_subsys_state * |
876 | css_next_descendant_post(struct cgroup_subsys_state *pos, | |
877 | struct cgroup_subsys_state *css); | |
574bd9f7 TH |
878 | |
879 | /** | |
492eb21b TH |
880 | * css_for_each_descendant_post - post-order walk of a css's descendants |
881 | * @pos: the css * to use as the loop cursor | |
882 | * @css: css whose descendants to walk | |
574bd9f7 | 883 | * |
492eb21b | 884 | * Similar to css_for_each_descendant_pre() but performs post-order |
574bd9f7 TH |
885 | * traversal instead. Note that the walk visibility guarantee described in |
886 | * pre-order walk doesn't apply the same to post-order walks. | |
887 | */ | |
492eb21b TH |
888 | #define css_for_each_descendant_post(pos, css) \ |
889 | for ((pos) = css_next_descendant_post(NULL, (css)); (pos); \ | |
890 | (pos) = css_next_descendant_post((pos), (css))) | |
574bd9f7 | 891 | |
817929ec PM |
892 | /* A cgroup_iter should be treated as an opaque object */ |
893 | struct cgroup_iter { | |
69d0206c | 894 | struct list_head *cset_link; |
817929ec PM |
895 | struct list_head *task; |
896 | }; | |
897 | ||
d20a390a PM |
898 | /* |
899 | * To iterate across the tasks in a cgroup: | |
817929ec | 900 | * |
b595076a | 901 | * 1) call cgroup_iter_start to initialize an iterator |
817929ec PM |
902 | * |
903 | * 2) call cgroup_iter_next() to retrieve member tasks until it | |
904 | * returns NULL or until you want to end the iteration | |
905 | * | |
906 | * 3) call cgroup_iter_end() to destroy the iterator. | |
31a7df01 | 907 | * |
d20a390a PM |
908 | * Or, call cgroup_scan_tasks() to iterate through every task in a |
909 | * cgroup - cgroup_scan_tasks() holds the css_set_lock when calling | |
910 | * the test_task() callback, but not while calling the process_task() | |
911 | * callback. | |
817929ec | 912 | */ |
ffd2d883 LZ |
913 | void cgroup_iter_start(struct cgroup *cgrp, struct cgroup_iter *it); |
914 | struct task_struct *cgroup_iter_next(struct cgroup *cgrp, | |
817929ec | 915 | struct cgroup_iter *it); |
ffd2d883 | 916 | void cgroup_iter_end(struct cgroup *cgrp, struct cgroup_iter *it); |
31a7df01 | 917 | int cgroup_scan_tasks(struct cgroup_scanner *scan); |
31583bb0 | 918 | int cgroup_attach_task_all(struct task_struct *from, struct task_struct *); |
8cc99345 | 919 | int cgroup_transfer_tasks(struct cgroup *to, struct cgroup *from); |
31583bb0 | 920 | |
38460b48 KH |
921 | /* |
922 | * CSS ID is ID for cgroup_subsys_state structs under subsys. This only works | |
923 | * if cgroup_subsys.use_id == true. It can be used for looking up and scanning. | |
924 | * CSS ID is assigned at cgroup allocation (create) automatically | |
925 | * and removed when subsys calls free_css_id() function. This is because | |
926 | * the lifetime of cgroup_subsys_state is subsys's matter. | |
927 | * | |
928 | * Looking up and scanning function should be called under rcu_read_lock(). | |
6be96a5c | 929 | * Taking cgroup_mutex is not necessary for following calls. |
38460b48 KH |
930 | * But the css returned by this routine can be "not populated yet" or "being |
931 | * destroyed". The caller should check css and cgroup's status. | |
932 | */ | |
933 | ||
934 | /* | |
935 | * Typically Called at ->destroy(), or somewhere the subsys frees | |
936 | * cgroup_subsys_state. | |
937 | */ | |
938 | void free_css_id(struct cgroup_subsys *ss, struct cgroup_subsys_state *css); | |
939 | ||
940 | /* Find a cgroup_subsys_state which has given ID */ | |
941 | ||
942 | struct cgroup_subsys_state *css_lookup(struct cgroup_subsys *ss, int id); | |
943 | ||
38460b48 KH |
944 | /* Returns true if root is ancestor of cg */ |
945 | bool css_is_ancestor(struct cgroup_subsys_state *cg, | |
0b7f569e | 946 | const struct cgroup_subsys_state *root); |
38460b48 KH |
947 | |
948 | /* Get id and depth of css */ | |
949 | unsigned short css_id(struct cgroup_subsys_state *css); | |
e5d1367f | 950 | struct cgroup_subsys_state *cgroup_css_from_dir(struct file *f, int id); |
38460b48 | 951 | |
ddbcc7e8 PM |
952 | #else /* !CONFIG_CGROUPS */ |
953 | ||
954 | static inline int cgroup_init_early(void) { return 0; } | |
955 | static inline int cgroup_init(void) { return 0; } | |
b4f48b63 | 956 | static inline void cgroup_fork(struct task_struct *p) {} |
817929ec | 957 | static inline void cgroup_post_fork(struct task_struct *p) {} |
b4f48b63 | 958 | static inline void cgroup_exit(struct task_struct *p, int callbacks) {} |
ddbcc7e8 | 959 | |
846c7bb0 BS |
960 | static inline int cgroupstats_build(struct cgroupstats *stats, |
961 | struct dentry *dentry) | |
962 | { | |
963 | return -EINVAL; | |
964 | } | |
ddbcc7e8 | 965 | |
d7926ee3 | 966 | /* No cgroups - nothing to do */ |
31583bb0 MT |
967 | static inline int cgroup_attach_task_all(struct task_struct *from, |
968 | struct task_struct *t) | |
969 | { | |
970 | return 0; | |
971 | } | |
d7926ee3 | 972 | |
ddbcc7e8 PM |
973 | #endif /* !CONFIG_CGROUPS */ |
974 | ||
975 | #endif /* _LINUX_CGROUP_H */ |