]> git.proxmox.com Git - mirror_ubuntu-bionic-kernel.git/blob - fs/fscache/object.c
UBUNTU: [Config] arm64: snapdragon: DRM_MSM=m
[mirror_ubuntu-bionic-kernel.git] / fs / fscache / object.c
1 /* FS-Cache object state machine handler
2 *
3 * Copyright (C) 2007 Red Hat, Inc. All Rights Reserved.
4 * Written by David Howells (dhowells@redhat.com)
5 *
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
10 *
11 * See Documentation/filesystems/caching/object.txt for a description of the
12 * object state machine and the in-kernel representations.
13 */
14
15 #define FSCACHE_DEBUG_LEVEL COOKIE
16 #include <linux/module.h>
17 #include <linux/slab.h>
18 #include <linux/prefetch.h>
19 #include "internal.h"
20
21 static const struct fscache_state *fscache_abort_initialisation(struct fscache_object *, int);
22 static const struct fscache_state *fscache_kill_dependents(struct fscache_object *, int);
23 static const struct fscache_state *fscache_drop_object(struct fscache_object *, int);
24 static const struct fscache_state *fscache_initialise_object(struct fscache_object *, int);
25 static const struct fscache_state *fscache_invalidate_object(struct fscache_object *, int);
26 static const struct fscache_state *fscache_jumpstart_dependents(struct fscache_object *, int);
27 static const struct fscache_state *fscache_kill_object(struct fscache_object *, int);
28 static const struct fscache_state *fscache_lookup_failure(struct fscache_object *, int);
29 static const struct fscache_state *fscache_look_up_object(struct fscache_object *, int);
30 static const struct fscache_state *fscache_object_available(struct fscache_object *, int);
31 static const struct fscache_state *fscache_parent_ready(struct fscache_object *, int);
32 static const struct fscache_state *fscache_update_object(struct fscache_object *, int);
33 static const struct fscache_state *fscache_object_dead(struct fscache_object *, int);
34
35 #define __STATE_NAME(n) fscache_osm_##n
36 #define STATE(n) (&__STATE_NAME(n))
37
38 /*
39 * Define a work state. Work states are execution states. No event processing
40 * is performed by them. The function attached to a work state returns a
41 * pointer indicating the next state to which the state machine should
42 * transition. Returning NO_TRANSIT repeats the current state, but goes back
43 * to the scheduler first.
44 */
45 #define WORK_STATE(n, sn, f) \
46 const struct fscache_state __STATE_NAME(n) = { \
47 .name = #n, \
48 .short_name = sn, \
49 .work = f \
50 }
51
52 /*
53 * Returns from work states.
54 */
55 #define transit_to(state) ({ prefetch(&STATE(state)->work); STATE(state); })
56
57 #define NO_TRANSIT ((struct fscache_state *)NULL)
58
59 /*
60 * Define a wait state. Wait states are event processing states. No execution
61 * is performed by them. Wait states are just tables of "if event X occurs,
62 * clear it and transition to state Y". The dispatcher returns to the
63 * scheduler if none of the events in which the wait state has an interest are
64 * currently pending.
65 */
66 #define WAIT_STATE(n, sn, ...) \
67 const struct fscache_state __STATE_NAME(n) = { \
68 .name = #n, \
69 .short_name = sn, \
70 .work = NULL, \
71 .transitions = { __VA_ARGS__, { 0, NULL } } \
72 }
73
74 #define TRANSIT_TO(state, emask) \
75 { .events = (emask), .transit_to = STATE(state) }
76
77 /*
78 * The object state machine.
79 */
80 static WORK_STATE(INIT_OBJECT, "INIT", fscache_initialise_object);
81 static WORK_STATE(PARENT_READY, "PRDY", fscache_parent_ready);
82 static WORK_STATE(ABORT_INIT, "ABRT", fscache_abort_initialisation);
83 static WORK_STATE(LOOK_UP_OBJECT, "LOOK", fscache_look_up_object);
84 static WORK_STATE(CREATE_OBJECT, "CRTO", fscache_look_up_object);
85 static WORK_STATE(OBJECT_AVAILABLE, "AVBL", fscache_object_available);
86 static WORK_STATE(JUMPSTART_DEPS, "JUMP", fscache_jumpstart_dependents);
87
88 static WORK_STATE(INVALIDATE_OBJECT, "INVL", fscache_invalidate_object);
89 static WORK_STATE(UPDATE_OBJECT, "UPDT", fscache_update_object);
90
91 static WORK_STATE(LOOKUP_FAILURE, "LCFL", fscache_lookup_failure);
92 static WORK_STATE(KILL_OBJECT, "KILL", fscache_kill_object);
93 static WORK_STATE(KILL_DEPENDENTS, "KDEP", fscache_kill_dependents);
94 static WORK_STATE(DROP_OBJECT, "DROP", fscache_drop_object);
95 static WORK_STATE(OBJECT_DEAD, "DEAD", fscache_object_dead);
96
97 static WAIT_STATE(WAIT_FOR_INIT, "?INI",
98 TRANSIT_TO(INIT_OBJECT, 1 << FSCACHE_OBJECT_EV_NEW_CHILD));
99
100 static WAIT_STATE(WAIT_FOR_PARENT, "?PRN",
101 TRANSIT_TO(PARENT_READY, 1 << FSCACHE_OBJECT_EV_PARENT_READY));
102
103 static WAIT_STATE(WAIT_FOR_CMD, "?CMD",
104 TRANSIT_TO(INVALIDATE_OBJECT, 1 << FSCACHE_OBJECT_EV_INVALIDATE),
105 TRANSIT_TO(UPDATE_OBJECT, 1 << FSCACHE_OBJECT_EV_UPDATE),
106 TRANSIT_TO(JUMPSTART_DEPS, 1 << FSCACHE_OBJECT_EV_NEW_CHILD));
107
108 static WAIT_STATE(WAIT_FOR_CLEARANCE, "?CLR",
109 TRANSIT_TO(KILL_OBJECT, 1 << FSCACHE_OBJECT_EV_CLEARED));
110
111 /*
112 * Out-of-band event transition tables. These are for handling unexpected
113 * events, such as an I/O error. If an OOB event occurs, the state machine
114 * clears and disables the event and forces a transition to the nominated work
115 * state (acurrently executing work states will complete first).
116 *
117 * In such a situation, object->state remembers the state the machine should
118 * have been in/gone to and returning NO_TRANSIT returns to that.
119 */
120 static const struct fscache_transition fscache_osm_init_oob[] = {
121 TRANSIT_TO(ABORT_INIT,
122 (1 << FSCACHE_OBJECT_EV_ERROR) |
123 (1 << FSCACHE_OBJECT_EV_KILL)),
124 { 0, NULL }
125 };
126
127 static const struct fscache_transition fscache_osm_lookup_oob[] = {
128 TRANSIT_TO(LOOKUP_FAILURE,
129 (1 << FSCACHE_OBJECT_EV_ERROR) |
130 (1 << FSCACHE_OBJECT_EV_KILL)),
131 { 0, NULL }
132 };
133
134 static const struct fscache_transition fscache_osm_run_oob[] = {
135 TRANSIT_TO(KILL_OBJECT,
136 (1 << FSCACHE_OBJECT_EV_ERROR) |
137 (1 << FSCACHE_OBJECT_EV_KILL)),
138 { 0, NULL }
139 };
140
141 static int fscache_get_object(struct fscache_object *);
142 static void fscache_put_object(struct fscache_object *);
143 static bool fscache_enqueue_dependents(struct fscache_object *, int);
144 static void fscache_dequeue_object(struct fscache_object *);
145
146 /*
147 * we need to notify the parent when an op completes that we had outstanding
148 * upon it
149 */
150 static inline void fscache_done_parent_op(struct fscache_object *object)
151 {
152 struct fscache_object *parent = object->parent;
153
154 _enter("OBJ%x {OBJ%x,%x}",
155 object->debug_id, parent->debug_id, parent->n_ops);
156
157 spin_lock_nested(&parent->lock, 1);
158 parent->n_obj_ops--;
159 parent->n_ops--;
160 if (parent->n_ops == 0)
161 fscache_raise_event(parent, FSCACHE_OBJECT_EV_CLEARED);
162 spin_unlock(&parent->lock);
163 }
164
165 /*
166 * Object state machine dispatcher.
167 */
168 static void fscache_object_sm_dispatcher(struct fscache_object *object)
169 {
170 const struct fscache_transition *t;
171 const struct fscache_state *state, *new_state;
172 unsigned long events, event_mask;
173 int event = -1;
174
175 ASSERT(object != NULL);
176
177 _enter("{OBJ%x,%s,%lx}",
178 object->debug_id, object->state->name, object->events);
179
180 event_mask = object->event_mask;
181 restart:
182 object->event_mask = 0; /* Mask normal event handling */
183 state = object->state;
184 restart_masked:
185 events = object->events;
186
187 /* Handle any out-of-band events (typically an error) */
188 if (events & object->oob_event_mask) {
189 _debug("{OBJ%x} oob %lx",
190 object->debug_id, events & object->oob_event_mask);
191 for (t = object->oob_table; t->events; t++) {
192 if (events & t->events) {
193 state = t->transit_to;
194 ASSERT(state->work != NULL);
195 event = fls(events & t->events) - 1;
196 __clear_bit(event, &object->oob_event_mask);
197 clear_bit(event, &object->events);
198 goto execute_work_state;
199 }
200 }
201 }
202
203 /* Wait states are just transition tables */
204 if (!state->work) {
205 if (events & event_mask) {
206 for (t = state->transitions; t->events; t++) {
207 if (events & t->events) {
208 new_state = t->transit_to;
209 event = fls(events & t->events) - 1;
210 clear_bit(event, &object->events);
211 _debug("{OBJ%x} ev %d: %s -> %s",
212 object->debug_id, event,
213 state->name, new_state->name);
214 object->state = state = new_state;
215 goto execute_work_state;
216 }
217 }
218
219 /* The event mask didn't include all the tabled bits */
220 BUG();
221 }
222 /* Randomly woke up */
223 goto unmask_events;
224 }
225
226 execute_work_state:
227 _debug("{OBJ%x} exec %s", object->debug_id, state->name);
228
229 new_state = state->work(object, event);
230 event = -1;
231 if (new_state == NO_TRANSIT) {
232 _debug("{OBJ%x} %s notrans", object->debug_id, state->name);
233 if (unlikely(state == STATE(OBJECT_DEAD))) {
234 _leave(" [dead]");
235 return;
236 }
237 fscache_enqueue_object(object);
238 event_mask = object->oob_event_mask;
239 goto unmask_events;
240 }
241
242 _debug("{OBJ%x} %s -> %s",
243 object->debug_id, state->name, new_state->name);
244 object->state = state = new_state;
245
246 if (state->work) {
247 if (unlikely(state == STATE(OBJECT_DEAD))) {
248 _leave(" [dead]");
249 return;
250 }
251 goto restart_masked;
252 }
253
254 /* Transited to wait state */
255 event_mask = object->oob_event_mask;
256 for (t = state->transitions; t->events; t++)
257 event_mask |= t->events;
258
259 unmask_events:
260 object->event_mask = event_mask;
261 smp_mb();
262 events = object->events;
263 if (events & event_mask)
264 goto restart;
265 _leave(" [msk %lx]", event_mask);
266 }
267
268 /*
269 * execute an object
270 */
271 static void fscache_object_work_func(struct work_struct *work)
272 {
273 struct fscache_object *object =
274 container_of(work, struct fscache_object, work);
275 unsigned long start;
276
277 _enter("{OBJ%x}", object->debug_id);
278
279 start = jiffies;
280 fscache_object_sm_dispatcher(object);
281 fscache_hist(fscache_objs_histogram, start);
282 fscache_put_object(object);
283 }
284
285 /**
286 * fscache_object_init - Initialise a cache object description
287 * @object: Object description
288 * @cookie: Cookie object will be attached to
289 * @cache: Cache in which backing object will be found
290 *
291 * Initialise a cache object description to its basic values.
292 *
293 * See Documentation/filesystems/caching/backend-api.txt for a complete
294 * description.
295 */
296 void fscache_object_init(struct fscache_object *object,
297 struct fscache_cookie *cookie,
298 struct fscache_cache *cache)
299 {
300 const struct fscache_transition *t;
301
302 atomic_inc(&cache->object_count);
303
304 object->state = STATE(WAIT_FOR_INIT);
305 object->oob_table = fscache_osm_init_oob;
306 object->flags = 1 << FSCACHE_OBJECT_IS_LIVE;
307 spin_lock_init(&object->lock);
308 INIT_LIST_HEAD(&object->cache_link);
309 INIT_HLIST_NODE(&object->cookie_link);
310 INIT_WORK(&object->work, fscache_object_work_func);
311 INIT_LIST_HEAD(&object->dependents);
312 INIT_LIST_HEAD(&object->dep_link);
313 INIT_LIST_HEAD(&object->pending_ops);
314 object->n_children = 0;
315 object->n_ops = object->n_in_progress = object->n_exclusive = 0;
316 object->events = 0;
317 object->store_limit = 0;
318 object->store_limit_l = 0;
319 object->cache = cache;
320 object->cookie = cookie;
321 atomic_inc(&cookie->usage);
322 object->parent = NULL;
323 #ifdef CONFIG_FSCACHE_OBJECT_LIST
324 RB_CLEAR_NODE(&object->objlist_link);
325 #endif
326
327 object->oob_event_mask = 0;
328 for (t = object->oob_table; t->events; t++)
329 object->oob_event_mask |= t->events;
330 object->event_mask = object->oob_event_mask;
331 for (t = object->state->transitions; t->events; t++)
332 object->event_mask |= t->events;
333 }
334 EXPORT_SYMBOL(fscache_object_init);
335
336 /*
337 * Mark the object as no longer being live, making sure that we synchronise
338 * against op submission.
339 */
340 static inline void fscache_mark_object_dead(struct fscache_object *object)
341 {
342 spin_lock(&object->lock);
343 clear_bit(FSCACHE_OBJECT_IS_LIVE, &object->flags);
344 spin_unlock(&object->lock);
345 }
346
347 /*
348 * Abort object initialisation before we start it.
349 */
350 static const struct fscache_state *fscache_abort_initialisation(struct fscache_object *object,
351 int event)
352 {
353 _enter("{OBJ%x},%d", object->debug_id, event);
354
355 object->oob_event_mask = 0;
356 fscache_dequeue_object(object);
357 return transit_to(KILL_OBJECT);
358 }
359
360 /*
361 * initialise an object
362 * - check the specified object's parent to see if we can make use of it
363 * immediately to do a creation
364 * - we may need to start the process of creating a parent and we need to wait
365 * for the parent's lookup and creation to complete if it's not there yet
366 */
367 static const struct fscache_state *fscache_initialise_object(struct fscache_object *object,
368 int event)
369 {
370 struct fscache_object *parent;
371 bool success;
372
373 _enter("{OBJ%x},%d", object->debug_id, event);
374
375 ASSERT(list_empty(&object->dep_link));
376
377 parent = object->parent;
378 if (!parent) {
379 _leave(" [no parent]");
380 return transit_to(DROP_OBJECT);
381 }
382
383 _debug("parent: %s of:%lx", parent->state->name, parent->flags);
384
385 if (fscache_object_is_dying(parent)) {
386 _leave(" [bad parent]");
387 return transit_to(DROP_OBJECT);
388 }
389
390 if (fscache_object_is_available(parent)) {
391 _leave(" [ready]");
392 return transit_to(PARENT_READY);
393 }
394
395 _debug("wait");
396
397 spin_lock(&parent->lock);
398 fscache_stat(&fscache_n_cop_grab_object);
399 success = false;
400 if (fscache_object_is_live(parent) &&
401 object->cache->ops->grab_object(object)) {
402 list_add(&object->dep_link, &parent->dependents);
403 success = true;
404 }
405 fscache_stat_d(&fscache_n_cop_grab_object);
406 spin_unlock(&parent->lock);
407 if (!success) {
408 _leave(" [grab failed]");
409 return transit_to(DROP_OBJECT);
410 }
411
412 /* fscache_acquire_non_index_cookie() uses this
413 * to wake the chain up */
414 fscache_raise_event(parent, FSCACHE_OBJECT_EV_NEW_CHILD);
415 _leave(" [wait]");
416 return transit_to(WAIT_FOR_PARENT);
417 }
418
419 /*
420 * Once the parent object is ready, we should kick off our lookup op.
421 */
422 static const struct fscache_state *fscache_parent_ready(struct fscache_object *object,
423 int event)
424 {
425 struct fscache_object *parent = object->parent;
426
427 _enter("{OBJ%x},%d", object->debug_id, event);
428
429 ASSERT(parent != NULL);
430
431 spin_lock(&parent->lock);
432 parent->n_ops++;
433 parent->n_obj_ops++;
434 object->lookup_jif = jiffies;
435 spin_unlock(&parent->lock);
436
437 _leave("");
438 return transit_to(LOOK_UP_OBJECT);
439 }
440
441 /*
442 * look an object up in the cache from which it was allocated
443 * - we hold an "access lock" on the parent object, so the parent object cannot
444 * be withdrawn by either party till we've finished
445 */
446 static const struct fscache_state *fscache_look_up_object(struct fscache_object *object,
447 int event)
448 {
449 struct fscache_cookie *cookie = object->cookie;
450 struct fscache_object *parent = object->parent;
451 int ret;
452
453 _enter("{OBJ%x},%d", object->debug_id, event);
454
455 object->oob_table = fscache_osm_lookup_oob;
456
457 ASSERT(parent != NULL);
458 ASSERTCMP(parent->n_ops, >, 0);
459 ASSERTCMP(parent->n_obj_ops, >, 0);
460
461 /* make sure the parent is still available */
462 ASSERT(fscache_object_is_available(parent));
463
464 if (fscache_object_is_dying(parent) ||
465 test_bit(FSCACHE_IOERROR, &object->cache->flags) ||
466 !fscache_use_cookie(object)) {
467 _leave(" [unavailable]");
468 return transit_to(LOOKUP_FAILURE);
469 }
470
471 _debug("LOOKUP \"%s\" in \"%s\"",
472 cookie->def->name, object->cache->tag->name);
473
474 fscache_stat(&fscache_n_object_lookups);
475 fscache_stat(&fscache_n_cop_lookup_object);
476 ret = object->cache->ops->lookup_object(object);
477 fscache_stat_d(&fscache_n_cop_lookup_object);
478
479 fscache_unuse_cookie(object);
480
481 if (ret == -ETIMEDOUT) {
482 /* probably stuck behind another object, so move this one to
483 * the back of the queue */
484 fscache_stat(&fscache_n_object_lookups_timed_out);
485 _leave(" [timeout]");
486 return NO_TRANSIT;
487 }
488
489 if (ret < 0) {
490 _leave(" [error]");
491 return transit_to(LOOKUP_FAILURE);
492 }
493
494 _leave(" [ok]");
495 return transit_to(OBJECT_AVAILABLE);
496 }
497
498 /**
499 * fscache_object_lookup_negative - Note negative cookie lookup
500 * @object: Object pointing to cookie to mark
501 *
502 * Note negative lookup, permitting those waiting to read data from an already
503 * existing backing object to continue as there's no data for them to read.
504 */
505 void fscache_object_lookup_negative(struct fscache_object *object)
506 {
507 struct fscache_cookie *cookie = object->cookie;
508
509 _enter("{OBJ%x,%s}", object->debug_id, object->state->name);
510
511 if (!test_and_set_bit(FSCACHE_OBJECT_IS_LOOKED_UP, &object->flags)) {
512 fscache_stat(&fscache_n_object_lookups_negative);
513
514 /* Allow write requests to begin stacking up and read requests to begin
515 * returning ENODATA.
516 */
517 set_bit(FSCACHE_COOKIE_NO_DATA_YET, &cookie->flags);
518 clear_bit(FSCACHE_COOKIE_UNAVAILABLE, &cookie->flags);
519
520 _debug("wake up lookup %p", &cookie->flags);
521 clear_bit_unlock(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags);
522 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP);
523 }
524 _leave("");
525 }
526 EXPORT_SYMBOL(fscache_object_lookup_negative);
527
528 /**
529 * fscache_obtained_object - Note successful object lookup or creation
530 * @object: Object pointing to cookie to mark
531 *
532 * Note successful lookup and/or creation, permitting those waiting to write
533 * data to a backing object to continue.
534 *
535 * Note that after calling this, an object's cookie may be relinquished by the
536 * netfs, and so must be accessed with object lock held.
537 */
538 void fscache_obtained_object(struct fscache_object *object)
539 {
540 struct fscache_cookie *cookie = object->cookie;
541
542 _enter("{OBJ%x,%s}", object->debug_id, object->state->name);
543
544 /* if we were still looking up, then we must have a positive lookup
545 * result, in which case there may be data available */
546 if (!test_and_set_bit(FSCACHE_OBJECT_IS_LOOKED_UP, &object->flags)) {
547 fscache_stat(&fscache_n_object_lookups_positive);
548
549 /* We do (presumably) have data */
550 clear_bit_unlock(FSCACHE_COOKIE_NO_DATA_YET, &cookie->flags);
551 clear_bit(FSCACHE_COOKIE_UNAVAILABLE, &cookie->flags);
552
553 /* Allow write requests to begin stacking up and read requests
554 * to begin shovelling data.
555 */
556 clear_bit_unlock(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags);
557 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP);
558 } else {
559 fscache_stat(&fscache_n_object_created);
560 }
561
562 set_bit(FSCACHE_OBJECT_IS_AVAILABLE, &object->flags);
563 _leave("");
564 }
565 EXPORT_SYMBOL(fscache_obtained_object);
566
567 /*
568 * handle an object that has just become available
569 */
570 static const struct fscache_state *fscache_object_available(struct fscache_object *object,
571 int event)
572 {
573 _enter("{OBJ%x},%d", object->debug_id, event);
574
575 object->oob_table = fscache_osm_run_oob;
576
577 spin_lock(&object->lock);
578
579 fscache_done_parent_op(object);
580 if (object->n_in_progress == 0) {
581 if (object->n_ops > 0) {
582 ASSERTCMP(object->n_ops, >=, object->n_obj_ops);
583 fscache_start_operations(object);
584 } else {
585 ASSERT(list_empty(&object->pending_ops));
586 }
587 }
588 spin_unlock(&object->lock);
589
590 fscache_stat(&fscache_n_cop_lookup_complete);
591 object->cache->ops->lookup_complete(object);
592 fscache_stat_d(&fscache_n_cop_lookup_complete);
593
594 fscache_hist(fscache_obj_instantiate_histogram, object->lookup_jif);
595 fscache_stat(&fscache_n_object_avail);
596
597 _leave("");
598 return transit_to(JUMPSTART_DEPS);
599 }
600
601 /*
602 * Wake up this object's dependent objects now that we've become available.
603 */
604 static const struct fscache_state *fscache_jumpstart_dependents(struct fscache_object *object,
605 int event)
606 {
607 _enter("{OBJ%x},%d", object->debug_id, event);
608
609 if (!fscache_enqueue_dependents(object, FSCACHE_OBJECT_EV_PARENT_READY))
610 return NO_TRANSIT; /* Not finished; requeue */
611 return transit_to(WAIT_FOR_CMD);
612 }
613
614 /*
615 * Handle lookup or creation failute.
616 */
617 static const struct fscache_state *fscache_lookup_failure(struct fscache_object *object,
618 int event)
619 {
620 struct fscache_cookie *cookie;
621
622 _enter("{OBJ%x},%d", object->debug_id, event);
623
624 object->oob_event_mask = 0;
625
626 fscache_stat(&fscache_n_cop_lookup_complete);
627 object->cache->ops->lookup_complete(object);
628 fscache_stat_d(&fscache_n_cop_lookup_complete);
629
630 set_bit(FSCACHE_OBJECT_KILLED_BY_CACHE, &object->flags);
631
632 cookie = object->cookie;
633 set_bit(FSCACHE_COOKIE_UNAVAILABLE, &cookie->flags);
634 if (test_and_clear_bit(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags))
635 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP);
636
637 fscache_done_parent_op(object);
638 return transit_to(KILL_OBJECT);
639 }
640
641 /*
642 * Wait for completion of all active operations on this object and the death of
643 * all child objects of this object.
644 */
645 static const struct fscache_state *fscache_kill_object(struct fscache_object *object,
646 int event)
647 {
648 _enter("{OBJ%x,%d,%d},%d",
649 object->debug_id, object->n_ops, object->n_children, event);
650
651 fscache_mark_object_dead(object);
652 object->oob_event_mask = 0;
653
654 if (test_bit(FSCACHE_OBJECT_RETIRED, &object->flags)) {
655 /* Reject any new read/write ops and abort any that are pending. */
656 clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags);
657 fscache_cancel_all_ops(object);
658 }
659
660 if (list_empty(&object->dependents) &&
661 object->n_ops == 0 &&
662 object->n_children == 0)
663 return transit_to(DROP_OBJECT);
664
665 if (object->n_in_progress == 0) {
666 spin_lock(&object->lock);
667 if (object->n_ops > 0 && object->n_in_progress == 0)
668 fscache_start_operations(object);
669 spin_unlock(&object->lock);
670 }
671
672 if (!list_empty(&object->dependents))
673 return transit_to(KILL_DEPENDENTS);
674
675 return transit_to(WAIT_FOR_CLEARANCE);
676 }
677
678 /*
679 * Kill dependent objects.
680 */
681 static const struct fscache_state *fscache_kill_dependents(struct fscache_object *object,
682 int event)
683 {
684 _enter("{OBJ%x},%d", object->debug_id, event);
685
686 if (!fscache_enqueue_dependents(object, FSCACHE_OBJECT_EV_KILL))
687 return NO_TRANSIT; /* Not finished */
688 return transit_to(WAIT_FOR_CLEARANCE);
689 }
690
691 /*
692 * Drop an object's attachments
693 */
694 static const struct fscache_state *fscache_drop_object(struct fscache_object *object,
695 int event)
696 {
697 struct fscache_object *parent = object->parent;
698 struct fscache_cookie *cookie = object->cookie;
699 struct fscache_cache *cache = object->cache;
700 bool awaken = false;
701
702 _enter("{OBJ%x,%d},%d", object->debug_id, object->n_children, event);
703
704 ASSERT(cookie != NULL);
705 ASSERT(!hlist_unhashed(&object->cookie_link));
706
707 /* Make sure the cookie no longer points here and that the netfs isn't
708 * waiting for us.
709 */
710 spin_lock(&cookie->lock);
711 hlist_del_init(&object->cookie_link);
712 if (hlist_empty(&cookie->backing_objects) &&
713 test_and_clear_bit(FSCACHE_COOKIE_INVALIDATING, &cookie->flags))
714 awaken = true;
715 spin_unlock(&cookie->lock);
716
717 if (awaken)
718 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_INVALIDATING);
719
720 /* Prevent a race with our last child, which has to signal EV_CLEARED
721 * before dropping our spinlock.
722 */
723 spin_lock(&object->lock);
724 spin_unlock(&object->lock);
725
726 /* Discard from the cache's collection of objects */
727 spin_lock(&cache->object_list_lock);
728 list_del_init(&object->cache_link);
729 spin_unlock(&cache->object_list_lock);
730
731 fscache_stat(&fscache_n_cop_drop_object);
732 cache->ops->drop_object(object);
733 fscache_stat_d(&fscache_n_cop_drop_object);
734
735 /* The parent object wants to know when all it dependents have gone */
736 if (parent) {
737 _debug("release parent OBJ%x {%d}",
738 parent->debug_id, parent->n_children);
739
740 spin_lock(&parent->lock);
741 parent->n_children--;
742 if (parent->n_children == 0)
743 fscache_raise_event(parent, FSCACHE_OBJECT_EV_CLEARED);
744 spin_unlock(&parent->lock);
745 object->parent = NULL;
746 }
747
748 /* this just shifts the object release to the work processor */
749 fscache_put_object(object);
750 fscache_stat(&fscache_n_object_dead);
751
752 _leave("");
753 return transit_to(OBJECT_DEAD);
754 }
755
756 /*
757 * get a ref on an object
758 */
759 static int fscache_get_object(struct fscache_object *object)
760 {
761 int ret;
762
763 fscache_stat(&fscache_n_cop_grab_object);
764 ret = object->cache->ops->grab_object(object) ? 0 : -EAGAIN;
765 fscache_stat_d(&fscache_n_cop_grab_object);
766 return ret;
767 }
768
769 /*
770 * Discard a ref on an object
771 */
772 static void fscache_put_object(struct fscache_object *object)
773 {
774 fscache_stat(&fscache_n_cop_put_object);
775 object->cache->ops->put_object(object);
776 fscache_stat_d(&fscache_n_cop_put_object);
777 }
778
779 /**
780 * fscache_object_destroy - Note that a cache object is about to be destroyed
781 * @object: The object to be destroyed
782 *
783 * Note the imminent destruction and deallocation of a cache object record.
784 */
785 void fscache_object_destroy(struct fscache_object *object)
786 {
787 fscache_objlist_remove(object);
788
789 /* We can get rid of the cookie now */
790 fscache_cookie_put(object->cookie);
791 object->cookie = NULL;
792 }
793 EXPORT_SYMBOL(fscache_object_destroy);
794
795 /*
796 * enqueue an object for metadata-type processing
797 */
798 void fscache_enqueue_object(struct fscache_object *object)
799 {
800 _enter("{OBJ%x}", object->debug_id);
801
802 if (fscache_get_object(object) >= 0) {
803 wait_queue_head_t *cong_wq =
804 &get_cpu_var(fscache_object_cong_wait);
805
806 if (queue_work(fscache_object_wq, &object->work)) {
807 if (fscache_object_congested())
808 wake_up(cong_wq);
809 } else
810 fscache_put_object(object);
811
812 put_cpu_var(fscache_object_cong_wait);
813 }
814 }
815
816 /**
817 * fscache_object_sleep_till_congested - Sleep until object wq is congested
818 * @timeoutp: Scheduler sleep timeout
819 *
820 * Allow an object handler to sleep until the object workqueue is congested.
821 *
822 * The caller must set up a wake up event before calling this and must have set
823 * the appropriate sleep mode (such as TASK_UNINTERRUPTIBLE) and tested its own
824 * condition before calling this function as no test is made here.
825 *
826 * %true is returned if the object wq is congested, %false otherwise.
827 */
828 bool fscache_object_sleep_till_congested(signed long *timeoutp)
829 {
830 wait_queue_head_t *cong_wq = this_cpu_ptr(&fscache_object_cong_wait);
831 DEFINE_WAIT(wait);
832
833 if (fscache_object_congested())
834 return true;
835
836 add_wait_queue_exclusive(cong_wq, &wait);
837 if (!fscache_object_congested())
838 *timeoutp = schedule_timeout(*timeoutp);
839 finish_wait(cong_wq, &wait);
840
841 return fscache_object_congested();
842 }
843 EXPORT_SYMBOL_GPL(fscache_object_sleep_till_congested);
844
845 /*
846 * Enqueue the dependents of an object for metadata-type processing.
847 *
848 * If we don't manage to finish the list before the scheduler wants to run
849 * again then return false immediately. We return true if the list was
850 * cleared.
851 */
852 static bool fscache_enqueue_dependents(struct fscache_object *object, int event)
853 {
854 struct fscache_object *dep;
855 bool ret = true;
856
857 _enter("{OBJ%x}", object->debug_id);
858
859 if (list_empty(&object->dependents))
860 return true;
861
862 spin_lock(&object->lock);
863
864 while (!list_empty(&object->dependents)) {
865 dep = list_entry(object->dependents.next,
866 struct fscache_object, dep_link);
867 list_del_init(&dep->dep_link);
868
869 fscache_raise_event(dep, event);
870 fscache_put_object(dep);
871
872 if (!list_empty(&object->dependents) && need_resched()) {
873 ret = false;
874 break;
875 }
876 }
877
878 spin_unlock(&object->lock);
879 return ret;
880 }
881
882 /*
883 * remove an object from whatever queue it's waiting on
884 */
885 static void fscache_dequeue_object(struct fscache_object *object)
886 {
887 _enter("{OBJ%x}", object->debug_id);
888
889 if (!list_empty(&object->dep_link)) {
890 spin_lock(&object->parent->lock);
891 list_del_init(&object->dep_link);
892 spin_unlock(&object->parent->lock);
893 }
894
895 _leave("");
896 }
897
898 /**
899 * fscache_check_aux - Ask the netfs whether an object on disk is still valid
900 * @object: The object to ask about
901 * @data: The auxiliary data for the object
902 * @datalen: The size of the auxiliary data
903 *
904 * This function consults the netfs about the coherency state of an object.
905 * The caller must be holding a ref on cookie->n_active (held by
906 * fscache_look_up_object() on behalf of the cache backend during object lookup
907 * and creation).
908 */
909 enum fscache_checkaux fscache_check_aux(struct fscache_object *object,
910 const void *data, uint16_t datalen)
911 {
912 enum fscache_checkaux result;
913
914 if (!object->cookie->def->check_aux) {
915 fscache_stat(&fscache_n_checkaux_none);
916 return FSCACHE_CHECKAUX_OKAY;
917 }
918
919 result = object->cookie->def->check_aux(object->cookie->netfs_data,
920 data, datalen);
921 switch (result) {
922 /* entry okay as is */
923 case FSCACHE_CHECKAUX_OKAY:
924 fscache_stat(&fscache_n_checkaux_okay);
925 break;
926
927 /* entry requires update */
928 case FSCACHE_CHECKAUX_NEEDS_UPDATE:
929 fscache_stat(&fscache_n_checkaux_update);
930 break;
931
932 /* entry requires deletion */
933 case FSCACHE_CHECKAUX_OBSOLETE:
934 fscache_stat(&fscache_n_checkaux_obsolete);
935 break;
936
937 default:
938 BUG();
939 }
940
941 return result;
942 }
943 EXPORT_SYMBOL(fscache_check_aux);
944
945 /*
946 * Asynchronously invalidate an object.
947 */
948 static const struct fscache_state *_fscache_invalidate_object(struct fscache_object *object,
949 int event)
950 {
951 struct fscache_operation *op;
952 struct fscache_cookie *cookie = object->cookie;
953
954 _enter("{OBJ%x},%d", object->debug_id, event);
955
956 /* We're going to need the cookie. If the cookie is not available then
957 * retire the object instead.
958 */
959 if (!fscache_use_cookie(object)) {
960 ASSERT(object->cookie->stores.rnode == NULL);
961 set_bit(FSCACHE_OBJECT_RETIRED, &object->flags);
962 _leave(" [no cookie]");
963 return transit_to(KILL_OBJECT);
964 }
965
966 /* Reject any new read/write ops and abort any that are pending. */
967 fscache_invalidate_writes(cookie);
968 clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags);
969 fscache_cancel_all_ops(object);
970
971 /* Now we have to wait for in-progress reads and writes */
972 op = kzalloc(sizeof(*op), GFP_KERNEL);
973 if (!op)
974 goto nomem;
975
976 fscache_operation_init(op, object->cache->ops->invalidate_object,
977 NULL, NULL);
978 op->flags = FSCACHE_OP_ASYNC |
979 (1 << FSCACHE_OP_EXCLUSIVE) |
980 (1 << FSCACHE_OP_UNUSE_COOKIE);
981
982 spin_lock(&cookie->lock);
983 if (fscache_submit_exclusive_op(object, op) < 0)
984 goto submit_op_failed;
985 spin_unlock(&cookie->lock);
986 fscache_put_operation(op);
987
988 /* Once we've completed the invalidation, we know there will be no data
989 * stored in the cache and thus we can reinstate the data-check-skip
990 * optimisation.
991 */
992 set_bit(FSCACHE_COOKIE_NO_DATA_YET, &cookie->flags);
993
994 /* We can allow read and write requests to come in once again. They'll
995 * queue up behind our exclusive invalidation operation.
996 */
997 if (test_and_clear_bit(FSCACHE_COOKIE_INVALIDATING, &cookie->flags))
998 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_INVALIDATING);
999 _leave(" [ok]");
1000 return transit_to(UPDATE_OBJECT);
1001
1002 nomem:
1003 fscache_mark_object_dead(object);
1004 fscache_unuse_cookie(object);
1005 _leave(" [ENOMEM]");
1006 return transit_to(KILL_OBJECT);
1007
1008 submit_op_failed:
1009 fscache_mark_object_dead(object);
1010 spin_unlock(&cookie->lock);
1011 fscache_unuse_cookie(object);
1012 kfree(op);
1013 _leave(" [EIO]");
1014 return transit_to(KILL_OBJECT);
1015 }
1016
1017 static const struct fscache_state *fscache_invalidate_object(struct fscache_object *object,
1018 int event)
1019 {
1020 const struct fscache_state *s;
1021
1022 fscache_stat(&fscache_n_invalidates_run);
1023 fscache_stat(&fscache_n_cop_invalidate_object);
1024 s = _fscache_invalidate_object(object, event);
1025 fscache_stat_d(&fscache_n_cop_invalidate_object);
1026 return s;
1027 }
1028
1029 /*
1030 * Asynchronously update an object.
1031 */
1032 static const struct fscache_state *fscache_update_object(struct fscache_object *object,
1033 int event)
1034 {
1035 _enter("{OBJ%x},%d", object->debug_id, event);
1036
1037 fscache_stat(&fscache_n_updates_run);
1038 fscache_stat(&fscache_n_cop_update_object);
1039 object->cache->ops->update_object(object);
1040 fscache_stat_d(&fscache_n_cop_update_object);
1041
1042 _leave("");
1043 return transit_to(WAIT_FOR_CMD);
1044 }
1045
1046 /**
1047 * fscache_object_retrying_stale - Note retrying stale object
1048 * @object: The object that will be retried
1049 *
1050 * Note that an object lookup found an on-disk object that was adjudged to be
1051 * stale and has been deleted. The lookup will be retried.
1052 */
1053 void fscache_object_retrying_stale(struct fscache_object *object)
1054 {
1055 fscache_stat(&fscache_n_cache_no_space_reject);
1056 }
1057 EXPORT_SYMBOL(fscache_object_retrying_stale);
1058
1059 /**
1060 * fscache_object_mark_killed - Note that an object was killed
1061 * @object: The object that was culled
1062 * @why: The reason the object was killed.
1063 *
1064 * Note that an object was killed. Returns true if the object was
1065 * already marked killed, false if it wasn't.
1066 */
1067 void fscache_object_mark_killed(struct fscache_object *object,
1068 enum fscache_why_object_killed why)
1069 {
1070 if (test_and_set_bit(FSCACHE_OBJECT_KILLED_BY_CACHE, &object->flags)) {
1071 pr_err("Error: Object already killed by cache [%s]\n",
1072 object->cache->identifier);
1073 return;
1074 }
1075
1076 switch (why) {
1077 case FSCACHE_OBJECT_NO_SPACE:
1078 fscache_stat(&fscache_n_cache_no_space_reject);
1079 break;
1080 case FSCACHE_OBJECT_IS_STALE:
1081 fscache_stat(&fscache_n_cache_stale_objects);
1082 break;
1083 case FSCACHE_OBJECT_WAS_RETIRED:
1084 fscache_stat(&fscache_n_cache_retired_objects);
1085 break;
1086 case FSCACHE_OBJECT_WAS_CULLED:
1087 fscache_stat(&fscache_n_cache_culled_objects);
1088 break;
1089 }
1090 }
1091 EXPORT_SYMBOL(fscache_object_mark_killed);
1092
1093 /*
1094 * The object is dead. We can get here if an object gets queued by an event
1095 * that would lead to its death (such as EV_KILL) when the dispatcher is
1096 * already running (and so can be requeued) but hasn't yet cleared the event
1097 * mask.
1098 */
1099 static const struct fscache_state *fscache_object_dead(struct fscache_object *object,
1100 int event)
1101 {
1102 if (!test_and_set_bit(FSCACHE_OBJECT_RUN_AFTER_DEAD,
1103 &object->flags))
1104 return NO_TRANSIT;
1105
1106 WARN(true, "FS-Cache object redispatched after death");
1107 return NO_TRANSIT;
1108 }