1 /* FS-Cache object state machine handler
3 * Copyright (C) 2007 Red Hat, Inc. All Rights Reserved.
4 * Written by David Howells (dhowells@redhat.com)
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
11 * See Documentation/filesystems/caching/object.txt for a description of the
12 * object state machine and the in-kernel representations.
15 #define FSCACHE_DEBUG_LEVEL COOKIE
16 #include <linux/module.h>
17 #include <linux/slab.h>
18 #include <linux/prefetch.h>
21 static const struct fscache_state
*fscache_abort_initialisation(struct fscache_object
*, int);
22 static const struct fscache_state
*fscache_kill_dependents(struct fscache_object
*, int);
23 static const struct fscache_state
*fscache_drop_object(struct fscache_object
*, int);
24 static const struct fscache_state
*fscache_initialise_object(struct fscache_object
*, int);
25 static const struct fscache_state
*fscache_invalidate_object(struct fscache_object
*, int);
26 static const struct fscache_state
*fscache_jumpstart_dependents(struct fscache_object
*, int);
27 static const struct fscache_state
*fscache_kill_object(struct fscache_object
*, int);
28 static const struct fscache_state
*fscache_lookup_failure(struct fscache_object
*, int);
29 static const struct fscache_state
*fscache_look_up_object(struct fscache_object
*, int);
30 static const struct fscache_state
*fscache_object_available(struct fscache_object
*, int);
31 static const struct fscache_state
*fscache_parent_ready(struct fscache_object
*, int);
32 static const struct fscache_state
*fscache_update_object(struct fscache_object
*, int);
33 static const struct fscache_state
*fscache_object_dead(struct fscache_object
*, int);
35 #define __STATE_NAME(n) fscache_osm_##n
36 #define STATE(n) (&__STATE_NAME(n))
39 * Define a work state. Work states are execution states. No event processing
40 * is performed by them. The function attached to a work state returns a
41 * pointer indicating the next state to which the state machine should
42 * transition. Returning NO_TRANSIT repeats the current state, but goes back
43 * to the scheduler first.
45 #define WORK_STATE(n, sn, f) \
46 const struct fscache_state __STATE_NAME(n) = { \
53 * Returns from work states.
55 #define transit_to(state) ({ prefetch(&STATE(state)->work); STATE(state); })
57 #define NO_TRANSIT ((struct fscache_state *)NULL)
60 * Define a wait state. Wait states are event processing states. No execution
61 * is performed by them. Wait states are just tables of "if event X occurs,
62 * clear it and transition to state Y". The dispatcher returns to the
63 * scheduler if none of the events in which the wait state has an interest are
66 #define WAIT_STATE(n, sn, ...) \
67 const struct fscache_state __STATE_NAME(n) = { \
71 .transitions = { __VA_ARGS__, { 0, NULL } } \
74 #define TRANSIT_TO(state, emask) \
75 { .events = (emask), .transit_to = STATE(state) }
78 * The object state machine.
80 static WORK_STATE(INIT_OBJECT
, "INIT", fscache_initialise_object
);
81 static WORK_STATE(PARENT_READY
, "PRDY", fscache_parent_ready
);
82 static WORK_STATE(ABORT_INIT
, "ABRT", fscache_abort_initialisation
);
83 static WORK_STATE(LOOK_UP_OBJECT
, "LOOK", fscache_look_up_object
);
84 static WORK_STATE(CREATE_OBJECT
, "CRTO", fscache_look_up_object
);
85 static WORK_STATE(OBJECT_AVAILABLE
, "AVBL", fscache_object_available
);
86 static WORK_STATE(JUMPSTART_DEPS
, "JUMP", fscache_jumpstart_dependents
);
88 static WORK_STATE(INVALIDATE_OBJECT
, "INVL", fscache_invalidate_object
);
89 static WORK_STATE(UPDATE_OBJECT
, "UPDT", fscache_update_object
);
91 static WORK_STATE(LOOKUP_FAILURE
, "LCFL", fscache_lookup_failure
);
92 static WORK_STATE(KILL_OBJECT
, "KILL", fscache_kill_object
);
93 static WORK_STATE(KILL_DEPENDENTS
, "KDEP", fscache_kill_dependents
);
94 static WORK_STATE(DROP_OBJECT
, "DROP", fscache_drop_object
);
95 static WORK_STATE(OBJECT_DEAD
, "DEAD", fscache_object_dead
);
97 static WAIT_STATE(WAIT_FOR_INIT
, "?INI",
98 TRANSIT_TO(INIT_OBJECT
, 1 << FSCACHE_OBJECT_EV_NEW_CHILD
));
100 static WAIT_STATE(WAIT_FOR_PARENT
, "?PRN",
101 TRANSIT_TO(PARENT_READY
, 1 << FSCACHE_OBJECT_EV_PARENT_READY
));
103 static WAIT_STATE(WAIT_FOR_CMD
, "?CMD",
104 TRANSIT_TO(INVALIDATE_OBJECT
, 1 << FSCACHE_OBJECT_EV_INVALIDATE
),
105 TRANSIT_TO(UPDATE_OBJECT
, 1 << FSCACHE_OBJECT_EV_UPDATE
),
106 TRANSIT_TO(JUMPSTART_DEPS
, 1 << FSCACHE_OBJECT_EV_NEW_CHILD
));
108 static WAIT_STATE(WAIT_FOR_CLEARANCE
, "?CLR",
109 TRANSIT_TO(KILL_OBJECT
, 1 << FSCACHE_OBJECT_EV_CLEARED
));
112 * Out-of-band event transition tables. These are for handling unexpected
113 * events, such as an I/O error. If an OOB event occurs, the state machine
114 * clears and disables the event and forces a transition to the nominated work
115 * state (acurrently executing work states will complete first).
117 * In such a situation, object->state remembers the state the machine should
118 * have been in/gone to and returning NO_TRANSIT returns to that.
120 static const struct fscache_transition fscache_osm_init_oob
[] = {
121 TRANSIT_TO(ABORT_INIT
,
122 (1 << FSCACHE_OBJECT_EV_ERROR
) |
123 (1 << FSCACHE_OBJECT_EV_KILL
)),
127 static const struct fscache_transition fscache_osm_lookup_oob
[] = {
128 TRANSIT_TO(LOOKUP_FAILURE
,
129 (1 << FSCACHE_OBJECT_EV_ERROR
) |
130 (1 << FSCACHE_OBJECT_EV_KILL
)),
134 static const struct fscache_transition fscache_osm_run_oob
[] = {
135 TRANSIT_TO(KILL_OBJECT
,
136 (1 << FSCACHE_OBJECT_EV_ERROR
) |
137 (1 << FSCACHE_OBJECT_EV_KILL
)),
141 static int fscache_get_object(struct fscache_object
*);
142 static void fscache_put_object(struct fscache_object
*);
143 static bool fscache_enqueue_dependents(struct fscache_object
*, int);
144 static void fscache_dequeue_object(struct fscache_object
*);
147 * we need to notify the parent when an op completes that we had outstanding
150 static inline void fscache_done_parent_op(struct fscache_object
*object
)
152 struct fscache_object
*parent
= object
->parent
;
154 _enter("OBJ%x {OBJ%x,%x}",
155 object
->debug_id
, parent
->debug_id
, parent
->n_ops
);
157 spin_lock_nested(&parent
->lock
, 1);
160 if (parent
->n_ops
== 0)
161 fscache_raise_event(parent
, FSCACHE_OBJECT_EV_CLEARED
);
162 spin_unlock(&parent
->lock
);
166 * Object state machine dispatcher.
168 static void fscache_object_sm_dispatcher(struct fscache_object
*object
)
170 const struct fscache_transition
*t
;
171 const struct fscache_state
*state
, *new_state
;
172 unsigned long events
, event_mask
;
175 ASSERT(object
!= NULL
);
177 _enter("{OBJ%x,%s,%lx}",
178 object
->debug_id
, object
->state
->name
, object
->events
);
180 event_mask
= object
->event_mask
;
182 object
->event_mask
= 0; /* Mask normal event handling */
183 state
= object
->state
;
185 events
= object
->events
;
187 /* Handle any out-of-band events (typically an error) */
188 if (events
& object
->oob_event_mask
) {
189 _debug("{OBJ%x} oob %lx",
190 object
->debug_id
, events
& object
->oob_event_mask
);
191 for (t
= object
->oob_table
; t
->events
; t
++) {
192 if (events
& t
->events
) {
193 state
= t
->transit_to
;
194 ASSERT(state
->work
!= NULL
);
195 event
= fls(events
& t
->events
) - 1;
196 __clear_bit(event
, &object
->oob_event_mask
);
197 clear_bit(event
, &object
->events
);
198 goto execute_work_state
;
203 /* Wait states are just transition tables */
205 if (events
& event_mask
) {
206 for (t
= state
->transitions
; t
->events
; t
++) {
207 if (events
& t
->events
) {
208 new_state
= t
->transit_to
;
209 event
= fls(events
& t
->events
) - 1;
210 clear_bit(event
, &object
->events
);
211 _debug("{OBJ%x} ev %d: %s -> %s",
212 object
->debug_id
, event
,
213 state
->name
, new_state
->name
);
214 object
->state
= state
= new_state
;
215 goto execute_work_state
;
219 /* The event mask didn't include all the tabled bits */
222 /* Randomly woke up */
227 _debug("{OBJ%x} exec %s", object
->debug_id
, state
->name
);
229 new_state
= state
->work(object
, event
);
231 if (new_state
== NO_TRANSIT
) {
232 _debug("{OBJ%x} %s notrans", object
->debug_id
, state
->name
);
233 if (unlikely(state
== STATE(OBJECT_DEAD
))) {
237 fscache_enqueue_object(object
);
238 event_mask
= object
->oob_event_mask
;
242 _debug("{OBJ%x} %s -> %s",
243 object
->debug_id
, state
->name
, new_state
->name
);
244 object
->state
= state
= new_state
;
247 if (unlikely(state
== STATE(OBJECT_DEAD
))) {
254 /* Transited to wait state */
255 event_mask
= object
->oob_event_mask
;
256 for (t
= state
->transitions
; t
->events
; t
++)
257 event_mask
|= t
->events
;
260 object
->event_mask
= event_mask
;
262 events
= object
->events
;
263 if (events
& event_mask
)
265 _leave(" [msk %lx]", event_mask
);
271 static void fscache_object_work_func(struct work_struct
*work
)
273 struct fscache_object
*object
=
274 container_of(work
, struct fscache_object
, work
);
277 _enter("{OBJ%x}", object
->debug_id
);
280 fscache_object_sm_dispatcher(object
);
281 fscache_hist(fscache_objs_histogram
, start
);
282 fscache_put_object(object
);
286 * fscache_object_init - Initialise a cache object description
287 * @object: Object description
288 * @cookie: Cookie object will be attached to
289 * @cache: Cache in which backing object will be found
291 * Initialise a cache object description to its basic values.
293 * See Documentation/filesystems/caching/backend-api.txt for a complete
296 void fscache_object_init(struct fscache_object
*object
,
297 struct fscache_cookie
*cookie
,
298 struct fscache_cache
*cache
)
300 const struct fscache_transition
*t
;
302 atomic_inc(&cache
->object_count
);
304 object
->state
= STATE(WAIT_FOR_INIT
);
305 object
->oob_table
= fscache_osm_init_oob
;
306 object
->flags
= 1 << FSCACHE_OBJECT_IS_LIVE
;
307 spin_lock_init(&object
->lock
);
308 INIT_LIST_HEAD(&object
->cache_link
);
309 INIT_HLIST_NODE(&object
->cookie_link
);
310 INIT_WORK(&object
->work
, fscache_object_work_func
);
311 INIT_LIST_HEAD(&object
->dependents
);
312 INIT_LIST_HEAD(&object
->dep_link
);
313 INIT_LIST_HEAD(&object
->pending_ops
);
314 object
->n_children
= 0;
315 object
->n_ops
= object
->n_in_progress
= object
->n_exclusive
= 0;
317 object
->store_limit
= 0;
318 object
->store_limit_l
= 0;
319 object
->cache
= cache
;
320 object
->cookie
= cookie
;
321 object
->parent
= NULL
;
322 #ifdef CONFIG_FSCACHE_OBJECT_LIST
323 RB_CLEAR_NODE(&object
->objlist_link
);
326 object
->oob_event_mask
= 0;
327 for (t
= object
->oob_table
; t
->events
; t
++)
328 object
->oob_event_mask
|= t
->events
;
329 object
->event_mask
= object
->oob_event_mask
;
330 for (t
= object
->state
->transitions
; t
->events
; t
++)
331 object
->event_mask
|= t
->events
;
333 EXPORT_SYMBOL(fscache_object_init
);
336 * Mark the object as no longer being live, making sure that we synchronise
337 * against op submission.
339 static inline void fscache_mark_object_dead(struct fscache_object
*object
)
341 spin_lock(&object
->lock
);
342 clear_bit(FSCACHE_OBJECT_IS_LIVE
, &object
->flags
);
343 spin_unlock(&object
->lock
);
347 * Abort object initialisation before we start it.
349 static const struct fscache_state
*fscache_abort_initialisation(struct fscache_object
*object
,
352 _enter("{OBJ%x},%d", object
->debug_id
, event
);
354 object
->oob_event_mask
= 0;
355 fscache_dequeue_object(object
);
356 return transit_to(KILL_OBJECT
);
360 * initialise an object
361 * - check the specified object's parent to see if we can make use of it
362 * immediately to do a creation
363 * - we may need to start the process of creating a parent and we need to wait
364 * for the parent's lookup and creation to complete if it's not there yet
366 static const struct fscache_state
*fscache_initialise_object(struct fscache_object
*object
,
369 struct fscache_object
*parent
;
372 _enter("{OBJ%x},%d", object
->debug_id
, event
);
374 ASSERT(list_empty(&object
->dep_link
));
376 parent
= object
->parent
;
378 _leave(" [no parent]");
379 return transit_to(DROP_OBJECT
);
382 _debug("parent: %s of:%lx", parent
->state
->name
, parent
->flags
);
384 if (fscache_object_is_dying(parent
)) {
385 _leave(" [bad parent]");
386 return transit_to(DROP_OBJECT
);
389 if (fscache_object_is_available(parent
)) {
391 return transit_to(PARENT_READY
);
396 spin_lock(&parent
->lock
);
397 fscache_stat(&fscache_n_cop_grab_object
);
399 if (fscache_object_is_live(parent
) &&
400 object
->cache
->ops
->grab_object(object
)) {
401 list_add(&object
->dep_link
, &parent
->dependents
);
404 fscache_stat_d(&fscache_n_cop_grab_object
);
405 spin_unlock(&parent
->lock
);
407 _leave(" [grab failed]");
408 return transit_to(DROP_OBJECT
);
411 /* fscache_acquire_non_index_cookie() uses this
412 * to wake the chain up */
413 fscache_raise_event(parent
, FSCACHE_OBJECT_EV_NEW_CHILD
);
415 return transit_to(WAIT_FOR_PARENT
);
419 * Once the parent object is ready, we should kick off our lookup op.
421 static const struct fscache_state
*fscache_parent_ready(struct fscache_object
*object
,
424 struct fscache_object
*parent
= object
->parent
;
426 _enter("{OBJ%x},%d", object
->debug_id
, event
);
428 ASSERT(parent
!= NULL
);
430 spin_lock(&parent
->lock
);
433 object
->lookup_jif
= jiffies
;
434 spin_unlock(&parent
->lock
);
437 return transit_to(LOOK_UP_OBJECT
);
441 * look an object up in the cache from which it was allocated
442 * - we hold an "access lock" on the parent object, so the parent object cannot
443 * be withdrawn by either party till we've finished
445 static const struct fscache_state
*fscache_look_up_object(struct fscache_object
*object
,
448 struct fscache_cookie
*cookie
= object
->cookie
;
449 struct fscache_object
*parent
= object
->parent
;
452 _enter("{OBJ%x},%d", object
->debug_id
, event
);
454 object
->oob_table
= fscache_osm_lookup_oob
;
456 ASSERT(parent
!= NULL
);
457 ASSERTCMP(parent
->n_ops
, >, 0);
458 ASSERTCMP(parent
->n_obj_ops
, >, 0);
460 /* make sure the parent is still available */
461 ASSERT(fscache_object_is_available(parent
));
463 if (fscache_object_is_dying(parent
) ||
464 test_bit(FSCACHE_IOERROR
, &object
->cache
->flags
) ||
465 !fscache_use_cookie(object
)) {
466 _leave(" [unavailable]");
467 return transit_to(LOOKUP_FAILURE
);
470 _debug("LOOKUP \"%s\" in \"%s\"",
471 cookie
->def
->name
, object
->cache
->tag
->name
);
473 fscache_stat(&fscache_n_object_lookups
);
474 fscache_stat(&fscache_n_cop_lookup_object
);
475 ret
= object
->cache
->ops
->lookup_object(object
);
476 fscache_stat_d(&fscache_n_cop_lookup_object
);
478 fscache_unuse_cookie(object
);
480 if (ret
== -ETIMEDOUT
) {
481 /* probably stuck behind another object, so move this one to
482 * the back of the queue */
483 fscache_stat(&fscache_n_object_lookups_timed_out
);
484 _leave(" [timeout]");
490 return transit_to(LOOKUP_FAILURE
);
494 return transit_to(OBJECT_AVAILABLE
);
498 * fscache_object_lookup_negative - Note negative cookie lookup
499 * @object: Object pointing to cookie to mark
501 * Note negative lookup, permitting those waiting to read data from an already
502 * existing backing object to continue as there's no data for them to read.
504 void fscache_object_lookup_negative(struct fscache_object
*object
)
506 struct fscache_cookie
*cookie
= object
->cookie
;
508 _enter("{OBJ%x,%s}", object
->debug_id
, object
->state
->name
);
510 if (!test_and_set_bit(FSCACHE_OBJECT_IS_LOOKED_UP
, &object
->flags
)) {
511 fscache_stat(&fscache_n_object_lookups_negative
);
513 /* Allow write requests to begin stacking up and read requests to begin
516 set_bit(FSCACHE_COOKIE_NO_DATA_YET
, &cookie
->flags
);
517 clear_bit(FSCACHE_COOKIE_UNAVAILABLE
, &cookie
->flags
);
519 _debug("wake up lookup %p", &cookie
->flags
);
520 clear_bit_unlock(FSCACHE_COOKIE_LOOKING_UP
, &cookie
->flags
);
521 wake_up_bit(&cookie
->flags
, FSCACHE_COOKIE_LOOKING_UP
);
525 EXPORT_SYMBOL(fscache_object_lookup_negative
);
528 * fscache_obtained_object - Note successful object lookup or creation
529 * @object: Object pointing to cookie to mark
531 * Note successful lookup and/or creation, permitting those waiting to write
532 * data to a backing object to continue.
534 * Note that after calling this, an object's cookie may be relinquished by the
535 * netfs, and so must be accessed with object lock held.
537 void fscache_obtained_object(struct fscache_object
*object
)
539 struct fscache_cookie
*cookie
= object
->cookie
;
541 _enter("{OBJ%x,%s}", object
->debug_id
, object
->state
->name
);
543 /* if we were still looking up, then we must have a positive lookup
544 * result, in which case there may be data available */
545 if (!test_and_set_bit(FSCACHE_OBJECT_IS_LOOKED_UP
, &object
->flags
)) {
546 fscache_stat(&fscache_n_object_lookups_positive
);
548 /* We do (presumably) have data */
549 clear_bit_unlock(FSCACHE_COOKIE_NO_DATA_YET
, &cookie
->flags
);
550 clear_bit(FSCACHE_COOKIE_UNAVAILABLE
, &cookie
->flags
);
552 /* Allow write requests to begin stacking up and read requests
553 * to begin shovelling data.
555 clear_bit_unlock(FSCACHE_COOKIE_LOOKING_UP
, &cookie
->flags
);
556 wake_up_bit(&cookie
->flags
, FSCACHE_COOKIE_LOOKING_UP
);
558 fscache_stat(&fscache_n_object_created
);
561 set_bit(FSCACHE_OBJECT_IS_AVAILABLE
, &object
->flags
);
564 EXPORT_SYMBOL(fscache_obtained_object
);
567 * handle an object that has just become available
569 static const struct fscache_state
*fscache_object_available(struct fscache_object
*object
,
572 _enter("{OBJ%x},%d", object
->debug_id
, event
);
574 object
->oob_table
= fscache_osm_run_oob
;
576 spin_lock(&object
->lock
);
578 fscache_done_parent_op(object
);
579 if (object
->n_in_progress
== 0) {
580 if (object
->n_ops
> 0) {
581 ASSERTCMP(object
->n_ops
, >=, object
->n_obj_ops
);
582 fscache_start_operations(object
);
584 ASSERT(list_empty(&object
->pending_ops
));
587 spin_unlock(&object
->lock
);
589 fscache_stat(&fscache_n_cop_lookup_complete
);
590 object
->cache
->ops
->lookup_complete(object
);
591 fscache_stat_d(&fscache_n_cop_lookup_complete
);
593 fscache_hist(fscache_obj_instantiate_histogram
, object
->lookup_jif
);
594 fscache_stat(&fscache_n_object_avail
);
597 return transit_to(JUMPSTART_DEPS
);
601 * Wake up this object's dependent objects now that we've become available.
603 static const struct fscache_state
*fscache_jumpstart_dependents(struct fscache_object
*object
,
606 _enter("{OBJ%x},%d", object
->debug_id
, event
);
608 if (!fscache_enqueue_dependents(object
, FSCACHE_OBJECT_EV_PARENT_READY
))
609 return NO_TRANSIT
; /* Not finished; requeue */
610 return transit_to(WAIT_FOR_CMD
);
614 * Handle lookup or creation failute.
616 static const struct fscache_state
*fscache_lookup_failure(struct fscache_object
*object
,
619 struct fscache_cookie
*cookie
;
621 _enter("{OBJ%x},%d", object
->debug_id
, event
);
623 object
->oob_event_mask
= 0;
625 fscache_stat(&fscache_n_cop_lookup_complete
);
626 object
->cache
->ops
->lookup_complete(object
);
627 fscache_stat_d(&fscache_n_cop_lookup_complete
);
629 set_bit(FSCACHE_OBJECT_KILLED_BY_CACHE
, &object
->flags
);
631 cookie
= object
->cookie
;
632 set_bit(FSCACHE_COOKIE_UNAVAILABLE
, &cookie
->flags
);
633 if (test_and_clear_bit(FSCACHE_COOKIE_LOOKING_UP
, &cookie
->flags
))
634 wake_up_bit(&cookie
->flags
, FSCACHE_COOKIE_LOOKING_UP
);
636 fscache_done_parent_op(object
);
637 return transit_to(KILL_OBJECT
);
641 * Wait for completion of all active operations on this object and the death of
642 * all child objects of this object.
644 static const struct fscache_state
*fscache_kill_object(struct fscache_object
*object
,
647 _enter("{OBJ%x,%d,%d},%d",
648 object
->debug_id
, object
->n_ops
, object
->n_children
, event
);
650 fscache_mark_object_dead(object
);
651 object
->oob_event_mask
= 0;
653 if (test_bit(FSCACHE_OBJECT_RETIRED
, &object
->flags
)) {
654 /* Reject any new read/write ops and abort any that are pending. */
655 clear_bit(FSCACHE_OBJECT_PENDING_WRITE
, &object
->flags
);
656 fscache_cancel_all_ops(object
);
659 if (list_empty(&object
->dependents
) &&
660 object
->n_ops
== 0 &&
661 object
->n_children
== 0)
662 return transit_to(DROP_OBJECT
);
664 if (object
->n_in_progress
== 0) {
665 spin_lock(&object
->lock
);
666 if (object
->n_ops
> 0 && object
->n_in_progress
== 0)
667 fscache_start_operations(object
);
668 spin_unlock(&object
->lock
);
671 if (!list_empty(&object
->dependents
))
672 return transit_to(KILL_DEPENDENTS
);
674 return transit_to(WAIT_FOR_CLEARANCE
);
678 * Kill dependent objects.
680 static const struct fscache_state
*fscache_kill_dependents(struct fscache_object
*object
,
683 _enter("{OBJ%x},%d", object
->debug_id
, event
);
685 if (!fscache_enqueue_dependents(object
, FSCACHE_OBJECT_EV_KILL
))
686 return NO_TRANSIT
; /* Not finished */
687 return transit_to(WAIT_FOR_CLEARANCE
);
691 * Drop an object's attachments
693 static const struct fscache_state
*fscache_drop_object(struct fscache_object
*object
,
696 struct fscache_object
*parent
= object
->parent
;
697 struct fscache_cookie
*cookie
= object
->cookie
;
698 struct fscache_cache
*cache
= object
->cache
;
701 _enter("{OBJ%x,%d},%d", object
->debug_id
, object
->n_children
, event
);
703 ASSERT(cookie
!= NULL
);
704 ASSERT(!hlist_unhashed(&object
->cookie_link
));
706 /* Make sure the cookie no longer points here and that the netfs isn't
709 spin_lock(&cookie
->lock
);
710 hlist_del_init(&object
->cookie_link
);
711 if (hlist_empty(&cookie
->backing_objects
) &&
712 test_and_clear_bit(FSCACHE_COOKIE_INVALIDATING
, &cookie
->flags
))
714 spin_unlock(&cookie
->lock
);
717 wake_up_bit(&cookie
->flags
, FSCACHE_COOKIE_INVALIDATING
);
718 if (test_and_clear_bit(FSCACHE_COOKIE_LOOKING_UP
, &cookie
->flags
))
719 wake_up_bit(&cookie
->flags
, FSCACHE_COOKIE_LOOKING_UP
);
722 /* Prevent a race with our last child, which has to signal EV_CLEARED
723 * before dropping our spinlock.
725 spin_lock(&object
->lock
);
726 spin_unlock(&object
->lock
);
728 /* Discard from the cache's collection of objects */
729 spin_lock(&cache
->object_list_lock
);
730 list_del_init(&object
->cache_link
);
731 spin_unlock(&cache
->object_list_lock
);
733 fscache_stat(&fscache_n_cop_drop_object
);
734 cache
->ops
->drop_object(object
);
735 fscache_stat_d(&fscache_n_cop_drop_object
);
737 /* The parent object wants to know when all it dependents have gone */
739 _debug("release parent OBJ%x {%d}",
740 parent
->debug_id
, parent
->n_children
);
742 spin_lock(&parent
->lock
);
743 parent
->n_children
--;
744 if (parent
->n_children
== 0)
745 fscache_raise_event(parent
, FSCACHE_OBJECT_EV_CLEARED
);
746 spin_unlock(&parent
->lock
);
747 object
->parent
= NULL
;
750 /* this just shifts the object release to the work processor */
751 fscache_put_object(object
);
752 fscache_stat(&fscache_n_object_dead
);
755 return transit_to(OBJECT_DEAD
);
759 * get a ref on an object
761 static int fscache_get_object(struct fscache_object
*object
)
765 fscache_stat(&fscache_n_cop_grab_object
);
766 ret
= object
->cache
->ops
->grab_object(object
) ? 0 : -EAGAIN
;
767 fscache_stat_d(&fscache_n_cop_grab_object
);
772 * Discard a ref on an object
774 static void fscache_put_object(struct fscache_object
*object
)
776 fscache_stat(&fscache_n_cop_put_object
);
777 object
->cache
->ops
->put_object(object
);
778 fscache_stat_d(&fscache_n_cop_put_object
);
782 * fscache_object_destroy - Note that a cache object is about to be destroyed
783 * @object: The object to be destroyed
785 * Note the imminent destruction and deallocation of a cache object record.
787 void fscache_object_destroy(struct fscache_object
*object
)
789 fscache_objlist_remove(object
);
791 /* We can get rid of the cookie now */
792 fscache_cookie_put(object
->cookie
);
793 object
->cookie
= NULL
;
795 EXPORT_SYMBOL(fscache_object_destroy
);
798 * enqueue an object for metadata-type processing
800 void fscache_enqueue_object(struct fscache_object
*object
)
802 _enter("{OBJ%x}", object
->debug_id
);
804 if (fscache_get_object(object
) >= 0) {
805 wait_queue_head_t
*cong_wq
=
806 &get_cpu_var(fscache_object_cong_wait
);
808 if (queue_work(fscache_object_wq
, &object
->work
)) {
809 if (fscache_object_congested())
812 fscache_put_object(object
);
814 put_cpu_var(fscache_object_cong_wait
);
819 * fscache_object_sleep_till_congested - Sleep until object wq is congested
820 * @timeoutp: Scheduler sleep timeout
822 * Allow an object handler to sleep until the object workqueue is congested.
824 * The caller must set up a wake up event before calling this and must have set
825 * the appropriate sleep mode (such as TASK_UNINTERRUPTIBLE) and tested its own
826 * condition before calling this function as no test is made here.
828 * %true is returned if the object wq is congested, %false otherwise.
830 bool fscache_object_sleep_till_congested(signed long *timeoutp
)
832 wait_queue_head_t
*cong_wq
= this_cpu_ptr(&fscache_object_cong_wait
);
835 if (fscache_object_congested())
838 add_wait_queue_exclusive(cong_wq
, &wait
);
839 if (!fscache_object_congested())
840 *timeoutp
= schedule_timeout(*timeoutp
);
841 finish_wait(cong_wq
, &wait
);
843 return fscache_object_congested();
845 EXPORT_SYMBOL_GPL(fscache_object_sleep_till_congested
);
848 * Enqueue the dependents of an object for metadata-type processing.
850 * If we don't manage to finish the list before the scheduler wants to run
851 * again then return false immediately. We return true if the list was
854 static bool fscache_enqueue_dependents(struct fscache_object
*object
, int event
)
856 struct fscache_object
*dep
;
859 _enter("{OBJ%x}", object
->debug_id
);
861 if (list_empty(&object
->dependents
))
864 spin_lock(&object
->lock
);
866 while (!list_empty(&object
->dependents
)) {
867 dep
= list_entry(object
->dependents
.next
,
868 struct fscache_object
, dep_link
);
869 list_del_init(&dep
->dep_link
);
871 fscache_raise_event(dep
, event
);
872 fscache_put_object(dep
);
874 if (!list_empty(&object
->dependents
) && need_resched()) {
880 spin_unlock(&object
->lock
);
885 * remove an object from whatever queue it's waiting on
887 static void fscache_dequeue_object(struct fscache_object
*object
)
889 _enter("{OBJ%x}", object
->debug_id
);
891 if (!list_empty(&object
->dep_link
)) {
892 spin_lock(&object
->parent
->lock
);
893 list_del_init(&object
->dep_link
);
894 spin_unlock(&object
->parent
->lock
);
901 * fscache_check_aux - Ask the netfs whether an object on disk is still valid
902 * @object: The object to ask about
903 * @data: The auxiliary data for the object
904 * @datalen: The size of the auxiliary data
906 * This function consults the netfs about the coherency state of an object.
907 * The caller must be holding a ref on cookie->n_active (held by
908 * fscache_look_up_object() on behalf of the cache backend during object lookup
911 enum fscache_checkaux
fscache_check_aux(struct fscache_object
*object
,
912 const void *data
, uint16_t datalen
)
914 enum fscache_checkaux result
;
916 if (!object
->cookie
->def
->check_aux
) {
917 fscache_stat(&fscache_n_checkaux_none
);
918 return FSCACHE_CHECKAUX_OKAY
;
921 result
= object
->cookie
->def
->check_aux(object
->cookie
->netfs_data
,
924 /* entry okay as is */
925 case FSCACHE_CHECKAUX_OKAY
:
926 fscache_stat(&fscache_n_checkaux_okay
);
929 /* entry requires update */
930 case FSCACHE_CHECKAUX_NEEDS_UPDATE
:
931 fscache_stat(&fscache_n_checkaux_update
);
934 /* entry requires deletion */
935 case FSCACHE_CHECKAUX_OBSOLETE
:
936 fscache_stat(&fscache_n_checkaux_obsolete
);
945 EXPORT_SYMBOL(fscache_check_aux
);
948 * Asynchronously invalidate an object.
950 static const struct fscache_state
*_fscache_invalidate_object(struct fscache_object
*object
,
953 struct fscache_operation
*op
;
954 struct fscache_cookie
*cookie
= object
->cookie
;
956 _enter("{OBJ%x},%d", object
->debug_id
, event
);
958 /* We're going to need the cookie. If the cookie is not available then
959 * retire the object instead.
961 if (!fscache_use_cookie(object
)) {
962 ASSERT(object
->cookie
->stores
.rnode
== NULL
);
963 set_bit(FSCACHE_OBJECT_RETIRED
, &object
->flags
);
964 _leave(" [no cookie]");
965 return transit_to(KILL_OBJECT
);
968 /* Reject any new read/write ops and abort any that are pending. */
969 fscache_invalidate_writes(cookie
);
970 clear_bit(FSCACHE_OBJECT_PENDING_WRITE
, &object
->flags
);
971 fscache_cancel_all_ops(object
);
973 /* Now we have to wait for in-progress reads and writes */
974 op
= kzalloc(sizeof(*op
), GFP_KERNEL
);
978 fscache_operation_init(op
, object
->cache
->ops
->invalidate_object
,
980 op
->flags
= FSCACHE_OP_ASYNC
|
981 (1 << FSCACHE_OP_EXCLUSIVE
) |
982 (1 << FSCACHE_OP_UNUSE_COOKIE
);
984 spin_lock(&cookie
->lock
);
985 if (fscache_submit_exclusive_op(object
, op
) < 0)
986 goto submit_op_failed
;
987 spin_unlock(&cookie
->lock
);
988 fscache_put_operation(op
);
990 /* Once we've completed the invalidation, we know there will be no data
991 * stored in the cache and thus we can reinstate the data-check-skip
994 set_bit(FSCACHE_COOKIE_NO_DATA_YET
, &cookie
->flags
);
996 /* We can allow read and write requests to come in once again. They'll
997 * queue up behind our exclusive invalidation operation.
999 if (test_and_clear_bit(FSCACHE_COOKIE_INVALIDATING
, &cookie
->flags
))
1000 wake_up_bit(&cookie
->flags
, FSCACHE_COOKIE_INVALIDATING
);
1002 return transit_to(UPDATE_OBJECT
);
1005 fscache_mark_object_dead(object
);
1006 fscache_unuse_cookie(object
);
1007 _leave(" [ENOMEM]");
1008 return transit_to(KILL_OBJECT
);
1011 fscache_mark_object_dead(object
);
1012 spin_unlock(&cookie
->lock
);
1013 fscache_unuse_cookie(object
);
1016 return transit_to(KILL_OBJECT
);
1019 static const struct fscache_state
*fscache_invalidate_object(struct fscache_object
*object
,
1022 const struct fscache_state
*s
;
1024 fscache_stat(&fscache_n_invalidates_run
);
1025 fscache_stat(&fscache_n_cop_invalidate_object
);
1026 s
= _fscache_invalidate_object(object
, event
);
1027 fscache_stat_d(&fscache_n_cop_invalidate_object
);
1032 * Asynchronously update an object.
1034 static const struct fscache_state
*fscache_update_object(struct fscache_object
*object
,
1037 _enter("{OBJ%x},%d", object
->debug_id
, event
);
1039 fscache_stat(&fscache_n_updates_run
);
1040 fscache_stat(&fscache_n_cop_update_object
);
1041 object
->cache
->ops
->update_object(object
);
1042 fscache_stat_d(&fscache_n_cop_update_object
);
1045 return transit_to(WAIT_FOR_CMD
);
1049 * fscache_object_retrying_stale - Note retrying stale object
1050 * @object: The object that will be retried
1052 * Note that an object lookup found an on-disk object that was adjudged to be
1053 * stale and has been deleted. The lookup will be retried.
1055 void fscache_object_retrying_stale(struct fscache_object
*object
)
1057 fscache_stat(&fscache_n_cache_no_space_reject
);
1059 EXPORT_SYMBOL(fscache_object_retrying_stale
);
1062 * fscache_object_mark_killed - Note that an object was killed
1063 * @object: The object that was culled
1064 * @why: The reason the object was killed.
1066 * Note that an object was killed. Returns true if the object was
1067 * already marked killed, false if it wasn't.
1069 void fscache_object_mark_killed(struct fscache_object
*object
,
1070 enum fscache_why_object_killed why
)
1072 if (test_and_set_bit(FSCACHE_OBJECT_KILLED_BY_CACHE
, &object
->flags
)) {
1073 pr_err("Error: Object already killed by cache [%s]\n",
1074 object
->cache
->identifier
);
1079 case FSCACHE_OBJECT_NO_SPACE
:
1080 fscache_stat(&fscache_n_cache_no_space_reject
);
1082 case FSCACHE_OBJECT_IS_STALE
:
1083 fscache_stat(&fscache_n_cache_stale_objects
);
1085 case FSCACHE_OBJECT_WAS_RETIRED
:
1086 fscache_stat(&fscache_n_cache_retired_objects
);
1088 case FSCACHE_OBJECT_WAS_CULLED
:
1089 fscache_stat(&fscache_n_cache_culled_objects
);
1093 EXPORT_SYMBOL(fscache_object_mark_killed
);
1096 * The object is dead. We can get here if an object gets queued by an event
1097 * that would lead to its death (such as EV_KILL) when the dispatcher is
1098 * already running (and so can be requeued) but hasn't yet cleared the event
1101 static const struct fscache_state
*fscache_object_dead(struct fscache_object
*object
,
1104 if (!test_and_set_bit(FSCACHE_OBJECT_RUN_AFTER_DEAD
,
1108 WARN(true, "FS-Cache object redispatched after death");