1 /* FS-Cache worker operation management routines
3 * Copyright (C) 2008 Red Hat, Inc. All Rights Reserved.
4 * Written by David Howells (dhowells@redhat.com)
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
11 * See Documentation/filesystems/caching/operations.txt
14 #define FSCACHE_DEBUG_LEVEL OPERATION
15 #include <linux/module.h>
16 #include <linux/seq_file.h>
17 #include <linux/slab.h>
20 atomic_t fscache_op_debug_id
;
21 EXPORT_SYMBOL(fscache_op_debug_id
);
23 static void fscache_operation_dummy_cancel(struct fscache_operation
*op
)
28 * fscache_operation_init - Do basic initialisation of an operation
29 * @op: The operation to initialise
30 * @release: The release function to assign
32 * Do basic initialisation of an operation. The caller must still set flags,
33 * object and processor if needed.
35 void fscache_operation_init(struct fscache_operation
*op
,
36 fscache_operation_processor_t processor
,
37 fscache_operation_cancel_t cancel
,
38 fscache_operation_release_t release
)
40 INIT_WORK(&op
->work
, fscache_op_work_func
);
41 atomic_set(&op
->usage
, 1);
42 op
->state
= FSCACHE_OP_ST_INITIALISED
;
43 op
->debug_id
= atomic_inc_return(&fscache_op_debug_id
);
44 op
->processor
= processor
;
45 op
->cancel
= cancel
?: fscache_operation_dummy_cancel
;
46 op
->release
= release
;
47 INIT_LIST_HEAD(&op
->pend_link
);
48 fscache_stat(&fscache_n_op_initialised
);
50 EXPORT_SYMBOL(fscache_operation_init
);
53 * fscache_enqueue_operation - Enqueue an operation for processing
54 * @op: The operation to enqueue
56 * Enqueue an operation for processing by the FS-Cache thread pool.
58 * This will get its own ref on the object.
60 void fscache_enqueue_operation(struct fscache_operation
*op
)
62 _enter("{OBJ%x OP%x,%u}",
63 op
->object
->debug_id
, op
->debug_id
, atomic_read(&op
->usage
));
65 ASSERT(list_empty(&op
->pend_link
));
66 ASSERT(op
->processor
!= NULL
);
67 ASSERT(fscache_object_is_available(op
->object
));
68 ASSERTCMP(atomic_read(&op
->usage
), >, 0);
69 ASSERTCMP(op
->state
, ==, FSCACHE_OP_ST_IN_PROGRESS
);
71 fscache_stat(&fscache_n_op_enqueue
);
72 switch (op
->flags
& FSCACHE_OP_TYPE
) {
73 case FSCACHE_OP_ASYNC
:
74 _debug("queue async");
75 atomic_inc(&op
->usage
);
76 if (!queue_work(fscache_op_wq
, &op
->work
))
77 fscache_put_operation(op
);
79 case FSCACHE_OP_MYTHREAD
:
80 _debug("queue for caller's attention");
83 pr_err("Unexpected op type %lx", op
->flags
);
88 EXPORT_SYMBOL(fscache_enqueue_operation
);
93 static void fscache_run_op(struct fscache_object
*object
,
94 struct fscache_operation
*op
)
96 ASSERTCMP(op
->state
, ==, FSCACHE_OP_ST_PENDING
);
98 op
->state
= FSCACHE_OP_ST_IN_PROGRESS
;
99 object
->n_in_progress
++;
100 if (test_and_clear_bit(FSCACHE_OP_WAITING
, &op
->flags
))
101 wake_up_bit(&op
->flags
, FSCACHE_OP_WAITING
);
103 fscache_enqueue_operation(op
);
104 fscache_stat(&fscache_n_op_run
);
108 * report an unexpected submission
110 static void fscache_report_unexpected_submission(struct fscache_object
*object
,
111 struct fscache_operation
*op
,
112 const struct fscache_state
*ostate
)
114 static bool once_only
;
115 struct fscache_operation
*p
;
122 kdebug("unexpected submission OP%x [OBJ%x %s]",
123 op
->debug_id
, object
->debug_id
, object
->state
->name
);
124 kdebug("objstate=%s [%s]", object
->state
->name
, ostate
->name
);
125 kdebug("objflags=%lx", object
->flags
);
126 kdebug("objevent=%lx [%lx]", object
->events
, object
->event_mask
);
127 kdebug("ops=%u inp=%u exc=%u",
128 object
->n_ops
, object
->n_in_progress
, object
->n_exclusive
);
130 if (!list_empty(&object
->pending_ops
)) {
132 list_for_each_entry(p
, &object
->pending_ops
, pend_link
) {
133 ASSERTCMP(p
->object
, ==, object
);
134 kdebug("%p %p", op
->processor
, op
->release
);
145 * submit an exclusive operation for an object
146 * - other ops are excluded from running simultaneously with this one
147 * - this gets any extra refs it needs on an op
149 int fscache_submit_exclusive_op(struct fscache_object
*object
,
150 struct fscache_operation
*op
)
152 const struct fscache_state
*ostate
;
156 _enter("{OBJ%x OP%x},", object
->debug_id
, op
->debug_id
);
158 ASSERTCMP(op
->state
, ==, FSCACHE_OP_ST_INITIALISED
);
159 ASSERTCMP(atomic_read(&op
->usage
), >, 0);
161 spin_lock(&object
->lock
);
162 ASSERTCMP(object
->n_ops
, >=, object
->n_in_progress
);
163 ASSERTCMP(object
->n_ops
, >=, object
->n_exclusive
);
164 ASSERT(list_empty(&op
->pend_link
));
166 ostate
= object
->state
;
169 op
->state
= FSCACHE_OP_ST_PENDING
;
170 flags
= READ_ONCE(object
->flags
);
171 if (unlikely(!(flags
& BIT(FSCACHE_OBJECT_IS_LIVE
)))) {
172 fscache_stat(&fscache_n_op_rejected
);
174 op
->state
= FSCACHE_OP_ST_CANCELLED
;
176 } else if (unlikely(fscache_cache_is_broken(object
))) {
178 op
->state
= FSCACHE_OP_ST_CANCELLED
;
180 } else if (flags
& BIT(FSCACHE_OBJECT_IS_AVAILABLE
)) {
183 object
->n_exclusive
++; /* reads and writes must wait */
185 if (object
->n_in_progress
> 0) {
186 atomic_inc(&op
->usage
);
187 list_add_tail(&op
->pend_link
, &object
->pending_ops
);
188 fscache_stat(&fscache_n_op_pend
);
189 } else if (!list_empty(&object
->pending_ops
)) {
190 atomic_inc(&op
->usage
);
191 list_add_tail(&op
->pend_link
, &object
->pending_ops
);
192 fscache_stat(&fscache_n_op_pend
);
193 fscache_start_operations(object
);
195 ASSERTCMP(object
->n_in_progress
, ==, 0);
196 fscache_run_op(object
, op
);
199 /* need to issue a new write op after this */
200 clear_bit(FSCACHE_OBJECT_PENDING_WRITE
, &object
->flags
);
202 } else if (flags
& BIT(FSCACHE_OBJECT_IS_LOOKED_UP
)) {
205 object
->n_exclusive
++; /* reads and writes must wait */
206 atomic_inc(&op
->usage
);
207 list_add_tail(&op
->pend_link
, &object
->pending_ops
);
208 fscache_stat(&fscache_n_op_pend
);
210 } else if (flags
& BIT(FSCACHE_OBJECT_KILLED_BY_CACHE
)) {
212 op
->state
= FSCACHE_OP_ST_CANCELLED
;
215 fscache_report_unexpected_submission(object
, op
, ostate
);
217 op
->state
= FSCACHE_OP_ST_CANCELLED
;
221 spin_unlock(&object
->lock
);
226 * submit an operation for an object
227 * - objects may be submitted only in the following states:
228 * - during object creation (write ops may be submitted)
229 * - whilst the object is active
230 * - after an I/O error incurred in one of the two above states (op rejected)
231 * - this gets any extra refs it needs on an op
233 int fscache_submit_op(struct fscache_object
*object
,
234 struct fscache_operation
*op
)
236 const struct fscache_state
*ostate
;
240 _enter("{OBJ%x OP%x},{%u}",
241 object
->debug_id
, op
->debug_id
, atomic_read(&op
->usage
));
243 ASSERTCMP(op
->state
, ==, FSCACHE_OP_ST_INITIALISED
);
244 ASSERTCMP(atomic_read(&op
->usage
), >, 0);
246 spin_lock(&object
->lock
);
247 ASSERTCMP(object
->n_ops
, >=, object
->n_in_progress
);
248 ASSERTCMP(object
->n_ops
, >=, object
->n_exclusive
);
249 ASSERT(list_empty(&op
->pend_link
));
251 ostate
= object
->state
;
254 op
->state
= FSCACHE_OP_ST_PENDING
;
255 flags
= READ_ONCE(object
->flags
);
256 if (unlikely(!(flags
& BIT(FSCACHE_OBJECT_IS_LIVE
)))) {
257 fscache_stat(&fscache_n_op_rejected
);
259 op
->state
= FSCACHE_OP_ST_CANCELLED
;
261 } else if (unlikely(fscache_cache_is_broken(object
))) {
263 op
->state
= FSCACHE_OP_ST_CANCELLED
;
265 } else if (flags
& BIT(FSCACHE_OBJECT_IS_AVAILABLE
)) {
269 if (object
->n_exclusive
> 0) {
270 atomic_inc(&op
->usage
);
271 list_add_tail(&op
->pend_link
, &object
->pending_ops
);
272 fscache_stat(&fscache_n_op_pend
);
273 } else if (!list_empty(&object
->pending_ops
)) {
274 atomic_inc(&op
->usage
);
275 list_add_tail(&op
->pend_link
, &object
->pending_ops
);
276 fscache_stat(&fscache_n_op_pend
);
277 fscache_start_operations(object
);
279 ASSERTCMP(object
->n_exclusive
, ==, 0);
280 fscache_run_op(object
, op
);
283 } else if (flags
& BIT(FSCACHE_OBJECT_IS_LOOKED_UP
)) {
286 atomic_inc(&op
->usage
);
287 list_add_tail(&op
->pend_link
, &object
->pending_ops
);
288 fscache_stat(&fscache_n_op_pend
);
290 } else if (flags
& BIT(FSCACHE_OBJECT_KILLED_BY_CACHE
)) {
292 op
->state
= FSCACHE_OP_ST_CANCELLED
;
295 fscache_report_unexpected_submission(object
, op
, ostate
);
296 ASSERT(!fscache_object_is_active(object
));
298 op
->state
= FSCACHE_OP_ST_CANCELLED
;
302 spin_unlock(&object
->lock
);
307 * queue an object for withdrawal on error, aborting all following asynchronous
310 void fscache_abort_object(struct fscache_object
*object
)
312 _enter("{OBJ%x}", object
->debug_id
);
314 fscache_raise_event(object
, FSCACHE_OBJECT_EV_ERROR
);
318 * Jump start the operation processing on an object. The caller must hold
321 void fscache_start_operations(struct fscache_object
*object
)
323 struct fscache_operation
*op
;
326 while (!list_empty(&object
->pending_ops
) && !stop
) {
327 op
= list_entry(object
->pending_ops
.next
,
328 struct fscache_operation
, pend_link
);
330 if (test_bit(FSCACHE_OP_EXCLUSIVE
, &op
->flags
)) {
331 if (object
->n_in_progress
> 0)
335 list_del_init(&op
->pend_link
);
336 fscache_run_op(object
, op
);
338 /* the pending queue was holding a ref on the object */
339 fscache_put_operation(op
);
342 ASSERTCMP(object
->n_in_progress
, <=, object
->n_ops
);
344 _debug("woke %d ops on OBJ%x",
345 object
->n_in_progress
, object
->debug_id
);
349 * cancel an operation that's pending on an object
351 int fscache_cancel_op(struct fscache_operation
*op
,
352 bool cancel_in_progress_op
)
354 struct fscache_object
*object
= op
->object
;
358 _enter("OBJ%x OP%x}", op
->object
->debug_id
, op
->debug_id
);
360 ASSERTCMP(op
->state
, >=, FSCACHE_OP_ST_PENDING
);
361 ASSERTCMP(op
->state
, !=, FSCACHE_OP_ST_CANCELLED
);
362 ASSERTCMP(atomic_read(&op
->usage
), >, 0);
364 spin_lock(&object
->lock
);
367 if (op
->state
== FSCACHE_OP_ST_PENDING
) {
368 ASSERT(!list_empty(&op
->pend_link
));
369 list_del_init(&op
->pend_link
);
372 fscache_stat(&fscache_n_op_cancelled
);
374 op
->state
= FSCACHE_OP_ST_CANCELLED
;
375 if (test_bit(FSCACHE_OP_EXCLUSIVE
, &op
->flags
))
376 object
->n_exclusive
--;
377 if (test_and_clear_bit(FSCACHE_OP_WAITING
, &op
->flags
))
378 wake_up_bit(&op
->flags
, FSCACHE_OP_WAITING
);
380 } else if (op
->state
== FSCACHE_OP_ST_IN_PROGRESS
&& cancel_in_progress_op
) {
381 ASSERTCMP(object
->n_in_progress
, >, 0);
382 if (test_bit(FSCACHE_OP_EXCLUSIVE
, &op
->flags
))
383 object
->n_exclusive
--;
384 object
->n_in_progress
--;
385 if (object
->n_in_progress
== 0)
386 fscache_start_operations(object
);
388 fscache_stat(&fscache_n_op_cancelled
);
390 op
->state
= FSCACHE_OP_ST_CANCELLED
;
391 if (test_bit(FSCACHE_OP_EXCLUSIVE
, &op
->flags
))
392 object
->n_exclusive
--;
393 if (test_and_clear_bit(FSCACHE_OP_WAITING
, &op
->flags
))
394 wake_up_bit(&op
->flags
, FSCACHE_OP_WAITING
);
399 fscache_put_operation(op
);
400 spin_unlock(&object
->lock
);
401 _leave(" = %d", ret
);
406 * Cancel all pending operations on an object
408 void fscache_cancel_all_ops(struct fscache_object
*object
)
410 struct fscache_operation
*op
;
412 _enter("OBJ%x", object
->debug_id
);
414 spin_lock(&object
->lock
);
416 while (!list_empty(&object
->pending_ops
)) {
417 op
= list_entry(object
->pending_ops
.next
,
418 struct fscache_operation
, pend_link
);
419 fscache_stat(&fscache_n_op_cancelled
);
420 list_del_init(&op
->pend_link
);
422 ASSERTCMP(op
->state
, ==, FSCACHE_OP_ST_PENDING
);
424 op
->state
= FSCACHE_OP_ST_CANCELLED
;
426 if (test_bit(FSCACHE_OP_EXCLUSIVE
, &op
->flags
))
427 object
->n_exclusive
--;
428 if (test_and_clear_bit(FSCACHE_OP_WAITING
, &op
->flags
))
429 wake_up_bit(&op
->flags
, FSCACHE_OP_WAITING
);
430 fscache_put_operation(op
);
431 cond_resched_lock(&object
->lock
);
434 spin_unlock(&object
->lock
);
439 * Record the completion or cancellation of an in-progress operation.
441 void fscache_op_complete(struct fscache_operation
*op
, bool cancelled
)
443 struct fscache_object
*object
= op
->object
;
445 _enter("OBJ%x", object
->debug_id
);
447 ASSERTCMP(op
->state
, ==, FSCACHE_OP_ST_IN_PROGRESS
);
448 ASSERTCMP(object
->n_in_progress
, >, 0);
449 ASSERTIFCMP(test_bit(FSCACHE_OP_EXCLUSIVE
, &op
->flags
),
450 object
->n_exclusive
, >, 0);
451 ASSERTIFCMP(test_bit(FSCACHE_OP_EXCLUSIVE
, &op
->flags
),
452 object
->n_in_progress
, ==, 1);
454 spin_lock(&object
->lock
);
457 op
->state
= FSCACHE_OP_ST_COMPLETE
;
460 op
->state
= FSCACHE_OP_ST_CANCELLED
;
463 if (test_bit(FSCACHE_OP_EXCLUSIVE
, &op
->flags
))
464 object
->n_exclusive
--;
465 object
->n_in_progress
--;
466 if (object
->n_in_progress
== 0)
467 fscache_start_operations(object
);
469 spin_unlock(&object
->lock
);
472 EXPORT_SYMBOL(fscache_op_complete
);
475 * release an operation
476 * - queues pending ops if this is the last in-progress op
478 void fscache_put_operation(struct fscache_operation
*op
)
480 struct fscache_object
*object
;
481 struct fscache_cache
*cache
;
483 _enter("{OBJ%x OP%x,%d}",
484 op
->object
->debug_id
, op
->debug_id
, atomic_read(&op
->usage
));
486 ASSERTCMP(atomic_read(&op
->usage
), >, 0);
488 if (!atomic_dec_and_test(&op
->usage
))
492 ASSERTIFCMP(op
->state
!= FSCACHE_OP_ST_INITIALISED
&&
493 op
->state
!= FSCACHE_OP_ST_COMPLETE
,
494 op
->state
, ==, FSCACHE_OP_ST_CANCELLED
);
496 fscache_stat(&fscache_n_op_release
);
502 op
->state
= FSCACHE_OP_ST_DEAD
;
505 if (likely(object
)) {
506 if (test_bit(FSCACHE_OP_DEC_READ_CNT
, &op
->flags
))
507 atomic_dec(&object
->n_reads
);
508 if (test_bit(FSCACHE_OP_UNUSE_COOKIE
, &op
->flags
))
509 fscache_unuse_cookie(object
);
511 /* now... we may get called with the object spinlock held, so we
512 * complete the cleanup here only if we can immediately acquire the
513 * lock, and defer it otherwise */
514 if (!spin_trylock(&object
->lock
)) {
516 fscache_stat(&fscache_n_op_deferred_release
);
518 cache
= object
->cache
;
519 spin_lock(&cache
->op_gc_list_lock
);
520 list_add_tail(&op
->pend_link
, &cache
->op_gc_list
);
521 spin_unlock(&cache
->op_gc_list_lock
);
522 schedule_work(&cache
->op_gc
);
527 ASSERTCMP(object
->n_ops
, >, 0);
529 if (object
->n_ops
== 0)
530 fscache_raise_event(object
, FSCACHE_OBJECT_EV_CLEARED
);
532 spin_unlock(&object
->lock
);
538 EXPORT_SYMBOL(fscache_put_operation
);
541 * garbage collect operations that have had their release deferred
543 void fscache_operation_gc(struct work_struct
*work
)
545 struct fscache_operation
*op
;
546 struct fscache_object
*object
;
547 struct fscache_cache
*cache
=
548 container_of(work
, struct fscache_cache
, op_gc
);
554 spin_lock(&cache
->op_gc_list_lock
);
555 if (list_empty(&cache
->op_gc_list
)) {
556 spin_unlock(&cache
->op_gc_list_lock
);
560 op
= list_entry(cache
->op_gc_list
.next
,
561 struct fscache_operation
, pend_link
);
562 list_del(&op
->pend_link
);
563 spin_unlock(&cache
->op_gc_list_lock
);
566 spin_lock(&object
->lock
);
568 _debug("GC DEFERRED REL OBJ%x OP%x",
569 object
->debug_id
, op
->debug_id
);
570 fscache_stat(&fscache_n_op_gc
);
572 ASSERTCMP(atomic_read(&op
->usage
), ==, 0);
573 ASSERTCMP(op
->state
, ==, FSCACHE_OP_ST_DEAD
);
575 ASSERTCMP(object
->n_ops
, >, 0);
577 if (object
->n_ops
== 0)
578 fscache_raise_event(object
, FSCACHE_OBJECT_EV_CLEARED
);
580 spin_unlock(&object
->lock
);
583 } while (count
++ < 20);
585 if (!list_empty(&cache
->op_gc_list
))
586 schedule_work(&cache
->op_gc
);
592 * execute an operation using fs_op_wq to provide processing context -
593 * the caller holds a ref to this object, so we don't need to hold one
595 void fscache_op_work_func(struct work_struct
*work
)
597 struct fscache_operation
*op
=
598 container_of(work
, struct fscache_operation
, work
);
601 _enter("{OBJ%x OP%x,%d}",
602 op
->object
->debug_id
, op
->debug_id
, atomic_read(&op
->usage
));
604 ASSERT(op
->processor
!= NULL
);
607 fscache_hist(fscache_ops_histogram
, start
);
608 fscache_put_operation(op
);