1// SPDX-License-Identifier: GPL-2.0-or-later 2/* FS-Cache object state machine handler 3 * 4 * Copyright (C) 2007 Red Hat, Inc. All Rights Reserved. 5 * Written by David Howells (dhowells@redhat.com) 6 * 7 * See Documentation/filesystems/caching/object.rst for a description of the 8 * object state machine and the in-kernel representations. 9 */ 10 11#define FSCACHE_DEBUG_LEVEL COOKIE 12#include <linux/module.h> 13#include <linux/slab.h> 14#include <linux/prefetch.h> 15#include "internal.h" 16 17static const struct fscache_state *fscache_abort_initialisation(struct fscache_object *, int); 18static const struct fscache_state *fscache_kill_dependents(struct fscache_object *, int); 19static const struct fscache_state *fscache_drop_object(struct fscache_object *, int); 20static const struct fscache_state *fscache_initialise_object(struct fscache_object *, int); 21static const struct fscache_state *fscache_invalidate_object(struct fscache_object *, int); 22static const struct fscache_state *fscache_jumpstart_dependents(struct fscache_object *, int); 23static const struct fscache_state *fscache_kill_object(struct fscache_object *, int); 24static const struct fscache_state *fscache_lookup_failure(struct fscache_object *, int); 25static const struct fscache_state *fscache_look_up_object(struct fscache_object *, int); 26static const struct fscache_state *fscache_object_available(struct fscache_object *, int); 27static const struct fscache_state *fscache_parent_ready(struct fscache_object *, int); 28static const struct fscache_state *fscache_update_object(struct fscache_object *, int); 29static const struct fscache_state *fscache_object_dead(struct fscache_object *, int); 30 31#define __STATE_NAME(n) fscache_osm_##n 32#define STATE(n) (&__STATE_NAME(n)) 33 34/* 35 * Define a work state. Work states are execution states. No event processing 36 * is performed by them. The function attached to a work state returns a 37 * pointer indicating the next state to which the state machine should 38 * transition. Returning NO_TRANSIT repeats the current state, but goes back 39 * to the scheduler first. 40 */ 41#define WORK_STATE(n, sn, f) \ 42 const struct fscache_state __STATE_NAME(n) = { \ 43 .name = #n, \ 44 .short_name = sn, \ 45 .work = f \ 46 } 47 48/* 49 * Returns from work states. 50 */ 51#define transit_to(state) ({ prefetch(&STATE(state)->work); STATE(state); }) 52 53#define NO_TRANSIT ((struct fscache_state *)NULL) 54 55/* 56 * Define a wait state. Wait states are event processing states. No execution 57 * is performed by them. Wait states are just tables of "if event X occurs, 58 * clear it and transition to state Y". The dispatcher returns to the 59 * scheduler if none of the events in which the wait state has an interest are 60 * currently pending. 61 */ 62#define WAIT_STATE(n, sn, ...) \ 63 const struct fscache_state __STATE_NAME(n) = { \ 64 .name = #n, \ 65 .short_name = sn, \ 66 .work = NULL, \ 67 .transitions = { __VA_ARGS__, { 0, NULL } } \ 68 } 69 70#define TRANSIT_TO(state, emask) \ 71 { .events = (emask), .transit_to = STATE(state) } 72 73/* 74 * The object state machine. 75 */ 76static WORK_STATE(INIT_OBJECT, "INIT", fscache_initialise_object); 77static WORK_STATE(PARENT_READY, "PRDY", fscache_parent_ready); 78static WORK_STATE(ABORT_INIT, "ABRT", fscache_abort_initialisation); 79static WORK_STATE(LOOK_UP_OBJECT, "LOOK", fscache_look_up_object); 80static WORK_STATE(CREATE_OBJECT, "CRTO", fscache_look_up_object); 81static WORK_STATE(OBJECT_AVAILABLE, "AVBL", fscache_object_available); 82static WORK_STATE(JUMPSTART_DEPS, "JUMP", fscache_jumpstart_dependents); 83 84static WORK_STATE(INVALIDATE_OBJECT, "INVL", fscache_invalidate_object); 85static WORK_STATE(UPDATE_OBJECT, "UPDT", fscache_update_object); 86 87static WORK_STATE(LOOKUP_FAILURE, "LCFL", fscache_lookup_failure); 88static WORK_STATE(KILL_OBJECT, "KILL", fscache_kill_object); 89static WORK_STATE(KILL_DEPENDENTS, "KDEP", fscache_kill_dependents); 90static WORK_STATE(DROP_OBJECT, "DROP", fscache_drop_object); 91static WORK_STATE(OBJECT_DEAD, "DEAD", fscache_object_dead); 92 93static WAIT_STATE(WAIT_FOR_INIT, "?INI", 94 TRANSIT_TO(INIT_OBJECT, 1 << FSCACHE_OBJECT_EV_NEW_CHILD)); 95 96static WAIT_STATE(WAIT_FOR_PARENT, "?PRN", 97 TRANSIT_TO(PARENT_READY, 1 << FSCACHE_OBJECT_EV_PARENT_READY)); 98 99static WAIT_STATE(WAIT_FOR_CMD, "?CMD", 100 TRANSIT_TO(INVALIDATE_OBJECT, 1 << FSCACHE_OBJECT_EV_INVALIDATE), 101 TRANSIT_TO(UPDATE_OBJECT, 1 << FSCACHE_OBJECT_EV_UPDATE), 102 TRANSIT_TO(JUMPSTART_DEPS, 1 << FSCACHE_OBJECT_EV_NEW_CHILD)); 103 104static WAIT_STATE(WAIT_FOR_CLEARANCE, "?CLR", 105 TRANSIT_TO(KILL_OBJECT, 1 << FSCACHE_OBJECT_EV_CLEARED)); 106 107/* 108 * Out-of-band event transition tables. These are for handling unexpected 109 * events, such as an I/O error. If an OOB event occurs, the state machine 110 * clears and disables the event and forces a transition to the nominated work 111 * state (acurrently executing work states will complete first). 112 * 113 * In such a situation, object->state remembers the state the machine should 114 * have been in/gone to and returning NO_TRANSIT returns to that. 115 */ 116static const struct fscache_transition fscache_osm_init_oob[] = { 117 TRANSIT_TO(ABORT_INIT, 118 (1 << FSCACHE_OBJECT_EV_ERROR) | 119 (1 << FSCACHE_OBJECT_EV_KILL)), 120 { 0, NULL } 121}; 122 123static const struct fscache_transition fscache_osm_lookup_oob[] = { 124 TRANSIT_TO(LOOKUP_FAILURE, 125 (1 << FSCACHE_OBJECT_EV_ERROR) | 126 (1 << FSCACHE_OBJECT_EV_KILL)), 127 { 0, NULL } 128}; 129 130static const struct fscache_transition fscache_osm_run_oob[] = { 131 TRANSIT_TO(KILL_OBJECT, 132 (1 << FSCACHE_OBJECT_EV_ERROR) | 133 (1 << FSCACHE_OBJECT_EV_KILL)), 134 { 0, NULL } 135}; 136 137static int fscache_get_object(struct fscache_object *, 138 enum fscache_obj_ref_trace); 139static void fscache_put_object(struct fscache_object *, 140 enum fscache_obj_ref_trace); 141static bool fscache_enqueue_dependents(struct fscache_object *, int); 142static void fscache_dequeue_object(struct fscache_object *); 143static void fscache_update_aux_data(struct fscache_object *); 144 145/* 146 * we need to notify the parent when an op completes that we had outstanding 147 * upon it 148 */ 149static inline void fscache_done_parent_op(struct fscache_object *object) 150{ 151 struct fscache_object *parent = object->parent; 152 153 _enter("OBJ%x {OBJ%x,%x}", 154 object->debug_id, parent->debug_id, parent->n_ops); 155 156 spin_lock_nested(&parent->lock, 1); 157 parent->n_obj_ops--; 158 parent->n_ops--; 159 if (parent->n_ops == 0) 160 fscache_raise_event(parent, FSCACHE_OBJECT_EV_CLEARED); 161 spin_unlock(&parent->lock); 162} 163 164/* 165 * Object state machine dispatcher. 166 */ 167static void fscache_object_sm_dispatcher(struct fscache_object *object) 168{ 169 const struct fscache_transition *t; 170 const struct fscache_state *state, *new_state; 171 unsigned long events, event_mask; 172 bool oob; 173 int event = -1; 174 175 ASSERT(object != NULL); 176 177 _enter("{OBJ%x,%s,%lx}", 178 object->debug_id, object->state->name, object->events); 179 180 event_mask = object->event_mask; 181restart: 182 object->event_mask = 0; /* Mask normal event handling */ 183 state = object->state; 184restart_masked: 185 events = object->events; 186 187 /* Handle any out-of-band events (typically an error) */ 188 if (events & object->oob_event_mask) { 189 _debug("{OBJ%x} oob %lx", 190 object->debug_id, events & object->oob_event_mask); 191 oob = true; 192 for (t = object->oob_table; t->events; t++) { 193 if (events & t->events) { 194 state = t->transit_to; 195 ASSERT(state->work != NULL); 196 event = fls(events & t->events) - 1; 197 __clear_bit(event, &object->oob_event_mask); 198 clear_bit(event, &object->events); 199 goto execute_work_state; 200 } 201 } 202 } 203 oob = false; 204 205 /* Wait states are just transition tables */ 206 if (!state->work) { 207 if (events & event_mask) { 208 for (t = state->transitions; t->events; t++) { 209 if (events & t->events) { 210 new_state = t->transit_to; 211 event = fls(events & t->events) - 1; 212 trace_fscache_osm(object, state, 213 true, false, event); 214 clear_bit(event, &object->events); 215 _debug("{OBJ%x} ev %d: %s -> %s", 216 object->debug_id, event, 217 state->name, new_state->name); 218 object->state = state = new_state; 219 goto execute_work_state; 220 } 221 } 222 223 /* The event mask didn't include all the tabled bits */ 224 BUG(); 225 } 226 /* Randomly woke up */ 227 goto unmask_events; 228 } 229 230execute_work_state: 231 _debug("{OBJ%x} exec %s", object->debug_id, state->name); 232 233 trace_fscache_osm(object, state, false, oob, event); 234 new_state = state->work(object, event); 235 event = -1; 236 if (new_state == NO_TRANSIT) { 237 _debug("{OBJ%x} %s notrans", object->debug_id, state->name); 238 if (unlikely(state == STATE(OBJECT_DEAD))) { 239 _leave(" [dead]"); 240 return; 241 } 242 fscache_enqueue_object(object); 243 event_mask = object->oob_event_mask; 244 goto unmask_events; 245 } 246 247 _debug("{OBJ%x} %s -> %s", 248 object->debug_id, state->name, new_state->name); 249 object->state = state = new_state; 250 251 if (state->work) { 252 if (unlikely(state == STATE(OBJECT_DEAD))) { 253 _leave(" [dead]"); 254 return; 255 } 256 goto restart_masked; 257 } 258 259 /* Transited to wait state */ 260 event_mask = object->oob_event_mask; 261 for (t = state->transitions; t->events; t++) 262 event_mask |= t->events; 263 264unmask_events: 265 object->event_mask = event_mask; 266 smp_mb(); 267 events = object->events; 268 if (events & event_mask) 269 goto restart; 270 _leave(" [msk %lx]", event_mask); 271} 272 273/* 274 * execute an object 275 */ 276static void fscache_object_work_func(struct work_struct *work) 277{ 278 struct fscache_object *object = 279 container_of(work, struct fscache_object, work); 280 unsigned long start; 281 282 _enter("{OBJ%x}", object->debug_id); 283 284 start = jiffies; 285 fscache_object_sm_dispatcher(object); 286 fscache_hist(fscache_objs_histogram, start); 287 fscache_put_object(object, fscache_obj_put_work); 288} 289 290/** 291 * fscache_object_init - Initialise a cache object description 292 * @object: Object description 293 * @cookie: Cookie object will be attached to 294 * @cache: Cache in which backing object will be found 295 * 296 * Initialise a cache object description to its basic values. 297 * 298 * See Documentation/filesystems/caching/backend-api.rst for a complete 299 * description. 300 */ 301void fscache_object_init(struct fscache_object *object, 302 struct fscache_cookie *cookie, 303 struct fscache_cache *cache) 304{ 305 const struct fscache_transition *t; 306 307 atomic_inc(&cache->object_count); 308 309 object->state = STATE(WAIT_FOR_INIT); 310 object->oob_table = fscache_osm_init_oob; 311 object->flags = 1 << FSCACHE_OBJECT_IS_LIVE; 312 spin_lock_init(&object->lock); 313 INIT_LIST_HEAD(&object->cache_link); 314 INIT_HLIST_NODE(&object->cookie_link); 315 INIT_WORK(&object->work, fscache_object_work_func); 316 INIT_LIST_HEAD(&object->dependents); 317 INIT_LIST_HEAD(&object->dep_link); 318 INIT_LIST_HEAD(&object->pending_ops); 319 object->n_children = 0; 320 object->n_ops = object->n_in_progress = object->n_exclusive = 0; 321 object->events = 0; 322 object->store_limit = 0; 323 object->store_limit_l = 0; 324 object->cache = cache; 325 object->cookie = cookie; 326 fscache_cookie_get(cookie, fscache_cookie_get_attach_object); 327 object->parent = NULL; 328#ifdef CONFIG_FSCACHE_OBJECT_LIST 329 RB_CLEAR_NODE(&object->objlist_link); 330#endif 331 332 object->oob_event_mask = 0; 333 for (t = object->oob_table; t->events; t++) 334 object->oob_event_mask |= t->events; 335 object->event_mask = object->oob_event_mask; 336 for (t = object->state->transitions; t->events; t++) 337 object->event_mask |= t->events; 338} 339EXPORT_SYMBOL(fscache_object_init); 340 341/* 342 * Mark the object as no longer being live, making sure that we synchronise 343 * against op submission. 344 */ 345static inline void fscache_mark_object_dead(struct fscache_object *object) 346{ 347 spin_lock(&object->lock); 348 clear_bit(FSCACHE_OBJECT_IS_LIVE, &object->flags); 349 spin_unlock(&object->lock); 350} 351 352/* 353 * Abort object initialisation before we start it. 354 */ 355static const struct fscache_state *fscache_abort_initialisation(struct fscache_object *object, 356 int event) 357{ 358 _enter("{OBJ%x},%d", object->debug_id, event); 359 360 object->oob_event_mask = 0; 361 fscache_dequeue_object(object); 362 return transit_to(KILL_OBJECT); 363} 364 365/* 366 * initialise an object 367 * - check the specified object's parent to see if we can make use of it 368 * immediately to do a creation 369 * - we may need to start the process of creating a parent and we need to wait 370 * for the parent's lookup and creation to complete if it's not there yet 371 */ 372static const struct fscache_state *fscache_initialise_object(struct fscache_object *object, 373 int event) 374{ 375 struct fscache_object *parent; 376 bool success; 377 378 _enter("{OBJ%x},%d", object->debug_id, event); 379 380 ASSERT(list_empty(&object->dep_link)); 381 382 parent = object->parent; 383 if (!parent) { 384 _leave(" [no parent]"); 385 return transit_to(DROP_OBJECT); 386 } 387 388 _debug("parent: %s of:%lx", parent->state->name, parent->flags); 389 390 if (fscache_object_is_dying(parent)) { 391 _leave(" [bad parent]"); 392 return transit_to(DROP_OBJECT); 393 } 394 395 if (fscache_object_is_available(parent)) { 396 _leave(" [ready]"); 397 return transit_to(PARENT_READY); 398 } 399 400 _debug("wait"); 401 402 spin_lock(&parent->lock); 403 fscache_stat(&fscache_n_cop_grab_object); 404 success = false; 405 if (fscache_object_is_live(parent) && 406 object->cache->ops->grab_object(object, fscache_obj_get_add_to_deps)) { 407 list_add(&object->dep_link, &parent->dependents); 408 success = true; 409 } 410 fscache_stat_d(&fscache_n_cop_grab_object); 411 spin_unlock(&parent->lock); 412 if (!success) { 413 _leave(" [grab failed]"); 414 return transit_to(DROP_OBJECT); 415 } 416 417 /* fscache_acquire_non_index_cookie() uses this 418 * to wake the chain up */ 419 fscache_raise_event(parent, FSCACHE_OBJECT_EV_NEW_CHILD); 420 _leave(" [wait]"); 421 return transit_to(WAIT_FOR_PARENT); 422} 423 424/* 425 * Once the parent object is ready, we should kick off our lookup op. 426 */ 427static const struct fscache_state *fscache_parent_ready(struct fscache_object *object, 428 int event) 429{ 430 struct fscache_object *parent = object->parent; 431 432 _enter("{OBJ%x},%d", object->debug_id, event); 433 434 ASSERT(parent != NULL); 435 436 spin_lock(&parent->lock); 437 parent->n_ops++; 438 parent->n_obj_ops++; 439 object->lookup_jif = jiffies; 440 spin_unlock(&parent->lock); 441 442 _leave(""); 443 return transit_to(LOOK_UP_OBJECT); 444} 445 446/* 447 * look an object up in the cache from which it was allocated 448 * - we hold an "access lock" on the parent object, so the parent object cannot 449 * be withdrawn by either party till we've finished 450 */ 451static const struct fscache_state *fscache_look_up_object(struct fscache_object *object, 452 int event) 453{ 454 struct fscache_cookie *cookie = object->cookie; 455 struct fscache_object *parent = object->parent; 456 int ret; 457 458 _enter("{OBJ%x},%d", object->debug_id, event); 459 460 object->oob_table = fscache_osm_lookup_oob; 461 462 ASSERT(parent != NULL); 463 ASSERTCMP(parent->n_ops, >, 0); 464 ASSERTCMP(parent->n_obj_ops, >, 0); 465 466 /* make sure the parent is still available */ 467 ASSERT(fscache_object_is_available(parent)); 468 469 if (fscache_object_is_dying(parent) || 470 test_bit(FSCACHE_IOERROR, &object->cache->flags) || 471 !fscache_use_cookie(object)) { 472 _leave(" [unavailable]"); 473 return transit_to(LOOKUP_FAILURE); 474 } 475 476 _debug("LOOKUP \"%s\" in \"%s\"", 477 cookie->def->name, object->cache->tag->name); 478 479 fscache_stat(&fscache_n_object_lookups); 480 fscache_stat(&fscache_n_cop_lookup_object); 481 ret = object->cache->ops->lookup_object(object); 482 fscache_stat_d(&fscache_n_cop_lookup_object); 483 484 fscache_unuse_cookie(object); 485 486 if (ret == -ETIMEDOUT) { 487 /* probably stuck behind another object, so move this one to 488 * the back of the queue */ 489 fscache_stat(&fscache_n_object_lookups_timed_out); 490 _leave(" [timeout]"); 491 return NO_TRANSIT; 492 } 493 494 if (ret < 0) { 495 _leave(" [error]"); 496 return transit_to(LOOKUP_FAILURE); 497 } 498 499 _leave(" [ok]"); 500 return transit_to(OBJECT_AVAILABLE); 501} 502 503/** 504 * fscache_object_lookup_negative - Note negative cookie lookup 505 * @object: Object pointing to cookie to mark 506 * 507 * Note negative lookup, permitting those waiting to read data from an already 508 * existing backing object to continue as there's no data for them to read. 509 */ 510void fscache_object_lookup_negative(struct fscache_object *object) 511{ 512 struct fscache_cookie *cookie = object->cookie; 513 514 _enter("{OBJ%x,%s}", object->debug_id, object->state->name); 515 516 if (!test_and_set_bit(FSCACHE_OBJECT_IS_LOOKED_UP, &object->flags)) { 517 fscache_stat(&fscache_n_object_lookups_negative); 518 519 /* Allow write requests to begin stacking up and read requests to begin 520 * returning ENODATA. 521 */ 522 set_bit(FSCACHE_COOKIE_NO_DATA_YET, &cookie->flags); 523 clear_bit(FSCACHE_COOKIE_UNAVAILABLE, &cookie->flags); 524 525 _debug("wake up lookup %p", &cookie->flags); 526 clear_bit_unlock(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags); 527 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP); 528 } 529 _leave(""); 530} 531EXPORT_SYMBOL(fscache_object_lookup_negative); 532 533/** 534 * fscache_obtained_object - Note successful object lookup or creation 535 * @object: Object pointing to cookie to mark 536 * 537 * Note successful lookup and/or creation, permitting those waiting to write 538 * data to a backing object to continue. 539 * 540 * Note that after calling this, an object's cookie may be relinquished by the 541 * netfs, and so must be accessed with object lock held. 542 */ 543void fscache_obtained_object(struct fscache_object *object) 544{ 545 struct fscache_cookie *cookie = object->cookie; 546 547 _enter("{OBJ%x,%s}", object->debug_id, object->state->name); 548 549 /* if we were still looking up, then we must have a positive lookup 550 * result, in which case there may be data available */ 551 if (!test_and_set_bit(FSCACHE_OBJECT_IS_LOOKED_UP, &object->flags)) { 552 fscache_stat(&fscache_n_object_lookups_positive); 553 554 /* We do (presumably) have data */ 555 clear_bit_unlock(FSCACHE_COOKIE_NO_DATA_YET, &cookie->flags); 556 clear_bit(FSCACHE_COOKIE_UNAVAILABLE, &cookie->flags); 557 558 /* Allow write requests to begin stacking up and read requests 559 * to begin shovelling data. 560 */ 561 clear_bit_unlock(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags); 562 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP); 563 } else { 564 fscache_stat(&fscache_n_object_created); 565 } 566 567 set_bit(FSCACHE_OBJECT_IS_AVAILABLE, &object->flags); 568 _leave(""); 569} 570EXPORT_SYMBOL(fscache_obtained_object); 571 572/* 573 * handle an object that has just become available 574 */ 575static const struct fscache_state *fscache_object_available(struct fscache_object *object, 576 int event) 577{ 578 _enter("{OBJ%x},%d", object->debug_id, event); 579 580 object->oob_table = fscache_osm_run_oob; 581 582 spin_lock(&object->lock); 583 584 fscache_done_parent_op(object); 585 if (object->n_in_progress == 0) { 586 if (object->n_ops > 0) { 587 ASSERTCMP(object->n_ops, >=, object->n_obj_ops); 588 fscache_start_operations(object); 589 } else { 590 ASSERT(list_empty(&object->pending_ops)); 591 } 592 } 593 spin_unlock(&object->lock); 594 595 fscache_stat(&fscache_n_cop_lookup_complete); 596 object->cache->ops->lookup_complete(object); 597 fscache_stat_d(&fscache_n_cop_lookup_complete); 598 599 fscache_hist(fscache_obj_instantiate_histogram, object->lookup_jif); 600 fscache_stat(&fscache_n_object_avail); 601 602 _leave(""); 603 return transit_to(JUMPSTART_DEPS); 604} 605 606/* 607 * Wake up this object's dependent objects now that we've become available. 608 */ 609static const struct fscache_state *fscache_jumpstart_dependents(struct fscache_object *object, 610 int event) 611{ 612 _enter("{OBJ%x},%d", object->debug_id, event); 613 614 if (!fscache_enqueue_dependents(object, FSCACHE_OBJECT_EV_PARENT_READY)) 615 return NO_TRANSIT; /* Not finished; requeue */ 616 return transit_to(WAIT_FOR_CMD); 617} 618 619/* 620 * Handle lookup or creation failute. 621 */ 622static const struct fscache_state *fscache_lookup_failure(struct fscache_object *object, 623 int event) 624{ 625 struct fscache_cookie *cookie; 626 627 _enter("{OBJ%x},%d", object->debug_id, event); 628 629 object->oob_event_mask = 0; 630 631 fscache_stat(&fscache_n_cop_lookup_complete); 632 object->cache->ops->lookup_complete(object); 633 fscache_stat_d(&fscache_n_cop_lookup_complete); 634 635 set_bit(FSCACHE_OBJECT_KILLED_BY_CACHE, &object->flags); 636 637 cookie = object->cookie; 638 set_bit(FSCACHE_COOKIE_UNAVAILABLE, &cookie->flags); 639 if (test_and_clear_bit(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags)) 640 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP); 641 642 fscache_done_parent_op(object); 643 return transit_to(KILL_OBJECT); 644} 645 646/* 647 * Wait for completion of all active operations on this object and the death of 648 * all child objects of this object. 649 */ 650static const struct fscache_state *fscache_kill_object(struct fscache_object *object, 651 int event) 652{ 653 _enter("{OBJ%x,%d,%d},%d", 654 object->debug_id, object->n_ops, object->n_children, event); 655 656 fscache_mark_object_dead(object); 657 object->oob_event_mask = 0; 658 659 if (test_bit(FSCACHE_OBJECT_RETIRED, &object->flags)) { 660 /* Reject any new read/write ops and abort any that are pending. */ 661 clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags); 662 fscache_cancel_all_ops(object); 663 } 664 665 if (list_empty(&object->dependents) && 666 object->n_ops == 0 && 667 object->n_children == 0) 668 return transit_to(DROP_OBJECT); 669 670 if (object->n_in_progress == 0) { 671 spin_lock(&object->lock); 672 if (object->n_ops > 0 && object->n_in_progress == 0) 673 fscache_start_operations(object); 674 spin_unlock(&object->lock); 675 } 676 677 if (!list_empty(&object->dependents)) 678 return transit_to(KILL_DEPENDENTS); 679 680 return transit_to(WAIT_FOR_CLEARANCE); 681} 682 683/* 684 * Kill dependent objects. 685 */ 686static const struct fscache_state *fscache_kill_dependents(struct fscache_object *object, 687 int event) 688{ 689 _enter("{OBJ%x},%d", object->debug_id, event); 690 691 if (!fscache_enqueue_dependents(object, FSCACHE_OBJECT_EV_KILL)) 692 return NO_TRANSIT; /* Not finished */ 693 return transit_to(WAIT_FOR_CLEARANCE); 694} 695 696/* 697 * Drop an object's attachments 698 */ 699static const struct fscache_state *fscache_drop_object(struct fscache_object *object, 700 int event) 701{ 702 struct fscache_object *parent = object->parent; 703 struct fscache_cookie *cookie = object->cookie; 704 struct fscache_cache *cache = object->cache; 705 bool awaken = false; 706 707 _enter("{OBJ%x,%d},%d", object->debug_id, object->n_children, event); 708 709 ASSERT(cookie != NULL); 710 ASSERT(!hlist_unhashed(&object->cookie_link)); 711 712 if (test_bit(FSCACHE_COOKIE_AUX_UPDATED, &cookie->flags)) { 713 _debug("final update"); 714 fscache_update_aux_data(object); 715 } 716 717 /* Make sure the cookie no longer points here and that the netfs isn't 718 * waiting for us. 719 */ 720 spin_lock(&cookie->lock); 721 hlist_del_init(&object->cookie_link); 722 if (hlist_empty(&cookie->backing_objects) && 723 test_and_clear_bit(FSCACHE_COOKIE_INVALIDATING, &cookie->flags)) 724 awaken = true; 725 spin_unlock(&cookie->lock); 726 727 if (awaken) 728 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_INVALIDATING); 729 if (test_and_clear_bit(FSCACHE_COOKIE_LOOKING_UP, &cookie->flags)) 730 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_LOOKING_UP); 731 732 733 /* Prevent a race with our last child, which has to signal EV_CLEARED 734 * before dropping our spinlock. 735 */ 736 spin_lock(&object->lock); 737 spin_unlock(&object->lock); 738 739 /* Discard from the cache's collection of objects */ 740 spin_lock(&cache->object_list_lock); 741 list_del_init(&object->cache_link); 742 spin_unlock(&cache->object_list_lock); 743 744 fscache_stat(&fscache_n_cop_drop_object); 745 cache->ops->drop_object(object); 746 fscache_stat_d(&fscache_n_cop_drop_object); 747 748 /* The parent object wants to know when all it dependents have gone */ 749 if (parent) { 750 _debug("release parent OBJ%x {%d}", 751 parent->debug_id, parent->n_children); 752 753 spin_lock(&parent->lock); 754 parent->n_children--; 755 if (parent->n_children == 0) 756 fscache_raise_event(parent, FSCACHE_OBJECT_EV_CLEARED); 757 spin_unlock(&parent->lock); 758 object->parent = NULL; 759 } 760 761 /* this just shifts the object release to the work processor */ 762 fscache_put_object(object, fscache_obj_put_drop_obj); 763 fscache_stat(&fscache_n_object_dead); 764 765 _leave(""); 766 return transit_to(OBJECT_DEAD); 767} 768 769/* 770 * get a ref on an object 771 */ 772static int fscache_get_object(struct fscache_object *object, 773 enum fscache_obj_ref_trace why) 774{ 775 int ret; 776 777 fscache_stat(&fscache_n_cop_grab_object); 778 ret = object->cache->ops->grab_object(object, why) ? 0 : -EAGAIN; 779 fscache_stat_d(&fscache_n_cop_grab_object); 780 return ret; 781} 782 783/* 784 * Discard a ref on an object 785 */ 786static void fscache_put_object(struct fscache_object *object, 787 enum fscache_obj_ref_trace why) 788{ 789 fscache_stat(&fscache_n_cop_put_object); 790 object->cache->ops->put_object(object, why); 791 fscache_stat_d(&fscache_n_cop_put_object); 792} 793 794/** 795 * fscache_object_destroy - Note that a cache object is about to be destroyed 796 * @object: The object to be destroyed 797 * 798 * Note the imminent destruction and deallocation of a cache object record. 799 */ 800void fscache_object_destroy(struct fscache_object *object) 801{ 802 fscache_objlist_remove(object); 803 804 /* We can get rid of the cookie now */ 805 fscache_cookie_put(object->cookie, fscache_cookie_put_object); 806 object->cookie = NULL; 807} 808EXPORT_SYMBOL(fscache_object_destroy); 809 810/* 811 * enqueue an object for metadata-type processing 812 */ 813void fscache_enqueue_object(struct fscache_object *object) 814{ 815 _enter("{OBJ%x}", object->debug_id); 816 817 if (fscache_get_object(object, fscache_obj_get_queue) >= 0) { 818 wait_queue_head_t *cong_wq = 819 &get_cpu_var(fscache_object_cong_wait); 820 821 if (queue_work(fscache_object_wq, &object->work)) { 822 if (fscache_object_congested()) 823 wake_up(cong_wq); 824 } else 825 fscache_put_object(object, fscache_obj_put_queue); 826 827 put_cpu_var(fscache_object_cong_wait); 828 } 829} 830 831/** 832 * fscache_object_sleep_till_congested - Sleep until object wq is congested 833 * @timeoutp: Scheduler sleep timeout 834 * 835 * Allow an object handler to sleep until the object workqueue is congested. 836 * 837 * The caller must set up a wake up event before calling this and must have set 838 * the appropriate sleep mode (such as TASK_UNINTERRUPTIBLE) and tested its own 839 * condition before calling this function as no test is made here. 840 * 841 * %true is returned if the object wq is congested, %false otherwise. 842 */ 843bool fscache_object_sleep_till_congested(signed long *timeoutp) 844{ 845 wait_queue_head_t *cong_wq = this_cpu_ptr(&fscache_object_cong_wait); 846 DEFINE_WAIT(wait); 847 848 if (fscache_object_congested()) 849 return true; 850 851 add_wait_queue_exclusive(cong_wq, &wait); 852 if (!fscache_object_congested()) 853 *timeoutp = schedule_timeout(*timeoutp); 854 finish_wait(cong_wq, &wait); 855 856 return fscache_object_congested(); 857} 858EXPORT_SYMBOL_GPL(fscache_object_sleep_till_congested); 859 860/* 861 * Enqueue the dependents of an object for metadata-type processing. 862 * 863 * If we don't manage to finish the list before the scheduler wants to run 864 * again then return false immediately. We return true if the list was 865 * cleared. 866 */ 867static bool fscache_enqueue_dependents(struct fscache_object *object, int event) 868{ 869 struct fscache_object *dep; 870 bool ret = true; 871 872 _enter("{OBJ%x}", object->debug_id); 873 874 if (list_empty(&object->dependents)) 875 return true; 876 877 spin_lock(&object->lock); 878 879 while (!list_empty(&object->dependents)) { 880 dep = list_entry(object->dependents.next, 881 struct fscache_object, dep_link); 882 list_del_init(&dep->dep_link); 883 884 fscache_raise_event(dep, event); 885 fscache_put_object(dep, fscache_obj_put_enq_dep); 886 887 if (!list_empty(&object->dependents) && need_resched()) { 888 ret = false; 889 break; 890 } 891 } 892 893 spin_unlock(&object->lock); 894 return ret; 895} 896 897/* 898 * remove an object from whatever queue it's waiting on 899 */ 900static void fscache_dequeue_object(struct fscache_object *object) 901{ 902 _enter("{OBJ%x}", object->debug_id); 903 904 if (!list_empty(&object->dep_link)) { 905 spin_lock(&object->parent->lock); 906 list_del_init(&object->dep_link); 907 spin_unlock(&object->parent->lock); 908 } 909 910 _leave(""); 911} 912 913/** 914 * fscache_check_aux - Ask the netfs whether an object on disk is still valid 915 * @object: The object to ask about 916 * @data: The auxiliary data for the object 917 * @datalen: The size of the auxiliary data 918 * 919 * This function consults the netfs about the coherency state of an object. 920 * The caller must be holding a ref on cookie->n_active (held by 921 * fscache_look_up_object() on behalf of the cache backend during object lookup 922 * and creation). 923 */ 924enum fscache_checkaux fscache_check_aux(struct fscache_object *object, 925 const void *data, uint16_t datalen, 926 loff_t object_size) 927{ 928 enum fscache_checkaux result; 929 930 if (!object->cookie->def->check_aux) { 931 fscache_stat(&fscache_n_checkaux_none); 932 return FSCACHE_CHECKAUX_OKAY; 933 } 934 935 result = object->cookie->def->check_aux(object->cookie->netfs_data, 936 data, datalen, object_size); 937 switch (result) { 938 /* entry okay as is */ 939 case FSCACHE_CHECKAUX_OKAY: 940 fscache_stat(&fscache_n_checkaux_okay); 941 break; 942 943 /* entry requires update */ 944 case FSCACHE_CHECKAUX_NEEDS_UPDATE: 945 fscache_stat(&fscache_n_checkaux_update); 946 break; 947 948 /* entry requires deletion */ 949 case FSCACHE_CHECKAUX_OBSOLETE: 950 fscache_stat(&fscache_n_checkaux_obsolete); 951 break; 952 953 default: 954 BUG(); 955 } 956 957 return result; 958} 959EXPORT_SYMBOL(fscache_check_aux); 960 961/* 962 * Asynchronously invalidate an object. 963 */ 964static const struct fscache_state *_fscache_invalidate_object(struct fscache_object *object, 965 int event) 966{ 967 struct fscache_operation *op; 968 struct fscache_cookie *cookie = object->cookie; 969 970 _enter("{OBJ%x},%d", object->debug_id, event); 971 972 /* We're going to need the cookie. If the cookie is not available then 973 * retire the object instead. 974 */ 975 if (!fscache_use_cookie(object)) { 976 ASSERT(radix_tree_empty(&object->cookie->stores)); 977 set_bit(FSCACHE_OBJECT_RETIRED, &object->flags); 978 _leave(" [no cookie]"); 979 return transit_to(KILL_OBJECT); 980 } 981 982 /* Reject any new read/write ops and abort any that are pending. */ 983 fscache_invalidate_writes(cookie); 984 clear_bit(FSCACHE_OBJECT_PENDING_WRITE, &object->flags); 985 fscache_cancel_all_ops(object); 986 987 /* Now we have to wait for in-progress reads and writes */ 988 op = kzalloc(sizeof(*op), GFP_KERNEL); 989 if (!op) 990 goto nomem; 991 992 fscache_operation_init(cookie, op, object->cache->ops->invalidate_object, 993 NULL, NULL); 994 op->flags = FSCACHE_OP_ASYNC | 995 (1 << FSCACHE_OP_EXCLUSIVE) | 996 (1 << FSCACHE_OP_UNUSE_COOKIE); 997 trace_fscache_page_op(cookie, NULL, op, fscache_page_op_invalidate); 998 999 spin_lock(&cookie->lock); 1000 if (fscache_submit_exclusive_op(object, op) < 0) 1001 goto submit_op_failed; 1002 spin_unlock(&cookie->lock); 1003 fscache_put_operation(op); 1004 1005 /* Once we've completed the invalidation, we know there will be no data 1006 * stored in the cache and thus we can reinstate the data-check-skip 1007 * optimisation. 1008 */ 1009 set_bit(FSCACHE_COOKIE_NO_DATA_YET, &cookie->flags); 1010 1011 /* We can allow read and write requests to come in once again. They'll 1012 * queue up behind our exclusive invalidation operation. 1013 */ 1014 if (test_and_clear_bit(FSCACHE_COOKIE_INVALIDATING, &cookie->flags)) 1015 wake_up_bit(&cookie->flags, FSCACHE_COOKIE_INVALIDATING); 1016 _leave(" [ok]"); 1017 return transit_to(UPDATE_OBJECT); 1018 1019nomem: 1020 fscache_mark_object_dead(object); 1021 fscache_unuse_cookie(object); 1022 _leave(" [ENOMEM]"); 1023 return transit_to(KILL_OBJECT); 1024 1025submit_op_failed: 1026 fscache_mark_object_dead(object); 1027 spin_unlock(&cookie->lock); 1028 fscache_unuse_cookie(object); 1029 kfree(op); 1030 _leave(" [EIO]"); 1031 return transit_to(KILL_OBJECT); 1032} 1033 1034static const struct fscache_state *fscache_invalidate_object(struct fscache_object *object, 1035 int event) 1036{ 1037 const struct fscache_state *s; 1038 1039 fscache_stat(&fscache_n_invalidates_run); 1040 fscache_stat(&fscache_n_cop_invalidate_object); 1041 s = _fscache_invalidate_object(object, event); 1042 fscache_stat_d(&fscache_n_cop_invalidate_object); 1043 return s; 1044} 1045 1046/* 1047 * Update auxiliary data. 1048 */ 1049static void fscache_update_aux_data(struct fscache_object *object) 1050{ 1051 fscache_stat(&fscache_n_updates_run); 1052 fscache_stat(&fscache_n_cop_update_object); 1053 object->cache->ops->update_object(object); 1054 fscache_stat_d(&fscache_n_cop_update_object); 1055} 1056 1057/* 1058 * Asynchronously update an object. 1059 */ 1060static const struct fscache_state *fscache_update_object(struct fscache_object *object, 1061 int event) 1062{ 1063 _enter("{OBJ%x},%d", object->debug_id, event); 1064 1065 fscache_update_aux_data(object); 1066 1067 _leave(""); 1068 return transit_to(WAIT_FOR_CMD); 1069} 1070 1071/** 1072 * fscache_object_retrying_stale - Note retrying stale object 1073 * @object: The object that will be retried 1074 * 1075 * Note that an object lookup found an on-disk object that was adjudged to be 1076 * stale and has been deleted. The lookup will be retried. 1077 */ 1078void fscache_object_retrying_stale(struct fscache_object *object) 1079{ 1080 fscache_stat(&fscache_n_cache_no_space_reject); 1081} 1082EXPORT_SYMBOL(fscache_object_retrying_stale); 1083 1084/** 1085 * fscache_object_mark_killed - Note that an object was killed 1086 * @object: The object that was culled 1087 * @why: The reason the object was killed. 1088 * 1089 * Note that an object was killed. Returns true if the object was 1090 * already marked killed, false if it wasn't. 1091 */ 1092void fscache_object_mark_killed(struct fscache_object *object, 1093 enum fscache_why_object_killed why) 1094{ 1095 if (test_and_set_bit(FSCACHE_OBJECT_KILLED_BY_CACHE, &object->flags)) { 1096 pr_err("Error: Object already killed by cache [%s]\n", 1097 object->cache->identifier); 1098 return; 1099 } 1100 1101 switch (why) { 1102 case FSCACHE_OBJECT_NO_SPACE: 1103 fscache_stat(&fscache_n_cache_no_space_reject); 1104 break; 1105 case FSCACHE_OBJECT_IS_STALE: 1106 fscache_stat(&fscache_n_cache_stale_objects); 1107 break; 1108 case FSCACHE_OBJECT_WAS_RETIRED: 1109 fscache_stat(&fscache_n_cache_retired_objects); 1110 break; 1111 case FSCACHE_OBJECT_WAS_CULLED: 1112 fscache_stat(&fscache_n_cache_culled_objects); 1113 break; 1114 } 1115} 1116EXPORT_SYMBOL(fscache_object_mark_killed); 1117 1118/* 1119 * The object is dead. We can get here if an object gets queued by an event 1120 * that would lead to its death (such as EV_KILL) when the dispatcher is 1121 * already running (and so can be requeued) but hasn't yet cleared the event 1122 * mask. 1123 */ 1124static const struct fscache_state *fscache_object_dead(struct fscache_object *object, 1125 int event) 1126{ 1127 if (!test_and_set_bit(FSCACHE_OBJECT_RUN_AFTER_DEAD, 1128 &object->flags)) 1129 return NO_TRANSIT; 1130 1131 WARN(true, "FS-Cache object redispatched after death"); 1132 return NO_TRANSIT; 1133} 1134