Lines Matching refs:set
53 * out of a cache set.
80 ca->set->need_gc = max(ca->set->need_gc, bucket_gc_gen(b));
81 WARN_ON_ONCE(ca->set->need_gc > BUCKET_GC_GEN_MAX);
132 BUG_ON(!ca->set->gc_mark_valid);
142 lockdep_assert_held(&ca->set->bucket_lock);
171 unsigned int min_prio = (INITIAL_PRIO - ca->set->min_prio) / 8; \
173 (b->prio - ca->set->min_prio + min_prio) * GC_SECTORS_USED(b); \
208 wake_up_gc(ca->set);
233 wake_up_gc(ca->set);
259 wake_up_gc(ca->set);
289 mutex_unlock(&(ca)->set->bucket_lock); \
291 test_bit(CACHE_SET_IO_DISABLE, &ca->set->flags)) { \
297 mutex_lock(&(ca)->set->bucket_lock); \
321 mutex_lock(&ca->set->bucket_lock);
336 mutex_unlock(&ca->set->bucket_lock);
338 bucket_to_sector(ca->set, bucket),
340 mutex_lock(&ca->set->bucket_lock);
344 wake_up(&ca->set->btree_cache_wait);
345 wake_up(&ca->set->bucket_wait);
355 allocator_wait(ca, ca->set->gc_mark_valid &&
363 allocator_wait(ca, !atomic_read(&ca->set->prio_blocked));
381 wake_up_gc(ca->set);
399 /* No allocation if CACHE_SET_IO_DISABLE bit is set */
400 if (unlikely(test_bit(CACHE_SET_IO_DISABLE, &ca->set->flags)))
414 prepare_to_wait(&ca->set->bucket_wait, &w,
417 mutex_unlock(&ca->set->bucket_lock);
419 mutex_lock(&ca->set->bucket_lock);
423 finish_wait(&ca->set->bucket_wait, &w);
430 if (expensive_debug_checks(ca->set)) {
461 if (ca->set->avail_nbuckets > 0) {
462 ca->set->avail_nbuckets--;
463 bch_update_bucket_in_use(ca->set, &ca->set->gc_stats);
474 if (ca->set->avail_nbuckets < ca->set->nbuckets) {
475 ca->set->avail_nbuckets++;
476 bch_update_bucket_in_use(ca->set, &ca->set->gc_stats);
495 /* No allocation if CACHE_SET_IO_DISABLE bit is set */