Lines Matching defs:cur

205 	struct btrfs_stripe_hash *cur;
230 cur = h + i;
231 INIT_LIST_HEAD(&cur->hash_list);
232 spin_lock_init(&cur->lock);
556 struct btrfs_raid_bio *cur)
559 test_bit(RBIO_RMW_LOCKED_BIT, &cur->flags))
570 test_bit(RBIO_CACHE_BIT, &cur->flags))
574 cur->bbio->raid_map[0])
578 if (last->operation != cur->operation)
597 int cur_fa = cur->faila;
598 int cur_fb = cur->failb;
605 if (cur->faila >= cur->failb) {
606 cur_fa = cur->failb;
607 cur_fb = cur->faila;
676 struct btrfs_raid_bio *cur;
686 list_for_each_entry(cur, &h->hash_list, hash_list) {
687 if (cur->bbio->raid_map[0] != rbio->bbio->raid_map[0])
690 spin_lock(&cur->bio_list_lock);
693 if (bio_list_empty(&cur->bio_list) &&
694 list_empty(&cur->plug_list) &&
695 test_bit(RBIO_CACHE_BIT, &cur->flags) &&
696 !test_bit(RBIO_RMW_LOCKED_BIT, &cur->flags)) {
697 list_del_init(&cur->hash_list);
698 refcount_dec(&cur->refs);
700 steal_rbio(cur, rbio);
701 cache_drop = cur;
702 spin_unlock(&cur->bio_list_lock);
708 if (rbio_can_merge(cur, rbio)) {
709 merge_rbio(cur, rbio);
710 spin_unlock(&cur->bio_list_lock);
722 list_for_each_entry(pending, &cur->plug_list, plug_list) {
725 spin_unlock(&cur->bio_list_lock);
736 list_add_tail(&rbio->plug_list, &cur->plug_list);
737 spin_unlock(&cur->bio_list_lock);
856 static void rbio_endio_bio_list(struct bio *cur, blk_status_t err)
860 while (cur) {
861 next = cur->bi_next;
862 cur->bi_next = NULL;
863 cur->bi_status = err;
864 bio_endio(cur);
865 cur = next;
875 struct bio *cur = bio_list_get(&rbio->bio_list);
899 rbio_endio_bio_list(cur, err);
1690 struct btrfs_raid_bio *cur;
1700 cur = list_entry(plug->rbio_list.next,
1702 list_del_init(&cur->plug_list);
1704 if (rbio_is_full(cur)) {
1708 ret = full_stripe_write(cur);
1713 if (rbio_can_merge(last, cur)) {
1714 merge_rbio(last, cur);
1715 __free_raid_bio(cur);
1721 last = cur;