Lines Matching refs:next
858 sums = list_entry(ordered_sums.next,
1146 goto next;
1183 next:
2053 goto next;
2066 next:
2067 /* check the next slot in the tree to see if it is a valid item */
2596 struct extent_buffer *next;
2618 next = btrfs_find_create_tree_block(fs_info, bytenr,
2621 if (IS_ERR(next))
2622 return PTR_ERR(next);
2625 ret = wc->process_func(root, next, wc, ptr_gen,
2628 free_extent_buffer(next);
2634 ret = btrfs_read_extent_buffer(next, &check);
2636 free_extent_buffer(next);
2640 ret = clean_log_buffer(trans, next);
2642 free_extent_buffer(next);
2646 free_extent_buffer(next);
2649 ret = btrfs_read_extent_buffer(next, &check);
2651 free_extent_buffer(next);
2657 path->nodes[*level-1] = next;
2658 *level = btrfs_header_level(next);
3350 * maximum possible value, so that the next attempt to log the inode does
3835 * find it anymore. If we can't find it, just move to the next key. We
3870 * look ahead to the next item and see if it is also
3888 * The next leaf was not changed in the current transaction
3890 * We check for the next key because there might have been
3892 * that next key. So the key range item we log (key type
3893 * BTRFS_DIR_LOG_INDEX_KEY) must end at the next key's
3915 * dir item and the first one on the next leaf (which did not
4610 struct btrfs_ordered_sum *sums = list_entry(ordered_sums.next,
4883 em = list_entry(extents.next, struct extent_map, list);
4982 * happens, causing the log to be replayed the next time the fs is mounted,
5509 struct btrfs_dir_list *next;
5511 list_for_each_entry_safe(dir_elem, next, &dir_list, list)
5527 struct btrfs_ino_list *next;
5529 list_for_each_entry_safe(curr, next, &ctx->conflict_inodes, list) {
6087 const struct btrfs_delayed_item *next;
6095 next = list_next_entry(curr, log_list);
6096 while (!list_entry_is_head(next, delayed_del_list, log_list)) {
6097 if (next->index != curr->index + 1)
6099 curr = next;
6100 next = list_next_entry(next, log_list);
6124 const struct btrfs_delayed_item *next;
6130 next = list_next_entry(first, log_list);
6133 !list_entry_is_head(next, delayed_del_list, log_list)) {
6139 key.offset != next->index)
6143 *last_ret = next;
6144 next = list_next_entry(next, log_list);
6633 * Reset the last_reflink_trans so that the next fsync does not need to
7198 * the next one.
7213 goto next;
7257 next:
7515 * then we know the next attempt to sync the log will fallback to a full