Lines Matching refs:io_tree

1099 	struct extent_io_tree *io_tree = &inode->io_tree;
1124 lock_extent(io_tree, start, end, NULL);
1629 unlock_extent(&inode->io_tree, start, end, NULL);
1752 struct extent_io_tree *io_tree = &inode->io_tree;
1789 count = count_range_bits(io_tree, &range_start, end, range_bytes,
1804 clear_extent_bit(io_tree, start, end, EXTENT_NORESERVE,
2238 test_range_bit(&inode->io_tree, start, end, EXTENT_DEFRAG,
2636 ret = set_extent_bit(&inode->io_tree, search_start,
2671 return set_extent_bit(&inode->io_tree, start, end,
2746 lock_extent(&inode->io_tree, page_start, page_end, &cached_state);
2754 unlock_extent(&inode->io_tree, page_start, page_end,
2781 unlock_extent(&inode->io_tree, page_start, page_end, &cached_state);
3018 struct extent_io_tree *io_tree = &inode->io_tree;
3080 lock_extent(io_tree, start, end, &cached_state);
3134 clear_extent_bit(&inode->io_tree, start, end,
3146 clear_extent_bit(&inode->io_tree, start, end, clear_bits,
3169 clear_extent_uptodate(io_tree, unwritten_start, end, NULL);
3299 test_range_bit(&inode->io_tree, file_offset, end, EXTENT_NODATASUM,
3302 clear_extent_bits(&inode->io_tree, file_offset, end,
4678 struct extent_io_tree *io_tree = &inode->io_tree;
4753 lock_extent(io_tree, block_start, block_end, &cached_state);
4757 unlock_extent(io_tree, block_start, block_end, &cached_state);
4765 clear_extent_bit(&inode->io_tree, block_start, block_end,
4772 unlock_extent(io_tree, block_start, block_end, &cached_state);
4789 unlock_extent(io_tree, block_start, block_end, &cached_state);
4792 set_extent_bit(&inode->io_tree, block_start, block_end,
4871 struct extent_io_tree *io_tree = &inode->io_tree;
4956 unlock_extent(io_tree, hole_start, block_end - 1, &cached_state);
5100 struct extent_io_tree *io_tree = &BTRFS_I(inode)->io_tree;
5121 * io_tree to unlock a range it may be too late, leading to an
5124 spin_lock(&io_tree->lock);
5125 while (!RB_EMPTY_ROOT(&io_tree->state)) {
5132 node = rb_first(&io_tree->state);
5137 spin_unlock(&io_tree->lock);
5139 lock_extent(io_tree, start, end, &cached_state);
5153 clear_extent_bit(io_tree, start, end,
5158 spin_lock(&io_tree->lock);
5160 spin_unlock(&io_tree->lock);
6743 * data directly into the page and marks the extent up to date in the io_tree.
7062 struct extent_io_tree *io_tree = &BTRFS_I(inode)->io_tree;
7132 ret = test_range_bit(io_tree, offset, range_end,
7158 struct extent_io_tree *io_tree = &BTRFS_I(inode)->io_tree;
7164 if (!try_lock_extent(io_tree, lockstart, lockend,
7168 lock_extent(io_tree, lockstart, lockend, cached_state);
7190 unlock_extent(io_tree, lockstart, lockend, cached_state);
7641 unlock_extent(&BTRFS_I(inode)->io_tree, lockstart, lockend,
7667 unlock_extent(&BTRFS_I(inode)->io_tree, lockstart, lockend,
7691 unlock_extent(&BTRFS_I(inode)->io_tree, pos, pos + length - 1,
7703 unlock_extent(&BTRFS_I(inode)->io_tree, pos,
7736 unlock_extent(&inode->io_tree, dip->file_offset,
7932 struct extent_io_tree *tree = &inode->io_tree;
8063 * Then its QGROUP_RESERVED bit in io_tree is already cleared.
8069 * QGROUP_RESERVED bit of its io_tree, and free the qgroup
8114 struct extent_io_tree *io_tree = &BTRFS_I(inode)->io_tree;
8169 lock_extent(io_tree, page_start, page_end, &cached_state);
8173 unlock_extent(io_tree, page_start, page_end, &cached_state);
8184 unlock_extent(io_tree, page_start, page_end, &cached_state);
8210 clear_extent_bit(&BTRFS_I(inode)->io_tree, page_start, end,
8217 unlock_extent(io_tree, page_start, page_end, &cached_state);
8237 unlock_extent(io_tree, page_start, page_end, &cached_state);
8347 lock_extent(&inode->io_tree, lock_start, (u64)-1, &cached_state);
8362 unlock_extent(&inode->io_tree, lock_start, (u64)-1, &cached_state);
8515 extent_io_tree_init(fs_info, &ei->io_tree, IO_TREE_INODE_IO);
8516 ei->io_tree.inode = ei;
9839 struct extent_io_tree *io_tree = &inode->io_tree;
9901 unlock_extent(io_tree, start, lockend, cached_state);
9995 struct extent_io_tree *io_tree = &inode->io_tree;
10017 unlock_extent(io_tree, start, lockend, cached_state);
10057 struct extent_io_tree *io_tree = &inode->io_tree;
10087 lock_extent(io_tree, start, lockend, &cached_state);
10093 unlock_extent(io_tree, start, lockend, &cached_state);
10167 unlock_extent(io_tree, start, lockend, &cached_state);
10188 unlock_extent(io_tree, start, lockend, &cached_state);
10201 struct extent_io_tree *io_tree = &inode->io_tree;
10334 lock_extent(io_tree, start, end, &cached_state);
10341 unlock_extent(io_tree, start, end, &cached_state);
10404 unlock_extent(io_tree, start, end, &cached_state);
10429 unlock_extent(io_tree, start, end, &cached_state);
10580 struct extent_io_tree *io_tree = &BTRFS_I(inode)->io_tree;
10669 lock_extent(io_tree, 0, isize - 1, &cached_state);
10806 unlock_extent(io_tree, 0, isize - 1, &cached_state);
10871 * the inode's io_tree.