Lines Matching defs:path

505 	struct btrfs_path *path;
514 path = btrfs_alloc_path();
515 if (!path)
535 path->skip_locking = 1;
536 path->search_commit_root = 1;
537 path->reada = READA_FORWARD;
544 ret = btrfs_search_slot(NULL, extent_root, &key, path, 0, 0);
548 leaf = path->nodes[0];
557 if (path->slots[0] < nritems) {
558 btrfs_item_key_to_cpu(leaf, &key, path->slots[0]);
560 ret = btrfs_find_next_key(extent_root, path, &key, 0, 0);
568 btrfs_release_path(path);
577 ret = btrfs_next_leaf(extent_root, path);
582 leaf = path->nodes[0];
594 btrfs_release_path(path);
599 path->slots[0]++;
622 path->slots[0]++;
631 btrfs_free_path(path);
880 struct btrfs_path *path,
893 ret = btrfs_search_slot(trans, root, &key, path, -1, 1);
899 ret = btrfs_del_item(trans, root, path);
907 struct btrfs_path *path;
952 path = btrfs_alloc_path();
953 if (!path) {
962 inode = lookup_free_space_inode(block_group, path);
976 btrfs_wait_cache_io(trans, block_group, path);
1014 ret = btrfs_search_slot(trans, tree_root, &key, path, -1, 1);
1018 btrfs_release_path(path);
1020 ret = btrfs_del_item(trans, tree_root, path);
1023 btrfs_release_path(path);
1122 ret = remove_block_group_item(trans, path, block_group);
1173 btrfs_free_path(path);
1382 * to the unused_bgs code path. Therefore, if it's not fully
1472 * The normal path here is an unused block group is passed here,
1473 * then trimming is handled in the transaction commit path.
1475 * before coming down the unused block group path as trimming
1476 * will no longer be done later in the transaction commit path.
1548 struct btrfs_path *path)
1558 slot = path->slots[0];
1559 leaf = path->nodes[0];
1599 struct btrfs_path *path,
1608 ret = btrfs_search_slot(NULL, root, key, path, 0, 0);
1613 slot = path->slots[0];
1614 leaf = path->nodes[0];
1616 ret = btrfs_next_leaf(root, path);
1627 ret = read_bg_from_eb(fs_info, &found_key, path);
1631 path->slots[0]++;
1889 struct btrfs_path *path,
1892 struct extent_buffer *leaf = path->nodes[0];
1894 int slot = path->slots[0];
1905 struct btrfs_path *path,
1920 read_block_group_item(cache, path, key);
2008 struct btrfs_path *path;
2019 path = btrfs_alloc_path();
2020 if (!path)
2031 ret = find_first_block_group(info, path, &key);
2037 btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
2038 ret = read_one_block_group(info, path, &key, need_clear);
2043 btrfs_release_path(path);
2045 btrfs_release_path(path);
2082 btrfs_free_path(path);
2344 struct btrfs_path *path,
2359 ret = btrfs_search_slot(trans, root, &key, path, 0, 1);
2366 leaf = path->nodes[0];
2367 bi = btrfs_item_ptr_offset(leaf, path->slots[0]);
2375 btrfs_release_path(path);
2382 struct btrfs_path *path)
2408 inode = lookup_free_space_inode(block_group, path);
2411 btrfs_release_path(path);
2422 ret = create_free_space_inode(trans, block_group, path);
2530 btrfs_release_path(path);
2547 struct btrfs_path *path;
2553 path = btrfs_alloc_path();
2554 if (!path)
2561 cache_save_setup(cache, trans, path);
2564 btrfs_free_path(path);
2587 struct btrfs_path *path = NULL;
2604 if (!path) {
2605 path = btrfs_alloc_path();
2606 if (!path) {
2630 btrfs_wait_cache_io(trans, cache, path);
2649 cache_save_setup(cache, trans, path);
2653 ret = btrfs_write_out_cache(trans, cache, path);
2672 ret = update_block_group_item(trans, path, cache);
2742 btrfs_free_path(path);
2753 struct btrfs_path *path;
2756 path = btrfs_alloc_path();
2757 if (!path)
2789 btrfs_wait_cache_io(trans, cache, path);
2802 cache_save_setup(cache, trans, path);
2810 ret = btrfs_write_out_cache(trans, cache, path);
2823 ret = update_block_group_item(trans, path, cache);
2840 ret = update_block_group_item(trans, path, cache);
2862 btrfs_wait_cache_io(trans, cache, path);
2866 btrfs_free_path(path);
2981 * @num_bytes except for the compress path.