Lines Matching refs:block_group
2607 struct btrfs_block_group *block_group;
2609 block_group = btrfs_lookup_block_group(fs_info, start);
2610 if (!block_group)
2613 ret = btrfs_cache_block_group(block_group, true);
2617 ret = btrfs_remove_free_space(block_group, start, num_bytes);
2619 btrfs_put_block_group(block_group);
2786 struct btrfs_block_group *block_group, *tmp;
2827 list_for_each_entry_safe(block_group, tmp, deleted_bgs, bg_list) {
2833 block_group->start,
2834 block_group->length,
2837 list_del_init(&block_group->bg_list);
2838 btrfs_unfreeze_block_group(block_group);
2839 btrfs_put_block_group(block_group);
3452 struct btrfs_block_group *block_group,
3461 used_bg = cluster->block_group;
3465 if (used_bg == block_group)
3482 if (used_bg == cluster->block_group)
3534 WARN_ON(last_ptr->block_group != cluster_bg);
3640 static int do_allocation_clustered(struct btrfs_block_group *block_group,
3648 ret = find_free_extent_clustered(block_group, ffe_ctl, bg_ret);
3654 return find_free_extent_unclustered(block_group, ffe_ctl);
3669 * block_group::lock
3678 static int do_allocation_zoned(struct btrfs_block_group *block_group,
3682 struct btrfs_fs_info *fs_info = block_group->fs_info;
3683 struct btrfs_space_info *space_info = block_group->space_info;
3684 struct btrfs_free_space_ctl *ctl = block_group->free_space_ctl;
3685 u64 start = block_group->start;
3688 u64 bytenr = block_group->start;
3694 ASSERT(btrfs_is_zoned(block_group->fs_info));
3724 spin_lock(&block_group->lock);
3725 if (block_group->ro || btrfs_zoned_bg_is_full(block_group)) {
3732 spin_unlock(&block_group->lock);
3735 if (!ret && (block_group->flags & BTRFS_BLOCK_GROUP_DATA) &&
3736 !btrfs_zone_activate(block_group)) {
3745 spin_lock(&block_group->lock);
3753 block_group->start == fs_info->treelog_bg ||
3756 block_group->start == fs_info->data_reloc_bg ||
3759 if (block_group->ro ||
3761 test_bit(BLOCK_GROUP_FLAG_ZONED_DATA_RELOC, &block_group->runtime_flags))) {
3771 (block_group->used || block_group->reserved)) {
3781 (block_group->used || block_group->reserved)) {
3786 WARN_ON_ONCE(block_group->alloc_offset > block_group->zone_capacity);
3787 avail = block_group->zone_capacity - block_group->alloc_offset;
3802 fs_info->treelog_bg = block_group->start;
3806 fs_info->data_reloc_bg = block_group->start;
3822 set_bit(BLOCK_GROUP_FLAG_ZONED_DATA_RELOC, &block_group->runtime_flags);
3825 ffe_ctl->found_offset = start + block_group->alloc_offset;
3826 block_group->alloc_offset += num_bytes;
3845 spin_unlock(&block_group->lock);
3850 static int do_allocation(struct btrfs_block_group *block_group,
3856 return do_allocation_clustered(block_group, ffe_ctl, bg_ret);
3858 return do_allocation_zoned(block_group, ffe_ctl, bg_ret);
3864 static void release_block_group(struct btrfs_block_group *block_group,
3879 BUG_ON(btrfs_bg_flags_to_raid_index(block_group->flags) !=
3881 btrfs_release_block_group(block_group, delalloc);
4125 if (last_ptr->block_group)
4156 struct btrfs_block_group *block_group;
4159 list_for_each_entry(block_group, &fs_info->zone_active_bgs, active_bg_list) {
4164 u64 avail = block_group->zone_capacity - block_group->alloc_offset;
4166 if (block_group_bits(block_group, ffe_ctl->flags) &&
4168 ffe_ctl->hint_byte = block_group->start;
4226 struct btrfs_block_group *block_group = NULL;
4272 block_group = btrfs_lookup_block_group(fs_info,
4281 if (block_group && block_group_bits(block_group, ffe_ctl->flags) &&
4282 block_group->cached != BTRFS_CACHE_NO) {
4284 if (list_empty(&block_group->list) ||
4285 block_group->ro) {
4292 btrfs_put_block_group(block_group);
4296 block_group->flags);
4297 btrfs_lock_block_group(block_group,
4302 } else if (block_group) {
4303 btrfs_put_block_group(block_group);
4313 list_for_each_entry(block_group,
4319 if (unlikely(block_group->ro)) {
4321 btrfs_clear_treelog_bg(block_group);
4323 btrfs_clear_data_reloc_bg(block_group);
4327 btrfs_grab_block_group(block_group, ffe_ctl->delalloc);
4328 ffe_ctl->search_start = block_group->start;
4335 if (!block_group_bits(block_group, ffe_ctl->flags)) {
4346 if ((ffe_ctl->flags & extra) && !(block_group->flags & extra))
4354 btrfs_release_block_group(block_group, ffe_ctl->delalloc);
4359 trace_find_free_extent_have_block_group(root, ffe_ctl, block_group);
4360 ffe_ctl->cached = btrfs_block_group_done(block_group);
4363 ret = btrfs_cache_block_group(block_group, false);
4381 if (unlikely(block_group->cached == BTRFS_CACHE_ERROR)) {
4387 if (!find_free_extent_check_size_class(ffe_ctl, block_group))
4391 ret = do_allocation(block_group, ffe_ctl, &bg_ret);
4395 if (bg_ret && bg_ret != block_group) {
4396 btrfs_release_block_group(block_group, ffe_ctl->delalloc);
4397 block_group = bg_ret;
4406 block_group->start + block_group->length) {
4407 btrfs_add_free_space_unused(block_group,
4414 btrfs_add_free_space_unused(block_group,
4418 ret = btrfs_add_reserved_bytes(block_group, ffe_ctl->ram_bytes,
4423 btrfs_add_free_space_unused(block_group,
4428 btrfs_inc_block_group_reservations(block_group);
4434 trace_btrfs_reserve_extent(block_group, ffe_ctl);
4435 btrfs_release_block_group(block_group, ffe_ctl->delalloc);
4441 btrfs_wait_block_group_cache_progress(block_group,
4447 release_block_group(block_group, ffe_ctl, ffe_ctl->delalloc);
4801 struct btrfs_block_group *block_group;
4815 block_group = btrfs_lookup_block_group(fs_info, ins->objectid);
4816 if (!block_group)
4819 space_info = block_group->space_info;
4821 spin_lock(&block_group->lock);
4823 block_group->reserved += ins->offset;
4824 spin_unlock(&block_group->lock);
4831 btrfs_put_block_group(block_group);