Lines Matching defs:work
498 * this does all the hard work for inserting an inline extent into
714 struct btrfs_work work;
819 * This is done inside an ordered work queue, and the compression is spread
820 * across many cpus. The actual IO submission is step two, and the ordered work
825 * work queue to write the uncompressed bytes. This makes sure that both
829 static void compress_file_range(struct btrfs_work *work)
832 container_of(work, struct async_chunk, work);
983 * we don't need to create any more async work items.
1017 * The async work queues will take care of doing actual allocation on
1139 * compressed size, how could it work for uncompressed size
1566 * which only gets called in the order the work was queued. We walk all the
1569 static noinline void submit_compressed_extents(struct btrfs_work *work)
1571 struct async_chunk *async_chunk = container_of(work, struct async_chunk,
1572 work);
1573 struct btrfs_fs_info *fs_info = btrfs_work_owner(work);
1594 static noinline void async_cow_free(struct btrfs_work *work)
1599 async_chunk = container_of(work, struct async_chunk, work);
1685 btrfs_init_work(&async_chunk[i].work, compress_file_range,
1691 btrfs_queue_work(fs_info->delalloc_workers, &async_chunk[i].work);
1811 * doesn't work.
2429 * list of inodes that have pending delalloc work to be done.
2679 struct btrfs_work work;
2682 static void btrfs_writepage_fixup_worker(struct btrfs_work *work)
2685 container_of(work, struct btrfs_writepage_fixup, work);
2850 btrfs_init_work(&fixup->work, btrfs_writepage_fixup_worker, NULL, NULL);
2853 btrfs_queue_work(fs_info->fixup_workers, &fixup->work);
5115 * submitting those bios, which are executed by a separate task (work
9011 * now so we don't add too much work to the end of the transaction
9206 struct btrfs_work work;
9209 static void btrfs_run_delalloc_work(struct btrfs_work *work)
9214 delalloc_work = container_of(work, struct btrfs_delalloc_work,
9215 work);
9228 struct btrfs_delalloc_work *work;
9230 work = kmalloc(sizeof(*work), GFP_NOFS);
9231 if (!work)
9234 init_completion(&work->completion);
9235 INIT_LIST_HEAD(&work->list);
9236 work->inode = inode;
9237 btrfs_init_work(&work->work, btrfs_run_delalloc_work, NULL, NULL);
9239 return work;
9252 struct btrfs_delalloc_work *work, *next;
9283 work = btrfs_alloc_delalloc_work(inode);
9284 if (!work) {
9289 list_add_tail(&work->list, &works);
9291 &work->work);
9304 list_for_each_entry_safe(work, next, &works, list) {
9305 list_del_init(&work->list);
9306 wait_for_completion(&work->completion);
9307 kfree(work);