Lines Matching defs:new_wb
358 struct bdi_writeback *new_wb;
374 struct bdi_writeback *new_wb)
391 trace_inode_switch_wbs(inode, old_wb, new_wb);
402 wb_stat_mod(new_wb, WB_RECLAIMABLE, nr);
411 wb_stat_mod(new_wb, WB_WRITEBACK, nr);
416 atomic_inc(&new_wb->writeback_inodes);
419 wb_get(new_wb);
422 * Transfer to @new_wb's IO list if necessary. If the @inode is dirty,
427 * the b_attached list of @new_wb.
430 inode->i_wb = new_wb;
435 list_for_each_entry(pos, &new_wb->b_dirty, i_io_list)
439 inode_io_list_move_locked(inode, new_wb,
442 inode_cgwb_move_to_attached(inode, new_wb);
445 inode->i_wb = new_wb;
472 struct bdi_writeback *new_wb = isw->new_wb;
492 if (old_wb < new_wb) {
494 spin_lock_nested(&new_wb->list_lock, SINGLE_DEPTH_NESTING);
496 spin_lock(&new_wb->list_lock);
502 if (inode_do_switch_wbs(*inodep, old_wb, new_wb))
506 spin_unlock(&new_wb->list_lock);
512 wb_wakeup(new_wb);
518 wb_put(new_wb);
524 struct bdi_writeback *new_wb)
541 inode_to_wb(inode) == new_wb) {
589 isw->new_wb = wb_get_create(bdi, memcg_css, GFP_ATOMIC);
591 if (!isw->new_wb)
594 if (!inode_prepare_wbs_switch(inode, isw->new_wb))
611 if (isw->new_wb)
612 wb_put(isw->new_wb);
622 if (!inode_prepare_wbs_switch(inode, isw->new_wb))
658 isw->new_wb = wb_get_create(wb->bdi, memcg_css, GFP_KERNEL);
659 if (isw->new_wb)
662 if (unlikely(!isw->new_wb))
663 isw->new_wb = &wb->bdi->wb; /* wb_get() is noop for bdi's wb */
683 wb_put(isw->new_wb);