Lines Matching defs:new_wb
383 struct bdi_writeback *new_wb;
399 struct bdi_writeback *new_wb)
416 trace_inode_switch_wbs(inode, old_wb, new_wb);
427 wb_stat_mod(new_wb, WB_RECLAIMABLE, nr);
436 wb_stat_mod(new_wb, WB_WRITEBACK, nr);
441 atomic_inc(&new_wb->writeback_inodes);
444 wb_get(new_wb);
447 * Transfer to @new_wb's IO list if necessary. If the @inode is dirty,
452 * the b_attached list of @new_wb.
455 inode->i_wb = new_wb;
460 list_for_each_entry(pos, &new_wb->b_dirty, i_io_list)
464 inode_io_list_move_locked(inode, new_wb,
467 inode_cgwb_move_to_attached(inode, new_wb);
470 inode->i_wb = new_wb;
497 struct bdi_writeback *new_wb = isw->new_wb;
517 if (old_wb < new_wb) {
519 spin_lock_nested(&new_wb->list_lock, SINGLE_DEPTH_NESTING);
521 spin_lock(&new_wb->list_lock);
527 if (inode_do_switch_wbs(*inodep, old_wb, new_wb))
531 spin_unlock(&new_wb->list_lock);
537 wb_wakeup(new_wb);
543 wb_put(new_wb);
549 struct bdi_writeback *new_wb)
566 inode_to_wb(inode) == new_wb) {
614 isw->new_wb = wb_get_create(bdi, memcg_css, GFP_ATOMIC);
616 if (!isw->new_wb)
619 if (!inode_prepare_wbs_switch(inode, isw->new_wb))
636 if (isw->new_wb)
637 wb_put(isw->new_wb);
647 if (!inode_prepare_wbs_switch(inode, isw->new_wb))
683 isw->new_wb = wb_get_create(wb->bdi, memcg_css, GFP_KERNEL);
684 if (isw->new_wb)
687 if (unlikely(!isw->new_wb))
688 isw->new_wb = &wb->bdi->wb; /* wb_get() is noop for bdi's wb */
708 wb_put(isw->new_wb);