new_wb 335 fs/fs-writeback.c struct bdi_writeback *new_wb; new_wb 359 fs/fs-writeback.c struct bdi_writeback *new_wb = isw->new_wb; new_wb 380 fs/fs-writeback.c if (old_wb < new_wb) { new_wb 382 fs/fs-writeback.c spin_lock_nested(&new_wb->list_lock, SINGLE_DEPTH_NESTING); new_wb 384 fs/fs-writeback.c spin_lock(&new_wb->list_lock); new_wb 397 fs/fs-writeback.c trace_inode_switch_wbs(inode, old_wb, new_wb); new_wb 407 fs/fs-writeback.c inc_wb_stat(new_wb, WB_RECLAIMABLE); new_wb 415 fs/fs-writeback.c inc_wb_stat(new_wb, WB_WRITEBACK); new_wb 418 fs/fs-writeback.c wb_get(new_wb); new_wb 430 fs/fs-writeback.c inode->i_wb = new_wb; new_wb 431 fs/fs-writeback.c list_for_each_entry(pos, &new_wb->b_dirty, i_io_list) new_wb 435 fs/fs-writeback.c inode_io_list_move_locked(inode, new_wb, pos->i_io_list.prev); new_wb 437 fs/fs-writeback.c inode->i_wb = new_wb; new_wb 454 fs/fs-writeback.c spin_unlock(&new_wb->list_lock); new_wb 460 fs/fs-writeback.c wb_wakeup(new_wb); new_wb 463 fs/fs-writeback.c wb_put(new_wb); new_wb 511 fs/fs-writeback.c isw->new_wb = wb_get_create(bdi, memcg_css, GFP_ATOMIC); new_wb 513 fs/fs-writeback.c if (!isw->new_wb) new_wb 520 fs/fs-writeback.c inode_to_wb(inode) == isw->new_wb) { new_wb 542 fs/fs-writeback.c if (isw->new_wb) new_wb 543 fs/fs-writeback.c wb_put(isw->new_wb); new_wb 212 include/trace/events/writeback.h struct bdi_writeback *new_wb), new_wb 214 include/trace/events/writeback.h TP_ARGS(inode, old_wb, new_wb), new_wb 227 include/trace/events/writeback.h __entry->new_cgroup_ino = __trace_wb_assign_cgroup(new_wb);