Lines Matching refs:sb

44 	struct super_block *sb;
1266 * Caller need not hold sb s_umount semaphore.
1298 * mark an inode as under writeback on the sb
1302 struct super_block *sb = inode->i_sb;
1306 spin_lock_irqsave(&sb->s_inode_wblist_lock, flags);
1308 list_add_tail(&inode->i_wb_list, &sb->s_inodes_wb);
1311 spin_unlock_irqrestore(&sb->s_inode_wblist_lock, flags);
1316 * clear an inode as under writeback on the sb
1320 struct super_block *sb = inode->i_sb;
1324 spin_lock_irqsave(&sb->s_inode_wblist_lock, flags);
1329 spin_unlock_irqrestore(&sb->s_inode_wblist_lock, flags);
1417 struct super_block *sb = NULL;
1433 if (sb && sb != inode->i_sb)
1435 sb = inode->i_sb;
1438 /* just one sb in list, splice to dispatch_queue and we're done */
1451 sb = wb_inode(tmp.prev)->i_sb;
1454 if (inode->i_sb == sb)
1841 * Write a portion of b_io inodes which belong to @sb.
1849 static long writeback_sb_inodes(struct super_block *sb,
1877 if (inode->i_sb != sb) {
1878 if (work->sb) {
2007 struct super_block *sb = inode->i_sb;
2009 if (!super_trylock_shared(sb)) {
2013 * requeue_io() to avoid busy retrying the inode/sb.
2018 wrote += writeback_sb_inodes(sb, wb, work);
2019 up_read(&sb->s_umount);
2126 if (work->sb)
2127 progress = writeback_sb_inodes(work->sb, wb, work);
2456 struct super_block *sb = inode->i_sb;
2485 if (sb->s_op->dirty_inode)
2486 sb->s_op->dirty_inode(inode,
2602 static void wait_sb_inodes(struct super_block *sb)
2610 WARN_ON(!rwsem_is_locked(&sb->s_umount));
2612 mutex_lock(&sb->s_sync_lock);
2619 * reference. s_inode_wblist_lock protects sb->s_inodes_wb as well as
2624 spin_lock_irq(&sb->s_inode_wblist_lock);
2625 list_splice_init(&sb->s_inodes_wb, &sync_list);
2645 list_move_tail(&inode->i_wb_list, &sb->s_inodes_wb);
2655 spin_unlock_irq(&sb->s_inode_wblist_lock);
2661 spin_lock_irq(&sb->s_inode_wblist_lock);
2680 spin_lock_irq(&sb->s_inode_wblist_lock);
2682 spin_unlock_irq(&sb->s_inode_wblist_lock);
2684 mutex_unlock(&sb->s_sync_lock);
2687 static void __writeback_inodes_sb_nr(struct super_block *sb, unsigned long nr,
2690 struct backing_dev_info *bdi = sb->s_bdi;
2693 .sb = sb,
2703 WARN_ON(!rwsem_is_locked(&sb->s_umount));
2705 bdi_split_work_to_wbs(sb->s_bdi, &work, skip_if_busy);
2711 * @sb: the superblock
2719 void writeback_inodes_sb_nr(struct super_block *sb,
2723 __writeback_inodes_sb_nr(sb, nr, reason, false);
2729 * @sb: the superblock
2736 void writeback_inodes_sb(struct super_block *sb, enum wb_reason reason)
2738 writeback_inodes_sb_nr(sb, get_nr_dirty_pages(), reason);
2744 * @sb: the superblock
2749 void try_to_writeback_inodes_sb(struct super_block *sb, enum wb_reason reason)
2751 if (!down_read_trylock(&sb->s_umount))
2754 __writeback_inodes_sb_nr(sb, get_nr_dirty_pages(), reason, true);
2755 up_read(&sb->s_umount);
2760 * sync_inodes_sb - sync sb inode pages
2761 * @sb: the superblock
2766 void sync_inodes_sb(struct super_block *sb)
2768 struct backing_dev_info *bdi = sb->s_bdi;
2771 .sb = sb,
2787 WARN_ON(!rwsem_is_locked(&sb->s_umount));
2795 wait_sb_inodes(sb);