11da177e4SLinus Torvalds /* 21da177e4SLinus Torvalds * include/linux/backing-dev.h 31da177e4SLinus Torvalds * 41da177e4SLinus Torvalds * low-level device information and state which is propagated up through 51da177e4SLinus Torvalds * to high-level code. 61da177e4SLinus Torvalds */ 71da177e4SLinus Torvalds 81da177e4SLinus Torvalds #ifndef _LINUX_BACKING_DEV_H 91da177e4SLinus Torvalds #define _LINUX_BACKING_DEV_H 101da177e4SLinus Torvalds 11cf0ca9feSPeter Zijlstra #include <linux/kernel.h> 12e4ad08feSMiklos Szeredi #include <linux/fs.h> 1303ba3782SJens Axboe #include <linux/sched.h> 14a212b105STejun Heo #include <linux/blkdev.h> 1503ba3782SJens Axboe #include <linux/writeback.h> 1652ebea74STejun Heo #include <linux/blk-cgroup.h> 1766114cadSTejun Heo #include <linux/backing-dev-defs.h> 181da177e4SLinus Torvalds 198077c0d9SMikulas Patocka int __must_check bdi_init(struct backing_dev_info *bdi); 20b2e8fb6eSPeter Zijlstra void bdi_destroy(struct backing_dev_info *bdi); 21b2e8fb6eSPeter Zijlstra 22d2cc4ddeSJoe Perches __printf(3, 4) 23cf0ca9feSPeter Zijlstra int bdi_register(struct backing_dev_info *bdi, struct device *parent, 24cf0ca9feSPeter Zijlstra const char *fmt, ...); 25cf0ca9feSPeter Zijlstra int bdi_register_dev(struct backing_dev_info *bdi, dev_t dev); 26cf0ca9feSPeter Zijlstra void bdi_unregister(struct backing_dev_info *bdi); 27b4caecd4SChristoph Hellwig int __must_check bdi_setup_and_register(struct backing_dev_info *, char *); 28c00ddad3STejun Heo void wb_start_writeback(struct bdi_writeback *wb, long nr_pages, 29c00ddad3STejun Heo bool range_cyclic, enum wb_reason reason); 309ecf4866STejun Heo void wb_start_background_writeback(struct bdi_writeback *wb); 31f0054bb1STejun Heo void wb_workfn(struct work_struct *work); 32f0054bb1STejun Heo void wb_wakeup_delayed(struct bdi_writeback *wb); 33cf0ca9feSPeter Zijlstra 3403ba3782SJens Axboe extern spinlock_t bdi_lock; 3566f3b8e2SJens Axboe extern struct list_head bdi_list; 3666f3b8e2SJens Axboe 37839a8e86STejun Heo extern struct workqueue_struct *bdi_wq; 38839a8e86STejun Heo 39d6c10f1fSTejun Heo static inline bool wb_has_dirty_io(struct bdi_writeback *wb) 4003ba3782SJens Axboe { 41d6c10f1fSTejun Heo return test_bit(WB_has_dirty_io, &wb->state); 4203ba3782SJens Axboe } 4303ba3782SJens Axboe 4495a46c65STejun Heo static inline bool bdi_has_dirty_io(struct backing_dev_info *bdi) 4595a46c65STejun Heo { 4695a46c65STejun Heo /* 4795a46c65STejun Heo * @bdi->tot_write_bandwidth is guaranteed to be > 0 if there are 4895a46c65STejun Heo * any dirty wbs. See wb_update_write_bandwidth(). 4995a46c65STejun Heo */ 5095a46c65STejun Heo return atomic_long_read(&bdi->tot_write_bandwidth); 5195a46c65STejun Heo } 5295a46c65STejun Heo 5393f78d88STejun Heo static inline void __add_wb_stat(struct bdi_writeback *wb, 5493f78d88STejun Heo enum wb_stat_item item, s64 amount) 55e0bf68ddSPeter Zijlstra { 5693f78d88STejun Heo __percpu_counter_add(&wb->stat[item], amount, WB_STAT_BATCH); 57e0bf68ddSPeter Zijlstra } 58e0bf68ddSPeter Zijlstra 5993f78d88STejun Heo static inline void __inc_wb_stat(struct bdi_writeback *wb, 6093f78d88STejun Heo enum wb_stat_item item) 61e0bf68ddSPeter Zijlstra { 6293f78d88STejun Heo __add_wb_stat(wb, item, 1); 63b2e8fb6eSPeter Zijlstra } 64b2e8fb6eSPeter Zijlstra 6593f78d88STejun Heo static inline void inc_wb_stat(struct bdi_writeback *wb, enum wb_stat_item item) 66b2e8fb6eSPeter Zijlstra { 67b2e8fb6eSPeter Zijlstra unsigned long flags; 68b2e8fb6eSPeter Zijlstra 69b2e8fb6eSPeter Zijlstra local_irq_save(flags); 7093f78d88STejun Heo __inc_wb_stat(wb, item); 71b2e8fb6eSPeter Zijlstra local_irq_restore(flags); 72b2e8fb6eSPeter Zijlstra } 73b2e8fb6eSPeter Zijlstra 7493f78d88STejun Heo static inline void __dec_wb_stat(struct bdi_writeback *wb, 7593f78d88STejun Heo enum wb_stat_item item) 76b2e8fb6eSPeter Zijlstra { 7793f78d88STejun Heo __add_wb_stat(wb, item, -1); 78b2e8fb6eSPeter Zijlstra } 79b2e8fb6eSPeter Zijlstra 8093f78d88STejun Heo static inline void dec_wb_stat(struct bdi_writeback *wb, enum wb_stat_item item) 81b2e8fb6eSPeter Zijlstra { 82b2e8fb6eSPeter Zijlstra unsigned long flags; 83b2e8fb6eSPeter Zijlstra 84b2e8fb6eSPeter Zijlstra local_irq_save(flags); 8593f78d88STejun Heo __dec_wb_stat(wb, item); 86b2e8fb6eSPeter Zijlstra local_irq_restore(flags); 87b2e8fb6eSPeter Zijlstra } 88b2e8fb6eSPeter Zijlstra 8993f78d88STejun Heo static inline s64 wb_stat(struct bdi_writeback *wb, enum wb_stat_item item) 90b2e8fb6eSPeter Zijlstra { 9193f78d88STejun Heo return percpu_counter_read_positive(&wb->stat[item]); 92b2e8fb6eSPeter Zijlstra } 93b2e8fb6eSPeter Zijlstra 9493f78d88STejun Heo static inline s64 __wb_stat_sum(struct bdi_writeback *wb, 9593f78d88STejun Heo enum wb_stat_item item) 96b2e8fb6eSPeter Zijlstra { 9793f78d88STejun Heo return percpu_counter_sum_positive(&wb->stat[item]); 98b2e8fb6eSPeter Zijlstra } 99b2e8fb6eSPeter Zijlstra 10093f78d88STejun Heo static inline s64 wb_stat_sum(struct bdi_writeback *wb, enum wb_stat_item item) 101b2e8fb6eSPeter Zijlstra { 102b2e8fb6eSPeter Zijlstra s64 sum; 103b2e8fb6eSPeter Zijlstra unsigned long flags; 104b2e8fb6eSPeter Zijlstra 105b2e8fb6eSPeter Zijlstra local_irq_save(flags); 10693f78d88STejun Heo sum = __wb_stat_sum(wb, item); 107b2e8fb6eSPeter Zijlstra local_irq_restore(flags); 108b2e8fb6eSPeter Zijlstra 109b2e8fb6eSPeter Zijlstra return sum; 110b2e8fb6eSPeter Zijlstra } 111b2e8fb6eSPeter Zijlstra 11293f78d88STejun Heo extern void wb_writeout_inc(struct bdi_writeback *wb); 113dd5656e5SMiklos Szeredi 114b2e8fb6eSPeter Zijlstra /* 115b2e8fb6eSPeter Zijlstra * maximal error of a stat counter. 116b2e8fb6eSPeter Zijlstra */ 11793f78d88STejun Heo static inline unsigned long wb_stat_error(struct bdi_writeback *wb) 118b2e8fb6eSPeter Zijlstra { 119b2e8fb6eSPeter Zijlstra #ifdef CONFIG_SMP 12093f78d88STejun Heo return nr_cpu_ids * WB_STAT_BATCH; 121b2e8fb6eSPeter Zijlstra #else 122b2e8fb6eSPeter Zijlstra return 1; 123b2e8fb6eSPeter Zijlstra #endif 124e0bf68ddSPeter Zijlstra } 1251da177e4SLinus Torvalds 126189d3c4aSPeter Zijlstra int bdi_set_min_ratio(struct backing_dev_info *bdi, unsigned int min_ratio); 127a42dde04SPeter Zijlstra int bdi_set_max_ratio(struct backing_dev_info *bdi, unsigned int max_ratio); 128189d3c4aSPeter Zijlstra 1291da177e4SLinus Torvalds /* 1301da177e4SLinus Torvalds * Flags in backing_dev_info::capability 131e4ad08feSMiklos Szeredi * 132e4ad08feSMiklos Szeredi * The first three flags control whether dirty pages will contribute to the 1331da177e4SLinus Torvalds * VM's accounting and whether writepages() should be called for dirty pages 1341da177e4SLinus Torvalds * (something that would not, for example, be appropriate for ramfs) 135e4ad08feSMiklos Szeredi * 136e4ad08feSMiklos Szeredi * WARNING: these flags are closely related and should not normally be 137e4ad08feSMiklos Szeredi * used separately. The BDI_CAP_NO_ACCT_AND_WRITEBACK combines these 138e4ad08feSMiklos Szeredi * three flags into a single convenience macro. 139e4ad08feSMiklos Szeredi * 140e4ad08feSMiklos Szeredi * BDI_CAP_NO_ACCT_DIRTY: Dirty pages shouldn't contribute to accounting 141e4ad08feSMiklos Szeredi * BDI_CAP_NO_WRITEBACK: Don't write pages back 142e4ad08feSMiklos Szeredi * BDI_CAP_NO_ACCT_WB: Don't automatically account writeback pages 1435a537485SMaxim Patlasov * BDI_CAP_STRICTLIMIT: Keep number of dirty pages below bdi threshold. 14489e9b9e0STejun Heo * 14589e9b9e0STejun Heo * BDI_CAP_CGROUP_WRITEBACK: Supports cgroup-aware writeback. 1461da177e4SLinus Torvalds */ 147e4ad08feSMiklos Szeredi #define BDI_CAP_NO_ACCT_DIRTY 0x00000001 148e4ad08feSMiklos Szeredi #define BDI_CAP_NO_WRITEBACK 0x00000002 149b4caecd4SChristoph Hellwig #define BDI_CAP_NO_ACCT_WB 0x00000004 150b4caecd4SChristoph Hellwig #define BDI_CAP_STABLE_WRITES 0x00000008 151b4caecd4SChristoph Hellwig #define BDI_CAP_STRICTLIMIT 0x00000010 15289e9b9e0STejun Heo #define BDI_CAP_CGROUP_WRITEBACK 0x00000020 1531da177e4SLinus Torvalds 154e4ad08feSMiklos Szeredi #define BDI_CAP_NO_ACCT_AND_WRITEBACK \ 155e4ad08feSMiklos Szeredi (BDI_CAP_NO_WRITEBACK | BDI_CAP_NO_ACCT_DIRTY | BDI_CAP_NO_ACCT_WB) 156e4ad08feSMiklos Szeredi 1575129a469SJörn Engel extern struct backing_dev_info noop_backing_dev_info; 1581da177e4SLinus Torvalds 159bc05873dSTejun Heo /** 160bc05873dSTejun Heo * writeback_in_progress - determine whether there is writeback in progress 161bc05873dSTejun Heo * @wb: bdi_writeback of interest 162bc05873dSTejun Heo * 163bc05873dSTejun Heo * Determine whether there is writeback waiting to be handled against a 164bc05873dSTejun Heo * bdi_writeback. 165bc05873dSTejun Heo */ 166bc05873dSTejun Heo static inline bool writeback_in_progress(struct bdi_writeback *wb) 167bc05873dSTejun Heo { 168bc05873dSTejun Heo return test_bit(WB_writeback_running, &wb->state); 169bc05873dSTejun Heo } 1701da177e4SLinus Torvalds 171a212b105STejun Heo static inline struct backing_dev_info *inode_to_bdi(struct inode *inode) 172a212b105STejun Heo { 173a212b105STejun Heo struct super_block *sb; 174a212b105STejun Heo 175a212b105STejun Heo if (!inode) 176a212b105STejun Heo return &noop_backing_dev_info; 177a212b105STejun Heo 178a212b105STejun Heo sb = inode->i_sb; 179a212b105STejun Heo #ifdef CONFIG_BLOCK 180a212b105STejun Heo if (sb_is_blkdev_sb(sb)) 181a212b105STejun Heo return blk_get_backing_dev_info(I_BDEV(inode)); 182a212b105STejun Heo #endif 183a212b105STejun Heo return sb->s_bdi; 184a212b105STejun Heo } 185a212b105STejun Heo 186ec8a6f26STejun Heo static inline int wb_congested(struct bdi_writeback *wb, int cong_bits) 1871da177e4SLinus Torvalds { 188ec8a6f26STejun Heo struct backing_dev_info *bdi = wb->bdi; 189ec8a6f26STejun Heo 1901da177e4SLinus Torvalds if (bdi->congested_fn) 191ec8a6f26STejun Heo return bdi->congested_fn(bdi->congested_data, cong_bits); 192ec8a6f26STejun Heo return wb->congested->state & cong_bits; 1931da177e4SLinus Torvalds } 1941da177e4SLinus Torvalds 1958aa7e847SJens Axboe long congestion_wait(int sync, long timeout); 1960e093d99SMel Gorman long wait_iff_congested(struct zone *zone, int sync, long timeout); 1973965c9aeSWanpeng Li int pdflush_proc_obsolete(struct ctl_table *table, int write, 1983965c9aeSWanpeng Li void __user *buffer, size_t *lenp, loff_t *ppos); 1991da177e4SLinus Torvalds 2007d311cdaSDarrick J. Wong static inline bool bdi_cap_stable_pages_required(struct backing_dev_info *bdi) 2017d311cdaSDarrick J. Wong { 2027d311cdaSDarrick J. Wong return bdi->capabilities & BDI_CAP_STABLE_WRITES; 2037d311cdaSDarrick J. Wong } 2047d311cdaSDarrick J. Wong 205e4ad08feSMiklos Szeredi static inline bool bdi_cap_writeback_dirty(struct backing_dev_info *bdi) 206e4ad08feSMiklos Szeredi { 207e4ad08feSMiklos Szeredi return !(bdi->capabilities & BDI_CAP_NO_WRITEBACK); 208e4ad08feSMiklos Szeredi } 2091da177e4SLinus Torvalds 210e4ad08feSMiklos Szeredi static inline bool bdi_cap_account_dirty(struct backing_dev_info *bdi) 211e4ad08feSMiklos Szeredi { 212e4ad08feSMiklos Szeredi return !(bdi->capabilities & BDI_CAP_NO_ACCT_DIRTY); 213e4ad08feSMiklos Szeredi } 2141da177e4SLinus Torvalds 215e4ad08feSMiklos Szeredi static inline bool bdi_cap_account_writeback(struct backing_dev_info *bdi) 216e4ad08feSMiklos Szeredi { 217e4ad08feSMiklos Szeredi /* Paranoia: BDI_CAP_NO_WRITEBACK implies BDI_CAP_NO_ACCT_WB */ 218e4ad08feSMiklos Szeredi return !(bdi->capabilities & (BDI_CAP_NO_ACCT_WB | 219e4ad08feSMiklos Szeredi BDI_CAP_NO_WRITEBACK)); 220e4ad08feSMiklos Szeredi } 2211da177e4SLinus Torvalds 222e4ad08feSMiklos Szeredi static inline bool mapping_cap_writeback_dirty(struct address_space *mapping) 223e4ad08feSMiklos Szeredi { 224de1414a6SChristoph Hellwig return bdi_cap_writeback_dirty(inode_to_bdi(mapping->host)); 225e4ad08feSMiklos Szeredi } 226e4ad08feSMiklos Szeredi 227e4ad08feSMiklos Szeredi static inline bool mapping_cap_account_dirty(struct address_space *mapping) 228e4ad08feSMiklos Szeredi { 229de1414a6SChristoph Hellwig return bdi_cap_account_dirty(inode_to_bdi(mapping->host)); 230e4ad08feSMiklos Szeredi } 2311da177e4SLinus Torvalds 23203ba3782SJens Axboe static inline int bdi_sched_wait(void *word) 23303ba3782SJens Axboe { 23403ba3782SJens Axboe schedule(); 23503ba3782SJens Axboe return 0; 23603ba3782SJens Axboe } 23703ba3782SJens Axboe 23889e9b9e0STejun Heo #ifdef CONFIG_CGROUP_WRITEBACK 23989e9b9e0STejun Heo 24052ebea74STejun Heo struct bdi_writeback_congested * 24152ebea74STejun Heo wb_congested_get_create(struct backing_dev_info *bdi, int blkcg_id, gfp_t gfp); 24252ebea74STejun Heo void wb_congested_put(struct bdi_writeback_congested *congested); 24352ebea74STejun Heo struct bdi_writeback *wb_get_create(struct backing_dev_info *bdi, 24452ebea74STejun Heo struct cgroup_subsys_state *memcg_css, 24552ebea74STejun Heo gfp_t gfp); 24652ebea74STejun Heo void wb_memcg_offline(struct mem_cgroup *memcg); 24752ebea74STejun Heo void wb_blkcg_offline(struct blkcg *blkcg); 248703c2708STejun Heo int inode_congested(struct inode *inode, int cong_bits); 24952ebea74STejun Heo 25089e9b9e0STejun Heo /** 25189e9b9e0STejun Heo * inode_cgwb_enabled - test whether cgroup writeback is enabled on an inode 25289e9b9e0STejun Heo * @inode: inode of interest 25389e9b9e0STejun Heo * 25489e9b9e0STejun Heo * cgroup writeback requires support from both the bdi and filesystem. 25589e9b9e0STejun Heo * Test whether @inode has both. 25689e9b9e0STejun Heo */ 25789e9b9e0STejun Heo static inline bool inode_cgwb_enabled(struct inode *inode) 25889e9b9e0STejun Heo { 25989e9b9e0STejun Heo struct backing_dev_info *bdi = inode_to_bdi(inode); 26089e9b9e0STejun Heo 26189e9b9e0STejun Heo return bdi_cap_account_dirty(bdi) && 26289e9b9e0STejun Heo (bdi->capabilities & BDI_CAP_CGROUP_WRITEBACK) && 26389e9b9e0STejun Heo (inode->i_sb->s_type->fs_flags & FS_CGROUP_WRITEBACK); 26489e9b9e0STejun Heo } 26589e9b9e0STejun Heo 26652ebea74STejun Heo /** 26752ebea74STejun Heo * wb_find_current - find wb for %current on a bdi 26852ebea74STejun Heo * @bdi: bdi of interest 26952ebea74STejun Heo * 27052ebea74STejun Heo * Find the wb of @bdi which matches both the memcg and blkcg of %current. 27152ebea74STejun Heo * Must be called under rcu_read_lock() which protects the returend wb. 27252ebea74STejun Heo * NULL if not found. 27352ebea74STejun Heo */ 27452ebea74STejun Heo static inline struct bdi_writeback *wb_find_current(struct backing_dev_info *bdi) 27552ebea74STejun Heo { 27652ebea74STejun Heo struct cgroup_subsys_state *memcg_css; 27752ebea74STejun Heo struct bdi_writeback *wb; 27852ebea74STejun Heo 27952ebea74STejun Heo memcg_css = task_css(current, memory_cgrp_id); 28052ebea74STejun Heo if (!memcg_css->parent) 28152ebea74STejun Heo return &bdi->wb; 28252ebea74STejun Heo 28352ebea74STejun Heo wb = radix_tree_lookup(&bdi->cgwb_tree, memcg_css->id); 28452ebea74STejun Heo 28552ebea74STejun Heo /* 28652ebea74STejun Heo * %current's blkcg equals the effective blkcg of its memcg. No 28752ebea74STejun Heo * need to use the relatively expensive cgroup_get_e_css(). 28852ebea74STejun Heo */ 28952ebea74STejun Heo if (likely(wb && wb->blkcg_css == task_css(current, blkio_cgrp_id))) 29052ebea74STejun Heo return wb; 29152ebea74STejun Heo return NULL; 29252ebea74STejun Heo } 29352ebea74STejun Heo 29452ebea74STejun Heo /** 29552ebea74STejun Heo * wb_get_create_current - get or create wb for %current on a bdi 29652ebea74STejun Heo * @bdi: bdi of interest 29752ebea74STejun Heo * @gfp: allocation mask 29852ebea74STejun Heo * 29952ebea74STejun Heo * Equivalent to wb_get_create() on %current's memcg. This function is 30052ebea74STejun Heo * called from a relatively hot path and optimizes the common cases using 30152ebea74STejun Heo * wb_find_current(). 30252ebea74STejun Heo */ 30352ebea74STejun Heo static inline struct bdi_writeback * 30452ebea74STejun Heo wb_get_create_current(struct backing_dev_info *bdi, gfp_t gfp) 30552ebea74STejun Heo { 30652ebea74STejun Heo struct bdi_writeback *wb; 30752ebea74STejun Heo 30852ebea74STejun Heo rcu_read_lock(); 30952ebea74STejun Heo wb = wb_find_current(bdi); 31052ebea74STejun Heo if (wb && unlikely(!wb_tryget(wb))) 31152ebea74STejun Heo wb = NULL; 31252ebea74STejun Heo rcu_read_unlock(); 31352ebea74STejun Heo 31452ebea74STejun Heo if (unlikely(!wb)) { 31552ebea74STejun Heo struct cgroup_subsys_state *memcg_css; 31652ebea74STejun Heo 31752ebea74STejun Heo memcg_css = task_get_css(current, memory_cgrp_id); 31852ebea74STejun Heo wb = wb_get_create(bdi, memcg_css, gfp); 31952ebea74STejun Heo css_put(memcg_css); 32052ebea74STejun Heo } 32152ebea74STejun Heo return wb; 32252ebea74STejun Heo } 32352ebea74STejun Heo 32452ebea74STejun Heo /** 325*aaa2cacfSTejun Heo * inode_to_wb_is_valid - test whether an inode has a wb associated 326*aaa2cacfSTejun Heo * @inode: inode of interest 327*aaa2cacfSTejun Heo * 328*aaa2cacfSTejun Heo * Returns %true if @inode has a wb associated. May be called without any 329*aaa2cacfSTejun Heo * locking. 330*aaa2cacfSTejun Heo */ 331*aaa2cacfSTejun Heo static inline bool inode_to_wb_is_valid(struct inode *inode) 332*aaa2cacfSTejun Heo { 333*aaa2cacfSTejun Heo return inode->i_wb; 334*aaa2cacfSTejun Heo } 335*aaa2cacfSTejun Heo 336*aaa2cacfSTejun Heo /** 33752ebea74STejun Heo * inode_to_wb - determine the wb of an inode 33852ebea74STejun Heo * @inode: inode of interest 33952ebea74STejun Heo * 340*aaa2cacfSTejun Heo * Returns the wb @inode is currently associated with. The caller must be 341*aaa2cacfSTejun Heo * holding either @inode->i_lock, @inode->i_mapping->tree_lock, or the 342*aaa2cacfSTejun Heo * associated wb's list_lock. 34352ebea74STejun Heo */ 34452ebea74STejun Heo static inline struct bdi_writeback *inode_to_wb(struct inode *inode) 34552ebea74STejun Heo { 346*aaa2cacfSTejun Heo #ifdef CONFIG_LOCKDEP 347*aaa2cacfSTejun Heo WARN_ON_ONCE(debug_locks && 348*aaa2cacfSTejun Heo (!lockdep_is_held(&inode->i_lock) && 349*aaa2cacfSTejun Heo !lockdep_is_held(&inode->i_mapping->tree_lock) && 350*aaa2cacfSTejun Heo !lockdep_is_held(&inode->i_wb->list_lock))); 351*aaa2cacfSTejun Heo #endif 35252ebea74STejun Heo return inode->i_wb; 35352ebea74STejun Heo } 35452ebea74STejun Heo 355682aa8e1STejun Heo /** 356682aa8e1STejun Heo * unlocked_inode_to_wb_begin - begin unlocked inode wb access transaction 357682aa8e1STejun Heo * @inode: target inode 358682aa8e1STejun Heo * @lockedp: temp bool output param, to be passed to the end function 359682aa8e1STejun Heo * 360682aa8e1STejun Heo * The caller wants to access the wb associated with @inode but isn't 361682aa8e1STejun Heo * holding inode->i_lock, mapping->tree_lock or wb->list_lock. This 362682aa8e1STejun Heo * function determines the wb associated with @inode and ensures that the 363682aa8e1STejun Heo * association doesn't change until the transaction is finished with 364682aa8e1STejun Heo * unlocked_inode_to_wb_end(). 365682aa8e1STejun Heo * 366682aa8e1STejun Heo * The caller must call unlocked_inode_to_wb_end() with *@lockdep 367682aa8e1STejun Heo * afterwards and can't sleep during transaction. IRQ may or may not be 368682aa8e1STejun Heo * disabled on return. 369682aa8e1STejun Heo */ 370682aa8e1STejun Heo static inline struct bdi_writeback * 371682aa8e1STejun Heo unlocked_inode_to_wb_begin(struct inode *inode, bool *lockedp) 372682aa8e1STejun Heo { 373682aa8e1STejun Heo rcu_read_lock(); 374682aa8e1STejun Heo 375682aa8e1STejun Heo /* 376682aa8e1STejun Heo * Paired with store_release in inode_switch_wb_work_fn() and 377682aa8e1STejun Heo * ensures that we see the new wb if we see cleared I_WB_SWITCH. 378682aa8e1STejun Heo */ 379682aa8e1STejun Heo *lockedp = smp_load_acquire(&inode->i_state) & I_WB_SWITCH; 380682aa8e1STejun Heo 381682aa8e1STejun Heo if (unlikely(*lockedp)) 382682aa8e1STejun Heo spin_lock_irq(&inode->i_mapping->tree_lock); 383*aaa2cacfSTejun Heo 384*aaa2cacfSTejun Heo /* 385*aaa2cacfSTejun Heo * Protected by either !I_WB_SWITCH + rcu_read_lock() or tree_lock. 386*aaa2cacfSTejun Heo * inode_to_wb() will bark. Deref directly. 387*aaa2cacfSTejun Heo */ 388*aaa2cacfSTejun Heo return inode->i_wb; 389682aa8e1STejun Heo } 390682aa8e1STejun Heo 391682aa8e1STejun Heo /** 392682aa8e1STejun Heo * unlocked_inode_to_wb_end - end inode wb access transaction 393682aa8e1STejun Heo * @inode: target inode 394682aa8e1STejun Heo * @locked: *@lockedp from unlocked_inode_to_wb_begin() 395682aa8e1STejun Heo */ 396682aa8e1STejun Heo static inline void unlocked_inode_to_wb_end(struct inode *inode, bool locked) 397682aa8e1STejun Heo { 398682aa8e1STejun Heo if (unlikely(locked)) 399682aa8e1STejun Heo spin_unlock_irq(&inode->i_mapping->tree_lock); 400682aa8e1STejun Heo 401682aa8e1STejun Heo rcu_read_unlock(); 402682aa8e1STejun Heo } 403682aa8e1STejun Heo 404ebe41ab0STejun Heo struct wb_iter { 405ebe41ab0STejun Heo int start_blkcg_id; 406ebe41ab0STejun Heo struct radix_tree_iter tree_iter; 407ebe41ab0STejun Heo void **slot; 408ebe41ab0STejun Heo }; 409ebe41ab0STejun Heo 410ebe41ab0STejun Heo static inline struct bdi_writeback *__wb_iter_next(struct wb_iter *iter, 411ebe41ab0STejun Heo struct backing_dev_info *bdi) 412ebe41ab0STejun Heo { 413ebe41ab0STejun Heo struct radix_tree_iter *titer = &iter->tree_iter; 414ebe41ab0STejun Heo 415ebe41ab0STejun Heo WARN_ON_ONCE(!rcu_read_lock_held()); 416ebe41ab0STejun Heo 417ebe41ab0STejun Heo if (iter->start_blkcg_id >= 0) { 418ebe41ab0STejun Heo iter->slot = radix_tree_iter_init(titer, iter->start_blkcg_id); 419ebe41ab0STejun Heo iter->start_blkcg_id = -1; 420ebe41ab0STejun Heo } else { 421ebe41ab0STejun Heo iter->slot = radix_tree_next_slot(iter->slot, titer, 0); 422ebe41ab0STejun Heo } 423ebe41ab0STejun Heo 424ebe41ab0STejun Heo if (!iter->slot) 425ebe41ab0STejun Heo iter->slot = radix_tree_next_chunk(&bdi->cgwb_tree, titer, 0); 426ebe41ab0STejun Heo if (iter->slot) 427ebe41ab0STejun Heo return *iter->slot; 428ebe41ab0STejun Heo return NULL; 429ebe41ab0STejun Heo } 430ebe41ab0STejun Heo 431ebe41ab0STejun Heo static inline struct bdi_writeback *__wb_iter_init(struct wb_iter *iter, 432ebe41ab0STejun Heo struct backing_dev_info *bdi, 433ebe41ab0STejun Heo int start_blkcg_id) 434ebe41ab0STejun Heo { 435ebe41ab0STejun Heo iter->start_blkcg_id = start_blkcg_id; 436ebe41ab0STejun Heo 437ebe41ab0STejun Heo if (start_blkcg_id) 438ebe41ab0STejun Heo return __wb_iter_next(iter, bdi); 439ebe41ab0STejun Heo else 440ebe41ab0STejun Heo return &bdi->wb; 441ebe41ab0STejun Heo } 442ebe41ab0STejun Heo 443ebe41ab0STejun Heo /** 444ebe41ab0STejun Heo * bdi_for_each_wb - walk all wb's of a bdi in ascending blkcg ID order 445ebe41ab0STejun Heo * @wb_cur: cursor struct bdi_writeback pointer 446ebe41ab0STejun Heo * @bdi: bdi to walk wb's of 447ebe41ab0STejun Heo * @iter: pointer to struct wb_iter to be used as iteration buffer 448ebe41ab0STejun Heo * @start_blkcg_id: blkcg ID to start iteration from 449ebe41ab0STejun Heo * 450ebe41ab0STejun Heo * Iterate @wb_cur through the wb's (bdi_writeback's) of @bdi in ascending 451ebe41ab0STejun Heo * blkcg ID order starting from @start_blkcg_id. @iter is struct wb_iter 452ebe41ab0STejun Heo * to be used as temp storage during iteration. rcu_read_lock() must be 453ebe41ab0STejun Heo * held throughout iteration. 454ebe41ab0STejun Heo */ 455ebe41ab0STejun Heo #define bdi_for_each_wb(wb_cur, bdi, iter, start_blkcg_id) \ 456ebe41ab0STejun Heo for ((wb_cur) = __wb_iter_init(iter, bdi, start_blkcg_id); \ 457ebe41ab0STejun Heo (wb_cur); (wb_cur) = __wb_iter_next(iter, bdi)) 458ebe41ab0STejun Heo 45989e9b9e0STejun Heo #else /* CONFIG_CGROUP_WRITEBACK */ 46089e9b9e0STejun Heo 46189e9b9e0STejun Heo static inline bool inode_cgwb_enabled(struct inode *inode) 46289e9b9e0STejun Heo { 46389e9b9e0STejun Heo return false; 46489e9b9e0STejun Heo } 46589e9b9e0STejun Heo 46652ebea74STejun Heo static inline struct bdi_writeback_congested * 46752ebea74STejun Heo wb_congested_get_create(struct backing_dev_info *bdi, int blkcg_id, gfp_t gfp) 46852ebea74STejun Heo { 46952ebea74STejun Heo return bdi->wb.congested; 47052ebea74STejun Heo } 47152ebea74STejun Heo 47252ebea74STejun Heo static inline void wb_congested_put(struct bdi_writeback_congested *congested) 47352ebea74STejun Heo { 47452ebea74STejun Heo } 47552ebea74STejun Heo 47652ebea74STejun Heo static inline struct bdi_writeback *wb_find_current(struct backing_dev_info *bdi) 47752ebea74STejun Heo { 47852ebea74STejun Heo return &bdi->wb; 47952ebea74STejun Heo } 48052ebea74STejun Heo 48152ebea74STejun Heo static inline struct bdi_writeback * 48252ebea74STejun Heo wb_get_create_current(struct backing_dev_info *bdi, gfp_t gfp) 48352ebea74STejun Heo { 48452ebea74STejun Heo return &bdi->wb; 48552ebea74STejun Heo } 48652ebea74STejun Heo 487*aaa2cacfSTejun Heo static inline bool inode_to_wb_is_valid(struct inode *inode) 488*aaa2cacfSTejun Heo { 489*aaa2cacfSTejun Heo return true; 490*aaa2cacfSTejun Heo } 491*aaa2cacfSTejun Heo 49252ebea74STejun Heo static inline struct bdi_writeback *inode_to_wb(struct inode *inode) 49352ebea74STejun Heo { 49452ebea74STejun Heo return &inode_to_bdi(inode)->wb; 49552ebea74STejun Heo } 49652ebea74STejun Heo 497682aa8e1STejun Heo static inline struct bdi_writeback * 498682aa8e1STejun Heo unlocked_inode_to_wb_begin(struct inode *inode, bool *lockedp) 499682aa8e1STejun Heo { 500682aa8e1STejun Heo return inode_to_wb(inode); 501682aa8e1STejun Heo } 502682aa8e1STejun Heo 503682aa8e1STejun Heo static inline void unlocked_inode_to_wb_end(struct inode *inode, bool locked) 504682aa8e1STejun Heo { 505682aa8e1STejun Heo } 506682aa8e1STejun Heo 50752ebea74STejun Heo static inline void wb_memcg_offline(struct mem_cgroup *memcg) 50852ebea74STejun Heo { 50952ebea74STejun Heo } 51052ebea74STejun Heo 51152ebea74STejun Heo static inline void wb_blkcg_offline(struct blkcg *blkcg) 51252ebea74STejun Heo { 51352ebea74STejun Heo } 51452ebea74STejun Heo 515ebe41ab0STejun Heo struct wb_iter { 516ebe41ab0STejun Heo int next_id; 517ebe41ab0STejun Heo }; 518ebe41ab0STejun Heo 519ebe41ab0STejun Heo #define bdi_for_each_wb(wb_cur, bdi, iter, start_blkcg_id) \ 520ebe41ab0STejun Heo for ((iter)->next_id = (start_blkcg_id); \ 521ebe41ab0STejun Heo ({ (wb_cur) = !(iter)->next_id++ ? &(bdi)->wb : NULL; }); ) 522ebe41ab0STejun Heo 523703c2708STejun Heo static inline int inode_congested(struct inode *inode, int cong_bits) 524703c2708STejun Heo { 525703c2708STejun Heo return wb_congested(&inode_to_bdi(inode)->wb, cong_bits); 526703c2708STejun Heo } 527703c2708STejun Heo 52889e9b9e0STejun Heo #endif /* CONFIG_CGROUP_WRITEBACK */ 52989e9b9e0STejun Heo 530703c2708STejun Heo static inline int inode_read_congested(struct inode *inode) 531703c2708STejun Heo { 532703c2708STejun Heo return inode_congested(inode, 1 << WB_sync_congested); 533703c2708STejun Heo } 534703c2708STejun Heo 535703c2708STejun Heo static inline int inode_write_congested(struct inode *inode) 536703c2708STejun Heo { 537703c2708STejun Heo return inode_congested(inode, 1 << WB_async_congested); 538703c2708STejun Heo } 539703c2708STejun Heo 540703c2708STejun Heo static inline int inode_rw_congested(struct inode *inode) 541703c2708STejun Heo { 542703c2708STejun Heo return inode_congested(inode, (1 << WB_sync_congested) | 543703c2708STejun Heo (1 << WB_async_congested)); 544703c2708STejun Heo } 545703c2708STejun Heo 546ec8a6f26STejun Heo static inline int bdi_congested(struct backing_dev_info *bdi, int cong_bits) 547ec8a6f26STejun Heo { 548ec8a6f26STejun Heo return wb_congested(&bdi->wb, cong_bits); 549ec8a6f26STejun Heo } 550ec8a6f26STejun Heo 551ec8a6f26STejun Heo static inline int bdi_read_congested(struct backing_dev_info *bdi) 552ec8a6f26STejun Heo { 553ec8a6f26STejun Heo return bdi_congested(bdi, 1 << WB_sync_congested); 554ec8a6f26STejun Heo } 555ec8a6f26STejun Heo 556ec8a6f26STejun Heo static inline int bdi_write_congested(struct backing_dev_info *bdi) 557ec8a6f26STejun Heo { 558ec8a6f26STejun Heo return bdi_congested(bdi, 1 << WB_async_congested); 559ec8a6f26STejun Heo } 560ec8a6f26STejun Heo 561ec8a6f26STejun Heo static inline int bdi_rw_congested(struct backing_dev_info *bdi) 562ec8a6f26STejun Heo { 563ec8a6f26STejun Heo return bdi_congested(bdi, (1 << WB_sync_congested) | 564ec8a6f26STejun Heo (1 << WB_async_congested)); 565ec8a6f26STejun Heo } 566ec8a6f26STejun Heo 5671da177e4SLinus Torvalds #endif /* _LINUX_BACKING_DEV_H */ 568