xref: /linux-6.15/include/linux/backing-dev.h (revision b4caecd4)
11da177e4SLinus Torvalds /*
21da177e4SLinus Torvalds  * include/linux/backing-dev.h
31da177e4SLinus Torvalds  *
41da177e4SLinus Torvalds  * low-level device information and state which is propagated up through
51da177e4SLinus Torvalds  * to high-level code.
61da177e4SLinus Torvalds  */
71da177e4SLinus Torvalds 
81da177e4SLinus Torvalds #ifndef _LINUX_BACKING_DEV_H
91da177e4SLinus Torvalds #define _LINUX_BACKING_DEV_H
101da177e4SLinus Torvalds 
11b2e8fb6eSPeter Zijlstra #include <linux/percpu_counter.h>
12b2e8fb6eSPeter Zijlstra #include <linux/log2.h>
13eb608e3aSJan Kara #include <linux/flex_proportions.h>
14cf0ca9feSPeter Zijlstra #include <linux/kernel.h>
15e4ad08feSMiklos Szeredi #include <linux/fs.h>
1603ba3782SJens Axboe #include <linux/sched.h>
1731373d09SMatthew Garrett #include <linux/timer.h>
1803ba3782SJens Axboe #include <linux/writeback.h>
1960063497SArun Sharma #include <linux/atomic.h>
203965c9aeSWanpeng Li #include <linux/sysctl.h>
21839a8e86STejun Heo #include <linux/workqueue.h>
221da177e4SLinus Torvalds 
233fcfab16SAndrew Morton struct page;
24cf0ca9feSPeter Zijlstra struct device;
2576f1418bSMiklos Szeredi struct dentry;
263fcfab16SAndrew Morton 
271da177e4SLinus Torvalds /*
281da177e4SLinus Torvalds  * Bits in backing_dev_info.state
291da177e4SLinus Torvalds  */
301da177e4SLinus Torvalds enum bdi_state {
311faa16d2SJens Axboe 	BDI_async_congested,	/* The async (write) queue is getting full */
321faa16d2SJens Axboe 	BDI_sync_congested,	/* The sync queue is getting full */
33500b067cSJens Axboe 	BDI_registered,		/* bdi_register() was done */
3481d73a32SJan Kara 	BDI_writeback_running,	/* Writeback is in progress */
351da177e4SLinus Torvalds };
361da177e4SLinus Torvalds 
371da177e4SLinus Torvalds typedef int (congested_fn)(void *, int);
381da177e4SLinus Torvalds 
39b2e8fb6eSPeter Zijlstra enum bdi_stat_item {
40c9e51e41SPeter Zijlstra 	BDI_RECLAIMABLE,
4169cb51d1SPeter Zijlstra 	BDI_WRITEBACK,
42c8e28ce0SWu Fengguang 	BDI_DIRTIED,
43f7d2b1ecSJan Kara 	BDI_WRITTEN,
44b2e8fb6eSPeter Zijlstra 	NR_BDI_STAT_ITEMS
45b2e8fb6eSPeter Zijlstra };
46b2e8fb6eSPeter Zijlstra 
47b2e8fb6eSPeter Zijlstra #define BDI_STAT_BATCH (8*(1+ilog2(nr_cpu_ids)))
48b2e8fb6eSPeter Zijlstra 
4903ba3782SJens Axboe struct bdi_writeback {
5003ba3782SJens Axboe 	struct backing_dev_info *bdi;	/* our parent bdi */
5103ba3782SJens Axboe 
5203ba3782SJens Axboe 	unsigned long last_old_flush;	/* last old data flush */
5303ba3782SJens Axboe 
54839a8e86STejun Heo 	struct delayed_work dwork;	/* work item used for writeback */
5503ba3782SJens Axboe 	struct list_head b_dirty;	/* dirty inodes */
5603ba3782SJens Axboe 	struct list_head b_io;		/* parked for writeback */
5703ba3782SJens Axboe 	struct list_head b_more_io;	/* parked for more writeback */
58f758eeabSChristoph Hellwig 	spinlock_t list_lock;		/* protects the b_* lists */
5903ba3782SJens Axboe };
6003ba3782SJens Axboe 
611da177e4SLinus Torvalds struct backing_dev_info {
6266f3b8e2SJens Axboe 	struct list_head bdi_list;
631da177e4SLinus Torvalds 	unsigned long ra_pages;	/* max readahead in PAGE_CACHE_SIZE units */
641da177e4SLinus Torvalds 	unsigned long state;	/* Always use atomic bitops on this */
651da177e4SLinus Torvalds 	unsigned int capabilities; /* Device capabilities */
661da177e4SLinus Torvalds 	congested_fn *congested_fn; /* Function pointer if device is md/dm */
671da177e4SLinus Torvalds 	void *congested_data;	/* Pointer to aux data for congested func */
68b2e8fb6eSPeter Zijlstra 
69d993831fSJens Axboe 	char *name;
70d993831fSJens Axboe 
71b2e8fb6eSPeter Zijlstra 	struct percpu_counter bdi_stat[NR_BDI_STAT_ITEMS];
7204fbfdc1SPeter Zijlstra 
73e98be2d5SWu Fengguang 	unsigned long bw_time_stamp;	/* last time write bw is updated */
74be3ffa27SWu Fengguang 	unsigned long dirtied_stamp;
75e98be2d5SWu Fengguang 	unsigned long written_stamp;	/* pages written at bw_time_stamp */
76e98be2d5SWu Fengguang 	unsigned long write_bandwidth;	/* the estimated write bandwidth */
77e98be2d5SWu Fengguang 	unsigned long avg_write_bandwidth; /* further smoothed write bw */
78e98be2d5SWu Fengguang 
79be3ffa27SWu Fengguang 	/*
80be3ffa27SWu Fengguang 	 * The base dirty throttle rate, re-calculated on every 200ms.
81be3ffa27SWu Fengguang 	 * All the bdi tasks' dirty rate will be curbed under it.
827381131cSWu Fengguang 	 * @dirty_ratelimit tracks the estimated @balanced_dirty_ratelimit
837381131cSWu Fengguang 	 * in small steps and is much more smooth/stable than the latter.
84be3ffa27SWu Fengguang 	 */
85be3ffa27SWu Fengguang 	unsigned long dirty_ratelimit;
867381131cSWu Fengguang 	unsigned long balanced_dirty_ratelimit;
87be3ffa27SWu Fengguang 
88eb608e3aSJan Kara 	struct fprop_local_percpu completions;
8904fbfdc1SPeter Zijlstra 	int dirty_exceeded;
90cf0ca9feSPeter Zijlstra 
91189d3c4aSPeter Zijlstra 	unsigned int min_ratio;
92a42dde04SPeter Zijlstra 	unsigned int max_ratio, max_prop_frac;
93189d3c4aSPeter Zijlstra 
9403ba3782SJens Axboe 	struct bdi_writeback wb;  /* default writeback info for this bdi */
955acda9d1SJan Kara 	spinlock_t wb_lock;	  /* protects work_list & wb.dwork scheduling */
9676f1418bSMiklos Szeredi 
9703ba3782SJens Axboe 	struct list_head work_list;
9803ba3782SJens Axboe 
9903ba3782SJens Axboe 	struct device *dev;
10066f3b8e2SJens Axboe 
10131373d09SMatthew Garrett 	struct timer_list laptop_mode_wb_timer;
10231373d09SMatthew Garrett 
10376f1418bSMiklos Szeredi #ifdef CONFIG_DEBUG_FS
10476f1418bSMiklos Szeredi 	struct dentry *debug_dir;
10576f1418bSMiklos Szeredi 	struct dentry *debug_stats;
10676f1418bSMiklos Szeredi #endif
1071da177e4SLinus Torvalds };
1081da177e4SLinus Torvalds 
1098077c0d9SMikulas Patocka int __must_check bdi_init(struct backing_dev_info *bdi);
110b2e8fb6eSPeter Zijlstra void bdi_destroy(struct backing_dev_info *bdi);
111b2e8fb6eSPeter Zijlstra 
112d2cc4ddeSJoe Perches __printf(3, 4)
113cf0ca9feSPeter Zijlstra int bdi_register(struct backing_dev_info *bdi, struct device *parent,
114cf0ca9feSPeter Zijlstra 		const char *fmt, ...);
115cf0ca9feSPeter Zijlstra int bdi_register_dev(struct backing_dev_info *bdi, dev_t dev);
116cf0ca9feSPeter Zijlstra void bdi_unregister(struct backing_dev_info *bdi);
117*b4caecd4SChristoph Hellwig int __must_check bdi_setup_and_register(struct backing_dev_info *, char *);
1180e175a18SCurt Wohlgemuth void bdi_start_writeback(struct backing_dev_info *bdi, long nr_pages,
1190e175a18SCurt Wohlgemuth 			enum wb_reason reason);
120c5444198SChristoph Hellwig void bdi_start_background_writeback(struct backing_dev_info *bdi);
121839a8e86STejun Heo void bdi_writeback_workfn(struct work_struct *work);
12203ba3782SJens Axboe int bdi_has_dirty_io(struct backing_dev_info *bdi);
1236467716aSArtem Bityutskiy void bdi_wakeup_thread_delayed(struct backing_dev_info *bdi);
124cf0ca9feSPeter Zijlstra 
12503ba3782SJens Axboe extern spinlock_t bdi_lock;
12666f3b8e2SJens Axboe extern struct list_head bdi_list;
12766f3b8e2SJens Axboe 
128839a8e86STejun Heo extern struct workqueue_struct *bdi_wq;
129839a8e86STejun Heo 
13003ba3782SJens Axboe static inline int wb_has_dirty_io(struct bdi_writeback *wb)
13103ba3782SJens Axboe {
13203ba3782SJens Axboe 	return !list_empty(&wb->b_dirty) ||
13303ba3782SJens Axboe 	       !list_empty(&wb->b_io) ||
13403ba3782SJens Axboe 	       !list_empty(&wb->b_more_io);
13503ba3782SJens Axboe }
13603ba3782SJens Axboe 
137b2e8fb6eSPeter Zijlstra static inline void __add_bdi_stat(struct backing_dev_info *bdi,
138b2e8fb6eSPeter Zijlstra 		enum bdi_stat_item item, s64 amount)
139e0bf68ddSPeter Zijlstra {
140b2e8fb6eSPeter Zijlstra 	__percpu_counter_add(&bdi->bdi_stat[item], amount, BDI_STAT_BATCH);
141e0bf68ddSPeter Zijlstra }
142e0bf68ddSPeter Zijlstra 
143b2e8fb6eSPeter Zijlstra static inline void __inc_bdi_stat(struct backing_dev_info *bdi,
144b2e8fb6eSPeter Zijlstra 		enum bdi_stat_item item)
145e0bf68ddSPeter Zijlstra {
146b2e8fb6eSPeter Zijlstra 	__add_bdi_stat(bdi, item, 1);
147b2e8fb6eSPeter Zijlstra }
148b2e8fb6eSPeter Zijlstra 
149b2e8fb6eSPeter Zijlstra static inline void inc_bdi_stat(struct backing_dev_info *bdi,
150b2e8fb6eSPeter Zijlstra 		enum bdi_stat_item item)
151b2e8fb6eSPeter Zijlstra {
152b2e8fb6eSPeter Zijlstra 	unsigned long flags;
153b2e8fb6eSPeter Zijlstra 
154b2e8fb6eSPeter Zijlstra 	local_irq_save(flags);
155b2e8fb6eSPeter Zijlstra 	__inc_bdi_stat(bdi, item);
156b2e8fb6eSPeter Zijlstra 	local_irq_restore(flags);
157b2e8fb6eSPeter Zijlstra }
158b2e8fb6eSPeter Zijlstra 
159b2e8fb6eSPeter Zijlstra static inline void __dec_bdi_stat(struct backing_dev_info *bdi,
160b2e8fb6eSPeter Zijlstra 		enum bdi_stat_item item)
161b2e8fb6eSPeter Zijlstra {
162b2e8fb6eSPeter Zijlstra 	__add_bdi_stat(bdi, item, -1);
163b2e8fb6eSPeter Zijlstra }
164b2e8fb6eSPeter Zijlstra 
165b2e8fb6eSPeter Zijlstra static inline void dec_bdi_stat(struct backing_dev_info *bdi,
166b2e8fb6eSPeter Zijlstra 		enum bdi_stat_item item)
167b2e8fb6eSPeter Zijlstra {
168b2e8fb6eSPeter Zijlstra 	unsigned long flags;
169b2e8fb6eSPeter Zijlstra 
170b2e8fb6eSPeter Zijlstra 	local_irq_save(flags);
171b2e8fb6eSPeter Zijlstra 	__dec_bdi_stat(bdi, item);
172b2e8fb6eSPeter Zijlstra 	local_irq_restore(flags);
173b2e8fb6eSPeter Zijlstra }
174b2e8fb6eSPeter Zijlstra 
175b2e8fb6eSPeter Zijlstra static inline s64 bdi_stat(struct backing_dev_info *bdi,
176b2e8fb6eSPeter Zijlstra 		enum bdi_stat_item item)
177b2e8fb6eSPeter Zijlstra {
178b2e8fb6eSPeter Zijlstra 	return percpu_counter_read_positive(&bdi->bdi_stat[item]);
179b2e8fb6eSPeter Zijlstra }
180b2e8fb6eSPeter Zijlstra 
181b2e8fb6eSPeter Zijlstra static inline s64 __bdi_stat_sum(struct backing_dev_info *bdi,
182b2e8fb6eSPeter Zijlstra 		enum bdi_stat_item item)
183b2e8fb6eSPeter Zijlstra {
184b2e8fb6eSPeter Zijlstra 	return percpu_counter_sum_positive(&bdi->bdi_stat[item]);
185b2e8fb6eSPeter Zijlstra }
186b2e8fb6eSPeter Zijlstra 
187b2e8fb6eSPeter Zijlstra static inline s64 bdi_stat_sum(struct backing_dev_info *bdi,
188b2e8fb6eSPeter Zijlstra 		enum bdi_stat_item item)
189b2e8fb6eSPeter Zijlstra {
190b2e8fb6eSPeter Zijlstra 	s64 sum;
191b2e8fb6eSPeter Zijlstra 	unsigned long flags;
192b2e8fb6eSPeter Zijlstra 
193b2e8fb6eSPeter Zijlstra 	local_irq_save(flags);
194b2e8fb6eSPeter Zijlstra 	sum = __bdi_stat_sum(bdi, item);
195b2e8fb6eSPeter Zijlstra 	local_irq_restore(flags);
196b2e8fb6eSPeter Zijlstra 
197b2e8fb6eSPeter Zijlstra 	return sum;
198b2e8fb6eSPeter Zijlstra }
199b2e8fb6eSPeter Zijlstra 
200dd5656e5SMiklos Szeredi extern void bdi_writeout_inc(struct backing_dev_info *bdi);
201dd5656e5SMiklos Szeredi 
202b2e8fb6eSPeter Zijlstra /*
203b2e8fb6eSPeter Zijlstra  * maximal error of a stat counter.
204b2e8fb6eSPeter Zijlstra  */
205b2e8fb6eSPeter Zijlstra static inline unsigned long bdi_stat_error(struct backing_dev_info *bdi)
206b2e8fb6eSPeter Zijlstra {
207b2e8fb6eSPeter Zijlstra #ifdef CONFIG_SMP
208b2e8fb6eSPeter Zijlstra 	return nr_cpu_ids * BDI_STAT_BATCH;
209b2e8fb6eSPeter Zijlstra #else
210b2e8fb6eSPeter Zijlstra 	return 1;
211b2e8fb6eSPeter Zijlstra #endif
212e0bf68ddSPeter Zijlstra }
2131da177e4SLinus Torvalds 
214189d3c4aSPeter Zijlstra int bdi_set_min_ratio(struct backing_dev_info *bdi, unsigned int min_ratio);
215a42dde04SPeter Zijlstra int bdi_set_max_ratio(struct backing_dev_info *bdi, unsigned int max_ratio);
216189d3c4aSPeter Zijlstra 
2171da177e4SLinus Torvalds /*
2181da177e4SLinus Torvalds  * Flags in backing_dev_info::capability
219e4ad08feSMiklos Szeredi  *
220e4ad08feSMiklos Szeredi  * The first three flags control whether dirty pages will contribute to the
2211da177e4SLinus Torvalds  * VM's accounting and whether writepages() should be called for dirty pages
2221da177e4SLinus Torvalds  * (something that would not, for example, be appropriate for ramfs)
223e4ad08feSMiklos Szeredi  *
224e4ad08feSMiklos Szeredi  * WARNING: these flags are closely related and should not normally be
225e4ad08feSMiklos Szeredi  * used separately.  The BDI_CAP_NO_ACCT_AND_WRITEBACK combines these
226e4ad08feSMiklos Szeredi  * three flags into a single convenience macro.
227e4ad08feSMiklos Szeredi  *
228e4ad08feSMiklos Szeredi  * BDI_CAP_NO_ACCT_DIRTY:  Dirty pages shouldn't contribute to accounting
229e4ad08feSMiklos Szeredi  * BDI_CAP_NO_WRITEBACK:   Don't write pages back
230e4ad08feSMiklos Szeredi  * BDI_CAP_NO_ACCT_WB:     Don't automatically account writeback pages
2315a537485SMaxim Patlasov  * BDI_CAP_STRICTLIMIT:    Keep number of dirty pages below bdi threshold.
2321da177e4SLinus Torvalds  */
233e4ad08feSMiklos Szeredi #define BDI_CAP_NO_ACCT_DIRTY	0x00000001
234e4ad08feSMiklos Szeredi #define BDI_CAP_NO_WRITEBACK	0x00000002
235*b4caecd4SChristoph Hellwig #define BDI_CAP_NO_ACCT_WB	0x00000004
236*b4caecd4SChristoph Hellwig #define BDI_CAP_STABLE_WRITES	0x00000008
237*b4caecd4SChristoph Hellwig #define BDI_CAP_STRICTLIMIT	0x00000010
2381da177e4SLinus Torvalds 
239e4ad08feSMiklos Szeredi #define BDI_CAP_NO_ACCT_AND_WRITEBACK \
240e4ad08feSMiklos Szeredi 	(BDI_CAP_NO_WRITEBACK | BDI_CAP_NO_ACCT_DIRTY | BDI_CAP_NO_ACCT_WB)
241e4ad08feSMiklos Szeredi 
2421da177e4SLinus Torvalds extern struct backing_dev_info default_backing_dev_info;
2435129a469SJörn Engel extern struct backing_dev_info noop_backing_dev_info;
2441da177e4SLinus Torvalds 
2451da177e4SLinus Torvalds int writeback_in_progress(struct backing_dev_info *bdi);
2461da177e4SLinus Torvalds 
2471da177e4SLinus Torvalds static inline int bdi_congested(struct backing_dev_info *bdi, int bdi_bits)
2481da177e4SLinus Torvalds {
2491da177e4SLinus Torvalds 	if (bdi->congested_fn)
2501da177e4SLinus Torvalds 		return bdi->congested_fn(bdi->congested_data, bdi_bits);
2511da177e4SLinus Torvalds 	return (bdi->state & bdi_bits);
2521da177e4SLinus Torvalds }
2531da177e4SLinus Torvalds 
2541da177e4SLinus Torvalds static inline int bdi_read_congested(struct backing_dev_info *bdi)
2551da177e4SLinus Torvalds {
2561faa16d2SJens Axboe 	return bdi_congested(bdi, 1 << BDI_sync_congested);
2571da177e4SLinus Torvalds }
2581da177e4SLinus Torvalds 
2591da177e4SLinus Torvalds static inline int bdi_write_congested(struct backing_dev_info *bdi)
2601da177e4SLinus Torvalds {
2611faa16d2SJens Axboe 	return bdi_congested(bdi, 1 << BDI_async_congested);
2621da177e4SLinus Torvalds }
2631da177e4SLinus Torvalds 
2641da177e4SLinus Torvalds static inline int bdi_rw_congested(struct backing_dev_info *bdi)
2651da177e4SLinus Torvalds {
2661faa16d2SJens Axboe 	return bdi_congested(bdi, (1 << BDI_sync_congested) |
2671faa16d2SJens Axboe 				  (1 << BDI_async_congested));
2681da177e4SLinus Torvalds }
2691da177e4SLinus Torvalds 
270373c0a7eSTrond Myklebust enum {
271373c0a7eSTrond Myklebust 	BLK_RW_ASYNC	= 0,
272373c0a7eSTrond Myklebust 	BLK_RW_SYNC	= 1,
273373c0a7eSTrond Myklebust };
274373c0a7eSTrond Myklebust 
2758aa7e847SJens Axboe void clear_bdi_congested(struct backing_dev_info *bdi, int sync);
2768aa7e847SJens Axboe void set_bdi_congested(struct backing_dev_info *bdi, int sync);
2778aa7e847SJens Axboe long congestion_wait(int sync, long timeout);
2780e093d99SMel Gorman long wait_iff_congested(struct zone *zone, int sync, long timeout);
2793965c9aeSWanpeng Li int pdflush_proc_obsolete(struct ctl_table *table, int write,
2803965c9aeSWanpeng Li 		void __user *buffer, size_t *lenp, loff_t *ppos);
2811da177e4SLinus Torvalds 
2827d311cdaSDarrick J. Wong static inline bool bdi_cap_stable_pages_required(struct backing_dev_info *bdi)
2837d311cdaSDarrick J. Wong {
2847d311cdaSDarrick J. Wong 	return bdi->capabilities & BDI_CAP_STABLE_WRITES;
2857d311cdaSDarrick J. Wong }
2867d311cdaSDarrick J. Wong 
287e4ad08feSMiklos Szeredi static inline bool bdi_cap_writeback_dirty(struct backing_dev_info *bdi)
288e4ad08feSMiklos Szeredi {
289e4ad08feSMiklos Szeredi 	return !(bdi->capabilities & BDI_CAP_NO_WRITEBACK);
290e4ad08feSMiklos Szeredi }
2911da177e4SLinus Torvalds 
292e4ad08feSMiklos Szeredi static inline bool bdi_cap_account_dirty(struct backing_dev_info *bdi)
293e4ad08feSMiklos Szeredi {
294e4ad08feSMiklos Szeredi 	return !(bdi->capabilities & BDI_CAP_NO_ACCT_DIRTY);
295e4ad08feSMiklos Szeredi }
2961da177e4SLinus Torvalds 
297e4ad08feSMiklos Szeredi static inline bool bdi_cap_account_writeback(struct backing_dev_info *bdi)
298e4ad08feSMiklos Szeredi {
299e4ad08feSMiklos Szeredi 	/* Paranoia: BDI_CAP_NO_WRITEBACK implies BDI_CAP_NO_ACCT_WB */
300e4ad08feSMiklos Szeredi 	return !(bdi->capabilities & (BDI_CAP_NO_ACCT_WB |
301e4ad08feSMiklos Szeredi 				      BDI_CAP_NO_WRITEBACK));
302e4ad08feSMiklos Szeredi }
3031da177e4SLinus Torvalds 
304e4ad08feSMiklos Szeredi static inline bool mapping_cap_writeback_dirty(struct address_space *mapping)
305e4ad08feSMiklos Szeredi {
306e4ad08feSMiklos Szeredi 	return bdi_cap_writeback_dirty(mapping->backing_dev_info);
307e4ad08feSMiklos Szeredi }
308e4ad08feSMiklos Szeredi 
309e4ad08feSMiklos Szeredi static inline bool mapping_cap_account_dirty(struct address_space *mapping)
310e4ad08feSMiklos Szeredi {
311e4ad08feSMiklos Szeredi 	return bdi_cap_account_dirty(mapping->backing_dev_info);
312e4ad08feSMiklos Szeredi }
3131da177e4SLinus Torvalds 
31403ba3782SJens Axboe static inline int bdi_sched_wait(void *word)
31503ba3782SJens Axboe {
31603ba3782SJens Axboe 	schedule();
31703ba3782SJens Axboe 	return 0;
31803ba3782SJens Axboe }
31903ba3782SJens Axboe 
3201da177e4SLinus Torvalds #endif		/* _LINUX_BACKING_DEV_H */
321