1b2441318SGreg Kroah-Hartman /* SPDX-License-Identifier: GPL-2.0 */ 21da177e4SLinus Torvalds /* 31da177e4SLinus Torvalds * include/linux/buffer_head.h 41da177e4SLinus Torvalds * 51da177e4SLinus Torvalds * Everything to do with buffer_heads. 61da177e4SLinus Torvalds */ 71da177e4SLinus Torvalds 81da177e4SLinus Torvalds #ifndef _LINUX_BUFFER_HEAD_H 91da177e4SLinus Torvalds #define _LINUX_BUFFER_HEAD_H 101da177e4SLinus Torvalds 111da177e4SLinus Torvalds #include <linux/types.h> 123ae72869SBart Van Assche #include <linux/blk_types.h> 131da177e4SLinus Torvalds #include <linux/fs.h> 141da177e4SLinus Torvalds #include <linux/linkage.h> 151da177e4SLinus Torvalds #include <linux/pagemap.h> 161da177e4SLinus Torvalds #include <linux/wait.h> 1760063497SArun Sharma #include <linux/atomic.h> 181da177e4SLinus Torvalds 199361401eSDavid Howells #ifdef CONFIG_BLOCK 209361401eSDavid Howells 211da177e4SLinus Torvalds enum bh_state_bits { 221da177e4SLinus Torvalds BH_Uptodate, /* Contains valid data */ 231da177e4SLinus Torvalds BH_Dirty, /* Is dirty */ 241da177e4SLinus Torvalds BH_Lock, /* Is locked */ 251da177e4SLinus Torvalds BH_Req, /* Has been submitted for I/O */ 261da177e4SLinus Torvalds 271da177e4SLinus Torvalds BH_Mapped, /* Has a disk mapping */ 281da177e4SLinus Torvalds BH_New, /* Disk mapping was newly created by get_block */ 291da177e4SLinus Torvalds BH_Async_Read, /* Is under end_buffer_async_read I/O */ 301da177e4SLinus Torvalds BH_Async_Write, /* Is under end_buffer_async_write I/O */ 311da177e4SLinus Torvalds BH_Delay, /* Buffer is not yet allocated on disk */ 321da177e4SLinus Torvalds BH_Boundary, /* Block is followed by a discontiguity */ 331da177e4SLinus Torvalds BH_Write_EIO, /* I/O error on write */ 3433a266ddSDavid Chinner BH_Unwritten, /* Buffer is allocated on disk but not written */ 3508bafc03SKeith Mannthey BH_Quiet, /* Buffer Error Prinks to be quiet */ 36877f962cSTheodore Ts'o BH_Meta, /* Buffer contains metadata */ 37877f962cSTheodore Ts'o BH_Prio, /* Buffer should be submitted with REQ_PRIO */ 387b7a8665SChristoph Hellwig BH_Defer_Completion, /* Defer AIO completion to workqueue */ 391da177e4SLinus Torvalds 401da177e4SLinus Torvalds BH_PrivateStart,/* not a state bit, but the first bit available 411da177e4SLinus Torvalds * for private allocation by other entities 421da177e4SLinus Torvalds */ 431da177e4SLinus Torvalds }; 441da177e4SLinus Torvalds 4509cbfeafSKirill A. Shutemov #define MAX_BUF_PER_PAGE (PAGE_SIZE / 512) 461da177e4SLinus Torvalds 471da177e4SLinus Torvalds struct page; 481da177e4SLinus Torvalds struct buffer_head; 491da177e4SLinus Torvalds struct address_space; 501da177e4SLinus Torvalds typedef void (bh_end_io_t)(struct buffer_head *bh, int uptodate); 511da177e4SLinus Torvalds 521da177e4SLinus Torvalds /* 53205f87f6SBadari Pulavarty * Historically, a buffer_head was used to map a single block 54205f87f6SBadari Pulavarty * within a page, and of course as the unit of I/O through the 55205f87f6SBadari Pulavarty * filesystem and block layers. Nowadays the basic I/O unit 56205f87f6SBadari Pulavarty * is the bio, and buffer_heads are used for extracting block 57205f87f6SBadari Pulavarty * mappings (via a get_block_t call), for tracking state within 58205f87f6SBadari Pulavarty * a page (via a page_mapping) and for wrapping bio submission 59205f87f6SBadari Pulavarty * for backward compatibility reasons (e.g. submit_bh). 601da177e4SLinus Torvalds */ 611da177e4SLinus Torvalds struct buffer_head { 621da177e4SLinus Torvalds unsigned long b_state; /* buffer state bitmap (see above) */ 631da177e4SLinus Torvalds struct buffer_head *b_this_page;/* circular list of page's buffers */ 64d685c668SMatthew Wilcox (Oracle) union { 651da177e4SLinus Torvalds struct page *b_page; /* the page this bh is mapped to */ 66d685c668SMatthew Wilcox (Oracle) struct folio *b_folio; /* the folio this bh is mapped to */ 67d685c668SMatthew Wilcox (Oracle) }; 681da177e4SLinus Torvalds 69205f87f6SBadari Pulavarty sector_t b_blocknr; /* start block number */ 70205f87f6SBadari Pulavarty size_t b_size; /* size of mapping */ 71205f87f6SBadari Pulavarty char *b_data; /* pointer to data within the page */ 721da177e4SLinus Torvalds 731da177e4SLinus Torvalds struct block_device *b_bdev; 741da177e4SLinus Torvalds bh_end_io_t *b_end_io; /* I/O completion */ 751da177e4SLinus Torvalds void *b_private; /* reserved for b_end_io */ 761da177e4SLinus Torvalds struct list_head b_assoc_buffers; /* associated with another mapping */ 7758ff407bSJan Kara struct address_space *b_assoc_map; /* mapping this buffer is 7858ff407bSJan Kara associated with */ 79205f87f6SBadari Pulavarty atomic_t b_count; /* users using this buffer_head */ 80f1e67e35SThomas Gleixner spinlock_t b_uptodate_lock; /* Used by the first bh in a page, to 81f1e67e35SThomas Gleixner * serialise IO completion of other 82f1e67e35SThomas Gleixner * buffers in the page */ 831da177e4SLinus Torvalds }; 841da177e4SLinus Torvalds 851da177e4SLinus Torvalds /* 861da177e4SLinus Torvalds * macro tricks to expand the set_buffer_foo(), clear_buffer_foo() 871da177e4SLinus Torvalds * and buffer_foo() functions. 8860f91826SKemi Wang * To avoid reset buffer flags that are already set, because that causes 8960f91826SKemi Wang * a costly cache line transition, check the flag first. 901da177e4SLinus Torvalds */ 911da177e4SLinus Torvalds #define BUFFER_FNS(bit, name) \ 92ee91ef61SDenys Vlasenko static __always_inline void set_buffer_##name(struct buffer_head *bh) \ 931da177e4SLinus Torvalds { \ 9460f91826SKemi Wang if (!test_bit(BH_##bit, &(bh)->b_state)) \ 951da177e4SLinus Torvalds set_bit(BH_##bit, &(bh)->b_state); \ 961da177e4SLinus Torvalds } \ 97ee91ef61SDenys Vlasenko static __always_inline void clear_buffer_##name(struct buffer_head *bh) \ 981da177e4SLinus Torvalds { \ 991da177e4SLinus Torvalds clear_bit(BH_##bit, &(bh)->b_state); \ 1001da177e4SLinus Torvalds } \ 101ee91ef61SDenys Vlasenko static __always_inline int buffer_##name(const struct buffer_head *bh) \ 1021da177e4SLinus Torvalds { \ 1031da177e4SLinus Torvalds return test_bit(BH_##bit, &(bh)->b_state); \ 1041da177e4SLinus Torvalds } 1051da177e4SLinus Torvalds 1061da177e4SLinus Torvalds /* 1071da177e4SLinus Torvalds * test_set_buffer_foo() and test_clear_buffer_foo() 1081da177e4SLinus Torvalds */ 1091da177e4SLinus Torvalds #define TAS_BUFFER_FNS(bit, name) \ 110ee91ef61SDenys Vlasenko static __always_inline int test_set_buffer_##name(struct buffer_head *bh) \ 1111da177e4SLinus Torvalds { \ 1121da177e4SLinus Torvalds return test_and_set_bit(BH_##bit, &(bh)->b_state); \ 1131da177e4SLinus Torvalds } \ 114ee91ef61SDenys Vlasenko static __always_inline int test_clear_buffer_##name(struct buffer_head *bh) \ 1151da177e4SLinus Torvalds { \ 1161da177e4SLinus Torvalds return test_and_clear_bit(BH_##bit, &(bh)->b_state); \ 1171da177e4SLinus Torvalds } \ 1181da177e4SLinus Torvalds 1191da177e4SLinus Torvalds /* 1201da177e4SLinus Torvalds * Emit the buffer bitops functions. Note that there are also functions 1211da177e4SLinus Torvalds * of the form "mark_buffer_foo()". These are higher-level functions which 1221da177e4SLinus Torvalds * do something in addition to setting a b_state bit. 1231da177e4SLinus Torvalds */ 1241da177e4SLinus Torvalds BUFFER_FNS(Dirty, dirty) 1251da177e4SLinus Torvalds TAS_BUFFER_FNS(Dirty, dirty) 1261da177e4SLinus Torvalds BUFFER_FNS(Lock, locked) 1271da177e4SLinus Torvalds BUFFER_FNS(Req, req) 1281da177e4SLinus Torvalds TAS_BUFFER_FNS(Req, req) 1291da177e4SLinus Torvalds BUFFER_FNS(Mapped, mapped) 1301da177e4SLinus Torvalds BUFFER_FNS(New, new) 1311da177e4SLinus Torvalds BUFFER_FNS(Async_Read, async_read) 1321da177e4SLinus Torvalds BUFFER_FNS(Async_Write, async_write) 1331da177e4SLinus Torvalds BUFFER_FNS(Delay, delay) 1341da177e4SLinus Torvalds BUFFER_FNS(Boundary, boundary) 1351da177e4SLinus Torvalds BUFFER_FNS(Write_EIO, write_io_error) 13633a266ddSDavid Chinner BUFFER_FNS(Unwritten, unwritten) 137877f962cSTheodore Ts'o BUFFER_FNS(Meta, meta) 138877f962cSTheodore Ts'o BUFFER_FNS(Prio, prio) 1397b7a8665SChristoph Hellwig BUFFER_FNS(Defer_Completion, defer_completion) 1401da177e4SLinus Torvalds 141d4252071SMikulas Patocka static __always_inline void set_buffer_uptodate(struct buffer_head *bh) 142d4252071SMikulas Patocka { 143d4252071SMikulas Patocka /* 1442f79cdfeSLinus Torvalds * If somebody else already set this uptodate, they will 1452f79cdfeSLinus Torvalds * have done the memory barrier, and a reader will thus 1462f79cdfeSLinus Torvalds * see *some* valid buffer state. 1472f79cdfeSLinus Torvalds * 1482f79cdfeSLinus Torvalds * Any other serialization (with IO errors or whatever that 1492f79cdfeSLinus Torvalds * might clear the bit) has to come from other state (eg BH_Lock). 1502f79cdfeSLinus Torvalds */ 1512f79cdfeSLinus Torvalds if (test_bit(BH_Uptodate, &bh->b_state)) 1522f79cdfeSLinus Torvalds return; 1532f79cdfeSLinus Torvalds 1542f79cdfeSLinus Torvalds /* 155d4252071SMikulas Patocka * make it consistent with folio_mark_uptodate 156d4252071SMikulas Patocka * pairs with smp_load_acquire in buffer_uptodate 157d4252071SMikulas Patocka */ 158d4252071SMikulas Patocka smp_mb__before_atomic(); 159d4252071SMikulas Patocka set_bit(BH_Uptodate, &bh->b_state); 160d4252071SMikulas Patocka } 161d4252071SMikulas Patocka 162d4252071SMikulas Patocka static __always_inline void clear_buffer_uptodate(struct buffer_head *bh) 163d4252071SMikulas Patocka { 164d4252071SMikulas Patocka clear_bit(BH_Uptodate, &bh->b_state); 165d4252071SMikulas Patocka } 166d4252071SMikulas Patocka 167d4252071SMikulas Patocka static __always_inline int buffer_uptodate(const struct buffer_head *bh) 168d4252071SMikulas Patocka { 169d4252071SMikulas Patocka /* 170d4252071SMikulas Patocka * make it consistent with folio_test_uptodate 171d4252071SMikulas Patocka * pairs with smp_mb__before_atomic in set_buffer_uptodate 172d4252071SMikulas Patocka */ 1738238b457SMikulas Patocka return test_bit_acquire(BH_Uptodate, &bh->b_state); 174d4252071SMikulas Patocka } 175d4252071SMikulas Patocka 1761da177e4SLinus Torvalds #define bh_offset(bh) ((unsigned long)(bh)->b_data & ~PAGE_MASK) 1771da177e4SLinus Torvalds 1781da177e4SLinus Torvalds /* If we *know* page->private refers to buffer_heads */ 1791da177e4SLinus Torvalds #define page_buffers(page) \ 1801da177e4SLinus Torvalds ({ \ 1811da177e4SLinus Torvalds BUG_ON(!PagePrivate(page)); \ 1824c21e2f2SHugh Dickins ((struct buffer_head *)page_private(page)); \ 1831da177e4SLinus Torvalds }) 1841da177e4SLinus Torvalds #define page_has_buffers(page) PagePrivate(page) 185cd1067beSMatthew Wilcox (Oracle) #define folio_buffers(folio) folio_get_private(folio) 1861da177e4SLinus Torvalds 187520f301cSMatthew Wilcox (Oracle) void buffer_check_dirty_writeback(struct folio *folio, 188b4597226SMel Gorman bool *dirty, bool *writeback); 189b4597226SMel Gorman 1901da177e4SLinus Torvalds /* 1911da177e4SLinus Torvalds * Declarations 1921da177e4SLinus Torvalds */ 1931da177e4SLinus Torvalds 194b3c97528SHarvey Harrison void mark_buffer_dirty(struct buffer_head *bh); 19587354e5dSJeff Layton void mark_buffer_write_io_error(struct buffer_head *bh); 196f0059afdSTejun Heo void touch_buffer(struct buffer_head *bh); 1971da177e4SLinus Torvalds void set_bh_page(struct buffer_head *bh, 1981da177e4SLinus Torvalds struct page *page, unsigned long offset); 199465e5e6aSPankaj Raghav void folio_set_bh(struct buffer_head *bh, struct folio *folio, 200465e5e6aSPankaj Raghav unsigned long offset); 20168189fefSMatthew Wilcox (Oracle) bool try_to_free_buffers(struct folio *); 202c71124a8SPankaj Raghav struct buffer_head *folio_alloc_buffers(struct folio *folio, unsigned long size, 203c71124a8SPankaj Raghav bool retry); 2041da177e4SLinus Torvalds struct buffer_head *alloc_page_buffers(struct page *page, unsigned long size, 205640ab98fSJens Axboe bool retry); 2061da177e4SLinus Torvalds void create_empty_buffers(struct page *, unsigned long, 2071da177e4SLinus Torvalds unsigned long b_state); 2088e2e1756SPankaj Raghav void folio_create_empty_buffers(struct folio *folio, unsigned long blocksize, 2098e2e1756SPankaj Raghav unsigned long b_state); 2101da177e4SLinus Torvalds void end_buffer_read_sync(struct buffer_head *bh, int uptodate); 2111da177e4SLinus Torvalds void end_buffer_write_sync(struct buffer_head *bh, int uptodate); 21235c80d5fSChris Mason void end_buffer_async_write(struct buffer_head *bh, int uptodate); 2131da177e4SLinus Torvalds 2141da177e4SLinus Torvalds /* Things to do with buffers at mapping->private_list */ 2151da177e4SLinus Torvalds void mark_buffer_dirty_inode(struct buffer_head *bh, struct inode *inode); 2161da177e4SLinus Torvalds int inode_has_buffers(struct inode *); 2171da177e4SLinus Torvalds void invalidate_inode_buffers(struct inode *); 2181da177e4SLinus Torvalds int remove_inode_buffers(struct inode *inode); 2191da177e4SLinus Torvalds int sync_mapping_buffers(struct address_space *mapping); 22031b2ebc0SRitesh Harjani (IBM) int generic_buffers_fsync_noflush(struct file *file, loff_t start, loff_t end, 22131b2ebc0SRitesh Harjani (IBM) bool datasync); 22231b2ebc0SRitesh Harjani (IBM) int generic_buffers_fsync(struct file *file, loff_t start, loff_t end, 22331b2ebc0SRitesh Harjani (IBM) bool datasync); 22429f3ad7dSJan Kara void clean_bdev_aliases(struct block_device *bdev, sector_t block, 22529f3ad7dSJan Kara sector_t len); 226e64855c6SJan Kara static inline void clean_bdev_bh_alias(struct buffer_head *bh) 227e64855c6SJan Kara { 228e64855c6SJan Kara clean_bdev_aliases(bh->b_bdev, bh->b_blocknr, 1); 229e64855c6SJan Kara } 2301da177e4SLinus Torvalds 2311da177e4SLinus Torvalds void mark_buffer_async_write(struct buffer_head *bh); 2321da177e4SLinus Torvalds void __wait_on_buffer(struct buffer_head *); 2331da177e4SLinus Torvalds wait_queue_head_t *bh_waitq_head(struct buffer_head *bh); 2343991d3bdSTomasz Kvarsin struct buffer_head *__find_get_block(struct block_device *bdev, sector_t block, 2353991d3bdSTomasz Kvarsin unsigned size); 2363b5e6454SGioh Kim struct buffer_head *__getblk_gfp(struct block_device *bdev, sector_t block, 2373b5e6454SGioh Kim unsigned size, gfp_t gfp); 2381da177e4SLinus Torvalds void __brelse(struct buffer_head *); 2391da177e4SLinus Torvalds void __bforget(struct buffer_head *); 2403991d3bdSTomasz Kvarsin void __breadahead(struct block_device *, sector_t block, unsigned int size); 2413b5e6454SGioh Kim struct buffer_head *__bread_gfp(struct block_device *, 2423b5e6454SGioh Kim sector_t block, unsigned size, gfp_t gfp); 243f9a14399SPeter Zijlstra void invalidate_bh_lrus(void); 244243418e3SMinchan Kim void invalidate_bh_lrus_cpu(void); 2458cc621d2SMinchan Kim bool has_bh_in_lru(int cpu, void *dummy); 246dd0fc66fSAl Viro struct buffer_head *alloc_buffer_head(gfp_t gfp_flags); 2471da177e4SLinus Torvalds void free_buffer_head(struct buffer_head * bh); 248b3c97528SHarvey Harrison void unlock_buffer(struct buffer_head *bh); 249b3c97528SHarvey Harrison void __lock_buffer(struct buffer_head *bh); 2501da177e4SLinus Torvalds int sync_dirty_buffer(struct buffer_head *bh); 2513ae72869SBart Van Assche int __sync_dirty_buffer(struct buffer_head *bh, blk_opf_t op_flags); 2523ae72869SBart Van Assche void write_dirty_buffer(struct buffer_head *bh, blk_opf_t op_flags); 2535bdf402aSRitesh Harjani (IBM) void submit_bh(blk_opf_t, struct buffer_head *); 2541da177e4SLinus Torvalds void write_boundary_block(struct block_device *bdev, 2551da177e4SLinus Torvalds sector_t bblock, unsigned blocksize); 256389d1b08SAneesh Kumar K.V int bh_uptodate_or_lock(struct buffer_head *bh); 257fdee117eSZhang Yi int __bh_read(struct buffer_head *bh, blk_opf_t op_flags, bool wait); 258fdee117eSZhang Yi void __bh_read_batch(int nr, struct buffer_head *bhs[], 259fdee117eSZhang Yi blk_opf_t op_flags, bool force_lock); 2601da177e4SLinus Torvalds 2611da177e4SLinus Torvalds extern int buffer_heads_over_limit; 2621da177e4SLinus Torvalds 2631da177e4SLinus Torvalds /* 2641da177e4SLinus Torvalds * Generic address_space_operations implementations for buffer_head-backed 2651da177e4SLinus Torvalds * address_spaces. 2661da177e4SLinus Torvalds */ 2677ba13abbSMatthew Wilcox (Oracle) void block_invalidate_folio(struct folio *folio, size_t offset, size_t length); 2681da177e4SLinus Torvalds int block_write_full_page(struct page *page, get_block_t *get_block, 2691da177e4SLinus Torvalds struct writeback_control *wbc); 27053418a18SMatthew Wilcox (Oracle) int __block_write_full_folio(struct inode *inode, struct folio *folio, 271b4bba389SBenjamin Marzinski get_block_t *get_block, struct writeback_control *wbc, 272b4bba389SBenjamin Marzinski bh_end_io_t *handler); 2732c69e205SMatthew Wilcox (Oracle) int block_read_full_folio(struct folio *, get_block_t *); 2742e7e80f7SMatthew Wilcox (Oracle) bool block_is_partially_uptodate(struct folio *, size_t from, size_t count); 275155130a4SChristoph Hellwig int block_write_begin(struct address_space *mapping, loff_t pos, unsigned len, 276b3992d1eSMatthew Wilcox (Oracle) struct page **pagep, get_block_t *get_block); 2776e1db88dSChristoph Hellwig int __block_write_begin(struct page *page, loff_t pos, unsigned len, 2786e1db88dSChristoph Hellwig get_block_t *get_block); 279afddba49SNick Piggin int block_write_end(struct file *, struct address_space *, 280afddba49SNick Piggin loff_t, unsigned, unsigned, 281afddba49SNick Piggin struct page *, void *); 282afddba49SNick Piggin int generic_write_end(struct file *, struct address_space *, 283afddba49SNick Piggin loff_t, unsigned, unsigned, 284afddba49SNick Piggin struct page *, void *); 2854a9622f2SMatthew Wilcox (Oracle) void folio_zero_new_buffers(struct folio *folio, size_t from, size_t to); 286f892760aSMatthew Wilcox void clean_page_buffers(struct page *page); 28789e10787SNick Piggin int cont_write_begin(struct file *, struct address_space *, loff_t, 288be3bbbc5SMatthew Wilcox (Oracle) unsigned, struct page **, void **, 28989e10787SNick Piggin get_block_t *, loff_t *); 29005eb0b51SOGAWA Hirofumi int generic_cont_expand_simple(struct inode *inode, loff_t size); 291*a524fcfeSBean Huo void block_commit_write(struct page *page, unsigned int from, unsigned int to); 292c2ec175cSNick Piggin int block_page_mkwrite(struct vm_area_struct *vma, struct vm_fault *vmf, 29354171690SDavid Chinner get_block_t get_block); 29424da4fabSJan Kara /* Convert errno to return value from ->page_mkwrite() call */ 295401b25aaSSouptick Joarder static inline vm_fault_t block_page_mkwrite_return(int err) 29624da4fabSJan Kara { 29724da4fabSJan Kara if (err == 0) 29824da4fabSJan Kara return VM_FAULT_LOCKED; 2990911d004SJan Kara if (err == -EFAULT || err == -EAGAIN) 30024da4fabSJan Kara return VM_FAULT_NOPAGE; 30124da4fabSJan Kara if (err == -ENOMEM) 30224da4fabSJan Kara return VM_FAULT_OOM; 30324da4fabSJan Kara /* -ENOSPC, -EDQUOT, -EIO ... */ 30424da4fabSJan Kara return VM_FAULT_SIGBUS; 30524da4fabSJan Kara } 3061da177e4SLinus Torvalds sector_t generic_block_bmap(struct address_space *, sector_t, get_block_t *); 3071da177e4SLinus Torvalds int block_truncate_page(struct address_space *, loff_t, get_block_t *); 3081da177e4SLinus Torvalds 30967235182SMatthew Wilcox (Oracle) #ifdef CONFIG_MIGRATION 31067235182SMatthew Wilcox (Oracle) extern int buffer_migrate_folio(struct address_space *, 31167235182SMatthew Wilcox (Oracle) struct folio *dst, struct folio *src, enum migrate_mode); 31267235182SMatthew Wilcox (Oracle) extern int buffer_migrate_folio_norefs(struct address_space *, 31367235182SMatthew Wilcox (Oracle) struct folio *dst, struct folio *src, enum migrate_mode); 31467235182SMatthew Wilcox (Oracle) #else 31567235182SMatthew Wilcox (Oracle) #define buffer_migrate_folio NULL 31667235182SMatthew Wilcox (Oracle) #define buffer_migrate_folio_norefs NULL 31767235182SMatthew Wilcox (Oracle) #endif 3181da177e4SLinus Torvalds 319b6cd0b77SAdrian Bunk void buffer_init(void); 3201da177e4SLinus Torvalds 3211da177e4SLinus Torvalds /* 3221da177e4SLinus Torvalds * inline definitions 3231da177e4SLinus Torvalds */ 3241da177e4SLinus Torvalds 3251da177e4SLinus Torvalds static inline void get_bh(struct buffer_head *bh) 3261da177e4SLinus Torvalds { 3271da177e4SLinus Torvalds atomic_inc(&bh->b_count); 3281da177e4SLinus Torvalds } 3291da177e4SLinus Torvalds 3301da177e4SLinus Torvalds static inline void put_bh(struct buffer_head *bh) 3311da177e4SLinus Torvalds { 3324e857c58SPeter Zijlstra smp_mb__before_atomic(); 3331da177e4SLinus Torvalds atomic_dec(&bh->b_count); 3341da177e4SLinus Torvalds } 3351da177e4SLinus Torvalds 3361da177e4SLinus Torvalds static inline void brelse(struct buffer_head *bh) 3371da177e4SLinus Torvalds { 3381da177e4SLinus Torvalds if (bh) 3391da177e4SLinus Torvalds __brelse(bh); 3401da177e4SLinus Torvalds } 3411da177e4SLinus Torvalds 3421da177e4SLinus Torvalds static inline void bforget(struct buffer_head *bh) 3431da177e4SLinus Torvalds { 3441da177e4SLinus Torvalds if (bh) 3451da177e4SLinus Torvalds __bforget(bh); 3461da177e4SLinus Torvalds } 3471da177e4SLinus Torvalds 3481da177e4SLinus Torvalds static inline struct buffer_head * 3491da177e4SLinus Torvalds sb_bread(struct super_block *sb, sector_t block) 3501da177e4SLinus Torvalds { 3513b5e6454SGioh Kim return __bread_gfp(sb->s_bdev, block, sb->s_blocksize, __GFP_MOVABLE); 3523b5e6454SGioh Kim } 3533b5e6454SGioh Kim 3543b5e6454SGioh Kim static inline struct buffer_head * 3553b5e6454SGioh Kim sb_bread_unmovable(struct super_block *sb, sector_t block) 3563b5e6454SGioh Kim { 3573b5e6454SGioh Kim return __bread_gfp(sb->s_bdev, block, sb->s_blocksize, 0); 3581da177e4SLinus Torvalds } 3591da177e4SLinus Torvalds 3601da177e4SLinus Torvalds static inline void 3611da177e4SLinus Torvalds sb_breadahead(struct super_block *sb, sector_t block) 3621da177e4SLinus Torvalds { 3631da177e4SLinus Torvalds __breadahead(sb->s_bdev, block, sb->s_blocksize); 3641da177e4SLinus Torvalds } 3651da177e4SLinus Torvalds 3661da177e4SLinus Torvalds static inline struct buffer_head * 3671da177e4SLinus Torvalds sb_getblk(struct super_block *sb, sector_t block) 3681da177e4SLinus Torvalds { 3693b5e6454SGioh Kim return __getblk_gfp(sb->s_bdev, block, sb->s_blocksize, __GFP_MOVABLE); 3701da177e4SLinus Torvalds } 3711da177e4SLinus Torvalds 372bd7ade3cSNikolay Borisov 373bd7ade3cSNikolay Borisov static inline struct buffer_head * 374bd7ade3cSNikolay Borisov sb_getblk_gfp(struct super_block *sb, sector_t block, gfp_t gfp) 375bd7ade3cSNikolay Borisov { 376bd7ade3cSNikolay Borisov return __getblk_gfp(sb->s_bdev, block, sb->s_blocksize, gfp); 377bd7ade3cSNikolay Borisov } 378bd7ade3cSNikolay Borisov 3791da177e4SLinus Torvalds static inline struct buffer_head * 3801da177e4SLinus Torvalds sb_find_get_block(struct super_block *sb, sector_t block) 3811da177e4SLinus Torvalds { 3821da177e4SLinus Torvalds return __find_get_block(sb->s_bdev, block, sb->s_blocksize); 3831da177e4SLinus Torvalds } 3841da177e4SLinus Torvalds 3851da177e4SLinus Torvalds static inline void 3861da177e4SLinus Torvalds map_bh(struct buffer_head *bh, struct super_block *sb, sector_t block) 3871da177e4SLinus Torvalds { 3881da177e4SLinus Torvalds set_buffer_mapped(bh); 3891da177e4SLinus Torvalds bh->b_bdev = sb->s_bdev; 3901da177e4SLinus Torvalds bh->b_blocknr = block; 391b0cf2321SBadari Pulavarty bh->b_size = sb->s_blocksize; 3921da177e4SLinus Torvalds } 3931da177e4SLinus Torvalds 3941da177e4SLinus Torvalds static inline void wait_on_buffer(struct buffer_head *bh) 3951da177e4SLinus Torvalds { 3961da177e4SLinus Torvalds might_sleep(); 397a9877cc2SRichard Kennedy if (buffer_locked(bh)) 3981da177e4SLinus Torvalds __wait_on_buffer(bh); 3991da177e4SLinus Torvalds } 4001da177e4SLinus Torvalds 401ca5de404SNick Piggin static inline int trylock_buffer(struct buffer_head *bh) 402ca5de404SNick Piggin { 40351b07fc3SNick Piggin return likely(!test_and_set_bit_lock(BH_Lock, &bh->b_state)); 404ca5de404SNick Piggin } 405ca5de404SNick Piggin 4061da177e4SLinus Torvalds static inline void lock_buffer(struct buffer_head *bh) 4071da177e4SLinus Torvalds { 4081da177e4SLinus Torvalds might_sleep(); 409ca5de404SNick Piggin if (!trylock_buffer(bh)) 4101da177e4SLinus Torvalds __lock_buffer(bh); 4111da177e4SLinus Torvalds } 4121da177e4SLinus Torvalds 4133b5e6454SGioh Kim static inline struct buffer_head *getblk_unmovable(struct block_device *bdev, 4143b5e6454SGioh Kim sector_t block, 4153b5e6454SGioh Kim unsigned size) 4163b5e6454SGioh Kim { 4173b5e6454SGioh Kim return __getblk_gfp(bdev, block, size, 0); 4183b5e6454SGioh Kim } 4193b5e6454SGioh Kim 4203b5e6454SGioh Kim static inline struct buffer_head *__getblk(struct block_device *bdev, 4213b5e6454SGioh Kim sector_t block, 4223b5e6454SGioh Kim unsigned size) 4233b5e6454SGioh Kim { 4243b5e6454SGioh Kim return __getblk_gfp(bdev, block, size, __GFP_MOVABLE); 4253b5e6454SGioh Kim } 4263b5e6454SGioh Kim 427fdee117eSZhang Yi static inline void bh_readahead(struct buffer_head *bh, blk_opf_t op_flags) 428fdee117eSZhang Yi { 429fdee117eSZhang Yi if (!buffer_uptodate(bh) && trylock_buffer(bh)) { 430fdee117eSZhang Yi if (!buffer_uptodate(bh)) 431fdee117eSZhang Yi __bh_read(bh, op_flags, false); 432fdee117eSZhang Yi else 433fdee117eSZhang Yi unlock_buffer(bh); 434fdee117eSZhang Yi } 435fdee117eSZhang Yi } 436fdee117eSZhang Yi 437fdee117eSZhang Yi static inline void bh_read_nowait(struct buffer_head *bh, blk_opf_t op_flags) 438fdee117eSZhang Yi { 439fdee117eSZhang Yi if (!bh_uptodate_or_lock(bh)) 440fdee117eSZhang Yi __bh_read(bh, op_flags, false); 441fdee117eSZhang Yi } 442fdee117eSZhang Yi 443fdee117eSZhang Yi /* Returns 1 if buffer uptodated, 0 on success, and -EIO on error. */ 444fdee117eSZhang Yi static inline int bh_read(struct buffer_head *bh, blk_opf_t op_flags) 445fdee117eSZhang Yi { 446fdee117eSZhang Yi if (bh_uptodate_or_lock(bh)) 447fdee117eSZhang Yi return 1; 448fdee117eSZhang Yi return __bh_read(bh, op_flags, true); 449fdee117eSZhang Yi } 450fdee117eSZhang Yi 451fdee117eSZhang Yi static inline void bh_read_batch(int nr, struct buffer_head *bhs[]) 452fdee117eSZhang Yi { 453fdee117eSZhang Yi __bh_read_batch(nr, bhs, 0, true); 454fdee117eSZhang Yi } 455fdee117eSZhang Yi 456fdee117eSZhang Yi static inline void bh_readahead_batch(int nr, struct buffer_head *bhs[], 457fdee117eSZhang Yi blk_opf_t op_flags) 458fdee117eSZhang Yi { 459fdee117eSZhang Yi __bh_read_batch(nr, bhs, op_flags, false); 460fdee117eSZhang Yi } 461fdee117eSZhang Yi 4623b5e6454SGioh Kim /** 4633b5e6454SGioh Kim * __bread() - reads a specified block and returns the bh 4643b5e6454SGioh Kim * @bdev: the block_device to read from 4653b5e6454SGioh Kim * @block: number of block 4663b5e6454SGioh Kim * @size: size (in bytes) to read 4673b5e6454SGioh Kim * 4683b5e6454SGioh Kim * Reads a specified block, and returns buffer head that contains it. 4693b5e6454SGioh Kim * The page cache is allocated from movable area so that it can be migrated. 4703b5e6454SGioh Kim * It returns NULL if the block was unreadable. 4713b5e6454SGioh Kim */ 4723b5e6454SGioh Kim static inline struct buffer_head * 4733b5e6454SGioh Kim __bread(struct block_device *bdev, sector_t block, unsigned size) 4743b5e6454SGioh Kim { 4753b5e6454SGioh Kim return __bread_gfp(bdev, block, size, __GFP_MOVABLE); 4763b5e6454SGioh Kim } 4773b5e6454SGioh Kim 478e621900aSMatthew Wilcox (Oracle) bool block_dirty_folio(struct address_space *mapping, struct folio *folio); 4799361401eSDavid Howells 4809361401eSDavid Howells #else /* CONFIG_BLOCK */ 4819361401eSDavid Howells 4829361401eSDavid Howells static inline void buffer_init(void) {} 48368189fefSMatthew Wilcox (Oracle) static inline bool try_to_free_buffers(struct folio *folio) { return true; } 4849361401eSDavid Howells static inline int inode_has_buffers(struct inode *inode) { return 0; } 4859361401eSDavid Howells static inline void invalidate_inode_buffers(struct inode *inode) {} 4869361401eSDavid Howells static inline int remove_inode_buffers(struct inode *inode) { return 1; } 4879361401eSDavid Howells static inline int sync_mapping_buffers(struct address_space *mapping) { return 0; } 488243418e3SMinchan Kim static inline void invalidate_bh_lrus_cpu(void) {} 4896de522d1SJing Yangyang static inline bool has_bh_in_lru(int cpu, void *dummy) { return false; } 490d2de7ea4SChristoph Hellwig #define buffer_heads_over_limit 0 4919361401eSDavid Howells 4929361401eSDavid Howells #endif /* CONFIG_BLOCK */ 4931da177e4SLinus Torvalds #endif /* _LINUX_BUFFER_HEAD_H */ 494