1b2441318SGreg Kroah-Hartman /* SPDX-License-Identifier: GPL-2.0 */ 21da177e4SLinus Torvalds /* 31da177e4SLinus Torvalds * include/linux/buffer_head.h 41da177e4SLinus Torvalds * 51da177e4SLinus Torvalds * Everything to do with buffer_heads. 61da177e4SLinus Torvalds */ 71da177e4SLinus Torvalds 81da177e4SLinus Torvalds #ifndef _LINUX_BUFFER_HEAD_H 91da177e4SLinus Torvalds #define _LINUX_BUFFER_HEAD_H 101da177e4SLinus Torvalds 111da177e4SLinus Torvalds #include <linux/types.h> 123ae72869SBart Van Assche #include <linux/blk_types.h> 131da177e4SLinus Torvalds #include <linux/fs.h> 141da177e4SLinus Torvalds #include <linux/linkage.h> 151da177e4SLinus Torvalds #include <linux/pagemap.h> 161da177e4SLinus Torvalds #include <linux/wait.h> 1760063497SArun Sharma #include <linux/atomic.h> 181da177e4SLinus Torvalds 199361401eSDavid Howells #ifdef CONFIG_BLOCK 209361401eSDavid Howells 211da177e4SLinus Torvalds enum bh_state_bits { 221da177e4SLinus Torvalds BH_Uptodate, /* Contains valid data */ 231da177e4SLinus Torvalds BH_Dirty, /* Is dirty */ 241da177e4SLinus Torvalds BH_Lock, /* Is locked */ 251da177e4SLinus Torvalds BH_Req, /* Has been submitted for I/O */ 261da177e4SLinus Torvalds 271da177e4SLinus Torvalds BH_Mapped, /* Has a disk mapping */ 281da177e4SLinus Torvalds BH_New, /* Disk mapping was newly created by get_block */ 291da177e4SLinus Torvalds BH_Async_Read, /* Is under end_buffer_async_read I/O */ 301da177e4SLinus Torvalds BH_Async_Write, /* Is under end_buffer_async_write I/O */ 311da177e4SLinus Torvalds BH_Delay, /* Buffer is not yet allocated on disk */ 321da177e4SLinus Torvalds BH_Boundary, /* Block is followed by a discontiguity */ 331da177e4SLinus Torvalds BH_Write_EIO, /* I/O error on write */ 3433a266ddSDavid Chinner BH_Unwritten, /* Buffer is allocated on disk but not written */ 3508bafc03SKeith Mannthey BH_Quiet, /* Buffer Error Prinks to be quiet */ 36877f962cSTheodore Ts'o BH_Meta, /* Buffer contains metadata */ 37877f962cSTheodore Ts'o BH_Prio, /* Buffer should be submitted with REQ_PRIO */ 387b7a8665SChristoph Hellwig BH_Defer_Completion, /* Defer AIO completion to workqueue */ 391da177e4SLinus Torvalds 401da177e4SLinus Torvalds BH_PrivateStart,/* not a state bit, but the first bit available 411da177e4SLinus Torvalds * for private allocation by other entities 421da177e4SLinus Torvalds */ 431da177e4SLinus Torvalds }; 441da177e4SLinus Torvalds 4509cbfeafSKirill A. Shutemov #define MAX_BUF_PER_PAGE (PAGE_SIZE / 512) 461da177e4SLinus Torvalds 471da177e4SLinus Torvalds struct page; 481da177e4SLinus Torvalds struct buffer_head; 491da177e4SLinus Torvalds struct address_space; 501da177e4SLinus Torvalds typedef void (bh_end_io_t)(struct buffer_head *bh, int uptodate); 511da177e4SLinus Torvalds 521da177e4SLinus Torvalds /* 53205f87f6SBadari Pulavarty * Historically, a buffer_head was used to map a single block 54205f87f6SBadari Pulavarty * within a page, and of course as the unit of I/O through the 55205f87f6SBadari Pulavarty * filesystem and block layers. Nowadays the basic I/O unit 56205f87f6SBadari Pulavarty * is the bio, and buffer_heads are used for extracting block 57205f87f6SBadari Pulavarty * mappings (via a get_block_t call), for tracking state within 58205f87f6SBadari Pulavarty * a page (via a page_mapping) and for wrapping bio submission 59205f87f6SBadari Pulavarty * for backward compatibility reasons (e.g. submit_bh). 601da177e4SLinus Torvalds */ 611da177e4SLinus Torvalds struct buffer_head { 621da177e4SLinus Torvalds unsigned long b_state; /* buffer state bitmap (see above) */ 631da177e4SLinus Torvalds struct buffer_head *b_this_page;/* circular list of page's buffers */ 641da177e4SLinus Torvalds struct page *b_page; /* the page this bh is mapped to */ 651da177e4SLinus Torvalds 66205f87f6SBadari Pulavarty sector_t b_blocknr; /* start block number */ 67205f87f6SBadari Pulavarty size_t b_size; /* size of mapping */ 68205f87f6SBadari Pulavarty char *b_data; /* pointer to data within the page */ 691da177e4SLinus Torvalds 701da177e4SLinus Torvalds struct block_device *b_bdev; 711da177e4SLinus Torvalds bh_end_io_t *b_end_io; /* I/O completion */ 721da177e4SLinus Torvalds void *b_private; /* reserved for b_end_io */ 731da177e4SLinus Torvalds struct list_head b_assoc_buffers; /* associated with another mapping */ 7458ff407bSJan Kara struct address_space *b_assoc_map; /* mapping this buffer is 7558ff407bSJan Kara associated with */ 76205f87f6SBadari Pulavarty atomic_t b_count; /* users using this buffer_head */ 77f1e67e35SThomas Gleixner spinlock_t b_uptodate_lock; /* Used by the first bh in a page, to 78f1e67e35SThomas Gleixner * serialise IO completion of other 79f1e67e35SThomas Gleixner * buffers in the page */ 801da177e4SLinus Torvalds }; 811da177e4SLinus Torvalds 821da177e4SLinus Torvalds /* 831da177e4SLinus Torvalds * macro tricks to expand the set_buffer_foo(), clear_buffer_foo() 841da177e4SLinus Torvalds * and buffer_foo() functions. 8560f91826SKemi Wang * To avoid reset buffer flags that are already set, because that causes 8660f91826SKemi Wang * a costly cache line transition, check the flag first. 871da177e4SLinus Torvalds */ 881da177e4SLinus Torvalds #define BUFFER_FNS(bit, name) \ 89ee91ef61SDenys Vlasenko static __always_inline void set_buffer_##name(struct buffer_head *bh) \ 901da177e4SLinus Torvalds { \ 9160f91826SKemi Wang if (!test_bit(BH_##bit, &(bh)->b_state)) \ 921da177e4SLinus Torvalds set_bit(BH_##bit, &(bh)->b_state); \ 931da177e4SLinus Torvalds } \ 94ee91ef61SDenys Vlasenko static __always_inline void clear_buffer_##name(struct buffer_head *bh) \ 951da177e4SLinus Torvalds { \ 961da177e4SLinus Torvalds clear_bit(BH_##bit, &(bh)->b_state); \ 971da177e4SLinus Torvalds } \ 98ee91ef61SDenys Vlasenko static __always_inline int buffer_##name(const struct buffer_head *bh) \ 991da177e4SLinus Torvalds { \ 1001da177e4SLinus Torvalds return test_bit(BH_##bit, &(bh)->b_state); \ 1011da177e4SLinus Torvalds } 1021da177e4SLinus Torvalds 1031da177e4SLinus Torvalds /* 1041da177e4SLinus Torvalds * test_set_buffer_foo() and test_clear_buffer_foo() 1051da177e4SLinus Torvalds */ 1061da177e4SLinus Torvalds #define TAS_BUFFER_FNS(bit, name) \ 107ee91ef61SDenys Vlasenko static __always_inline int test_set_buffer_##name(struct buffer_head *bh) \ 1081da177e4SLinus Torvalds { \ 1091da177e4SLinus Torvalds return test_and_set_bit(BH_##bit, &(bh)->b_state); \ 1101da177e4SLinus Torvalds } \ 111ee91ef61SDenys Vlasenko static __always_inline int test_clear_buffer_##name(struct buffer_head *bh) \ 1121da177e4SLinus Torvalds { \ 1131da177e4SLinus Torvalds return test_and_clear_bit(BH_##bit, &(bh)->b_state); \ 1141da177e4SLinus Torvalds } \ 1151da177e4SLinus Torvalds 1161da177e4SLinus Torvalds /* 1171da177e4SLinus Torvalds * Emit the buffer bitops functions. Note that there are also functions 1181da177e4SLinus Torvalds * of the form "mark_buffer_foo()". These are higher-level functions which 1191da177e4SLinus Torvalds * do something in addition to setting a b_state bit. 1201da177e4SLinus Torvalds */ 1211da177e4SLinus Torvalds BUFFER_FNS(Dirty, dirty) 1221da177e4SLinus Torvalds TAS_BUFFER_FNS(Dirty, dirty) 1231da177e4SLinus Torvalds BUFFER_FNS(Lock, locked) 1241da177e4SLinus Torvalds BUFFER_FNS(Req, req) 1251da177e4SLinus Torvalds TAS_BUFFER_FNS(Req, req) 1261da177e4SLinus Torvalds BUFFER_FNS(Mapped, mapped) 1271da177e4SLinus Torvalds BUFFER_FNS(New, new) 1281da177e4SLinus Torvalds BUFFER_FNS(Async_Read, async_read) 1291da177e4SLinus Torvalds BUFFER_FNS(Async_Write, async_write) 1301da177e4SLinus Torvalds BUFFER_FNS(Delay, delay) 1311da177e4SLinus Torvalds BUFFER_FNS(Boundary, boundary) 1321da177e4SLinus Torvalds BUFFER_FNS(Write_EIO, write_io_error) 13333a266ddSDavid Chinner BUFFER_FNS(Unwritten, unwritten) 134877f962cSTheodore Ts'o BUFFER_FNS(Meta, meta) 135877f962cSTheodore Ts'o BUFFER_FNS(Prio, prio) 1367b7a8665SChristoph Hellwig BUFFER_FNS(Defer_Completion, defer_completion) 1371da177e4SLinus Torvalds 138d4252071SMikulas Patocka static __always_inline void set_buffer_uptodate(struct buffer_head *bh) 139d4252071SMikulas Patocka { 140d4252071SMikulas Patocka /* 141*2f79cdfeSLinus Torvalds * If somebody else already set this uptodate, they will 142*2f79cdfeSLinus Torvalds * have done the memory barrier, and a reader will thus 143*2f79cdfeSLinus Torvalds * see *some* valid buffer state. 144*2f79cdfeSLinus Torvalds * 145*2f79cdfeSLinus Torvalds * Any other serialization (with IO errors or whatever that 146*2f79cdfeSLinus Torvalds * might clear the bit) has to come from other state (eg BH_Lock). 147*2f79cdfeSLinus Torvalds */ 148*2f79cdfeSLinus Torvalds if (test_bit(BH_Uptodate, &bh->b_state)) 149*2f79cdfeSLinus Torvalds return; 150*2f79cdfeSLinus Torvalds 151*2f79cdfeSLinus Torvalds /* 152d4252071SMikulas Patocka * make it consistent with folio_mark_uptodate 153d4252071SMikulas Patocka * pairs with smp_load_acquire in buffer_uptodate 154d4252071SMikulas Patocka */ 155d4252071SMikulas Patocka smp_mb__before_atomic(); 156d4252071SMikulas Patocka set_bit(BH_Uptodate, &bh->b_state); 157d4252071SMikulas Patocka } 158d4252071SMikulas Patocka 159d4252071SMikulas Patocka static __always_inline void clear_buffer_uptodate(struct buffer_head *bh) 160d4252071SMikulas Patocka { 161d4252071SMikulas Patocka clear_bit(BH_Uptodate, &bh->b_state); 162d4252071SMikulas Patocka } 163d4252071SMikulas Patocka 164d4252071SMikulas Patocka static __always_inline int buffer_uptodate(const struct buffer_head *bh) 165d4252071SMikulas Patocka { 166d4252071SMikulas Patocka /* 167d4252071SMikulas Patocka * make it consistent with folio_test_uptodate 168d4252071SMikulas Patocka * pairs with smp_mb__before_atomic in set_buffer_uptodate 169d4252071SMikulas Patocka */ 1708238b457SMikulas Patocka return test_bit_acquire(BH_Uptodate, &bh->b_state); 171d4252071SMikulas Patocka } 172d4252071SMikulas Patocka 1731da177e4SLinus Torvalds #define bh_offset(bh) ((unsigned long)(bh)->b_data & ~PAGE_MASK) 1741da177e4SLinus Torvalds 1751da177e4SLinus Torvalds /* If we *know* page->private refers to buffer_heads */ 1761da177e4SLinus Torvalds #define page_buffers(page) \ 1771da177e4SLinus Torvalds ({ \ 1781da177e4SLinus Torvalds BUG_ON(!PagePrivate(page)); \ 1794c21e2f2SHugh Dickins ((struct buffer_head *)page_private(page)); \ 1801da177e4SLinus Torvalds }) 1811da177e4SLinus Torvalds #define page_has_buffers(page) PagePrivate(page) 182cd1067beSMatthew Wilcox (Oracle) #define folio_buffers(folio) folio_get_private(folio) 1831da177e4SLinus Torvalds 184520f301cSMatthew Wilcox (Oracle) void buffer_check_dirty_writeback(struct folio *folio, 185b4597226SMel Gorman bool *dirty, bool *writeback); 186b4597226SMel Gorman 1871da177e4SLinus Torvalds /* 1881da177e4SLinus Torvalds * Declarations 1891da177e4SLinus Torvalds */ 1901da177e4SLinus Torvalds 191b3c97528SHarvey Harrison void mark_buffer_dirty(struct buffer_head *bh); 19287354e5dSJeff Layton void mark_buffer_write_io_error(struct buffer_head *bh); 193f0059afdSTejun Heo void touch_buffer(struct buffer_head *bh); 1941da177e4SLinus Torvalds void set_bh_page(struct buffer_head *bh, 1951da177e4SLinus Torvalds struct page *page, unsigned long offset); 19668189fefSMatthew Wilcox (Oracle) bool try_to_free_buffers(struct folio *); 1971da177e4SLinus Torvalds struct buffer_head *alloc_page_buffers(struct page *page, unsigned long size, 198640ab98fSJens Axboe bool retry); 1991da177e4SLinus Torvalds void create_empty_buffers(struct page *, unsigned long, 2001da177e4SLinus Torvalds unsigned long b_state); 2011da177e4SLinus Torvalds void end_buffer_read_sync(struct buffer_head *bh, int uptodate); 2021da177e4SLinus Torvalds void end_buffer_write_sync(struct buffer_head *bh, int uptodate); 20335c80d5fSChris Mason void end_buffer_async_write(struct buffer_head *bh, int uptodate); 2041da177e4SLinus Torvalds 2051da177e4SLinus Torvalds /* Things to do with buffers at mapping->private_list */ 2061da177e4SLinus Torvalds void mark_buffer_dirty_inode(struct buffer_head *bh, struct inode *inode); 2071da177e4SLinus Torvalds int inode_has_buffers(struct inode *); 2081da177e4SLinus Torvalds void invalidate_inode_buffers(struct inode *); 2091da177e4SLinus Torvalds int remove_inode_buffers(struct inode *inode); 2101da177e4SLinus Torvalds int sync_mapping_buffers(struct address_space *mapping); 21129f3ad7dSJan Kara void clean_bdev_aliases(struct block_device *bdev, sector_t block, 21229f3ad7dSJan Kara sector_t len); 213e64855c6SJan Kara static inline void clean_bdev_bh_alias(struct buffer_head *bh) 214e64855c6SJan Kara { 215e64855c6SJan Kara clean_bdev_aliases(bh->b_bdev, bh->b_blocknr, 1); 216e64855c6SJan Kara } 2171da177e4SLinus Torvalds 2181da177e4SLinus Torvalds void mark_buffer_async_write(struct buffer_head *bh); 2191da177e4SLinus Torvalds void __wait_on_buffer(struct buffer_head *); 2201da177e4SLinus Torvalds wait_queue_head_t *bh_waitq_head(struct buffer_head *bh); 2213991d3bdSTomasz Kvarsin struct buffer_head *__find_get_block(struct block_device *bdev, sector_t block, 2223991d3bdSTomasz Kvarsin unsigned size); 2233b5e6454SGioh Kim struct buffer_head *__getblk_gfp(struct block_device *bdev, sector_t block, 2243b5e6454SGioh Kim unsigned size, gfp_t gfp); 2251da177e4SLinus Torvalds void __brelse(struct buffer_head *); 2261da177e4SLinus Torvalds void __bforget(struct buffer_head *); 2273991d3bdSTomasz Kvarsin void __breadahead(struct block_device *, sector_t block, unsigned int size); 228d87f6392SRoman Gushchin void __breadahead_gfp(struct block_device *, sector_t block, unsigned int size, 229d87f6392SRoman Gushchin gfp_t gfp); 2303b5e6454SGioh Kim struct buffer_head *__bread_gfp(struct block_device *, 2313b5e6454SGioh Kim sector_t block, unsigned size, gfp_t gfp); 232f9a14399SPeter Zijlstra void invalidate_bh_lrus(void); 233243418e3SMinchan Kim void invalidate_bh_lrus_cpu(void); 2348cc621d2SMinchan Kim bool has_bh_in_lru(int cpu, void *dummy); 235dd0fc66fSAl Viro struct buffer_head *alloc_buffer_head(gfp_t gfp_flags); 2361da177e4SLinus Torvalds void free_buffer_head(struct buffer_head * bh); 237b3c97528SHarvey Harrison void unlock_buffer(struct buffer_head *bh); 238b3c97528SHarvey Harrison void __lock_buffer(struct buffer_head *bh); 2391420c4a5SBart Van Assche void ll_rw_block(blk_opf_t, int, struct buffer_head * bh[]); 2401da177e4SLinus Torvalds int sync_dirty_buffer(struct buffer_head *bh); 2413ae72869SBart Van Assche int __sync_dirty_buffer(struct buffer_head *bh, blk_opf_t op_flags); 2423ae72869SBart Van Assche void write_dirty_buffer(struct buffer_head *bh, blk_opf_t op_flags); 2431420c4a5SBart Van Assche int submit_bh(blk_opf_t, struct buffer_head *); 2441da177e4SLinus Torvalds void write_boundary_block(struct block_device *bdev, 2451da177e4SLinus Torvalds sector_t bblock, unsigned blocksize); 246389d1b08SAneesh Kumar K.V int bh_uptodate_or_lock(struct buffer_head *bh); 247389d1b08SAneesh Kumar K.V int bh_submit_read(struct buffer_head *bh); 2481da177e4SLinus Torvalds 2491da177e4SLinus Torvalds extern int buffer_heads_over_limit; 2501da177e4SLinus Torvalds 2511da177e4SLinus Torvalds /* 2521da177e4SLinus Torvalds * Generic address_space_operations implementations for buffer_head-backed 2531da177e4SLinus Torvalds * address_spaces. 2541da177e4SLinus Torvalds */ 2557ba13abbSMatthew Wilcox (Oracle) void block_invalidate_folio(struct folio *folio, size_t offset, size_t length); 2561da177e4SLinus Torvalds int block_write_full_page(struct page *page, get_block_t *get_block, 2571da177e4SLinus Torvalds struct writeback_control *wbc); 258b4bba389SBenjamin Marzinski int __block_write_full_page(struct inode *inode, struct page *page, 259b4bba389SBenjamin Marzinski get_block_t *get_block, struct writeback_control *wbc, 260b4bba389SBenjamin Marzinski bh_end_io_t *handler); 2612c69e205SMatthew Wilcox (Oracle) int block_read_full_folio(struct folio *, get_block_t *); 2622e7e80f7SMatthew Wilcox (Oracle) bool block_is_partially_uptodate(struct folio *, size_t from, size_t count); 263155130a4SChristoph Hellwig int block_write_begin(struct address_space *mapping, loff_t pos, unsigned len, 264b3992d1eSMatthew Wilcox (Oracle) struct page **pagep, get_block_t *get_block); 2656e1db88dSChristoph Hellwig int __block_write_begin(struct page *page, loff_t pos, unsigned len, 2666e1db88dSChristoph Hellwig get_block_t *get_block); 267afddba49SNick Piggin int block_write_end(struct file *, struct address_space *, 268afddba49SNick Piggin loff_t, unsigned, unsigned, 269afddba49SNick Piggin struct page *, void *); 270afddba49SNick Piggin int generic_write_end(struct file *, struct address_space *, 271afddba49SNick Piggin loff_t, unsigned, unsigned, 272afddba49SNick Piggin struct page *, void *); 273afddba49SNick Piggin void page_zero_new_buffers(struct page *page, unsigned from, unsigned to); 274f892760aSMatthew Wilcox void clean_page_buffers(struct page *page); 27589e10787SNick Piggin int cont_write_begin(struct file *, struct address_space *, loff_t, 276be3bbbc5SMatthew Wilcox (Oracle) unsigned, struct page **, void **, 27789e10787SNick Piggin get_block_t *, loff_t *); 27805eb0b51SOGAWA Hirofumi int generic_cont_expand_simple(struct inode *inode, loff_t size); 2791da177e4SLinus Torvalds int block_commit_write(struct page *page, unsigned from, unsigned to); 280c2ec175cSNick Piggin int block_page_mkwrite(struct vm_area_struct *vma, struct vm_fault *vmf, 28154171690SDavid Chinner get_block_t get_block); 28224da4fabSJan Kara /* Convert errno to return value from ->page_mkwrite() call */ 283401b25aaSSouptick Joarder static inline vm_fault_t block_page_mkwrite_return(int err) 28424da4fabSJan Kara { 28524da4fabSJan Kara if (err == 0) 28624da4fabSJan Kara return VM_FAULT_LOCKED; 2870911d004SJan Kara if (err == -EFAULT || err == -EAGAIN) 28824da4fabSJan Kara return VM_FAULT_NOPAGE; 28924da4fabSJan Kara if (err == -ENOMEM) 29024da4fabSJan Kara return VM_FAULT_OOM; 29124da4fabSJan Kara /* -ENOSPC, -EDQUOT, -EIO ... */ 29224da4fabSJan Kara return VM_FAULT_SIGBUS; 29324da4fabSJan Kara } 2941da177e4SLinus Torvalds sector_t generic_block_bmap(struct address_space *, sector_t, get_block_t *); 2951da177e4SLinus Torvalds int block_truncate_page(struct address_space *, loff_t, get_block_t *); 2961da177e4SLinus Torvalds 29767235182SMatthew Wilcox (Oracle) #ifdef CONFIG_MIGRATION 29867235182SMatthew Wilcox (Oracle) extern int buffer_migrate_folio(struct address_space *, 29967235182SMatthew Wilcox (Oracle) struct folio *dst, struct folio *src, enum migrate_mode); 30067235182SMatthew Wilcox (Oracle) extern int buffer_migrate_folio_norefs(struct address_space *, 30167235182SMatthew Wilcox (Oracle) struct folio *dst, struct folio *src, enum migrate_mode); 30267235182SMatthew Wilcox (Oracle) #else 30367235182SMatthew Wilcox (Oracle) #define buffer_migrate_folio NULL 30467235182SMatthew Wilcox (Oracle) #define buffer_migrate_folio_norefs NULL 30567235182SMatthew Wilcox (Oracle) #endif 3061da177e4SLinus Torvalds 307b6cd0b77SAdrian Bunk void buffer_init(void); 3081da177e4SLinus Torvalds 3091da177e4SLinus Torvalds /* 3101da177e4SLinus Torvalds * inline definitions 3111da177e4SLinus Torvalds */ 3121da177e4SLinus Torvalds 3131da177e4SLinus Torvalds static inline void get_bh(struct buffer_head *bh) 3141da177e4SLinus Torvalds { 3151da177e4SLinus Torvalds atomic_inc(&bh->b_count); 3161da177e4SLinus Torvalds } 3171da177e4SLinus Torvalds 3181da177e4SLinus Torvalds static inline void put_bh(struct buffer_head *bh) 3191da177e4SLinus Torvalds { 3204e857c58SPeter Zijlstra smp_mb__before_atomic(); 3211da177e4SLinus Torvalds atomic_dec(&bh->b_count); 3221da177e4SLinus Torvalds } 3231da177e4SLinus Torvalds 3241da177e4SLinus Torvalds static inline void brelse(struct buffer_head *bh) 3251da177e4SLinus Torvalds { 3261da177e4SLinus Torvalds if (bh) 3271da177e4SLinus Torvalds __brelse(bh); 3281da177e4SLinus Torvalds } 3291da177e4SLinus Torvalds 3301da177e4SLinus Torvalds static inline void bforget(struct buffer_head *bh) 3311da177e4SLinus Torvalds { 3321da177e4SLinus Torvalds if (bh) 3331da177e4SLinus Torvalds __bforget(bh); 3341da177e4SLinus Torvalds } 3351da177e4SLinus Torvalds 3361da177e4SLinus Torvalds static inline struct buffer_head * 3371da177e4SLinus Torvalds sb_bread(struct super_block *sb, sector_t block) 3381da177e4SLinus Torvalds { 3393b5e6454SGioh Kim return __bread_gfp(sb->s_bdev, block, sb->s_blocksize, __GFP_MOVABLE); 3403b5e6454SGioh Kim } 3413b5e6454SGioh Kim 3423b5e6454SGioh Kim static inline struct buffer_head * 3433b5e6454SGioh Kim sb_bread_unmovable(struct super_block *sb, sector_t block) 3443b5e6454SGioh Kim { 3453b5e6454SGioh Kim return __bread_gfp(sb->s_bdev, block, sb->s_blocksize, 0); 3461da177e4SLinus Torvalds } 3471da177e4SLinus Torvalds 3481da177e4SLinus Torvalds static inline void 3491da177e4SLinus Torvalds sb_breadahead(struct super_block *sb, sector_t block) 3501da177e4SLinus Torvalds { 3511da177e4SLinus Torvalds __breadahead(sb->s_bdev, block, sb->s_blocksize); 3521da177e4SLinus Torvalds } 3531da177e4SLinus Torvalds 354d87f6392SRoman Gushchin static inline void 355d87f6392SRoman Gushchin sb_breadahead_unmovable(struct super_block *sb, sector_t block) 356d87f6392SRoman Gushchin { 357d87f6392SRoman Gushchin __breadahead_gfp(sb->s_bdev, block, sb->s_blocksize, 0); 358d87f6392SRoman Gushchin } 359d87f6392SRoman Gushchin 3601da177e4SLinus Torvalds static inline struct buffer_head * 3611da177e4SLinus Torvalds sb_getblk(struct super_block *sb, sector_t block) 3621da177e4SLinus Torvalds { 3633b5e6454SGioh Kim return __getblk_gfp(sb->s_bdev, block, sb->s_blocksize, __GFP_MOVABLE); 3641da177e4SLinus Torvalds } 3651da177e4SLinus Torvalds 366bd7ade3cSNikolay Borisov 367bd7ade3cSNikolay Borisov static inline struct buffer_head * 368bd7ade3cSNikolay Borisov sb_getblk_gfp(struct super_block *sb, sector_t block, gfp_t gfp) 369bd7ade3cSNikolay Borisov { 370bd7ade3cSNikolay Borisov return __getblk_gfp(sb->s_bdev, block, sb->s_blocksize, gfp); 371bd7ade3cSNikolay Borisov } 372bd7ade3cSNikolay Borisov 3731da177e4SLinus Torvalds static inline struct buffer_head * 3741da177e4SLinus Torvalds sb_find_get_block(struct super_block *sb, sector_t block) 3751da177e4SLinus Torvalds { 3761da177e4SLinus Torvalds return __find_get_block(sb->s_bdev, block, sb->s_blocksize); 3771da177e4SLinus Torvalds } 3781da177e4SLinus Torvalds 3791da177e4SLinus Torvalds static inline void 3801da177e4SLinus Torvalds map_bh(struct buffer_head *bh, struct super_block *sb, sector_t block) 3811da177e4SLinus Torvalds { 3821da177e4SLinus Torvalds set_buffer_mapped(bh); 3831da177e4SLinus Torvalds bh->b_bdev = sb->s_bdev; 3841da177e4SLinus Torvalds bh->b_blocknr = block; 385b0cf2321SBadari Pulavarty bh->b_size = sb->s_blocksize; 3861da177e4SLinus Torvalds } 3871da177e4SLinus Torvalds 3881da177e4SLinus Torvalds static inline void wait_on_buffer(struct buffer_head *bh) 3891da177e4SLinus Torvalds { 3901da177e4SLinus Torvalds might_sleep(); 391a9877cc2SRichard Kennedy if (buffer_locked(bh)) 3921da177e4SLinus Torvalds __wait_on_buffer(bh); 3931da177e4SLinus Torvalds } 3941da177e4SLinus Torvalds 395ca5de404SNick Piggin static inline int trylock_buffer(struct buffer_head *bh) 396ca5de404SNick Piggin { 39751b07fc3SNick Piggin return likely(!test_and_set_bit_lock(BH_Lock, &bh->b_state)); 398ca5de404SNick Piggin } 399ca5de404SNick Piggin 4001da177e4SLinus Torvalds static inline void lock_buffer(struct buffer_head *bh) 4011da177e4SLinus Torvalds { 4021da177e4SLinus Torvalds might_sleep(); 403ca5de404SNick Piggin if (!trylock_buffer(bh)) 4041da177e4SLinus Torvalds __lock_buffer(bh); 4051da177e4SLinus Torvalds } 4061da177e4SLinus Torvalds 4073b5e6454SGioh Kim static inline struct buffer_head *getblk_unmovable(struct block_device *bdev, 4083b5e6454SGioh Kim sector_t block, 4093b5e6454SGioh Kim unsigned size) 4103b5e6454SGioh Kim { 4113b5e6454SGioh Kim return __getblk_gfp(bdev, block, size, 0); 4123b5e6454SGioh Kim } 4133b5e6454SGioh Kim 4143b5e6454SGioh Kim static inline struct buffer_head *__getblk(struct block_device *bdev, 4153b5e6454SGioh Kim sector_t block, 4163b5e6454SGioh Kim unsigned size) 4173b5e6454SGioh Kim { 4183b5e6454SGioh Kim return __getblk_gfp(bdev, block, size, __GFP_MOVABLE); 4193b5e6454SGioh Kim } 4203b5e6454SGioh Kim 4213b5e6454SGioh Kim /** 4223b5e6454SGioh Kim * __bread() - reads a specified block and returns the bh 4233b5e6454SGioh Kim * @bdev: the block_device to read from 4243b5e6454SGioh Kim * @block: number of block 4253b5e6454SGioh Kim * @size: size (in bytes) to read 4263b5e6454SGioh Kim * 4273b5e6454SGioh Kim * Reads a specified block, and returns buffer head that contains it. 4283b5e6454SGioh Kim * The page cache is allocated from movable area so that it can be migrated. 4293b5e6454SGioh Kim * It returns NULL if the block was unreadable. 4303b5e6454SGioh Kim */ 4313b5e6454SGioh Kim static inline struct buffer_head * 4323b5e6454SGioh Kim __bread(struct block_device *bdev, sector_t block, unsigned size) 4333b5e6454SGioh Kim { 4343b5e6454SGioh Kim return __bread_gfp(bdev, block, size, __GFP_MOVABLE); 4353b5e6454SGioh Kim } 4363b5e6454SGioh Kim 437e621900aSMatthew Wilcox (Oracle) bool block_dirty_folio(struct address_space *mapping, struct folio *folio); 4389361401eSDavid Howells 4399361401eSDavid Howells #else /* CONFIG_BLOCK */ 4409361401eSDavid Howells 4419361401eSDavid Howells static inline void buffer_init(void) {} 44268189fefSMatthew Wilcox (Oracle) static inline bool try_to_free_buffers(struct folio *folio) { return true; } 4439361401eSDavid Howells static inline int inode_has_buffers(struct inode *inode) { return 0; } 4449361401eSDavid Howells static inline void invalidate_inode_buffers(struct inode *inode) {} 4459361401eSDavid Howells static inline int remove_inode_buffers(struct inode *inode) { return 1; } 4469361401eSDavid Howells static inline int sync_mapping_buffers(struct address_space *mapping) { return 0; } 447243418e3SMinchan Kim static inline void invalidate_bh_lrus_cpu(void) {} 4486de522d1SJing Yangyang static inline bool has_bh_in_lru(int cpu, void *dummy) { return false; } 449d2de7ea4SChristoph Hellwig #define buffer_heads_over_limit 0 4509361401eSDavid Howells 4519361401eSDavid Howells #endif /* CONFIG_BLOCK */ 4521da177e4SLinus Torvalds #endif /* _LINUX_BUFFER_HEAD_H */ 453