| /linux-6.15/include/linux/ |
| H A D | pagevec.h | 43 fbatch->nr = 0; in folio_batch_init() 44 fbatch->i = 0; in folio_batch_init() 50 fbatch->nr = 0; in folio_batch_reinit() 51 fbatch->i = 0; in folio_batch_reinit() 56 return fbatch->nr; in folio_batch_count() 61 return PAGEVEC_SIZE - fbatch->nr; in folio_batch_space() 77 fbatch->folios[fbatch->nr++] = folio; in folio_batch_add() 91 if (fbatch->i == fbatch->nr) in folio_batch_next() 93 return fbatch->folios[fbatch->i++]; in folio_batch_next() 100 if (folio_batch_count(fbatch)) in folio_batch_release() [all …]
|
| H A D | pagemap.h | 962 pgoff_t end, struct folio_batch *fbatch); 964 pgoff_t *start, pgoff_t end, struct folio_batch *fbatch); 966 pgoff_t end, xa_mark_t tag, struct folio_batch *fbatch); 1269 struct folio_batch *fbatch);
|
| /linux-6.15/mm/ |
| H A D | swap.c | 175 folios_put(fbatch); in folio_batch_move_lru() 320 if (folio_batch_count(fbatch)) in folio_activate_drain() 353 struct folio_batch *fbatch; in __lru_cache_activate_folio() local 641 if (folio_batch_count(fbatch)) in lru_add_drain_cpu() 656 if (folio_batch_count(fbatch)) in lru_add_drain_cpu() 1010 struct folio_batch fbatch; in release_pages() local 1015 folio_batch_init(&fbatch); in release_pages() 1021 refs[fbatch.nr] = 1; in release_pages() 1031 if (fbatch.nr) in release_pages() 1052 folios_put(fbatch); in __folio_batch_release() [all …]
|
| H A D | truncate.c | 346 struct folio_batch fbatch; in truncate_inode_pages_range() local 373 folio_batch_init(&fbatch); in truncate_inode_pages_range() 376 &fbatch, indices)) { in truncate_inode_pages_range() 383 folio_batch_release(&fbatch); in truncate_inode_pages_range() 440 folio_batch_release(&fbatch); in truncate_inode_pages_range() 513 struct folio_batch fbatch; in mapping_try_invalidate() local 519 folio_batch_init(&fbatch); in mapping_try_invalidate() 554 folio_batch_release(&fbatch); in mapping_try_invalidate() 651 struct folio_batch fbatch; in invalidate_inode_pages2_range() local 661 folio_batch_init(&fbatch); in invalidate_inode_pages2_range() [all …]
|
| H A D | mlock.c | 33 struct folio_batch fbatch; member 194 folio = fbatch->folios[i]; in mlock_folio_batch() 197 fbatch->folios[i] = folio; in mlock_folio_batch() 209 folios_put(fbatch); in mlock_folio_batch() 214 struct folio_batch *fbatch; in mlock_drain_local() local 217 fbatch = this_cpu_ptr(&mlock_fbatch.fbatch); in mlock_drain_local() 219 mlock_folio_batch(fbatch); in mlock_drain_local() 228 fbatch = &per_cpu(mlock_fbatch.fbatch, cpu); in mlock_drain_remote() 247 fbatch = this_cpu_ptr(&mlock_fbatch.fbatch); in mlock_folio() 273 fbatch = this_cpu_ptr(&mlock_fbatch.fbatch); in mlock_new_folio() [all …]
|
| H A D | filemap.c | 522 struct folio_batch fbatch; in __filemap_fdatawait_range() local 525 folio_batch_init(&fbatch); in __filemap_fdatawait_range() 541 folio_batch_release(&fbatch); in __filemap_fdatawait_range() 2603 folio = fbatch->folios[folio_batch_count(fbatch) - 1]; in filemap_get_pages() 2624 if (likely(--fbatch->nr)) in filemap_get_pages() 2672 struct folio_batch fbatch; in filemap_read() local 2686 folio_batch_init(&fbatch); in filemap_read() 2730 fbatch.folios[0])) in filemap_read() 2771 folio_batch_init(&fbatch); in filemap_read() 2954 struct folio_batch fbatch; in filemap_splice_read() local [all …]
|
| H A D | shmem.c | 1037 struct folio_batch fbatch; in shmem_unlock_mapping() local 1040 folio_batch_init(&fbatch); in shmem_unlock_mapping() 1047 folio_batch_release(&fbatch); in shmem_unlock_mapping() 1091 struct folio_batch fbatch; in shmem_undo_range() local 1105 folio_batch_init(&fbatch); in shmem_undo_range() 1108 &fbatch, indices)) { in shmem_undo_range() 1110 folio = fbatch.folios[i]; in shmem_undo_range() 1125 folio_batch_release(&fbatch); in shmem_undo_range() 1179 folio = fbatch.folios[i]; in shmem_undo_range() 1472 struct folio_batch fbatch; in shmem_unuse_inode() local [all …]
|
| /linux-6.15/fs/nilfs2/ |
| H A D | page.c | 246 struct folio_batch fbatch; in nilfs_copy_dirty_pages() local 251 folio_batch_init(&fbatch); in nilfs_copy_dirty_pages() 282 folio_batch_release(&fbatch); in nilfs_copy_dirty_pages() 301 struct folio_batch fbatch; in nilfs_copy_back_pages() local 305 folio_batch_init(&fbatch); in nilfs_copy_back_pages() 363 struct folio_batch fbatch; in nilfs_clear_dirty_pages() local 367 folio_batch_init(&fbatch); in nilfs_clear_dirty_pages() 513 struct folio_batch fbatch; in nilfs_find_uncommitted_extent() local 521 folio_batch_init(&fbatch); in nilfs_find_uncommitted_extent() 525 &fbatch); in nilfs_find_uncommitted_extent() [all …]
|
| H A D | segment.c | 704 struct folio_batch fbatch; in nilfs_lookup_dirty_data_buffers() local 718 folio_batch_init(&fbatch); in nilfs_lookup_dirty_data_buffers() 722 PAGECACHE_TAG_DIRTY, &fbatch)) in nilfs_lookup_dirty_data_buffers() 727 struct folio *folio = fbatch.folios[i]; in nilfs_lookup_dirty_data_buffers() 749 folio_batch_release(&fbatch); in nilfs_lookup_dirty_data_buffers() 757 folio_batch_release(&fbatch); in nilfs_lookup_dirty_data_buffers() 767 struct folio_batch fbatch; in nilfs_lookup_dirty_node_buffers() local 774 folio_batch_init(&fbatch); in nilfs_lookup_dirty_node_buffers() 777 (pgoff_t)-1, PAGECACHE_TAG_DIRTY, &fbatch)) { in nilfs_lookup_dirty_node_buffers() 779 bh = head = folio_buffers(fbatch.folios[i]); in nilfs_lookup_dirty_node_buffers() [all …]
|
| /linux-6.15/fs/ramfs/ |
| H A D | file-nommu.c | 208 struct folio_batch fbatch; in ramfs_nommu_get_unmapped_area() local 224 folio_batch_init(&fbatch); in ramfs_nommu_get_unmapped_area() 228 ULONG_MAX, &fbatch); in ramfs_nommu_get_unmapped_area() 235 ret = (unsigned long) folio_address(fbatch.folios[0]); in ramfs_nommu_get_unmapped_area() 236 pfn = folio_pfn(fbatch.folios[0]); in ramfs_nommu_get_unmapped_area() 240 if (pfn + nr_pages != folio_pfn(fbatch.folios[loop])) { in ramfs_nommu_get_unmapped_area() 244 nr_pages += folio_nr_pages(fbatch.folios[loop]); in ramfs_nommu_get_unmapped_area() 250 folio_batch_release(&fbatch); in ramfs_nommu_get_unmapped_area() 256 folio_batch_release(&fbatch); in ramfs_nommu_get_unmapped_area()
|
| /linux-6.15/fs/bcachefs/ |
| H A D | fs-io-pagecache.c | 255 struct folio_batch fbatch; in bch2_mark_pagecache_unallocated() local 261 folio_batch_init(&fbatch); in bch2_mark_pagecache_unallocated() 287 folio_batch_release(&fbatch); in bch2_mark_pagecache_unallocated() 299 struct folio_batch fbatch; in bch2_mark_pagecache_reserved() local 306 folio_batch_init(&fbatch); in bch2_mark_pagecache_reserved() 316 folio_batch_release(&fbatch); in bch2_mark_pagecache_reserved() 344 folio_batch_release(&fbatch); in bch2_mark_pagecache_reserved() 699 struct folio_batch fbatch; in bch2_seek_pagecache_data() local 707 folio_batch_init(&fbatch); in bch2_seek_pagecache_data() 717 folio_batch_release(&fbatch); in bch2_seek_pagecache_data() [all …]
|
| /linux-6.15/fs/netfs/ |
| H A D | rolling_buffer.c | 199 struct folio_batch fbatch; in rolling_buffer_clear() local 202 folio_batch_init(&fbatch); in rolling_buffer_clear() 213 if (!folio_batch_add(&fbatch, folio)) in rolling_buffer_clear() 214 folio_batch_release(&fbatch); in rolling_buffer_clear() 221 folio_batch_release(&fbatch); in rolling_buffer_clear()
|
| H A D | misc.c | 81 struct folio_batch fbatch; in netfs_free_folioq_buffer() local 83 folio_batch_init(&fbatch); in netfs_free_folioq_buffer() 94 if (folio_batch_add(&fbatch, folio)) in netfs_free_folioq_buffer() 95 folio_batch_release(&fbatch); in netfs_free_folioq_buffer() 103 folio_batch_release(&fbatch); in netfs_free_folioq_buffer()
|
| /linux-6.15/drivers/gpu/drm/ |
| H A D | drm_gem.c | 533 check_move_unevictable_folios(fbatch); in drm_gem_check_release_batch() 534 __folio_batch_release(fbatch); in drm_gem_check_release_batch() 568 struct folio_batch fbatch; in drm_gem_get_pages() local 615 folio_batch_init(&fbatch); in drm_gem_get_pages() 619 if (!folio_batch_add(&fbatch, f)) in drm_gem_get_pages() 623 if (fbatch.nr) in drm_gem_get_pages() 624 drm_gem_check_release_batch(&fbatch); in drm_gem_get_pages() 643 struct folio_batch fbatch; in drm_gem_put_pages() local 656 folio_batch_init(&fbatch); in drm_gem_put_pages() 671 if (!folio_batch_add(&fbatch, folio)) in drm_gem_put_pages() [all …]
|
| /linux-6.15/drivers/gpu/drm/i915/gem/ |
| H A D | i915_gem_shmem.c | 25 static void check_release_folio_batch(struct folio_batch *fbatch) in check_release_folio_batch() argument 27 check_move_unevictable_folios(fbatch); in check_release_folio_batch() 28 __folio_batch_release(fbatch); in check_release_folio_batch() 36 struct folio_batch fbatch; in shmem_sg_free_table() local 42 folio_batch_init(&fbatch); in shmem_sg_free_table() 54 if (!folio_batch_add(&fbatch, folio)) in shmem_sg_free_table() 55 check_release_folio_batch(&fbatch); in shmem_sg_free_table() 57 if (fbatch.nr) in shmem_sg_free_table() 58 check_release_folio_batch(&fbatch); in shmem_sg_free_table()
|
| /linux-6.15/fs/gfs2/ |
| H A D | aops.c | 179 struct folio_batch *fbatch, in gfs2_write_jdata_batch() argument 188 int nr_folios = folio_batch_count(fbatch); in gfs2_write_jdata_batch() 191 size += folio_size(fbatch->folios[i]); in gfs2_write_jdata_batch() 199 struct folio *folio = fbatch->folios[i]; in gfs2_write_jdata_batch() 282 struct folio_batch fbatch; in gfs2_write_cache_jdata() local 292 folio_batch_init(&fbatch); in gfs2_write_cache_jdata() 319 tag, &fbatch); in gfs2_write_cache_jdata() 323 ret = gfs2_write_jdata_batch(mapping, wbc, &fbatch, in gfs2_write_cache_jdata() 329 folio_batch_release(&fbatch); in gfs2_write_cache_jdata()
|
| /linux-6.15/fs/ceph/ |
| H A D | addr.c | 598 struct folio_batch fbatch; member 1003 folio_batch_init(&ceph_wbc->fbatch); in ceph_folio_batch_init() 1289 folio = ceph_wbc->fbatch.folios[i]; in ceph_process_folio_batch() 1368 ceph_wbc->fbatch.folios[i] = NULL; in ceph_process_folio_batch() 1384 if (!fbatch->folios[j]) in ceph_shift_unused_folios_left() 1388 fbatch->folios[n] = fbatch->folios[j]; in ceph_shift_unused_folios_left() 1394 fbatch->nr = n; in ceph_shift_unused_folios_left() 1602 &ceph_wbc->fbatch))) { in ceph_wait_until_current_writes_complete() 1670 &ceph_wbc.fbatch); in ceph_writepages_start() 1720 (int)ceph_wbc.fbatch.nr, in ceph_writepages_start() [all …]
|
| /linux-6.15/fs/f2fs/ |
| H A D | node.c | 1614 struct folio_batch fbatch; in last_fsync_dnode() local 1618 folio_batch_init(&fbatch); in last_fsync_dnode() 1623 &fbatch))) { in last_fsync_dnode() 1824 struct folio_batch fbatch; in f2fs_fsync_node_pages() local 1838 folio_batch_init(&fbatch); in f2fs_fsync_node_pages() 1843 &fbatch))) { in f2fs_fsync_node_pages() 1983 struct folio_batch fbatch; in f2fs_flush_inline_data() local 1986 folio_batch_init(&fbatch); in f2fs_flush_inline_data() 1990 &fbatch))) { in f2fs_flush_inline_data() 2026 struct folio_batch fbatch; in f2fs_sync_node_pages() local [all …]
|
| H A D | checkpoint.c | 421 struct folio_batch fbatch; in f2fs_sync_meta_pages() local 429 folio_batch_init(&fbatch); in f2fs_sync_meta_pages() 435 PAGECACHE_TAG_DIRTY, &fbatch))) { in f2fs_sync_meta_pages() 439 struct folio *folio = fbatch.folios[i]; in f2fs_sync_meta_pages() 443 folio_nr_pages(fbatch.folios[i-1])) { in f2fs_sync_meta_pages() 444 folio_batch_release(&fbatch); in f2fs_sync_meta_pages() 476 folio_batch_release(&fbatch); in f2fs_sync_meta_pages()
|
| /linux-6.15/fs/btrfs/tests/ |
| H A D | extent-io-tests.c | 25 struct folio_batch fbatch; in process_page_range() local 32 folio_batch_init(&fbatch); in process_page_range() 36 end_index, &fbatch); in process_page_range() 38 struct folio *folio = fbatch.folios[i]; in process_page_range() 48 folio_batch_release(&fbatch); in process_page_range()
|
| /linux-6.15/drivers/gpu/drm/i915/ |
| H A D | i915_gpu_error.c | 191 static void pool_fini(struct folio_batch *fbatch) in pool_fini() argument 193 folio_batch_release(fbatch); in pool_fini() 198 while (folio_batch_space(fbatch)) { in pool_refill() 205 folio_batch_add(fbatch, folio); in pool_refill() 215 folio_batch_init(fbatch); in pool_init() 217 err = pool_refill(fbatch, gfp); in pool_init() 219 pool_fini(fbatch); in pool_init() 229 if (!folio && folio_batch_count(fbatch)) in pool_alloc() 230 folio = fbatch->folios[--fbatch->nr]; in pool_alloc() 239 if (folio_batch_space(fbatch)) in pool_free() [all …]
|
| /linux-6.15/fs/btrfs/ |
| H A D | extent_io.c | 203 struct folio_batch fbatch; in __process_folios_contig() local 206 folio_batch_init(&fbatch); in __process_folios_contig() 211 end_index, &fbatch); in __process_folios_contig() 247 struct folio_batch fbatch; in lock_delalloc_folios() local 252 folio_batch_init(&fbatch); in lock_delalloc_folios() 257 end_index, &fbatch); in lock_delalloc_folios() 287 folio_batch_release(&fbatch); in lock_delalloc_folios() 2174 struct folio_batch fbatch; in btree_write_cache_pages() local 2181 folio_batch_init(&fbatch); in btree_write_cache_pages() 2205 tag, &fbatch))) { in btree_write_cache_pages() [all …]
|
| H A D | compression.c | 287 struct folio_batch fbatch; in end_compressed_writeback() local 295 folio_batch_init(&fbatch); in end_compressed_writeback() 298 &fbatch); in end_compressed_writeback() 304 struct folio *folio = fbatch.folios[i]; in end_compressed_writeback() 309 folio_batch_release(&fbatch); in end_compressed_writeback()
|
| /linux-6.15/fs/hugetlbfs/ |
| H A D | inode.c | 577 struct folio_batch fbatch; in remove_inode_hugepages() local 582 folio_batch_init(&fbatch); in remove_inode_hugepages() 584 while (filemap_get_folios(mapping, &next, end - 1, &fbatch)) { in remove_inode_hugepages() 585 for (i = 0; i < folio_batch_count(&fbatch); ++i) { in remove_inode_hugepages() 586 struct folio *folio = fbatch.folios[i]; in remove_inode_hugepages() 602 folio_batch_release(&fbatch); in remove_inode_hugepages()
|
| /linux-6.15/fs/ext4/ |
| H A D | inode.c | 1572 struct folio_batch fbatch; in mpage_release_unused_pages() local 1597 folio_batch_init(&fbatch); in mpage_release_unused_pages() 1620 folio_batch_release(&fbatch); in mpage_release_unused_pages() 2150 struct folio_batch fbatch; in mpage_map_and_submit_buffers() local 2165 folio_batch_init(&fbatch); in mpage_map_and_submit_buffers() 2188 folio_batch_release(&fbatch); in mpage_map_and_submit_buffers() 2195 folio_batch_release(&fbatch); in mpage_map_and_submit_buffers() 2426 struct folio_batch fbatch; in mpage_prepare_extent_to_map() local 2451 folio_batch_init(&fbatch); in mpage_prepare_extent_to_map() 2454 tag, &fbatch); in mpage_prepare_extent_to_map() [all …]
|