Home
last modified time | relevance | path

Searched refs:fbatch (Results 1 – 25 of 35) sorted by relevance

12

/linux-6.15/include/linux/
H A Dpagevec.h43 fbatch->nr = 0; in folio_batch_init()
44 fbatch->i = 0; in folio_batch_init()
50 fbatch->nr = 0; in folio_batch_reinit()
51 fbatch->i = 0; in folio_batch_reinit()
56 return fbatch->nr; in folio_batch_count()
61 return PAGEVEC_SIZE - fbatch->nr; in folio_batch_space()
77 fbatch->folios[fbatch->nr++] = folio; in folio_batch_add()
91 if (fbatch->i == fbatch->nr) in folio_batch_next()
93 return fbatch->folios[fbatch->i++]; in folio_batch_next()
100 if (folio_batch_count(fbatch)) in folio_batch_release()
[all …]
H A Dpagemap.h962 pgoff_t end, struct folio_batch *fbatch);
964 pgoff_t *start, pgoff_t end, struct folio_batch *fbatch);
966 pgoff_t end, xa_mark_t tag, struct folio_batch *fbatch);
1269 struct folio_batch *fbatch);
/linux-6.15/mm/
H A Dswap.c175 folios_put(fbatch); in folio_batch_move_lru()
320 if (folio_batch_count(fbatch)) in folio_activate_drain()
353 struct folio_batch *fbatch; in __lru_cache_activate_folio() local
641 if (folio_batch_count(fbatch)) in lru_add_drain_cpu()
656 if (folio_batch_count(fbatch)) in lru_add_drain_cpu()
1010 struct folio_batch fbatch; in release_pages() local
1015 folio_batch_init(&fbatch); in release_pages()
1021 refs[fbatch.nr] = 1; in release_pages()
1031 if (fbatch.nr) in release_pages()
1052 folios_put(fbatch); in __folio_batch_release()
[all …]
H A Dtruncate.c346 struct folio_batch fbatch; in truncate_inode_pages_range() local
373 folio_batch_init(&fbatch); in truncate_inode_pages_range()
376 &fbatch, indices)) { in truncate_inode_pages_range()
383 folio_batch_release(&fbatch); in truncate_inode_pages_range()
440 folio_batch_release(&fbatch); in truncate_inode_pages_range()
513 struct folio_batch fbatch; in mapping_try_invalidate() local
519 folio_batch_init(&fbatch); in mapping_try_invalidate()
554 folio_batch_release(&fbatch); in mapping_try_invalidate()
651 struct folio_batch fbatch; in invalidate_inode_pages2_range() local
661 folio_batch_init(&fbatch); in invalidate_inode_pages2_range()
[all …]
H A Dmlock.c33 struct folio_batch fbatch; member
194 folio = fbatch->folios[i]; in mlock_folio_batch()
197 fbatch->folios[i] = folio; in mlock_folio_batch()
209 folios_put(fbatch); in mlock_folio_batch()
214 struct folio_batch *fbatch; in mlock_drain_local() local
217 fbatch = this_cpu_ptr(&mlock_fbatch.fbatch); in mlock_drain_local()
219 mlock_folio_batch(fbatch); in mlock_drain_local()
228 fbatch = &per_cpu(mlock_fbatch.fbatch, cpu); in mlock_drain_remote()
247 fbatch = this_cpu_ptr(&mlock_fbatch.fbatch); in mlock_folio()
273 fbatch = this_cpu_ptr(&mlock_fbatch.fbatch); in mlock_new_folio()
[all …]
H A Dfilemap.c522 struct folio_batch fbatch; in __filemap_fdatawait_range() local
525 folio_batch_init(&fbatch); in __filemap_fdatawait_range()
541 folio_batch_release(&fbatch); in __filemap_fdatawait_range()
2603 folio = fbatch->folios[folio_batch_count(fbatch) - 1]; in filemap_get_pages()
2624 if (likely(--fbatch->nr)) in filemap_get_pages()
2672 struct folio_batch fbatch; in filemap_read() local
2686 folio_batch_init(&fbatch); in filemap_read()
2730 fbatch.folios[0])) in filemap_read()
2771 folio_batch_init(&fbatch); in filemap_read()
2954 struct folio_batch fbatch; in filemap_splice_read() local
[all …]
H A Dshmem.c1037 struct folio_batch fbatch; in shmem_unlock_mapping() local
1040 folio_batch_init(&fbatch); in shmem_unlock_mapping()
1047 folio_batch_release(&fbatch); in shmem_unlock_mapping()
1091 struct folio_batch fbatch; in shmem_undo_range() local
1105 folio_batch_init(&fbatch); in shmem_undo_range()
1108 &fbatch, indices)) { in shmem_undo_range()
1110 folio = fbatch.folios[i]; in shmem_undo_range()
1125 folio_batch_release(&fbatch); in shmem_undo_range()
1179 folio = fbatch.folios[i]; in shmem_undo_range()
1472 struct folio_batch fbatch; in shmem_unuse_inode() local
[all …]
/linux-6.15/fs/nilfs2/
H A Dpage.c246 struct folio_batch fbatch; in nilfs_copy_dirty_pages() local
251 folio_batch_init(&fbatch); in nilfs_copy_dirty_pages()
282 folio_batch_release(&fbatch); in nilfs_copy_dirty_pages()
301 struct folio_batch fbatch; in nilfs_copy_back_pages() local
305 folio_batch_init(&fbatch); in nilfs_copy_back_pages()
363 struct folio_batch fbatch; in nilfs_clear_dirty_pages() local
367 folio_batch_init(&fbatch); in nilfs_clear_dirty_pages()
513 struct folio_batch fbatch; in nilfs_find_uncommitted_extent() local
521 folio_batch_init(&fbatch); in nilfs_find_uncommitted_extent()
525 &fbatch); in nilfs_find_uncommitted_extent()
[all …]
H A Dsegment.c704 struct folio_batch fbatch; in nilfs_lookup_dirty_data_buffers() local
718 folio_batch_init(&fbatch); in nilfs_lookup_dirty_data_buffers()
722 PAGECACHE_TAG_DIRTY, &fbatch)) in nilfs_lookup_dirty_data_buffers()
727 struct folio *folio = fbatch.folios[i]; in nilfs_lookup_dirty_data_buffers()
749 folio_batch_release(&fbatch); in nilfs_lookup_dirty_data_buffers()
757 folio_batch_release(&fbatch); in nilfs_lookup_dirty_data_buffers()
767 struct folio_batch fbatch; in nilfs_lookup_dirty_node_buffers() local
774 folio_batch_init(&fbatch); in nilfs_lookup_dirty_node_buffers()
777 (pgoff_t)-1, PAGECACHE_TAG_DIRTY, &fbatch)) { in nilfs_lookup_dirty_node_buffers()
779 bh = head = folio_buffers(fbatch.folios[i]); in nilfs_lookup_dirty_node_buffers()
[all …]
/linux-6.15/fs/ramfs/
H A Dfile-nommu.c208 struct folio_batch fbatch; in ramfs_nommu_get_unmapped_area() local
224 folio_batch_init(&fbatch); in ramfs_nommu_get_unmapped_area()
228 ULONG_MAX, &fbatch); in ramfs_nommu_get_unmapped_area()
235 ret = (unsigned long) folio_address(fbatch.folios[0]); in ramfs_nommu_get_unmapped_area()
236 pfn = folio_pfn(fbatch.folios[0]); in ramfs_nommu_get_unmapped_area()
240 if (pfn + nr_pages != folio_pfn(fbatch.folios[loop])) { in ramfs_nommu_get_unmapped_area()
244 nr_pages += folio_nr_pages(fbatch.folios[loop]); in ramfs_nommu_get_unmapped_area()
250 folio_batch_release(&fbatch); in ramfs_nommu_get_unmapped_area()
256 folio_batch_release(&fbatch); in ramfs_nommu_get_unmapped_area()
/linux-6.15/fs/bcachefs/
H A Dfs-io-pagecache.c255 struct folio_batch fbatch; in bch2_mark_pagecache_unallocated() local
261 folio_batch_init(&fbatch); in bch2_mark_pagecache_unallocated()
287 folio_batch_release(&fbatch); in bch2_mark_pagecache_unallocated()
299 struct folio_batch fbatch; in bch2_mark_pagecache_reserved() local
306 folio_batch_init(&fbatch); in bch2_mark_pagecache_reserved()
316 folio_batch_release(&fbatch); in bch2_mark_pagecache_reserved()
344 folio_batch_release(&fbatch); in bch2_mark_pagecache_reserved()
699 struct folio_batch fbatch; in bch2_seek_pagecache_data() local
707 folio_batch_init(&fbatch); in bch2_seek_pagecache_data()
717 folio_batch_release(&fbatch); in bch2_seek_pagecache_data()
[all …]
/linux-6.15/fs/netfs/
H A Drolling_buffer.c199 struct folio_batch fbatch; in rolling_buffer_clear() local
202 folio_batch_init(&fbatch); in rolling_buffer_clear()
213 if (!folio_batch_add(&fbatch, folio)) in rolling_buffer_clear()
214 folio_batch_release(&fbatch); in rolling_buffer_clear()
221 folio_batch_release(&fbatch); in rolling_buffer_clear()
H A Dmisc.c81 struct folio_batch fbatch; in netfs_free_folioq_buffer() local
83 folio_batch_init(&fbatch); in netfs_free_folioq_buffer()
94 if (folio_batch_add(&fbatch, folio)) in netfs_free_folioq_buffer()
95 folio_batch_release(&fbatch); in netfs_free_folioq_buffer()
103 folio_batch_release(&fbatch); in netfs_free_folioq_buffer()
/linux-6.15/drivers/gpu/drm/
H A Ddrm_gem.c533 check_move_unevictable_folios(fbatch); in drm_gem_check_release_batch()
534 __folio_batch_release(fbatch); in drm_gem_check_release_batch()
568 struct folio_batch fbatch; in drm_gem_get_pages() local
615 folio_batch_init(&fbatch); in drm_gem_get_pages()
619 if (!folio_batch_add(&fbatch, f)) in drm_gem_get_pages()
623 if (fbatch.nr) in drm_gem_get_pages()
624 drm_gem_check_release_batch(&fbatch); in drm_gem_get_pages()
643 struct folio_batch fbatch; in drm_gem_put_pages() local
656 folio_batch_init(&fbatch); in drm_gem_put_pages()
671 if (!folio_batch_add(&fbatch, folio)) in drm_gem_put_pages()
[all …]
/linux-6.15/drivers/gpu/drm/i915/gem/
H A Di915_gem_shmem.c25 static void check_release_folio_batch(struct folio_batch *fbatch) in check_release_folio_batch() argument
27 check_move_unevictable_folios(fbatch); in check_release_folio_batch()
28 __folio_batch_release(fbatch); in check_release_folio_batch()
36 struct folio_batch fbatch; in shmem_sg_free_table() local
42 folio_batch_init(&fbatch); in shmem_sg_free_table()
54 if (!folio_batch_add(&fbatch, folio)) in shmem_sg_free_table()
55 check_release_folio_batch(&fbatch); in shmem_sg_free_table()
57 if (fbatch.nr) in shmem_sg_free_table()
58 check_release_folio_batch(&fbatch); in shmem_sg_free_table()
/linux-6.15/fs/gfs2/
H A Daops.c179 struct folio_batch *fbatch, in gfs2_write_jdata_batch() argument
188 int nr_folios = folio_batch_count(fbatch); in gfs2_write_jdata_batch()
191 size += folio_size(fbatch->folios[i]); in gfs2_write_jdata_batch()
199 struct folio *folio = fbatch->folios[i]; in gfs2_write_jdata_batch()
282 struct folio_batch fbatch; in gfs2_write_cache_jdata() local
292 folio_batch_init(&fbatch); in gfs2_write_cache_jdata()
319 tag, &fbatch); in gfs2_write_cache_jdata()
323 ret = gfs2_write_jdata_batch(mapping, wbc, &fbatch, in gfs2_write_cache_jdata()
329 folio_batch_release(&fbatch); in gfs2_write_cache_jdata()
/linux-6.15/fs/ceph/
H A Daddr.c598 struct folio_batch fbatch; member
1003 folio_batch_init(&ceph_wbc->fbatch); in ceph_folio_batch_init()
1289 folio = ceph_wbc->fbatch.folios[i]; in ceph_process_folio_batch()
1368 ceph_wbc->fbatch.folios[i] = NULL; in ceph_process_folio_batch()
1384 if (!fbatch->folios[j]) in ceph_shift_unused_folios_left()
1388 fbatch->folios[n] = fbatch->folios[j]; in ceph_shift_unused_folios_left()
1394 fbatch->nr = n; in ceph_shift_unused_folios_left()
1602 &ceph_wbc->fbatch))) { in ceph_wait_until_current_writes_complete()
1670 &ceph_wbc.fbatch); in ceph_writepages_start()
1720 (int)ceph_wbc.fbatch.nr, in ceph_writepages_start()
[all …]
/linux-6.15/fs/f2fs/
H A Dnode.c1614 struct folio_batch fbatch; in last_fsync_dnode() local
1618 folio_batch_init(&fbatch); in last_fsync_dnode()
1623 &fbatch))) { in last_fsync_dnode()
1824 struct folio_batch fbatch; in f2fs_fsync_node_pages() local
1838 folio_batch_init(&fbatch); in f2fs_fsync_node_pages()
1843 &fbatch))) { in f2fs_fsync_node_pages()
1983 struct folio_batch fbatch; in f2fs_flush_inline_data() local
1986 folio_batch_init(&fbatch); in f2fs_flush_inline_data()
1990 &fbatch))) { in f2fs_flush_inline_data()
2026 struct folio_batch fbatch; in f2fs_sync_node_pages() local
[all …]
H A Dcheckpoint.c421 struct folio_batch fbatch; in f2fs_sync_meta_pages() local
429 folio_batch_init(&fbatch); in f2fs_sync_meta_pages()
435 PAGECACHE_TAG_DIRTY, &fbatch))) { in f2fs_sync_meta_pages()
439 struct folio *folio = fbatch.folios[i]; in f2fs_sync_meta_pages()
443 folio_nr_pages(fbatch.folios[i-1])) { in f2fs_sync_meta_pages()
444 folio_batch_release(&fbatch); in f2fs_sync_meta_pages()
476 folio_batch_release(&fbatch); in f2fs_sync_meta_pages()
/linux-6.15/fs/btrfs/tests/
H A Dextent-io-tests.c25 struct folio_batch fbatch; in process_page_range() local
32 folio_batch_init(&fbatch); in process_page_range()
36 end_index, &fbatch); in process_page_range()
38 struct folio *folio = fbatch.folios[i]; in process_page_range()
48 folio_batch_release(&fbatch); in process_page_range()
/linux-6.15/drivers/gpu/drm/i915/
H A Di915_gpu_error.c191 static void pool_fini(struct folio_batch *fbatch) in pool_fini() argument
193 folio_batch_release(fbatch); in pool_fini()
198 while (folio_batch_space(fbatch)) { in pool_refill()
205 folio_batch_add(fbatch, folio); in pool_refill()
215 folio_batch_init(fbatch); in pool_init()
217 err = pool_refill(fbatch, gfp); in pool_init()
219 pool_fini(fbatch); in pool_init()
229 if (!folio && folio_batch_count(fbatch)) in pool_alloc()
230 folio = fbatch->folios[--fbatch->nr]; in pool_alloc()
239 if (folio_batch_space(fbatch)) in pool_free()
[all …]
/linux-6.15/fs/btrfs/
H A Dextent_io.c203 struct folio_batch fbatch; in __process_folios_contig() local
206 folio_batch_init(&fbatch); in __process_folios_contig()
211 end_index, &fbatch); in __process_folios_contig()
247 struct folio_batch fbatch; in lock_delalloc_folios() local
252 folio_batch_init(&fbatch); in lock_delalloc_folios()
257 end_index, &fbatch); in lock_delalloc_folios()
287 folio_batch_release(&fbatch); in lock_delalloc_folios()
2174 struct folio_batch fbatch; in btree_write_cache_pages() local
2181 folio_batch_init(&fbatch); in btree_write_cache_pages()
2205 tag, &fbatch))) { in btree_write_cache_pages()
[all …]
H A Dcompression.c287 struct folio_batch fbatch; in end_compressed_writeback() local
295 folio_batch_init(&fbatch); in end_compressed_writeback()
298 &fbatch); in end_compressed_writeback()
304 struct folio *folio = fbatch.folios[i]; in end_compressed_writeback()
309 folio_batch_release(&fbatch); in end_compressed_writeback()
/linux-6.15/fs/hugetlbfs/
H A Dinode.c577 struct folio_batch fbatch; in remove_inode_hugepages() local
582 folio_batch_init(&fbatch); in remove_inode_hugepages()
584 while (filemap_get_folios(mapping, &next, end - 1, &fbatch)) { in remove_inode_hugepages()
585 for (i = 0; i < folio_batch_count(&fbatch); ++i) { in remove_inode_hugepages()
586 struct folio *folio = fbatch.folios[i]; in remove_inode_hugepages()
602 folio_batch_release(&fbatch); in remove_inode_hugepages()
/linux-6.15/fs/ext4/
H A Dinode.c1572 struct folio_batch fbatch; in mpage_release_unused_pages() local
1597 folio_batch_init(&fbatch); in mpage_release_unused_pages()
1620 folio_batch_release(&fbatch); in mpage_release_unused_pages()
2150 struct folio_batch fbatch; in mpage_map_and_submit_buffers() local
2165 folio_batch_init(&fbatch); in mpage_map_and_submit_buffers()
2188 folio_batch_release(&fbatch); in mpage_map_and_submit_buffers()
2195 folio_batch_release(&fbatch); in mpage_map_and_submit_buffers()
2426 struct folio_batch fbatch; in mpage_prepare_extent_to_map() local
2451 folio_batch_init(&fbatch); in mpage_prepare_extent_to_map()
2454 tag, &fbatch); in mpage_prepare_extent_to_map()
[all …]

12