| /linux-6.15/drivers/net/wireless/realtek/rtlwifi/ |
| H A D | ps.h | 14 void rtl_lps_enter(struct ieee80211_hw *hw, bool may_block); 15 void rtl_lps_leave(struct ieee80211_hw *hw, bool may_block);
|
| H A D | ps.c | 656 void rtl_lps_enter(struct ieee80211_hw *hw, bool may_block) in rtl_lps_enter() argument 660 if (may_block) in rtl_lps_enter() 667 void rtl_lps_leave(struct ieee80211_hw *hw, bool may_block) in rtl_lps_leave() argument 671 if (may_block) in rtl_lps_leave()
|
| /linux-6.15/include/linux/ |
| H A D | dm-region-hash.h | 67 int dm_rh_get_state(struct dm_region_hash *rh, region_t region, int may_block); 69 enum dm_rh_region_states state, int may_block);
|
| H A D | nfs_fs.h | 581 u32 *mask, bool may_block);
|
| H A D | kvm_host.h | 269 bool may_block; member
|
| /linux-6.15/arch/riscv/kvm/ |
| H A D | mmu.c | 270 gpa_t size, bool may_block) in gstage_unmap_range() argument 300 if (may_block && addr < end) in gstage_unmap_range() 549 range->may_block); in kvm_unmap_gfn_range()
|
| /linux-6.15/arch/arm64/kvm/ |
| H A D | mmu.c | 333 bool may_block) in __unmap_stage2_range() argument 341 may_block)); in __unmap_stage2_range() 345 u64 size, bool may_block) in kvm_stage2_unmap_range() argument 347 __unmap_stage2_range(mmu, start, size, may_block); in kvm_stage2_unmap_range() 1973 range->may_block); in kvm_unmap_gfn_range() 1975 kvm_nested_s2_unmap(kvm, range->may_block); in kvm_unmap_gfn_range()
|
| H A D | nested.c | 760 void kvm_nested_s2_unmap(struct kvm *kvm, bool may_block) in kvm_nested_s2_unmap() argument 770 kvm_stage2_unmap_range(mmu, 0, kvm_phys_size(mmu), may_block); in kvm_nested_s2_unmap()
|
| /linux-6.15/arch/arm64/include/asm/ |
| H A D | kvm_nested.h | 130 extern void kvm_nested_s2_unmap(struct kvm *kvm, bool may_block);
|
| H A D | kvm_mmu.h | 172 u64 size, bool may_block);
|
| /linux-6.15/drivers/md/ |
| H A D | dm-region-hash.c | 341 int dm_rh_get_state(struct dm_region_hash *rh, region_t region, int may_block) in dm_rh_get_state() argument 357 r = rh->log->type->in_sync(rh->log, region, may_block); in dm_rh_get_state()
|
| H A D | dm-raid1.c | 561 int may_block) in region_in_sync() argument 563 int state = dm_rh_get_state(ms->rh, region, may_block); in region_in_sync()
|
| /linux-6.15/fs/nfs/ |
| H A D | dir.c | 3044 …s_access_get_cached_locked(struct inode *inode, const struct cred *cred, u32 *mask, bool may_block) in nfs_access_get_cached_locked() argument 3066 if (!may_block) in nfs_access_get_cached_locked() 3123 u32 *mask, bool may_block) in nfs_access_get_cached() argument 3130 may_block); in nfs_access_get_cached() 3250 bool may_block = (mask & MAY_NOT_BLOCK) == 0; in nfs_do_access() local 3256 status = nfs_access_get_cached(inode, cred, &cache.mask, may_block); in nfs_do_access() 3261 if (!may_block) in nfs_do_access()
|
| /linux-6.15/virt/kvm/ |
| H A D | kvm_main.c | 527 bool may_block; member 609 gfn_range.may_block = range->may_block; in kvm_handle_hva_range() 665 .may_block = false, in kvm_age_hva_range() 738 .may_block = mmu_notifier_range_blockable(range), in kvm_mmu_notifier_invalidate_range_start() 814 .may_block = mmu_notifier_range_blockable(range), in kvm_mmu_notifier_invalidate_range_end() 2418 gfn_range.may_block = range->may_block; in kvm_handle_gfn_range() 2488 .may_block = true, in kvm_vm_set_mem_attributes() 2496 .may_block = true, in kvm_vm_set_mem_attributes()
|
| H A D | guest_memfd.c | 120 .may_block = true, in kvm_gmem_invalidate_begin()
|
| /linux-6.15/fs/fuse/ |
| H A D | fuse_i.h | 329 bool may_block:1; member
|
| H A D | virtio_fs.c | 842 if (req->args->may_block) { in virtio_fs_requests_done_work()
|
| H A D | file.c | 815 ia->ap.args.may_block = io->should_dirty; in fuse_async_req_send()
|
| /linux-6.15/arch/x86/kvm/mmu/ |
| H A D | tdp_mmu.c | 1338 range->may_block, flush); in kvm_tdp_mmu_unmap_gfn_range()
|
| H A D | mmu.c | 1660 range->may_block, flush); in kvm_unmap_gfn_range() 7241 .may_block = true, in kvm_mmu_zap_memslot()
|