| /linux-6.15/include/linux/ |
| H A D | blkdev.h | 1227 return &bdev_get_queue(bdev)->limits; in bdev_limits() 1288 return queue_max_segments(bdev_get_queue(bdev)); in bdev_max_segments() 1298 return queue_logical_block_size(bdev_get_queue(bdev)); in bdev_logical_block_size() 1318 return queue_io_min(bdev_get_queue(bdev)); in bdev_io_min() 1328 return queue_io_opt(bdev_get_queue(bdev)); in bdev_io_opt() 1369 return blk_queue_nonrot(bdev_get_queue(bdev)); in bdev_nonrot() 1379 struct request_queue *q = bdev_get_queue(bdev); in bdev_stable_writes() 1395 return blk_queue_write_cache(bdev_get_queue(bdev)); in bdev_write_cache() 1410 return blk_queue_is_zoned(bdev_get_queue(bdev)); in bdev_is_zoned() 1420 struct request_queue *q = bdev_get_queue(bdev); in bdev_zone_sectors() [all …]
|
| /linux-6.15/block/ |
| H A D | blk-crypto.c | 388 return __blk_crypto_cfg_supported(bdev_get_queue(bdev)->crypto_profile, in blk_crypto_config_supported_natively() 453 struct request_queue *q = bdev_get_queue(bdev); in blk_crypto_evict_key() 599 bdev_get_queue(bdev)->crypto_profile; in blk_crypto_ioctl()
|
| H A D | blk-core.c | 681 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in __submit_bio_noacct() 699 if (q == bdev_get_queue(bio->bi_bdev)) in __submit_bio_noacct() 780 struct request_queue *q = bdev_get_queue(bdev); in submit_bio_noacct() 943 q = bdev_get_queue(bdev); in bio_poll()
|
| H A D | bio-integrity.c | 125 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in bio_integrity_add_page() 264 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in bio_integrity_map_user()
|
| H A D | blk-rq-qos.h | 141 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in rq_qos_done_bio()
|
| H A D | blk-settings.c | 927 struct request_queue *q = bdev_get_queue(bdev); in bdev_alignment_offset() 940 struct request_queue *q = bdev_get_queue(bdev); in bdev_discard_alignment()
|
| H A D | blk-crypto-profile.c | 489 bdev_get_queue(bdev)->crypto_profile; in blk_crypto_derive_sw_secret()
|
| H A D | blk-cgroup.c | 997 spin_unlock_irq(&bdev_get_queue(ctx->bdev)->queue_lock); in blkg_conf_exit() 2150 blkg_get(bdev_get_queue(bio->bi_bdev)->root_blkg); in bio_associate_blkg_from_css() 2151 bio->bi_blkg = bdev_get_queue(bio->bi_bdev)->root_blkg; in bio_associate_blkg_from_css()
|
| H A D | blk.h | 75 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in bio_queue_enter()
|
| H A D | ioctl.c | 609 queue_max_sectors(bdev_get_queue(bdev))); in blkdev_common_ioctl()
|
| H A D | genhd.c | 1049 struct request_queue *q = bdev_get_queue(bdev); in part_inflight_show()
|
| H A D | bio.c | 1521 trace_block_bio_complete(bdev_get_queue(bio->bi_bdev), bio); in bio_endio()
|
| H A D | blk-throttle.c | 1618 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in __blk_throtl_bio()
|
| H A D | blk-iocost.c | 3427 q = bdev_get_queue(ctx.bdev); in ioc_cost_model_write()
|
| /linux-6.15/drivers/md/ |
| H A D | dm-table.c | 426 struct request_queue *q = bdev_get_queue(bdev); in dm_set_device_limits() 899 struct request_queue *q = bdev_get_queue(bdev); in device_is_rq_stackable() 1198 bdev_get_queue(dev->bdev)->crypto_profile; in device_intersect_crypto_capabilities() 1723 struct request_queue *q = bdev_get_queue(dev->bdev); in device_not_write_zeroes_capable()
|
| H A D | dm-mpath.c | 539 q = bdev_get_queue(bdev); in multipath_clone_and_map() 884 struct request_queue *q = bdev_get_queue(bdev); in setup_scsi_dh() 960 q = bdev_get_queue(p->path.dev->bdev); in parse_path() 1624 struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev); in activate_or_offline_path() 2096 struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev); in pgpath_busy()
|
| H A D | dm-io.c | 316 struct request_queue *q = bdev_get_queue(where->bdev); in do_region()
|
| H A D | dm-zoned-target.c | 589 if (blk_queue_dying(bdev_get_queue(dmz_dev->bdev))) { in dmz_bdev_is_dying()
|
| /linux-6.15/kernel/trace/ |
| H A D | blktrace.c | 723 struct request_queue *q = bdev_get_queue(bdev); in blk_trace_ioctl() 1755 struct request_queue *q = bdev_get_queue(bdev); in sysfs_blk_trace_attr_show() 1789 struct request_queue *q = bdev_get_queue(bdev); in sysfs_blk_trace_attr_store()
|
| /linux-6.15/drivers/block/rnbd/ |
| H A D | rnbd-srv.c | 553 cpu_to_le32(queue_max_hw_sectors(bdev_get_queue(bdev))); in rnbd_srv_fill_msg_open_rsp()
|
| /linux-6.15/drivers/target/ |
| H A D | target_core_iblock.c | 129 q = bdev_get_queue(bd); in iblock_configure_device()
|
| /linux-6.15/fs/xfs/ |
| H A D | xfs_zone_gc.c | 760 &bdev_get_queue(chunk->bio.bi_bdev)->limits; in xfs_zone_gc_split_write()
|
| /linux-6.15/drivers/md/bcache/ |
| H A D | super.c | 1016 q = bdev_get_queue(dc->bdev); in cached_dev_status_update() 1398 struct request_queue *q = bdev_get_queue(dc->bdev); in cached_dev_init()
|
| /linux-6.15/drivers/block/ |
| H A D | pktcdvd.c | 716 struct request_queue *q = bdev_get_queue(file_bdev(pd->bdev_file)); in pkt_generic_packet() 2193 q = bdev_get_queue(file_bdev(pd->bdev_file)); in pkt_open_dev()
|
| /linux-6.15/drivers/block/drbd/ |
| H A D | drbd_main.c | 918 struct request_queue *q = bdev_get_queue(bdev); in drbd_send_sizes()
|