| /linux-6.15/drivers/pinctrl/ti/ |
| H A D | pinctrl-ti-iodelay.c | 315 iod); in ti_iodelay_pinconf_init_dev() 445 if (!iod) in ti_iodelay_node_iterator() 504 if (!iod) in ti_iodelay_dt_node_to_map() 621 dev = iod->dev; in ti_iodelay_pinconf_group_set() 763 if (!iod->pa) in ti_iodelay_alloc_pins() 766 iod->desc.pins = iod->pa; in ti_iodelay_alloc_pins() 841 iod = devm_kzalloc(dev, sizeof(*iod), GFP_KERNEL); in ti_iodelay_probe() 842 if (!iod) in ti_iodelay_probe() 845 iod->dev = dev; in ti_iodelay_probe() 859 iod->regmap = devm_regmap_init_mmio(dev, iod->reg_base, in ti_iodelay_probe() [all …]
|
| /linux-6.15/drivers/soc/rockchip/ |
| H A D | io-domain.c | 88 struct rockchip_iodomain *iod = supply->iod; in rk3568_iodomain_write() local 130 struct rockchip_iodomain *iod = supply->iod; in rockchip_iodomain_write() local 141 ret = regmap_write(iod->grf, iod->soc_data->grf_offset, val); in rockchip_iodomain_write() 622 iod = devm_kzalloc(&pdev->dev, sizeof(*iod), GFP_KERNEL); in rockchip_iodomain_probe() 623 if (!iod) in rockchip_iodomain_probe() 626 iod->dev = &pdev->dev; in rockchip_iodomain_probe() 633 iod->write = iod->soc_data->write; in rockchip_iodomain_probe() 645 if (IS_ERR(iod->grf)) { in rockchip_iodomain_probe() 684 dev_crit(iod->dev, in rockchip_iodomain_probe() 693 supply->iod = iod; in rockchip_iodomain_probe() [all …]
|
| /linux-6.15/tools/lib/api/ |
| H A D | io_dir.h | 61 iod->dirfd = dirfd; in io_dir__init() 62 iod->available_bytes = 0; in io_dir__init() 65 static inline void io_dir__rewinddir(struct io_dir *iod) in io_dir__rewinddir() argument 67 lseek(iod->dirfd, 0, SEEK_SET); in io_dir__rewinddir() 68 iod->available_bytes = 0; in io_dir__rewinddir() 75 if (iod->available_bytes <= 0) { in io_dir__readdir() 76 ssize_t rc = perf_getdents64(iod->dirfd, iod->buff, sizeof(iod->buff)); in io_dir__readdir() 80 iod->available_bytes = rc; in io_dir__readdir() 81 iod->next = iod->buff; in io_dir__readdir() 83 entry = iod->next; in io_dir__readdir() [all …]
|
| /linux-6.15/drivers/nvme/target/ |
| H A D | pci-epf.c | 683 memset(iod, 0, sizeof(*iod)); in nvmet_pci_epf_alloc_iod() 684 iod->req.cmd = &iod->cmd; in nvmet_pci_epf_alloc_iod() 685 iod->req.cqe = &iod->cqe; in nvmet_pci_epf_alloc_iod() 726 if (iod->data_segs != &iod->data_seg) in nvmet_pci_epf_free_iod() 731 mempool_free(iod, &iod->ctrl->iod_pool); in nvmet_pci_epf_free_iod() 951 iod->data_segs = &iod->data_seg; in nvmet_pci_epf_iod_parse_prp_simple() 1103 iod->data_segs = &iod->data_seg; in nvmet_pci_epf_iod_parse_sgls() 1139 iod->data_sgt.sgl = &iod->data_sgl; in nvmet_pci_epf_alloc_iod_data_buf() 1174 iod->status = le16_to_cpu(iod->cqe.status) >> 1; in nvmet_pci_epf_complete_iod() 1178 iod->sq->qid, nvmet_pci_epf_iod_name(iod), in nvmet_pci_epf_complete_iod() [all …]
|
| H A D | loop.c | 125 struct nvme_loop_iod *iod = in nvme_loop_execute_work() local 128 iod->req.execute(&iod->req); in nvme_loop_execute_work() 156 iod->sg_table.sgl = iod->first_sgl; in nvme_loop_queue_rq() 164 iod->req.sg = iod->sg_table.sgl; in nvme_loop_queue_rq() 165 iod->req.sg_cnt = blk_rq_map_sg(req, iod->sg_table.sgl); in nvme_loop_queue_rq() 169 queue_work(nvmet_wq, &iod->work); in nvme_loop_queue_rq() 179 memset(&iod->cmd, 0, sizeof(iod->cmd)); in nvme_loop_submit_async_event() 190 queue_work(nvmet_wq, &iod->work); in nvme_loop_submit_async_event() 196 iod->req.cmd = &iod->cmd; in nvme_loop_init_iod() 197 iod->req.cqe = &iod->cqe; in nvme_loop_init_iod() [all …]
|
| H A D | fc.c | 535 if (!iod) in nvmet_fc_alloc_ls_iodlist() 538 tgtport->iod = iod; in nvmet_fc_alloc_ls_iodlist() 565 for (iod--, i--; i >= 0; iod--, i--) { in nvmet_fc_alloc_ls_iodlist() 572 kfree(iod); in nvmet_fc_alloc_ls_iodlist() 580 struct nvmet_fc_ls_iod *iod = tgtport->iod; in nvmet_fc_free_ls_iodlist() local 585 iod->rspdma, sizeof(*iod->rspbuf), in nvmet_fc_free_ls_iodlist() 602 if (iod) in nvmet_fc_alloc_ls_iod() 605 return iod; in nvmet_fc_alloc_ls_iod() 1935 iod->lsrsp->nvme_fc_private = iod; in nvmet_fc_handle_ls_rqst() 1936 iod->lsrsp->rspbuf = iod->rspbuf; in nvmet_fc_handle_ls_rqst() [all …]
|
| /linux-6.15/tools/testing/selftests/ublk/ |
| H A D | file_backed.c | 7 unsigned ublk_op = ublksrv_get_op(iod); in ublk_to_uring_op() 18 unsigned ublk_op = ublksrv_get_op(iod); in loop_queue_flush_io() 31 unsigned ublk_op = ublksrv_get_op(iod); in loop_queue_tgt_rw_io() 42 (void *)iod->addr, in loop_queue_tgt_rw_io() 43 iod->nr_sectors << 9, in loop_queue_tgt_rw_io() 44 iod->start_sector << 9); in loop_queue_tgt_rw_io() 59 iod->nr_sectors << 9, in loop_queue_tgt_rw_io() 60 iod->start_sector << 9); in loop_queue_tgt_rw_io() 74 unsigned ublk_op = ublksrv_get_op(iod); in loop_queue_tgt_io() 79 ret = loop_queue_flush_io(q, iod, tag); in loop_queue_tgt_io() [all …]
|
| H A D | stripe.c | 34 const struct ublksrv_io_desc *iod) in calculate_nr_vec() argument 38 loff_t start = iod->start_sector; in calculate_nr_vec() 39 loff_t end = start + iod->nr_sectors; in calculate_nr_vec() 45 const struct ublksrv_io_desc *iod) in alloc_stripe_array() argument 78 off64_t start = iod->start_sector; in calculate_stripe_array() 79 off64_t end = start + iod->nr_sectors; in calculate_stripe_array() 117 unsigned ublk_op = ublksrv_get_op(iod); in stripe_to_uring_op() 137 calculate_stripe_array(conf, iod, s); in stripe_queue_tgt_rw_io() 200 ret = handle_flush(q, iod, tag); in stripe_queue_tgt_io() 215 iod->op_flags, iod->start_sector, iod->nr_sectors << 9, ret); in stripe_queue_tgt_io() [all …]
|
| H A D | null.c | 47 const struct ublksrv_io_desc *iod = ublk_get_iod(q, tag); in null_queue_zc_io() local 48 unsigned ublk_op = ublksrv_get_op(iod); in null_queue_zc_io() 62 sqe[1]->len = iod->nr_sectors << 9; /* injected result */ in null_queue_zc_io() 96 const struct ublksrv_io_desc *iod = ublk_get_iod(q, tag); in ublk_null_queue_io() local 101 ublk_complete_io(q, tag, iod->nr_sectors << 9); in ublk_null_queue_io()
|
| H A D | fault_inject.c | 38 const struct ublksrv_io_desc *iod = ublk_get_iod(q, tag); in ublk_fault_inject_queue_io() local 46 sqe->user_data = build_user_data(tag, ublksrv_get_op(iod), 0, 1); in ublk_fault_inject_queue_io() 56 const struct ublksrv_io_desc *iod = ublk_get_iod(q, tag); in ublk_fault_inject_tgt_io_done() local 62 ublk_complete_io(q, tag, iod->nr_sectors << 9); in ublk_fault_inject_tgt_io_done()
|
| /linux-6.15/drivers/nvme/host/ |
| H A D | apple.c | 363 dma_unmap_page(anv->dev, iod->first_dma, iod->dma_len, in apple_nvme_unmap_data() 370 dma_unmap_sg(anv->dev, iod->sg, iod->nents, rq_dma_dir(req)); in apple_nvme_unmap_data() 432 iod->npages = 0; in apple_nvme_setup_prps() 481 WARN(DO_ONCE(apple_nvme_print_sgl, iod->sg, iod->nents), in apple_nvme_setup_prps() 525 if (!iod->sg) in apple_nvme_map_data() 528 iod->nents = blk_rq_map_sg(req, iod->sg); in apple_nvme_map_data() 529 if (!iod->nents) in apple_nvme_map_data() 532 nr_mapped = dma_map_sg_attrs(anv->dev, iod->sg, iod->nents, in apple_nvme_map_data() 543 dma_unmap_sg(anv->dev, iod->sg, iod->nents, rq_dma_dir(req)); in apple_nvme_map_data() 745 iod->nents = 0; in apple_nvme_queue_rq() [all …]
|
| H A D | pci.c | 560 if (iod->dma_len) { in nvme_unmap_data() 561 dma_unmap_page(dev->dev, iod->first_dma, iod->dma_len, in nvme_unmap_data() 611 iod->first_dma = 0; in nvme_pci_setup_prps() 652 iod->list[iod->nr_allocations++].prp_list = prp_list; in nvme_pci_setup_prps() 679 WARN(DO_ONCE(nvme_print_sgl, iod->sgt.sgl, iod->sgt.nents), in nvme_pci_setup_prps() 810 iod->dma_len = 0; in nvme_map_data() 812 if (!iod->sgt.sgl) in nvme_map_data() 815 iod->sgt.orig_nents = blk_rq_map_sg(req, iod->sgt.sgl); in nvme_map_data() 925 iod->aborted = false; in nvme_prep_rq() 927 iod->sgt.nents = 0; in nvme_prep_rq() [all …]
|
| /linux-6.15/drivers/clk/ |
| H A D | clk-versaclock7.c | 172 struct vc7_iod_data *iod; member 379 map->src.iod = &vc7->clk_iod[0]; in vc7_get_bank_clk() 383 map->src.iod = &vc7->clk_iod[1]; in vc7_get_bank_clk() 400 map->src.iod = &vc7->clk_iod[1]; in vc7_get_bank_clk() 437 map->src.iod = &vc7->clk_iod[2]; in vc7_get_bank_clk() 463 map->src.iod = &vc7->clk_iod[2]; in vc7_get_bank_clk() 467 map->src.iod = &vc7->clk_iod[3]; in vc7_get_bank_clk() 490 map->src.iod = &vc7->clk_iod[2]; in vc7_get_bank_clk() 494 map->src.iod = &vc7->clk_iod[3]; in vc7_get_bank_clk() 966 err = vc7_read_iod(vc7, iod->num); in vc7_iod_recalc_rate() [all …]
|
| /linux-6.15/include/uapi/linux/ |
| H A D | ublk_cmd.h | 324 static inline __u8 ublksrv_get_op(const struct ublksrv_io_desc *iod) in ublksrv_get_op() argument 326 return iod->op_flags & 0xff; in ublksrv_get_op() 329 static inline __u32 ublksrv_get_flags(const struct ublksrv_io_desc *iod) in ublksrv_get_flags() argument 331 return iod->op_flags >> 8; in ublksrv_get_flags()
|
| /linux-6.15/arch/alpha/kernel/ |
| H A D | io.c | 661 u16 __iomem *iod = (u16 __iomem *) d; in scr_memcpyw() local 673 __raw_writew(tmp, iod++); in scr_memcpyw() 680 memcpy_toio(iod, s, count); in scr_memcpyw()
|
| /linux-6.15/drivers/block/ |
| H A D | ublk_drv.c | 431 iod->op_flags = ublk_op | ublk_req_build_flags(req); in ublk_setup_iod_zoned() 432 iod->nr_zones = desc->nr_zones; in ublk_setup_iod_zoned() 433 iod->start_sector = desc->sector; in ublk_setup_iod_zoned() 445 iod->op_flags = ublk_op | ublk_req_build_flags(req); in ublk_setup_iod_zoned() 446 iod->nr_sectors = blk_rq_sectors(req); in ublk_setup_iod_zoned() 447 iod->start_sector = blk_rq_pos(req); in ublk_setup_iod_zoned() 448 iod->addr = io->addr; in ublk_setup_iod_zoned() 1053 iod->op_flags = ublk_op | ublk_req_build_flags(req); in ublk_setup_iod() 1054 iod->nr_sectors = blk_rq_sectors(req); in ublk_setup_iod() 1055 iod->start_sector = blk_rq_pos(req); in ublk_setup_iod() [all …]
|
| /linux-6.15/tools/perf/util/ |
| H A D | synthetic-events.c | 771 struct io_dir iod; in __event__synthesize_thread() local 806 io_dir__init(&iod, open(filename, O_CLOEXEC | O_DIRECTORY | O_RDONLY)); in __event__synthesize_thread() 807 if (iod.dirfd < 0) in __event__synthesize_thread() 810 while ((dent = io_dir__readdir(&iod)) != NULL) { in __event__synthesize_thread() 852 close(iod.dirfd); in __event__synthesize_thread()
|
| H A D | machine.c | 1348 struct io_dir iod; in maps__set_modules_path_dir() local 1352 io_dir__init(&iod, open(path, O_CLOEXEC | O_DIRECTORY | O_RDONLY)); in maps__set_modules_path_dir() 1353 if (iod.dirfd < 0) { in maps__set_modules_path_dir() 1361 while ((dent = io_dir__readdir(&iod)) != NULL) { in maps__set_modules_path_dir() 1362 if (io_dir__is_dir(&iod, dent)) { in maps__set_modules_path_dir() 1405 close(iod.dirfd); in maps__set_modules_path_dir()
|
| /linux-6.15/Documentation/admin-guide/ |
| H A D | devices.txt | 2625 2 = /dev/vx/iod Volume i/o daemon access node
|