Home
last modified time | relevance | path

Searched refs:dchan (Results 1 – 25 of 27) sorted by relevance

12

/linux-6.15/drivers/dma/
H A Dloongson1-apb-dma.c86 #define to_ls1x_dma_chan(dchan) \ argument
100 struct dma_chan *dchan = &chan->vc.chan; in ls1x_dma_query() local
105 val |= dchan->chan_id; in ls1x_dma_query()
119 struct device *dev = chan2dev(dchan); in ls1x_dma_start()
124 val |= dchan->chan_id; in ls1x_dma_start()
146 struct device *dev = chan2dev(dchan); in ls1x_dma_free_chan_resources()
159 struct device *dev = chan2dev(dchan); in ls1x_dma_alloc_chan_resources()
222 struct device *dev = chan2dev(dchan); in ls1x_dma_prep_lli()
328 dev_dbg(chan2dev(dchan), in ls1x_dma_prep_dma_cyclic()
492 struct device *dev = chan2dev(dchan); in ls1x_dma_irq_handler()
[all …]
H A Dnbpfaxi.c603 struct nbpf_channel *chan = nbpf_to_chan(dchan); in nbpf_issue_pending()
628 struct nbpf_channel *chan = nbpf_to_chan(dchan); in nbpf_tx_status()
691 struct dma_chan *dchan = &chan->dma_chan; in nbpf_desc_page_alloc() local
699 struct device *dev = dchan->device->dev; in nbpf_desc_page_alloc()
863 static int nbpf_pause(struct dma_chan *dchan) in nbpf_pause() argument
882 dev_dbg(dchan->device->dev, "Terminating\n"); in nbpf_terminate_all()
890 static int nbpf_config(struct dma_chan *dchan, in nbpf_config() argument
1091 struct dma_chan *dchan; in nbpf_of_xlate() local
1098 if (!dchan) in nbpf_of_xlate()
1104 chan = nbpf_to_chan(dchan); in nbpf_of_xlate()
[all …]
H A Dmmp_pdma.c134 #define to_mmp_pdma_chan(dchan) \ argument
135 container_of(dchan, struct mmp_pdma_chan, chan)
443 mmp_pdma_prep_memcpy(struct dma_chan *dchan, in mmp_pdma_prep_memcpy() argument
451 if (!dchan) in mmp_pdma_prep_memcpy()
457 chan = to_mmp_pdma_chan(dchan); in mmp_pdma_prep_memcpy()
604 mmp_pdma_prep_dma_cyclic(struct dma_chan *dchan, in mmp_pdma_prep_dma_cyclic() argument
613 if (!dchan || !len || !period_len) in mmp_pdma_prep_dma_cyclic()
623 chan = to_mmp_pdma_chan(dchan); in mmp_pdma_prep_dma_cyclic()
698 if (!dchan) in mmp_pdma_config_write()
747 if (!dchan) in mmp_pdma_terminate_all()
[all …]
H A Dfsldma.c373 if (!dchan) in fsl_dma_external_start()
376 chan = to_fsl_chan(dchan); in fsl_dma_external_start()
692 struct fsldma_chan *chan = to_fsl_chan(dchan); in fsl_dma_alloc_chan_resources()
745 struct fsldma_chan *chan = to_fsl_chan(dchan); in fsl_dma_free_chan_resources()
760 fsl_dma_prep_memcpy(struct dma_chan *dchan, in fsl_dma_prep_memcpy() argument
768 if (!dchan) in fsl_dma_prep_memcpy()
774 chan = to_fsl_chan(dchan); in fsl_dma_prep_memcpy()
828 if (!dchan) in fsl_dma_device_terminate_all()
831 chan = to_fsl_chan(dchan); in fsl_dma_device_terminate_all()
854 if (!dchan) in fsl_dma_device_config()
[all …]
H A Dpxa_dma.c139 #define to_pxad_chan(dchan) \ argument
140 container_of(dchan, struct pxad_chan, vc.chan)
686 struct pxad_chan *chan = to_pxad_chan(dchan); in pxad_alloc_chan_resources()
709 struct pxad_chan *chan = to_pxad_chan(dchan); in pxad_free_chan_resources()
827 struct pxad_chan *chan = to_pxad_chan(dchan); in pxad_issue_pending()
916 pxad_prep_memcpy(struct dma_chan *dchan, in pxad_prep_memcpy() argument
920 struct pxad_chan *chan = to_pxad_chan(dchan); in pxad_prep_memcpy()
927 if (!dchan || !len) in pxad_prep_memcpy()
1013 pxad_prep_dma_cyclic(struct dma_chan *dchan, in pxad_prep_dma_cyclic() argument
1024 if (!dchan || !len || !period_len) in pxad_prep_dma_cyclic()
[all …]
H A Daltera-msgdma.c332 msgdma_prep_memcpy(struct dma_chan *dchan, dma_addr_t dma_dst, in msgdma_prep_memcpy() argument
335 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_prep_memcpy()
388 msgdma_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl, in msgdma_prep_slave_sg() argument
393 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_prep_slave_sg()
461 static int msgdma_dma_config(struct dma_chan *dchan, in msgdma_dma_config() argument
464 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_dma_config()
639 static void msgdma_free_chan_resources(struct dma_chan *dchan) in msgdma_free_chan_resources() argument
641 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_free_chan_resources()
656 static int msgdma_alloc_chan_resources(struct dma_chan *dchan) in msgdma_alloc_chan_resources() argument
658 struct msgdma_device *mdev = to_mdev(dchan); in msgdma_alloc_chan_resources()
H A Dxgene-dma.c191 #define to_dma_chan(dchan) \ argument
192 container_of(dchan, struct xgene_dma_chan, dma_chan)
777 struct xgene_dma_chan *chan = to_dma_chan(dchan); in xgene_dma_alloc_chan_resources()
814 struct xgene_dma_chan *chan = to_dma_chan(dchan); in xgene_dma_free_chan_resources()
839 struct dma_chan *dchan, dma_addr_t dst, dma_addr_t *src, in xgene_dma_prep_xor() argument
847 if (unlikely(!dchan || !len)) in xgene_dma_prep_xor()
850 chan = to_dma_chan(dchan); in xgene_dma_prep_xor()
896 if (unlikely(!dchan || !len)) in xgene_dma_prep_pq()
899 chan = to_dma_chan(dchan); in xgene_dma_prep_pq()
965 struct xgene_dma_chan *chan = to_dma_chan(dchan); in xgene_dma_issue_pending()
[all …]
H A Dbcm-sba-raid.c87 #define to_sba_device(dchan) \ argument
88 container_of(dchan, struct sba_device, dma_chan)
499 sba_cleanup_pending_requests(to_sba_device(dchan)); in sba_device_terminate_all()
507 struct sba_device *sba = to_sba_device(dchan); in sba_issue_pending()
544 struct sba_device *sba = to_sba_device(dchan); in sba_tx_status()
546 ret = dma_cookie_status(dchan, cookie, txstate); in sba_tx_status()
552 return dma_cookie_status(dchan, cookie, txstate); in sba_tx_status()
618 struct sba_device *sba = to_sba_device(dchan); in sba_prep_dma_interrupt()
733 struct sba_device *sba = to_sba_device(dchan); in sba_prep_dma_memcpy()
872 struct sba_device *sba = to_sba_device(dchan); in sba_prep_dma_xor()
[all …]
H A Dmmp_tdma.c137 #define to_mmp_tdma_chan(dchan) container_of(dchan, struct mmp_tdma_chan, chan) argument
/linux-6.15/drivers/rapidio/devices/
H A Dtsi721_dma.c308 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_start_dma()
315 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_start_dma()
413 struct dma_chan *dchan = desc->txd.chan; in tsi721_submit_sg() local
588 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_dma_tasklet()
652 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_dma_tasklet()
701 tsi_err(&bdma_chan->dchan.dev->device, in tsi721_tx_submit()
758 dma_cookie_init(dchan); in tsi721_alloc_chan_resources()
862 tsi_err(&dchan->dev->device, in tsi721_prep_rio_sg()
887 tsi_debug(DMA, &dchan->dev->device, in tsi721_prep_rio_sg()
982 bdma_chan->dchan.cookie = 1; in tsi721_register_dma()
[all …]
H A Dtsi721.h701 struct dma_chan dchan; member
/linux-6.15/drivers/dma/dw-edma/
H A Ddw-edma-core.c27 struct device *dchan2dev(struct dma_chan *dchan) in dchan2dev() argument
29 return &dchan->dev->device; in dchan2dev()
216 static void dw_edma_device_caps(struct dma_chan *dchan, in dw_edma_device_caps() argument
219 struct dw_edma_chan *chan = dchan2dw_edma_chan(dchan); in dw_edma_device_caps()
237 struct dw_edma_chan *chan = dchan2dw_edma_chan(dchan); in dw_edma_device_config()
245 static int dw_edma_device_pause(struct dma_chan *dchan) in dw_edma_device_pause() argument
337 ret = dma_cookie_status(dchan, cookie, txstate); in dw_edma_device_tx_status()
554 xfer.dchan = dchan; in dw_edma_device_prep_slave_sg()
572 xfer.dchan = dchan; in dw_edma_device_prep_dma_cyclic()
590 xfer.dchan = dchan; in dw_edma_device_prep_interleaved_dma()
[all …]
H A Ddw-edma-core.h143 struct dma_chan *dchan; member
161 struct dw_edma_chan *dchan2dw_edma_chan(struct dma_chan *dchan) in dchan2dw_edma_chan() argument
163 return vc2dw_edma_chan(to_virt_chan(dchan)); in dchan2dw_edma_chan()
/linux-6.15/drivers/dma/sf-pdma/
H A Dsf-pdma.c88 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_prep_dma_memcpy()
112 static int sf_pdma_slave_config(struct dma_chan *dchan, in sf_pdma_slave_config() argument
115 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_slave_config()
124 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_alloc_chan_resources()
127 dma_cookie_init(dchan); in sf_pdma_alloc_chan_resources()
142 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_free_chan_resources()
195 sf_pdma_tx_status(struct dma_chan *dchan, in sf_pdma_tx_status() argument
199 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_tx_status()
202 status = dma_cookie_status(dchan, cookie, txstate); in sf_pdma_tx_status()
212 struct sf_pdma_chan *chan = to_sf_pdma_chan(dchan); in sf_pdma_terminate_all()
[all …]
/linux-6.15/drivers/dma/dw-axi-dmac/
H A Ddw-axi-dmac-platform.c368 status = dma_cookie_status(dchan, cookie, txstate); in dma_chan_tx_status()
549 dev_err(dchan2dev(dchan), "%s is non-idle!\n", in dma_chan_free_chan_resources()
559 dev_vdbg(dchan2dev(dchan), in dma_chan_free_chan_resources()
1195 dev_warn(dchan2dev(dchan), in dma_chan_terminate_all()
1217 static int dma_chan_pause(struct dma_chan *dchan) in dma_chan_pause() argument
1302 static int dma_chan_resume(struct dma_chan *dchan) in dma_chan_resume() argument
1365 struct dma_chan *dchan; in dw_axi_dma_of_xlate() local
1367 dchan = dma_get_any_slave_channel(&dw->dma); in dw_axi_dma_of_xlate()
1368 if (!dchan) in dw_axi_dma_of_xlate()
1371 chan = dchan_to_axi_dma_chan(dchan); in dw_axi_dma_of_xlate()
[all …]
H A Ddw-axi-dmac.h121 static inline struct device *dchan2dev(struct dma_chan *dchan) in dchan2dev() argument
123 return &dchan->dev->device; in dchan2dev()
141 static inline struct axi_dma_chan *dchan_to_axi_dma_chan(struct dma_chan *dchan) in dchan_to_axi_dma_chan() argument
143 return vc_to_axi_dma_chan(to_virt_chan(dchan)); in dchan_to_axi_dma_chan()
/linux-6.15/drivers/dma/xilinx/
H A Dxilinx_dpdma.c1276 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_prep_dma_cyclic()
1293 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_prep_interleaved_dma()
1324 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_alloc_chan_resources()
1350 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_free_chan_resources()
1360 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_issue_pending()
1371 static int xilinx_dpdma_config(struct dma_chan *dchan, in xilinx_dpdma_config() argument
1374 struct xilinx_dpdma_chan *chan = to_xilinx_chan(dchan); in xilinx_dpdma_config()
1402 static int xilinx_dpdma_pause(struct dma_chan *dchan) in xilinx_dpdma_pause() argument
1404 xilinx_dpdma_chan_pause(to_xilinx_chan(dchan)); in xilinx_dpdma_pause()
1409 static int xilinx_dpdma_resume(struct dma_chan *dchan) in xilinx_dpdma_resume() argument
[all …]
H A Dzynqmp_dma.c477 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_alloc_chan_resources()
574 static int zynqmp_dma_device_config(struct dma_chan *dchan, in zynqmp_dma_device_config() argument
577 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_device_config()
660 static void zynqmp_dma_issue_pending(struct dma_chan *dchan) in zynqmp_dma_issue_pending() argument
662 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_issue_pending()
691 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_free_chan_resources()
804 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_device_terminate_all()
816 static void zynqmp_dma_synchronize(struct dma_chan *dchan) in zynqmp_dma_synchronize() argument
818 struct zynqmp_dma_chan *chan = to_chan(dchan); in zynqmp_dma_synchronize()
834 struct dma_chan *dchan, dma_addr_t dma_dst, in zynqmp_dma_prep_memcpy() argument
[all …]
H A Dxilinx_dma.c933 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_free_chan_resources()
1116 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_alloc_chan_resources()
1217 dma_cookie_init(dchan); in xilinx_dma_alloc_chan_resources()
1274 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_tx_status()
1280 ret = dma_cookie_status(dchan, cookie, txstate); in xilinx_dma_tx_status()
1688 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_issue_pending()
2047 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_vdma_dma_prep_interleaved()
2133 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_cdma_prep_memcpy()
2191 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_prep_slave_sg()
2293 struct xilinx_dma_chan *chan = to_xilinx_chan(dchan); in xilinx_dma_prep_dma_cyclic()
[all …]
/linux-6.15/include/linux/
H A Dfsldma.h8 int fsl_dma_external_start(struct dma_chan *dchan, int enable);
H A Drio_drv.h396 extern void rio_release_dma(struct dma_chan *dchan);
398 struct rio_dev *rdev, struct dma_chan *dchan,
402 struct dma_chan *dchan, u16 destid,
/linux-6.15/include/linux/dma/
H A Dxilinx_dma.h42 int xilinx_vdma_channel_set_config(struct dma_chan *dchan,
/linux-6.15/arch/mips/alchemy/common/
H A Dirq.c581 void au1300_set_dbdma_gpio(int dchan, unsigned int gpio) in au1300_set_dbdma_gpio() argument
585 if ((dchan >= 0) && (dchan <= 1)) { in au1300_set_dbdma_gpio()
587 r &= ~(0xff << (8 * dchan)); in au1300_set_dbdma_gpio()
588 r |= (gpio & 0x7f) << (8 * dchan); in au1300_set_dbdma_gpio()
/linux-6.15/drivers/rapidio/
H A Drio.c1794 void rio_release_dma(struct dma_chan *dchan) in rio_release_dma() argument
1796 dma_release_channel(dchan); in rio_release_dma()
1816 struct dma_async_tx_descriptor *rio_dma_prep_xfer(struct dma_chan *dchan, in rio_dma_prep_xfer() argument
1822 if (!dchan->device->device_prep_slave_sg) { in rio_dma_prep_xfer()
1832 return dmaengine_prep_rio_sg(dchan, data->sg, data->sg_len, in rio_dma_prep_xfer()
1854 struct dma_chan *dchan, struct rio_dma_data *data, in rio_dma_prep_slave_sg() argument
1857 return rio_dma_prep_xfer(dchan, rdev->destid, data, direction, flags); in rio_dma_prep_slave_sg()
/linux-6.15/arch/mips/include/asm/mach-au1x00/
H A Dau1000.h882 extern void au1300_set_dbdma_gpio(int dchan, unsigned int gpio);

12