Home
last modified time | relevance | path

Searched refs:vdesc (Results 1 – 25 of 27) sorted by relevance

12

/linux-6.6.21/drivers/dma/
Ddma-jz4780.c119 struct virt_dma_desc vdesc; member
171 struct virt_dma_desc *vdesc) in to_jz4780_dma_desc() argument
173 return container_of(vdesc, struct jz4780_dma_desc, vdesc); in to_jz4780_dma_desc()
262 static void jz4780_dma_desc_free(struct virt_dma_desc *vdesc) in jz4780_dma_desc_free() argument
264 struct jz4780_dma_desc *desc = to_jz4780_dma_desc(vdesc); in jz4780_dma_desc_free()
265 struct jz4780_dma_chan *jzchan = to_jz4780_dma_chan(vdesc->tx.chan); in jz4780_dma_desc_free()
382 jz4780_dma_desc_free(&jzchan->desc->vdesc); in jz4780_dma_prep_slave_sg()
403 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); in jz4780_dma_prep_slave_sg()
429 jz4780_dma_desc_free(&jzchan->desc->vdesc); in jz4780_dma_prep_dma_cyclic()
455 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); in jz4780_dma_prep_dma_cyclic()
[all …]
Ddma-axi-dmac.c113 struct virt_dma_desc vdesc; member
165 static struct axi_dmac_desc *to_axi_dmac_desc(struct virt_dma_desc *vdesc) in to_axi_dmac_desc() argument
167 return container_of(vdesc, struct axi_dmac_desc, vdesc); in to_axi_dmac_desc()
210 struct virt_dma_desc *vdesc; in axi_dmac_start_transfer() local
223 vdesc = vchan_next_desc(&chan->vchan); in axi_dmac_start_transfer()
224 if (!vdesc) in axi_dmac_start_transfer()
226 list_move_tail(&vdesc->node, &chan->active_descs); in axi_dmac_start_transfer()
227 desc = to_axi_dmac_desc(vdesc); in axi_dmac_start_transfer()
266 if (chan->hw_cyclic && desc->cyclic && !desc->vdesc.tx.callback && in axi_dmac_start_transfer()
282 struct axi_dmac_desc, vdesc.node); in axi_dmac_active_desc()
[all …]
Didma64.c115 struct virt_dma_desc *vdesc; in idma64_start_transfer() local
118 vdesc = vchan_next_desc(&idma64c->vchan); in idma64_start_transfer()
119 if (!vdesc) { in idma64_start_transfer()
124 list_del(&vdesc->node); in idma64_start_transfer()
125 idma64c->desc = to_idma64_desc(vdesc); in idma64_start_transfer()
154 vchan_cookie_complete(&desc->vdesc); in idma64_chan_irq()
226 static void idma64_vdesc_free(struct virt_dma_desc *vdesc) in idma64_vdesc_free() argument
228 struct idma64_chan *idma64c = to_idma64_chan(vdesc->tx.chan); in idma64_vdesc_free()
230 idma64_desc_free(idma64c, to_idma64_desc(vdesc)); in idma64_vdesc_free()
330 return vchan_tx_prep(&idma64c->vchan, &desc->vdesc, flags); in idma64_prep_slave_sg()
[all …]
Dfsl-edma-common.c56 list_del(&fsl_chan->edesc->vdesc.node); in fsl_edma_tx_chan_handler()
57 vchan_cookie_complete(&fsl_chan->edesc->vdesc); in fsl_edma_tx_chan_handler()
62 vchan_cyclic_callback(&fsl_chan->edesc->vdesc); in fsl_edma_tx_chan_handler()
227 void fsl_edma_free_desc(struct virt_dma_desc *vdesc) in fsl_edma_free_desc() argument
232 fsl_desc = to_fsl_edma_desc(vdesc); in fsl_edma_free_desc()
350 struct virt_dma_desc *vdesc, bool in_progress) in fsl_edma_desc_residue() argument
402 struct virt_dma_desc *vdesc; in fsl_edma_tx_status() local
414 vdesc = vchan_find_desc(&fsl_chan->vchan, cookie); in fsl_edma_tx_status()
415 if (fsl_chan->edesc && cookie == fsl_chan->edesc->vdesc.tx.cookie) in fsl_edma_tx_status()
417 fsl_edma_desc_residue(fsl_chan, vdesc, true); in fsl_edma_tx_status()
[all …]
Dst_fdma.c29 return container_of(vd, struct st_fdma_desc, vdesc); in to_st_fdma_desc()
76 struct virt_dma_desc *vdesc; in st_fdma_xfer_desc() local
79 vdesc = vchan_next_desc(&fchan->vchan); in st_fdma_xfer_desc()
80 if (!vdesc) in st_fdma_xfer_desc()
83 fchan->fdesc = to_st_fdma_desc(vdesc); in st_fdma_xfer_desc()
144 list_del(&fchan->fdesc->vdesc.node); in st_fdma_irq_handler()
145 vchan_cookie_complete(&fchan->fdesc->vdesc); in st_fdma_irq_handler()
149 vchan_cyclic_callback(&fchan->fdesc->vdesc); in st_fdma_irq_handler()
225 static void st_fdma_free_desc(struct virt_dma_desc *vdesc) in st_fdma_free_desc() argument
230 fdesc = to_st_fdma_desc(vdesc); in st_fdma_free_desc()
[all …]
Didma64.h114 struct virt_dma_desc vdesc; member
122 static inline struct idma64_desc *to_idma64_desc(struct virt_dma_desc *vdesc) in to_idma64_desc() argument
124 return container_of(vdesc, struct idma64_desc, vdesc); in to_idma64_desc()
Dstm32-dma.c191 struct virt_dma_desc vdesc; member
247 static struct stm32_dma_desc *to_stm32_dma_desc(struct virt_dma_desc *vdesc) in to_stm32_dma_desc() argument
249 return container_of(vdesc, struct stm32_dma_desc, vdesc); in to_stm32_dma_desc()
509 dma_cookie_complete(&chan->desc->vdesc.tx); in stm32_dma_terminate_all()
510 vchan_terminate_vdesc(&chan->desc->vdesc); in stm32_dma_terminate_all()
560 struct virt_dma_desc *vdesc; in stm32_dma_start_transfer() local
571 vdesc = vchan_next_desc(&chan->vchan); in stm32_dma_start_transfer()
572 if (!vdesc) in stm32_dma_start_transfer()
575 list_del(&vdesc->node); in stm32_dma_start_transfer()
577 chan->desc = to_stm32_dma_desc(vdesc); in stm32_dma_start_transfer()
[all …]
Dfsl-qdma.c178 struct virt_dma_desc vdesc; member
205 struct virt_dma_desc vdesc; member
302 return container_of(vd, struct fsl_qdma_comp, vdesc); in to_fsl_qdma_comp()
713 fsl_comp->vdesc.tx_result.result = in fsl_qdma_queue_transfer_complete()
717 fsl_comp->vdesc.tx_result.result = in fsl_qdma_queue_transfer_complete()
723 fsl_comp->vdesc.tx_result.result = in fsl_qdma_queue_transfer_complete()
732 vchan_cookie_complete(&fsl_comp->vdesc); in fsl_qdma_queue_transfer_complete()
987 return vchan_tx_prep(&fsl_chan->vchan, &fsl_comp->vdesc, flags); in fsl_qdma_prep_memcpy()
993 struct virt_dma_desc *vdesc; in fsl_qdma_enqueue_desc() local
1001 vdesc = vchan_next_desc(&fsl_chan->vchan); in fsl_qdma_enqueue_desc()
[all …]
Dstm32-mdma.c223 struct virt_dma_desc vdesc; member
274 static struct stm32_mdma_desc *to_stm32_mdma_desc(struct virt_dma_desc *vdesc) in to_stm32_mdma_desc() argument
276 return container_of(vdesc, struct stm32_mdma_desc, vdesc); in to_stm32_mdma_desc()
346 static void stm32_mdma_desc_free(struct virt_dma_desc *vdesc) in stm32_mdma_desc_free() argument
348 struct stm32_mdma_desc *desc = to_stm32_mdma_desc(vdesc); in stm32_mdma_desc_free()
349 struct stm32_mdma_chan *chan = to_stm32_mdma_chan(vdesc->tx.chan); in stm32_mdma_desc_free()
831 return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); in stm32_mdma_prep_slave_sg()
927 return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); in stm32_mdma_prep_dma_cyclic()
1113 return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); in stm32_mdma_prep_dma_memcpy()
1145 struct virt_dma_desc *vdesc; in stm32_mdma_start_transfer() local
[all …]
Dfsl-edma-common.h163 struct virt_dma_desc vdesc; member
312 return container_of(vd, struct fsl_edma_desc, vdesc); in to_fsl_edma_desc()
325 void fsl_edma_free_desc(struct virt_dma_desc *vdesc);
Dst_fdma.h96 struct virt_dma_desc vdesc; member
Dtegra186-gpc-dma.c509 struct virt_dma_desc *vdesc; in tegra_dma_start() local
512 vdesc = vchan_next_desc(&tdc->vc); in tegra_dma_start()
513 if (!vdesc) in tegra_dma_start()
516 dma_desc = vd_to_tegra_dma_desc(vdesc); in tegra_dma_start()
517 list_del(&vdesc->node); in tegra_dma_start()
/linux-6.6.21/fs/nilfs2/
Dioctl.c541 struct nilfs_vdesc *vdesc, in nilfs_ioctl_move_inode_block() argument
547 if (vdesc->vd_flags == 0) in nilfs_ioctl_move_inode_block()
549 inode, vdesc->vd_offset, vdesc->vd_blocknr, in nilfs_ioctl_move_inode_block()
550 vdesc->vd_vblocknr, &bh); in nilfs_ioctl_move_inode_block()
553 inode, vdesc->vd_blocknr, vdesc->vd_vblocknr, &bh); in nilfs_ioctl_move_inode_block()
559 __func__, vdesc->vd_flags ? "node" : "data", in nilfs_ioctl_move_inode_block()
560 (unsigned long long)vdesc->vd_ino, in nilfs_ioctl_move_inode_block()
561 (unsigned long long)vdesc->vd_cno, in nilfs_ioctl_move_inode_block()
562 (unsigned long long)vdesc->vd_offset, in nilfs_ioctl_move_inode_block()
563 (unsigned long long)vdesc->vd_blocknr, in nilfs_ioctl_move_inode_block()
[all …]
/linux-6.6.21/drivers/dma/hsu/
Dhsu.c118 struct virt_dma_desc *vdesc; in hsu_dma_start_transfer() local
121 vdesc = vchan_next_desc(&hsuc->vchan); in hsu_dma_start_transfer()
122 if (!vdesc) { in hsu_dma_start_transfer()
127 list_del(&vdesc->node); in hsu_dma_start_transfer()
128 hsuc->desc = to_hsu_dma_desc(vdesc); in hsu_dma_start_transfer()
232 vchan_cookie_complete(&desc->vdesc); in hsu_dma_do_irq()
261 static void hsu_dma_desc_free(struct virt_dma_desc *vdesc) in hsu_dma_desc_free() argument
263 struct hsu_dma_desc *desc = to_hsu_dma_desc(vdesc); in hsu_dma_desc_free()
295 return vchan_tx_prep(&hsuc->vchan, &desc->vdesc, flags); in hsu_dma_prep_slave_sg()
330 struct virt_dma_desc *vdesc; in hsu_dma_tx_status() local
[all …]
Dhsu.h72 struct virt_dma_desc vdesc; member
81 static inline struct hsu_dma_desc *to_hsu_dma_desc(struct virt_dma_desc *vdesc) in to_hsu_dma_desc() argument
83 return container_of(vdesc, struct hsu_dma_desc, vdesc); in to_hsu_dma_desc()
/linux-6.6.21/drivers/dma/sf-pdma/
Dsf-pdma.c49 return container_of(vd, struct sf_pdma_desc, vdesc); in to_sf_pdma_desc()
100 desc->async_tx = vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); in sf_pdma_prep_dma_memcpy()
241 struct virt_dma_desc *vdesc; in sf_pdma_get_first_pending_desc() local
246 vdesc = list_first_entry(&vchan->desc_issued, struct virt_dma_desc, node); in sf_pdma_get_first_pending_desc()
248 return container_of(vdesc, struct sf_pdma_desc, vdesc); in sf_pdma_get_first_pending_desc()
287 static void sf_pdma_free_desc(struct virt_dma_desc *vdesc) in sf_pdma_free_desc() argument
291 desc = to_sf_pdma_desc(vdesc); in sf_pdma_free_desc()
309 list_del(&chan->desc->vdesc.node); in sf_pdma_donebh_tasklet()
310 vchan_cookie_complete(&chan->desc->vdesc); in sf_pdma_donebh_tasklet()
Dsf-pdma.h79 struct virt_dma_desc vdesc; member
/linux-6.6.21/drivers/dma/xilinx/
Dxilinx_dpdma.c199 struct virt_dma_desc vdesc; member
206 container_of(_desc, struct xilinx_dpdma_tx_desc, vdesc)
654 static void xilinx_dpdma_chan_free_tx_desc(struct virt_dma_desc *vdesc) in xilinx_dpdma_chan_free_tx_desc() argument
659 if (!vdesc) in xilinx_dpdma_chan_free_tx_desc()
662 desc = to_dpdma_tx_desc(vdesc); in xilinx_dpdma_chan_free_tx_desc()
706 xilinx_dpdma_chan_free_tx_desc(&tx_desc->vdesc); in xilinx_dpdma_chan_prep_interleaved_dma()
830 struct virt_dma_desc *vdesc; in xilinx_dpdma_chan_queue_transfer() local
846 vdesc = vchan_next_desc(&chan->vchan); in xilinx_dpdma_chan_queue_transfer()
847 if (!vdesc) in xilinx_dpdma_chan_queue_transfer()
850 desc = to_dpdma_tx_desc(vdesc); in xilinx_dpdma_chan_queue_transfer()
[all …]
Dxdma.c88 struct virt_dma_desc vdesc; member
182 static inline struct xdma_desc *to_xdma_desc(struct virt_dma_desc *vdesc) in to_xdma_desc() argument
184 return container_of(vdesc, struct xdma_desc, vdesc); in to_xdma_desc()
213 static void xdma_free_desc(struct virt_dma_desc *vdesc) in xdma_free_desc() argument
218 sw_desc = to_xdma_desc(vdesc); in xdma_free_desc()
274 xdma_free_desc(&sw_desc->vdesc); in xdma_alloc_desc()
512 tx_desc = vchan_tx_prep(&xdma_chan->vchan, &sw_desc->vdesc, flags); in xdma_prep_device_sg()
519 xdma_free_desc(&sw_desc->vdesc); in xdma_prep_device_sg()
/linux-6.6.21/drivers/sh/intc/
Dvirq.c122 struct irq_desc *vdesc = irq_to_desc(entry->irq); in intc_virq_handler() local
124 if (vdesc) { in intc_virq_handler()
125 handle = (unsigned long)irq_desc_get_handler_data(vdesc); in intc_virq_handler()
128 generic_handle_irq_desc(vdesc); in intc_virq_handler()
/linux-6.6.21/drivers/dma/fsl-dpaa2-qdma/
Ddpaa2-qdma.c26 return container_of(vd, struct dpaa2_qdma_comp, vdesc); in to_fsl_qdma_comp()
269 return vchan_tx_prep(&dpaa2_chan->vchan, &dpaa2_comp->vdesc, flags); in dpaa2_qdma_prep_memcpy()
276 struct virt_dma_desc *vdesc; in dpaa2_qdma_issue_pending() local
284 vdesc = vchan_next_desc(&dpaa2_chan->vchan); in dpaa2_qdma_issue_pending()
285 if (!vdesc) in dpaa2_qdma_issue_pending()
287 dpaa2_comp = to_fsl_qdma_comp(vdesc); in dpaa2_qdma_issue_pending()
291 list_del(&vdesc->node); in dpaa2_qdma_issue_pending()
443 dpaa2_comp->vdesc); in dpaa2_qdma_fqdan_cb()
621 static void dpaa2_qdma_free_desc(struct virt_dma_desc *vdesc) in dpaa2_qdma_free_desc() argument
627 dpaa2_comp = to_fsl_qdma_comp(vdesc); in dpaa2_qdma_free_desc()
Ddpaa2-qdma.h72 struct virt_dma_desc vdesc; member
95 struct virt_dma_desc vdesc; member
/linux-6.6.21/drivers/dma/ti/
Dedma.c170 struct virt_dma_desc vdesc; member
709 return container_of(tx, struct edma_desc, vdesc.tx); in to_edma_desc()
712 static void edma_desc_free(struct virt_dma_desc *vdesc) in edma_desc_free() argument
714 kfree(container_of(vdesc, struct edma_desc, vdesc)); in edma_desc_free()
721 struct virt_dma_desc *vdesc; in edma_execute() local
728 vdesc = vchan_next_desc(&echan->vchan); in edma_execute()
729 if (!vdesc) in edma_execute()
731 list_del(&vdesc->node); in edma_execute()
732 echan->edesc = to_edma_desc(&vdesc->tx); in edma_execute()
830 vchan_terminate_vdesc(&echan->edesc->vdesc); in edma_terminate_all()
[all …]
/linux-6.6.21/drivers/dma/lgm/
Dlgm-dma.c270 struct virt_dma_desc vdesc; member
300 static inline struct dw2_desc_sw *to_lgm_dma_desc(struct virt_dma_desc *vdesc) in to_lgm_dma_desc() argument
302 return container_of(vdesc, struct dw2_desc_sw, vdesc); in to_lgm_dma_desc()
688 tx = &ds->vdesc.tx; in ldma_chan_desc_cfg()
965 static void dma_free_desc_resource(struct virt_dma_desc *vdesc) in dma_free_desc_resource() argument
967 struct dw2_desc_sw *ds = to_lgm_dma_desc(vdesc); in dma_free_desc_resource()
1023 struct virt_dma_desc *vdesc; in ldma_issue_pending() local
1026 vdesc = vchan_next_desc(&c->vchan); in ldma_issue_pending()
1027 if (!vdesc) { in ldma_issue_pending()
1032 list_del(&vdesc->node); in ldma_issue_pending()
[all …]
/linux-6.6.21/drivers/dma/dw-axi-dmac/
Ddw-axi-dmac-platform.c302 static void vchan_desc_put(struct virt_dma_desc *vdesc) in vchan_desc_put() argument
304 axi_desc_put(vd_to_axi_desc(vdesc)); in vchan_desc_put()
312 struct virt_dma_desc *vdesc; in dma_chan_tx_status() local
327 vdesc = vchan_find_desc(&chan->vc, cookie); in dma_chan_tx_status()
328 if (vdesc) { in dma_chan_tx_status()
329 length = vd_to_axi_desc(vdesc)->length; in dma_chan_tx_status()
330 completed_blocks = vd_to_axi_desc(vdesc)->completed_blocks; in dma_chan_tx_status()
331 len = vd_to_axi_desc(vdesc)->hw_desc[0].len; in dma_chan_tx_status()

12