Home
last modified time | relevance | path

Searched refs:dma_dev (Results 1 – 25 of 168) sorted by relevance

1234567

/linux-6.1.9/drivers/dma/ptdma/
Dptdma-dmaengine.c323 struct dma_device *dma_dev = &pt->dma_dev; in pt_dmaengine_register() local
355 dma_dev->dev = pt->dev; in pt_dmaengine_register()
356 dma_dev->src_addr_widths = DMA_SLAVE_BUSWIDTH_64_BYTES; in pt_dmaengine_register()
357 dma_dev->dst_addr_widths = DMA_SLAVE_BUSWIDTH_64_BYTES; in pt_dmaengine_register()
358 dma_dev->directions = DMA_MEM_TO_MEM; in pt_dmaengine_register()
359 dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; in pt_dmaengine_register()
360 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in pt_dmaengine_register()
361 dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); in pt_dmaengine_register()
367 dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask); in pt_dmaengine_register()
369 INIT_LIST_HEAD(&dma_dev->channels); in pt_dmaengine_register()
[all …]
/linux-6.1.9/drivers/dma/
Dmcf-edma.c222 INIT_LIST_HEAD(&mcf_edma->dma_dev.channels); in mcf_edma_probe()
231 vchan_init(&mcf_chan->vchan, &mcf_edma->dma_dev); in mcf_edma_probe()
242 dma_cap_set(DMA_PRIVATE, mcf_edma->dma_dev.cap_mask); in mcf_edma_probe()
243 dma_cap_set(DMA_SLAVE, mcf_edma->dma_dev.cap_mask); in mcf_edma_probe()
244 dma_cap_set(DMA_CYCLIC, mcf_edma->dma_dev.cap_mask); in mcf_edma_probe()
246 mcf_edma->dma_dev.dev = &pdev->dev; in mcf_edma_probe()
247 mcf_edma->dma_dev.device_alloc_chan_resources = in mcf_edma_probe()
249 mcf_edma->dma_dev.device_free_chan_resources = in mcf_edma_probe()
251 mcf_edma->dma_dev.device_config = fsl_edma_slave_config; in mcf_edma_probe()
252 mcf_edma->dma_dev.device_prep_dma_cyclic = in mcf_edma_probe()
[all …]
Dfsl-edma.c117 list_for_each_entry_safe(chan, _chan, &fsl_edma->dma_dev.channels, device_node) { in fsl_edma_xlate()
352 INIT_LIST_HEAD(&fsl_edma->dma_dev.channels); in fsl_edma_probe()
362 vchan_init(&fsl_chan->vchan, &fsl_edma->dma_dev); in fsl_edma_probe()
373 dma_cap_set(DMA_PRIVATE, fsl_edma->dma_dev.cap_mask); in fsl_edma_probe()
374 dma_cap_set(DMA_SLAVE, fsl_edma->dma_dev.cap_mask); in fsl_edma_probe()
375 dma_cap_set(DMA_CYCLIC, fsl_edma->dma_dev.cap_mask); in fsl_edma_probe()
376 dma_cap_set(DMA_MEMCPY, fsl_edma->dma_dev.cap_mask); in fsl_edma_probe()
378 fsl_edma->dma_dev.dev = &pdev->dev; in fsl_edma_probe()
379 fsl_edma->dma_dev.device_alloc_chan_resources in fsl_edma_probe()
381 fsl_edma->dma_dev.device_free_chan_resources in fsl_edma_probe()
[all …]
Ddma-axi-dmac.c150 struct dma_device dma_dev; member
157 dma_dev); in chan_to_axi_dmac()
324 dev_dbg(dmac->dma_dev.dev, in axi_dmac_dequeue_partial_xfers()
328 dev_warn(dmac->dma_dev.dev, in axi_dmac_dequeue_partial_xfers()
882 dev_err(dmac->dma_dev.dev, in axi_dmac_detect_caps()
890 dev_err(dmac->dma_dev.dev, in axi_dmac_detect_caps()
911 struct dma_device *dma_dev; in axi_dmac_probe() local
955 dma_dev = &dmac->dma_dev; in axi_dmac_probe()
956 dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); in axi_dmac_probe()
957 dma_cap_set(DMA_CYCLIC, dma_dev->cap_mask); in axi_dmac_probe()
[all …]
Daltera-msgdma.c810 struct dma_device *dma_dev; in msgdma_probe() local
858 dma_dev = &mdev->dmadev; in msgdma_probe()
861 dma_cap_zero(dma_dev->cap_mask); in msgdma_probe()
862 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in msgdma_probe()
863 dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); in msgdma_probe()
865 dma_dev->src_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES); in msgdma_probe()
866 dma_dev->dst_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES); in msgdma_probe()
867 dma_dev->directions = BIT(DMA_MEM_TO_DEV) | BIT(DMA_DEV_TO_MEM) | in msgdma_probe()
869 dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; in msgdma_probe()
872 INIT_LIST_HEAD(&dma_dev->channels); in msgdma_probe()
[all …]
Dimg-mdc-dma.c134 struct dma_device dma_dev; member
183 return mdma->dma_dev.dev; in mdma2dev()
817 list_for_each_entry(chan, &mdma->dma_dev.channels, device_node) { in mdc_of_xlate()
915 dma_cap_zero(mdma->dma_dev.cap_mask); in mdc_dma_probe()
916 dma_cap_set(DMA_SLAVE, mdma->dma_dev.cap_mask); in mdc_dma_probe()
917 dma_cap_set(DMA_PRIVATE, mdma->dma_dev.cap_mask); in mdc_dma_probe()
918 dma_cap_set(DMA_CYCLIC, mdma->dma_dev.cap_mask); in mdc_dma_probe()
919 dma_cap_set(DMA_MEMCPY, mdma->dma_dev.cap_mask); in mdc_dma_probe()
949 mdma->dma_dev.dev = &pdev->dev; in mdc_dma_probe()
950 mdma->dma_dev.device_prep_slave_sg = mdc_prep_slave_sg; in mdc_dma_probe()
[all …]
Dnbpfaxi.c230 struct dma_device dma_dev; member
326 dev_dbg(nbpf->dma_dev.dev, "%s(0x%p + 0x%x) = 0x%x\n", in nbpf_read()
335 dev_dbg(nbpf->dma_dev.dev, "%s(0x%p + 0x%x) = 0x%x\n", in nbpf_write()
401 dev_dbg(chan->nbpf->dma_dev.dev, "%s(): next 0x%x, cur 0x%x\n", __func__, in nbpf_start()
857 dev_dbg(chan->nbpf->dma_dev.dev, "%s(): force-free desc %p cookie %d\n", in nbpf_chan_idle()
1098 dchan = dma_get_any_slave_channel(&nbpf->dma_dev); in nbpf_of_xlate()
1229 dev_warn(nbpf->dma_dev.dev, "DMA error IRQ %u\n", irq); in nbpf_err_irq()
1247 struct dma_device *dma_dev = &nbpf->dma_dev; in nbpf_chan_probe() local
1255 chan->dma_chan.device = dma_dev; in nbpf_chan_probe()
1259 dev_dbg(dma_dev->dev, "%s(): channel %d: -> %p\n", __func__, n, chan->base); in nbpf_chan_probe()
[all …]
Dfsl_raid.c628 struct dma_device *dma_dev; in fsl_re_chan_probe() local
636 dma_dev = &re_priv->dma_dev; in fsl_re_chan_probe()
682 chan->chan.device = dma_dev; in fsl_re_chan_probe()
752 struct dma_device *dma_dev; in fsl_re_probe() local
781 dma_dev = &re_priv->dma_dev; in fsl_re_probe()
782 dma_dev->dev = dev; in fsl_re_probe()
783 INIT_LIST_HEAD(&dma_dev->channels); in fsl_re_probe()
786 dma_dev->device_alloc_chan_resources = fsl_re_alloc_chan_resources; in fsl_re_probe()
787 dma_dev->device_tx_status = fsl_re_tx_status; in fsl_re_probe()
788 dma_dev->device_issue_pending = fsl_re_issue_pending; in fsl_re_probe()
[all …]
Diop-adma.c1271 struct dma_device *dma_dev; in iop_adma_probe() local
1285 dma_dev = &adev->common; in iop_adma_probe()
1306 dma_dev->cap_mask = plat_data->cap_mask; in iop_adma_probe()
1311 INIT_LIST_HEAD(&dma_dev->channels); in iop_adma_probe()
1314 dma_dev->device_alloc_chan_resources = iop_adma_alloc_chan_resources; in iop_adma_probe()
1315 dma_dev->device_free_chan_resources = iop_adma_free_chan_resources; in iop_adma_probe()
1316 dma_dev->device_tx_status = iop_adma_status; in iop_adma_probe()
1317 dma_dev->device_issue_pending = iop_adma_issue_pending; in iop_adma_probe()
1318 dma_dev->dev = &pdev->dev; in iop_adma_probe()
1321 if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) in iop_adma_probe()
[all …]
Dtegra210-adma.c150 struct dma_device dma_dev; member
720 chan = dma_get_any_slave_channel(&tdma->dma_dev); in tegra_dma_of_xlate()
871 INIT_LIST_HEAD(&tdma->dma_dev.channels); in tegra_adma_probe()
884 vchan_init(&tdc->vc, &tdma->dma_dev); in tegra_adma_probe()
899 dma_cap_set(DMA_SLAVE, tdma->dma_dev.cap_mask); in tegra_adma_probe()
900 dma_cap_set(DMA_PRIVATE, tdma->dma_dev.cap_mask); in tegra_adma_probe()
901 dma_cap_set(DMA_CYCLIC, tdma->dma_dev.cap_mask); in tegra_adma_probe()
903 tdma->dma_dev.dev = &pdev->dev; in tegra_adma_probe()
904 tdma->dma_dev.device_alloc_chan_resources = in tegra_adma_probe()
906 tdma->dma_dev.device_free_chan_resources = in tegra_adma_probe()
[all …]
Dmv_xor_v2.c719 struct dma_device *dma_dev; in mv_xor_v2_probe() local
828 dma_dev = &xor_dev->dmadev; in mv_xor_v2_probe()
831 dma_cap_zero(dma_dev->cap_mask); in mv_xor_v2_probe()
832 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in mv_xor_v2_probe()
833 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in mv_xor_v2_probe()
834 dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); in mv_xor_v2_probe()
837 INIT_LIST_HEAD(&dma_dev->channels); in mv_xor_v2_probe()
840 dma_dev->device_tx_status = dma_cookie_status; in mv_xor_v2_probe()
841 dma_dev->device_issue_pending = mv_xor_v2_issue_pending; in mv_xor_v2_probe()
842 dma_dev->dev = &pdev->dev; in mv_xor_v2_probe()
[all …]
Dsprd-dma.c209 struct dma_device dma_dev; member
358 dev_warn(sdev->dma_dev.dev, in sprd_dma_pause_resume()
422 dev_warn(sdev->dma_dev.dev, "incorrect dma interrupt type\n"); in sprd_dma_get_int_type()
481 dev_err(sdev->dma_dev.dev, "invalid channel mode setting %d\n", in sprd_dma_set_2stage_config()
764 dev_err(sdev->dma_dev.dev, "invalid source step\n"); in sprd_dma_fill_desc()
780 dev_err(sdev->dma_dev.dev, "invalid destination step\n"); in sprd_dma_fill_desc()
788 dev_err(sdev->dma_dev.dev, "invalid source datawidth\n"); in sprd_dma_fill_desc()
794 dev_err(sdev->dma_dev.dev, "invalid destination datawidth\n"); in sprd_dma_fill_desc()
1170 dma_cap_set(DMA_MEMCPY, sdev->dma_dev.cap_mask); in sprd_dma_probe()
1172 sdev->dma_dev.chancnt = chn_count; in sprd_dma_probe()
[all …]
Dep93xx_dma.c204 struct dma_device dma_dev; member
1322 struct dma_device *dma_dev; in ep93xx_dma_probe() local
1331 dma_dev = &edma->dma_dev; in ep93xx_dma_probe()
1335 INIT_LIST_HEAD(&dma_dev->channels); in ep93xx_dma_probe()
1340 edmac->chan.device = dma_dev; in ep93xx_dma_probe()
1359 &dma_dev->channels); in ep93xx_dma_probe()
1362 dma_cap_zero(dma_dev->cap_mask); in ep93xx_dma_probe()
1363 dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); in ep93xx_dma_probe()
1364 dma_cap_set(DMA_CYCLIC, dma_dev->cap_mask); in ep93xx_dma_probe()
1366 dma_dev->dev = &pdev->dev; in ep93xx_dma_probe()
[all …]
Dhisi_dma.c161 struct dma_device dma_dev; member
712 vchan_init(&hdma_dev->chan[i].vc, &hdma_dev->dma_dev); in hisi_dma_enable_qps()
824 struct dma_device *dma_dev; in hisi_dma_init_dma_dev() local
826 dma_dev = &hdma_dev->dma_dev; in hisi_dma_init_dma_dev()
827 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in hisi_dma_init_dma_dev()
828 dma_dev->device_free_chan_resources = hisi_dma_free_chan_resources; in hisi_dma_init_dma_dev()
829 dma_dev->device_prep_dma_memcpy = hisi_dma_prep_dma_memcpy; in hisi_dma_init_dma_dev()
830 dma_dev->device_tx_status = hisi_dma_tx_status; in hisi_dma_init_dma_dev()
831 dma_dev->device_issue_pending = hisi_dma_issue_pending; in hisi_dma_init_dma_dev()
832 dma_dev->device_terminate_all = hisi_dma_terminate_all; in hisi_dma_init_dma_dev()
[all …]
Dxgene-dma.c310 struct dma_device dma_dev[XGENE_DMA_MAX_CHANNEL]; member
1485 struct dma_device *dma_dev) in xgene_dma_set_caps() argument
1488 dma_cap_zero(dma_dev->cap_mask); in xgene_dma_set_caps()
1503 dma_cap_set(DMA_PQ, dma_dev->cap_mask); in xgene_dma_set_caps()
1504 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in xgene_dma_set_caps()
1507 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in xgene_dma_set_caps()
1511 dma_dev->dev = chan->dev; in xgene_dma_set_caps()
1512 dma_dev->device_alloc_chan_resources = xgene_dma_alloc_chan_resources; in xgene_dma_set_caps()
1513 dma_dev->device_free_chan_resources = xgene_dma_free_chan_resources; in xgene_dma_set_caps()
1514 dma_dev->device_issue_pending = xgene_dma_issue_pending; in xgene_dma_set_caps()
[all …]
Dtegra186-gpc-dma.c251 struct dma_device dma_dev; member
1281 chan = dma_get_any_slave_channel(&tdma->dma_dev); in tegra_dma_of_xlate()
1375 tdma->dma_dev.dev = &pdev->dev; in tegra_dma_probe()
1384 INIT_LIST_HEAD(&tdma->dma_dev.channels); in tegra_dma_probe()
1399 vchan_init(&tdc->vc, &tdma->dma_dev); in tegra_dma_probe()
1407 dma_cap_set(DMA_SLAVE, tdma->dma_dev.cap_mask); in tegra_dma_probe()
1408 dma_cap_set(DMA_PRIVATE, tdma->dma_dev.cap_mask); in tegra_dma_probe()
1409 dma_cap_set(DMA_MEMCPY, tdma->dma_dev.cap_mask); in tegra_dma_probe()
1410 dma_cap_set(DMA_MEMSET, tdma->dma_dev.cap_mask); in tegra_dma_probe()
1411 dma_cap_set(DMA_CYCLIC, tdma->dma_dev.cap_mask); in tegra_dma_probe()
[all …]
/linux-6.1.9/drivers/spi/
Dspi-pxa2xx-pci.c78 if (dws->dma_dev != chan->device->dev) in lpss_dma_filter()
85 static void lpss_dma_put_device(void *dma_dev) in lpss_dma_put_device() argument
87 pci_dev_put(dma_dev); in lpss_dma_put_device()
94 struct pci_dev *dma_dev; in lpss_spi_setup() local
146 dma_dev = pci_get_slot(dev->bus, PCI_DEVFN(PCI_SLOT(dev->devfn), 0)); in lpss_spi_setup()
147 ret = devm_add_action_or_reset(&dev->dev, lpss_dma_put_device, dma_dev); in lpss_spi_setup()
152 tx->dma_dev = &dma_dev->dev; in lpss_spi_setup()
157 rx->dma_dev = &dma_dev->dev; in lpss_spi_setup()
190 struct pci_dev *dma_dev; in mrfld_spi_setup() local
222 dma_dev = pci_get_slot(dev->bus, PCI_DEVFN(21, 0)); in mrfld_spi_setup()
[all …]
/linux-6.1.9/drivers/crypto/ccp/
Dccp-dmaengine.c656 struct dma_device *dma_dev = &ccp->dma_dev; in ccp_dmaengine_register() local
702 dma_dev->dev = ccp->dev; in ccp_dmaengine_register()
703 dma_dev->src_addr_widths = CCP_DMA_WIDTH(dma_get_mask(ccp->dev)); in ccp_dmaengine_register()
704 dma_dev->dst_addr_widths = CCP_DMA_WIDTH(dma_get_mask(ccp->dev)); in ccp_dmaengine_register()
705 dma_dev->directions = DMA_MEM_TO_MEM; in ccp_dmaengine_register()
706 dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; in ccp_dmaengine_register()
707 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in ccp_dmaengine_register()
708 dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); in ccp_dmaengine_register()
717 dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask); in ccp_dmaengine_register()
719 INIT_LIST_HEAD(&dma_dev->channels); in ccp_dmaengine_register()
[all …]
/linux-6.1.9/drivers/mmc/host/
Dmmc_spi.c123 struct device *dma_dev; member
162 if (host->dma_dev) in mmc_spi_readbytes()
163 dma_sync_single_for_device(host->dma_dev, in mmc_spi_readbytes()
169 if (host->dma_dev) in mmc_spi_readbytes()
170 dma_sync_single_for_cpu(host->dma_dev, in mmc_spi_readbytes()
514 if (host->dma_dev) { in mmc_spi_command_send()
516 dma_sync_single_for_device(host->dma_dev, in mmc_spi_command_send()
522 if (host->dma_dev) in mmc_spi_command_send()
523 dma_sync_single_for_cpu(host->dma_dev, in mmc_spi_command_send()
656 if (host->dma_dev) in mmc_spi_writeblock()
[all …]
/linux-6.1.9/drivers/dma/sf-pdma/
Dsf-pdma.c90 dev_err(chan->pdma->dma_dev.dev, in sf_pdma_prep_dma_memcpy()
258 dev_err(chan->pdma->dma_dev.dev, "NULL desc.\n"); in sf_pdma_xfer_desc()
454 INIT_LIST_HEAD(&pdma->dma_dev.channels); in sf_pdma_setup_chans()
485 vchan_init(&chan->vchan, &pdma->dma_dev); in sf_pdma_setup_chans()
533 pdma->dma_dev.dev = &pdev->dev; in sf_pdma_probe()
536 dma_cap_set(DMA_MEMCPY, pdma->dma_dev.cap_mask); in sf_pdma_probe()
537 pdma->dma_dev.copy_align = 2; in sf_pdma_probe()
538 pdma->dma_dev.src_addr_widths = widths; in sf_pdma_probe()
539 pdma->dma_dev.dst_addr_widths = widths; in sf_pdma_probe()
540 pdma->dma_dev.directions = BIT(DMA_MEM_TO_MEM); in sf_pdma_probe()
[all …]
/linux-6.1.9/drivers/tty/serial/8250/
D8250_mid.c45 struct pci_dev *dma_dev; member
71 mid->dma_dev = pci_get_slot(pdev->bus, in pnw_setup()
78 pci_dev_put(mid->dma_dev); in pnw_exit()
90 chip = pci_get_drvdata(mid->dma_dev); in tng_handle_irq()
126 mid->dma_dev = pci_get_slot(pdev->bus, PCI_DEVFN(5, 0)); in tng_setup()
134 pci_dev_put(mid->dma_dev); in tng_exit()
194 mid->dma_dev = pdev; in dnv_setup()
202 if (!mid->dma_dev) in dnv_exit()
248 if (s->dma_dev != chan->device->dev || s->chan_id != chan->chan_id) in mid8250_dma_filter()
262 if (!mid->dma_dev) in mid8250_dma_setup()
[all …]
/linux-6.1.9/drivers/gpu/drm/exynos/
Dexynos_drm_dma.c48 if (get_dma_ops(priv->dma_dev) != get_dma_ops(subdrv_dev)) { in drm_iommu_attach_device()
100 if (!priv->dma_dev) { in exynos_drm_register_dma()
101 priv->dma_dev = dev; in exynos_drm_register_dma()
116 mapping = iommu_get_domain_for_dev(priv->dma_dev); in exynos_drm_register_dma()
144 priv->dma_dev = NULL; in exynos_drm_cleanup_dma()
/linux-6.1.9/drivers/net/ethernet/broadcom/
Dbgmac.c135 struct device *dma_dev = bgmac->dma_dev; in bgmac_dma_tx_add() local
162 slot->dma_addr = dma_map_single(dma_dev, skb->data, skb_headlen(skb), in bgmac_dma_tx_add()
164 if (unlikely(dma_mapping_error(dma_dev, slot->dma_addr))) in bgmac_dma_tx_add()
180 slot->dma_addr = skb_frag_dma_map(dma_dev, frag, 0, in bgmac_dma_tx_add()
182 if (unlikely(dma_mapping_error(dma_dev, slot->dma_addr))) in bgmac_dma_tx_add()
211 dma_unmap_single(dma_dev, slot->dma_addr, skb_headlen(skb), in bgmac_dma_tx_add()
220 dma_unmap_page(dma_dev, slot->dma_addr, len, DMA_TO_DEVICE); in bgmac_dma_tx_add()
237 struct device *dma_dev = bgmac->dma_dev; in bgmac_dma_tx_free() local
262 dma_unmap_single(dma_dev, slot->dma_addr, len, in bgmac_dma_tx_free()
265 dma_unmap_page(dma_dev, slot->dma_addr, len, in bgmac_dma_tx_free()
[all …]
/linux-6.1.9/drivers/dma/sh/
Dshdmac.c165 dev_warn(shdev->shdma_dev.dma_dev.dev, "Can't initialize DMAOR.\n"); in sh_dmae_rst()
169 dev_warn(shdev->shdma_dev.dma_dev.dev, in sh_dmae_rst()
522 struct platform_device *pdev = to_platform_device(sdev->dma_dev.dev); in sh_dmae_chan_probe()
527 sh_chan = devm_kzalloc(sdev->dma_dev.dev, sizeof(struct sh_dmae_chan), in sh_dmae_chan_probe()
549 dev_err(sdev->dma_dev.dev, in sh_dmae_chan_probe()
680 struct dma_device *dma_dev; in sh_dmae_probe() local
720 dma_dev = &shdev->shdma_dev.dma_dev; in sh_dmae_probe()
731 dma_dev->src_addr_widths = widths; in sh_dmae_probe()
732 dma_dev->dst_addr_widths = widths; in sh_dmae_probe()
733 dma_dev->directions = BIT(DMA_MEM_TO_DEV) | BIT(DMA_DEV_TO_MEM); in sh_dmae_probe()
[all …]
/linux-6.1.9/drivers/media/usb/stk1160/
Dstk1160-video.c300 struct device *dma_dev = stk1160_get_dmadev(dev); in stk1160_isoc_irq() local
317 dma_sync_sgtable_for_cpu(dma_dev, stk_urb->sgt, DMA_FROM_DEVICE); in stk1160_isoc_irq()
327 dma_sync_sgtable_for_device(dma_dev, stk_urb->sgt, DMA_FROM_DEVICE); in stk1160_isoc_irq()
365 struct device *dma_dev = stk1160_get_dmadev(dev); in stk_free_urb() local
367 dma_vunmap_noncontiguous(dma_dev, stk_urb->transfer_buffer); in stk_free_urb()
368 dma_free_noncontiguous(dma_dev, stk_urb->urb->transfer_buffer_length, in stk_free_urb()
410 struct device *dma_dev = stk1160_get_dmadev(dev); in stk1160_fill_urb() local
415 stk_urb->sgt = dma_alloc_noncontiguous(dma_dev, sb_size, in stk1160_fill_urb()
425 stk_urb->transfer_buffer = dma_vmap_noncontiguous(dma_dev, sb_size, in stk1160_fill_urb()
434 dma_free_noncontiguous(dma_dev, sb_size, stk_urb->sgt, DMA_FROM_DEVICE); in stk1160_fill_urb()

1234567