/linux-6.1.9/drivers/media/common/videobuf2/ |
D | videobuf2-dma-contig.c | 53 static unsigned long vb2_dc_get_contiguous_size(struct sg_table *sgt) in vb2_dc_get_contiguous_size() argument 56 dma_addr_t expected = sg_dma_address(sgt->sgl); in vb2_dc_get_contiguous_size() 60 for_each_sgtable_dma_sg(sgt, s, i) { in vb2_dc_get_contiguous_size() 126 struct sg_table *sgt = buf->dma_sgt; in vb2_dc_prepare() local 140 dma_sync_sgtable_for_device(buf->dev, sgt, buf->dma_dir); in vb2_dc_prepare() 146 struct sg_table *sgt = buf->dma_sgt; in vb2_dc_finish() local 160 dma_sync_sgtable_for_cpu(buf->dev, sgt, buf->dma_dir); in vb2_dc_finish() 313 struct sg_table sgt; member 323 struct sg_table *sgt; in vb2_dc_dmabuf_ops_attach() local 331 sgt = &attach->sgt; in vb2_dc_dmabuf_ops_attach() [all …]
|
D | videobuf2-dma-sg.c | 105 struct sg_table *sgt; in vb2_dma_sg_alloc() local 145 sgt = &buf->sg_table; in vb2_dma_sg_alloc() 150 if (dma_map_sgtable(buf->dev, sgt, buf->dma_dir, in vb2_dma_sg_alloc() 182 struct sg_table *sgt = &buf->sg_table; in vb2_dma_sg_put() local 188 dma_unmap_sgtable(buf->dev, sgt, buf->dma_dir, in vb2_dma_sg_put() 204 struct sg_table *sgt = buf->dma_sgt; in vb2_dma_sg_prepare() local 209 dma_sync_sgtable_for_device(buf->dev, sgt, buf->dma_dir); in vb2_dma_sg_prepare() 215 struct sg_table *sgt = buf->dma_sgt; in vb2_dma_sg_finish() local 220 dma_sync_sgtable_for_cpu(buf->dev, sgt, buf->dma_dir); in vb2_dma_sg_finish() 227 struct sg_table *sgt; in vb2_dma_sg_get_userptr() local [all …]
|
D | videobuf2-vmalloc.c | 204 struct sg_table sgt; member 214 struct sg_table *sgt; in vb2_vmalloc_dmabuf_ops_attach() local 224 sgt = &attach->sgt; in vb2_vmalloc_dmabuf_ops_attach() 225 ret = sg_alloc_table(sgt, num_pages, GFP_KERNEL); in vb2_vmalloc_dmabuf_ops_attach() 230 for_each_sgtable_sg(sgt, sg, i) { in vb2_vmalloc_dmabuf_ops_attach() 234 sg_free_table(sgt); in vb2_vmalloc_dmabuf_ops_attach() 251 struct sg_table *sgt; in vb2_vmalloc_dmabuf_ops_detach() local 256 sgt = &attach->sgt; in vb2_vmalloc_dmabuf_ops_detach() 260 dma_unmap_sgtable(db_attach->dev, sgt, attach->dma_dir, 0); in vb2_vmalloc_dmabuf_ops_detach() 261 sg_free_table(sgt); in vb2_vmalloc_dmabuf_ops_detach() [all …]
|
/linux-6.1.9/drivers/gpu/drm/tegra/ |
D | gem.c | 46 static inline unsigned int sgt_dma_count_chunks(struct sg_table *sgt) in sgt_dma_count_chunks() argument 48 return sg_dma_count_chunks(sgt->sgl, sgt->nents); in sgt_dma_count_chunks() 87 map->sgt = dma_buf_map_attachment(map->attach, direction); in tegra_bo_pin() 88 if (IS_ERR(map->sgt)) { in tegra_bo_pin() 90 err = PTR_ERR(map->sgt); in tegra_bo_pin() 91 map->sgt = NULL; in tegra_bo_pin() 95 err = sgt_dma_count_chunks(map->sgt); in tegra_bo_pin() 105 map->sgt = kzalloc(sizeof(*map->sgt), GFP_KERNEL); in tegra_bo_pin() 106 if (!map->sgt) { in tegra_bo_pin() 116 err = sg_alloc_table_from_pages(map->sgt, obj->pages, obj->num_pages, 0, gem->size, in tegra_bo_pin() [all …]
|
/linux-6.1.9/drivers/gpu/drm/armada/ |
D | armada_gem.c | 68 if (dobj->sgt) in armada_gem_free_object() 70 dobj->sgt, DMA_TO_DEVICE); in armada_gem_free_object() 392 struct sg_table *sgt; in armada_gem_prime_map_dma_buf() local 395 sgt = kmalloc(sizeof(*sgt), GFP_KERNEL); in armada_gem_prime_map_dma_buf() 396 if (!sgt) in armada_gem_prime_map_dma_buf() 404 if (sg_alloc_table(sgt, count, GFP_KERNEL)) in armada_gem_prime_map_dma_buf() 409 for_each_sgtable_sg(sgt, sg, i) { in armada_gem_prime_map_dma_buf() 419 if (dma_map_sgtable(attach->dev, sgt, dir, 0)) in armada_gem_prime_map_dma_buf() 423 if (sg_alloc_table(sgt, 1, GFP_KERNEL)) in armada_gem_prime_map_dma_buf() 426 sg_set_page(sgt->sgl, dobj->page, dobj->obj.size, 0); in armada_gem_prime_map_dma_buf() [all …]
|
/linux-6.1.9/drivers/gpu/drm/virtio/ |
D | virtgpu_vram.c | 75 struct sg_table *sgt; in virtio_gpu_vram_map_dma_buf() local 79 sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); in virtio_gpu_vram_map_dma_buf() 80 if (!sgt) in virtio_gpu_vram_map_dma_buf() 90 return sgt; in virtio_gpu_vram_map_dma_buf() 93 ret = sg_alloc_table(sgt, 1, GFP_KERNEL); in virtio_gpu_vram_map_dma_buf() 104 sg_set_page(sgt->sgl, NULL, vram->vram_node.size, 0); in virtio_gpu_vram_map_dma_buf() 105 sg_dma_address(sgt->sgl) = addr; in virtio_gpu_vram_map_dma_buf() 106 sg_dma_len(sgt->sgl) = vram->vram_node.size; in virtio_gpu_vram_map_dma_buf() 108 return sgt; in virtio_gpu_vram_map_dma_buf() 110 sg_free_table(sgt); in virtio_gpu_vram_map_dma_buf() [all …]
|
/linux-6.1.9/drivers/gpu/drm/i915/ |
D | i915_mm.c | 37 struct sgt_iter sgt; member 46 return (r->sgt.dma + r->sgt.curr + r->iobase) >> PAGE_SHIFT; in sgt_pfn() 48 return r->sgt.pfn + (r->sgt.curr >> PAGE_SHIFT); in sgt_pfn() 55 if (GEM_WARN_ON(!r->sgt.sgp)) in remap_sg() 63 r->sgt.curr += PAGE_SIZE; in remap_sg() 64 if (r->sgt.curr >= r->sgt.max) in remap_sg() 65 r->sgt = __sgt_iter(__sg_next(r->sgt.sgp), use_dma(r->iobase)); in remap_sg() 136 .sgt = __sgt_iter(sgl, use_dma(iobase)), in remap_io_sg()
|
/linux-6.1.9/drivers/hwtracing/intel_th/ |
D | msu-sink.c | 51 static int msu_sink_alloc_window(void *data, struct sg_table **sgt, size_t size) in msu_sink_alloc_window() argument 64 ret = sg_alloc_table(*sgt, nents, GFP_KERNEL); in msu_sink_alloc_window() 68 priv->sgts[priv->nr_sgts++] = *sgt; in msu_sink_alloc_window() 70 for_each_sg((*sgt)->sgl, sg_ptr, nents, i) { in msu_sink_alloc_window() 84 static void msu_sink_free_window(void *data, struct sg_table *sgt) in msu_sink_free_window() argument 90 for_each_sg(sgt->sgl, sg_ptr, sgt->nents, i) { in msu_sink_free_window() 95 sg_free_table(sgt); in msu_sink_free_window() 99 static int msu_sink_ready(void *data, struct sg_table *sgt, size_t bytes) in msu_sink_ready() argument 103 intel_th_msc_window_unlock(priv->dev, sgt); in msu_sink_ready()
|
/linux-6.1.9/drivers/xen/ |
D | gntdev-dmabuf.c | 45 struct sg_table *sgt; member 64 struct sg_table *sgt; member 197 struct sg_table *sgt; in dmabuf_pages_to_sgt() local 200 sgt = kmalloc(sizeof(*sgt), GFP_KERNEL); in dmabuf_pages_to_sgt() 201 if (!sgt) { in dmabuf_pages_to_sgt() 206 ret = sg_alloc_table_from_pages(sgt, pages, nr_pages, 0, in dmabuf_pages_to_sgt() 212 return sgt; in dmabuf_pages_to_sgt() 215 kfree(sgt); in dmabuf_pages_to_sgt() 240 struct sg_table *sgt = gntdev_dmabuf_attach->sgt; in dmabuf_exp_ops_detach() local 242 if (sgt) { in dmabuf_exp_ops_detach() [all …]
|
/linux-6.1.9/kernel/dma/ |
D | mapping.c | 266 int dma_map_sgtable(struct device *dev, struct sg_table *sgt, in dma_map_sgtable() argument 271 nents = __dma_map_sg_attrs(dev, sgt->sgl, sgt->orig_nents, dir, attrs); in dma_map_sgtable() 274 sgt->nents = nents; in dma_map_sgtable() 395 int dma_get_sgtable_attrs(struct device *dev, struct sg_table *sgt, in dma_get_sgtable_attrs() argument 402 return dma_direct_get_sgtable(dev, sgt, cpu_addr, dma_addr, in dma_get_sgtable_attrs() 406 return ops->get_sgtable(dev, sgt, cpu_addr, dma_addr, size, attrs); in dma_get_sgtable_attrs() 611 struct sg_table *sgt; in alloc_single_sgt() local 614 sgt = kmalloc(sizeof(*sgt), gfp); in alloc_single_sgt() 615 if (!sgt) in alloc_single_sgt() 617 if (sg_alloc_table(sgt, 1, gfp)) in alloc_single_sgt() [all …]
|
/linux-6.1.9/include/linux/ |
D | scatterlist.h | 46 struct sg_table sgt; /* The scatter list table */ member 177 #define for_each_sgtable_sg(sgt, sg, i) \ argument 178 for_each_sg((sgt)->sgl, sg, (sgt)->orig_nents, i) 185 #define for_each_sgtable_dma_sg(sgt, sg, i) \ argument 186 for_each_sg((sgt)->sgl, sg, (sgt)->nents, i) 377 void sg_free_append_table(struct sg_append_table *sgt); 381 int sg_alloc_append_table_from_pages(struct sg_append_table *sgt, 386 int sg_alloc_table_from_pages_segment(struct sg_table *sgt, struct page **pages, 411 static inline int sg_alloc_table_from_pages(struct sg_table *sgt, in sg_alloc_table_from_pages() argument 417 return sg_alloc_table_from_pages_segment(sgt, pages, n_pages, offset, in sg_alloc_table_from_pages() [all …]
|
/linux-6.1.9/drivers/media/platform/nvidia/tegra-vde/ |
D | dmabuf-cache.c | 27 struct sg_table *sgt; member 41 dma_buf_unmap_attachment(entry->a, entry->sgt, entry->dma_dir); in tegra_vde_release_entry() 72 struct sg_table *sgt; in tegra_vde_dmabuf_cache_map() local 93 *addrp = sg_dma_address(entry->sgt->sgl); in tegra_vde_dmabuf_cache_map() 105 sgt = dma_buf_map_attachment(attachment, dma_dir); in tegra_vde_dmabuf_cache_map() 106 if (IS_ERR(sgt)) { in tegra_vde_dmabuf_cache_map() 108 err = PTR_ERR(sgt); in tegra_vde_dmabuf_cache_map() 112 if (!vde->domain && sgt->nents > 1) { in tegra_vde_dmabuf_cache_map() 125 err = tegra_vde_iommu_map(vde, sgt, &iova, dmabuf->size); in tegra_vde_dmabuf_cache_map() 131 *addrp = sg_dma_address(sgt->sgl); in tegra_vde_dmabuf_cache_map() [all …]
|
/linux-6.1.9/net/ceph/ |
D | crypto.c | 161 static int setup_sgtable(struct sg_table *sgt, struct scatterlist *prealloc_sg, in setup_sgtable() argument 173 memset(sgt, 0, sizeof(*sgt)); in setup_sgtable() 183 ret = sg_alloc_table(sgt, chunk_cnt, GFP_NOFS); in setup_sgtable() 189 sgt->sgl = prealloc_sg; in setup_sgtable() 190 sgt->nents = sgt->orig_nents = 1; in setup_sgtable() 193 for_each_sg(sgt->sgl, sg, sgt->orig_nents, i) { in setup_sgtable() 213 static void teardown_sgtable(struct sg_table *sgt) in teardown_sgtable() argument 215 if (sgt->orig_nents > 1) in teardown_sgtable() 216 sg_free_table(sgt); in teardown_sgtable() 223 struct sg_table sgt; in ceph_aes_crypt() local [all …]
|
/linux-6.1.9/drivers/gpu/drm/mediatek/ |
D | mtk_drm_gem.c | 183 struct sg_table *sgt; in mtk_gem_prime_get_sg_table() local 186 sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); in mtk_gem_prime_get_sg_table() 187 if (!sgt) in mtk_gem_prime_get_sg_table() 190 ret = dma_get_sgtable_attrs(priv->dma_dev, sgt, mtk_gem->cookie, in mtk_gem_prime_get_sg_table() 195 kfree(sgt); in mtk_gem_prime_get_sg_table() 199 return sgt; in mtk_gem_prime_get_sg_table() 226 struct sg_table *sgt = NULL; in mtk_drm_gem_prime_vmap() local 232 sgt = mtk_gem_prime_get_sg_table(obj); in mtk_drm_gem_prime_vmap() 233 if (IS_ERR(sgt)) in mtk_drm_gem_prime_vmap() 234 return PTR_ERR(sgt); in mtk_drm_gem_prime_vmap() [all …]
|
/linux-6.1.9/drivers/infiniband/core/ |
D | umem_dmabuf.c | 17 struct sg_table *sgt; in ib_umem_dmabuf_map_pages() local 26 if (umem_dmabuf->sgt) in ib_umem_dmabuf_map_pages() 29 sgt = dma_buf_map_attachment(umem_dmabuf->attach, DMA_BIDIRECTIONAL); in ib_umem_dmabuf_map_pages() 30 if (IS_ERR(sgt)) in ib_umem_dmabuf_map_pages() 31 return PTR_ERR(sgt); in ib_umem_dmabuf_map_pages() 38 for_each_sgtable_dma_sg(sgt, sg, i) { in ib_umem_dmabuf_map_pages() 61 umem_dmabuf->umem.sgt_append.sgt.sgl = umem_dmabuf->first_sg; in ib_umem_dmabuf_map_pages() 62 umem_dmabuf->umem.sgt_append.sgt.nents = nmap; in ib_umem_dmabuf_map_pages() 63 umem_dmabuf->sgt = sgt; in ib_umem_dmabuf_map_pages() 86 if (!umem_dmabuf->sgt) in ib_umem_dmabuf_unmap_pages() [all …]
|
/linux-6.1.9/drivers/gpu/drm/ |
D | drm_prime.c | 624 struct sg_table *sgt; in drm_gem_map_dma_buf() local 633 sgt = obj->funcs->get_sg_table(obj); in drm_gem_map_dma_buf() 634 if (IS_ERR(sgt)) in drm_gem_map_dma_buf() 635 return sgt; in drm_gem_map_dma_buf() 637 ret = dma_map_sgtable(attach->dev, sgt, dir, in drm_gem_map_dma_buf() 640 sg_free_table(sgt); in drm_gem_map_dma_buf() 641 kfree(sgt); in drm_gem_map_dma_buf() 642 sgt = ERR_PTR(ret); in drm_gem_map_dma_buf() 645 return sgt; in drm_gem_map_dma_buf() 658 struct sg_table *sgt, in drm_gem_unmap_dma_buf() argument [all …]
|
D | drm_gem_shmem_helper.c | 145 drm_prime_gem_destroy(obj, shmem->sgt); in drm_gem_shmem_free() 147 if (shmem->sgt) { in drm_gem_shmem_free() 148 dma_unmap_sgtable(obj->dev->dev, shmem->sgt, in drm_gem_shmem_free() 150 sg_free_table(shmem->sgt); in drm_gem_shmem_free() 151 kfree(shmem->sgt); in drm_gem_shmem_free() 466 dma_unmap_sgtable(dev->dev, shmem->sgt, DMA_BIDIRECTIONAL, 0); in drm_gem_shmem_purge_locked() 467 sg_free_table(shmem->sgt); in drm_gem_shmem_purge_locked() 468 kfree(shmem->sgt); in drm_gem_shmem_purge_locked() 469 shmem->sgt = NULL; in drm_gem_shmem_purge_locked() 704 struct sg_table *sgt; in drm_gem_shmem_get_pages_sgt() local [all …]
|
D | drm_gem_dma_helper.c | 234 drm_prime_gem_destroy(gem_obj, dma_obj->sgt); in drm_gem_dma_free() 424 struct sg_table *sgt; in drm_gem_dma_get_sg_table() local 427 sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); in drm_gem_dma_get_sg_table() 428 if (!sgt) in drm_gem_dma_get_sg_table() 431 ret = dma_get_sgtable(obj->dev->dev, sgt, dma_obj->vaddr, in drm_gem_dma_get_sg_table() 436 return sgt; in drm_gem_dma_get_sg_table() 439 kfree(sgt); in drm_gem_dma_get_sg_table() 464 struct sg_table *sgt) in drm_gem_dma_prime_import_sg_table() argument 469 if (drm_prime_get_contiguous_size(sgt) < attach->dmabuf->size) in drm_gem_dma_prime_import_sg_table() 477 dma_obj->dma_addr = sg_dma_address(sgt->sgl); in drm_gem_dma_prime_import_sg_table() [all …]
|
/linux-6.1.9/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_dma_buf.c | 145 struct sg_table *sgt; in amdgpu_dma_buf_map() local 170 sgt = drm_prime_pages_to_sg(obj->dev, in amdgpu_dma_buf_map() 173 if (IS_ERR(sgt)) in amdgpu_dma_buf_map() 174 return sgt; in amdgpu_dma_buf_map() 176 if (dma_map_sgtable(attach->dev, sgt, dir, in amdgpu_dma_buf_map() 184 dir, &sgt); in amdgpu_dma_buf_map() 192 return sgt; in amdgpu_dma_buf_map() 195 sg_free_table(sgt); in amdgpu_dma_buf_map() 196 kfree(sgt); in amdgpu_dma_buf_map() 210 struct sg_table *sgt, in amdgpu_dma_buf_unmap() argument [all …]
|
/linux-6.1.9/drivers/gpu/drm/lima/ |
D | lima_gem.c | 29 struct sg_table sgt; in lima_heap_alloc() local 67 ret = sg_alloc_table_from_pages(&sgt, pages, i, 0, in lima_heap_alloc() 72 if (bo->base.sgt) { in lima_heap_alloc() 73 dma_unmap_sgtable(dev, bo->base.sgt, DMA_BIDIRECTIONAL, 0); in lima_heap_alloc() 74 sg_free_table(bo->base.sgt); in lima_heap_alloc() 76 bo->base.sgt = kmalloc(sizeof(*bo->base.sgt), GFP_KERNEL); in lima_heap_alloc() 77 if (!bo->base.sgt) { in lima_heap_alloc() 78 sg_free_table(&sgt); in lima_heap_alloc() 83 ret = dma_map_sgtable(dev, &sgt, DMA_BIDIRECTIONAL, 0); in lima_heap_alloc() 85 sg_free_table(&sgt); in lima_heap_alloc() [all …]
|
/linux-6.1.9/drivers/gpu/drm/etnaviv/ |
D | etnaviv_gem.c | 23 struct sg_table *sgt = etnaviv_obj->sgt; in etnaviv_gem_scatter_map() local 30 dma_map_sgtable(dev->dev, sgt, DMA_BIDIRECTIONAL, 0); in etnaviv_gem_scatter_map() 36 struct sg_table *sgt = etnaviv_obj->sgt; in etnaviv_gem_scatterlist_unmap() local 54 dma_unmap_sgtable(dev->dev, sgt, DMA_BIDIRECTIONAL, 0); in etnaviv_gem_scatterlist_unmap() 75 if (etnaviv_obj->sgt) { in put_pages() 77 sg_free_table(etnaviv_obj->sgt); in put_pages() 78 kfree(etnaviv_obj->sgt); in put_pages() 79 etnaviv_obj->sgt = NULL; in put_pages() 101 if (!etnaviv_obj->sgt) { in etnaviv_gem_get_pages() 104 struct sg_table *sgt; in etnaviv_gem_get_pages() local [all …]
|
/linux-6.1.9/drivers/gpu/drm/rockchip/ |
D | rockchip_drm_gem.c | 40 ret = iommu_map_sgtable(private->domain, rk_obj->dma_addr, rk_obj->sgt, in rockchip_gem_iommu_map() 89 rk_obj->sgt = drm_prime_pages_to_sg(rk_obj->base.dev, in rockchip_gem_get_pages() 91 if (IS_ERR(rk_obj->sgt)) { in rockchip_gem_get_pages() 92 ret = PTR_ERR(rk_obj->sgt); in rockchip_gem_get_pages() 103 for_each_sgtable_sg(rk_obj->sgt, s, i) in rockchip_gem_get_pages() 106 dma_sync_sgtable_for_device(drm->dev, rk_obj->sgt, DMA_TO_DEVICE); in rockchip_gem_get_pages() 117 sg_free_table(rk_obj->sgt); in rockchip_gem_put_pages() 118 kfree(rk_obj->sgt); in rockchip_gem_put_pages() 342 dma_unmap_sgtable(drm->dev, rk_obj->sgt, in rockchip_gem_free_object() 345 drm_prime_gem_destroy(obj, rk_obj->sgt); in rockchip_gem_free_object() [all …]
|
/linux-6.1.9/drivers/fpga/ |
D | fpga-mgr.c | 96 struct sg_table *sgt) in fpga_mgr_write_sg() argument 99 return mgr->mops->write_sg(mgr, sgt); in fpga_mgr_write_sg() 182 struct sg_table *sgt) in fpga_mgr_parse_header_sg_first() argument 189 sg_miter_start(&miter, sgt->sgl, sgt->nents, SG_MITER_FROM_SG); in fpga_mgr_parse_header_sg_first() 214 struct sg_table *sgt, size_t *ret_size) in fpga_mgr_parse_header_sg() argument 236 len = sg_pcopy_to_buffer(sgt->sgl, sgt->nents, in fpga_mgr_parse_header_sg() 295 struct sg_table *sgt) in fpga_mgr_prepare_sg() argument 310 ret = fpga_mgr_parse_header_sg_first(mgr, info, sgt); in fpga_mgr_prepare_sg() 313 sg_miter_start(&miter, sgt->sgl, sgt->nents, SG_MITER_FROM_SG); in fpga_mgr_prepare_sg() 333 buf = fpga_mgr_parse_header_sg(mgr, info, sgt, &len); in fpga_mgr_prepare_sg() [all …]
|
/linux-6.1.9/drivers/gpu/drm/exynos/ |
D | exynos_drm_gem.c | 125 drm_prime_gem_destroy(obj, exynos_gem->sgt); in exynos_drm_gem_destroy() 409 struct sg_table *sgt; in exynos_drm_gem_prime_get_sg_table() local 412 sgt = kzalloc(sizeof(*sgt), GFP_KERNEL); in exynos_drm_gem_prime_get_sg_table() 413 if (!sgt) in exynos_drm_gem_prime_get_sg_table() 416 ret = dma_get_sgtable_attrs(to_dma_dev(drm_dev), sgt, exynos_gem->cookie, in exynos_drm_gem_prime_get_sg_table() 421 kfree(sgt); in exynos_drm_gem_prime_get_sg_table() 425 return sgt; in exynos_drm_gem_prime_get_sg_table() 431 struct sg_table *sgt) in exynos_drm_gem_prime_import_sg_table() argument 436 if (drm_prime_get_contiguous_size(sgt) < attach->dmabuf->size) { in exynos_drm_gem_prime_import_sg_table() 455 exynos_gem->dma_addr = sg_dma_address(sgt->sgl); in exynos_drm_gem_prime_import_sg_table() [all …]
|
/linux-6.1.9/drivers/gpu/drm/rcar-du/ |
D | rcar_du_vsp.c | 206 struct sg_table *sgt = &sg_tables[i]; in rcar_du_vsp_map_fb() local 208 if (gem->sgt) { in rcar_du_vsp_map_fb() 219 ret = sg_alloc_table(sgt, gem->sgt->orig_nents, in rcar_du_vsp_map_fb() 224 src = gem->sgt->sgl; in rcar_du_vsp_map_fb() 225 dst = sgt->sgl; in rcar_du_vsp_map_fb() 226 for (j = 0; j < gem->sgt->orig_nents; ++j) { in rcar_du_vsp_map_fb() 233 ret = dma_get_sgtable(rcdu->dev, sgt, gem->vaddr, in rcar_du_vsp_map_fb() 239 ret = vsp1_du_map_sg(vsp->vsp, sgt); in rcar_du_vsp_map_fb() 241 sg_free_table(sgt); in rcar_du_vsp_map_fb() 250 struct sg_table *sgt = &sg_tables[i]; in rcar_du_vsp_map_fb() local [all …]
|