/linux-6.6.21/drivers/gpu/drm/rockchip/ |
D | rockchip_drm_gem.c | 137 rk_obj->kvaddr = vmap(rk_obj->pages, rk_obj->num_pages, VM_MAP, in rockchip_gem_alloc_iommu() 139 if (!rk_obj->kvaddr) { in rockchip_gem_alloc_iommu() 167 rk_obj->kvaddr = dma_alloc_attrs(drm->dev, obj->size, in rockchip_gem_alloc_dma() 170 if (!rk_obj->kvaddr) { in rockchip_gem_alloc_dma() 193 vunmap(rk_obj->kvaddr); in rockchip_gem_free_iommu() 203 dma_free_attrs(drm->dev, obj->size, rk_obj->kvaddr, rk_obj->dma_addr, in rockchip_gem_free_dma() 234 return dma_mmap_attrs(drm->dev, vma, rk_obj->kvaddr, rk_obj->dma_addr, in rockchip_drm_gem_object_mmap_dma() 440 ret = dma_get_sgtable_attrs(drm->dev, sgt, rk_obj->kvaddr, in rockchip_gem_prime_get_sg_table() 520 if (rk_obj->kvaddr) in rockchip_gem_prime_vmap() 521 vaddr = rk_obj->kvaddr; in rockchip_gem_prime_vmap() [all …]
|
D | rockchip_drm_gem.h | 16 void *kvaddr; member
|
/linux-6.6.21/drivers/accel/ivpu/ |
D | ivpu_gem.h | 28 void *kvaddr; member 113 return bo->kvaddr + (vpu_addr - bo->vpu_addr); in ivpu_to_cpu_addr() 118 if (cpu_addr < bo->kvaddr) in cpu_to_vpu_addr() 121 if (cpu_addr >= (bo->kvaddr + bo->base.size)) in cpu_to_vpu_addr() 124 return bo->vpu_addr + (cpu_addr - bo->kvaddr); in cpu_to_vpu_addr()
|
D | ivpu_fw_log.c | 37 log = bo->kvaddr + *offset; in fw_log_ptr() 46 if ((char *)log + log->size > (char *)bo->kvaddr + bo->base.size) { in fw_log_ptr()
|
D | ivpu_fw.c | 322 memset(fw->mem->kvaddr, 0, fw->image_load_offset); in ivpu_fw_load() 323 memcpy(fw->mem->kvaddr + fw->image_load_offset, in ivpu_fw_load() 327 u8 *start = fw->mem->kvaddr + image_end_offset; in ivpu_fw_load()
|
D | ivpu_gem.c | 418 vunmap(bo->kvaddr); in ivpu_bo_free() 601 bo->kvaddr = vmap(bo->pages, bo->base.size >> PAGE_SHIFT, VM_MAP, prot); in ivpu_bo_alloc_internal() 602 if (!bo->kvaddr) { in ivpu_bo_alloc_internal()
|
D | ivpu_ipc.c | 500 memset(ipc->mem_tx->kvaddr, 0, ipc->mem_tx->base.size); in ivpu_ipc_reset() 501 memset(ipc->mem_rx->kvaddr, 0, ipc->mem_rx->base.size); in ivpu_ipc_reset()
|
D | ivpu_pm.c | 40 struct vpu_boot_params *bp = fw->mem->kvaddr; in ivpu_pm_prepare_warm_boot()
|
D | ivpu_drv.c | 349 ivpu_fw_boot_params_setup(vdev, vdev->fw->mem->kvaddr); in ivpu_boot()
|
D | ivpu_job.c | 54 cmdq->jobq = (struct vpu_job_queue *)cmdq->mem->kvaddr; in ivpu_cmdq_alloc()
|
/linux-6.6.21/arch/xtensa/mm/ |
D | cache.c | 63 unsigned long kvaddr; in kmap_invalidate_coherent() local 66 kvaddr = (unsigned long)page_to_virt(page); in kmap_invalidate_coherent() 68 __invalidate_dcache_page(kvaddr); in kmap_invalidate_coherent() 70 kvaddr = TLBTEMP_BASE_1 + in kmap_invalidate_coherent() 74 __invalidate_dcache_page_alias(kvaddr, in kmap_invalidate_coherent() 91 void *kvaddr = coherent_kvaddr(page, TLBTEMP_BASE_1, vaddr, &paddr); in clear_user_highpage() local 96 clear_page_alias(kvaddr, paddr); in clear_user_highpage()
|
/linux-6.6.21/arch/sh/mm/ |
D | kmap.c | 51 void kunmap_coherent(void *kvaddr) in kunmap_coherent() argument 53 if (kvaddr >= (void *)FIXADDR_START) { in kunmap_coherent() 54 unsigned long vaddr = (unsigned long)kvaddr & PAGE_MASK; in kunmap_coherent()
|
D | nommu.c | 86 void kunmap_coherent(void *kvaddr) in kunmap_coherent() argument
|
/linux-6.6.21/drivers/gpu/drm/mediatek/ |
D | mtk_drm_gem.c | 86 mtk_gem->kvaddr = mtk_gem->cookie; in mtk_drm_gem_create() 239 if (mtk_gem->kvaddr) in mtk_drm_gem_prime_vmap() 256 mtk_gem->kvaddr = vmap(mtk_gem->pages, npages, VM_MAP, in mtk_drm_gem_prime_vmap() 258 if (!mtk_gem->kvaddr) { in mtk_drm_gem_prime_vmap() 268 iosys_map_set_vaddr(map, mtk_gem->kvaddr); in mtk_drm_gem_prime_vmap() 283 mtk_gem->kvaddr = NULL; in mtk_drm_gem_prime_vunmap()
|
D | mtk_drm_gem.h | 28 void *kvaddr; member
|
/linux-6.6.21/arch/arc/mm/ |
D | highmem.c | 50 static noinline pte_t * __init alloc_kmap_pgtable(unsigned long kvaddr) in alloc_kmap_pgtable() argument 52 pmd_t *pmd_k = pmd_off_k(kvaddr); in alloc_kmap_pgtable()
|
/linux-6.6.21/drivers/infiniband/hw/qib/ |
D | qib_user_sdma.c | 94 void *kvaddr; /* FIXME: only for pio hack */ member 258 struct page *page, void *kvaddr, in qib_user_sdma_init_frag() argument 268 pkt->addr[i].kvaddr = kvaddr; in qib_user_sdma_init_frag() 299 u16 offset, u16 len, void *kvaddr) in qib_user_sdma_page_to_frags() argument 367 page, kvaddr, /* struct page, virt addr */ in qib_user_sdma_page_to_frags() 384 pkt->addr[pkt->index].kvaddr, in qib_user_sdma_page_to_frags() 438 pbc16 = (__le16 *)pkt->addr[pkt->index].kvaddr; in qib_user_sdma_page_to_frags() 475 pkt->addr[pkt->index].kvaddr, in qib_user_sdma_page_to_frags() 543 kvaddr = NULL; in qib_user_sdma_page_to_frags() 632 } else if (pkt->addr[i].kvaddr) { in qib_user_sdma_free_pkt_frag() [all …]
|
D | qib_file_ops.c | 709 unsigned len, void *kvaddr, u32 write_ok, char *what) in qib_mmap_mem() argument 739 pfn = virt_to_phys(kvaddr) >> PAGE_SHIFT; in qib_mmap_mem()
|
/linux-6.6.21/drivers/infiniband/hw/hfi1/ |
D | sdma.h | 627 int type, void *kvaddr, struct page *page, 798 void *kvaddr, in sdma_txadd_kvaddr() argument 806 kvaddr, NULL, 0, len); in sdma_txadd_kvaddr() 813 kvaddr, in sdma_txadd_kvaddr()
|
D | sdma.c | 3070 int type, void *kvaddr, struct page *page, in ext_coal_sdma_tx_descs() argument 3090 kvaddr = kmap_local_page(page); in ext_coal_sdma_tx_descs() 3091 kvaddr += offset; in ext_coal_sdma_tx_descs() 3092 } else if (WARN_ON(!kvaddr)) { in ext_coal_sdma_tx_descs() 3097 memcpy(tx->coalesce_buf + tx->coalesce_idx, kvaddr, len); in ext_coal_sdma_tx_descs() 3100 kunmap_local(kvaddr); in ext_coal_sdma_tx_descs()
|
/linux-6.6.21/drivers/gpu/drm/exynos/ |
D | exynos_drm_gem.h | 43 void *kvaddr; member
|
D | exynos_drm_fbdev.c | 83 fbi->screen_buffer = exynos_gem->kvaddr + offset; in exynos_drm_fbdev_update()
|
D | exynos_drm_gem.c | 65 exynos_gem->kvaddr = exynos_gem->cookie; in exynos_drm_alloc_buf()
|
/linux-6.6.21/arch/sh/include/asm/ |
D | cacheflush.h | 101 void kunmap_coherent(void *kvaddr);
|