Home
last modified time | relevance | path

Searched refs:bo (Results 1 – 25 of 291) sorted by relevance

12345678910>>...12

/linux-6.1.9/drivers/staging/media/atomisp/pci/hmm/
Dhmm_bo.c47 static int __bo_init(struct hmm_bo_device *bdev, struct hmm_buffer_object *bo, in __bo_init() argument
59 memset(bo, 0, sizeof(*bo)); in __bo_init()
60 mutex_init(&bo->mutex); in __bo_init()
63 INIT_LIST_HEAD(&bo->list); in __bo_init()
65 bo->bdev = bdev; in __bo_init()
66 bo->vmap_addr = NULL; in __bo_init()
67 bo->status = HMM_BO_FREE; in __bo_init()
68 bo->start = bdev->start; in __bo_init()
69 bo->pgnr = pgnr; in __bo_init()
70 bo->end = bo->start + pgnr_to_size(pgnr); in __bo_init()
[all …]
Dhmm.c54 struct hmm_buffer_object *bo; in bo_show() local
69 list_for_each_entry(bo, bo_list, list) { in bo_show()
70 if ((active && (bo->status & HMM_BO_ALLOCED)) || in bo_show()
71 (!active && !(bo->status & HMM_BO_ALLOCED))) { in bo_show()
74 hmm_bo_type_string[bo->type], bo->pgnr); in bo_show()
76 total[bo->type] += bo->pgnr; in bo_show()
77 count[bo->type]++; in bo_show()
174 struct hmm_buffer_object *bo; in __hmm_alloc() local
188 bo = hmm_bo_alloc(&bo_device, pgnr); in __hmm_alloc()
189 if (!bo) { in __hmm_alloc()
[all …]
/linux-6.1.9/drivers/gpu/drm/ttm/
Dttm_bo.c47 static void ttm_bo_mem_space_debug(struct ttm_buffer_object *bo, in ttm_bo_mem_space_debug() argument
55 bo, bo->resource->num_pages, bo->base.size >> 10, in ttm_bo_mem_space_debug()
56 bo->base.size >> 20); in ttm_bo_mem_space_debug()
61 man = ttm_manager_type(bo->bdev, mem_type); in ttm_bo_mem_space_debug()
75 void ttm_bo_move_to_lru_tail(struct ttm_buffer_object *bo) in ttm_bo_move_to_lru_tail() argument
77 dma_resv_assert_held(bo->base.resv); in ttm_bo_move_to_lru_tail()
79 if (bo->resource) in ttm_bo_move_to_lru_tail()
80 ttm_resource_move_to_lru_tail(bo->resource); in ttm_bo_move_to_lru_tail()
97 void ttm_bo_set_bulk_move(struct ttm_buffer_object *bo, in ttm_bo_set_bulk_move() argument
100 dma_resv_assert_held(bo->base.resv); in ttm_bo_set_bulk_move()
[all …]
Dttm_bo_util.c47 struct ttm_buffer_object *bo; member
131 int ttm_bo_move_memcpy(struct ttm_buffer_object *bo, in ttm_bo_move_memcpy() argument
135 struct ttm_device *bdev = bo->bdev; in ttm_bo_move_memcpy()
137 ttm_manager_type(bo->bdev, dst_mem->mem_type); in ttm_bo_move_memcpy()
138 struct ttm_tt *ttm = bo->ttm; in ttm_bo_move_memcpy()
139 struct ttm_resource *src_mem = bo->resource; in ttm_bo_move_memcpy()
162 dst_iter = ttm_kmap_iter_tt_init(&_dst_iter.tt, bo->ttm); in ttm_bo_move_memcpy()
168 src_iter = ttm_kmap_iter_tt_init(&_src_iter.tt, bo->ttm); in ttm_bo_move_memcpy()
180 ttm_bo_move_sync_cleanup(bo, dst_mem); in ttm_bo_move_memcpy()
190 static void ttm_transfered_destroy(struct ttm_buffer_object *bo) in ttm_transfered_destroy() argument
[all …]
Dttm_bo_vm.c46 static vm_fault_t ttm_bo_vm_fault_idle(struct ttm_buffer_object *bo, in ttm_bo_vm_fault_idle() argument
54 if (dma_resv_test_signaled(bo->base.resv, DMA_RESV_USAGE_KERNEL)) in ttm_bo_vm_fault_idle()
66 ttm_bo_get(bo); in ttm_bo_vm_fault_idle()
68 (void)dma_resv_wait_timeout(bo->base.resv, in ttm_bo_vm_fault_idle()
71 dma_resv_unlock(bo->base.resv); in ttm_bo_vm_fault_idle()
72 ttm_bo_put(bo); in ttm_bo_vm_fault_idle()
79 err = dma_resv_wait_timeout(bo->base.resv, DMA_RESV_USAGE_KERNEL, true, in ttm_bo_vm_fault_idle()
89 static unsigned long ttm_bo_io_mem_pfn(struct ttm_buffer_object *bo, in ttm_bo_io_mem_pfn() argument
92 struct ttm_device *bdev = bo->bdev; in ttm_bo_io_mem_pfn()
95 return bdev->funcs->io_mem_pfn(bo, page_offset); in ttm_bo_io_mem_pfn()
[all …]
/linux-6.1.9/drivers/gpu/drm/qxl/
Dqxl_object.c32 static int __qxl_bo_pin(struct qxl_bo *bo);
33 static void __qxl_bo_unpin(struct qxl_bo *bo);
37 struct qxl_bo *bo; in qxl_ttm_bo_destroy() local
40 bo = to_qxl_bo(tbo); in qxl_ttm_bo_destroy()
41 qdev = to_qxl(bo->tbo.base.dev); in qxl_ttm_bo_destroy()
43 qxl_surface_evict(qdev, bo, false); in qxl_ttm_bo_destroy()
44 WARN_ON_ONCE(bo->map_count > 0); in qxl_ttm_bo_destroy()
46 list_del_init(&bo->list); in qxl_ttm_bo_destroy()
48 drm_gem_object_release(&bo->tbo.base); in qxl_ttm_bo_destroy()
49 kfree(bo); in qxl_ttm_bo_destroy()
[all …]
Dqxl_object.h30 static inline int qxl_bo_reserve(struct qxl_bo *bo) in qxl_bo_reserve() argument
34 r = ttm_bo_reserve(&bo->tbo, true, false, NULL); in qxl_bo_reserve()
37 struct drm_device *ddev = bo->tbo.base.dev; in qxl_bo_reserve()
39 dev_err(ddev->dev, "%p reserve failed\n", bo); in qxl_bo_reserve()
46 static inline void qxl_bo_unreserve(struct qxl_bo *bo) in qxl_bo_unreserve() argument
48 ttm_bo_unreserve(&bo->tbo); in qxl_bo_unreserve()
51 static inline unsigned long qxl_bo_size(struct qxl_bo *bo) in qxl_bo_size() argument
53 return bo->tbo.base.size; in qxl_bo_size()
62 int qxl_bo_vmap(struct qxl_bo *bo, struct iosys_map *map);
63 int qxl_bo_vmap_locked(struct qxl_bo *bo, struct iosys_map *map);
[all …]
Dqxl_release.c124 struct qxl_bo *bo; in qxl_release_free_list() local
128 bo = to_qxl_bo(entry->tv.bo); in qxl_release_free_list()
129 qxl_bo_unref(&bo); in qxl_release_free_list()
163 struct qxl_bo **bo, in qxl_release_bo_alloc() argument
168 QXL_GEM_DOMAIN_VRAM, priority, NULL, bo); in qxl_release_bo_alloc()
171 int qxl_release_list_add(struct qxl_release *release, struct qxl_bo *bo) in qxl_release_list_add() argument
176 if (entry->tv.bo == &bo->tbo) in qxl_release_list_add()
184 qxl_bo_ref(bo); in qxl_release_list_add()
185 entry->tv.bo = &bo->tbo; in qxl_release_list_add()
191 static int qxl_release_validate_bo(struct qxl_bo *bo) in qxl_release_validate_bo() argument
[all …]
/linux-6.1.9/drivers/gpu/drm/radeon/
Dradeon_object.c45 static void radeon_bo_clear_surface_reg(struct radeon_bo *bo);
54 struct radeon_bo *bo; in radeon_ttm_bo_destroy() local
56 bo = container_of(tbo, struct radeon_bo, tbo); in radeon_ttm_bo_destroy()
58 mutex_lock(&bo->rdev->gem.mutex); in radeon_ttm_bo_destroy()
59 list_del_init(&bo->list); in radeon_ttm_bo_destroy()
60 mutex_unlock(&bo->rdev->gem.mutex); in radeon_ttm_bo_destroy()
61 radeon_bo_clear_surface_reg(bo); in radeon_ttm_bo_destroy()
62 WARN_ON_ONCE(!list_empty(&bo->va)); in radeon_ttm_bo_destroy()
63 if (bo->tbo.base.import_attach) in radeon_ttm_bo_destroy()
64 drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg); in radeon_ttm_bo_destroy()
[all …]
Dradeon_object.h64 static inline int radeon_bo_reserve(struct radeon_bo *bo, bool no_intr) in radeon_bo_reserve() argument
68 r = ttm_bo_reserve(&bo->tbo, !no_intr, false, NULL); in radeon_bo_reserve()
71 dev_err(bo->rdev->dev, "%p reserve failed\n", bo); in radeon_bo_reserve()
77 static inline void radeon_bo_unreserve(struct radeon_bo *bo) in radeon_bo_unreserve() argument
79 ttm_bo_unreserve(&bo->tbo); in radeon_bo_unreserve()
91 static inline u64 radeon_bo_gpu_offset(struct radeon_bo *bo) in radeon_bo_gpu_offset() argument
96 rdev = radeon_get_rdev(bo->tbo.bdev); in radeon_bo_gpu_offset()
98 switch (bo->tbo.resource->mem_type) { in radeon_bo_gpu_offset()
107 return (bo->tbo.resource->start << PAGE_SHIFT) + start; in radeon_bo_gpu_offset()
110 static inline unsigned long radeon_bo_size(struct radeon_bo *bo) in radeon_bo_size() argument
[all …]
Dradeon_prime.c37 struct radeon_bo *bo = gem_to_radeon_bo(obj); in radeon_gem_prime_get_sg_table() local
39 return drm_prime_pages_to_sg(obj->dev, bo->tbo.ttm->pages, in radeon_gem_prime_get_sg_table()
40 bo->tbo.ttm->num_pages); in radeon_gem_prime_get_sg_table()
49 struct radeon_bo *bo; in radeon_gem_prime_import_sg_table() local
54 RADEON_GEM_DOMAIN_GTT, 0, sg, resv, &bo); in radeon_gem_prime_import_sg_table()
59 bo->tbo.base.funcs = &radeon_gem_object_funcs; in radeon_gem_prime_import_sg_table()
62 list_add_tail(&bo->list, &rdev->gem.objects); in radeon_gem_prime_import_sg_table()
65 bo->prime_shared_count = 1; in radeon_gem_prime_import_sg_table()
66 return &bo->tbo.base; in radeon_gem_prime_import_sg_table()
71 struct radeon_bo *bo = gem_to_radeon_bo(obj); in radeon_gem_prime_pin() local
[all …]
Dradeon_mn.c53 struct radeon_bo *bo = container_of(mn, struct radeon_bo, notifier); in radeon_mn_invalidate() local
57 if (!bo->tbo.ttm || !radeon_ttm_tt_is_bound(bo->tbo.bdev, bo->tbo.ttm)) in radeon_mn_invalidate()
63 r = radeon_bo_reserve(bo, true); in radeon_mn_invalidate()
69 r = dma_resv_wait_timeout(bo->tbo.base.resv, DMA_RESV_USAGE_BOOKKEEP, in radeon_mn_invalidate()
74 radeon_ttm_placement_from_domain(bo, RADEON_GEM_DOMAIN_CPU); in radeon_mn_invalidate()
75 r = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); in radeon_mn_invalidate()
79 radeon_bo_unreserve(bo); in radeon_mn_invalidate()
96 int radeon_mn_register(struct radeon_bo *bo, unsigned long addr) in radeon_mn_register() argument
100 ret = mmu_interval_notifier_insert(&bo->notifier, current->mm, addr, in radeon_mn_register()
101 radeon_bo_size(bo), &radeon_mn_ops); in radeon_mn_register()
[all …]
/linux-6.1.9/drivers/gpu/drm/tegra/
Dgem.c51 static void tegra_bo_put(struct host1x_bo *bo) in tegra_bo_put() argument
53 struct tegra_bo *obj = host1x_to_tegra_bo(bo); in tegra_bo_put()
58 static struct host1x_bo_mapping *tegra_bo_pin(struct device *dev, struct host1x_bo *bo, in tegra_bo_pin() argument
61 struct tegra_bo *obj = host1x_to_tegra_bo(bo); in tegra_bo_pin()
71 map->bo = host1x_bo_get(bo); in tegra_bo_pin()
171 host1x_bo_put(map->bo); in tegra_bo_unpin()
175 static void *tegra_bo_mmap(struct host1x_bo *bo) in tegra_bo_mmap() argument
177 struct tegra_bo *obj = host1x_to_tegra_bo(bo); in tegra_bo_mmap()
192 static void tegra_bo_munmap(struct host1x_bo *bo, void *addr) in tegra_bo_munmap() argument
194 struct tegra_bo *obj = host1x_to_tegra_bo(bo); in tegra_bo_munmap()
[all …]
/linux-6.1.9/drivers/gpu/drm/amd/amdgpu/
Damdgpu_object.c58 struct amdgpu_bo *bo = ttm_to_amdgpu_bo(tbo); in amdgpu_bo_destroy() local
60 amdgpu_bo_kunmap(bo); in amdgpu_bo_destroy()
62 if (bo->tbo.base.import_attach) in amdgpu_bo_destroy()
63 drm_prime_gem_destroy(&bo->tbo.base, bo->tbo.sg); in amdgpu_bo_destroy()
64 drm_gem_object_release(&bo->tbo.base); in amdgpu_bo_destroy()
65 amdgpu_bo_unref(&bo->parent); in amdgpu_bo_destroy()
66 kvfree(bo); in amdgpu_bo_destroy()
71 struct amdgpu_bo *bo = ttm_to_amdgpu_bo(tbo); in amdgpu_bo_user_destroy() local
74 ubo = to_amdgpu_bo_user(bo); in amdgpu_bo_user_destroy()
82 struct amdgpu_bo *bo = ttm_to_amdgpu_bo(tbo); in amdgpu_bo_vm_destroy() local
[all …]
Damdgpu_object.h45 #define to_amdgpu_bo_user(abo) container_of((abo), struct amdgpu_bo_user, bo)
46 #define to_amdgpu_bo_vm(abo) container_of((abo), struct amdgpu_bo_vm, bo)
58 void (*destroy)(struct ttm_buffer_object *bo);
114 struct amdgpu_bo bo; member
123 struct amdgpu_bo bo; member
170 static inline int amdgpu_bo_reserve(struct amdgpu_bo *bo, bool no_intr) in amdgpu_bo_reserve() argument
172 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_bo_reserve()
175 r = ttm_bo_reserve(&bo->tbo, !no_intr, false, NULL); in amdgpu_bo_reserve()
178 dev_err(adev->dev, "%p reserve failed\n", bo); in amdgpu_bo_reserve()
184 static inline void amdgpu_bo_unreserve(struct amdgpu_bo *bo) in amdgpu_bo_unreserve() argument
[all …]
Damdgpu_dma_buf.c57 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); in amdgpu_dma_buf_attach() local
58 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_dma_buf_attach()
87 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); in amdgpu_dma_buf_detach() local
88 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_dma_buf_detach()
104 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); in amdgpu_dma_buf_pin() local
107 return amdgpu_bo_pin(bo, AMDGPU_GEM_DOMAIN_GTT); in amdgpu_dma_buf_pin()
120 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); in amdgpu_dma_buf_unpin() local
122 amdgpu_bo_unpin(bo); in amdgpu_dma_buf_unpin()
143 struct amdgpu_bo *bo = gem_to_amdgpu_bo(obj); in amdgpu_dma_buf_map() local
144 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_dma_buf_map()
[all …]
Damdgpu_amdkfd_gpuvm.c230 void amdgpu_amdkfd_release_notify(struct amdgpu_bo *bo) in amdgpu_amdkfd_release_notify() argument
232 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_amdkfd_release_notify()
233 u32 alloc_flags = bo->kfd_bo->alloc_flags; in amdgpu_amdkfd_release_notify()
234 u64 size = amdgpu_bo_size(bo); in amdgpu_amdkfd_release_notify()
238 kfree(bo->kfd_bo); in amdgpu_amdkfd_release_notify()
256 ret = amdgpu_bo_reserve(mem->bo, false); in create_dmamap_sg_bo()
261 ret = amdgpu_gem_object_create(adev, mem->bo->tbo.base.size, align, in create_dmamap_sg_bo()
263 ttm_bo_type_sg, mem->bo->tbo.base.resv, &gem_obj); in create_dmamap_sg_bo()
265 amdgpu_bo_unreserve(mem->bo); in create_dmamap_sg_bo()
273 (*bo_out)->parent = amdgpu_bo_ref(mem->bo); in create_dmamap_sg_bo()
[all …]
/linux-6.1.9/drivers/gpu/drm/vc4/
Dvc4_bo.c134 struct vc4_bo *bo = to_vc4_bo(gem_obj); in vc4_bo_set_label() local
144 vc4->bo_labels[bo->label].num_allocated--; in vc4_bo_set_label()
145 vc4->bo_labels[bo->label].size_allocated -= gem_obj->size; in vc4_bo_set_label()
147 if (vc4->bo_labels[bo->label].num_allocated == 0 && in vc4_bo_set_label()
148 is_user_label(bo->label)) { in vc4_bo_set_label()
154 kfree(vc4->bo_labels[bo->label].name); in vc4_bo_set_label()
155 vc4->bo_labels[bo->label].name = NULL; in vc4_bo_set_label()
158 bo->label = label; in vc4_bo_set_label()
166 static void vc4_bo_destroy(struct vc4_bo *bo) in vc4_bo_destroy() argument
168 struct drm_gem_object *obj = &bo->base.base; in vc4_bo_destroy()
[all …]
/linux-6.1.9/drivers/gpu/drm/vmwgfx/
Dvmwgfx_bo.c44 vmw_buffer_object(struct ttm_buffer_object *bo) in vmw_buffer_object() argument
46 return container_of(bo, struct vmw_buffer_object, base); in vmw_buffer_object()
59 static bool bo_is_vmw(struct ttm_buffer_object *bo) in bo_is_vmw() argument
61 return bo->destroy == &vmw_bo_bo_free || in bo_is_vmw()
62 bo->destroy == &vmw_gem_destroy; in bo_is_vmw()
81 struct ttm_buffer_object *bo = &buf->base; in vmw_bo_pin_in_placement() local
86 ret = ttm_bo_reserve(bo, interruptible, false, NULL); in vmw_bo_pin_in_placement()
91 ret = ttm_resource_compat(bo->resource, placement) in vmw_bo_pin_in_placement()
94 ret = ttm_bo_validate(bo, placement, &ctx); in vmw_bo_pin_in_placement()
99 ttm_bo_unreserve(bo); in vmw_bo_pin_in_placement()
[all …]
/linux-6.1.9/include/drm/ttm/
Dttm_bo_driver.h71 int ttm_bo_mem_space(struct ttm_buffer_object *bo,
81 void ttm_bo_unmap_virtual(struct ttm_buffer_object *bo);
105 static inline int ttm_bo_reserve(struct ttm_buffer_object *bo, in ttm_bo_reserve() argument
116 success = dma_resv_trylock(bo->base.resv); in ttm_bo_reserve()
121 ret = dma_resv_lock_interruptible(bo->base.resv, ticket); in ttm_bo_reserve()
123 ret = dma_resv_lock(bo->base.resv, ticket); in ttm_bo_reserve()
139 static inline int ttm_bo_reserve_slowpath(struct ttm_buffer_object *bo, in ttm_bo_reserve_slowpath() argument
144 int ret = dma_resv_lock_slow_interruptible(bo->base.resv, in ttm_bo_reserve_slowpath()
150 dma_resv_lock_slow(bo->base.resv, ticket); in ttm_bo_reserve_slowpath()
155 ttm_bo_move_to_lru_tail_unlocked(struct ttm_buffer_object *bo) in ttm_bo_move_to_lru_tail_unlocked() argument
[all …]
/linux-6.1.9/drivers/gpu/drm/lima/
Dlima_gem.c21 int lima_heap_alloc(struct lima_bo *bo, struct lima_vm *vm) in lima_heap_alloc() argument
24 struct address_space *mapping = bo->base.base.filp->f_mapping; in lima_heap_alloc()
25 struct device *dev = bo->base.base.dev->dev; in lima_heap_alloc()
26 size_t old_size = bo->heap_size; in lima_heap_alloc()
27 size_t new_size = bo->heap_size ? bo->heap_size * 2 : in lima_heap_alloc()
32 if (bo->heap_size >= bo->base.base.size) in lima_heap_alloc()
35 new_size = min(new_size, bo->base.base.size); in lima_heap_alloc()
37 mutex_lock(&bo->base.pages_lock); in lima_heap_alloc()
39 if (bo->base.pages) { in lima_heap_alloc()
40 pages = bo->base.pages; in lima_heap_alloc()
[all …]
/linux-6.1.9/drivers/gpu/drm/panfrost/
Dpanfrost_gem.c19 struct panfrost_gem_object *bo = to_panfrost_bo(obj); in panfrost_gem_free_object() local
29 list_del_init(&bo->base.madv_list); in panfrost_gem_free_object()
36 WARN_ON_ONCE(!list_empty(&bo->mappings.list)); in panfrost_gem_free_object()
38 if (bo->sgts) { in panfrost_gem_free_object()
40 int n_sgt = bo->base.base.size / SZ_2M; in panfrost_gem_free_object()
43 if (bo->sgts[i].sgl) { in panfrost_gem_free_object()
44 dma_unmap_sgtable(pfdev->dev, &bo->sgts[i], in panfrost_gem_free_object()
46 sg_free_table(&bo->sgts[i]); in panfrost_gem_free_object()
49 kvfree(bo->sgts); in panfrost_gem_free_object()
52 drm_gem_shmem_free(&bo->base); in panfrost_gem_free_object()
[all …]
/linux-6.1.9/drivers/gpu/drm/v3d/
Dv3d_bo.c34 struct v3d_bo *bo = to_v3d_bo(obj); in v3d_free_object() local
36 v3d_mmu_remove_ptes(bo); in v3d_free_object()
44 drm_mm_remove_node(&bo->node); in v3d_free_object()
48 bo->base.pages_mark_dirty_on_put = true; in v3d_free_object()
50 drm_gem_shmem_free(&bo->base); in v3d_free_object()
70 struct v3d_bo *bo; in v3d_create_object() local
76 bo = kzalloc(sizeof(*bo), GFP_KERNEL); in v3d_create_object()
77 if (!bo) in v3d_create_object()
79 obj = &bo->base.base; in v3d_create_object()
82 bo->base.map_wc = true; in v3d_create_object()
[all …]
/linux-6.1.9/drivers/staging/media/atomisp/include/hmm/
Dhmm_bo.h41 #define check_bo_status_yes_goto(bo, _status, label) \ argument
42 var_not_equal_goto((bo->status & (_status)), (_status), \
47 #define check_bo_status_no_goto(bo, _status, label) \ argument
48 var_equal_goto((bo->status & (_status)), (_status), \
62 #define check_bo_null_return(bo, exp) \ argument
63 check_null_return(bo, exp, "NULL hmm buffer object.\n")
65 #define check_bo_null_return_void(bo) \ argument
66 check_null_return_void(bo, "NULL hmm buffer object.\n")
145 void hmm_bo_release(struct hmm_buffer_object *bo);
164 void hmm_bo_ref(struct hmm_buffer_object *bo);
[all …]
/linux-6.1.9/drivers/media/platform/nvidia/tegra-vde/
Dvde.c60 struct tegra_vde_bo *bo; in tegra_vde_alloc_bo() local
63 bo = kzalloc(sizeof(*bo), GFP_KERNEL); in tegra_vde_alloc_bo()
64 if (!bo) in tegra_vde_alloc_bo()
67 bo->vde = vde; in tegra_vde_alloc_bo()
68 bo->size = size; in tegra_vde_alloc_bo()
69 bo->dma_dir = dma_dir; in tegra_vde_alloc_bo()
70 bo->dma_attrs = DMA_ATTR_WRITE_COMBINE | in tegra_vde_alloc_bo()
74 bo->dma_attrs |= DMA_ATTR_FORCE_CONTIGUOUS; in tegra_vde_alloc_bo()
76 bo->dma_cookie = dma_alloc_attrs(dev, bo->size, &bo->dma_handle, in tegra_vde_alloc_bo()
77 GFP_KERNEL, bo->dma_attrs); in tegra_vde_alloc_bo()
[all …]

12345678910>>...12