/linux-5.19.10/drivers/gpu/drm/i915/ |
D | i915_gem_gtt.c | 47 GEM_BUG_ON(obj->mm.pages == pages); in i915_gem_gtt_prepare_pages() 105 GEM_BUG_ON(!size); in i915_gem_gtt_reserve() 106 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); in i915_gem_gtt_reserve() 107 GEM_BUG_ON(!IS_ALIGNED(offset, I915_GTT_MIN_ALIGNMENT)); in i915_gem_gtt_reserve() 108 GEM_BUG_ON(range_overflows(offset, size, vm->total)); in i915_gem_gtt_reserve() 109 GEM_BUG_ON(vm == &to_gt(vm->i915)->ggtt->alias->vm); in i915_gem_gtt_reserve() 110 GEM_BUG_ON(drm_mm_node_allocated(node)); in i915_gem_gtt_reserve() 134 GEM_BUG_ON(range_overflows(start, len, end)); in random_offset() 135 GEM_BUG_ON(round_up(start, align) > round_down(end - len, align)); in random_offset() 202 GEM_BUG_ON(!size); in i915_gem_gtt_insert() [all …]
|
D | i915_vma.c | 120 GEM_BUG_ON(vm == &vm->gt->ggtt->alias->vm); in vma_create() 147 GEM_BUG_ON(range_overflows_t(u64, in vma_create() 153 GEM_BUG_ON(vma->size > obj->base.size); in vma_create() 166 GEM_BUG_ON(!IS_ALIGNED(vma->size, I915_GTT_PAGE_SIZE)); in vma_create() 189 GEM_BUG_ON(!IS_ALIGNED(vma->fence_size, I915_GTT_MIN_ALIGNMENT)); in vma_create() 194 GEM_BUG_ON(!is_power_of_2(vma->fence_alignment)); in vma_create() 293 GEM_BUG_ON(view && !i915_is_ggtt_or_dpt(vm)); in i915_vma_instance() 294 GEM_BUG_ON(!kref_read(&vm->ref)); in i915_vma_instance() 304 GEM_BUG_ON(!IS_ERR(vma) && i915_vma_compare(vma, vm, view)); in i915_vma_instance() 439 GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); in i915_vma_bind() [all …]
|
D | i915_scheduler.c | 47 GEM_BUG_ON(rb_first_cached(&sched_engine->queue) != in assert_priolists() 54 GEM_BUG_ON(p->priority > last_prio); in assert_priolists() 136 GEM_BUG_ON(!locked); in lock_sched_engine() 151 GEM_BUG_ON(locked != sched_engine); in lock_sched_engine() 167 GEM_BUG_ON(prio == I915_PRIORITY_INVALID); in __i915_schedule() 207 GEM_BUG_ON(p == dep); /* no cycles! */ in __i915_schedule() 224 GEM_BUG_ON(!list_empty(&node->link)); in __i915_schedule() 253 GEM_BUG_ON(node_to_request(node)->engine->sched_engine != in __i915_schedule() 311 GEM_BUG_ON(!list_empty(&node->signalers_list)); in i915_sched_node_reinit() 312 GEM_BUG_ON(!list_empty(&node->waiters_list)); in i915_sched_node_reinit() [all …]
|
D | i915_active.c | 50 GEM_BUG_ON(!is_barrier(&node->base)); in barrier_to_ll() 63 GEM_BUG_ON(!is_barrier(&node->base)); in barrier_to_engine() 133 GEM_BUG_ON(i915_active_is_idle(ref)); in __active_retire() 139 GEM_BUG_ON(rcu_access_pointer(ref->excl.fence)); in __active_retire() 155 GEM_BUG_ON(ref->tree.rb_node != &ref->cache->node); in __active_retire() 172 GEM_BUG_ON(i915_active_fence_isset(&it->base)); in __active_retire() 182 GEM_BUG_ON(!atomic_read(&ref->count)); in active_work() 192 GEM_BUG_ON(!atomic_read(&ref->count)); in active_retire() 237 GEM_BUG_ON(idx == 0); /* 0 is the unordered timeline, rsvd for cache */ in __active_lookup() 271 GEM_BUG_ON(i915_active_is_idle(ref)); in __active_lookup() [all …]
|
D | i915_vma.h | 89 GEM_BUG_ON(!i915_vma_is_ggtt(vma)); in i915_vma_set_ggtt_write() 108 GEM_BUG_ON(!i915_vma_is_map_and_fenceable(vma)); in i915_vma_set_userfault() 129 GEM_BUG_ON(!i915_vma_is_ggtt(vma)); in i915_ggtt_offset() 130 GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); in i915_ggtt_offset() 131 GEM_BUG_ON(upper_32_bits(vma->node.start)); in i915_ggtt_offset() 132 GEM_BUG_ON(upper_32_bits(vma->node.start + vma->node.size - 1)); in i915_ggtt_offset() 167 GEM_BUG_ON(view && !i915_is_ggtt_or_dpt(vm)); in i915_vma_compare() 282 GEM_BUG_ON(!i915_vma_is_pinned(vma)); in __i915_vma_pin() 287 GEM_BUG_ON(!i915_vma_is_pinned(vma)); in __i915_vma_unpin() 293 GEM_BUG_ON(!drm_mm_node_allocated(&vma->node)); in i915_vma_unpin() [all …]
|
D | intel_wopcm.c | 243 GEM_BUG_ON(!wopcm_size); in intel_wopcm_init() 244 GEM_BUG_ON(wopcm->guc.base); in intel_wopcm_init() 245 GEM_BUG_ON(wopcm->guc.size); in intel_wopcm_init() 246 GEM_BUG_ON(guc_fw_size >= wopcm_size); in intel_wopcm_init() 247 GEM_BUG_ON(huc_fw_size >= wopcm_size); in intel_wopcm_init() 248 GEM_BUG_ON(ctx_rsvd + WOPCM_RESERVED_SIZE >= wopcm_size); in intel_wopcm_init() 302 GEM_BUG_ON(!wopcm->guc.base); in intel_wopcm_init() 303 GEM_BUG_ON(!wopcm->guc.size); in intel_wopcm_init()
|
D | i915_request.c | 117 GEM_BUG_ON(rq->guc_prio != GUC_PRIO_INIT && in i915_fence_release() 321 #define assert_capture_list_is_null(_rq) GEM_BUG_ON((_rq)->capture_list) 342 GEM_BUG_ON(!i915_sw_fence_signaled(&rq->submit)); in i915_request_retire() 357 GEM_BUG_ON(!list_is_first(&rq->link, in i915_request_retire() 384 GEM_BUG_ON(!llist_empty(&rq->execute_cb)); in i915_request_retire() 403 GEM_BUG_ON(!__i915_request_is_complete(rq)); in i915_request_retire_upto() 407 GEM_BUG_ON(!i915_request_completed(tmp)); in i915_request_retire_upto() 533 GEM_BUG_ON(!fatal_error(rq->fence.error)); in __i915_request_skip() 553 GEM_BUG_ON(!IS_ERR_VALUE((long)error)); in i915_request_set_error_once() 572 GEM_BUG_ON(i915_request_signaled(rq)); in i915_request_mark_eio() [all …]
|
D | i915_syncmap.c | 101 GEM_BUG_ON(p->height); in __sync_seqno() 107 GEM_BUG_ON(!p->height); in __sync_child() 120 GEM_BUG_ON(p->height); in __sync_leaf_idx() 131 GEM_BUG_ON(p->height); in __sync_leaf_prefix() 244 GEM_BUG_ON(__sync_leaf_prefix(p, id) == p->prefix); in __sync_set() 300 GEM_BUG_ON(!(p->parent->bitmap & BIT(idx))); in __sync_set() 317 GEM_BUG_ON(!p->height); in __sync_set() 334 GEM_BUG_ON(p->prefix != __sync_leaf_prefix(p, id)); in __sync_set()
|
D | i915_scatterlist.c | 35 GEM_BUG_ON(new_sg); /* Should walk exactly nents and hit the end */ in i915_sg_trim() 91 GEM_BUG_ON(!max_segment); in i915_rsgt_from_mm_node() 119 GEM_BUG_ON(!IS_ALIGNED(sg_dma_address(sg), in i915_rsgt_from_mm_node() 171 GEM_BUG_ON(list_empty(blocks)); in i915_rsgt_from_buddy_resource() 172 GEM_BUG_ON(!max_segment); in i915_rsgt_from_buddy_resource() 203 GEM_BUG_ON(!IS_ALIGNED(sg_dma_address(sg), in i915_rsgt_from_buddy_resource()
|
/linux-5.19.10/drivers/gpu/drm/i915/gt/ |
D | intel_ring.c | 30 GEM_BUG_ON(!atomic_read(&ring->pin_count)); in __intel_ring_pin() 151 GEM_BUG_ON(!is_power_of_2(size)); in intel_engine_create_ring() 152 GEM_BUG_ON(RING_CTL_SIZE(size) & ~RING_NR_PAGES); in intel_engine_create_ring() 202 GEM_BUG_ON(list_empty(&tl->requests)); in wait_for_space() 225 GEM_BUG_ON(ring->space < bytes); in wait_for_space() 239 GEM_BUG_ON(num_dwords & 1); in intel_ring_begin() 242 GEM_BUG_ON(total_bytes > ring->effective_size); in intel_ring_begin() 278 GEM_BUG_ON(!rq->reserved_space); in intel_ring_begin() 289 GEM_BUG_ON(need_wrap > ring->space); in intel_ring_begin() 290 GEM_BUG_ON(ring->emit + need_wrap > ring->size); in intel_ring_begin() [all …]
|
D | gen8_ppgtt.c | 123 GEM_BUG_ON(start >= end); in gen8_pd_range() 137 GEM_BUG_ON(start >= end); in gen8_pd_contains() 143 GEM_BUG_ON(start >= end); in gen8_pt_count() 210 GEM_BUG_ON(end > vm->total >> GEN8_PTE_SHIFT); in __gen8_ppgtt_clear() 216 GEM_BUG_ON(!len || len >= atomic_read(px_used(pd))); in __gen8_ppgtt_clear() 245 GEM_BUG_ON(!count || count >= atomic_read(&pt->used)); in __gen8_ppgtt_clear() 249 GEM_BUG_ON(num_ptes % 16); in __gen8_ppgtt_clear() 250 GEM_BUG_ON(pte % 16); in __gen8_ppgtt_clear() 274 GEM_BUG_ON(!IS_ALIGNED(start, BIT_ULL(GEN8_PTE_SHIFT))); in gen8_ppgtt_clear() 275 GEM_BUG_ON(!IS_ALIGNED(length, BIT_ULL(GEN8_PTE_SHIFT))); in gen8_ppgtt_clear() [all …]
|
D | intel_ring.h | 51 GEM_BUG_ON((rq->ring->vaddr + rq->ring->emit) != cs); in intel_ring_advance() 52 GEM_BUG_ON(!IS_ALIGNED(rq->ring->emit, 8)); /* RING_TAIL qword align */ in intel_ring_advance() 86 GEM_BUG_ON(offset > rq->ring->size); in intel_ring_offset() 95 GEM_BUG_ON(!intel_ring_offset_valid(ring, tail)); in assert_ring_tail_valid() 112 GEM_BUG_ON(cacheline(tail) == cacheline(head) && tail < head); in assert_ring_tail_valid() 138 GEM_BUG_ON(!is_power_of_2(size)); in __intel_ring_space()
|
D | intel_timeline.c | 99 GEM_BUG_ON(timeline->hwsp_offset >= hwsp->size); in intel_timeline_init() 187 GEM_BUG_ON(!atomic_read(&tl->pin_count)); in __intel_timeline_pin() 227 GEM_BUG_ON(!atomic_read(&tl->pin_count)); in intel_timeline_reset_seqno() 282 GEM_BUG_ON(!atomic_read(&tl->active_count)); in intel_timeline_exit() 301 GEM_BUG_ON(!atomic_read(&tl->pin_count)); in timeline_advance() 302 GEM_BUG_ON(tl->seqno & tl->has_initial_breadcrumb); in timeline_advance() 322 GEM_BUG_ON(i915_seqno_passed(*tl->hwsp_seqno, *seqno)); in __intel_timeline_get_seqno() 383 GEM_BUG_ON(!atomic_read(&tl->pin_count)); in intel_timeline_unpin() 396 GEM_BUG_ON(atomic_read(&timeline->pin_count)); in __intel_timeline_free() 397 GEM_BUG_ON(!list_empty(&timeline->requests)); in __intel_timeline_free() [all …]
|
D | intel_execlists_submission.c | 202 GEM_BUG_ON(!intel_engine_is_virtual(engine)); in to_virtual_engine() 384 GEM_BUG_ON(rq_prio(rq) == I915_PRIORITY_INVALID); in __unwind_incomplete_requests() 390 GEM_BUG_ON(i915_sched_engine_is_empty(engine->sched_engine)); in __unwind_incomplete_requests() 493 GEM_BUG_ON(ce->tag <= BITS_PER_LONG); in __execlists_schedule_in() 499 GEM_BUG_ON(tag == 0 || tag >= BITS_PER_LONG); in __execlists_schedule_in() 509 GEM_BUG_ON(tag >= BITS_PER_LONG); in __execlists_schedule_in() 534 GEM_BUG_ON(!intel_engine_pm_is_awake(rq->engine)); in execlists_schedule_in() 542 GEM_BUG_ON(intel_context_inflight(ce) != rq->engine); in execlists_schedule_in() 601 GEM_BUG_ON(ce->inflight != engine); in __execlists_schedule_out() 624 GEM_BUG_ON(ccid == 0); in __execlists_schedule_out() [all …]
|
D | intel_ggtt_fencing.c | 79 GEM_BUG_ON(!IS_ALIGNED(stride, 128)); in i965_write_fence_reg() 126 GEM_BUG_ON(!is_power_of_2(stride)); in i915_write_fence_reg() 210 GEM_BUG_ON(!i915_gem_object_get_stride(vma->obj) || in fence_update() 246 GEM_BUG_ON(old->fence != fence); in fence_update() 266 GEM_BUG_ON(vma); in fence_update() 298 GEM_BUG_ON(fence->vma != vma); in i915_vma_revoke_fence() 299 GEM_BUG_ON(!i915_active_is_idle(&fence->active)); in i915_vma_revoke_fence() 300 GEM_BUG_ON(atomic_read(&fence->pin_count)); in i915_vma_revoke_fence() 332 GEM_BUG_ON(fence->vma && fence->vma->fence != fence); in fence_find() 371 GEM_BUG_ON(fence->vma != vma); in __i915_vma_pin_fence() [all …]
|
D | intel_engine_pm.c | 55 GEM_BUG_ON(test_bit(CONTEXT_VALID_BIT, &ce->flags)); in __engine_unpark() 71 GEM_BUG_ON(ce->timeline->seqno != in __engine_unpark() 148 GEM_BUG_ON(!intel_context_is_barrier(ce)); in switch_to_kernel_context() 149 GEM_BUG_ON(ce->timeline->hwsp_ggtt != engine->status_page.vma); in switch_to_kernel_context() 184 GEM_BUG_ON(atomic_read(&ce->timeline->active_count) < 0); in switch_to_kernel_context() 257 GEM_BUG_ON(engine->sched_engine->queue_priority_hint != INT_MIN); in __engine_park()
|
D | intel_context.c | 271 GEM_BUG_ON(!intel_context_is_pinned(ce)); /* no overflow! */ in __intel_context_do_pin_ww() 378 GEM_BUG_ON(!engine->cops); in intel_context_init() 379 GEM_BUG_ON(!engine->gt->vm); in intel_context_init() 473 GEM_BUG_ON(rq->context == ce); in intel_context_prepare_remote_request() 489 GEM_BUG_ON(i915_active_is_idle(&ce->active)); in intel_context_prepare_remote_request() 537 GEM_BUG_ON(!intel_engine_uses_guc(ce->engine)); in intel_context_find_active_request() 567 GEM_BUG_ON(intel_context_is_pinned(parent)); in intel_context_bind_parent_child() 568 GEM_BUG_ON(intel_context_is_child(parent)); in intel_context_bind_parent_child() 569 GEM_BUG_ON(intel_context_is_pinned(child)); in intel_context_bind_parent_child() 570 GEM_BUG_ON(intel_context_is_child(child)); in intel_context_bind_parent_child() [all …]
|
/linux-5.19.10/drivers/gpu/drm/i915/gem/ |
D | i915_gem_object.c | 90 GEM_BUG_ON(flags & ~I915_BO_ALLOC_FLAGS); in i915_gem_object_init() 213 GEM_BUG_ON(vma->obj != obj); in i915_gem_close_object() 214 GEM_BUG_ON(!atomic_read(&vma->open_count)); in i915_gem_close_object() 233 GEM_BUG_ON(!atomic_read(&i915->mm.free_count)); in __i915_gem_free_object_rcu() 280 GEM_BUG_ON(vma->obj != obj); in __i915_gem_object_pages_fini() 294 GEM_BUG_ON(i915_gem_object_has_pages(obj)); in __i915_gem_object_pages_fini() 301 GEM_BUG_ON(!list_empty(&obj->lut_list)); in __i915_gem_free_object() 364 GEM_BUG_ON(i915_gem_object_is_framebuffer(obj)); in i915_gem_free_object() 463 GEM_BUG_ON(offset >= obj->base.size); in i915_gem_object_read_from_page() 464 GEM_BUG_ON(offset_in_page(offset) > PAGE_SIZE - size); in i915_gem_object_read_from_page() [all …]
|
D | i915_gem_region.c | 49 GEM_BUG_ON(flags & ~I915_BO_ALLOC_FLAGS); in __i915_gem_object_create_region() 64 GEM_BUG_ON(overflows_type(default_page_size, u32)); in __i915_gem_object_create_region() 65 GEM_BUG_ON(!is_power_of_2_u64(default_page_size)); in __i915_gem_object_create_region() 66 GEM_BUG_ON(default_page_size < PAGE_SIZE); in __i915_gem_object_create_region() 73 GEM_BUG_ON(!size); in __i915_gem_object_create_region() 74 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_MIN_ALIGNMENT)); in __i915_gem_object_create_region() 122 GEM_BUG_ON(offset == I915_BO_INVALID_OFFSET); in i915_gem_object_create_region_at()
|
D | i915_gem_pages.c | 48 GEM_BUG_ON(!sg_page_sizes); in __i915_gem_object_set_pages() 64 GEM_BUG_ON(!HAS_PAGE_SIZES(i915, obj->mm.page_sizes.sg)); in __i915_gem_object_set_pages() 70 GEM_BUG_ON(i915_gem_object_has_tiling_quirk(obj)); in __i915_gem_object_set_pages() 72 GEM_BUG_ON(!list_empty(&obj->mm.link)); in __i915_gem_object_set_pages() 112 GEM_BUG_ON(!err && !i915_gem_object_has_pages(obj)); in ____i915_gem_object_get_pages() 133 GEM_BUG_ON(i915_gem_object_has_pinned_pages(obj)); in __i915_gem_object_get_pages() 336 GEM_BUG_ON(type != I915_MAP_WC); in i915_gem_object_map_pfn() 378 GEM_BUG_ON(i915_gem_object_has_pinned_pages(obj)); in i915_gem_object_pin_map() 389 GEM_BUG_ON(!i915_gem_object_has_pages(obj)); in i915_gem_object_pin_map() 476 GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj)); in __i915_gem_object_flush_map() [all …]
|
D | i915_gem_tiling.c | 60 GEM_BUG_ON(!size); in i915_gem_fence_size() 65 GEM_BUG_ON(!stride); in i915_gem_fence_size() 69 GEM_BUG_ON(!IS_ALIGNED(stride, I965_FENCE_PAGE)); in i915_gem_fence_size() 98 GEM_BUG_ON(!size); in i915_gem_fence_alignment() 199 GEM_BUG_ON(vma->vm != &ggtt->vm); in i915_gem_object_fence_prepare() 241 GEM_BUG_ON(!i915_tiling_ok(obj, tiling, stride)); in i915_gem_object_set_tiling() 242 GEM_BUG_ON(!stride ^ (tiling == I915_TILING_NONE)); in i915_gem_object_set_tiling() 283 GEM_BUG_ON(!i915_gem_object_has_tiling_quirk(obj)); in i915_gem_object_set_tiling() 288 GEM_BUG_ON(i915_gem_object_has_tiling_quirk(obj)); in i915_gem_object_set_tiling()
|
/linux-5.19.10/drivers/gpu/drm/i915/gt/uc/ |
D | intel_uc.c | 89 GEM_BUG_ON(intel_uc_wants_guc(uc)); in __confirm_options() 90 GEM_BUG_ON(intel_uc_wants_guc_submission(uc)); in __confirm_options() 91 GEM_BUG_ON(intel_uc_wants_huc(uc)); in __confirm_options() 92 GEM_BUG_ON(intel_uc_wants_guc_slpc(uc)); in __confirm_options() 211 GEM_BUG_ON(!intel_guc_ct_enabled(&guc->ct)); in guc_handle_mmio_msg() 227 GEM_BUG_ON(intel_guc_ct_enabled(&guc->ct)); in guc_enable_communication() 283 GEM_BUG_ON(!intel_uc_wants_guc(uc)); in __uc_fetch_firmwares() 314 GEM_BUG_ON(!intel_uc_wants_guc(uc)); in __uc_init() 350 GEM_BUG_ON(!intel_uc_supports_guc(uc)); in __uc_sanitize() 374 GEM_BUG_ON(!intel_uc_supports_guc(uc)); in uc_init_wopcm() [all …]
|
D | intel_guc_submission.c | 334 GEM_BUG_ON(!context_blocked(ce)); /* Overflow check */ in incr_context_blocked() 341 GEM_BUG_ON(!context_blocked(ce)); /* Underflow check */ in decr_context_blocked() 355 GEM_BUG_ON(ce->guc_state.number_committed_requests < 0); in incr_context_committed_requests() 362 GEM_BUG_ON(ce->guc_state.number_committed_requests < 0); in decr_context_committed_requests() 433 GEM_BUG_ON(!ce->parallel.guc.parent_page); in __get_parent_scratch_offset() 502 GEM_BUG_ON(id >= GUC_MAX_CONTEXT_ID); in __get_context() 514 GEM_BUG_ON(index >= GUC_MAX_CONTEXT_ID); in __get_lrc_desc_v69() 610 GEM_BUG_ON(g2h_len_dw && !loop); in guc_submission_send_busy_loop() 628 GEM_BUG_ON(timeout < 0); in intel_guc_wait_for_pending_msg() 693 GEM_BUG_ON(!atomic_read(&ce->guc_id.ref)); in __guc_add_request() [all …]
|
D | intel_guc_ct.c | 138 GEM_BUG_ON(size_in_bytes % 4); in guc_ct_buffer_init() 158 GEM_BUG_ON(control != GUC_CTB_CONTROL_DISABLE && control != GUC_CTB_CONTROL_ENABLE); in guc_action_control_ctb() 233 GEM_BUG_ON(ct->vma); in intel_guc_ct_init() 278 GEM_BUG_ON(ct->enabled); in intel_guc_ct_fini() 298 GEM_BUG_ON(ct->enabled); in intel_guc_ct_enable() 301 GEM_BUG_ON(!ct->vma); in intel_guc_ct_enable() 302 GEM_BUG_ON(!i915_gem_object_has_pinned_pages(ct->vma->obj)); in intel_guc_ct_enable() 307 GEM_BUG_ON(blob != ct->ctbs.send.desc); in intel_guc_ct_enable() 353 GEM_BUG_ON(!ct->enabled); in intel_guc_ct_disable() 386 GEM_BUG_ON(tail > size); in ct_write() [all …]
|
D | intel_uc_fw.c | 25 GEM_BUG_ON(type != INTEL_UC_FW_TYPE_HUC); in ____uc_fw_to_gt() 31 GEM_BUG_ON(uc_fw->status == INTEL_UC_FIRMWARE_UNINITIALIZED); in __uc_fw_to_gt() 182 GEM_BUG_ON(uc_fw->type >= ARRAY_SIZE(blobs_all)); in __uc_fw_auto_select() 285 GEM_BUG_ON(uc_fw->status); in intel_uc_fw_init_early() 286 GEM_BUG_ON(uc_fw->path); in intel_uc_fw_init_early() 356 GEM_BUG_ON(!i915->wopcm.size); in intel_uc_fw_fetch() 357 GEM_BUG_ON(!intel_uc_fw_is_enabled(uc_fw)); in intel_uc_fw_fetch() 496 GEM_BUG_ON(!drm_mm_node_allocated(node)); in uc_fw_ggtt_offset() 497 GEM_BUG_ON(upper_32_bits(node->start)); in uc_fw_ggtt_offset() 498 GEM_BUG_ON(upper_32_bits(node->start + node->size - 1)); in uc_fw_ggtt_offset() [all …]
|