/linux-6.1.9/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_uvd.c | 155 if (adev->uvd.address_64_bit) in amdgpu_uvd_create_msg_bo_helper() 192 INIT_DELAYED_WORK(&adev->uvd.idle_work, amdgpu_uvd_idle_work_handler); in amdgpu_uvd_sw_init() 263 r = request_firmware(&adev->uvd.fw, fw_name, adev->dev); in amdgpu_uvd_sw_init() 270 r = amdgpu_ucode_validate(adev->uvd.fw); in amdgpu_uvd_sw_init() 274 release_firmware(adev->uvd.fw); in amdgpu_uvd_sw_init() 275 adev->uvd.fw = NULL; in amdgpu_uvd_sw_init() 280 adev->uvd.max_handles = AMDGPU_DEFAULT_UVD_HANDLES; in amdgpu_uvd_sw_init() 282 hdr = (const struct common_firmware_header *)adev->uvd.fw->data; in amdgpu_uvd_sw_init() 301 adev->uvd.max_handles = AMDGPU_MAX_UVD_HANDLES; in amdgpu_uvd_sw_init() 303 adev->uvd.fw_version = ((version_major << 24) | (version_minor << 16) | in amdgpu_uvd_sw_init() [all …]
|
D | uvd_v7_0.c | 89 if (ring == &adev->uvd.inst[ring->me].ring_enc[0]) in uvd_v7_0_enc_ring_get_rptr() 123 if (ring == &adev->uvd.inst[ring->me].ring_enc[0]) in uvd_v7_0_enc_ring_get_wptr() 161 if (ring == &adev->uvd.inst[ring->me].ring_enc[0]) in uvd_v7_0_enc_ring_set_wptr() 342 struct amdgpu_bo *bo = ring->adev->uvd.ib_bo; in uvd_v7_0_enc_ring_test_ib() 372 adev->uvd.num_uvd_inst = UVD7_MAX_HW_INSTANCES_VEGA20; in uvd_v7_0_early_init() 373 for (i = 0; i < adev->uvd.num_uvd_inst; i++) { in uvd_v7_0_early_init() 376 adev->uvd.harvest_config |= 1 << i; in uvd_v7_0_early_init() 379 if (adev->uvd.harvest_config == (AMDGPU_UVD_HARVEST_UVD0 | in uvd_v7_0_early_init() 384 adev->uvd.num_uvd_inst = 1; in uvd_v7_0_early_init() 388 adev->uvd.num_enc_rings = 1; in uvd_v7_0_early_init() [all …]
|
D | uvd_v6_0.c | 67 (!adev->uvd.fw_version || adev->uvd.fw_version >= FW_1_130_16)); in uvd_v6_0_enc_support() 95 if (ring == &adev->uvd.inst->ring_enc[0]) in uvd_v6_0_enc_ring_get_rptr() 125 if (ring == &adev->uvd.inst->ring_enc[0]) in uvd_v6_0_enc_ring_get_wptr() 156 if (ring == &adev->uvd.inst->ring_enc[0]) in uvd_v6_0_enc_ring_set_wptr() 335 struct amdgpu_bo *bo = ring->adev->uvd.ib_bo; in uvd_v6_0_enc_ring_test_ib() 360 adev->uvd.num_uvd_inst = 1; in uvd_v6_0_early_init() 369 adev->uvd.num_enc_rings = 2; in uvd_v6_0_early_init() 385 …id(adev, AMDGPU_IRQ_CLIENTID_LEGACY, VISLANDS30_IV_SRCID_UVD_SYSTEM_MESSAGE, &adev->uvd.inst->irq); in uvd_v6_0_sw_init() 391 for (i = 0; i < adev->uvd.num_enc_rings; ++i) { in uvd_v6_0_sw_init() 392 …(adev, AMDGPU_IRQ_CLIENTID_LEGACY, i + VISLANDS30_IV_SRCID_UVD_ENC_GEN_PURP, &adev->uvd.inst->irq); in uvd_v6_0_sw_init() [all …]
|
D | uvd_v3_1.c | 203 adev->uvd.inst->ring.funcs = &uvd_v3_1_ring_funcs; in uvd_v3_1_set_ring_funcs() 246 addr = (adev->uvd.inst->gpu_addr + AMDGPU_UVD_FIRMWARE_OFFSET) >> 3; in uvd_v3_1_mc_resume() 258 (AMDGPU_UVD_SESSION_SIZE * adev->uvd.max_handles)) >> 3; in uvd_v3_1_mc_resume() 263 addr = (adev->uvd.inst->gpu_addr >> 28) & 0xF; in uvd_v3_1_mc_resume() 267 addr = (adev->uvd.inst->gpu_addr >> 32) & 0xFF; in uvd_v3_1_mc_resume() 285 uint32_t keysel = adev->uvd.keyselect; in uvd_v3_1_fw_validate() 322 struct amdgpu_ring *ring = &adev->uvd.inst->ring; in uvd_v3_1_start() 517 amdgpu_fence_process(&adev->uvd.inst->ring); in uvd_v3_1_process_interrupt() 529 adev->uvd.inst->irq.num_types = 1; in uvd_v3_1_set_irq_funcs() 530 adev->uvd.inst->irq.funcs = &uvd_v3_1_irq_funcs; in uvd_v3_1_set_irq_funcs() [all …]
|
D | uvd_v4_2.c | 96 adev->uvd.num_uvd_inst = 1; in uvd_v4_2_early_init() 111 r = amdgpu_irq_add_id(adev, AMDGPU_IRQ_CLIENTID_LEGACY, 124, &adev->uvd.inst->irq); in uvd_v4_2_sw_init() 119 ring = &adev->uvd.inst->ring; in uvd_v4_2_sw_init() 121 r = amdgpu_ring_init(adev, ring, 512, &adev->uvd.inst->irq, 0, in uvd_v4_2_sw_init() 159 struct amdgpu_ring *ring = &adev->uvd.inst->ring; in uvd_v4_2_hw_init() 215 cancel_delayed_work_sync(&adev->uvd.idle_work); in uvd_v4_2_hw_fini() 239 cancel_delayed_work_sync(&adev->uvd.idle_work); in uvd_v4_2_suspend() 280 struct amdgpu_ring *ring = &adev->uvd.inst->ring; in uvd_v4_2_start() 575 addr = (adev->uvd.inst->gpu_addr + AMDGPU_UVD_FIRMWARE_OFFSET) >> 3; in uvd_v4_2_mc_resume() 587 (AMDGPU_UVD_SESSION_SIZE * adev->uvd.max_handles)) >> 3; in uvd_v4_2_mc_resume() [all …]
|
D | uvd_v5_0.c | 94 adev->uvd.num_uvd_inst = 1; in uvd_v5_0_early_init() 109 …id(adev, AMDGPU_IRQ_CLIENTID_LEGACY, VISLANDS30_IV_SRCID_UVD_SYSTEM_MESSAGE, &adev->uvd.inst->irq); in uvd_v5_0_sw_init() 117 ring = &adev->uvd.inst->ring; in uvd_v5_0_sw_init() 119 r = amdgpu_ring_init(adev, ring, 512, &adev->uvd.inst->irq, 0, in uvd_v5_0_sw_init() 155 struct amdgpu_ring *ring = &adev->uvd.inst->ring; in uvd_v5_0_hw_init() 213 cancel_delayed_work_sync(&adev->uvd.idle_work); in uvd_v5_0_hw_fini() 237 cancel_delayed_work_sync(&adev->uvd.idle_work); in uvd_v5_0_suspend() 283 lower_32_bits(adev->uvd.inst->gpu_addr)); in uvd_v5_0_mc_resume() 285 upper_32_bits(adev->uvd.inst->gpu_addr)); in uvd_v5_0_mc_resume() 299 (AMDGPU_UVD_SESSION_SIZE * adev->uvd.max_handles); in uvd_v5_0_mc_resume() [all …]
|
D | amdgpu_kms.c | 211 fw_info->ver = adev->uvd.fw_version; in amdgpu_firmware_info() 399 for (i = 0; i < adev->uvd.num_uvd_inst; i++) { in amdgpu_hw_ip_info() 400 if (adev->uvd.harvest_config & (1 << i)) in amdgpu_hw_ip_info() 403 if (adev->uvd.inst[i].ring.sched.ready) in amdgpu_hw_ip_info() 419 for (i = 0; i < adev->uvd.num_uvd_inst; i++) { in amdgpu_hw_ip_info() 420 if (adev->uvd.harvest_config & (1 << i)) in amdgpu_hw_ip_info() 423 for (j = 0; j < adev->uvd.num_enc_rings; j++) in amdgpu_hw_ip_info() 424 if (adev->uvd.inst[i].ring_enc[j].sched.ready) in amdgpu_hw_ip_info() 433 if (adev->uvd.harvest_config & (1 << i)) in amdgpu_hw_ip_info() 445 if (adev->uvd.harvest_config & (1 << i)) in amdgpu_hw_ip_info() [all …]
|
D | amdgpu_uvd.h | 37 …(AMDGPU_GPU_PAGE_ALIGN(le32_to_cpu(((const struct common_firmware_header *)(adev)->uvd.fw->data)->…
|
D | amdgpu_fence.c | 434 index = ALIGN(adev->uvd.fw->size, 8); in amdgpu_fence_driver_start_ring() 435 ring->fence_drv.cpu_addr = adev->uvd.inst[ring->me].cpu_addr + index; in amdgpu_fence_driver_start_ring() 436 ring->fence_drv.gpu_addr = adev->uvd.inst[ring->me].gpu_addr + index; in amdgpu_fence_driver_start_ring()
|
D | amdgpu_virt.c | 539 POPULATE_UCODE_INFO(vf2pf_info, AMD_SRIOV_UCODE_ID_UVD, adev->uvd.fw_version); in amdgpu_virt_populate_vf2pf_ucode_info()
|
D | amdgpu.h | 925 struct amdgpu_uvd uvd; member
|
D | amdgpu_ucode.c | 690 FW_VERSION_ATTR(uvd_fw_version, 0444, uvd.fw_version);
|
/linux-6.1.9/drivers/gpu/drm/radeon/ |
D | radeon_uvd.c | 72 INIT_DELAYED_WORK(&rdev->uvd.idle_work, radeon_uvd_idle_work_handler); in radeon_uvd_init() 137 rdev->uvd.fw_header_present = false; in radeon_uvd_init() 138 rdev->uvd.max_handles = RADEON_DEFAULT_UVD_HANDLES; in radeon_uvd_init() 153 rdev->uvd.fw_header_present = true; in radeon_uvd_init() 168 rdev->uvd.max_handles = RADEON_MAX_UVD_HANDLES; in radeon_uvd_init() 188 RADEON_UVD_SESSION_SIZE * rdev->uvd.max_handles; in radeon_uvd_init() 191 NULL, &rdev->uvd.vcpu_bo); in radeon_uvd_init() 197 r = radeon_bo_reserve(rdev->uvd.vcpu_bo, false); in radeon_uvd_init() 199 radeon_bo_unref(&rdev->uvd.vcpu_bo); in radeon_uvd_init() 204 r = radeon_bo_pin(rdev->uvd.vcpu_bo, RADEON_GEM_DOMAIN_VRAM, in radeon_uvd_init() [all …]
|
D | uvd_v4_2.c | 46 if (rdev->uvd.fw_header_present) in uvd_v4_2_resume() 47 addr = (rdev->uvd.gpu_addr + 0x200) >> 3; in uvd_v4_2_resume() 49 addr = rdev->uvd.gpu_addr >> 3; in uvd_v4_2_resume() 62 (RADEON_UVD_SESSION_SIZE * rdev->uvd.max_handles)) >> 3; in uvd_v4_2_resume() 67 addr = (rdev->uvd.gpu_addr >> 28) & 0xF; in uvd_v4_2_resume() 71 addr = (rdev->uvd.gpu_addr >> 32) & 0xFF; in uvd_v4_2_resume() 74 if (rdev->uvd.fw_header_present) in uvd_v4_2_resume() 75 WREG32(UVD_GP_SCRATCH4, rdev->uvd.max_handles); in uvd_v4_2_resume()
|
D | uvd_v2_2.c | 113 addr = rdev->uvd.gpu_addr >> 3; in uvd_v2_2_resume() 125 (RADEON_UVD_SESSION_SIZE * rdev->uvd.max_handles)) >> 3; in uvd_v2_2_resume() 130 addr = (rdev->uvd.gpu_addr >> 28) & 0xF; in uvd_v2_2_resume() 134 addr = (rdev->uvd.gpu_addr >> 32) & 0xFF; in uvd_v2_2_resume()
|
D | uvd_v1_0.c | 121 addr = (rdev->uvd.gpu_addr >> 3) + 16; in uvd_v1_0_resume() 133 (RADEON_UVD_SESSION_SIZE * rdev->uvd.max_handles)) >> 3; in uvd_v1_0_resume() 138 addr = (rdev->uvd.gpu_addr >> 28) & 0xF; in uvd_v1_0_resume() 142 addr = (rdev->uvd.gpu_addr >> 32) & 0xFF; in uvd_v1_0_resume() 145 WREG32(UVD_FW_START, *((uint32_t*)rdev->uvd.cpu_addr)); in uvd_v1_0_resume()
|
D | radeon_drv.c | 265 MODULE_PARM_DESC(uvd, "uvd enable/disable uvd support (1 = enable, 0 = disable)"); 266 module_param_named(uvd, radeon_uvd, int, 0444);
|
D | radeon_fence.c | 833 rdev->fence_drv[ring].cpu_addr = rdev->uvd.cpu_addr + index; in radeon_fence_driver_start_ring() 834 rdev->fence_drv[ring].gpu_addr = rdev->uvd.gpu_addr + index; in radeon_fence_driver_start_ring()
|
D | radeon.h | 2418 struct radeon_uvd uvd; member
|
/linux-6.1.9/drivers/gpu/drm/amd/pm/powerplay/hwmgr/ |
D | smu10_hwmgr.h | 112 uint32_t uvd : 1; member
|
D | smu8_hwmgr.h | 135 uint32_t uvd : 1; member
|
D | smu8_hwmgr.c | 1984 adev->uvd.decode_image_width >= WIDTH_4K) in smu8_dpm_powergate_uvd()
|