/linux-6.1.9/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_flat_memory.c | 316 static void kfd_init_apertures_vi(struct kfd_process_device *pdd, uint8_t id) in kfd_init_apertures_vi() argument 322 pdd->lds_base = MAKE_LDS_APP_BASE_VI(); in kfd_init_apertures_vi() 323 pdd->lds_limit = MAKE_LDS_APP_LIMIT(pdd->lds_base); in kfd_init_apertures_vi() 325 if (!pdd->dev->use_iommu_v2) { in kfd_init_apertures_vi() 330 pdd->gpuvm_base = SVM_USER_BASE; in kfd_init_apertures_vi() 331 pdd->gpuvm_limit = in kfd_init_apertures_vi() 332 pdd->dev->shared_resources.gpuvm_size - 1; in kfd_init_apertures_vi() 337 pdd->gpuvm_base = MAKE_GPUVM_APP_BASE_VI(id + 1); in kfd_init_apertures_vi() 338 pdd->gpuvm_limit = MAKE_GPUVM_APP_LIMIT(pdd->gpuvm_base, in kfd_init_apertures_vi() 339 pdd->dev->shared_resources.gpuvm_size); in kfd_init_apertures_vi() [all …]
|
D | kfd_process.c | 77 static void kfd_process_device_destroy_cwsr_dgpu(struct kfd_process_device *pdd); 90 struct kfd_process_device *pdd; member 104 struct kfd_process_device *pdd; in kfd_sdma_activity_worker() local 117 pdd = workarea->pdd; in kfd_sdma_activity_worker() 118 if (!pdd) in kfd_sdma_activity_worker() 120 dqm = pdd->dev->dqm; in kfd_sdma_activity_worker() 121 qpd = &pdd->qpd; in kfd_sdma_activity_worker() 176 workarea->sdma_activity_counter = pdd->sdma_past_activity_counter; in kfd_sdma_activity_worker() 186 mm = get_task_mm(pdd->process->lead_thread); in kfd_sdma_activity_worker() 213 workarea->sdma_activity_counter += pdd->sdma_past_activity_counter; in kfd_sdma_activity_worker() [all …]
|
D | kfd_chardev.c | 70 struct kfd_process_device *pdd; in kfd_lock_pdd_by_id() local 73 pdd = kfd_process_device_data_by_id(p, gpu_id); in kfd_lock_pdd_by_id() 75 if (pdd) in kfd_lock_pdd_by_id() 76 return pdd; in kfd_lock_pdd_by_id() 82 static inline void kfd_unlock_pdd(struct kfd_process_device *pdd) in kfd_unlock_pdd() argument 84 mutex_unlock(&pdd->process->mutex); in kfd_unlock_pdd() 299 struct kfd_process_device *pdd; in kfd_ioctl_create_queue() local 316 pdd = kfd_process_device_data_by_id(p, args->gpu_id); in kfd_ioctl_create_queue() 317 if (!pdd) { in kfd_ioctl_create_queue() 322 dev = pdd->dev; in kfd_ioctl_create_queue() [all …]
|
D | kfd_iommu.c | 110 int kfd_iommu_bind_process_to_device(struct kfd_process_device *pdd) in kfd_iommu_bind_process_to_device() argument 112 struct kfd_dev *dev = pdd->dev; in kfd_iommu_bind_process_to_device() 113 struct kfd_process *p = pdd->process; in kfd_iommu_bind_process_to_device() 116 if (!dev->use_iommu_v2 || pdd->bound == PDD_BOUND) in kfd_iommu_bind_process_to_device() 119 if (unlikely(pdd->bound == PDD_BOUND_SUSPENDED)) { in kfd_iommu_bind_process_to_device() 126 pdd->bound = PDD_BOUND; in kfd_iommu_bind_process_to_device() 150 struct kfd_process_device *pdd; in iommu_pasid_shutdown_callback() local 168 pdd = kfd_get_process_device_data(dev, p); in iommu_pasid_shutdown_callback() 169 if (pdd) in iommu_pasid_shutdown_callback() 173 kfd_process_dequeue_from_device(pdd); in iommu_pasid_shutdown_callback() [all …]
|
D | kfd_process_queue_manager.c | 82 void kfd_process_dequeue_from_device(struct kfd_process_device *pdd) in kfd_process_dequeue_from_device() argument 84 struct kfd_dev *dev = pdd->dev; in kfd_process_dequeue_from_device() 86 if (pdd->already_dequeued) in kfd_process_dequeue_from_device() 89 dev->dqm->ops.process_termination(dev->dqm, &pdd->qpd); in kfd_process_dequeue_from_device() 90 pdd->already_dequeued = true; in kfd_process_dequeue_from_device() 98 struct kfd_process_device *pdd; in pqm_set_gws() local 113 pdd = kfd_get_process_device_data(dev, pqm->process); in pqm_set_gws() 114 if (!pdd) { in pqm_set_gws() 120 if (gws && pdd->qpd.num_gws) in pqm_set_gws() 123 if (!gws && pdd->qpd.num_gws == 0) in pqm_set_gws() [all …]
|
D | kfd_doorbell.c | 145 struct kfd_process_device *pdd; in kfd_doorbell_mmap() local 154 pdd = kfd_get_process_device_data(dev, process); in kfd_doorbell_mmap() 155 if (!pdd) in kfd_doorbell_mmap() 159 address = kfd_get_process_doorbells(pdd); in kfd_doorbell_mmap() 247 struct kfd_process_device *pdd, in kfd_get_doorbell_dw_offset_in_bar() argument 258 pdd->doorbell_index in kfd_get_doorbell_dw_offset_in_bar() 265 pdd->doorbell_index, doorbell_id); in kfd_get_doorbell_dw_offset_in_bar() 278 phys_addr_t kfd_get_process_doorbells(struct kfd_process_device *pdd) in kfd_get_process_doorbells() argument 280 if (!pdd->doorbell_index) { in kfd_get_process_doorbells() 281 int r = kfd_alloc_process_doorbells(pdd->dev, in kfd_get_process_doorbells() [all …]
|
D | kfd_device_queue_manager_v9.c | 44 static uint32_t compute_sh_mem_bases_64bit(struct kfd_process_device *pdd) in compute_sh_mem_bases_64bit() argument 46 uint32_t shared_base = pdd->lds_base >> 48; in compute_sh_mem_bases_64bit() 47 uint32_t private_base = pdd->scratch_base >> 48; in compute_sh_mem_bases_64bit() 56 struct kfd_process_device *pdd; in update_qpd_v9() local 58 pdd = qpd_to_pdd(qpd); in update_qpd_v9() 70 if (!pdd->process->xnack_enabled) in update_qpd_v9() 83 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(pdd); in update_qpd_v9()
|
D | kfd_device_queue_manager_v10.c | 43 static uint32_t compute_sh_mem_bases_64bit(struct kfd_process_device *pdd) in compute_sh_mem_bases_64bit() argument 45 uint32_t shared_base = pdd->lds_base >> 48; in compute_sh_mem_bases_64bit() 46 uint32_t private_base = pdd->scratch_base >> 48; in compute_sh_mem_bases_64bit() 55 struct kfd_process_device *pdd; in update_qpd_v10() local 57 pdd = qpd_to_pdd(qpd); in update_qpd_v10() 69 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(pdd); in update_qpd_v10()
|
D | kfd_device_queue_manager_v11.c | 42 static uint32_t compute_sh_mem_bases_64bit(struct kfd_process_device *pdd) in compute_sh_mem_bases_64bit() argument 44 uint32_t shared_base = pdd->lds_base >> 48; in compute_sh_mem_bases_64bit() 45 uint32_t private_base = pdd->scratch_base >> 48; in compute_sh_mem_bases_64bit() 54 struct kfd_process_device *pdd; in update_qpd_v11() local 56 pdd = qpd_to_pdd(qpd); in update_qpd_v11() 69 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(pdd); in update_qpd_v11()
|
D | kfd_device_queue_manager_cik.c | 120 struct kfd_process_device *pdd; in update_qpd_cik() local 123 pdd = qpd_to_pdd(qpd); in update_qpd_cik() 136 temp = get_sh_mem_bases_32(pdd); in update_qpd_cik() 140 temp = get_sh_mem_bases_nybble_64(pdd); in update_qpd_cik() 154 struct kfd_process_device *pdd; in update_qpd_cik_hawaii() local 157 pdd = qpd_to_pdd(qpd); in update_qpd_cik_hawaii() 172 temp = get_sh_mem_bases_nybble_64(pdd); in update_qpd_cik_hawaii()
|
D | kfd_device_queue_manager_vi.c | 157 struct kfd_process_device *pdd; in update_qpd_vi() local 160 pdd = qpd_to_pdd(qpd); in update_qpd_vi() 176 temp = get_sh_mem_bases_32(pdd); in update_qpd_vi() 181 temp = get_sh_mem_bases_nybble_64(pdd); in update_qpd_vi() 198 struct kfd_process_device *pdd; in update_qpd_vi_tonga() local 201 pdd = qpd_to_pdd(qpd); in update_qpd_vi_tonga() 220 temp = get_sh_mem_bases_nybble_64(pdd); in update_qpd_vi_tonga()
|
D | kfd_device_queue_manager.c | 177 struct kfd_process_device *pdd = qpd_to_pdd(qpd); in add_queue_mes() local 192 queue_input.process_context_addr = pdd->proc_ctx_gpu_addr; in add_queue_mes() 671 struct kfd_process_device *pdd; in dbgdev_wave_reset_wavefronts() local 707 pdd = kfd_get_process_device_data(dev, p); in dbgdev_wave_reset_wavefronts() 708 if (!pdd) in dbgdev_wave_reset_wavefronts() 793 struct kfd_process_device *pdd = qpd_to_pdd(qpd); in destroy_queue_nocpsch() local 810 pdd->sdma_past_activity_counter += sdma_val; in destroy_queue_nocpsch() 823 struct kfd_process_device *pdd; in update_queue() local 827 pdd = kfd_get_process_device_data(q->device, q->process); in update_queue() 828 if (!pdd) { in update_queue() [all …]
|
D | kfd_svm.c | 206 struct kfd_process_device *pdd; in svm_range_dma_map() local 209 pdd = kfd_process_device_from_gpuidx(p, gpuidx); in svm_range_dma_map() 210 if (!pdd) { in svm_range_dma_map() 215 r = svm_range_dma_map_dev(pdd->dev->adev, prange, offset, npages, in svm_range_dma_map() 244 struct kfd_process_device *pdd; in svm_range_free_dma_mappings() local 257 pdd = kfd_process_device_from_gpuidx(p, gpuidx); in svm_range_free_dma_mappings() 258 if (!pdd) { in svm_range_free_dma_mappings() 262 dev = &pdd->dev->pdev->dev; in svm_range_free_dma_mappings() 622 struct kfd_process_device *pdd; in svm_range_get_adev_by_id() local 633 pdd = kfd_process_device_from_gpuidx(p, gpu_idx); in svm_range_get_adev_by_id() [all …]
|
D | kfd_device_queue_manager.h | 283 static inline unsigned int get_sh_mem_bases_32(struct kfd_process_device *pdd) in get_sh_mem_bases_32() argument 285 return (pdd->lds_base >> 16) & 0xFF; in get_sh_mem_bases_32() 289 get_sh_mem_bases_nybble_64(struct kfd_process_device *pdd) in get_sh_mem_bases_nybble_64() argument 291 return (pdd->lds_base >> 60) & 0x0E; in get_sh_mem_bases_nybble_64()
|
D | kfd_iommu.h | 36 int kfd_iommu_bind_process_to_device(struct kfd_process_device *pdd); 59 struct kfd_process_device *pdd) in kfd_iommu_bind_process_to_device() argument
|
D | kfd_migrate.c | 411 struct kfd_process_device *pdd; in svm_migrate_vma_to_vram() local 480 pdd = svm_range_get_pdd_by_adev(prange, adev); in svm_migrate_vma_to_vram() 481 if (pdd) in svm_migrate_vma_to_vram() 482 WRITE_ONCE(pdd->page_in, pdd->page_in + cpages); in svm_migrate_vma_to_vram() 675 struct kfd_process_device *pdd; in svm_migrate_vma_to_ram() local 749 pdd = svm_range_get_pdd_by_adev(prange, adev); in svm_migrate_vma_to_ram() 750 if (pdd) in svm_migrate_vma_to_ram() 751 WRITE_ONCE(pdd->page_out, pdd->page_out + cpages); in svm_migrate_vma_to_ram()
|
D | kfd_priv.h | 960 int kfd_process_device_init_vm(struct kfd_process_device *pdd, 975 int kfd_process_device_create_obj_handle(struct kfd_process_device *pdd, 979 void kfd_process_device_remove_obj_handle(struct kfd_process_device *pdd, 1004 struct kfd_process_device *pdd, 1006 phys_addr_t kfd_get_process_doorbells(struct kfd_process_device *pdd); 1202 void kfd_process_dequeue_from_device(struct kfd_process_device *pdd); 1347 void kfd_flush_tlb(struct kfd_process_device *pdd, enum TLB_FLUSH_TYPE type);
|
D | kfd_events.c | 352 struct kfd_process_device *pdd; in kfd_kmap_event_page() local 362 pdd = kfd_process_device_data_by_id(p, GET_GPU_ID(event_page_offset)); in kfd_kmap_event_page() 363 if (!pdd) { in kfd_kmap_event_page() 367 kfd = pdd->dev; in kfd_kmap_event_page() 369 pdd = kfd_bind_process_to_device(kfd, p); in kfd_kmap_event_page() 370 if (IS_ERR(pdd)) in kfd_kmap_event_page() 371 return PTR_ERR(pdd); in kfd_kmap_event_page() 373 mem = kfd_process_device_translate_handle(pdd, in kfd_kmap_event_page()
|
D | kfd_smi_events.c | 317 struct kfd_process_device *pdd = p->pdds[i]; in kfd_smi_event_queue_restore_rescheduled() local 319 kfd_smi_event_add(p->lead_thread->pid, pdd->dev, in kfd_smi_event_queue_restore_rescheduled() 322 p->lead_thread->pid, pdd->dev->id, 'R'); in kfd_smi_event_queue_restore_rescheduled()
|
/linux-6.1.9/drivers/base/power/ |
D | domain_governor.c | 127 struct pm_domain_data *pdd; in update_domain_next_wakeup() local 141 list_for_each_entry(pdd, &genpd->dev_list, list_node) { in update_domain_next_wakeup() 142 next_wakeup = to_gpd_data(pdd)->td->next_wakeup; in update_domain_next_wakeup() 179 struct pm_domain_data *pdd; in __default_power_down_ok() local 215 list_for_each_entry(pdd, &genpd->dev_list, list_node) { in __default_power_down_ok() 224 td = to_gpd_data(pdd)->td; in __default_power_down_ok()
|
D | domain.c | 264 struct pm_domain_data *pdd; in _genpd_reeval_performance_state() local 276 list_for_each_entry(pdd, &genpd->dev_list, list_node) { in _genpd_reeval_performance_state() 277 pd_data = to_gpd_data(pdd); in _genpd_reeval_performance_state() 624 struct pm_domain_data *pdd; in genpd_power_off() local 659 list_for_each_entry(pdd, &genpd->dev_list, list_node) { in genpd_power_off() 664 if (!pm_runtime_suspended(pdd->dev) || in genpd_power_off() 665 irq_safe_dev_in_sleep_domain(pdd->dev, genpd)) in genpd_power_off() 781 struct pm_domain_data *pdd; in genpd_dev_pm_qos_notifier() local 786 pdd = dev->power.subsys_data ? in genpd_dev_pm_qos_notifier() 788 if (pdd) { in genpd_dev_pm_qos_notifier() [all …]
|
/linux-6.1.9/drivers/input/touchscreen/ |
D | wm9705.c | 77 static int pdd = 8; variable 78 module_param(pdd, int, 0); 79 MODULE_PARM_DESC(pdd, "Set pen detect comparator threshold"); 168 dig2 |= (pdd & 0x000f); in wm9705_phy_init() 169 dev_dbg(wm->dev, "setting pdd to Vmid/%d", 1 - (pdd & 0x000f)); in wm9705_phy_init()
|
/linux-6.1.9/drivers/soc/xilinx/ |
D | zynqmp_pm_domains.c | 108 struct pm_domain_data *pdd, *tmp; in zynqmp_gpd_power_off() local 119 list_for_each_entry_safe(pdd, tmp, &domain->dev_list, list_node) { in zynqmp_gpd_power_off() 121 may_wakeup = zynqmp_gpd_is_active_wakeup_path(pdd->dev, NULL); in zynqmp_gpd_power_off() 123 dev_dbg(pdd->dev, "device is in wakeup path in %s\n", in zynqmp_gpd_power_off()
|
/linux-6.1.9/include/linux/ |
D | pm_domain.h | 212 static inline struct generic_pm_domain_data *to_gpd_data(struct pm_domain_data *pdd) in to_gpd_data() argument 214 return container_of(pdd, struct generic_pm_domain_data, base); in to_gpd_data()
|
/linux-6.1.9/Documentation/admin-guide/ |
D | devices.txt | 883 48 = /dev/pdd Fourth parallel port IDE disk
|