Home
last modified time | relevance | path

Searched refs:pdd (Results 1 – 25 of 26) sorted by relevance

12

/linux-6.6.21/drivers/gpu/drm/amd/amdkfd/
Dkfd_flat_memory.c316 static void kfd_init_apertures_vi(struct kfd_process_device *pdd, uint8_t id) in kfd_init_apertures_vi() argument
322 pdd->lds_base = MAKE_LDS_APP_BASE_VI(); in kfd_init_apertures_vi()
323 pdd->lds_limit = MAKE_LDS_APP_LIMIT(pdd->lds_base); in kfd_init_apertures_vi()
329 pdd->gpuvm_base = SVM_USER_BASE; in kfd_init_apertures_vi()
330 pdd->gpuvm_limit = in kfd_init_apertures_vi()
331 pdd->dev->kfd->shared_resources.gpuvm_size - 1; in kfd_init_apertures_vi()
333 pdd->scratch_base = MAKE_SCRATCH_APP_BASE_VI(); in kfd_init_apertures_vi()
334 pdd->scratch_limit = MAKE_SCRATCH_APP_LIMIT(pdd->scratch_base); in kfd_init_apertures_vi()
337 static void kfd_init_apertures_v9(struct kfd_process_device *pdd, uint8_t id) in kfd_init_apertures_v9() argument
339 pdd->lds_base = MAKE_LDS_APP_BASE_V9(); in kfd_init_apertures_v9()
[all …]
Dkfd_debug.c71 struct kfd_process_device *pdd = process->pdds[i]; in kfd_dbg_ev_query_debug_event() local
73 & pdd->exception_status; in kfd_dbg_ev_query_debug_event()
78 *event_status = pdd->exception_status; in kfd_dbg_ev_query_debug_event()
79 *gpu_id = pdd->dev->id; in kfd_dbg_ev_query_debug_event()
80 pdd->exception_status &= ~exception_clear_mask; in kfd_dbg_ev_query_debug_event()
131 struct kfd_process_device *pdd = process->pdds[i]; in kfd_dbg_ev_raise() local
133 if (pdd->dev != dev) in kfd_dbg_ev_raise()
136 pdd->exception_status |= event_mask & KFD_EC_MASK_DEVICE; in kfd_dbg_ev_raise()
139 if (!pdd->vm_fault_exc_data) { in kfd_dbg_ev_raise()
140 pdd->vm_fault_exc_data = kmemdup( in kfd_dbg_ev_raise()
[all …]
Dkfd_process.c75 static void kfd_process_device_destroy_cwsr_dgpu(struct kfd_process_device *pdd);
88 struct kfd_process_device *pdd; member
102 struct kfd_process_device *pdd; in kfd_sdma_activity_worker() local
115 pdd = workarea->pdd; in kfd_sdma_activity_worker()
116 if (!pdd) in kfd_sdma_activity_worker()
118 dqm = pdd->dev->dqm; in kfd_sdma_activity_worker()
119 qpd = &pdd->qpd; in kfd_sdma_activity_worker()
174 workarea->sdma_activity_counter = pdd->sdma_past_activity_counter; in kfd_sdma_activity_worker()
184 mm = get_task_mm(pdd->process->lead_thread); in kfd_sdma_activity_worker()
211 workarea->sdma_activity_counter += pdd->sdma_past_activity_counter; in kfd_sdma_activity_worker()
[all …]
Dkfd_chardev.c71 struct kfd_process_device *pdd; in kfd_lock_pdd_by_id() local
74 pdd = kfd_process_device_data_by_id(p, gpu_id); in kfd_lock_pdd_by_id()
76 if (pdd) in kfd_lock_pdd_by_id()
77 return pdd; in kfd_lock_pdd_by_id()
83 static inline void kfd_unlock_pdd(struct kfd_process_device *pdd) in kfd_unlock_pdd() argument
85 mutex_unlock(&pdd->process->mutex); in kfd_unlock_pdd()
305 struct kfd_process_device *pdd; in kfd_ioctl_create_queue() local
322 pdd = kfd_process_device_data_by_id(p, args->gpu_id); in kfd_ioctl_create_queue()
323 if (!pdd) { in kfd_ioctl_create_queue()
328 dev = pdd->dev; in kfd_ioctl_create_queue()
[all …]
Dkfd_process_queue_manager.c82 void kfd_process_dequeue_from_device(struct kfd_process_device *pdd) in kfd_process_dequeue_from_device() argument
84 struct kfd_node *dev = pdd->dev; in kfd_process_dequeue_from_device()
86 if (pdd->already_dequeued) in kfd_process_dequeue_from_device()
89 dev->dqm->ops.process_termination(dev->dqm, &pdd->qpd); in kfd_process_dequeue_from_device()
91 amdgpu_mes_flush_shader_debugger(dev->adev, pdd->proc_ctx_gpu_addr); in kfd_process_dequeue_from_device()
92 pdd->already_dequeued = true; in kfd_process_dequeue_from_device()
100 struct kfd_process_device *pdd; in pqm_set_gws() local
115 pdd = kfd_get_process_device_data(dev, pqm->process); in pqm_set_gws()
116 if (!pdd) { in pqm_set_gws()
122 if (gws && pdd->qpd.num_gws) in pqm_set_gws()
[all …]
Dkfd_doorbell.c110 struct kfd_process_device *pdd; in kfd_doorbell_mmap() local
119 pdd = kfd_get_process_device_data(dev, process); in kfd_doorbell_mmap()
120 if (!pdd) in kfd_doorbell_mmap()
124 address = kfd_get_process_doorbells(pdd); in kfd_doorbell_mmap()
235 phys_addr_t kfd_get_process_doorbells(struct kfd_process_device *pdd) in kfd_get_process_doorbells() argument
237 struct amdgpu_device *adev = pdd->dev->adev; in kfd_get_process_doorbells()
240 if (!pdd->qpd.proc_doorbells) { in kfd_get_process_doorbells()
241 if (kfd_alloc_process_doorbells(pdd->dev->kfd, pdd)) in kfd_get_process_doorbells()
247 pdd->qpd.proc_doorbells, in kfd_get_process_doorbells()
249 pdd->dev->kfd->device_info.doorbell_size); in kfd_get_process_doorbells()
[all …]
Dkfd_device_queue_manager_v9.c42 static uint32_t compute_sh_mem_bases_64bit(struct kfd_process_device *pdd) in compute_sh_mem_bases_64bit() argument
44 uint32_t shared_base = pdd->lds_base >> 48; in compute_sh_mem_bases_64bit()
45 uint32_t private_base = pdd->scratch_base >> 48; in compute_sh_mem_bases_64bit()
54 struct kfd_process_device *pdd; in update_qpd_v9() local
56 pdd = qpd_to_pdd(qpd); in update_qpd_v9()
75 if (!pdd->process->xnack_enabled) in update_qpd_v9()
81 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(pdd); in update_qpd_v9()
Dkfd_device_queue_manager_v10.c43 static uint32_t compute_sh_mem_bases_64bit(struct kfd_process_device *pdd) in compute_sh_mem_bases_64bit() argument
45 uint32_t shared_base = pdd->lds_base >> 48; in compute_sh_mem_bases_64bit()
46 uint32_t private_base = pdd->scratch_base >> 48; in compute_sh_mem_bases_64bit()
55 struct kfd_process_device *pdd; in update_qpd_v10() local
57 pdd = qpd_to_pdd(qpd); in update_qpd_v10()
69 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(pdd); in update_qpd_v10()
Dkfd_device_queue_manager_v11.c42 static uint32_t compute_sh_mem_bases_64bit(struct kfd_process_device *pdd) in compute_sh_mem_bases_64bit() argument
44 uint32_t shared_base = pdd->lds_base >> 48; in compute_sh_mem_bases_64bit()
45 uint32_t private_base = pdd->scratch_base >> 48; in compute_sh_mem_bases_64bit()
54 struct kfd_process_device *pdd; in update_qpd_v11() local
56 pdd = qpd_to_pdd(qpd); in update_qpd_v11()
69 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(pdd); in update_qpd_v11()
Dkfd_device_queue_manager.c192 struct kfd_process_device *pdd = qpd_to_pdd(qpd); in add_queue_mes() local
207 queue_input.process_context_addr = pdd->proc_ctx_gpu_addr; in add_queue_mes()
700 struct kfd_process_device *pdd; in dbgdev_wave_reset_wavefronts() local
738 pdd = kfd_get_process_device_data(dev, p); in dbgdev_wave_reset_wavefronts()
739 if (!pdd) in dbgdev_wave_reset_wavefronts()
825 struct kfd_process_device *pdd = qpd_to_pdd(qpd); in destroy_queue_nocpsch() local
842 pdd->sdma_past_activity_counter += sdma_val; in destroy_queue_nocpsch()
855 struct kfd_process_device *pdd; in update_queue() local
859 pdd = kfd_get_process_device_data(q->device, q->process); in update_queue()
860 if (!pdd) { in update_queue()
[all …]
Dkfd_svm.c216 struct kfd_process_device *pdd; in svm_range_dma_map() local
219 pdd = kfd_process_device_from_gpuidx(p, gpuidx); in svm_range_dma_map()
220 if (!pdd) { in svm_range_dma_map()
225 r = svm_range_dma_map_dev(pdd->dev->adev, prange, offset, npages, in svm_range_dma_map()
254 struct kfd_process_device *pdd; in svm_range_free_dma_mappings() local
267 pdd = kfd_process_device_from_gpuidx(p, gpuidx); in svm_range_free_dma_mappings()
268 if (!pdd) { in svm_range_free_dma_mappings()
272 dev = &pdd->dev->adev->pdev->dev; in svm_range_free_dma_mappings()
641 struct kfd_process_device *pdd; in svm_range_get_node_by_id() local
644 pdd = kfd_process_device_data_by_id(p, gpu_id); in svm_range_get_node_by_id()
[all …]
Dkfd_packet_manager_v9.c38 struct kfd_process_device *pdd = in pm_map_process_v9() local
55 if (kfd->dqm->trap_debug_vmid && pdd->process->debug_trap_enabled && in pm_map_process_v9()
56 pdd->process->runtime_info.runtime_state == DEBUG_RUNTIME_STATE_ENABLED) { in pm_map_process_v9()
92 struct kfd_process_device *pdd = in pm_map_process_aldebaran() local
109 packet->spi_gdbg_per_vmid_cntl = pdd->spi_dbg_override | in pm_map_process_aldebaran()
110 pdd->spi_dbg_launch_mode; in pm_map_process_aldebaran()
112 if (pdd->process->debug_trap_enabled) { in pm_map_process_aldebaran()
114 packet->tcp_watch_cntl[i] = pdd->watch_points[i]; in pm_map_process_aldebaran()
117 !!(pdd->process->dbg_flags & KFD_DBG_TRAP_FLAG_SINGLE_MEM_OP); in pm_map_process_aldebaran()
Dkfd_device_queue_manager_vi.c109 struct kfd_process_device *pdd; in update_qpd_vi() local
112 pdd = qpd_to_pdd(qpd); in update_qpd_vi()
131 temp = get_sh_mem_bases_nybble_64(pdd); in update_qpd_vi()
Dkfd_device_queue_manager_cik.c107 struct kfd_process_device *pdd; in update_qpd_cik() local
110 pdd = qpd_to_pdd(qpd); in update_qpd_cik()
125 temp = get_sh_mem_bases_nybble_64(pdd); in update_qpd_cik()
Dkfd_debug.h58 int kfd_dbg_trap_clear_dev_address_watch(struct kfd_process_device *pdd,
60 int kfd_dbg_trap_set_dev_address_watch(struct kfd_process_device *pdd,
129 int kfd_dbg_set_mes_debug_mode(struct kfd_process_device *pdd, bool sq_trap_en);
Dkfd_device_queue_manager.h306 static inline unsigned int get_sh_mem_bases_32(struct kfd_process_device *pdd) in get_sh_mem_bases_32() argument
308 return (pdd->lds_base >> 16) & 0xFF; in get_sh_mem_bases_32()
312 get_sh_mem_bases_nybble_64(struct kfd_process_device *pdd) in get_sh_mem_bases_nybble_64() argument
314 return (pdd->lds_base >> 60) & 0x0E; in get_sh_mem_bases_nybble_64()
Dkfd_migrate.c401 struct kfd_process_device *pdd; in svm_migrate_vma_to_vram() local
469 pdd = svm_range_get_pdd_by_node(prange, node); in svm_migrate_vma_to_vram()
470 if (pdd) in svm_migrate_vma_to_vram()
471 WRITE_ONCE(pdd->page_in, pdd->page_in + cpages); in svm_migrate_vma_to_vram()
680 struct kfd_process_device *pdd; in svm_migrate_vma_to_ram() local
754 pdd = svm_range_get_pdd_by_node(prange, node); in svm_migrate_vma_to_ram()
755 if (pdd) in svm_migrate_vma_to_ram()
756 WRITE_ONCE(pdd->page_out, pdd->page_out + cpages); in svm_migrate_vma_to_ram()
Dkfd_priv.h1042 int kfd_process_device_init_vm(struct kfd_process_device *pdd,
1057 int kfd_process_device_create_obj_handle(struct kfd_process_device *pdd,
1061 void kfd_process_device_remove_obj_handle(struct kfd_process_device *pdd,
1086 struct kfd_process_device *pdd,
1088 phys_addr_t kfd_get_process_doorbells(struct kfd_process_device *pdd);
1090 struct kfd_process_device *pdd);
1092 struct kfd_process_device *pdd);
1155 int kfd_process_drain_interrupts(struct kfd_process_device *pdd);
1310 void kfd_process_dequeue_from_device(struct kfd_process_device *pdd);
1465 void kfd_flush_tlb(struct kfd_process_device *pdd, enum TLB_FLUSH_TYPE type);
Dkfd_events.c352 struct kfd_process_device *pdd; in kfd_kmap_event_page() local
362 pdd = kfd_process_device_data_by_id(p, GET_GPU_ID(event_page_offset)); in kfd_kmap_event_page()
363 if (!pdd) { in kfd_kmap_event_page()
367 kfd = pdd->dev; in kfd_kmap_event_page()
369 pdd = kfd_bind_process_to_device(kfd, p); in kfd_kmap_event_page()
370 if (IS_ERR(pdd)) in kfd_kmap_event_page()
371 return PTR_ERR(pdd); in kfd_kmap_event_page()
373 mem = kfd_process_device_translate_handle(pdd, in kfd_kmap_event_page()
Dkfd_smi_events.c317 struct kfd_process_device *pdd = p->pdds[i]; in kfd_smi_event_queue_restore_rescheduled() local
319 kfd_smi_event_add(p->lead_thread->pid, pdd->dev, in kfd_smi_event_queue_restore_rescheduled()
322 p->lead_thread->pid, pdd->dev->id, 'R'); in kfd_smi_event_queue_restore_rescheduled()
/linux-6.6.21/drivers/base/power/
Ddomain_governor.c127 struct pm_domain_data *pdd; in update_domain_next_wakeup() local
141 list_for_each_entry(pdd, &genpd->dev_list, list_node) { in update_domain_next_wakeup()
142 next_wakeup = to_gpd_data(pdd)->td->next_wakeup; in update_domain_next_wakeup()
179 struct pm_domain_data *pdd; in __default_power_down_ok() local
215 list_for_each_entry(pdd, &genpd->dev_list, list_node) { in __default_power_down_ok()
224 td = to_gpd_data(pdd)->td; in __default_power_down_ok()
Ddomain.c261 struct pm_domain_data *pdd; in _genpd_reeval_performance_state() local
273 list_for_each_entry(pdd, &genpd->dev_list, list_node) { in _genpd_reeval_performance_state()
274 pd_data = to_gpd_data(pdd); in _genpd_reeval_performance_state()
672 struct pm_domain_data *pdd; in genpd_power_off() local
707 list_for_each_entry(pdd, &genpd->dev_list, list_node) { in genpd_power_off()
712 if (!pm_runtime_suspended(pdd->dev) || in genpd_power_off()
713 irq_safe_dev_in_sleep_domain(pdd->dev, genpd)) in genpd_power_off()
829 struct pm_domain_data *pdd; in genpd_dev_pm_qos_notifier() local
834 pdd = dev->power.subsys_data ? in genpd_dev_pm_qos_notifier()
836 if (pdd) { in genpd_dev_pm_qos_notifier()
[all …]
/linux-6.6.21/drivers/input/touchscreen/
Dwm9705.c77 static int pdd = 8; variable
78 module_param(pdd, int, 0);
79 MODULE_PARM_DESC(pdd, "Set pen detect comparator threshold");
168 dig2 |= (pdd & 0x000f); in wm9705_phy_init()
169 dev_dbg(wm->dev, "setting pdd to Vmid/%d", 1 - (pdd & 0x000f)); in wm9705_phy_init()
/linux-6.6.21/drivers/pmdomain/xilinx/
Dzynqmp-pm-domains.c108 struct pm_domain_data *pdd, *tmp; in zynqmp_gpd_power_off() local
119 list_for_each_entry_safe(pdd, tmp, &domain->dev_list, list_node) { in zynqmp_gpd_power_off()
121 may_wakeup = zynqmp_gpd_is_active_wakeup_path(pdd->dev, NULL); in zynqmp_gpd_power_off()
123 dev_dbg(pdd->dev, "device is in wakeup path in %s\n", in zynqmp_gpd_power_off()
/linux-6.6.21/include/linux/
Dpm_domain.h215 static inline struct generic_pm_domain_data *to_gpd_data(struct pm_domain_data *pdd) in to_gpd_data() argument
217 return container_of(pdd, struct generic_pm_domain_data, base); in to_gpd_data()

12