/linux-6.6.21/drivers/gpu/drm/amd/amdkfd/ |
D | kfd_device_queue_manager_v9.c | 30 struct qcm_process_device *qpd); 32 struct qcm_process_device *qpd); 52 struct qcm_process_device *qpd) in update_qpd_v9() argument 56 pdd = qpd_to_pdd(qpd); in update_qpd_v9() 59 if (qpd->sh_mem_config == 0) { in update_qpd_v9() 60 qpd->sh_mem_config = SH_MEM_ALIGNMENT_MODE_UNALIGNED << in update_qpd_v9() 64 qpd->sh_mem_config |= 1 << SH_MEM_CONFIG__RETRY_DISABLE__SHIFT; in update_qpd_v9() 67 qpd->sh_mem_config |= in update_qpd_v9() 70 qpd->sh_mem_ape1_limit = 0; in update_qpd_v9() 71 qpd->sh_mem_ape1_base = 0; in update_qpd_v9() [all …]
|
D | kfd_device_queue_manager_cik.c | 31 struct qcm_process_device *qpd, 37 struct qcm_process_device *qpd); 40 struct qcm_process_device *qpd); 79 struct qcm_process_device *qpd, in set_cache_memory_policy_cik() argument 96 qpd->sh_mem_config = (qpd->sh_mem_config & PTR32) in set_cache_memory_policy_cik() 105 struct qcm_process_device *qpd) in update_qpd_cik() argument 110 pdd = qpd_to_pdd(qpd); in update_qpd_cik() 113 if (qpd->sh_mem_config == 0) { in update_qpd_cik() 114 qpd->sh_mem_config = in update_qpd_cik() 118 qpd->sh_mem_ape1_limit = 0; in update_qpd_cik() [all …]
|
D | kfd_packet_manager_v9.c | 33 uint32_t *buffer, struct qcm_process_device *qpd) in pm_map_process_v9() argument 36 uint64_t vm_page_table_base_addr = qpd->page_table_base; in pm_map_process_v9() 39 container_of(qpd, struct kfd_process_device, qpd); in pm_map_process_v9() 45 packet->bitfields2.diq_enable = (qpd->is_debug) ? 1 : 0; in pm_map_process_v9() 47 packet->bitfields2.pasid = qpd->pqm->process->pasid; in pm_map_process_v9() 48 packet->bitfields14.gds_size = qpd->gds_size & 0x3F; in pm_map_process_v9() 49 packet->bitfields14.gds_size_hi = (qpd->gds_size >> 6) & 0xF; in pm_map_process_v9() 50 packet->bitfields14.num_gws = (qpd->mapped_gws_queue) ? qpd->num_gws : 0; in pm_map_process_v9() 51 packet->bitfields14.num_oac = qpd->num_oac; in pm_map_process_v9() 53 packet->bitfields14.num_queues = (qpd->is_debug) ? 0 : qpd->queue_count; in pm_map_process_v9() [all …]
|
D | kfd_device_queue_manager.c | 141 struct qcm_process_device *qpd) in program_sh_mem_settings() argument 148 dqm->dev->adev, qpd->vmid, qpd->sh_mem_config, in program_sh_mem_settings() 149 qpd->sh_mem_ape1_base, qpd->sh_mem_ape1_limit, in program_sh_mem_settings() 150 qpd->sh_mem_bases, xcc_id); in program_sh_mem_settings() 189 struct qcm_process_device *qpd) in add_queue_mes() argument 192 struct kfd_process_device *pdd = qpd_to_pdd(qpd); in add_queue_mes() 201 queue_input.process_id = qpd->pqm->process->pasid; in add_queue_mes() 202 queue_input.page_table_base_addr = qpd->page_table_base; in add_queue_mes() 227 queue_input.tba_addr = qpd->tba_addr; in add_queue_mes() 228 queue_input.tma_addr = qpd->tma_addr; in add_queue_mes() [all …]
|
D | kfd_device_queue_manager_vi.c | 31 struct qcm_process_device *qpd, 37 struct qcm_process_device *qpd); 40 struct qcm_process_device *qpd); 80 struct qcm_process_device *qpd, in set_cache_memory_policy_vi() argument 97 qpd->sh_mem_config = in set_cache_memory_policy_vi() 107 struct qcm_process_device *qpd) in update_qpd_vi() argument 112 pdd = qpd_to_pdd(qpd); in update_qpd_vi() 115 if (qpd->sh_mem_config == 0) { in update_qpd_vi() 116 qpd->sh_mem_config = in update_qpd_vi() 124 qpd->sh_mem_ape1_limit = 0; in update_qpd_vi() [all …]
|
D | kfd_device_queue_manager_v10.c | 31 struct qcm_process_device *qpd); 33 struct qcm_process_device *qpd); 53 struct qcm_process_device *qpd) in update_qpd_v10() argument 57 pdd = qpd_to_pdd(qpd); in update_qpd_v10() 60 if (qpd->sh_mem_config == 0) { in update_qpd_v10() 61 qpd->sh_mem_config = in update_qpd_v10() 65 qpd->sh_mem_ape1_limit = 0; in update_qpd_v10() 66 qpd->sh_mem_ape1_base = 0; in update_qpd_v10() 69 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(pdd); in update_qpd_v10() 71 pr_debug("sh_mem_bases 0x%X\n", qpd->sh_mem_bases); in update_qpd_v10() [all …]
|
D | kfd_device_queue_manager_v11.c | 30 struct qcm_process_device *qpd); 32 struct qcm_process_device *qpd); 52 struct qcm_process_device *qpd) in update_qpd_v11() argument 56 pdd = qpd_to_pdd(qpd); in update_qpd_v11() 59 if (qpd->sh_mem_config == 0) { in update_qpd_v11() 60 qpd->sh_mem_config = in update_qpd_v11() 65 qpd->sh_mem_ape1_limit = 0; in update_qpd_v11() 66 qpd->sh_mem_ape1_base = 0; in update_qpd_v11() 69 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(pdd); in update_qpd_v11() 71 pr_debug("sh_mem_bases 0x%X\n", qpd->sh_mem_bases); in update_qpd_v11() [all …]
|
D | kfd_doorbell.c | 208 static int init_doorbell_bitmap(struct qcm_process_device *qpd, in init_doorbell_bitmap() argument 226 __set_bit(i, qpd->doorbell_bitmap); in init_doorbell_bitmap() 228 qpd->doorbell_bitmap); in init_doorbell_bitmap() 240 if (!pdd->qpd.proc_doorbells) { in kfd_get_process_doorbells() 247 pdd->qpd.proc_doorbells, in kfd_get_process_doorbells() 256 struct qcm_process_device *qpd = &pdd->qpd; in kfd_alloc_process_doorbells() local 259 qpd->doorbell_bitmap = bitmap_zalloc(KFD_MAX_NUM_OF_QUEUES_PER_PROCESS, in kfd_alloc_process_doorbells() 261 if (!qpd->doorbell_bitmap) { in kfd_alloc_process_doorbells() 266 r = init_doorbell_bitmap(&pdd->qpd, kfd); in kfd_alloc_process_doorbells() 278 &qpd->proc_doorbells, in kfd_alloc_process_doorbells() [all …]
|
D | kfd_device_queue_manager.h | 43 struct qcm_process_device *qpd; member 134 struct qcm_process_device *qpd, 140 struct qcm_process_device *qpd, 147 struct qcm_process_device *qpd); 150 struct qcm_process_device *qpd); 159 struct qcm_process_device *qpd); 163 struct qcm_process_device *qpd); 166 struct qcm_process_device *qpd, 173 struct qcm_process_device *qpd); 176 struct qcm_process_device *qpd); [all …]
|
D | kfd_process.c | 106 struct qcm_process_device *qpd; in kfd_sdma_activity_worker() local 119 qpd = &pdd->qpd; in kfd_sdma_activity_worker() 120 if (!dqm || !qpd) in kfd_sdma_activity_worker() 151 list_for_each_entry(q, &qpd->queues_list, list) { in kfd_sdma_activity_worker() 213 list_for_each_entry(q, &qpd->queues_list, list) { in kfd_sdma_activity_worker() 281 if (pdd->qpd.queue_count == 0) { in kfd_get_cu_occupancy() 765 struct qcm_process_device *qpd = &pdd->qpd; in kfd_process_device_reserve_ib_mem() local 774 if (qpd->ib_kaddr || !qpd->ib_base) in kfd_process_device_reserve_ib_mem() 778 ret = kfd_process_alloc_gpuvm(pdd, qpd->ib_base, PAGE_SIZE, flags, in kfd_process_device_reserve_ib_mem() 783 qpd->ib_mem = mem; in kfd_process_device_reserve_ib_mem() [all …]
|
D | kfd_packet_manager_vi.c | 43 struct qcm_process_device *qpd) in pm_map_process_vi() argument 53 packet->bitfields2.diq_enable = (qpd->is_debug) ? 1 : 0; in pm_map_process_vi() 55 packet->bitfields2.pasid = qpd->pqm->process->pasid; in pm_map_process_vi() 56 packet->bitfields3.page_table_base = qpd->page_table_base; in pm_map_process_vi() 57 packet->bitfields10.gds_size = qpd->gds_size; in pm_map_process_vi() 58 packet->bitfields10.num_gws = qpd->num_gws; in pm_map_process_vi() 59 packet->bitfields10.num_oac = qpd->num_oac; in pm_map_process_vi() 60 packet->bitfields10.num_queues = (qpd->is_debug) ? 0 : qpd->queue_count; in pm_map_process_vi() 62 packet->sh_mem_config = qpd->sh_mem_config; in pm_map_process_vi() 63 packet->sh_mem_bases = qpd->sh_mem_bases; in pm_map_process_vi() [all …]
|
D | kfd_process_queue_manager.c | 89 dev->dqm->ops.process_termination(dev->dqm, &pdd->qpd); in kfd_process_dequeue_from_device() 122 if (gws && pdd->qpd.num_gws) in pqm_set_gws() 125 if (!gws && pdd->qpd.num_gws == 0) in pqm_set_gws() 148 pdd->qpd.num_gws = gws ? dev->adev->gds.gws_size : 0; in pqm_set_gws() 193 pdd->qpd.num_gws = 0; in pqm_clean_queue_resource() 309 if ((pdd->qpd.is_debug) || (type == KFD_QUEUE_TYPE_DIQ)) in pqm_create_queue() 312 if (pdd->qpd.queue_count >= max_queues) in pqm_create_queue() 324 if (list_empty(&pdd->qpd.queues_list) && in pqm_create_queue() 325 list_empty(&pdd->qpd.priv_queue_list)) in pqm_create_queue() 326 dev->dqm->ops.register_process(dev->dqm, &pdd->qpd); in pqm_create_queue() [all …]
|
D | kfd_packet_manager.c | 130 struct qcm_process_device *qpd; in pm_create_runlist_ib() local 150 qpd = cur->qpd; in pm_create_runlist_ib() 158 retval = pm->pmf->map_process(pm, &rl_buffer[rl_wptr], qpd); in pm_create_runlist_ib() 166 list_for_each_entry(kq, &qpd->priv_queue_list, list) { in pm_create_runlist_ib() 171 kq->queue->queue, qpd->is_debug); in pm_create_runlist_ib() 176 qpd->is_debug); in pm_create_runlist_ib() 185 list_for_each_entry(q, &qpd->queues_list, list) { in pm_create_runlist_ib() 190 q->queue, qpd->is_debug); in pm_create_runlist_ib() 195 qpd->is_debug); in pm_create_runlist_ib()
|
D | kfd_flat_memory.c | 414 pdd->qpd.cwsr_base = SVM_CWSR_BASE; in kfd_init_apertures() 415 pdd->qpd.ib_base = SVM_IB_BASE; in kfd_init_apertures()
|
D | kfd_debug.c | 584 kfd_process_set_trap_debug_flag(&pdd->qpd, false); in kfd_dbg_trap_deactivate() 597 release_debug_trap_vmid(pdd->dev->dqm, &pdd->qpd)) in kfd_dbg_trap_deactivate() 675 r = reserve_debug_trap_vmid(pdd->dev->dqm, &pdd->qpd); in kfd_dbg_trap_activate() 714 kfd_process_set_trap_debug_flag(&pdd->qpd, true); in kfd_dbg_trap_activate() 755 if (pdd->qpd.num_gws && (!kfd_dbg_has_gws_support(pdd->dev) || in kfd_dbg_trap_enable() 921 struct qcm_process_device *qpd = &pdd->qpd; in kfd_dbg_trap_query_exception_info() local 923 list_for_each_entry(queue, &qpd->queues_list, list) { in kfd_dbg_trap_query_exception_info()
|
D | kfd_priv.h | 738 struct qcm_process_device qpd; member 841 #define qpd_to_pdd(x) container_of(x, struct kfd_process_device, qpd) 1161 void kfd_process_set_trap_handler(struct qcm_process_device *qpd, 1164 void kfd_process_set_trap_debug_flag(struct qcm_process_device *qpd, 1374 struct qcm_process_device *qpd);
|
D | kfd_chardev.c | 336 if (!pdd->qpd.proc_doorbells) { in kfd_ioctl_create_queue() 611 &pdd->qpd, in kfd_ioctl_set_memory_policy() 646 kfd_process_set_trap_handler(&pdd->qpd, args->tba_addr, args->tma_addr); in kfd_ioctl_set_trap_handler() 920 pdd->qpd.sh_hidden_private_base = args->va_addr; in kfd_ioctl_set_scratch_backing_va() 925 pdd->qpd.vmid != 0 && dev->kfd2kgd->set_scratch_backing_va) in kfd_ioctl_set_scratch_backing_va() 927 dev->adev, args->va_addr, pdd->qpd.vmid); in kfd_ioctl_set_scratch_backing_va() 2267 if (!pdd->qpd.proc_doorbells) { in criu_restore_devices() 2758 if (pdd->qpd.queue_count) in runtime_enable()
|
/linux-6.6.21/drivers/infiniband/hw/cxgb4/ |
D | device.c | 246 struct c4iw_debugfs_data *qpd) in dump_qp() argument 253 space = qpd->bufsize - qpd->pos - 1; in dump_qp() 267 cc = snprintf(qpd->buf + qpd->pos, space, in dump_qp() 288 cc = snprintf(qpd->buf + qpd->pos, space, in dump_qp() 304 cc = snprintf(qpd->buf + qpd->pos, space, in dump_qp() 310 qpd->pos += cc; in dump_qp() 316 struct c4iw_debugfs_data *qpd = file->private_data; in qp_release() local 317 if (!qpd) { in qp_release() 321 vfree(qpd->buf); in qp_release() 322 kfree(qpd); in qp_release() [all …]
|