/linux-6.6.21/drivers/gpu/drm/ |
D | drm_exec.c | 54 static void drm_exec_unlock_all(struct drm_exec *exec) in drm_exec_unlock_all() argument 59 drm_exec_for_each_locked_object_reverse(exec, index, obj) { in drm_exec_unlock_all() 64 drm_gem_object_put(exec->prelocked); in drm_exec_unlock_all() 65 exec->prelocked = NULL; in drm_exec_unlock_all() 75 void drm_exec_init(struct drm_exec *exec, uint32_t flags) in drm_exec_init() argument 77 exec->flags = flags; in drm_exec_init() 78 exec->objects = kmalloc(PAGE_SIZE, GFP_KERNEL); in drm_exec_init() 81 exec->max_objects = exec->objects ? PAGE_SIZE / sizeof(void *) : 0; in drm_exec_init() 82 exec->num_objects = 0; in drm_exec_init() 83 exec->contended = DRM_EXEC_DUMMY; in drm_exec_init() [all …]
|
/linux-6.6.21/drivers/gpu/drm/tests/ |
D | drm_exec_test.c | 47 struct drm_exec exec; in sanitycheck() local 49 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT); in sanitycheck() 50 drm_exec_fini(&exec); in sanitycheck() 58 struct drm_exec exec; in test_lock() local 63 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT); in test_lock() 64 drm_exec_until_all_locked(&exec) { in test_lock() 65 ret = drm_exec_lock_obj(&exec, &gobj); in test_lock() 66 drm_exec_retry_on_contention(&exec); in test_lock() 71 drm_exec_fini(&exec); in test_lock() 78 struct drm_exec exec; in test_lock_unlock() local [all …]
|
/linux-6.6.21/drivers/gpu/drm/vc4/ |
D | vc4_validate.c | 51 struct vc4_exec_info *exec, \ 106 vc4_use_bo(struct vc4_exec_info *exec, uint32_t hindex) in vc4_use_bo() argument 108 struct vc4_dev *vc4 = exec->dev; in vc4_use_bo() 115 if (hindex >= exec->bo_count) { in vc4_use_bo() 117 hindex, exec->bo_count); in vc4_use_bo() 120 obj = to_drm_gem_dma_obj(exec->bo[hindex]); in vc4_use_bo() 133 vc4_use_handle(struct vc4_exec_info *exec, uint32_t gem_handles_packet_index) in vc4_use_handle() argument 135 return vc4_use_bo(exec, exec->bo_index[gem_handles_packet_index]); in vc4_use_handle() 139 validate_bin_pos(struct vc4_exec_info *exec, void *untrusted, uint32_t pos) in validate_bin_pos() argument 144 return (untrusted - 1 == exec->bin_u + pos); in validate_bin_pos() [all …]
|
D | vc4_gem.c | 157 struct vc4_exec_info *exec[2]; in vc4_save_hang_state() local 169 exec[0] = vc4_first_bin_job(vc4); in vc4_save_hang_state() 170 exec[1] = vc4_first_render_job(vc4); in vc4_save_hang_state() 171 if (!exec[0] && !exec[1]) { in vc4_save_hang_state() 179 if (!exec[i]) in vc4_save_hang_state() 183 list_for_each_entry(bo, &exec[i]->unref_list, unref_head) in vc4_save_hang_state() 185 state->bo_count += exec[i]->bo_count + unref_list_count; in vc4_save_hang_state() 198 if (!exec[i]) in vc4_save_hang_state() 201 for (j = 0; j < exec[i]->bo_count; j++) { in vc4_save_hang_state() 202 bo = to_vc4_bo(exec[i]->bo[j]); in vc4_save_hang_state() [all …]
|
D | vc4_irq.c | 67 struct vc4_exec_info *exec; in vc4_overflow_mem_work() local 92 exec = vc4_first_bin_job(vc4); in vc4_overflow_mem_work() 93 if (!exec) in vc4_overflow_mem_work() 94 exec = vc4_last_render_job(vc4); in vc4_overflow_mem_work() 95 if (exec) { in vc4_overflow_mem_work() 96 exec->bin_slots |= vc4->bin_alloc_overflow; in vc4_overflow_mem_work() 120 struct vc4_exec_info *next, *exec = vc4_first_bin_job(vc4); in vc4_irq_finish_bin_job() local 122 if (!exec) in vc4_irq_finish_bin_job() 125 trace_vc4_bcl_end_irq(dev, exec->seqno); in vc4_irq_finish_bin_job() 127 vc4_move_job_to_render(dev, exec); in vc4_irq_finish_bin_job() [all …]
|
D | vc4_render_cl.c | 99 static uint32_t vc4_full_res_offset(struct vc4_exec_info *exec, in vc4_full_res_offset() argument 105 (DIV_ROUND_UP(exec->args->width, 32) * y + x); in vc4_full_res_offset() 123 static void emit_tile(struct vc4_exec_info *exec, in emit_tile() argument 127 struct drm_vc4_submit_cl *args = exec->args; in emit_tile() 139 vc4_full_res_offset(exec, setup->color_read, in emit_tile() 161 vc4_full_res_offset(exec, setup->zs_read, in emit_tile() 185 rcl_u32(setup, (exec->tile_alloc_offset + in emit_tile() 186 (y * exec->bin_tiles_x + x) * 32)); in emit_tile() 201 vc4_full_res_offset(exec, setup->msaa_color_write, in emit_tile() 219 vc4_full_res_offset(exec, setup->msaa_zs_write, in emit_tile() [all …]
|
/linux-6.6.21/include/drm/ |
D | drm_exec.h | 63 drm_exec_obj(struct drm_exec *exec, unsigned long index) in drm_exec_obj() argument 65 return index < exec->num_objects ? exec->objects[index] : NULL; in drm_exec_obj() 76 #define drm_exec_for_each_locked_object(exec, index, obj) \ argument 77 for ((index) = 0; ((obj) = drm_exec_obj(exec, index)); ++(index)) 90 #define drm_exec_for_each_locked_object_reverse(exec, index, obj) \ argument 91 for ((index) = (exec)->num_objects - 1; \ 92 ((obj) = drm_exec_obj(exec, index)); --(index)) 105 #define drm_exec_until_all_locked(exec) \ argument 110 drm_exec_cleanup(exec); \ 120 #define drm_exec_retry_on_contention(exec) \ argument [all …]
|
/linux-6.6.21/include/uapi/linux/ |
D | a.out.h | 44 #define N_MAGIC(exec) ((exec).a_info & 0xffff) argument 46 #define N_MACHTYPE(exec) ((enum machine_type)(((exec).a_info >> 16) & 0xff)) argument 47 #define N_FLAGS(exec) (((exec).a_info >> 24) & 0xff) argument 48 #define N_SET_INFO(exec, magic, type, flags) \ argument 49 ((exec).a_info = ((magic) & 0xffff) \ 52 #define N_SET_MAGIC(exec, magic) \ argument 53 ((exec).a_info = (((exec).a_info & 0xffff0000) | ((magic) & 0xffff))) 55 #define N_SET_MACHTYPE(exec, machtype) \ argument 56 ((exec).a_info = \ 57 ((exec).a_info&0xff00ffff) | ((((int)(machtype))&0xff) << 16)) [all …]
|
/linux-6.6.21/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_csa.c | 70 struct drm_exec exec; in amdgpu_map_static_csa() local 73 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT); in amdgpu_map_static_csa() 74 drm_exec_until_all_locked(&exec) { in amdgpu_map_static_csa() 75 r = amdgpu_vm_lock_pd(vm, &exec, 0); in amdgpu_map_static_csa() 77 r = drm_exec_lock_obj(&exec, &bo->tbo.base); in amdgpu_map_static_csa() 78 drm_exec_retry_on_contention(&exec); in amdgpu_map_static_csa() 102 drm_exec_fini(&exec); in amdgpu_map_static_csa() 110 struct drm_exec exec; in amdgpu_unmap_static_csa() local 113 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT); in amdgpu_unmap_static_csa() 114 drm_exec_until_all_locked(&exec) { in amdgpu_unmap_static_csa() [all …]
|
D | amdgpu_gem.c | 203 struct drm_exec exec; in amdgpu_gem_object_close() local 206 drm_exec_init(&exec, DRM_EXEC_IGNORE_DUPLICATES); in amdgpu_gem_object_close() 207 drm_exec_until_all_locked(&exec) { in amdgpu_gem_object_close() 208 r = drm_exec_prepare_obj(&exec, &bo->tbo.base, 1); in amdgpu_gem_object_close() 209 drm_exec_retry_on_contention(&exec); in amdgpu_gem_object_close() 213 r = amdgpu_vm_lock_pd(vm, &exec, 0); in amdgpu_gem_object_close() 214 drm_exec_retry_on_contention(&exec); in amdgpu_gem_object_close() 240 drm_exec_fini(&exec); in amdgpu_gem_object_close() 680 struct drm_exec exec; in amdgpu_gem_va_ioctl() local 741 drm_exec_init(&exec, DRM_EXEC_INTERRUPTIBLE_WAIT | in amdgpu_gem_va_ioctl() [all …]
|
D | amdgpu_amdkfd_gpuvm.c | 1069 struct drm_exec exec; member 1099 drm_exec_init(&ctx->exec, DRM_EXEC_INTERRUPTIBLE_WAIT); in reserve_bo_and_vm() 1100 drm_exec_until_all_locked(&ctx->exec) { in reserve_bo_and_vm() 1101 ret = amdgpu_vm_lock_pd(vm, &ctx->exec, 2); in reserve_bo_and_vm() 1102 drm_exec_retry_on_contention(&ctx->exec); in reserve_bo_and_vm() 1106 ret = drm_exec_prepare_obj(&ctx->exec, &bo->tbo.base, 1); in reserve_bo_and_vm() 1107 drm_exec_retry_on_contention(&ctx->exec); in reserve_bo_and_vm() 1115 drm_exec_fini(&ctx->exec); in reserve_bo_and_vm() 1138 drm_exec_init(&ctx->exec, DRM_EXEC_INTERRUPTIBLE_WAIT); in reserve_bo_and_cond_vms() 1139 drm_exec_until_all_locked(&ctx->exec) { in reserve_bo_and_cond_vms() [all …]
|
/linux-6.6.21/drivers/gpu/drm/nouveau/ |
D | nouveau_exec.c | 94 struct drm_exec *exec = &job->exec; in nouveau_exec_job_submit() local 105 drm_exec_init(exec, DRM_EXEC_INTERRUPTIBLE_WAIT | in nouveau_exec_job_submit() 107 drm_exec_until_all_locked(exec) { in nouveau_exec_job_submit() 114 ret = drm_exec_prepare_obj(exec, va->gem.obj, 1); in nouveau_exec_job_submit() 115 drm_exec_retry_on_contention(exec); in nouveau_exec_job_submit() 122 drm_exec_for_each_locked_object(exec, index, obj) { in nouveau_exec_job_submit() 135 drm_exec_fini(exec); in nouveau_exec_job_submit() 143 struct drm_exec *exec = &job->exec; in nouveau_exec_job_armed_submit() local 147 drm_exec_for_each_locked_object(exec, index, obj) in nouveau_exec_job_armed_submit() 150 drm_exec_fini(exec); in nouveau_exec_job_armed_submit()
|
/linux-6.6.21/tools/perf/util/ |
D | comm.c | 108 struct comm *comm__new(const char *str, u64 timestamp, bool exec) in comm__new() argument 116 comm->exec = exec; in comm__new() 127 int comm__override(struct comm *comm, const char *str, u64 timestamp, bool exec) in comm__override() argument 138 if (exec) in comm__override() 139 comm->exec = true; in comm__override()
|
D | comm.h | 15 bool exec; member 23 struct comm *comm__new(const char *str, u64 timestamp, bool exec); 26 bool exec);
|
D | thread.c | 216 if (comm->exec) in thread__exec_comm() 235 u64 timestamp, bool exec) in ____thread__set_comm() argument 241 int err = comm__override(curr, str, timestamp, exec); in ____thread__set_comm() 245 new = comm__new(str, timestamp, exec); in ____thread__set_comm() 250 if (exec) in ____thread__set_comm() 260 bool exec) in __thread__set_comm() argument 265 ret = ____thread__set_comm(thread, str, timestamp, exec); in __thread__set_comm()
|
/linux-6.6.21/tools/testing/selftests/exec/ |
D | null-argv.c | 10 #define FORK(exec) \ argument 15 exec; /* Some kind of exec */ \ 16 perror("# " #exec); \ 19 check_result(pid, #exec); \
|
/linux-6.6.21/drivers/leds/ |
D | leds-lp8501.c | 162 u8 exec; in lp8501_run_engine() local 180 ret = lp55xx_read(chip, LP8501_REG_ENABLE, &exec); in lp8501_run_engine() 187 exec = (exec & ~LP8501_EXEC_ENG1_M) | LP8501_RUN_ENG1; in lp8501_run_engine() 192 exec = (exec & ~LP8501_EXEC_ENG2_M) | LP8501_RUN_ENG2; in lp8501_run_engine() 197 exec = (exec & ~LP8501_EXEC_ENG3_M) | LP8501_RUN_ENG3; in lp8501_run_engine() 203 lp55xx_update_bits(chip, LP8501_REG_ENABLE, LP8501_EXEC_M, exec); in lp8501_run_engine()
|
D | leds-lp5562.c | 157 u8 exec; in lp5562_run_engine() local 179 ret = lp55xx_read(chip, LP5562_REG_ENABLE, &exec); in lp5562_run_engine() 186 exec = (exec & ~LP5562_EXEC_ENG1_M) | LP5562_RUN_ENG1; in lp5562_run_engine() 191 exec = (exec & ~LP5562_EXEC_ENG2_M) | LP5562_RUN_ENG2; in lp5562_run_engine() 196 exec = (exec & ~LP5562_EXEC_ENG3_M) | LP5562_RUN_ENG3; in lp5562_run_engine() 202 lp55xx_update_bits(chip, LP5562_REG_ENABLE, LP5562_EXEC_M, exec); in lp5562_run_engine()
|
D | leds-lp5521.c | 162 u8 exec; in lp5521_run_engine() local 181 ret = lp55xx_read(chip, LP5521_REG_ENABLE, &exec); in lp5521_run_engine() 188 exec = (exec & ~LP5521_EXEC_R_M) | LP5521_RUN_R; in lp5521_run_engine() 193 exec = (exec & ~LP5521_EXEC_G_M) | LP5521_RUN_G; in lp5521_run_engine() 198 exec = (exec & ~LP5521_EXEC_B_M) | LP5521_RUN_B; in lp5521_run_engine() 204 lp55xx_update_bits(chip, LP5521_REG_ENABLE, LP5521_EXEC_M, exec); in lp5521_run_engine()
|
/linux-6.6.21/drivers/gpu/drm/nouveau/nvkm/subdev/timer/ |
D | base.c | 73 LIST_HEAD(exec); in nvkm_timer_alarm_trigger() 90 list_add(&alarm->exec, &exec); in nvkm_timer_alarm_trigger() 99 list_for_each_entry_safe(alarm, atemp, &exec, exec) { in nvkm_timer_alarm_trigger() 100 list_del(&alarm->exec); in nvkm_timer_alarm_trigger()
|
/linux-6.6.21/tools/testing/selftests/bpf/progs/ |
D | test_overhead.c | 13 int BPF_KPROBE(prog1, struct task_struct *tsk, const char *buf, bool exec) in BPF_KPROBE() argument 31 int BPF_PROG(prog4, struct task_struct *tsk, const char *buf, bool exec) in BPF_PROG() argument 37 int BPF_PROG(prog5, struct task_struct *tsk, const char *buf, bool exec) in BPF_PROG() argument
|
/linux-6.6.21/drivers/sbus/char/ |
D | oradax.c | 861 ctx->result.exec.status = DAX_SUBMIT_ERR_THR_INIT; in dax_ccb_exec() 868 ctx->result.exec.status = DAX_SUBMIT_ERR_NO_CA_AVAIL; in dax_ccb_exec() 878 ctx->result.exec.status = DAX_SUBMIT_ERR_CCB_ARR_MMU_MISS; in dax_ccb_exec() 886 ctx->result.exec.status = DAX_SUBMIT_ERR_NO_CA_AVAIL; in dax_ccb_exec() 892 ctx->result.exec.status = dax_preprocess_usr_ccbs(ctx, idx, nccbs); in dax_ccb_exec() 893 if (ctx->result.exec.status != DAX_SUBMIT_OK) in dax_ccb_exec() 896 ctx->result.exec.status = dax_lock_pages(ctx, idx, nccbs, in dax_ccb_exec() 897 &ctx->result.exec.status_data); in dax_ccb_exec() 898 if (ctx->result.exec.status != DAX_SUBMIT_OK) in dax_ccb_exec() 906 &accepted_len, &ctx->result.exec.status_data); in dax_ccb_exec() [all …]
|
/linux-6.6.21/arch/sparc/mm/ |
D | tlb.c | 70 bool exec, unsigned int hugepage_shift) in tlb_batch_add_one() argument 76 if (exec) in tlb_batch_add_one() 159 bool exec = pte_exec(*pte); in tlb_batch_pmd_scan() local 161 tlb_batch_add_one(mm, vaddr, exec, PAGE_SHIFT); in tlb_batch_pmd_scan() 211 bool exec = pte_exec(orig_pte); in __set_pmd_acct() local 213 tlb_batch_add_one(mm, addr, exec, REAL_HPAGE_SHIFT); in __set_pmd_acct() 214 tlb_batch_add_one(mm, addr + REAL_HPAGE_SIZE, exec, in __set_pmd_acct()
|
/linux-6.6.21/drivers/gpu/drm/i915/gem/ |
D | i915_gem_execbuffer.c | 42 struct drm_i915_gem_exec_object2 *exec; member 250 struct drm_i915_gem_exec_object2 *exec; /** ioctl execobj[] */ member 566 struct drm_i915_gem_exec_object2 *entry = &eb->exec[i]; in eb_add_vma() 570 ev->exec = entry; in eb_add_vma() 658 struct drm_i915_gem_exec_object2 *entry = ev->exec; in eb_reserve_vma() 950 vma = eb_lookup_vma(eb, eb->exec[i].handle); in eb_lookup_vmas() 956 err = eb_validate_vma(eb, &eb->exec[i], vma); in eb_lookup_vmas() 1023 struct drm_i915_gem_exec_object2 *entry = &eb->exec[i]; in eb_validate_vmas() 1053 eb_vma_misplaced(&eb->exec[i], vma, ev->flags)); in eb_validate_vmas() 1528 const struct drm_i915_gem_exec_object2 *entry = ev->exec; in eb_relocate_vma() [all …]
|
/linux-6.6.21/arch/arc/mm/ |
D | fault.c | 79 unsigned int write = 0, exec = 0, mask; in do_page_fault() local 107 exec = 1; in do_page_fault() 127 if (exec) in do_page_fault()
|