/linux-5.19.10/arch/arm64/include/asm/ |
D | preempt.h | 13 return READ_ONCE(current_thread_info()->preempt.count); in preempt_count() 19 WRITE_ONCE(current_thread_info()->preempt.count, pc); in preempt_count_set() 32 current_thread_info()->preempt.need_resched = 0; in set_preempt_need_resched() 37 current_thread_info()->preempt.need_resched = 1; in clear_preempt_need_resched() 42 return !current_thread_info()->preempt.need_resched; in test_preempt_need_resched() 47 u32 pc = READ_ONCE(current_thread_info()->preempt.count); in __preempt_count_add() 49 WRITE_ONCE(current_thread_info()->preempt.count, pc); in __preempt_count_add() 54 u32 pc = READ_ONCE(current_thread_info()->preempt.count); in __preempt_count_sub() 56 WRITE_ONCE(current_thread_info()->preempt.count, pc); in __preempt_count_sub() 65 WRITE_ONCE(ti->preempt.count, --pc); in __preempt_count_dec_and_test()
|
D | thread_info.h | 39 } preempt; member
|
/linux-5.19.10/tools/testing/selftests/ftrace/test.d/preemptirq/ |
D | irqsoff_tracer.tc | 36 modprobe $MOD test_mode=preempt delay=500000 || fail 38 modprobe $MOD test_mode=preempt delay=500000 || fail 40 modprobe $MOD test_mode=preempt delay=500000 || fail
|
/linux-5.19.10/include/linux/ |
D | rcupdate.h | 158 # define rcu_tasks_classic_qs(t, preempt) \ argument 160 if (!(preempt) && READ_ONCE((t)->rcu_tasks_holdout)) \ 166 # define rcu_tasks_classic_qs(t, preempt) do { } while (0) argument 184 #define rcu_tasks_qs(t, preempt) \ argument 186 rcu_tasks_classic_qs((t), (preempt)); \ 199 #define rcu_tasks_classic_qs(t, preempt) do { } while (0) argument 200 #define rcu_tasks_qs(t, preempt) do { } while (0) argument
|
D | rcutiny.h | 61 #define rcu_note_context_switch(preempt) \ argument 64 rcu_tasks_qs(current, (preempt)); \
|
D | rcutree.h | 21 void rcu_note_context_switch(bool preempt);
|
D | parport.h | 142 int (*preempt)(void *); member 362 int (*preempt)(void *); member
|
/linux-5.19.10/Documentation/locking/ |
D | preempt-locking.rst | 53 Note, some FPU functions are already explicitly preempt safe. For example, 77 preempt_enable() decrement the preempt counter 78 preempt_disable() increment the preempt counter 79 preempt_enable_no_resched() decrement, but do not immediately preempt 81 preempt_count() return the preempt counter 85 call to preempt_enable. The preempt statements define to nothing if 94 a reschedule if the preempt count is 0. A simple printk() might trigger a 121 This code is not preempt-safe, but see how easily we can fix it by simply
|
D | index.rst | 20 preempt-locking
|
/linux-5.19.10/drivers/gpu/drm/msm/adreno/ |
D | a5xx_preempt.c | 135 a5xx_gpu->preempt[ring->id]->wptr = get_wptr(ring); in a5xx_preempt_trigger() 208 a5xx_gpu->preempt[i]->wptr = 0; in a5xx_preempt_hw_init() 209 a5xx_gpu->preempt[i]->rptr = 0; in a5xx_preempt_hw_init() 210 a5xx_gpu->preempt[i]->rbase = gpu->rb[i]->iova; in a5xx_preempt_hw_init() 253 a5xx_gpu->preempt[ring->id] = ptr; in preempt_init_ring()
|
D | a5xx_gpu.h | 35 struct a5xx_preempt_record *preempt[MSM_GPU_MAX_RINGS]; member
|
/linux-5.19.10/samples/trace_events/ |
D | trace_custom_sched.h | 27 TP_PROTO(bool preempt, 32 TP_ARGS(preempt, prev, next, prev_state),
|
/linux-5.19.10/tools/lib/traceevent/Documentation/ |
D | libtraceevent-event_print.txt | 54 . : The preempt count is zero. 57 preempt count, when non zero, will prevent the kernel 58 from scheduling out the current task. The preempt count
|
/linux-5.19.10/include/trace/events/ |
D | sched.h | 190 static inline long __trace_sched_switch_state(bool preempt, in __trace_sched_switch_state() argument 204 if (preempt) in __trace_sched_switch_state() 224 TP_PROTO(bool preempt, 229 TP_ARGS(preempt, prev, next, prev_state), 245 __entry->prev_state = __trace_sched_switch_state(preempt, prev_state, prev);
|
/linux-5.19.10/Documentation/translations/zh_CN/locking/ |
D | index.rst | 29 * preempt-locking
|
/linux-5.19.10/kernel/rcu/ |
D | tree_plugin.h | 310 void rcu_note_context_switch(bool preempt) in rcu_note_context_switch() argument 318 …WARN_ONCE(!preempt && rcu_preempt_depth() > 0, "Voluntary context switch within RCU read-side crit… in rcu_note_context_switch() 357 rcu_tasks_qs(current, preempt); in rcu_note_context_switch() 892 void rcu_note_context_switch(bool preempt) in rcu_note_context_switch() argument 902 rcu_tasks_qs(current, preempt); in rcu_note_context_switch()
|
/linux-5.19.10/drivers/parport/ |
D | share.c | 707 if (!par_dev_cb->preempt || !par_dev_cb->wakeup) { in parport_register_dev_model() 748 par_dev->preempt = par_dev_cb->preempt; in parport_register_dev_model() 992 if (oldcad->preempt) { in parport_claim() 993 if (oldcad->preempt(oldcad->private)) in parport_claim()
|
/linux-5.19.10/arch/um/include/asm/ |
D | Kbuild | 23 generic-y += preempt.h
|
/linux-5.19.10/kernel/trace/ |
D | trace_sched_switch.c | 24 probe_sched_switch(void *ignore, bool preempt, in probe_sched_switch() argument
|
/linux-5.19.10/Documentation/power/ |
D | tricks.rst | 12 * turn off APIC and preempt
|
/linux-5.19.10/Documentation/trace/ |
D | ftrace.rst | 867 # || / _--=> preempt-depth 908 # latency: 259 us, #4/4, CPU#2 | (M:preempt VP:0, KP:0, SP:0 HP:0 #P:4) 920 # ||| / _--=> preempt-depth 985 preempt-depth: The level of preempt_disabled 1206 Shows the interrupt, preempt count, need resched data. 1380 # latency: 16 us, #4/4, CPU#0 | (M:preempt VP:0, KP:0, SP:0 HP:0 #P:4) 1392 # ||| / _--=> preempt-depth 1431 # latency: 71 us, #168/168, CPU#3 | (M:preempt VP:0, KP:0, SP:0 HP:0 #P:4) 1443 # ||| / _--=> preempt-depth 1520 # || / _--=> preempt-depth [all …]
|
D | ring-buffer-design.rst | 98 same time, nor can a reader preempt/interrupt another reader. A reader 99 cannot preempt/interrupt a writer, but it may read/consume from the 104 A writer can preempt a reader, but a reader cannot preempt a writer. 486 and writes only preempt in "stack" formation. 604 must spin, and this is why the reader cannot preempt the writer:: 657 As stated before, if enough writes preempt the first write, the
|
/linux-5.19.10/drivers/gpu/drm/i915/gt/ |
D | selftest_lrc.c | 614 bool preempt) in __live_lrc_gpr() argument 644 if (preempt) { in __live_lrc_gpr() 795 static int __lrc_timestamp(const struct lrc_timestamp *arg, bool preempt) in __lrc_timestamp() argument 811 if (preempt) { in __lrc_timestamp() 828 arg->engine->name, preempt ? "preempt" : "simple", in __lrc_timestamp() 836 arg->engine->name, preempt ? "preempt" : "simple", in __lrc_timestamp()
|
/linux-5.19.10/Documentation/RCU/Design/Expedited-Grace-Periods/ |
D | Expedited-Grace-Periods.rst | 13 There are two flavors of RCU (RCU-preempt and RCU-sched), with an earlier 38 RCU-preempt Expedited Grace Periods 41 ``CONFIG_PREEMPTION=y`` kernels implement RCU-preempt. 42 The overall flow of the handling of a given CPU by an RCU-preempt 121 As with RCU-preempt, RCU-sched's ``synchronize_rcu_expedited()`` ignores 132 NEED_RESCHED flag of the current task's thread flag and the CPU preempt 235 For RCU-preempt, there is no specific check for idle in the IPI handler 507 idle CPUs, at least in the common case. RCU-preempt and RCU-sched use
|
/linux-5.19.10/include/asm-generic/ |
D | Kbuild | 48 mandatory-y += preempt.h
|