Home
last modified time | relevance | path

Searched refs:this_cpu (Results 1 – 25 of 64) sorted by relevance

123

/linux-6.1.9/net/netfilter/
Dnft_counter.c34 struct nft_counter *this_cpu; in nft_counter_do_eval() local
38 this_cpu = this_cpu_ptr(priv->counter); in nft_counter_do_eval()
43 this_cpu->bytes += pkt->skb->len; in nft_counter_do_eval()
44 this_cpu->packets++; in nft_counter_do_eval()
63 struct nft_counter *this_cpu; in nft_counter_do_init() local
70 this_cpu = this_cpu_ptr(cpu_stats); in nft_counter_do_init()
72 this_cpu->packets = in nft_counter_do_init()
76 this_cpu->bytes = in nft_counter_do_init()
109 struct nft_counter *this_cpu; in nft_counter_reset() local
112 this_cpu = this_cpu_ptr(priv->counter); in nft_counter_reset()
[all …]
/linux-6.1.9/arch/parisc/kernel/
Dsmp.c122 int this_cpu = smp_processor_id(); in ipi_interrupt() local
123 struct cpuinfo_parisc *p = &per_cpu(cpu_data, this_cpu); in ipi_interrupt()
128 spinlock_t *lock = &per_cpu(ipi_lock, this_cpu); in ipi_interrupt()
146 smp_debug(100, KERN_DEBUG "CPU%d IPI_NOP\n", this_cpu); in ipi_interrupt()
150 smp_debug(100, KERN_DEBUG "CPU%d IPI_RESCHEDULE\n", this_cpu); in ipi_interrupt()
156 smp_debug(100, KERN_DEBUG "CPU%d IPI_CALL_FUNC\n", this_cpu); in ipi_interrupt()
162 smp_debug(100, KERN_DEBUG "CPU%d IPI_CPU_START\n", this_cpu); in ipi_interrupt()
166 smp_debug(100, KERN_DEBUG "CPU%d IPI_CPU_STOP\n", this_cpu); in ipi_interrupt()
171 smp_debug(100, KERN_DEBUG "CPU%d is alive!\n", this_cpu); in ipi_interrupt()
175 smp_debug(100, KERN_DEBUG "CPU%d ENTER_KGDB\n", this_cpu); in ipi_interrupt()
[all …]
/linux-6.1.9/lib/
Dnmi_backtrace.c40 int i, this_cpu = get_cpu(); in nmi_trigger_cpumask_backtrace() local
53 cpumask_clear_cpu(this_cpu, to_cpumask(backtrace_mask)); in nmi_trigger_cpumask_backtrace()
61 if (cpumask_test_cpu(this_cpu, to_cpumask(backtrace_mask))) in nmi_trigger_cpumask_backtrace()
66 this_cpu, nr_cpumask_bits, to_cpumask(backtrace_mask)); in nmi_trigger_cpumask_backtrace()
Dsmp_processor_id.c14 int this_cpu = raw_smp_processor_id(); in check_preemption_disabled() local
55 return this_cpu; in check_preemption_disabled()
/linux-6.1.9/arch/sparc/kernel/
Dprocess_64.c191 int this_cpu) in __global_reg_self() argument
197 rp = &global_cpu_snapshot[this_cpu].reg; in __global_reg_self()
243 int this_cpu, cpu; in arch_trigger_cpumask_backtrace() local
250 this_cpu = raw_smp_processor_id(); in arch_trigger_cpumask_backtrace()
254 if (cpumask_test_cpu(this_cpu, mask) && !exclude_self) in arch_trigger_cpumask_backtrace()
255 __global_reg_self(tp, regs, this_cpu); in arch_trigger_cpumask_backtrace()
262 if (exclude_self && cpu == this_cpu) in arch_trigger_cpumask_backtrace()
271 (cpu == this_cpu ? '*' : ' '), cpu, in arch_trigger_cpumask_backtrace()
308 static void __global_pmu_self(int this_cpu) in __global_pmu_self() argument
316 pp = &global_cpu_snapshot[this_cpu].pmu; in __global_pmu_self()
[all …]
Dsmp_64.c647 int this_cpu, tot_cpus, prev_sent, i, rem; in hypervisor_xcall_deliver() local
657 this_cpu = smp_processor_id(); in hypervisor_xcall_deliver()
764 this_cpu, ecpuerror_id - 1); in hypervisor_xcall_deliver()
767 this_cpu, enocpu_id - 1); in hypervisor_xcall_deliver()
774 this_cpu, tot_cpus, tb->cpu_list_pa, tb->cpu_mondo_block_pa); in hypervisor_xcall_deliver()
780 this_cpu, first_cpu, (tot_retries + retries), tot_cpus); in hypervisor_xcall_deliver()
789 int this_cpu, i, cnt; in xcall_deliver() local
806 this_cpu = smp_processor_id(); in xcall_deliver()
807 tb = &trap_block[this_cpu]; in xcall_deliver()
820 if (i == this_cpu || !cpu_online(i)) in xcall_deliver()
[all …]
/linux-6.1.9/kernel/
Dsmp.c418 unsigned int this_cpu = smp_processor_id(); in __smp_call_single_queue_debug() local
423 cfd_seq_store(pcpu->seq_queue, this_cpu, cpu, CFD_SEQ_QUEUE); in __smp_call_single_queue_debug()
425 cfd_seq_store(pcpu->seq_ipi, this_cpu, cpu, CFD_SEQ_IPI); in __smp_call_single_queue_debug()
426 cfd_seq_store(seq->ping, this_cpu, cpu, CFD_SEQ_PING); in __smp_call_single_queue_debug()
428 cfd_seq_store(seq->pinged, this_cpu, cpu, CFD_SEQ_PINGED); in __smp_call_single_queue_debug()
430 cfd_seq_store(pcpu->seq_noipi, this_cpu, cpu, CFD_SEQ_NOIPI); in __smp_call_single_queue_debug()
734 int this_cpu; in smp_call_function_single() local
741 this_cpu = get_cpu(); in smp_call_function_single()
749 WARN_ON_ONCE(cpu_online(this_cpu) && irqs_disabled() in smp_call_function_single()
887 int cpu, last_cpu, this_cpu = smp_processor_id(); in smp_call_function_many_cond() local
[all …]
/linux-6.1.9/arch/alpha/kernel/
Dsmp.c517 int this_cpu = smp_processor_id(); in handle_ipi() local
518 unsigned long *pending_ipis = &ipi_data[this_cpu].bits; in handle_ipi()
523 this_cpu, *pending_ipis, regs->pc)); in handle_ipi()
550 this_cpu, which); in handle_ipi()
558 cpu_data[this_cpu].ipi_count++; in handle_ipi()
646 int cpu, this_cpu = smp_processor_id(); in flush_tlb_mm() local
648 if (!cpu_online(cpu) || cpu == this_cpu) in flush_tlb_mm()
693 int cpu, this_cpu = smp_processor_id(); in flush_tlb_page() local
695 if (!cpu_online(cpu) || cpu == this_cpu) in flush_tlb_page()
747 int cpu, this_cpu = smp_processor_id(); in flush_icache_user_page() local
[all …]
/linux-6.1.9/tools/power/cpupower/utils/idle_monitor/
Dcpuidle_sysfs.c145 int this_cpu; in cpuidle_register() local
147 this_cpu = sched_getcpu(); in cpuidle_register()
150 cpuidle_sysfs_monitor.hw_states_num = cpuidle_state_count(this_cpu); in cpuidle_register()
156 tmp = cpuidle_state_name(this_cpu, num); in cpuidle_register()
165 tmp = cpuidle_state_desc(this_cpu, num); in cpuidle_register()
/linux-6.1.9/arch/x86/lib/
Dmsr-smp.c12 int this_cpu = raw_smp_processor_id(); in __rdmsr_on_cpu() local
15 reg = per_cpu_ptr(rv->msrs, this_cpu); in __rdmsr_on_cpu()
26 int this_cpu = raw_smp_processor_id(); in __wrmsr_on_cpu() local
29 reg = per_cpu_ptr(rv->msrs, this_cpu); in __wrmsr_on_cpu()
104 int this_cpu; in __rwmsr_on_cpus() local
111 this_cpu = get_cpu(); in __rwmsr_on_cpus()
113 if (cpumask_test_cpu(this_cpu, mask)) in __rwmsr_on_cpus()
/linux-6.1.9/kernel/trace/
Dtrace_clock.c97 int this_cpu; in trace_clock_global() local
102 this_cpu = raw_smp_processor_id(); in trace_clock_global()
116 now = sched_clock_cpu(this_cpu); in trace_clock_global()
/linux-6.1.9/Documentation/arm/
Dvlocks.rst36 bool vlock_trylock(int this_cpu)
39 currently_voting[this_cpu] = 1;
42 currently_voting[this_cpu] = 0;
47 last_vote = this_cpu;
48 currently_voting[this_cpu] = 0;
57 if (last_vote == this_cpu)
100 my_town = towns[(this_cpu >> 4) & 0xf];
101 I_won = vlock_trylock(my_town, this_cpu & 0xf);
104 my_state = states[(this_cpu >> 8) & 0xf];
105 I_won = vlock_lock(my_state, this_cpu & 0xf));
[all …]
/linux-6.1.9/Documentation/core-api/
Dthis_cpu_ops.rst2 this_cpu operations
8 this_cpu operations are a way of optimizing access to per cpu
14 this_cpu operations add a per cpu variable offset to the processor
41 The main use of the this_cpu operations has been to optimize counter
44 The following this_cpu() operations with implied preemption protection
65 Inner working of this_cpu operations
91 Consider the following this_cpu operation::
115 after the this_cpu instruction is executed. In general this means that
167 cpu variable. Most this_cpu operations take a cpu variable.
207 if we do not make use of this_cpu ops later to manipulate fields::
[all …]
/linux-6.1.9/arch/arm/common/
DbL_switcher.c149 unsigned int mpidr, this_cpu, that_cpu; in bL_switch_to() local
155 this_cpu = smp_processor_id(); in bL_switch_to()
159 BUG_ON(cpu_logical_map(this_cpu) != ob_mpidr); in bL_switch_to()
164 that_cpu = bL_switcher_cpu_pairing[this_cpu]; in bL_switch_to()
170 this_cpu, ob_mpidr, ib_mpidr); in bL_switch_to()
172 this_cpu = smp_processor_id(); in bL_switch_to()
180 ipi_nr = register_ipi_completion(&inbound_alive, this_cpu); in bL_switch_to()
227 cpu_logical_map(this_cpu) = ib_mpidr; in bL_switch_to()
237 pr_debug("after switch: CPU %d MPIDR %#x\n", this_cpu, mpidr); in bL_switch_to()
/linux-6.1.9/init/
Dcalibrate.c279 int this_cpu = smp_processor_id(); in calibrate_delay() local
281 if (per_cpu(cpu_loops_per_jiffy, this_cpu)) { in calibrate_delay()
282 lpj = per_cpu(cpu_loops_per_jiffy, this_cpu); in calibrate_delay()
306 per_cpu(cpu_loops_per_jiffy, this_cpu) = lpj; in calibrate_delay()
/linux-6.1.9/arch/x86/kernel/apic/
Dipi.c206 unsigned int this_cpu = smp_processor_id(); in default_send_IPI_mask_allbutself_phys() local
214 if (query_cpu == this_cpu) in default_send_IPI_mask_allbutself_phys()
272 unsigned int this_cpu = smp_processor_id(); in default_send_IPI_mask_allbutself_logical() local
278 if (query_cpu == this_cpu) in default_send_IPI_mask_allbutself_logical()
Dapic_numachip.c130 unsigned int this_cpu = smp_processor_id(); in numachip_send_IPI_mask_allbutself() local
134 if (cpu != this_cpu) in numachip_send_IPI_mask_allbutself()
141 unsigned int this_cpu = smp_processor_id(); in numachip_send_IPI_allbutself() local
145 if (cpu != this_cpu) in numachip_send_IPI_allbutself()
Dx2apic_phys.c55 unsigned long this_cpu; in __x2apic_send_IPI_mask() local
63 this_cpu = smp_processor_id(); in __x2apic_send_IPI_mask()
65 if (apic_dest == APIC_DEST_ALLBUT && this_cpu == query_cpu) in __x2apic_send_IPI_mask()
/linux-6.1.9/drivers/thermal/intel/
Dtherm_throt.c243 unsigned int i, avg, this_cpu = smp_processor_id(); in throttle_active_work() local
253 this_cpu, in throttle_active_work()
289 this_cpu, in throttle_active_work()
299 schedule_delayed_work_on(this_cpu, &state->therm_work, THERM_THROT_POLL_INTERVAL); in throttle_active_work()
316 unsigned int this_cpu = smp_processor_id(); in therm_throt_process() local
319 struct thermal_state *pstate = &per_cpu(thermal_state, this_cpu); in therm_throt_process()
363 schedule_delayed_work_on(this_cpu, &state->therm_work, THERM_THROT_POLL_INTERVAL); in therm_throt_process()
378 unsigned int this_cpu = smp_processor_id(); in thresh_event_valid() local
379 struct thermal_state *pstate = &per_cpu(thermal_state, this_cpu); in thresh_event_valid()
/linux-6.1.9/drivers/cpuidle/
Dcpuidle-ux500.c26 int this_cpu = smp_processor_id(); in ux500_enter_idle() local
49 if (!prcmu_is_cpu_in_wfi(this_cpu ? 0 : 1)) in ux500_enter_idle()
/linux-6.1.9/arch/arm/kernel/
Dsmp_tlb.c166 int this_cpu; in broadcast_tlb_mm_a15_erratum() local
172 this_cpu = get_cpu(); in broadcast_tlb_mm_a15_erratum()
173 a15_erratum_get_cpumask(this_cpu, mm, &mask); in broadcast_tlb_mm_a15_erratum()
/linux-6.1.9/arch/s390/mm/
Dmaccess.c167 int this_cpu, cpu; in xlate_dev_mem_ptr() local
170 this_cpu = get_cpu(); in xlate_dev_mem_ptr()
185 } else if (cpu == this_cpu) { in xlate_dev_mem_ptr()
/linux-6.1.9/arch/x86/kernel/cpu/
Dcommon.c202 static const struct cpu_dev *this_cpu = &default_cpu; variable
686 if (!this_cpu) in table_lookup_model()
689 info = this_cpu->legacy_models; in table_lookup_model()
824 if (this_cpu->legacy_cache_size) in cpu_detect_cache_sizes()
825 l2size = this_cpu->legacy_cache_size(c, l2size); in cpu_detect_cache_sizes()
848 if (this_cpu->c_detect_tlb) in cpu_detect_tlb()
849 this_cpu->c_detect_tlb(c); in cpu_detect_tlb()
918 this_cpu = cpu_devs[i]; in get_cpu_vendor()
919 c->x86_vendor = this_cpu->c_x86_vendor; in get_cpu_vendor()
928 this_cpu = &default_cpu; in get_cpu_vendor()
[all …]
/linux-6.1.9/arch/s390/kernel/
Dprocessor.c69 int cpu, this_cpu; in stop_machine_yield() local
71 this_cpu = smp_processor_id(); in stop_machine_yield()
74 cpu = cpumask_next_wrap(this_cpu, cpumask, this_cpu, false); in stop_machine_yield()
Dmachine_kexec.c79 int this_cpu, cpu; in __machine_kdump() local
83 this_cpu = smp_find_processor_id(stap()); in __machine_kdump()
85 if (cpu == this_cpu) in __machine_kdump()

123