Home
last modified time | relevance | path

Searched refs:per_cpu (Results 1 – 25 of 425) sorted by relevance

12345678910>>...17

/linux-6.6.21/arch/x86/xen/
Dsmp.c35 kfree(per_cpu(xen_resched_irq, cpu).name); in xen_smp_intr_free()
36 per_cpu(xen_resched_irq, cpu).name = NULL; in xen_smp_intr_free()
37 if (per_cpu(xen_resched_irq, cpu).irq >= 0) { in xen_smp_intr_free()
38 unbind_from_irqhandler(per_cpu(xen_resched_irq, cpu).irq, NULL); in xen_smp_intr_free()
39 per_cpu(xen_resched_irq, cpu).irq = -1; in xen_smp_intr_free()
41 kfree(per_cpu(xen_callfunc_irq, cpu).name); in xen_smp_intr_free()
42 per_cpu(xen_callfunc_irq, cpu).name = NULL; in xen_smp_intr_free()
43 if (per_cpu(xen_callfunc_irq, cpu).irq >= 0) { in xen_smp_intr_free()
44 unbind_from_irqhandler(per_cpu(xen_callfunc_irq, cpu).irq, NULL); in xen_smp_intr_free()
45 per_cpu(xen_callfunc_irq, cpu).irq = -1; in xen_smp_intr_free()
[all …]
Dspinlock.c25 int irq = per_cpu(lock_kicker_irq, cpu); in xen_qlock_kick()
74 WARN(per_cpu(lock_kicker_irq, cpu) >= 0, "spinlock on CPU%d exists on IRQ%d!\n", in xen_init_lock_cpu()
75 cpu, per_cpu(lock_kicker_irq, cpu)); in xen_init_lock_cpu()
78 per_cpu(irq_name, cpu) = name; in xen_init_lock_cpu()
88 per_cpu(lock_kicker_irq, cpu) = irq; in xen_init_lock_cpu()
101 kfree(per_cpu(irq_name, cpu)); in xen_uninit_lock_cpu()
102 per_cpu(irq_name, cpu) = NULL; in xen_uninit_lock_cpu()
107 irq = per_cpu(lock_kicker_irq, cpu); in xen_uninit_lock_cpu()
112 per_cpu(lock_kicker_irq, cpu) = -1; in xen_uninit_lock_cpu()
Dsmp_pv.c101 kfree(per_cpu(xen_irq_work, cpu).name); in xen_smp_intr_free_pv()
102 per_cpu(xen_irq_work, cpu).name = NULL; in xen_smp_intr_free_pv()
103 if (per_cpu(xen_irq_work, cpu).irq >= 0) { in xen_smp_intr_free_pv()
104 unbind_from_irqhandler(per_cpu(xen_irq_work, cpu).irq, NULL); in xen_smp_intr_free_pv()
105 per_cpu(xen_irq_work, cpu).irq = -1; in xen_smp_intr_free_pv()
108 kfree(per_cpu(xen_pmu_irq, cpu).name); in xen_smp_intr_free_pv()
109 per_cpu(xen_pmu_irq, cpu).name = NULL; in xen_smp_intr_free_pv()
110 if (per_cpu(xen_pmu_irq, cpu).irq >= 0) { in xen_smp_intr_free_pv()
111 unbind_from_irqhandler(per_cpu(xen_pmu_irq, cpu).irq, NULL); in xen_smp_intr_free_pv()
112 per_cpu(xen_pmu_irq, cpu).irq = -1; in xen_smp_intr_free_pv()
[all …]
/linux-6.6.21/arch/powerpc/kernel/
Dirq.c101 seq_printf(p, "%10u ", per_cpu(irq_stat, j).timer_irqs_event); in arch_show_interrupts()
106 seq_printf(p, "%10u ", per_cpu(irq_stat, j).broadcast_irqs_event); in arch_show_interrupts()
111 seq_printf(p, "%10u ", per_cpu(irq_stat, j).timer_irqs_others); in arch_show_interrupts()
116 seq_printf(p, "%10u ", per_cpu(irq_stat, j).spurious_irqs); in arch_show_interrupts()
121 seq_printf(p, "%10u ", per_cpu(irq_stat, j).pmu_irqs); in arch_show_interrupts()
126 seq_printf(p, "%10u ", per_cpu(irq_stat, j).mce_exceptions); in arch_show_interrupts()
140 seq_printf(p, "%10u ", per_cpu(irq_stat, j).sreset_irqs); in arch_show_interrupts()
146 seq_printf(p, "%10u ", per_cpu(irq_stat, j).soft_nmi_irqs); in arch_show_interrupts()
154 seq_printf(p, "%10u ", per_cpu(irq_stat, j).doorbell_irqs); in arch_show_interrupts()
167 u64 sum = per_cpu(irq_stat, cpu).timer_irqs_event; in arch_irq_stat_cpu()
[all …]
/linux-6.6.21/drivers/cpufreq/
Dspeedstep-centrino.c261 per_cpu(centrino_model, policy->cpu) = model; in centrino_cpu_init_table()
296 if ((per_cpu(centrino_cpu, cpu) == &cpu_ids[CPU_BANIAS]) || in extract_clock()
297 (per_cpu(centrino_cpu, cpu) == &cpu_ids[CPU_DOTHAN_A1]) || in extract_clock()
298 (per_cpu(centrino_cpu, cpu) == &cpu_ids[CPU_DOTHAN_B0])) { in extract_clock()
303 if ((!per_cpu(centrino_model, cpu)) || in extract_clock()
304 (!per_cpu(centrino_model, cpu)->op_points)) in extract_clock()
309 per_cpu(centrino_model, cpu)->op_points[i].frequency in extract_clock()
312 if (msr == per_cpu(centrino_model, cpu)->op_points[i].driver_data) in extract_clock()
313 return per_cpu(centrino_model, cpu)-> in extract_clock()
317 return per_cpu(centrino_model, cpu)->op_points[i-1].frequency; in extract_clock()
[all …]
Dvexpress-spc-cpufreq.c77 cpu_freq = per_cpu(cpu_last_req_freq, j); in find_cluster_maxfreq()
79 if (cluster == per_cpu(physical_cluster, j) && in find_cluster_maxfreq()
89 u32 cur_cluster = per_cpu(physical_cluster, cpu); in clk_get_cpu_rate()
102 return per_cpu(cpu_last_req_freq, cpu); in ve_spc_cpufreq_get_rate()
117 prev_rate = per_cpu(cpu_last_req_freq, cpu); in ve_spc_cpufreq_set_rate()
118 per_cpu(cpu_last_req_freq, cpu) = rate; in ve_spc_cpufreq_set_rate()
119 per_cpu(physical_cluster, cpu) = new_cluster; in ve_spc_cpufreq_set_rate()
143 per_cpu(cpu_last_req_freq, cpu) = prev_rate; in ve_spc_cpufreq_set_rate()
144 per_cpu(physical_cluster, cpu) = old_cluster; in ve_spc_cpufreq_set_rate()
184 new_cluster = actual_cluster = per_cpu(physical_cluster, cpu); in ve_spc_cpufreq_set_target()
[all …]
/linux-6.6.21/drivers/perf/
Darm_pmu_acpi.c220 per_cpu(pmu_irqs, cpu) = irq; in arm_pmu_acpi_parse_irqs()
230 irq = per_cpu(pmu_irqs, cpu); in arm_pmu_acpi_parse_irqs()
241 if (per_cpu(pmu_irqs, irq_cpu) == irq) in arm_pmu_acpi_parse_irqs()
242 per_cpu(pmu_irqs, irq_cpu) = 0; in arm_pmu_acpi_parse_irqs()
256 pmu = per_cpu(probed_pmus, cpu); in arm_pmu_acpi_find_pmu()
279 int other_irq = per_cpu(hw_events->irq, cpu); in pmu_irq_matches()
298 int irq = per_cpu(pmu_irqs, cpu); in arm_pmu_acpi_associate_pmu_cpu()
300 per_cpu(probed_pmus, cpu) = pmu; in arm_pmu_acpi_associate_pmu_cpu()
305 per_cpu(hw_events->irq, cpu) = irq; in arm_pmu_acpi_associate_pmu_cpu()
325 if (per_cpu(probed_pmus, cpu)) in arm_pmu_acpi_cpu_starting()
[all …]
/linux-6.6.21/arch/arm/mm/
Dcontext.c67 asid = per_cpu(active_asids, cpu).counter; in a15_erratum_get_cpumask()
69 asid = per_cpu(reserved_asids, cpu); in a15_erratum_get_cpumask()
144 asid = atomic64_xchg(&per_cpu(active_asids, i), 0); in flush_context()
153 asid = per_cpu(reserved_asids, i); in flush_context()
155 per_cpu(reserved_asids, i) = asid; in flush_context()
180 if (per_cpu(reserved_asids, cpu) == asid) { in check_update_reserved_asid()
182 per_cpu(reserved_asids, cpu) = newasid; in check_update_reserved_asid()
254 && atomic64_xchg(&per_cpu(active_asids, cpu), asid)) in check_and_switch_context()
270 atomic64_set(&per_cpu(active_asids, cpu), asid); in check_and_switch_context()
/linux-6.6.21/kernel/
Dprofile.c175 per_cpu(cpu_profile_flip, cpu) = !per_cpu(cpu_profile_flip, cpu); in __profile_flip_buffers()
183 j = per_cpu(cpu_profile_flip, get_cpu()); in profile_flip_buffers()
187 struct profile_hit *hits = per_cpu(cpu_profile_hits, cpu)[j]; in profile_flip_buffers()
206 i = per_cpu(cpu_profile_flip, get_cpu()); in profile_discard_flip_buffers()
210 struct profile_hit *hits = per_cpu(cpu_profile_hits, cpu)[i]; in profile_discard_flip_buffers()
226 hits = per_cpu(cpu_profile_hits, cpu)[per_cpu(cpu_profile_flip, cpu)]; in do_profile_hits()
274 if (per_cpu(cpu_profile_hits, cpu)[i]) { in profile_dead_cpu()
275 page = virt_to_page(per_cpu(cpu_profile_hits, cpu)[i]); in profile_dead_cpu()
276 per_cpu(cpu_profile_hits, cpu)[i] = NULL; in profile_dead_cpu()
288 per_cpu(cpu_profile_flip, cpu) = 0; in profile_prepare_cpu()
[all …]
Dsoftirq.c898 per_cpu(tasklet_vec, cpu).tail = in softirq_init()
899 &per_cpu(tasklet_vec, cpu).head; in softirq_init()
900 per_cpu(tasklet_hi_vec, cpu).tail = in softirq_init()
901 &per_cpu(tasklet_hi_vec, cpu).head; in softirq_init()
936 if (&per_cpu(tasklet_vec, cpu).head != per_cpu(tasklet_vec, cpu).tail) { in takeover_tasklets()
937 *__this_cpu_read(tasklet_vec.tail) = per_cpu(tasklet_vec, cpu).head; in takeover_tasklets()
938 __this_cpu_write(tasklet_vec.tail, per_cpu(tasklet_vec, cpu).tail); in takeover_tasklets()
939 per_cpu(tasklet_vec, cpu).head = NULL; in takeover_tasklets()
940 per_cpu(tasklet_vec, cpu).tail = &per_cpu(tasklet_vec, cpu).head; in takeover_tasklets()
944 if (&per_cpu(tasklet_hi_vec, cpu).head != per_cpu(tasklet_hi_vec, cpu).tail) { in takeover_tasklets()
[all …]
/linux-6.6.21/arch/ia64/mm/
Dtlb.c101 per_cpu(ia64_need_tlb_flush, i) = 1; in wrap_mmu_context()
397 per_cpu(ia64_tr_num, cpu) = 8; in ia64_tlb_init()
400 per_cpu(ia64_tr_num, cpu) = vm_info_1.pal_vm_info_1_s.max_itr_entry+1; in ia64_tlb_init()
401 if (per_cpu(ia64_tr_num, cpu) > in ia64_tlb_init()
403 per_cpu(ia64_tr_num, cpu) = in ia64_tlb_init()
405 if (per_cpu(ia64_tr_num, cpu) > IA64_TR_ALLOC_MAX) { in ia64_tlb_init()
407 per_cpu(ia64_tr_num, cpu) = IA64_TR_ALLOC_MAX; in ia64_tlb_init()
472 for (i = IA64_TR_ALLOC_BASE; i <= per_cpu(ia64_tr_used, cpu); in ia64_itr_entry()
484 for (i = IA64_TR_ALLOC_BASE; i <= per_cpu(ia64_tr_used, cpu); in ia64_itr_entry()
495 for (i = IA64_TR_ALLOC_BASE; i < per_cpu(ia64_tr_num, cpu); i++) { in ia64_itr_entry()
[all …]
/linux-6.6.21/arch/x86/kernel/apic/
Dx2apic_cluster.c58 struct cpumask *cmsk = per_cpu(cluster_masks, cpu); in __x2apic_send_IPI_mask()
110 struct cpumask **cpu_cmsk = &per_cpu(cluster_masks, cpu_i); in prefill_clustermask()
135 if (per_cpu(cluster_masks, cpu)) in alloc_clustermask()
150 cmsk = per_cpu(cluster_masks, cpu_i); in alloc_clustermask()
156 per_cpu(cluster_masks, cpu) = cmsk; in alloc_clustermask()
170 per_cpu(cluster_masks, cpu) = cmsk; in alloc_clustermask()
186 if (!zalloc_cpumask_var(&per_cpu(ipi_mask, cpu), GFP_KERNEL)) in x2apic_prepare_cpu()
193 struct cpumask *cmsk = per_cpu(cluster_masks, dead_cpu); in x2apic_dead_cpu()
197 free_cpumask_var(per_cpu(ipi_mask, dead_cpu)); in x2apic_dead_cpu()
/linux-6.6.21/arch/parisc/kernel/
Dirq.c75 per_cpu(local_ack_eiem, cpu) &= ~mask; in cpu_ack_irq()
78 set_eiem(cpu_eiem & per_cpu(local_ack_eiem, cpu)); in cpu_ack_irq()
90 per_cpu(local_ack_eiem, cpu) |= mask; in cpu_eoi_irq()
93 set_eiem(cpu_eiem & per_cpu(local_ack_eiem, cpu)); in cpu_eoi_irq()
126 #define irq_stats(x) (&per_cpu(irq_stat, x))
318 return per_cpu(cpu_data, cpu).txn_addr; in txn_affinity_addr()
330 (!per_cpu(cpu_data, next_cpu).txn_addr || in txn_alloc_addr()
403 stack_start = (unsigned long) &per_cpu(irq_stack_union, cpu).stack; in stack_overflow_check()
406 last_usage = &per_cpu(irq_stat.irq_stack_usage, cpu); in stack_overflow_check()
422 last_usage = &per_cpu(irq_stat.kernel_stack_usage, cpu); in stack_overflow_check()
[all …]
Dtopology.c40 per_cpu(cpu_devices, cpuid).hotpluggable = 1; in store_cpu_topology()
42 if (register_cpu(&per_cpu(cpu_devices, cpuid), cpuid)) in store_cpu_topology()
49 p = &per_cpu(cpu_data, cpuid); in store_cpu_topology()
51 const struct cpuinfo_parisc *cpuinfo = &per_cpu(cpu_data, cpu); in store_cpu_topology()
Dsmp.c123 struct cpuinfo_parisc *p = &per_cpu(cpu_data, this_cpu); in ipi_interrupt()
128 spinlock_t *lock = &per_cpu(ipi_lock, this_cpu); in ipi_interrupt()
199 struct cpuinfo_parisc *p = &per_cpu(cpu_data, cpu); in ipi_send()
200 spinlock_t *lock = &per_cpu(ipi_lock, cpu); in ipi_send()
335 const struct cpuinfo_parisc *p = &per_cpu(cpu_data, cpuid); in smp_boot_one_cpu()
342 memset(&per_cpu(irq_stat, cpuid), 0, sizeof(irq_cpustat_t)); in smp_boot_one_cpu()
407 int bootstrap_processor = per_cpu(cpu_data, 0).cpuid; in smp_prepare_boot_cpu()
427 spin_lock_init(&per_cpu(ipi_lock, cpu)); in smp_prepare_cpus()
/linux-6.6.21/arch/mips/kernel/
Dmips-cpc.c77 spin_lock_init(&per_cpu(cpc_core_lock, cpu)); in mips_cpc_probe()
100 spin_lock_irqsave(&per_cpu(cpc_core_lock, curr_core), in mips_cpc_lock_other()
101 per_cpu(cpc_core_lock_flags, curr_core)); in mips_cpc_lock_other()
120 spin_unlock_irqrestore(&per_cpu(cpc_core_lock, curr_core), in mips_cpc_unlock_other()
121 per_cpu(cpc_core_lock_flags, curr_core)); in mips_cpc_unlock_other()
/linux-6.6.21/arch/x86/kvm/vmx/
Dposted_intr.c92 raw_spin_lock(&per_cpu(wakeup_vcpus_on_cpu_lock, vcpu->cpu)); in vmx_vcpu_pi_load()
94 raw_spin_unlock(&per_cpu(wakeup_vcpus_on_cpu_lock, vcpu->cpu)); in vmx_vcpu_pi_load()
155 raw_spin_lock(&per_cpu(wakeup_vcpus_on_cpu_lock, vcpu->cpu)); in pi_enable_wakeup_handler()
157 &per_cpu(wakeup_vcpus_on_cpu, vcpu->cpu)); in pi_enable_wakeup_handler()
158 raw_spin_unlock(&per_cpu(wakeup_vcpus_on_cpu_lock, vcpu->cpu)); in pi_enable_wakeup_handler()
221 struct list_head *wakeup_list = &per_cpu(wakeup_vcpus_on_cpu, cpu); in pi_wakeup_handler()
222 raw_spinlock_t *spinlock = &per_cpu(wakeup_vcpus_on_cpu_lock, cpu); in pi_wakeup_handler()
236 INIT_LIST_HEAD(&per_cpu(wakeup_vcpus_on_cpu, cpu)); in pi_init_cpu()
237 raw_spin_lock_init(&per_cpu(wakeup_vcpus_on_cpu_lock, cpu)); in pi_init_cpu()
/linux-6.6.21/arch/x86/mm/
Dcpu_entry_area.c25 return per_cpu(_cea_offset, cpu); in cea_offset()
35 per_cpu(_cea_offset, i) = i; in init_cea_offsets()
56 per_cpu(_cea_offset, i) = cea; in init_cea_offsets()
117 cea_map_percpu_pages(cea, &per_cpu(cpu_debug_store, cpu), npages, in percpu_setup_debug_store()
147 per_cpu(cea_exception_stacks, cpu) = &cea->estacks; in percpu_setup_exception_stacks()
172 &per_cpu(doublefault_stack, cpu), 1, PAGE_KERNEL); in percpu_setup_exception_stacks()
233 cea_map_percpu_pages(&cea->tss, &per_cpu(cpu_tss_rw, cpu), in setup_cpu_entry_area()
237 per_cpu(cpu_entry_area, cpu) = cea; in setup_cpu_entry_area()
/linux-6.6.21/arch/riscv/mm/
Dcontext.c52 if (per_cpu(reserved_context, cpu) == cntx) { in check_update_reserved_context()
54 per_cpu(reserved_context, cpu) = newcntx; in check_update_reserved_context()
74 cntx = atomic_long_xchg_relaxed(&per_cpu(active_context, i), 0); in __flush_context()
82 cntx = per_cpu(reserved_context, i); in __flush_context()
85 per_cpu(reserved_context, i) = cntx; in __flush_context()
169 old_active_cntx = atomic_long_read(&per_cpu(active_context, cpu)); in set_mm_asid()
172 atomic_long_cmpxchg_relaxed(&per_cpu(active_context, cpu), in set_mm_asid()
188 atomic_long_set(&per_cpu(active_context, cpu), cntx); in set_mm_asid()
/linux-6.6.21/drivers/acpi/
Dcppc_acpi.c333 struct cpc_desc *desc = per_cpu(cpc_desc_ptr, i); in send_pcc_cmd()
431 cpc_ptr = per_cpu(cpc_desc_ptr, cpu); in acpi_cpc_valid()
447 cpc_ptr = per_cpu(cpc_desc_ptr, cpu); in cppc_allow_fast_switch()
476 cpc_ptr = per_cpu(cpc_desc_ptr, cpu); in acpi_get_psd_map()
498 match_cpc_ptr = per_cpu(cpc_desc_ptr, i); in acpi_get_psd_map()
831 per_cpu(cpu_pcc_subspace_idx, pr->id) = pcc_subspace_id; in acpi_cppc_processor_probe()
873 per_cpu(cpc_desc_ptr, pr->id) = cpc_ptr; in acpi_cppc_processor_probe()
878 per_cpu(cpc_desc_ptr, pr->id) = NULL; in acpi_cppc_processor_probe()
915 int pcc_ss_id = per_cpu(cpu_pcc_subspace_idx, pr->id); in acpi_cppc_processor_exit()
928 cpc_ptr = per_cpu(cpc_desc_ptr, pr->id); in acpi_cppc_processor_exit()
[all …]
/linux-6.6.21/arch/arm64/kernel/
Dsdei.c57 p = per_cpu(*ptr, cpu); in _free_sdei_stack()
59 per_cpu(*ptr, cpu) = NULL; in _free_sdei_stack()
84 per_cpu(*ptr, cpu) = p; in _init_sdei_stack()
116 s = per_cpu(*ptr, cpu); in _free_sdei_scs()
118 per_cpu(*ptr, cpu) = NULL; in _free_sdei_scs()
140 per_cpu(*ptr, cpu) = s; in _init_sdei_scs()
/linux-6.6.21/arch/riscv/kernel/
Dirq.c47 per_cpu(irq_stack_ptr, cpu) = p; in init_irq_stacks()
59 per_cpu(irq_stack_ptr, cpu) = per_cpu(irq_stack, cpu); in init_irq_stacks()
68 ulong *sp = per_cpu(irq_stack_ptr, smp_processor_id()) in do_softirq_own_stack()
/linux-6.6.21/drivers/xen/events/
Devents_2l.c52 clear_bit(evtchn, BM(per_cpu(cpu_evtchn_mask, cpu))); in evtchn_2l_remove()
58 clear_bit(evtchn, BM(per_cpu(cpu_evtchn_mask, old_cpu))); in evtchn_2l_bind_to_cpu()
59 set_bit(evtchn, BM(per_cpu(cpu_evtchn_mask, cpu))); in evtchn_2l_bind_to_cpu()
152 per_cpu(cpu_evtchn_mask, cpu)[idx] & in active_evtchns()
268 xen_ulong_t *cpu_evtchn = per_cpu(cpu_evtchn_mask, cpu); in xen_debug_interrupt()
280 v = per_cpu(xen_vcpu, i); in xen_debug_interrupt()
289 v = per_cpu(xen_vcpu, cpu); in xen_debug_interrupt()
353 memset(per_cpu(cpu_evtchn_mask, i), 0, sizeof(xen_ulong_t) * in evtchn_2l_resume()
359 memset(per_cpu(cpu_evtchn_mask, cpu), 0, sizeof(xen_ulong_t) * in evtchn_2l_percpu_deinit()
/linux-6.6.21/arch/powerpc/include/asm/
Dsmp.h115 return per_cpu(cpu_sibling_map, cpu); in cpu_sibling_mask()
120 return per_cpu(cpu_core_map, cpu); in cpu_core_mask()
125 return per_cpu(cpu_l2_cache_map, cpu); in cpu_l2_cache_mask()
130 return per_cpu(cpu_smallcore_map, cpu); in cpu_smallcore_mask()
144 return per_cpu(cpu_smallcore_map, cpu); in cpu_smt_mask()
146 return per_cpu(cpu_sibling_map, cpu); in cpu_smt_mask()
/linux-6.6.21/arch/x86/kernel/
Dsetup_percpu.c171 per_cpu(this_cpu_off, cpu) = per_cpu_offset(cpu); in setup_per_cpu_areas()
172 per_cpu(pcpu_hot.cpu_number, cpu) = cpu; in setup_per_cpu_areas()
182 per_cpu(x86_cpu_to_apicid, cpu) = in setup_per_cpu_areas()
184 per_cpu(x86_cpu_to_acpiid, cpu) = in setup_per_cpu_areas()
188 per_cpu(x86_cpu_to_node_map, cpu) = in setup_per_cpu_areas()

12345678910>>...17