Home
last modified time | relevance | path

Searched refs:cpu_mask (Results 1 – 25 of 72) sorted by relevance

123

/linux-6.1.9/arch/riscv/kernel/
Dsbi.c20 static int (*__sbi_send_ipi)(const struct cpumask *cpu_mask) __ro_after_init;
21 static int (*__sbi_rfence)(int fid, const struct cpumask *cpu_mask,
71 static unsigned long __sbi_v01_cpumask_to_hartmask(const struct cpumask *cpu_mask) in __sbi_v01_cpumask_to_hartmask() argument
83 for_each_cpu(cpuid, cpu_mask) { in __sbi_v01_cpumask_to_hartmask()
160 static int __sbi_send_ipi_v01(const struct cpumask *cpu_mask) in __sbi_send_ipi_v01() argument
164 if (!cpu_mask || cpumask_empty(cpu_mask)) in __sbi_send_ipi_v01()
165 cpu_mask = cpu_online_mask; in __sbi_send_ipi_v01()
166 hart_mask = __sbi_v01_cpumask_to_hartmask(cpu_mask); in __sbi_send_ipi_v01()
173 static int __sbi_rfence_v01(int fid, const struct cpumask *cpu_mask, in __sbi_rfence_v01() argument
180 if (!cpu_mask || cpumask_empty(cpu_mask)) in __sbi_rfence_v01()
[all …]
/linux-6.1.9/arch/riscv/include/asm/
Dsbi.h272 int sbi_send_ipi(const struct cpumask *cpu_mask);
273 int sbi_remote_fence_i(const struct cpumask *cpu_mask);
274 int sbi_remote_sfence_vma(const struct cpumask *cpu_mask,
278 int sbi_remote_sfence_vma_asid(const struct cpumask *cpu_mask,
282 int sbi_remote_hfence_gvma(const struct cpumask *cpu_mask,
285 int sbi_remote_hfence_gvma_vmid(const struct cpumask *cpu_mask,
289 int sbi_remote_hfence_vvma(const struct cpumask *cpu_mask,
292 int sbi_remote_hfence_vvma_asid(const struct cpumask *cpu_mask,
327 static inline int sbi_remote_fence_i(const struct cpumask *cpu_mask) { return -1; } in sbi_remote_fence_i() argument
/linux-6.1.9/arch/powerpc/platforms/pseries/
Dhotplug-cpu.c153 cpumask_var_t *cpu_mask) in find_cpu_id_range() argument
162 cpumask_clear(*cpu_mask); in find_cpu_id_range()
164 cpumask_set_cpu(cpu, *cpu_mask); in find_cpu_id_range()
188 while (!cpumask_empty(*cpu_mask)) { in find_cpu_id_range()
189 if (cpumask_subset(*cpu_mask, candidate_mask)) in find_cpu_id_range()
192 cpumask_shift_left(*cpu_mask, *cpu_mask, nthreads); in find_cpu_id_range()
195 if (!cpumask_empty(*cpu_mask)) in find_cpu_id_range()
214 cpumask_var_t cpu_mask; in pseries_add_processor() local
223 if (!alloc_cpumask_var(&cpu_mask, GFP_KERNEL)) in pseries_add_processor()
239 rc = find_cpu_id_range(nthreads, node, &cpu_mask); in pseries_add_processor()
[all …]
/linux-6.1.9/scripts/gdb/linux/
Dcpus.py56 cpu_mask = {} variable
60 global cpu_mask
61 cpu_mask = {}
68 global cpu_mask
70 if mask_name in cpu_mask:
71 mask = cpu_mask[mask_name]
75 cpu_mask[mask_name] = mask
/linux-6.1.9/arch/x86/kernel/cpu/resctrl/
Drdtgroup.c280 mask = &rdtgrp->plr->d->cpu_mask; in rdtgroup_cpus_show()
287 cpumask_pr_args(&rdtgrp->cpu_mask)); in rdtgroup_cpus_show()
326 update_closid_rmid(const struct cpumask *cpu_mask, struct rdtgroup *r) in update_closid_rmid() argument
330 if (cpumask_test_cpu(cpu, cpu_mask)) in update_closid_rmid()
332 smp_call_function_many(cpu_mask, update_cpu_closid_rmid, r, 1); in update_closid_rmid()
343 cpumask_andnot(tmpmask, newmask, &prgrp->cpu_mask); in cpus_mon_write()
350 cpumask_andnot(tmpmask, &rdtgrp->cpu_mask, newmask); in cpus_mon_write()
353 cpumask_or(&prgrp->cpu_mask, &prgrp->cpu_mask, tmpmask); in cpus_mon_write()
361 cpumask_andnot(tmpmask, newmask, &rdtgrp->cpu_mask); in cpus_mon_write()
367 cpumask_andnot(&crgrp->cpu_mask, &crgrp->cpu_mask, in cpus_mon_write()
[all …]
Dctrlmondata.c271 cpumask_var_t cpu_mask) in apply_config() argument
276 cpumask_set_cpu(cpumask_any(&dom->cpu_mask), cpu_mask); in apply_config()
293 if (!cpumask_test_cpu(smp_processor_id(), &d->cpu_mask)) in resctrl_arch_update_one()
312 cpumask_var_t cpu_mask; in resctrl_arch_update_domains() local
317 if (!zalloc_cpumask_var(&cpu_mask, GFP_KERNEL)) in resctrl_arch_update_domains()
329 if (!apply_config(hw_dom, cfg, idx, cpu_mask)) in resctrl_arch_update_domains()
343 if (cpumask_empty(cpu_mask)) in resctrl_arch_update_domains()
347 if (cpumask_test_cpu(cpu, cpu_mask)) in resctrl_arch_update_domains()
350 smp_call_function_many(cpu_mask, rdt_ctrl_update, &msr_param, 1); in resctrl_arch_update_domains()
354 free_cpumask_var(cpu_mask); in resctrl_arch_update_domains()
[all …]
Dcore.c332 if (cpumask_test_cpu(cpu, &d->cpu_mask)) in get_domain_from_cpu()
493 cpumask_set_cpu(cpu, &d->cpu_mask); in domain_add_cpu()
505 cpumask_set_cpu(cpu, &d->cpu_mask); in domain_add_cpu()
541 cpumask_clear_cpu(cpu, &d->cpu_mask); in domain_remove_cpu()
542 if (cpumask_empty(&d->cpu_mask)) { in domain_remove_cpu()
589 cpumask_set_cpu(cpu, &rdtgroup_default.cpu_mask); in resctrl_online_cpu()
601 if (cpumask_test_and_clear_cpu(cpu, &cr->cpu_mask)) { in clear_childcpus()
616 if (cpumask_test_and_clear_cpu(cpu, &rdtgrp->cpu_mask)) { in resctrl_offline_cpu()
Dpseudo_lock.c210 for_each_cpu(cpu, &plr->d->cpu_mask) { in pseudo_lock_cstates_constrain()
286 plr->cpu = cpumask_first(&plr->d->cpu_mask); in pseudo_lock_region_init()
719 if (!cpumask_empty(&rdtgrp->cpu_mask)) { in rdtgroup_locksetup_enter()
844 &d_i->cpu_mask); in rdtgroup_pseudo_locked_in_hierarchy()
852 if (cpumask_intersects(&d->cpu_mask, cpu_with_psl)) in rdtgroup_pseudo_locked_in_hierarchy()
1184 cpu = cpumask_first(&plr->d->cpu_mask); in pseudo_lock_measure_cycles()
1514 if (!cpumask_subset(current->cpus_ptr, &plr->d->cpu_mask)) { in pseudo_lock_dev_mmap()
/linux-6.1.9/arch/sparc/mm/
Dsrmmu.c1658 cpumask_t cpu_mask; in smp_flush_cache_mm() local
1659 cpumask_copy(&cpu_mask, mm_cpumask(mm)); in smp_flush_cache_mm()
1660 cpumask_clear_cpu(smp_processor_id(), &cpu_mask); in smp_flush_cache_mm()
1661 if (!cpumask_empty(&cpu_mask)) in smp_flush_cache_mm()
1670 cpumask_t cpu_mask; in smp_flush_tlb_mm() local
1671 cpumask_copy(&cpu_mask, mm_cpumask(mm)); in smp_flush_tlb_mm()
1672 cpumask_clear_cpu(smp_processor_id(), &cpu_mask); in smp_flush_tlb_mm()
1673 if (!cpumask_empty(&cpu_mask)) { in smp_flush_tlb_mm()
1690 cpumask_t cpu_mask; in smp_flush_cache_range() local
1691 cpumask_copy(&cpu_mask, mm_cpumask(mm)); in smp_flush_cache_range()
[all …]
/linux-6.1.9/tools/power/x86/amd_pstate_tracer/
Damd_pstate_trace.py156 …, freq_ghz, mperf, aperf, tsc, common_comm, load, duration_ms, sample_num, elapsed_time, cpu_mask): argument
161 if cpu_mask[cpu_int] == 0:
186 def read_trace_data(file_name, cpu_mask): argument
235 …f, freq_ghz, mperf, aperf, tsc, common_comm, load, duration_ms, sample_num, elapsed_time, cpu_mask)
240 ipt.split_csv(current_max_cpu, cpu_mask)
264 cpu_mask = zeros((MAX_CPUS,), dtype=int) variable
298 cpu_mask[int(p)] = 1
301 cpu_mask[i] = 1
331 read_trace_data(file_name, cpu_mask)
/linux-6.1.9/kernel/irq/
Daffinity.c130 const struct cpumask *cpu_mask, in alloc_nodes_vectors() argument
145 cpumask_and(nmsk, cpu_mask, node_to_cpumask[n]); in alloc_nodes_vectors()
251 const struct cpumask *cpu_mask, in __irq_build_affinity_masks() argument
261 if (cpumask_empty(cpu_mask)) in __irq_build_affinity_masks()
264 nodes = get_nodes_in_cpumask(node_to_cpumask, cpu_mask, &nodemsk); in __irq_build_affinity_masks()
273 cpumask_and(nmsk, cpu_mask, node_to_cpumask[n]); in __irq_build_affinity_masks()
288 alloc_nodes_vectors(numvecs, node_to_cpumask, cpu_mask, in __irq_build_affinity_masks()
299 cpumask_and(nmsk, cpu_mask, node_to_cpumask[nv->id]); in __irq_build_affinity_masks()
/linux-6.1.9/tools/power/x86/intel_pstate_tracer/
Dintel_pstate_tracer.py326 …, freq_ghz, io_boost, common_comm, load, duration_ms, sample_num, elapsed_time, tsc_ghz, cpu_mask): argument
331 if cpu_mask[cpu_int] == 0:
345 def split_csv(current_max_cpu, cpu_mask): argument
350 if cpu_mask[int(index)] != 0:
420 def read_trace_data(filename, cpu_mask): argument
480 …c, freq_ghz, io_boost, common_comm, load, duration_ms, sample_num, elapsed_time, tsc_ghz, cpu_mask)
486 split_csv(current_max_cpu, cpu_mask)
511 cpu_mask = zeros((MAX_CPUS,), dtype=int) variable
544 cpu_mask[int(p)] = 1
547 cpu_mask[i] = 1
[all …]
/linux-6.1.9/arch/x86/events/amd/
Dpower.c41 static cpumask_t cpu_mask; variable
149 return cpumap_print_to_pagebuf(true, buf, &cpu_mask); in get_attr_cpumask()
223 if (!cpumask_test_and_clear_cpu(cpu, &cpu_mask)) in power_cpu_exit()
233 cpumask_set_cpu(target, &cpu_mask); in power_cpu_exit()
254 cpumask_set_cpu(cpu, &cpu_mask); in power_cpu_init()
/linux-6.1.9/kernel/trace/
Dpreemptirq_delay_test.c122 struct cpumask cpu_mask; in preemptirq_delay_run() local
125 cpumask_clear(&cpu_mask); in preemptirq_delay_run()
126 cpumask_set_cpu(cpu_affinity, &cpu_mask); in preemptirq_delay_run()
127 if (set_cpus_allowed_ptr(current, &cpu_mask)) in preemptirq_delay_run()
/linux-6.1.9/drivers/irqchip/
Dirq-vf610-mscm-ir.c43 u16 cpu_mask; member
95 WARN_ON(irsprc & ~chip_data->cpu_mask); in vf610_mscm_ir_enable()
97 writew_relaxed(chip_data->cpu_mask, in vf610_mscm_ir_enable()
210 mscm_ir_data->cpu_mask = 0x1 << cpuid; in vf610_mscm_ir_of_init()
Dirq-hip04.c272 unsigned int cpu_mask, cpu = smp_processor_id(); in hip04_irq_cpu_init() local
279 cpu_mask = hip04_get_cpumask(intc); in hip04_irq_cpu_init()
280 hip04_cpu_map[cpu] = cpu_mask; in hip04_irq_cpu_init()
288 hip04_cpu_map[i] &= ~cpu_mask; in hip04_irq_cpu_init()
/linux-6.1.9/arch/arm/mach-omap2/
Dclkt2xxx_virt_prcm_set.c80 if (!(ptr->flags & cpu_mask)) in omap2_round_to_table_rate()
104 if (!(prcm->flags & cpu_mask)) in omap2_select_table_rate()
177 if (!(prcm->flags & cpu_mask)) in omap2xxx_clkt_vps_check_bootloader_rates()
/linux-6.1.9/tools/testing/selftests/bpf/prog_tests/
Dcpu_mask.c33 const char *cpu_mask; member
65 err = parse_cpu_mask_str(test_cases[i].cpu_mask, &mask, &n); in test_cpu_mask()
/linux-6.1.9/arch/arm/mm/
Dproc-arm9tdmi.S71 .macro arm9tdmi_proc_info name:req, cpu_val:req, cpu_mask:req, cpu_name:req
75 .long \cpu_mask
Dproc-arm7tdmi.S77 .macro arm7tdmi_proc_info name:req, cpu_val:req, cpu_mask:req, cpu_name:req, \
82 .long \cpu_mask
/linux-6.1.9/tools/perf/tests/
Dperf-record.c52 cpu_set_t cpu_mask; in test__PERF_RECORD() local
53 size_t cpu_mask_size = sizeof(cpu_mask); in test__PERF_RECORD()
112 err = sched__get_first_possible_cpu(evlist->workload.pid, &cpu_mask); in test__PERF_RECORD()
124 if (sched_setaffinity(evlist->workload.pid, cpu_mask_size, &cpu_mask) < 0) { in test__PERF_RECORD()
/linux-6.1.9/drivers/soc/ti/
Dknav_qmss_acc.c200 struct cpumask *cpu_mask; in knav_range_setup_acc_irq() local
207 cpu_mask = range->irqs[0].cpu_mask; in knav_range_setup_acc_irq()
211 cpu_mask = range->irqs[queue].cpu_mask; in knav_range_setup_acc_irq()
234 if (!ret && cpu_mask) { in knav_range_setup_acc_irq()
235 ret = irq_set_affinity_hint(irq, cpu_mask); in knav_range_setup_acc_irq()
/linux-6.1.9/arch/x86/hyperv/
Dhv_apic.c196 ipi_arg.cpu_mask = 0; in __send_ipi_mask()
212 __set_bit(vcpu, (unsigned long *)&ipi_arg.cpu_mask); in __send_ipi_mask()
216 ipi_arg.cpu_mask); in __send_ipi_mask()
/linux-6.1.9/arch/arm/mach-omap1/
Dclock_data.c692 u32 cpu_mask; variable
716 cpu_mask = 0; in omap1_clk_init()
718 cpu_mask |= CK_1710; in omap1_clk_init()
720 cpu_mask |= CK_16XX; in omap1_clk_init()
722 cpu_mask |= CK_1510; in omap1_clk_init()
724 cpu_mask |= CK_7XX; in omap1_clk_init()
726 cpu_mask |= CK_310; in omap1_clk_init()
806 if (!(c->cpu & cpu_mask)) in omap1_clk_init()
/linux-6.1.9/io_uring/
Dio-wq.c109 cpumask_var_t cpu_mask; member
706 set_cpus_allowed_ptr(tsk, wqe->cpu_mask); in io_init_new_worker()
1168 if (!alloc_cpumask_var(&wqe->cpu_mask, GFP_KERNEL)) in io_wq_create()
1170 cpumask_copy(wqe->cpu_mask, cpumask_of_node(node)); in io_wq_create()
1201 free_cpumask_var(wq->wqes[node]->cpu_mask); in io_wq_create()
1283 free_cpumask_var(wqe->cpu_mask); in io_wq_destroy()
1308 cpumask_set_cpu(od->cpu, worker->wqe->cpu_mask); in io_wq_worker_affinity()
1310 cpumask_clear_cpu(od->cpu, worker->wqe->cpu_mask); in io_wq_worker_affinity()
1352 cpumask_copy(wqe->cpu_mask, mask); in io_wq_cpu_affinity()
1354 cpumask_copy(wqe->cpu_mask, cpumask_of_node(i)); in io_wq_cpu_affinity()

123