/linux-6.6.21/kernel/ |
D | watchdog_buddy.c | 16 if (next_cpu >= nr_cpu_ids) in watchdog_next_cpu() 20 return nr_cpu_ids; in watchdog_next_cpu() 51 if (next_cpu < nr_cpu_ids) in watchdog_hardlockup_enable() 75 if (next_cpu < nr_cpu_ids) in watchdog_hardlockup_disable() 102 if (next_cpu >= nr_cpu_ids) in watchdog_buddy_check_hardlockup()
|
D | scftorture.c | 342 cpu = torture_random(trsp) % nr_cpu_ids; in scftorture_invoke_one() 349 cpu = torture_random(trsp) % nr_cpu_ids; in scftorture_invoke_one() 372 cpu = torture_random(trsp) % nr_cpu_ids; in scftorture_invoke_one() 455 cpu = scfp->cpu % nr_cpu_ids; in scftorture_invoker() 465 WARN_ONCE(curcpu != scfp->cpu % nr_cpu_ids, in scftorture_invoker() 467 __func__, scfp->cpu, curcpu, nr_cpu_ids); in scftorture_invoker() 564 weight_resched1 = weight_resched == 0 ? 0 : 2 * nr_cpu_ids; in scf_torture_init() 565 weight_single1 = weight_single == 0 ? 0 : 2 * nr_cpu_ids; in scf_torture_init() 566 weight_single_rpc1 = weight_single_rpc == 0 ? 0 : 2 * nr_cpu_ids; in scf_torture_init() 567 weight_single_wait1 = weight_single_wait == 0 ? 0 : 2 * nr_cpu_ids; in scf_torture_init()
|
D | smp.c | 244 if (WARN_ONCE(cpu < 0 || cpu >= nr_cpu_ids, "%s: cpu = %d\n", __func__, cpu)) in csd_lock_wait_toolong() 410 if ((unsigned)cpu >= nr_cpu_ids || !cpu_online(cpu)) { in generic_exec_single() 732 for (cpu = cpumask_first_and(nodemask, mask); cpu < nr_cpu_ids; in smp_call_function_any() 796 if (cpu < nr_cpu_ids) in smp_call_function_many_cond() 939 if (get_option(&str, &nr_cpus) && nr_cpus > 0 && nr_cpus < nr_cpu_ids) in nrcpus() 960 unsigned int nr_cpu_ids __read_mostly = NR_CPUS; 961 EXPORT_SYMBOL(nr_cpu_ids); 1117 if (cpu >= nr_cpu_ids || !cpu_online(cpu)) in smp_call_on_cpu()
|
/linux-6.6.21/lib/ |
D | cpumask_kunit.c | 44 for_each_cpu_wrap(cpu, m, nr_cpu_ids / 2) \ 68 KUNIT_EXPECT_EQ_MSG(test, nr_cpu_ids, cpumask_weight(cpu_possible_mask), in test_cpumask_weight() 70 KUNIT_EXPECT_EQ_MSG(test, nr_cpu_ids, cpumask_weight(&mask_all), MASK_MSG(&mask_all)); in test_cpumask_weight() 75 KUNIT_EXPECT_LE_MSG(test, nr_cpu_ids, cpumask_first(&mask_empty), MASK_MSG(&mask_empty)); in test_cpumask_first() 79 KUNIT_EXPECT_LE_MSG(test, nr_cpu_ids, cpumask_first_zero(cpu_possible_mask), in test_cpumask_first() 87 KUNIT_EXPECT_EQ_MSG(test, nr_cpu_ids - 1, cpumask_last(cpu_possible_mask), in test_cpumask_last() 94 KUNIT_EXPECT_LE_MSG(test, nr_cpu_ids, cpumask_next_zero(-1, cpu_possible_mask), in test_cpumask_next() 97 KUNIT_EXPECT_LE_MSG(test, nr_cpu_ids, cpumask_next(-1, &mask_empty), in test_cpumask_next()
|
D | cpumask.c | 153 WARN_ON(cpu >= nr_cpu_ids); in cpumask_local_spread() 180 if (next < nr_cpu_ids) in cpumask_any_and_distribute() 194 if (next < nr_cpu_ids) in cpumask_any_distribute()
|
/linux-6.6.21/arch/powerpc/kernel/ |
D | paca.c | 62 size_t shared_lppaca_total_size = PAGE_ALIGN(nr_cpu_ids * LPPACA_SIZE); in alloc_shared_lppaca() 245 paca_nr_cpu_ids = nr_cpu_ids; in allocate_paca_ptrs() 247 paca_ptrs_size = sizeof(struct paca_struct *) * nr_cpu_ids; in allocate_paca_ptrs() 291 new_ptrs_size = sizeof(struct paca_struct *) * nr_cpu_ids; in free_unused_pacas() 296 paca_nr_cpu_ids = nr_cpu_ids; in free_unused_pacas() 309 paca_ptrs_size + paca_struct_size, nr_cpu_ids); in free_unused_pacas()
|
D | setup-common.c | 340 if (cpumask_next(cpu_id, cpu_online_mask) >= nr_cpu_ids) in show_cpuinfo() 352 if ((*pos) < nr_cpu_ids) in c_start() 456 cpu_to_phys_id = memblock_alloc(nr_cpu_ids * sizeof(u32), in smp_setup_cpu_maps() 460 __func__, nr_cpu_ids * sizeof(u32), __alignof__(u32)); in smp_setup_cpu_maps() 487 for (j = 0; j < nthreads && cpu < nr_cpu_ids; j++) { in smp_setup_cpu_maps() 504 if (cpu >= nr_cpu_ids) { in smp_setup_cpu_maps() 540 if (maxcpus > nr_cpu_ids) { in smp_setup_cpu_maps() 544 maxcpus, nr_cpu_ids); in smp_setup_cpu_maps() 545 maxcpus = nr_cpu_ids; in smp_setup_cpu_maps() 889 memblock_free(cpu_to_phys_id, nr_cpu_ids * sizeof(u32)); in smp_setup_pacas()
|
/linux-6.6.21/arch/arm/mach-spear/ |
D | platsmp.c | 102 if (ncores > nr_cpu_ids) { in spear13xx_smp_init_cpus() 104 ncores, nr_cpu_ids); in spear13xx_smp_init_cpus() 105 ncores = nr_cpu_ids; in spear13xx_smp_init_cpus()
|
/linux-6.6.21/include/linux/ |
D | cpumask.h | 36 #define cpumask_pr_args(maskp) nr_cpu_ids, cpumask_bits(maskp) 39 #define nr_cpu_ids ((unsigned int)NR_CPUS) macro 41 extern unsigned int nr_cpu_ids; 47 WARN_ON(nr != nr_cpu_ids); in set_nr_cpu_ids() 49 nr_cpu_ids = nr; in set_nr_cpu_ids() 81 #define small_cpumask_bits nr_cpu_ids 84 #define small_cpumask_bits nr_cpu_ids 85 #define large_cpumask_bits nr_cpu_ids 87 #define nr_cpumask_bits nr_cpu_ids 1170 nr_cpu_ids); in cpumap_print_to_pagebuf() [all …]
|
/linux-6.6.21/arch/riscv/kernel/ |
D | smpboot.c | 132 for (cpuid = 1; cpuid < nr_cpu_ids; cpuid++) { in acpi_parse_and_init_cpus() 177 if (cpuid > nr_cpu_ids) in of_parse_and_init_cpus() 179 cpuid, nr_cpu_ids); in of_parse_and_init_cpus() 181 for (cpuid = 1; cpuid < nr_cpu_ids; cpuid++) { in of_parse_and_init_cpus()
|
/linux-6.6.21/arch/arm/mach-bcm/ |
D | bcm63xx_smp.c | 64 if (ncores > nr_cpu_ids) { in scu_a9_enable() 66 ncores, nr_cpu_ids); in scu_a9_enable() 67 ncores = nr_cpu_ids; in scu_a9_enable()
|
/linux-6.6.21/arch/arm/kernel/ |
D | devtree.c | 130 if (WARN(cpuidx > nr_cpu_ids, "DT /cpu %u nodes greater than " in arm_dt_init_cpu_maps() 132 cpuidx, nr_cpu_ids)) { in arm_dt_init_cpu_maps() 133 cpuidx = nr_cpu_ids; in arm_dt_init_cpu_maps()
|
/linux-6.6.21/kernel/irq/ |
D | cpuhotplug.c | 39 if (cpumask_any_but(m, cpu) < nr_cpu_ids && in irq_needs_fixup() 40 cpumask_any_and(m, cpu_online_mask) >= nr_cpu_ids) { in irq_needs_fixup() 113 if (cpumask_any_and(affinity, cpu_online_mask) >= nr_cpu_ids) { in migrate_one_irq()
|
D | ipi.c | 70 if (next < nr_cpu_ids) in irq_reserve_ipi() 72 if (next < nr_cpu_ids) { in irq_reserve_ipi() 168 if (!data || cpu >= nr_cpu_ids) in ipi_get_hwirq() 199 if (cpu >= nr_cpu_ids) in ipi_send_verify()
|
D | migration.c | 29 if (cpumask_any_and(desc->pending_mask, cpu_online_mask) >= nr_cpu_ids) { in irq_fixup_move_pending() 77 if (cpumask_any_and(desc->pending_mask, cpu_online_mask) < nr_cpu_ids) { in irq_move_masked_irq()
|
/linux-6.6.21/scripts/gdb/linux/ |
D | timerlist.py | 151 nr_cpu_ids = 1 153 nr_cpu_ids = gdb.parse_and_eval("nr_cpu_ids") 157 num_bytes = (nr_cpu_ids + 7) / 8 173 extra = nr_cpu_ids % 8
|
/linux-6.6.21/net/netfilter/ |
D | nf_flow_table_procfs.c | 14 for (cpu = *pos - 1; cpu < nr_cpu_ids; ++cpu) { in nf_flow_table_cpu_seq_start() 29 for (cpu = *pos; cpu < nr_cpu_ids; ++cpu) { in nf_flow_table_cpu_seq_next()
|
/linux-6.6.21/arch/x86/kernel/apic/ |
D | probe_32.c | 98 if (nr_cpu_ids <= 8 || xen_pv_domain()) in x86_32_probe_bigsmp_early() 121 if (nr_cpu_ids > 8 && !xen_pv_domain()) in x86_32_install_bigsmp()
|
/linux-6.6.21/arch/arm/mach-omap2/ |
D | omap-smp.c | 278 if (ncores > nr_cpu_ids) { in omap4_smp_init_cpus() 280 ncores, nr_cpu_ids); in omap4_smp_init_cpus() 281 ncores = nr_cpu_ids; in omap4_smp_init_cpus()
|
/linux-6.6.21/drivers/perf/ |
D | arm_pmu_platform.c | 87 cpu = nr_cpu_ids; in pmu_parse_irq_affinity() 123 if (nr_cpu_ids != 1 && !pmu_has_irq_affinity(dev->of_node)) in pmu_parse_irqs() 141 if (cpu >= nr_cpu_ids) in pmu_parse_irqs()
|
/linux-6.6.21/drivers/md/ |
D | dm-ps-io-affinity.c | 80 if (cpu >= nr_cpu_ids) { in ioa_add_path() 82 cpu, nr_cpu_ids); in ioa_add_path() 119 s->path_map = kzalloc(nr_cpu_ids * sizeof(struct path_info *), in ioa_create()
|
/linux-6.6.21/arch/x86/kernel/ |
D | tsc_sync.c | 102 if (next_cpu >= nr_cpu_ids) in tsc_sync_check_timer_fn() 210 refcpu = mask ? cpumask_any_but(mask, cpu) : nr_cpu_ids; in tsc_store_and_check_tsc_adjust() 212 if (refcpu >= nr_cpu_ids) { in tsc_store_and_check_tsc_adjust()
|
/linux-6.6.21/drivers/base/ |
D | cpu.c | 260 if (total_cpus && nr_cpu_ids < total_cpus) { in print_cpus_offline() 263 if (nr_cpu_ids == total_cpus-1) in print_cpus_offline() 264 len += sysfs_emit_at(buf, len, "%u", nr_cpu_ids); in print_cpus_offline() 267 nr_cpu_ids, total_cpus - 1); in print_cpus_offline() 422 if (cpu < nr_cpu_ids && cpu_possible(cpu)) in get_cpu_device()
|
/linux-6.6.21/kernel/sched/ |
D | cpupri.c | 99 if (cpumask_any_and(&p->cpus_mask, vec->mask) >= nr_cpu_ids) in __cpupri_find() 290 cp->cpu_to_pri = kcalloc(nr_cpu_ids, sizeof(int), GFP_KERNEL); in cpupri_init()
|
/linux-6.6.21/drivers/nvdimm/ |
D | nd_perf.c | 155 if (target >= nr_cpu_ids) { in nvdimm_pmu_cpu_offline() 163 if (target >= 0 && target < nr_cpu_ids) in nvdimm_pmu_cpu_offline() 175 if (nd_pmu->cpu >= nr_cpu_ids) in nvdimm_pmu_cpu_online()
|