/linux-6.1.9/drivers/net/wireguard/ |
D | queueing.h | 114 cpu = cpumask_next(cpu, cpu_online_mask); in wg_cpumask_choose_online() 132 cpu = cpumask_next(cpu, cpu_online_mask) % nr_cpumask_bits; in wg_cpumask_next_online() 133 *next = cpumask_next(cpu, cpu_online_mask) % nr_cpumask_bits; in wg_cpumask_next_online()
|
/linux-6.1.9/drivers/cpuidle/ |
D | dt_idle_states.c | 111 for (cpu = cpumask_next(cpumask_first(cpumask), cpumask); in idle_state_valid() 112 cpu < nr_cpu_ids; cpu = cpumask_next(cpu, cpumask)) { in idle_state_valid()
|
/linux-6.1.9/lib/ |
D | cpumask_kunit.c | 107 KUNIT_EXPECT_LE_MSG(test, nr_cpu_ids, cpumask_next(-1, &mask_empty), in test_cpumask_next() 109 KUNIT_EXPECT_EQ_MSG(test, 0, cpumask_next(-1, cpu_possible_mask), in test_cpumask_next()
|
D | cpumask.c | 27 next = cpumask_next(n, mask); in cpumask_next_wrap()
|
/linux-6.1.9/drivers/infiniband/hw/hfi1/ |
D | affinity.c | 115 curr_cpu = cpumask_next(curr_cpu, &node_affinity.real_cpu_mask); in init_real_cpu_mask() 122 curr_cpu = cpumask_next(curr_cpu, &node_affinity.real_cpu_mask); in init_real_cpu_mask() 645 curr_cpu = cpumask_next(curr_cpu, in hfi1_dev_affinity_init() 660 curr_cpu = cpumask_next(curr_cpu, in hfi1_dev_affinity_init() 981 curr_cpu = cpumask_next(curr_cpu, hw_thread_mask); in find_hw_thread_mask() 985 curr_cpu = cpumask_next(curr_cpu, hw_thread_mask); in find_hw_thread_mask()
|
/linux-6.1.9/kernel/rcu/ |
D | rcu.h | 363 (cpu) = cpumask_next((rnp)->grplo - 1, cpu_possible_mask); \ 365 (cpu) = cpumask_next((cpu), cpu_possible_mask))
|
/linux-6.1.9/arch/x86/kernel/cpu/ |
D | proc.c | 156 *pos = cpumask_next(*pos - 1, cpu_online_mask); in c_start()
|
/linux-6.1.9/arch/riscv/kernel/ |
D | cpu.c | 219 *pos = cpumask_next(*pos - 1, cpu_online_mask); in c_start()
|
/linux-6.1.9/arch/powerpc/kernel/ |
D | setup-common.c | 335 if (cpumask_next(cpu_id, cpu_online_mask) >= nr_cpu_ids) in show_cpuinfo() 346 *pos = cpumask_next(*pos - 1, cpu_online_mask); in c_start()
|
D | irq.c | 366 irq_rover = cpumask_next(irq_rover, cpu_online_mask); in irq_choose_cpu()
|
D | rtasd.c | 434 cpu = cpumask_next(raw_smp_processor_id(), cpu_online_mask); in rtas_event_scan()
|
/linux-6.1.9/kernel/sched/ |
D | stats.c | 196 n = cpumask_next(n - 1, cpu_online_mask); in schedstat_start()
|
/linux-6.1.9/kernel/time/ |
D | clocksource.c | 296 cpu = cpumask_next(cpu, cpu_online_mask); in clocksource_verify_choose_cpus() 314 cpu = cpumask_next(cpu - 1, cpu_online_mask); in clocksource_verify_choose_cpus() 506 next_cpu = cpumask_next(raw_smp_processor_id(), cpu_online_mask); in clocksource_watchdog()
|
D | timer_list.c | 304 iter->cpu = cpumask_next(iter->cpu, cpu_online_mask); in move_iter()
|
/linux-6.1.9/kernel/irq/ |
D | ipi.c | 71 next = cpumask_next(next, dest); in irq_reserve_ipi()
|
D | affinity.c | 32 sibl = cpumask_next(sibl, siblmsk); in irq_spread_init_one()
|
/linux-6.1.9/arch/x86/kernel/ |
D | tsc_sync.c | 100 next_cpu = cpumask_next(raw_smp_processor_id(), cpu_online_mask); in tsc_sync_check_timer_fn()
|
/linux-6.1.9/arch/s390/kernel/ |
D | processor.c | 338 *pos = cpumask_next(*pos - 1, cpu_online_mask); in c_update()
|
/linux-6.1.9/arch/openrisc/kernel/ |
D | setup.c | 371 *pos = cpumask_next(*pos - 1, cpu_online_mask); in c_start()
|
/linux-6.1.9/crypto/ |
D | pcrypt.c | 184 ctx->cb_cpu = cpumask_next(ctx->cb_cpu, cpu_online_mask); in pcrypt_aead_init_tfm()
|
/linux-6.1.9/fs/ |
D | seq_file.c | 1136 for (*cpu = cpumask_next(*cpu, cpu_possible_mask); *cpu < nr_cpu_ids; in seq_hlist_next_percpu() 1137 *cpu = cpumask_next(*cpu, cpu_possible_mask)) { in seq_hlist_next_percpu()
|
/linux-6.1.9/kernel/ |
D | padata.c | 56 target_cpu = cpumask_next(target_cpu, pd->cpumask.pcpu); in padata_index_to_cpu() 192 cpu = cpumask_next(cpu, pd->cpumask.cbcpu); in padata_do_parallel()
|
/linux-6.1.9/drivers/thermal/intel/ |
D | intel_powerclamp.c | 572 control_cpu = cpumask_next(cpu, cpu_online_mask); in powerclamp_cpu_predown()
|
/linux-6.1.9/drivers/infiniband/sw/siw/ |
D | siw_main.c | 209 i++, cpu = cpumask_next(cpu, tx_cpumask)) { in siw_get_tx_cpu()
|
/linux-6.1.9/net/openvswitch/ |
D | flow.c | 138 for (cpu = 0; cpu < nr_cpu_ids; cpu = cpumask_next(cpu, &flow->cpu_used_mask)) { in ovs_flow_stats_get() 162 for (cpu = 0; cpu < nr_cpu_ids; cpu = cpumask_next(cpu, &flow->cpu_used_mask)) { in ovs_flow_stats_clear()
|