/linux-6.6.21/arch/powerpc/include/asm/ |
D | cputhreads.h | 23 extern int threads_per_core; 28 #define threads_per_core 1 macro 50 return cpu & (threads_per_core - 1); in cpu_thread_in_core() 60 return cpu & ~(threads_per_core - 1); in cpu_first_thread_sibling() 65 return cpu | (threads_per_core - 1); in cpu_last_thread_sibling() 76 if (cpu_has_feature(CPU_FTR_ARCH_300) && (threads_per_core == 8)) in cpu_first_tlb_thread_sibling() 84 if (cpu_has_feature(CPU_FTR_ARCH_300) && (threads_per_core == 8)) in cpu_last_tlb_thread_sibling() 92 if (cpu_has_feature(CPU_FTR_ARCH_300) && (threads_per_core == 8)) in cpu_tlb_thread_sibling_step()
|
/linux-6.6.21/arch/powerpc/platforms/powernv/ |
D | subcore.c | 154 for (i = cpu + 1; i < cpu + threads_per_core; i++) in wait_for_sync_step() 209 for (i = cpu + 1; i < cpu + threads_per_core; i++) in unsplit_core() 328 threads_per_subcore = threads_per_core / subcores_per_core; in cpu_update_split_mode() 435 if (setup_max_cpus % threads_per_core) in subcore_init()
|
D | idle.c | 320 unsigned long core_thread_mask = (1UL << threads_per_core) - 1; in power7_idle_insn() 371 == threads_per_core) in power7_idle_insn() 613 unsigned long core_thread_mask = (1UL << threads_per_core) - 1; in power9_idle_stop() 822 int need_awake = threads_per_core; in pnv_power9_force_smt4_catch() 825 cpu0 = cpu & ~(threads_per_core - 1); in pnv_power9_force_smt4_catch() 826 for (thr = 0; thr < threads_per_core; ++thr) { in pnv_power9_force_smt4_catch() 832 for (thr = 0; thr < threads_per_core; ++thr) { in pnv_power9_force_smt4_catch() 842 for (thr = 0; thr < threads_per_core; ++thr) { in pnv_power9_force_smt4_catch() 851 for (thr = 0; thr < threads_per_core; ++thr) { in pnv_power9_force_smt4_catch() 868 cpu0 = cpu & ~(threads_per_core - 1); in pnv_power9_force_smt4_release() [all …]
|
/linux-6.6.21/drivers/virt/nitro_enclaves/ |
D | ne_misc_dev.c | 554 cpumask_set_cpu(cpu, ne_enclave->threads_per_core[core_id]); in ne_set_enclave_threads_per_core() 586 for_each_cpu(cpu, ne_enclave->threads_per_core[i]) in ne_get_cpu_from_cpu_pool() 605 *vcpu_id = cpumask_any(ne_enclave->threads_per_core[core_id]); in ne_get_cpu_from_cpu_pool() 670 if (cpumask_test_cpu(vcpu_id, ne_enclave->threads_per_core[i])) in ne_check_cpu_in_cpu_pool() 1105 for_each_cpu(cpu, ne_enclave->threads_per_core[i]) in ne_start_enclave_ioctl() 1425 for_each_cpu(cpu, ne_enclave->threads_per_core[i]) in ne_enclave_remove_all_vcpu_id_entries() 1429 free_cpumask_var(ne_enclave->threads_per_core[i]); in ne_enclave_remove_all_vcpu_id_entries() 1434 kfree(ne_enclave->threads_per_core); in ne_enclave_remove_all_vcpu_id_entries() 1632 ne_enclave->threads_per_core = kcalloc(ne_enclave->nr_parent_vm_cores, in ne_create_vm_ioctl() 1633 sizeof(*ne_enclave->threads_per_core), in ne_create_vm_ioctl() [all …]
|
D | ne_misc_dev.h | 80 cpumask_var_t *threads_per_core; member
|
/linux-6.6.21/arch/powerpc/perf/ |
D | imc-pmu.c | 601 int nid, rc = 0, core_id = (cpu / threads_per_core); in core_imc_mem_init() 638 int core_id = (cpu / threads_per_core); in is_core_imc_mem_inited() 716 core_id = cpu / threads_per_core; in ppc_core_imc_cpu_offline() 774 core_id = event->cpu / threads_per_core; in core_imc_counters_release() 842 core_id = event->cpu / threads_per_core; in core_imc_event_init() 1134 core_id = smp_processor_id() / threads_per_core; in thread_imc_event_add() 1169 core_id = smp_processor_id() / threads_per_core; in thread_imc_event_del() 1208 int core_id = (cpu_id / threads_per_core); in trace_imc_mem_alloc() 1359 int core_id = smp_processor_id() / threads_per_core; in trace_imc_event_add() 1409 int core_id = smp_processor_id() / threads_per_core; in trace_imc_event_del() [all …]
|
D | hv-24x7.c | 1724 if (threads_per_core == 8) in hv_24x7_init()
|
/linux-6.6.21/arch/powerpc/platforms/85xx/ |
D | smp.c | 288 if (threads_per_core == 2) { in smp_85xx_kick_cpu() 319 } else if (threads_per_core == 1) { in smp_85xx_kick_cpu() 326 } else if (threads_per_core > 2) { in smp_85xx_kick_cpu()
|
/linux-6.6.21/arch/powerpc/kernel/ |
D | setup_64.c | 134 smt_enabled_at_boot = threads_per_core; in check_smt_enabled() 139 smt_enabled_at_boot = threads_per_core; in check_smt_enabled() 149 min(threads_per_core, smt); in check_smt_enabled() 159 smt_enabled_at_boot = threads_per_core; in check_smt_enabled()
|
D | setup-common.c | 398 int threads_per_core, threads_per_subcore, threads_shift __read_mostly; variable 400 EXPORT_SYMBOL_GPL(threads_per_core); 409 threads_per_core = tpc; in cpu_init_thread_core_maps()
|
D | smp.c | 937 for (i = first_thread; i < first_thread + threads_per_core; i++) { in update_mask_from_threadgroup() 1143 int idx = DIV_ROUND_UP(num_possible_cpus(), threads_per_core); in smp_prepare_cpus() 1163 cpu_smt_set_num_threads(num_threads, threads_per_core); in smp_prepare_cpus() 1279 if (threads_per_core > 1 && secondaries_inhibited() && in __cpu_up() 1573 for (i = first_thread; i < first_thread + threads_per_core; i++) in add_cpu_to_masks()
|
D | prom.c | 976 idx = cpu / threads_per_core; in cpu_to_chip_id()
|
/linux-6.6.21/arch/powerpc/kvm/ |
D | e500mc.c | 358 if (threads_per_core == 2) in kvmppc_core_init_vm_e500mc() 369 if (threads_per_core == 2) in kvmppc_core_destroy_vm_e500mc() 410 kvmppc_init_lpid(KVMPPC_NR_LPIDS/threads_per_core); in kvmppc_e500mc_init()
|
D | e500.h | 299 if (threads_per_core == 2) in get_thread_specific_lpid()
|
D | book3s_hv.c | 5361 for (cpu = 0; cpu < nr_cpu_ids; cpu += threads_per_core) { in kvmppc_alloc_host_rm_ops() 6238 int first_cpu = i * threads_per_core; in kvm_init_subcore_bitmap() 6252 for (j = 0; j < threads_per_core; j++) { in kvm_init_subcore_bitmap()
|
/linux-6.6.21/arch/powerpc/sysdev/ |
D | fsl_rcpm.c | 148 if (threads_per_core == 2) { in rcpm_v2_cpu_die() 160 if (threads_per_core == 1) in rcpm_v2_cpu_die()
|
/linux-6.6.21/arch/ia64/kernel/ |
D | setup.c | 723 if (c->threads_per_core > 1 || c->cores_per_socket > 1) in show_cpuinfo() 835 c->threads_per_core = c->cores_per_socket = c->num_log = 1; in identify_cpu() 840 if (c->threads_per_core > smp_num_siblings) in identify_cpu() 841 smp_num_siblings = c->threads_per_core; in identify_cpu()
|
D | smpboot.c | 579 if (cpu_data(cpu)->threads_per_core == 1 && in remove_siblinginfo() 735 if (cpu_data(cpu)->threads_per_core == 1 && in __cpu_up() 810 c->threads_per_core = info.overview_tpc; in identify_siblings()
|
D | topology.c | 134 if (cpu_data(cpu)->threads_per_core <= 1 && in cache_shared_cpu_map_setup()
|
/linux-6.6.21/arch/powerpc/platforms/pseries/ |
D | lpar.c | 209 vcpu_associativity = kcalloc(num_possible_cpus() / threads_per_core, in init_cpu_associativity() 211 pcpu_associativity = kcalloc(NR_CPUS_H / threads_per_core, in init_cpu_associativity() 234 assoc = &cpu_assoc[(int)(cpu / threads_per_core) * VPHN_ASSOC_BUFSIZE]; in __get_cpu_associativity()
|
/linux-6.6.21/arch/ia64/include/asm/ |
D | processor.h | 214 unsigned char threads_per_core; /* Threads per core */ member
|
/linux-6.6.21/tools/power/x86/turbostat/ |
D | turbostat.c | 410 topo.nodes_per_pkg * topo.cores_per_node * topo.threads_per_core) + \ 411 ((node_no) * topo.cores_per_node * topo.threads_per_core) + \ 412 ((core_no) * topo.threads_per_core) + \ 570 int threads_per_core; member 602 for (thread_no = 0; thread_no < topo.threads_per_core; ++thread_no) { in for_all_cpus() 2994 for (thread_no = 0; thread_no < topo.threads_per_core; ++thread_no) { in for_all_cpus_2() 5978 topo.threads_per_core = max_siblings; in topology_probe() 6000 int num_threads = topo.threads_per_core * num_cores; in allocate_counters()
|
/linux-6.6.21/tools/arch/x86/kcpuid/ |
D | cpuid.csv | 438 0x8000001E, 0, EBX, 15:8, threads_per_core, The number of threads per core is threads_per_core + 1
|
/linux-6.6.21/drivers/cpufreq/ |
D | powernv-cpufreq.c | 836 for (i = 0; i < threads_per_core; i++) in powernv_cpufreq_cpu_init()
|
/linux-6.6.21/arch/powerpc/mm/ |
D | numa.c | 775 for (i = 0; i < threads_per_core; i++) { in verify_cpu_node_mapping()
|