/linux-5.19.10/tools/perf/tests/ |
D | topology.c | 119 if (cpu__get_socket_id(perf_cpu_map__cpu(map, 0)) == -1) in check_cpu_topology() 137 id = aggr_cpu_id__cpu(perf_cpu_map__cpu(map, i), NULL); in check_cpu_topology() 139 perf_cpu_map__cpu(map, i).cpu == id.cpu.cpu); in check_cpu_topology() 142 session->header.env.cpu[perf_cpu_map__cpu(map, i).cpu].core_id == id.core); in check_cpu_topology() 144 session->header.env.cpu[perf_cpu_map__cpu(map, i).cpu].socket_id == in check_cpu_topology() 148 session->header.env.cpu[perf_cpu_map__cpu(map, i).cpu].die_id == id.die); in check_cpu_topology() 155 id = aggr_cpu_id__core(perf_cpu_map__cpu(map, i), NULL); in check_cpu_topology() 157 session->header.env.cpu[perf_cpu_map__cpu(map, i).cpu].core_id == id.core); in check_cpu_topology() 160 session->header.env.cpu[perf_cpu_map__cpu(map, i).cpu].socket_id == in check_cpu_topology() 164 session->header.env.cpu[perf_cpu_map__cpu(map, i).cpu].die_id == id.die); in check_cpu_topology() [all …]
|
D | event_update.c | 79 TEST_ASSERT_VAL("wrong cpus", perf_cpu_map__cpu(map, 0).cpu == 1); in process_event_cpus() 80 TEST_ASSERT_VAL("wrong cpus", perf_cpu_map__cpu(map, 1).cpu == 2); in process_event_cpus() 81 TEST_ASSERT_VAL("wrong cpus", perf_cpu_map__cpu(map, 2).cpu == 3); in process_event_cpus()
|
D | cpumap.c | 41 TEST_ASSERT_VAL("wrong cpu", perf_cpu_map__cpu(map, i).cpu == i); in process_event_mask() 70 TEST_ASSERT_VAL("wrong cpu", perf_cpu_map__cpu(map, 0).cpu == 1); in process_event_cpus() 71 TEST_ASSERT_VAL("wrong cpu", perf_cpu_map__cpu(map, 1).cpu == 256); in process_event_cpus()
|
D | mmap-basic.c | 62 CPU_SET(perf_cpu_map__cpu(cpus, 0).cpu, &cpu_set); in test__basic_mmap() 66 perf_cpu_map__cpu(cpus, 0).cpu, in test__basic_mmap()
|
D | bitmap.c | 21 set_bit(perf_cpu_map__cpu(map, i).cpu, bm); in get_bitmap()
|
/linux-5.19.10/tools/lib/perf/include/perf/ |
D | cpumap.h | 23 LIBPERF_API struct perf_cpu perf_cpu_map__cpu(const struct perf_cpu_map *cpus, int idx); 30 for ((idx) = 0, (cpu) = perf_cpu_map__cpu(cpus, idx); \ 32 (idx)++, (cpu) = perf_cpu_map__cpu(cpus, idx))
|
/linux-5.19.10/tools/perf/util/ |
D | record.c | 109 if (perf_cpu_map__cpu(evlist->core.user_requested_cpus, 0).cpu < 0) in evlist__config() 251 cpu = perf_cpu_map__cpu(cpus, 0); in evlist__can_select_event() 255 cpu = perf_cpu_map__cpu(evlist->core.user_requested_cpus, 0); in evlist__can_select_event()
|
D | perf_api_probe.c | 70 cpu = perf_cpu_map__cpu(cpus, 0); in perf_probe_api() 147 cpu = perf_cpu_map__cpu(cpus, 0); in perf_can_record_cpu_wide()
|
D | cpumap.h | 59 return perf_cpu_map__nr(cpus) == 1 && perf_cpu_map__cpu(cpus, 0).cpu == -1; in cpu_map__is_dummy()
|
D | bpf_ftrace.c | 66 cpu = perf_cpu_map__cpu(ftrace->evlist->core.user_requested_cpus, i).cpu; in perf_ftrace__latency_prepare_bpf()
|
D | cpumap.c | 548 struct perf_cpu last_cpu = perf_cpu_map__cpu(map, map->nr - 1); in cpu_map__snprint_mask() 560 cpu = perf_cpu_map__cpu(map, i).cpu; in cpu_map__snprint_mask()
|
D | bpf_off_cpu.c | 182 cpu = perf_cpu_map__cpu(evlist->core.user_requested_cpus, i).cpu; in off_cpu_prepare()
|
D | cputopo.c | 337 if (load_numa_node(&tp->nodes[i], perf_cpu_map__cpu(node_map, i).cpu)) { in numa_topology__new()
|
D | mmap.c | 257 cpu = perf_cpu_map__cpu(cpu_map, idx); /* map c index to online cpu index */ in build_node_mask()
|
/linux-5.19.10/tools/lib/perf/ |
D | libperf.map | 11 perf_cpu_map__cpu;
|
D | evlist.c | 390 sid->cpu = perf_cpu_map__cpu(evsel->cpus, cpu); in perf_evsel__set_sid_idx() 438 struct perf_cpu evlist_cpu = perf_cpu_map__cpu(evlist->all_cpus, cpu_idx); in mmap_per_evsel()
|
D | cpumap.c | 260 struct perf_cpu perf_cpu_map__cpu(const struct perf_cpu_map *cpus, int idx) in perf_cpu_map__cpu() function
|
/linux-5.19.10/tools/perf/bench/ |
D | futex-wake.c | 113 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in block_threads()
|
D | futex-hash.c | 189 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in bench_futex_hash()
|
D | futex-lock-pi.c | 145 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in create_threads()
|
D | futex-requeue.c | 140 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in block_threads()
|
D | futex-wake-parallel.c | 161 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in block_threads()
|
D | epoll-ctl.c | 263 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, in do_threads()
|
/linux-5.19.10/tools/lib/perf/Documentation/ |
D | libperf.txt | 47 int perf_cpu_map__cpu(const struct perf_cpu_map *cpus, int idx);
|
/linux-5.19.10/tools/perf/ |
D | builtin-stat.c | 238 if (perf_cpu_map__cpu(a->core.cpus, i).cpu != in cpus_map_matched() 239 perf_cpu_map__cpu(b->core.cpus, i).cpu) in cpus_map_matched() 333 struct perf_cpu cpu = perf_cpu_map__cpu(evsel__cpus(counter), cpu_map_idx); in evsel__write_stat_event() 415 perf_cpu_map__cpu(evsel__cpus(counter), in read_counter_cpu()
|