/linux-6.1.9/tools/perf/util/ |
D | evlist-hybrid.c | 127 if (perf_cpu_map__nr(matched_cpus) > 0 && in evlist__fix_hybrid_cpus() 128 (perf_cpu_map__nr(unmatched_cpus) > 0 || in evlist__fix_hybrid_cpus() 129 perf_cpu_map__nr(matched_cpus) < perf_cpu_map__nr(cpus) || in evlist__fix_hybrid_cpus() 130 perf_cpu_map__nr(matched_cpus) < perf_cpu_map__nr(pmu->cpus))) { in evlist__fix_hybrid_cpus() 136 if (perf_cpu_map__nr(unmatched_cpus) > 0) { in evlist__fix_hybrid_cpus() 143 if (perf_cpu_map__nr(matched_cpus) == 0) { in evlist__fix_hybrid_cpus()
|
D | top.c | 98 perf_cpu_map__nr(top->evlist->core.user_requested_cpus) > 1 in perf_top__header_snprintf() 106 perf_cpu_map__nr(top->evlist->core.user_requested_cpus), in perf_top__header_snprintf() 107 perf_cpu_map__nr(top->evlist->core.user_requested_cpus) > 1 in perf_top__header_snprintf()
|
D | counts.c | 64 evsel->counts = perf_counts__new(perf_cpu_map__nr(cpus), nthreads); in evsel__alloc_counts()
|
D | cpumap.h | 61 return perf_cpu_map__nr(cpus) == 1 && perf_cpu_map__cpu(cpus, 0).cpu == -1; in cpu_map__is_dummy()
|
D | bpf_ftrace.c | 41 ncpus = perf_cpu_map__nr(ftrace->evlist->core.user_requested_cpus); in perf_ftrace__latency_prepare_bpf()
|
D | bpf_lock_contention.c | 44 ncpus = perf_cpu_map__nr(evlist->core.user_requested_cpus); in lock_contention_prepare()
|
D | bpf_kwork.c | 165 for (i = 0; i < perf_cpu_map__nr(map); i++) { in setup_filters()
|
/linux-6.1.9/tools/lib/perf/include/perf/ |
D | cpumap.h | 24 LIBPERF_API int perf_cpu_map__nr(const struct perf_cpu_map *cpus); 31 (idx) < perf_cpu_map__nr(cpus); \ 35 for ((idx) = 0; (idx) < perf_cpu_map__nr(cpus); (idx)++)
|
/linux-6.1.9/tools/perf/tests/ |
D | cpumap.c | 40 TEST_ASSERT_VAL("wrong nr", perf_cpu_map__nr(map) == 20); in process_event_mask() 68 TEST_ASSERT_VAL("wrong nr", perf_cpu_map__nr(map) == 2); in process_event_cpus() 94 TEST_ASSERT_VAL("wrong nr", perf_cpu_map__nr(map) == 256); in process_event_range_cpus() 166 TEST_ASSERT_VAL("failed to merge map: bad nr", perf_cpu_map__nr(c) == 5); in test__cpu_map_merge()
|
D | topology.c | 136 for (i = 0; i < perf_cpu_map__nr(map); i++) { in check_cpu_topology() 154 for (i = 0; i < perf_cpu_map__nr(map); i++) { in check_cpu_topology() 170 for (i = 0; i < perf_cpu_map__nr(map); i++) { in check_cpu_topology() 186 for (i = 0; i < perf_cpu_map__nr(map); i++) { in check_cpu_topology() 200 for (i = 0; i < perf_cpu_map__nr(map); i++) { in check_cpu_topology()
|
D | bitmap.c | 20 for (i = 0; i < perf_cpu_map__nr(map); i++) in get_bitmap()
|
D | event_update.c | 72 TEST_ASSERT_VAL("wrong cpus", perf_cpu_map__nr(map) == 3); in process_event_cpus()
|
/linux-6.1.9/tools/perf/bench/ |
D | futex-wake.c | 104 int nrcpus = perf_cpu_map__nr(cpu); in block_threads() 114 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in block_threads() 165 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_wake()
|
D | futex-hash.c | 156 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_hash() 177 nrcpus = perf_cpu_map__nr(cpu); in bench_futex_hash() 190 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in bench_futex_hash()
|
D | futex-lock-pi.c | 126 int nrcpus = perf_cpu_map__nr(cpu); in create_threads() 146 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in create_threads() 188 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_lock_pi()
|
D | futex-requeue.c | 129 int nrcpus = perf_cpu_map__nr(cpu); in block_threads() 141 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in block_threads() 190 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_requeue()
|
D | futex-wake-parallel.c | 150 int nrcpus = perf_cpu_map__nr(cpu); in block_threads() 162 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, size, cpuset); in block_threads() 267 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_wake_parallel()
|
D | epoll-ctl.c | 235 nrcpus = perf_cpu_map__nr(cpu); in do_threads() 264 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, in do_threads() 350 nthreads = perf_cpu_map__nr(cpu); in bench_epoll_ctl()
|
D | evlist-open-close.c | 74 cnt += evsel->core.threads->nr * perf_cpu_map__nr(evsel->core.cpus); in evlist__count_evsel_fds() 154 printf(" Number of cpus:\t%d\n", perf_cpu_map__nr(evlist->core.user_requested_cpus)); in bench_evlist_open_close__run()
|
D | epoll-wait.c | 312 nrcpus = perf_cpu_map__nr(cpu); in do_threads() 353 CPU_SET_S(perf_cpu_map__cpu(cpu, i % perf_cpu_map__nr(cpu)).cpu, in do_threads() 469 nthreads = perf_cpu_map__nr(cpu) - 1; in bench_epoll_wait()
|
/linux-6.1.9/tools/lib/perf/ |
D | libperf.map | 10 perf_cpu_map__nr;
|
D | evlist.c | 306 int nr_cpus = perf_cpu_map__nr(evlist->all_cpus); in perf_evlist__alloc_pollfd() 532 int nr_cpus = perf_cpu_map__nr(evlist->all_cpus); in mmap_per_thread() 574 int nr_cpus = perf_cpu_map__nr(evlist->all_cpus); in mmap_per_cpu() 606 nr_mmaps = perf_cpu_map__nr(evlist->all_cpus); in perf_evlist__nr_mmaps()
|
D | evsel.c | 144 perf_evsel__alloc_fd(evsel, perf_cpu_map__nr(cpus), threads->nr) < 0) in perf_evsel__open() 491 for (i = 0; i < perf_cpu_map__nr(evsel->cpus) && !err; i++) in perf_evsel__apply_filter()
|
/linux-6.1.9/tools/perf/arch/arm64/util/ |
D | header.c | 28 for (cpu = 0; cpu < perf_cpu_map__nr(cpus); cpu++) { in _get_cpuid()
|
/linux-6.1.9/tools/lib/perf/tests/ |
D | test-evlist.c | 77 for (idx = 0; idx < perf_cpu_map__nr(cpus); idx++) { in test_stat_cpu() 414 __T("failed count", count >= perf_cpu_map__nr(cpus)); in test_mmap_cpus()
|