Lines Matching refs:thread

57 		int idx, thread;  in perf_evsel__alloc_fd()  local
60 for (thread = 0; thread < nthreads; thread++) { in perf_evsel__alloc_fd()
61 int *fd = FD(evsel, idx, thread); in perf_evsel__alloc_fd()
87 static int get_group_fd(struct perf_evsel *evsel, int cpu_map_idx, int thread, int *group_fd) in get_group_fd() argument
104 fd = FD(leader, cpu_map_idx, thread); in get_group_fd()
117 int idx, thread, err = 0; in perf_evsel__open() local
148 for (thread = 0; thread < threads->nr; thread++) { in perf_evsel__open()
151 evsel_fd = FD(evsel, idx, thread); in perf_evsel__open()
157 err = get_group_fd(evsel, idx, thread, &group_fd); in perf_evsel__open()
162 threads->map[thread].pid, in perf_evsel__open()
182 int thread; in perf_evsel__close_fd_cpu() local
184 for (thread = 0; thread < xyarray__max_y(evsel->fd); ++thread) { in perf_evsel__close_fd_cpu()
185 int *fd = FD(evsel, cpu_map_idx, thread); in perf_evsel__close_fd_cpu()
225 int idx, thread; in perf_evsel__munmap() local
231 for (thread = 0; thread < xyarray__max_y(evsel->fd); thread++) { in perf_evsel__munmap()
232 int *fd = FD(evsel, idx, thread); in perf_evsel__munmap()
237 perf_mmap__munmap(MMAP(evsel, idx, thread)); in perf_evsel__munmap()
247 int ret, idx, thread; in perf_evsel__mmap() local
260 for (thread = 0; thread < xyarray__max_y(evsel->fd); thread++) { in perf_evsel__mmap()
261 int *fd = FD(evsel, idx, thread); in perf_evsel__mmap()
268 map = MMAP(evsel, idx, thread); in perf_evsel__mmap()
282 void *perf_evsel__mmap_base(struct perf_evsel *evsel, int cpu_map_idx, int thread) in perf_evsel__mmap_base() argument
284 int *fd = FD(evsel, cpu_map_idx, thread); in perf_evsel__mmap_base()
286 if (fd == NULL || *fd < 0 || MMAP(evsel, cpu_map_idx, thread) == NULL) in perf_evsel__mmap_base()
289 return MMAP(evsel, cpu_map_idx, thread)->base; in perf_evsel__mmap_base()
322 int thread, struct perf_counts_values *count) in perf_evsel__read_group() argument
325 int *fd = FD(evsel, cpu_map_idx, thread); in perf_evsel__read_group()
387 int perf_evsel__read(struct perf_evsel *evsel, int cpu_map_idx, int thread, in perf_evsel__read() argument
391 int *fd = FD(evsel, cpu_map_idx, thread); in perf_evsel__read()
401 return perf_evsel__read_group(evsel, cpu_map_idx, thread, count); in perf_evsel__read()
403 if (MMAP(evsel, cpu_map_idx, thread) && in perf_evsel__read()
405 !perf_mmap__read_self(MMAP(evsel, cpu_map_idx, thread), count)) in perf_evsel__read()
416 int cpu_map_idx, int thread) in perf_evsel__ioctl() argument
418 int *fd = FD(evsel, cpu_map_idx, thread); in perf_evsel__ioctl()
430 int thread; in perf_evsel__run_ioctl() local
432 for (thread = 0; thread < xyarray__max_y(evsel->fd); thread++) { in perf_evsel__run_ioctl()
433 int err = perf_evsel__ioctl(evsel, ioc, arg, cpu_map_idx, thread); in perf_evsel__run_ioctl()
447 int perf_evsel__enable_thread(struct perf_evsel *evsel, int thread) in perf_evsel__enable_thread() argument
454 err = perf_evsel__ioctl(evsel, PERF_EVENT_IOC_ENABLE, NULL, idx, thread); in perf_evsel__enable_thread()