1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * builtin-inject.c
4 *
5 * Builtin inject command: Examine the live mode (stdin) event stream
6 * and repipe it to stdout while optionally injecting additional
7 * events into it.
8 */
9 #include "builtin.h"
10
11 #include "util/color.h"
12 #include "util/dso.h"
13 #include "util/vdso.h"
14 #include "util/evlist.h"
15 #include "util/evsel.h"
16 #include "util/map.h"
17 #include "util/session.h"
18 #include "util/tool.h"
19 #include "util/debug.h"
20 #include "util/build-id.h"
21 #include "util/data.h"
22 #include "util/auxtrace.h"
23 #include "util/jit.h"
24 #include "util/symbol.h"
25 #include "util/synthetic-events.h"
26 #include "util/thread.h"
27 #include "util/namespaces.h"
28 #include "util/util.h"
29
30 #include <internal/lib.h>
31
32 #include <linux/err.h>
33 #include <subcmd/parse-options.h>
34 #include <uapi/linux/mman.h> /* To get things like MAP_HUGETLB even on older libc headers */
35
36 #include <linux/list.h>
37 #include <linux/string.h>
38 #include <errno.h>
39 #include <signal.h>
40
41 struct perf_inject {
42 struct perf_tool tool;
43 struct perf_session *session;
44 bool build_ids;
45 bool build_id_all;
46 bool sched_stat;
47 bool have_auxtrace;
48 bool strip;
49 bool jit_mode;
50 bool in_place_update;
51 bool in_place_update_dry_run;
52 bool is_pipe;
53 bool copy_kcore_dir;
54 const char *input_name;
55 struct perf_data output;
56 u64 bytes_written;
57 u64 aux_id;
58 struct list_head samples;
59 struct itrace_synth_opts itrace_synth_opts;
60 char event_copy[PERF_SAMPLE_MAX_SIZE];
61 struct perf_file_section secs[HEADER_FEAT_BITS];
62 };
63
64 struct event_entry {
65 struct list_head node;
66 u32 tid;
67 union perf_event event[];
68 };
69
70 static int dso__inject_build_id(struct dso *dso, struct perf_tool *tool,
71 struct machine *machine, u8 cpumode, u32 flags);
72
output_bytes(struct perf_inject * inject,void * buf,size_t sz)73 static int output_bytes(struct perf_inject *inject, void *buf, size_t sz)
74 {
75 ssize_t size;
76
77 size = perf_data__write(&inject->output, buf, sz);
78 if (size < 0)
79 return -errno;
80
81 inject->bytes_written += size;
82 return 0;
83 }
84
perf_event__repipe_synth(struct perf_tool * tool,union perf_event * event)85 static int perf_event__repipe_synth(struct perf_tool *tool,
86 union perf_event *event)
87 {
88 struct perf_inject *inject = container_of(tool, struct perf_inject,
89 tool);
90
91 return output_bytes(inject, event, event->header.size);
92 }
93
perf_event__repipe_oe_synth(struct perf_tool * tool,union perf_event * event,struct ordered_events * oe __maybe_unused)94 static int perf_event__repipe_oe_synth(struct perf_tool *tool,
95 union perf_event *event,
96 struct ordered_events *oe __maybe_unused)
97 {
98 return perf_event__repipe_synth(tool, event);
99 }
100
101 #ifdef HAVE_JITDUMP
perf_event__drop_oe(struct perf_tool * tool __maybe_unused,union perf_event * event __maybe_unused,struct ordered_events * oe __maybe_unused)102 static int perf_event__drop_oe(struct perf_tool *tool __maybe_unused,
103 union perf_event *event __maybe_unused,
104 struct ordered_events *oe __maybe_unused)
105 {
106 return 0;
107 }
108 #endif
109
perf_event__repipe_op2_synth(struct perf_session * session,union perf_event * event)110 static int perf_event__repipe_op2_synth(struct perf_session *session,
111 union perf_event *event)
112 {
113 return perf_event__repipe_synth(session->tool, event);
114 }
115
perf_event__repipe_op4_synth(struct perf_session * session,union perf_event * event,u64 data __maybe_unused,const char * str __maybe_unused)116 static int perf_event__repipe_op4_synth(struct perf_session *session,
117 union perf_event *event,
118 u64 data __maybe_unused,
119 const char *str __maybe_unused)
120 {
121 return perf_event__repipe_synth(session->tool, event);
122 }
123
perf_event__repipe_attr(struct perf_tool * tool,union perf_event * event,struct evlist ** pevlist)124 static int perf_event__repipe_attr(struct perf_tool *tool,
125 union perf_event *event,
126 struct evlist **pevlist)
127 {
128 struct perf_inject *inject = container_of(tool, struct perf_inject,
129 tool);
130 int ret;
131
132 ret = perf_event__process_attr(tool, event, pevlist);
133 if (ret)
134 return ret;
135
136 if (!inject->is_pipe)
137 return 0;
138
139 return perf_event__repipe_synth(tool, event);
140 }
141
perf_event__repipe_event_update(struct perf_tool * tool,union perf_event * event,struct evlist ** pevlist __maybe_unused)142 static int perf_event__repipe_event_update(struct perf_tool *tool,
143 union perf_event *event,
144 struct evlist **pevlist __maybe_unused)
145 {
146 return perf_event__repipe_synth(tool, event);
147 }
148
149 #ifdef HAVE_AUXTRACE_SUPPORT
150
copy_bytes(struct perf_inject * inject,int fd,off_t size)151 static int copy_bytes(struct perf_inject *inject, int fd, off_t size)
152 {
153 char buf[4096];
154 ssize_t ssz;
155 int ret;
156
157 while (size > 0) {
158 ssz = read(fd, buf, min(size, (off_t)sizeof(buf)));
159 if (ssz < 0)
160 return -errno;
161 ret = output_bytes(inject, buf, ssz);
162 if (ret)
163 return ret;
164 size -= ssz;
165 }
166
167 return 0;
168 }
169
perf_event__repipe_auxtrace(struct perf_session * session,union perf_event * event)170 static s64 perf_event__repipe_auxtrace(struct perf_session *session,
171 union perf_event *event)
172 {
173 struct perf_tool *tool = session->tool;
174 struct perf_inject *inject = container_of(tool, struct perf_inject,
175 tool);
176 int ret;
177
178 inject->have_auxtrace = true;
179
180 if (!inject->output.is_pipe) {
181 off_t offset;
182
183 offset = lseek(inject->output.file.fd, 0, SEEK_CUR);
184 if (offset == -1)
185 return -errno;
186 ret = auxtrace_index__auxtrace_event(&session->auxtrace_index,
187 event, offset);
188 if (ret < 0)
189 return ret;
190 }
191
192 if (perf_data__is_pipe(session->data) || !session->one_mmap) {
193 ret = output_bytes(inject, event, event->header.size);
194 if (ret < 0)
195 return ret;
196 ret = copy_bytes(inject, perf_data__fd(session->data),
197 event->auxtrace.size);
198 } else {
199 ret = output_bytes(inject, event,
200 event->header.size + event->auxtrace.size);
201 }
202 if (ret < 0)
203 return ret;
204
205 return event->auxtrace.size;
206 }
207
208 #else
209
210 static s64
perf_event__repipe_auxtrace(struct perf_session * session __maybe_unused,union perf_event * event __maybe_unused)211 perf_event__repipe_auxtrace(struct perf_session *session __maybe_unused,
212 union perf_event *event __maybe_unused)
213 {
214 pr_err("AUX area tracing not supported\n");
215 return -EINVAL;
216 }
217
218 #endif
219
perf_event__repipe(struct perf_tool * tool,union perf_event * event,struct perf_sample * sample __maybe_unused,struct machine * machine __maybe_unused)220 static int perf_event__repipe(struct perf_tool *tool,
221 union perf_event *event,
222 struct perf_sample *sample __maybe_unused,
223 struct machine *machine __maybe_unused)
224 {
225 return perf_event__repipe_synth(tool, event);
226 }
227
perf_event__drop(struct perf_tool * tool __maybe_unused,union perf_event * event __maybe_unused,struct perf_sample * sample __maybe_unused,struct machine * machine __maybe_unused)228 static int perf_event__drop(struct perf_tool *tool __maybe_unused,
229 union perf_event *event __maybe_unused,
230 struct perf_sample *sample __maybe_unused,
231 struct machine *machine __maybe_unused)
232 {
233 return 0;
234 }
235
perf_event__drop_aux(struct perf_tool * tool,union perf_event * event __maybe_unused,struct perf_sample * sample,struct machine * machine __maybe_unused)236 static int perf_event__drop_aux(struct perf_tool *tool,
237 union perf_event *event __maybe_unused,
238 struct perf_sample *sample,
239 struct machine *machine __maybe_unused)
240 {
241 struct perf_inject *inject = container_of(tool, struct perf_inject, tool);
242
243 if (!inject->aux_id)
244 inject->aux_id = sample->id;
245
246 return 0;
247 }
248
249 static union perf_event *
perf_inject__cut_auxtrace_sample(struct perf_inject * inject,union perf_event * event,struct perf_sample * sample)250 perf_inject__cut_auxtrace_sample(struct perf_inject *inject,
251 union perf_event *event,
252 struct perf_sample *sample)
253 {
254 size_t sz1 = sample->aux_sample.data - (void *)event;
255 size_t sz2 = event->header.size - sample->aux_sample.size - sz1;
256 union perf_event *ev = (union perf_event *)inject->event_copy;
257
258 if (sz1 > event->header.size || sz2 > event->header.size ||
259 sz1 + sz2 > event->header.size ||
260 sz1 < sizeof(struct perf_event_header) + sizeof(u64))
261 return event;
262
263 memcpy(ev, event, sz1);
264 memcpy((void *)ev + sz1, (void *)event + event->header.size - sz2, sz2);
265 ev->header.size = sz1 + sz2;
266 ((u64 *)((void *)ev + sz1))[-1] = 0;
267
268 return ev;
269 }
270
271 typedef int (*inject_handler)(struct perf_tool *tool,
272 union perf_event *event,
273 struct perf_sample *sample,
274 struct evsel *evsel,
275 struct machine *machine);
276
perf_event__repipe_sample(struct perf_tool * tool,union perf_event * event,struct perf_sample * sample,struct evsel * evsel,struct machine * machine)277 static int perf_event__repipe_sample(struct perf_tool *tool,
278 union perf_event *event,
279 struct perf_sample *sample,
280 struct evsel *evsel,
281 struct machine *machine)
282 {
283 struct perf_inject *inject = container_of(tool, struct perf_inject,
284 tool);
285
286 if (evsel && evsel->handler) {
287 inject_handler f = evsel->handler;
288 return f(tool, event, sample, evsel, machine);
289 }
290
291 build_id__mark_dso_hit(tool, event, sample, evsel, machine);
292
293 if (inject->itrace_synth_opts.set && sample->aux_sample.size)
294 event = perf_inject__cut_auxtrace_sample(inject, event, sample);
295
296 return perf_event__repipe_synth(tool, event);
297 }
298
perf_event__repipe_mmap(struct perf_tool * tool,union perf_event * event,struct perf_sample * sample,struct machine * machine)299 static int perf_event__repipe_mmap(struct perf_tool *tool,
300 union perf_event *event,
301 struct perf_sample *sample,
302 struct machine *machine)
303 {
304 int err;
305
306 err = perf_event__process_mmap(tool, event, sample, machine);
307 perf_event__repipe(tool, event, sample, machine);
308
309 return err;
310 }
311
312 #ifdef HAVE_JITDUMP
perf_event__jit_repipe_mmap(struct perf_tool * tool,union perf_event * event,struct perf_sample * sample,struct machine * machine)313 static int perf_event__jit_repipe_mmap(struct perf_tool *tool,
314 union perf_event *event,
315 struct perf_sample *sample,
316 struct machine *machine)
317 {
318 struct perf_inject *inject = container_of(tool, struct perf_inject, tool);
319 u64 n = 0;
320 int ret;
321
322 /*
323 * if jit marker, then inject jit mmaps and generate ELF images
324 */
325 ret = jit_process(inject->session, &inject->output, machine,
326 event->mmap.filename, event->mmap.pid, event->mmap.tid, &n);
327 if (ret < 0)
328 return ret;
329 if (ret) {
330 inject->bytes_written += n;
331 return 0;
332 }
333 return perf_event__repipe_mmap(tool, event, sample, machine);
334 }
335 #endif
336
findnew_dso(int pid,int tid,const char * filename,struct dso_id * id,struct machine * machine)337 static struct dso *findnew_dso(int pid, int tid, const char *filename,
338 struct dso_id *id, struct machine *machine)
339 {
340 struct thread *thread;
341 struct nsinfo *nsi = NULL;
342 struct nsinfo *nnsi;
343 struct dso *dso;
344 bool vdso;
345
346 thread = machine__findnew_thread(machine, pid, tid);
347 if (thread == NULL) {
348 pr_err("cannot find or create a task %d/%d.\n", tid, pid);
349 return NULL;
350 }
351
352 vdso = is_vdso_map(filename);
353 nsi = nsinfo__get(thread->nsinfo);
354
355 if (vdso) {
356 /* The vdso maps are always on the host and not the
357 * container. Ensure that we don't use setns to look
358 * them up.
359 */
360 nnsi = nsinfo__copy(nsi);
361 if (nnsi) {
362 nsinfo__put(nsi);
363 nsinfo__clear_need_setns(nnsi);
364 nsi = nnsi;
365 }
366 dso = machine__findnew_vdso(machine, thread);
367 } else {
368 dso = machine__findnew_dso_id(machine, filename, id);
369 }
370
371 if (dso) {
372 nsinfo__put(dso->nsinfo);
373 dso->nsinfo = nsi;
374 } else
375 nsinfo__put(nsi);
376
377 thread__put(thread);
378 return dso;
379 }
380
perf_event__repipe_buildid_mmap(struct perf_tool * tool,union perf_event * event,struct perf_sample * sample,struct machine * machine)381 static int perf_event__repipe_buildid_mmap(struct perf_tool *tool,
382 union perf_event *event,
383 struct perf_sample *sample,
384 struct machine *machine)
385 {
386 struct dso *dso;
387
388 dso = findnew_dso(event->mmap.pid, event->mmap.tid,
389 event->mmap.filename, NULL, machine);
390
391 if (dso && !dso->hit) {
392 dso->hit = 1;
393 dso__inject_build_id(dso, tool, machine, sample->cpumode, 0);
394 }
395 dso__put(dso);
396
397 return perf_event__repipe(tool, event, sample, machine);
398 }
399
perf_event__repipe_mmap2(struct perf_tool * tool,union perf_event * event,struct perf_sample * sample,struct machine * machine)400 static int perf_event__repipe_mmap2(struct perf_tool *tool,
401 union perf_event *event,
402 struct perf_sample *sample,
403 struct machine *machine)
404 {
405 int err;
406
407 err = perf_event__process_mmap2(tool, event, sample, machine);
408 perf_event__repipe(tool, event, sample, machine);
409
410 if (event->header.misc & PERF_RECORD_MISC_MMAP_BUILD_ID) {
411 struct dso *dso;
412
413 dso = findnew_dso(event->mmap2.pid, event->mmap2.tid,
414 event->mmap2.filename, NULL, machine);
415 if (dso) {
416 /* mark it not to inject build-id */
417 dso->hit = 1;
418 }
419 dso__put(dso);
420 }
421
422 return err;
423 }
424
425 #ifdef HAVE_JITDUMP
perf_event__jit_repipe_mmap2(struct perf_tool * tool,union perf_event * event,struct perf_sample * sample,struct machine * machine)426 static int perf_event__jit_repipe_mmap2(struct perf_tool *tool,
427 union perf_event *event,
428 struct perf_sample *sample,
429 struct machine *machine)
430 {
431 struct perf_inject *inject = container_of(tool, struct perf_inject, tool);
432 u64 n = 0;
433 int ret;
434
435 /*
436 * if jit marker, then inject jit mmaps and generate ELF images
437 */
438 ret = jit_process(inject->session, &inject->output, machine,
439 event->mmap2.filename, event->mmap2.pid, event->mmap2.tid, &n);
440 if (ret < 0)
441 return ret;
442 if (ret) {
443 inject->bytes_written += n;
444 return 0;
445 }
446 return perf_event__repipe_mmap2(tool, event, sample, machine);
447 }
448 #endif
449
perf_event__repipe_buildid_mmap2(struct perf_tool * tool,union perf_event * event,struct perf_sample * sample,struct machine * machine)450 static int perf_event__repipe_buildid_mmap2(struct perf_tool *tool,
451 union perf_event *event,
452 struct perf_sample *sample,
453 struct machine *machine)
454 {
455 struct dso_id dso_id = {
456 .maj = event->mmap2.maj,
457 .min = event->mmap2.min,
458 .ino = event->mmap2.ino,
459 .ino_generation = event->mmap2.ino_generation,
460 };
461 struct dso *dso;
462
463 if (event->header.misc & PERF_RECORD_MISC_MMAP_BUILD_ID) {
464 /* cannot use dso_id since it'd have invalid info */
465 dso = findnew_dso(event->mmap2.pid, event->mmap2.tid,
466 event->mmap2.filename, NULL, machine);
467 if (dso) {
468 /* mark it not to inject build-id */
469 dso->hit = 1;
470 }
471 dso__put(dso);
472 return 0;
473 }
474
475 dso = findnew_dso(event->mmap2.pid, event->mmap2.tid,
476 event->mmap2.filename, &dso_id, machine);
477
478 if (dso && !dso->hit) {
479 dso->hit = 1;
480 dso__inject_build_id(dso, tool, machine, sample->cpumode,
481 event->mmap2.flags);
482 }
483 dso__put(dso);
484
485 perf_event__repipe(tool, event, sample, machine);
486
487 return 0;
488 }
489
perf_event__repipe_fork(struct perf_tool * tool,union perf_event * event,struct perf_sample * sample,struct machine * machine)490 static int perf_event__repipe_fork(struct perf_tool *tool,
491 union perf_event *event,
492 struct perf_sample *sample,
493 struct machine *machine)
494 {
495 int err;
496
497 err = perf_event__process_fork(tool, event, sample, machine);
498 perf_event__repipe(tool, event, sample, machine);
499
500 return err;
501 }
502
perf_event__repipe_comm(struct perf_tool * tool,union perf_event * event,struct perf_sample * sample,struct machine * machine)503 static int perf_event__repipe_comm(struct perf_tool *tool,
504 union perf_event *event,
505 struct perf_sample *sample,
506 struct machine *machine)
507 {
508 int err;
509
510 err = perf_event__process_comm(tool, event, sample, machine);
511 perf_event__repipe(tool, event, sample, machine);
512
513 return err;
514 }
515
perf_event__repipe_namespaces(struct perf_tool * tool,union perf_event * event,struct perf_sample * sample,struct machine * machine)516 static int perf_event__repipe_namespaces(struct perf_tool *tool,
517 union perf_event *event,
518 struct perf_sample *sample,
519 struct machine *machine)
520 {
521 int err = perf_event__process_namespaces(tool, event, sample, machine);
522
523 perf_event__repipe(tool, event, sample, machine);
524
525 return err;
526 }
527
perf_event__repipe_exit(struct perf_tool * tool,union perf_event * event,struct perf_sample * sample,struct machine * machine)528 static int perf_event__repipe_exit(struct perf_tool *tool,
529 union perf_event *event,
530 struct perf_sample *sample,
531 struct machine *machine)
532 {
533 int err;
534
535 err = perf_event__process_exit(tool, event, sample, machine);
536 perf_event__repipe(tool, event, sample, machine);
537
538 return err;
539 }
540
perf_event__repipe_tracing_data(struct perf_session * session,union perf_event * event)541 static int perf_event__repipe_tracing_data(struct perf_session *session,
542 union perf_event *event)
543 {
544 perf_event__repipe_synth(session->tool, event);
545
546 return perf_event__process_tracing_data(session, event);
547 }
548
dso__read_build_id(struct dso * dso)549 static int dso__read_build_id(struct dso *dso)
550 {
551 struct nscookie nsc;
552
553 if (dso->has_build_id)
554 return 0;
555
556 nsinfo__mountns_enter(dso->nsinfo, &nsc);
557 if (filename__read_build_id(dso->long_name, &dso->bid) > 0)
558 dso->has_build_id = true;
559 else if (dso->nsinfo) {
560 char *new_name;
561
562 new_name = filename_with_chroot(dso->nsinfo->pid,
563 dso->long_name);
564 if (new_name && filename__read_build_id(new_name, &dso->bid) > 0)
565 dso->has_build_id = true;
566 free(new_name);
567 }
568 nsinfo__mountns_exit(&nsc);
569
570 return dso->has_build_id ? 0 : -1;
571 }
572
dso__inject_build_id(struct dso * dso,struct perf_tool * tool,struct machine * machine,u8 cpumode,u32 flags)573 static int dso__inject_build_id(struct dso *dso, struct perf_tool *tool,
574 struct machine *machine, u8 cpumode, u32 flags)
575 {
576 int err;
577
578 if (is_anon_memory(dso->long_name) || flags & MAP_HUGETLB)
579 return 0;
580 if (is_no_dso_memory(dso->long_name))
581 return 0;
582
583 if (dso__read_build_id(dso) < 0) {
584 pr_debug("no build_id found for %s\n", dso->long_name);
585 return -1;
586 }
587
588 err = perf_event__synthesize_build_id(tool, dso, cpumode,
589 perf_event__repipe, machine);
590 if (err) {
591 pr_err("Can't synthesize build_id event for %s\n", dso->long_name);
592 return -1;
593 }
594
595 return 0;
596 }
597
perf_event__inject_buildid(struct perf_tool * tool,union perf_event * event,struct perf_sample * sample,struct evsel * evsel __maybe_unused,struct machine * machine)598 int perf_event__inject_buildid(struct perf_tool *tool, union perf_event *event,
599 struct perf_sample *sample,
600 struct evsel *evsel __maybe_unused,
601 struct machine *machine)
602 {
603 struct addr_location al;
604 struct thread *thread;
605
606 thread = machine__findnew_thread(machine, sample->pid, sample->tid);
607 if (thread == NULL) {
608 pr_err("problem processing %d event, skipping it.\n",
609 event->header.type);
610 goto repipe;
611 }
612
613 if (thread__find_map(thread, sample->cpumode, sample->ip, &al)) {
614 if (!al.map->dso->hit) {
615 al.map->dso->hit = 1;
616 dso__inject_build_id(al.map->dso, tool, machine,
617 sample->cpumode, al.map->flags);
618 }
619 }
620
621 thread__put(thread);
622 repipe:
623 perf_event__repipe(tool, event, sample, machine);
624 return 0;
625 }
626
perf_inject__sched_process_exit(struct perf_tool * tool,union perf_event * event __maybe_unused,struct perf_sample * sample,struct evsel * evsel __maybe_unused,struct machine * machine __maybe_unused)627 static int perf_inject__sched_process_exit(struct perf_tool *tool,
628 union perf_event *event __maybe_unused,
629 struct perf_sample *sample,
630 struct evsel *evsel __maybe_unused,
631 struct machine *machine __maybe_unused)
632 {
633 struct perf_inject *inject = container_of(tool, struct perf_inject, tool);
634 struct event_entry *ent;
635
636 list_for_each_entry(ent, &inject->samples, node) {
637 if (sample->tid == ent->tid) {
638 list_del_init(&ent->node);
639 free(ent);
640 break;
641 }
642 }
643
644 return 0;
645 }
646
perf_inject__sched_switch(struct perf_tool * tool,union perf_event * event,struct perf_sample * sample,struct evsel * evsel,struct machine * machine)647 static int perf_inject__sched_switch(struct perf_tool *tool,
648 union perf_event *event,
649 struct perf_sample *sample,
650 struct evsel *evsel,
651 struct machine *machine)
652 {
653 struct perf_inject *inject = container_of(tool, struct perf_inject, tool);
654 struct event_entry *ent;
655
656 perf_inject__sched_process_exit(tool, event, sample, evsel, machine);
657
658 ent = malloc(event->header.size + sizeof(struct event_entry));
659 if (ent == NULL) {
660 color_fprintf(stderr, PERF_COLOR_RED,
661 "Not enough memory to process sched switch event!");
662 return -1;
663 }
664
665 ent->tid = sample->tid;
666 memcpy(&ent->event, event, event->header.size);
667 list_add(&ent->node, &inject->samples);
668 return 0;
669 }
670
perf_inject__sched_stat(struct perf_tool * tool,union perf_event * event __maybe_unused,struct perf_sample * sample,struct evsel * evsel,struct machine * machine)671 static int perf_inject__sched_stat(struct perf_tool *tool,
672 union perf_event *event __maybe_unused,
673 struct perf_sample *sample,
674 struct evsel *evsel,
675 struct machine *machine)
676 {
677 struct event_entry *ent;
678 union perf_event *event_sw;
679 struct perf_sample sample_sw;
680 struct perf_inject *inject = container_of(tool, struct perf_inject, tool);
681 u32 pid = evsel__intval(evsel, sample, "pid");
682
683 list_for_each_entry(ent, &inject->samples, node) {
684 if (pid == ent->tid)
685 goto found;
686 }
687
688 return 0;
689 found:
690 event_sw = &ent->event[0];
691 evsel__parse_sample(evsel, event_sw, &sample_sw);
692
693 sample_sw.period = sample->period;
694 sample_sw.time = sample->time;
695 perf_event__synthesize_sample(event_sw, evsel->core.attr.sample_type,
696 evsel->core.attr.read_format, &sample_sw);
697 build_id__mark_dso_hit(tool, event_sw, &sample_sw, evsel, machine);
698 return perf_event__repipe(tool, event_sw, &sample_sw, machine);
699 }
700
sig_handler(int sig __maybe_unused)701 static void sig_handler(int sig __maybe_unused)
702 {
703 session_done = 1;
704 }
705
evsel__check_stype(struct evsel * evsel,u64 sample_type,const char * sample_msg)706 static int evsel__check_stype(struct evsel *evsel, u64 sample_type, const char *sample_msg)
707 {
708 struct perf_event_attr *attr = &evsel->core.attr;
709 const char *name = evsel__name(evsel);
710
711 if (!(attr->sample_type & sample_type)) {
712 pr_err("Samples for %s event do not have %s attribute set.",
713 name, sample_msg);
714 return -EINVAL;
715 }
716
717 return 0;
718 }
719
drop_sample(struct perf_tool * tool __maybe_unused,union perf_event * event __maybe_unused,struct perf_sample * sample __maybe_unused,struct evsel * evsel __maybe_unused,struct machine * machine __maybe_unused)720 static int drop_sample(struct perf_tool *tool __maybe_unused,
721 union perf_event *event __maybe_unused,
722 struct perf_sample *sample __maybe_unused,
723 struct evsel *evsel __maybe_unused,
724 struct machine *machine __maybe_unused)
725 {
726 return 0;
727 }
728
strip_init(struct perf_inject * inject)729 static void strip_init(struct perf_inject *inject)
730 {
731 struct evlist *evlist = inject->session->evlist;
732 struct evsel *evsel;
733
734 inject->tool.context_switch = perf_event__drop;
735
736 evlist__for_each_entry(evlist, evsel)
737 evsel->handler = drop_sample;
738 }
739
parse_vm_time_correlation(const struct option * opt,const char * str,int unset)740 static int parse_vm_time_correlation(const struct option *opt, const char *str, int unset)
741 {
742 struct perf_inject *inject = opt->value;
743 const char *args;
744 char *dry_run;
745
746 if (unset)
747 return 0;
748
749 inject->itrace_synth_opts.set = true;
750 inject->itrace_synth_opts.vm_time_correlation = true;
751 inject->in_place_update = true;
752
753 if (!str)
754 return 0;
755
756 dry_run = skip_spaces(str);
757 if (!strncmp(dry_run, "dry-run", strlen("dry-run"))) {
758 inject->itrace_synth_opts.vm_tm_corr_dry_run = true;
759 inject->in_place_update_dry_run = true;
760 args = dry_run + strlen("dry-run");
761 } else {
762 args = str;
763 }
764
765 inject->itrace_synth_opts.vm_tm_corr_args = strdup(args);
766
767 return inject->itrace_synth_opts.vm_tm_corr_args ? 0 : -ENOMEM;
768 }
769
save_section_info_cb(struct perf_file_section * section,struct perf_header * ph __maybe_unused,int feat,int fd __maybe_unused,void * data)770 static int save_section_info_cb(struct perf_file_section *section,
771 struct perf_header *ph __maybe_unused,
772 int feat, int fd __maybe_unused, void *data)
773 {
774 struct perf_inject *inject = data;
775
776 inject->secs[feat] = *section;
777 return 0;
778 }
779
save_section_info(struct perf_inject * inject)780 static int save_section_info(struct perf_inject *inject)
781 {
782 struct perf_header *header = &inject->session->header;
783 int fd = perf_data__fd(inject->session->data);
784
785 return perf_header__process_sections(header, fd, inject, save_section_info_cb);
786 }
787
keep_feat(int feat)788 static bool keep_feat(int feat)
789 {
790 switch (feat) {
791 /* Keep original information that describes the machine or software */
792 case HEADER_TRACING_DATA:
793 case HEADER_HOSTNAME:
794 case HEADER_OSRELEASE:
795 case HEADER_VERSION:
796 case HEADER_ARCH:
797 case HEADER_NRCPUS:
798 case HEADER_CPUDESC:
799 case HEADER_CPUID:
800 case HEADER_TOTAL_MEM:
801 case HEADER_CPU_TOPOLOGY:
802 case HEADER_NUMA_TOPOLOGY:
803 case HEADER_PMU_MAPPINGS:
804 case HEADER_CACHE:
805 case HEADER_MEM_TOPOLOGY:
806 case HEADER_CLOCKID:
807 case HEADER_BPF_PROG_INFO:
808 case HEADER_BPF_BTF:
809 case HEADER_CPU_PMU_CAPS:
810 case HEADER_CLOCK_DATA:
811 case HEADER_HYBRID_TOPOLOGY:
812 case HEADER_HYBRID_CPU_PMU_CAPS:
813 return true;
814 /* Information that can be updated */
815 case HEADER_BUILD_ID:
816 case HEADER_CMDLINE:
817 case HEADER_EVENT_DESC:
818 case HEADER_BRANCH_STACK:
819 case HEADER_GROUP_DESC:
820 case HEADER_AUXTRACE:
821 case HEADER_STAT:
822 case HEADER_SAMPLE_TIME:
823 case HEADER_DIR_FORMAT:
824 case HEADER_COMPRESSED:
825 default:
826 return false;
827 };
828 }
829
read_file(int fd,u64 offs,void * buf,size_t sz)830 static int read_file(int fd, u64 offs, void *buf, size_t sz)
831 {
832 ssize_t ret = preadn(fd, buf, sz, offs);
833
834 if (ret < 0)
835 return -errno;
836 if ((size_t)ret != sz)
837 return -EINVAL;
838 return 0;
839 }
840
feat_copy(struct perf_inject * inject,int feat,struct feat_writer * fw)841 static int feat_copy(struct perf_inject *inject, int feat, struct feat_writer *fw)
842 {
843 int fd = perf_data__fd(inject->session->data);
844 u64 offs = inject->secs[feat].offset;
845 size_t sz = inject->secs[feat].size;
846 void *buf = malloc(sz);
847 int ret;
848
849 if (!buf)
850 return -ENOMEM;
851
852 ret = read_file(fd, offs, buf, sz);
853 if (ret)
854 goto out_free;
855
856 ret = fw->write(fw, buf, sz);
857 out_free:
858 free(buf);
859 return ret;
860 }
861
862 struct inject_fc {
863 struct feat_copier fc;
864 struct perf_inject *inject;
865 };
866
feat_copy_cb(struct feat_copier * fc,int feat,struct feat_writer * fw)867 static int feat_copy_cb(struct feat_copier *fc, int feat, struct feat_writer *fw)
868 {
869 struct inject_fc *inj_fc = container_of(fc, struct inject_fc, fc);
870 struct perf_inject *inject = inj_fc->inject;
871 int ret;
872
873 if (!inject->secs[feat].offset ||
874 !keep_feat(feat))
875 return 0;
876
877 ret = feat_copy(inject, feat, fw);
878 if (ret < 0)
879 return ret;
880
881 return 1; /* Feature section copied */
882 }
883
copy_kcore_dir(struct perf_inject * inject)884 static int copy_kcore_dir(struct perf_inject *inject)
885 {
886 char *cmd;
887 int ret;
888
889 ret = asprintf(&cmd, "cp -r -n %s/kcore_dir* %s >/dev/null 2>&1",
890 inject->input_name, inject->output.path);
891 if (ret < 0)
892 return ret;
893 pr_debug("%s\n", cmd);
894 ret = system(cmd);
895 free(cmd);
896 return ret;
897 }
898
output_fd(struct perf_inject * inject)899 static int output_fd(struct perf_inject *inject)
900 {
901 return inject->in_place_update ? -1 : perf_data__fd(&inject->output);
902 }
903
__cmd_inject(struct perf_inject * inject)904 static int __cmd_inject(struct perf_inject *inject)
905 {
906 int ret = -EINVAL;
907 struct perf_session *session = inject->session;
908 int fd = output_fd(inject);
909 u64 output_data_offset;
910
911 signal(SIGINT, sig_handler);
912
913 if (inject->build_ids || inject->sched_stat ||
914 inject->itrace_synth_opts.set || inject->build_id_all) {
915 inject->tool.mmap = perf_event__repipe_mmap;
916 inject->tool.mmap2 = perf_event__repipe_mmap2;
917 inject->tool.fork = perf_event__repipe_fork;
918 inject->tool.tracing_data = perf_event__repipe_tracing_data;
919 }
920
921 output_data_offset = perf_session__data_offset(session->evlist);
922
923 if (inject->build_id_all) {
924 inject->tool.mmap = perf_event__repipe_buildid_mmap;
925 inject->tool.mmap2 = perf_event__repipe_buildid_mmap2;
926 } else if (inject->build_ids) {
927 inject->tool.sample = perf_event__inject_buildid;
928 } else if (inject->sched_stat) {
929 struct evsel *evsel;
930
931 evlist__for_each_entry(session->evlist, evsel) {
932 const char *name = evsel__name(evsel);
933
934 if (!strcmp(name, "sched:sched_switch")) {
935 if (evsel__check_stype(evsel, PERF_SAMPLE_TID, "TID"))
936 return -EINVAL;
937
938 evsel->handler = perf_inject__sched_switch;
939 } else if (!strcmp(name, "sched:sched_process_exit"))
940 evsel->handler = perf_inject__sched_process_exit;
941 else if (!strncmp(name, "sched:sched_stat_", 17))
942 evsel->handler = perf_inject__sched_stat;
943 }
944 } else if (inject->itrace_synth_opts.vm_time_correlation) {
945 session->itrace_synth_opts = &inject->itrace_synth_opts;
946 memset(&inject->tool, 0, sizeof(inject->tool));
947 inject->tool.id_index = perf_event__process_id_index;
948 inject->tool.auxtrace_info = perf_event__process_auxtrace_info;
949 inject->tool.auxtrace = perf_event__process_auxtrace;
950 inject->tool.auxtrace_error = perf_event__process_auxtrace_error;
951 inject->tool.ordered_events = true;
952 inject->tool.ordering_requires_timestamps = true;
953 } else if (inject->itrace_synth_opts.set) {
954 session->itrace_synth_opts = &inject->itrace_synth_opts;
955 inject->itrace_synth_opts.inject = true;
956 inject->tool.comm = perf_event__repipe_comm;
957 inject->tool.namespaces = perf_event__repipe_namespaces;
958 inject->tool.exit = perf_event__repipe_exit;
959 inject->tool.id_index = perf_event__process_id_index;
960 inject->tool.auxtrace_info = perf_event__process_auxtrace_info;
961 inject->tool.auxtrace = perf_event__process_auxtrace;
962 inject->tool.aux = perf_event__drop_aux;
963 inject->tool.itrace_start = perf_event__drop_aux;
964 inject->tool.aux_output_hw_id = perf_event__drop_aux;
965 inject->tool.ordered_events = true;
966 inject->tool.ordering_requires_timestamps = true;
967 /* Allow space in the header for new attributes */
968 output_data_offset = roundup(8192 + session->header.data_offset, 4096);
969 if (inject->strip)
970 strip_init(inject);
971 }
972
973 if (!inject->itrace_synth_opts.set)
974 auxtrace_index__free(&session->auxtrace_index);
975
976 if (!inject->is_pipe && !inject->in_place_update)
977 lseek(fd, output_data_offset, SEEK_SET);
978
979 ret = perf_session__process_events(session);
980 if (ret)
981 return ret;
982
983 if (!inject->is_pipe && !inject->in_place_update) {
984 struct inject_fc inj_fc = {
985 .fc.copy = feat_copy_cb,
986 .inject = inject,
987 };
988
989 if (inject->build_ids)
990 perf_header__set_feat(&session->header,
991 HEADER_BUILD_ID);
992 /*
993 * Keep all buildids when there is unprocessed AUX data because
994 * it is not known which ones the AUX trace hits.
995 */
996 if (perf_header__has_feat(&session->header, HEADER_BUILD_ID) &&
997 inject->have_auxtrace && !inject->itrace_synth_opts.set)
998 dsos__hit_all(session);
999 /*
1000 * The AUX areas have been removed and replaced with
1001 * synthesized hardware events, so clear the feature flag.
1002 */
1003 if (inject->itrace_synth_opts.set) {
1004 perf_header__clear_feat(&session->header,
1005 HEADER_AUXTRACE);
1006 if (inject->itrace_synth_opts.last_branch ||
1007 inject->itrace_synth_opts.add_last_branch)
1008 perf_header__set_feat(&session->header,
1009 HEADER_BRANCH_STACK);
1010 }
1011 session->header.data_offset = output_data_offset;
1012 session->header.data_size = inject->bytes_written;
1013 perf_session__inject_header(session, session->evlist, fd, &inj_fc.fc);
1014
1015 if (inject->copy_kcore_dir) {
1016 ret = copy_kcore_dir(inject);
1017 if (ret)
1018 return ret;
1019 }
1020 }
1021
1022 return ret;
1023 }
1024
cmd_inject(int argc,const char ** argv)1025 int cmd_inject(int argc, const char **argv)
1026 {
1027 struct perf_inject inject = {
1028 .tool = {
1029 .sample = perf_event__repipe_sample,
1030 .read = perf_event__repipe_sample,
1031 .mmap = perf_event__repipe,
1032 .mmap2 = perf_event__repipe,
1033 .comm = perf_event__repipe,
1034 .namespaces = perf_event__repipe,
1035 .cgroup = perf_event__repipe,
1036 .fork = perf_event__repipe,
1037 .exit = perf_event__repipe,
1038 .lost = perf_event__repipe,
1039 .lost_samples = perf_event__repipe,
1040 .aux = perf_event__repipe,
1041 .itrace_start = perf_event__repipe,
1042 .aux_output_hw_id = perf_event__repipe,
1043 .context_switch = perf_event__repipe,
1044 .throttle = perf_event__repipe,
1045 .unthrottle = perf_event__repipe,
1046 .ksymbol = perf_event__repipe,
1047 .bpf = perf_event__repipe,
1048 .text_poke = perf_event__repipe,
1049 .attr = perf_event__repipe_attr,
1050 .event_update = perf_event__repipe_event_update,
1051 .tracing_data = perf_event__repipe_op2_synth,
1052 .finished_round = perf_event__repipe_oe_synth,
1053 .build_id = perf_event__repipe_op2_synth,
1054 .id_index = perf_event__repipe_op2_synth,
1055 .auxtrace_info = perf_event__repipe_op2_synth,
1056 .auxtrace_error = perf_event__repipe_op2_synth,
1057 .time_conv = perf_event__repipe_op2_synth,
1058 .thread_map = perf_event__repipe_op2_synth,
1059 .cpu_map = perf_event__repipe_op2_synth,
1060 .stat_config = perf_event__repipe_op2_synth,
1061 .stat = perf_event__repipe_op2_synth,
1062 .stat_round = perf_event__repipe_op2_synth,
1063 .feature = perf_event__repipe_op2_synth,
1064 .compressed = perf_event__repipe_op4_synth,
1065 .auxtrace = perf_event__repipe_auxtrace,
1066 },
1067 .input_name = "-",
1068 .samples = LIST_HEAD_INIT(inject.samples),
1069 .output = {
1070 .path = "-",
1071 .mode = PERF_DATA_MODE_WRITE,
1072 .use_stdio = true,
1073 },
1074 };
1075 struct perf_data data = {
1076 .mode = PERF_DATA_MODE_READ,
1077 .use_stdio = true,
1078 };
1079 int ret;
1080 bool repipe = true;
1081
1082 struct option options[] = {
1083 OPT_BOOLEAN('b', "build-ids", &inject.build_ids,
1084 "Inject build-ids into the output stream"),
1085 OPT_BOOLEAN(0, "buildid-all", &inject.build_id_all,
1086 "Inject build-ids of all DSOs into the output stream"),
1087 OPT_STRING('i', "input", &inject.input_name, "file",
1088 "input file name"),
1089 OPT_STRING('o', "output", &inject.output.path, "file",
1090 "output file name"),
1091 OPT_BOOLEAN('s', "sched-stat", &inject.sched_stat,
1092 "Merge sched-stat and sched-switch for getting events "
1093 "where and how long tasks slept"),
1094 #ifdef HAVE_JITDUMP
1095 OPT_BOOLEAN('j', "jit", &inject.jit_mode, "merge jitdump files into perf.data file"),
1096 #endif
1097 OPT_INCR('v', "verbose", &verbose,
1098 "be more verbose (show build ids, etc)"),
1099 OPT_STRING('k', "vmlinux", &symbol_conf.vmlinux_name,
1100 "file", "vmlinux pathname"),
1101 OPT_BOOLEAN(0, "ignore-vmlinux", &symbol_conf.ignore_vmlinux,
1102 "don't load vmlinux even if found"),
1103 OPT_STRING(0, "kallsyms", &symbol_conf.kallsyms_name, "file",
1104 "kallsyms pathname"),
1105 OPT_BOOLEAN('f', "force", &data.force, "don't complain, do it"),
1106 OPT_CALLBACK_OPTARG(0, "itrace", &inject.itrace_synth_opts,
1107 NULL, "opts", "Instruction Tracing options\n"
1108 ITRACE_HELP,
1109 itrace_parse_synth_opts),
1110 OPT_BOOLEAN(0, "strip", &inject.strip,
1111 "strip non-synthesized events (use with --itrace)"),
1112 OPT_CALLBACK_OPTARG(0, "vm-time-correlation", &inject, NULL, "opts",
1113 "correlate time between VM guests and the host",
1114 parse_vm_time_correlation),
1115 OPT_END()
1116 };
1117 const char * const inject_usage[] = {
1118 "perf inject [<options>]",
1119 NULL
1120 };
1121 #ifndef HAVE_JITDUMP
1122 set_option_nobuild(options, 'j', "jit", "NO_LIBELF=1", true);
1123 #endif
1124 argc = parse_options(argc, argv, options, inject_usage, 0);
1125
1126 /*
1127 * Any (unrecognized) arguments left?
1128 */
1129 if (argc)
1130 usage_with_options(inject_usage, options);
1131
1132 if (inject.strip && !inject.itrace_synth_opts.set) {
1133 pr_err("--strip option requires --itrace option\n");
1134 return -1;
1135 }
1136
1137 if (symbol__validate_sym_arguments())
1138 return -1;
1139
1140 if (inject.in_place_update) {
1141 if (!strcmp(inject.input_name, "-")) {
1142 pr_err("Input file name required for in-place updating\n");
1143 return -1;
1144 }
1145 if (strcmp(inject.output.path, "-")) {
1146 pr_err("Output file name must not be specified for in-place updating\n");
1147 return -1;
1148 }
1149 if (!data.force && !inject.in_place_update_dry_run) {
1150 pr_err("The input file would be updated in place, "
1151 "the --force option is required.\n");
1152 return -1;
1153 }
1154 if (!inject.in_place_update_dry_run)
1155 data.in_place_update = true;
1156 } else {
1157 if (strcmp(inject.output.path, "-") && !inject.strip &&
1158 has_kcore_dir(inject.input_name)) {
1159 inject.output.is_dir = true;
1160 inject.copy_kcore_dir = true;
1161 }
1162 if (perf_data__open(&inject.output)) {
1163 perror("failed to create output file");
1164 return -1;
1165 }
1166 }
1167
1168 data.path = inject.input_name;
1169 if (!strcmp(inject.input_name, "-") || inject.output.is_pipe) {
1170 inject.is_pipe = true;
1171 /*
1172 * Do not repipe header when input is a regular file
1173 * since either it can rewrite the header at the end
1174 * or write a new pipe header.
1175 */
1176 if (strcmp(inject.input_name, "-"))
1177 repipe = false;
1178 }
1179
1180 inject.session = __perf_session__new(&data, repipe,
1181 output_fd(&inject),
1182 &inject.tool);
1183 if (IS_ERR(inject.session)) {
1184 ret = PTR_ERR(inject.session);
1185 goto out_close_output;
1186 }
1187
1188 if (zstd_init(&(inject.session->zstd_data), 0) < 0)
1189 pr_warning("Decompression initialization failed.\n");
1190
1191 /* Save original section info before feature bits change */
1192 ret = save_section_info(&inject);
1193 if (ret)
1194 goto out_delete;
1195
1196 if (!data.is_pipe && inject.output.is_pipe) {
1197 ret = perf_header__write_pipe(perf_data__fd(&inject.output));
1198 if (ret < 0) {
1199 pr_err("Couldn't write a new pipe header.\n");
1200 goto out_delete;
1201 }
1202
1203 ret = perf_event__synthesize_for_pipe(&inject.tool,
1204 inject.session,
1205 &inject.output,
1206 perf_event__repipe);
1207 if (ret < 0)
1208 goto out_delete;
1209 }
1210
1211 if (inject.build_ids && !inject.build_id_all) {
1212 /*
1213 * to make sure the mmap records are ordered correctly
1214 * and so that the correct especially due to jitted code
1215 * mmaps. We cannot generate the buildid hit list and
1216 * inject the jit mmaps at the same time for now.
1217 */
1218 inject.tool.ordered_events = true;
1219 inject.tool.ordering_requires_timestamps = true;
1220 }
1221
1222 if (inject.sched_stat) {
1223 inject.tool.ordered_events = true;
1224 }
1225
1226 #ifdef HAVE_JITDUMP
1227 if (inject.jit_mode) {
1228 inject.tool.mmap2 = perf_event__jit_repipe_mmap2;
1229 inject.tool.mmap = perf_event__jit_repipe_mmap;
1230 inject.tool.ordered_events = true;
1231 inject.tool.ordering_requires_timestamps = true;
1232 /*
1233 * JIT MMAP injection injects all MMAP events in one go, so it
1234 * does not obey finished_round semantics.
1235 */
1236 inject.tool.finished_round = perf_event__drop_oe;
1237 }
1238 #endif
1239 ret = symbol__init(&inject.session->header.env);
1240 if (ret < 0)
1241 goto out_delete;
1242
1243 ret = __cmd_inject(&inject);
1244
1245 out_delete:
1246 zstd_fini(&(inject.session->zstd_data));
1247 perf_session__delete(inject.session);
1248 out_close_output:
1249 if (!inject.in_place_update)
1250 perf_data__close(&inject.output);
1251 free(inject.itrace_synth_opts.vm_tm_corr_args);
1252 return ret;
1253 }
1254