Lines Matching refs:event
8 * event may be specified using --event.
35 #include "util/event.h"
58 #include "trace-event.h"
88 #include <traceevent/event-parse.h>
1607 union perf_event *event, struct perf_sample *sample)
1611 switch (event->header.type) {
1614 "LOST %" PRIu64 " events!\n", event->lost.lost);
1615 ret = machine__process_lost_event(machine, event, sample);
1618 ret = machine__process_event(machine, event, sample);
1626 union perf_event *event,
1631 return trace__process_event(trace, machine, event, sample);
1985 * 8-byte unaligned accesses. args points to raw_data within the event
2124 union perf_event *event,
2329 union perf_event *event __maybe_unused,
2368 * raw_syscalls:sys_enter event, when we expected just 4, we end up mistakenly
2479 union perf_event *event __maybe_unused,
2596 * We only consider an 'event' for the sake of --max-events a non-filtered
2615 union perf_event *event __maybe_unused,
2676 union perf_event *event __maybe_unused,
2810 union perf_event *event __maybe_unused,
2817 * this event's max_events having been hit and this is an entry coming
2860 * fall thru and print it as a bpf_output event.
2919 union perf_event *event __maybe_unused,
3003 * some other event in the future without PERF_SAMPLE_TIME for good
3006 * appears in our event stream (vfs_getname comes to mind).
3014 union perf_event *event,
3033 handler(trace, evsel, event, sample);
3060 /* +3 is for the event string below and the pid filter */
3076 /* event string may be different for older kernels - e.g., RHEL6 */
3171 static void trace__handle_event(struct trace *trace, union perf_event *event, struct perf_sample *sample)
3173 const u32 type = event->header.type;
3177 trace__process_event(trace, trace->host, event, sample);
3199 handler(trace, evsel, event, sample);
3627 static int __trace__deliver_event(struct trace *trace, union perf_event *event)
3631 int err = evlist__parse_sample(evlist, event, &sample);
3636 trace__handle_event(trace, event, &sample);
3658 static int trace__deliver_event(struct trace *trace, union perf_event *event)
3663 return __trace__deliver_event(trace, event);
3665 err = evlist__parse_sample_timestamp(trace->evlist, event, &trace->oe.last);
3669 err = ordered_events__queue(&trace->oe.data, event, trace->oe.last, 0, NULL);
3677 struct ordered_event *event)
3681 return __trace__deliver_event(trace, event->event);
3867 * will only set the sched:sched_switch event to the 'A' cgroup, all the
3916 * CPU the bpf-output event's file descriptor.
3942 pr_debug("event qualifier tracepoint filter: %s\n",
3989 * trace__resolve_callchain(), allowing per-event max-stack settings
4001 union perf_event *event;
4008 while ((event = perf_mmap__read_event(&md->core)) != NULL) {
4011 err = trace__deliver_event(trace, event);
4097 "Failed to set filter \"%s\" on event %s with %d (%s)\n",
4170 pr_err("Error during initialize raw_syscalls:sys_enter event\n");
4181 pr_err("Error during initialize raw_syscalls:sys_exit event\n");
4489 * XXX: Hackish, just splitting the combined -e+--event (syscalls
4494 * list with the terms it didn't match to an event...
4559 fputs("Not enough memory to parse event qualifier", trace->output);
4669 pr_debug("ERROR: failed to create the \"__augmented_syscalls__\" bpf-output event\n");
4708 OPT_CALLBACK('e', "event", &trace, "event",
4709 "event/syscall selector. use 'perf list' to list available events",
4712 "event filter", parse_filter),
4777 OPT_CALLBACK('G', "cgroup", &trace, "name", "monitor event in cgroup name only",
4812 * Parsing .perfconfig may entail creating a BPF event, that may need
4829 * already figured out if -e syscall_name, if not but if --event
4898 pr_err("ERROR: Setup BPF output event failed: %s\n", bf);
4973 * Augmented is __augmented_syscalls__ BPF_OUTPUT event
4979 * Now we do the same for the *syscalls:sys_enter event so that