Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 | // SPDX-License-Identifier: GPL-2.0 #include "debug.h" #include "evlist.h" #include "evsel.h" #include "target.h" #include "thread_map.h" #include "tests.h" #include "util/mmap.h" #include <errno.h> #include <signal.h> #include <linux/string.h> #include <perf/cpumap.h> #include <perf/evlist.h> #include <perf/mmap.h> static int exited; static int nr_exit; static void sig_handler(int sig __maybe_unused) { exited = 1; } /* * perf_evlist__prepare_workload will send a SIGUSR1 if the fork fails, since * we asked by setting its exec_error to this handler. */ static void workload_exec_failed_signal(int signo __maybe_unused, siginfo_t *info __maybe_unused, void *ucontext __maybe_unused) { exited = 1; nr_exit = -1; } /* * This test will start a workload that does nothing then it checks * if the number of exit event reported by the kernel is 1 or not * in order to check the kernel returns correct number of event. */ int test__task_exit(struct test *test __maybe_unused, int subtest __maybe_unused) { int err = -1; union perf_event *event; struct evsel *evsel; struct evlist *evlist; struct target target = { .uid = UINT_MAX, .uses_mmap = true, }; const char *argv[] = { "true", NULL }; char sbuf[STRERR_BUFSIZE]; struct perf_cpu_map *cpus; struct perf_thread_map *threads; struct mmap *md; int retry_count = 0; signal(SIGCHLD, sig_handler); evlist = perf_evlist__new_default(); if (evlist == NULL) { pr_debug("perf_evlist__new_default\n"); return -1; } /* * Create maps of threads and cpus to monitor. In this case * we start with all threads and cpus (-1, -1) but then in * perf_evlist__prepare_workload we'll fill in the only thread * we're monitoring, the one forked there. */ cpus = perf_cpu_map__dummy_new(); threads = thread_map__new_by_tid(-1); if (!cpus || !threads) { err = -ENOMEM; pr_debug("Not enough memory to create thread/cpu maps\n"); goto out_free_maps; } perf_evlist__set_maps(&evlist->core, cpus, threads); cpus = NULL; threads = NULL; err = perf_evlist__prepare_workload(evlist, &target, argv, false, workload_exec_failed_signal); if (err < 0) { pr_debug("Couldn't run the workload!\n"); goto out_delete_evlist; } evsel = evlist__first(evlist); evsel->core.attr.task = 1; #ifdef __s390x__ evsel->core.attr.sample_freq = 1000000; #else evsel->core.attr.sample_freq = 1; #endif evsel->core.attr.inherit = 0; evsel->core.attr.watermark = 0; evsel->core.attr.wakeup_events = 1; evsel->core.attr.exclude_kernel = 1; err = evlist__open(evlist); if (err < 0) { pr_debug("Couldn't open the evlist: %s\n", str_error_r(-err, sbuf, sizeof(sbuf))); goto out_delete_evlist; } if (evlist__mmap(evlist, 128) < 0) { pr_debug("failed to mmap events: %d (%s)\n", errno, str_error_r(errno, sbuf, sizeof(sbuf))); err = -1; goto out_delete_evlist; } perf_evlist__start_workload(evlist); retry: md = &evlist->mmap[0]; if (perf_mmap__read_init(&md->core) < 0) goto out_init; while ((event = perf_mmap__read_event(&md->core)) != NULL) { if (event->header.type == PERF_RECORD_EXIT) nr_exit++; perf_mmap__consume(&md->core); } perf_mmap__read_done(&md->core); out_init: if (!exited || !nr_exit) { evlist__poll(evlist, -1); if (retry_count++ > 1000) { pr_debug("Failed after retrying 1000 times\n"); err = -1; goto out_free_maps; } goto retry; } if (nr_exit != 1) { pr_debug("received %d EXIT records\n", nr_exit); err = -1; } out_free_maps: perf_cpu_map__put(cpus); perf_thread_map__put(threads); out_delete_evlist: evlist__delete(evlist); return err; } |