 f66a889dbc
			
		
	
	
	f66a889dbc
	
	
	
		
			
			Since we have access two evlist members in all these poll calls, provide a helper. This will also help to make the patch introducing the pollfd class more clear, as the evlist specific uses will be hiden away perf_evlist__poll(). Acked-by: Jiri Olsa <jolsa@kernel.org> Cc: Adrian Hunter <adrian.hunter@intel.com> Cc: Corey Ashford <cjashfor@linux.vnet.ibm.com> Cc: David Ahern <dsahern@gmail.com> Cc: Frederic Weisbecker <fweisbec@gmail.com> Cc: Ingo Molnar <mingo@kernel.org> Cc: Jean Pihet <jean.pihet@linaro.org> Cc: Jiri Olsa <jolsa@kernel.org> Cc: Namhyung Kim <namhyung@kernel.org> Cc: Paul Mackerras <paulus@samba.org> Cc: Peter Zijlstra <a.p.zijlstra@chello.nl> Link: http://lkml.kernel.org/n/tip-jr9d4aop4lvy9453qahbcgp0@git.kernel.org Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
		
			
				
	
	
		
			120 lines
		
	
	
	
		
			2.8 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			120 lines
		
	
	
	
		
			2.8 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| #include "evlist.h"
 | |
| #include "evsel.h"
 | |
| #include "thread_map.h"
 | |
| #include "cpumap.h"
 | |
| #include "tests.h"
 | |
| 
 | |
| #include <signal.h>
 | |
| 
 | |
| static int exited;
 | |
| static int nr_exit;
 | |
| 
 | |
| static void sig_handler(int sig __maybe_unused)
 | |
| {
 | |
| 	exited = 1;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * perf_evlist__prepare_workload will send a SIGUSR1 if the fork fails, since
 | |
|  * we asked by setting its exec_error to this handler.
 | |
|  */
 | |
| static void workload_exec_failed_signal(int signo __maybe_unused,
 | |
| 					siginfo_t *info __maybe_unused,
 | |
| 					void *ucontext __maybe_unused)
 | |
| {
 | |
| 	exited	= 1;
 | |
| 	nr_exit = -1;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * This test will start a workload that does nothing then it checks
 | |
|  * if the number of exit event reported by the kernel is 1 or not
 | |
|  * in order to check the kernel returns correct number of event.
 | |
|  */
 | |
| int test__task_exit(void)
 | |
| {
 | |
| 	int err = -1;
 | |
| 	union perf_event *event;
 | |
| 	struct perf_evsel *evsel;
 | |
| 	struct perf_evlist *evlist;
 | |
| 	struct target target = {
 | |
| 		.uid		= UINT_MAX,
 | |
| 		.uses_mmap	= true,
 | |
| 	};
 | |
| 	const char *argv[] = { "true", NULL };
 | |
| 	char sbuf[STRERR_BUFSIZE];
 | |
| 
 | |
| 	signal(SIGCHLD, sig_handler);
 | |
| 
 | |
| 	evlist = perf_evlist__new_default();
 | |
| 	if (evlist == NULL) {
 | |
| 		pr_debug("perf_evlist__new_default\n");
 | |
| 		return -1;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * Create maps of threads and cpus to monitor. In this case
 | |
| 	 * we start with all threads and cpus (-1, -1) but then in
 | |
| 	 * perf_evlist__prepare_workload we'll fill in the only thread
 | |
| 	 * we're monitoring, the one forked there.
 | |
| 	 */
 | |
| 	evlist->cpus = cpu_map__dummy_new();
 | |
| 	evlist->threads = thread_map__new_by_tid(-1);
 | |
| 	if (!evlist->cpus || !evlist->threads) {
 | |
| 		err = -ENOMEM;
 | |
| 		pr_debug("Not enough memory to create thread/cpu maps\n");
 | |
| 		goto out_delete_evlist;
 | |
| 	}
 | |
| 
 | |
| 	err = perf_evlist__prepare_workload(evlist, &target, argv, false,
 | |
| 					    workload_exec_failed_signal);
 | |
| 	if (err < 0) {
 | |
| 		pr_debug("Couldn't run the workload!\n");
 | |
| 		goto out_delete_evlist;
 | |
| 	}
 | |
| 
 | |
| 	evsel = perf_evlist__first(evlist);
 | |
| 	evsel->attr.task = 1;
 | |
| 	evsel->attr.sample_freq = 0;
 | |
| 	evsel->attr.inherit = 0;
 | |
| 	evsel->attr.watermark = 0;
 | |
| 	evsel->attr.wakeup_events = 1;
 | |
| 	evsel->attr.exclude_kernel = 1;
 | |
| 
 | |
| 	err = perf_evlist__open(evlist);
 | |
| 	if (err < 0) {
 | |
| 		pr_debug("Couldn't open the evlist: %s\n",
 | |
| 			 strerror_r(-err, sbuf, sizeof(sbuf)));
 | |
| 		goto out_delete_evlist;
 | |
| 	}
 | |
| 
 | |
| 	if (perf_evlist__mmap(evlist, 128, true) < 0) {
 | |
| 		pr_debug("failed to mmap events: %d (%s)\n", errno,
 | |
| 			 strerror_r(errno, sbuf, sizeof(sbuf)));
 | |
| 		goto out_delete_evlist;
 | |
| 	}
 | |
| 
 | |
| 	perf_evlist__start_workload(evlist);
 | |
| 
 | |
| retry:
 | |
| 	while ((event = perf_evlist__mmap_read(evlist, 0)) != NULL) {
 | |
| 		if (event->header.type == PERF_RECORD_EXIT)
 | |
| 			nr_exit++;
 | |
| 
 | |
| 		perf_evlist__mmap_consume(evlist, 0);
 | |
| 	}
 | |
| 
 | |
| 	if (!exited || !nr_exit) {
 | |
| 		perf_evlist__poll(evlist, -1);
 | |
| 		goto retry;
 | |
| 	}
 | |
| 
 | |
| 	if (nr_exit != 1) {
 | |
| 		pr_debug("received %d EXIT records\n", nr_exit);
 | |
| 		err = -1;
 | |
| 	}
 | |
| 
 | |
| out_delete_evlist:
 | |
| 	perf_evlist__delete(evlist);
 | |
| 	return err;
 | |
| }
 |