 430ad5a600
			
		
	
	
	430ad5a600
	
	
	
		
			
			Introduce ftrace_perf_buf_prepare() and ftrace_perf_buf_submit() to gather the common code that operates on raw events sampling buffer. This cleans up redundant code between regular trace events, syscall events and kprobe events. Changelog v1->v2: - Rename function name as per Masami and Frederic's suggestion - Add __kprobes for ftrace_perf_buf_prepare() and make ftrace_perf_buf_submit() inline as per Masami's suggestion - Export ftrace_perf_buf_prepare since modules will use it Signed-off-by: Xiao Guangrong <xiaoguangrong@cn.fujitsu.com> Acked-by: Masami Hiramatsu <mhiramat@redhat.com> Cc: Ingo Molnar <mingo@elte.hu> Cc: Steven Rostedt <rostedt@goodmis.org> Cc: Paul Mackerras <paulus@samba.org> Cc: Jason Baron <jbaron@redhat.com> Cc: Peter Zijlstra <peterz@infradead.org> LKML-Reference: <4B60E92D.9000808@cn.fujitsu.com> Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
		
			
				
	
	
		
			164 lines
		
	
	
	
		
			3.3 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			164 lines
		
	
	
	
		
			3.3 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * trace event based perf counter profiling
 | |
|  *
 | |
|  * Copyright (C) 2009 Red Hat Inc, Peter Zijlstra <pzijlstr@redhat.com>
 | |
|  *
 | |
|  */
 | |
| 
 | |
| #include <linux/module.h>
 | |
| #include <linux/kprobes.h>
 | |
| #include "trace.h"
 | |
| 
 | |
| 
 | |
| static char *perf_trace_buf;
 | |
| static char *perf_trace_buf_nmi;
 | |
| 
 | |
| typedef typeof(char [FTRACE_MAX_PROFILE_SIZE]) perf_trace_t ;
 | |
| 
 | |
| /* Count the events in use (per event id, not per instance) */
 | |
| static int	total_profile_count;
 | |
| 
 | |
| static int ftrace_profile_enable_event(struct ftrace_event_call *event)
 | |
| {
 | |
| 	char *buf;
 | |
| 	int ret = -ENOMEM;
 | |
| 
 | |
| 	if (event->profile_count++ > 0)
 | |
| 		return 0;
 | |
| 
 | |
| 	if (!total_profile_count) {
 | |
| 		buf = (char *)alloc_percpu(perf_trace_t);
 | |
| 		if (!buf)
 | |
| 			goto fail_buf;
 | |
| 
 | |
| 		rcu_assign_pointer(perf_trace_buf, buf);
 | |
| 
 | |
| 		buf = (char *)alloc_percpu(perf_trace_t);
 | |
| 		if (!buf)
 | |
| 			goto fail_buf_nmi;
 | |
| 
 | |
| 		rcu_assign_pointer(perf_trace_buf_nmi, buf);
 | |
| 	}
 | |
| 
 | |
| 	ret = event->profile_enable(event);
 | |
| 	if (!ret) {
 | |
| 		total_profile_count++;
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| fail_buf_nmi:
 | |
| 	if (!total_profile_count) {
 | |
| 		free_percpu(perf_trace_buf_nmi);
 | |
| 		free_percpu(perf_trace_buf);
 | |
| 		perf_trace_buf_nmi = NULL;
 | |
| 		perf_trace_buf = NULL;
 | |
| 	}
 | |
| fail_buf:
 | |
| 	event->profile_count--;
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int ftrace_profile_enable(int event_id)
 | |
| {
 | |
| 	struct ftrace_event_call *event;
 | |
| 	int ret = -EINVAL;
 | |
| 
 | |
| 	mutex_lock(&event_mutex);
 | |
| 	list_for_each_entry(event, &ftrace_events, list) {
 | |
| 		if (event->id == event_id && event->profile_enable &&
 | |
| 		    try_module_get(event->mod)) {
 | |
| 			ret = ftrace_profile_enable_event(event);
 | |
| 			break;
 | |
| 		}
 | |
| 	}
 | |
| 	mutex_unlock(&event_mutex);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static void ftrace_profile_disable_event(struct ftrace_event_call *event)
 | |
| {
 | |
| 	char *buf, *nmi_buf;
 | |
| 
 | |
| 	if (--event->profile_count > 0)
 | |
| 		return;
 | |
| 
 | |
| 	event->profile_disable(event);
 | |
| 
 | |
| 	if (!--total_profile_count) {
 | |
| 		buf = perf_trace_buf;
 | |
| 		rcu_assign_pointer(perf_trace_buf, NULL);
 | |
| 
 | |
| 		nmi_buf = perf_trace_buf_nmi;
 | |
| 		rcu_assign_pointer(perf_trace_buf_nmi, NULL);
 | |
| 
 | |
| 		/*
 | |
| 		 * Ensure every events in profiling have finished before
 | |
| 		 * releasing the buffers
 | |
| 		 */
 | |
| 		synchronize_sched();
 | |
| 
 | |
| 		free_percpu(buf);
 | |
| 		free_percpu(nmi_buf);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| void ftrace_profile_disable(int event_id)
 | |
| {
 | |
| 	struct ftrace_event_call *event;
 | |
| 
 | |
| 	mutex_lock(&event_mutex);
 | |
| 	list_for_each_entry(event, &ftrace_events, list) {
 | |
| 		if (event->id == event_id) {
 | |
| 			ftrace_profile_disable_event(event);
 | |
| 			module_put(event->mod);
 | |
| 			break;
 | |
| 		}
 | |
| 	}
 | |
| 	mutex_unlock(&event_mutex);
 | |
| }
 | |
| 
 | |
| __kprobes void *ftrace_perf_buf_prepare(int size, unsigned short type,
 | |
| 					int *rctxp, unsigned long *irq_flags)
 | |
| {
 | |
| 	struct trace_entry *entry;
 | |
| 	char *trace_buf, *raw_data;
 | |
| 	int pc, cpu;
 | |
| 
 | |
| 	pc = preempt_count();
 | |
| 
 | |
| 	/* Protect the per cpu buffer, begin the rcu read side */
 | |
| 	local_irq_save(*irq_flags);
 | |
| 
 | |
| 	*rctxp = perf_swevent_get_recursion_context();
 | |
| 	if (*rctxp < 0)
 | |
| 		goto err_recursion;
 | |
| 
 | |
| 	cpu = smp_processor_id();
 | |
| 
 | |
| 	if (in_nmi())
 | |
| 		trace_buf = rcu_dereference(perf_trace_buf_nmi);
 | |
| 	else
 | |
| 		trace_buf = rcu_dereference(perf_trace_buf);
 | |
| 
 | |
| 	if (!trace_buf)
 | |
| 		goto err;
 | |
| 
 | |
| 	raw_data = per_cpu_ptr(trace_buf, cpu);
 | |
| 
 | |
| 	/* zero the dead bytes from align to not leak stack to user */
 | |
| 	*(u64 *)(&raw_data[size - sizeof(u64)]) = 0ULL;
 | |
| 
 | |
| 	entry = (struct trace_entry *)raw_data;
 | |
| 	tracing_generic_entry_update(entry, *irq_flags, pc);
 | |
| 	entry->type = type;
 | |
| 
 | |
| 	return raw_data;
 | |
| err:
 | |
| 	perf_swevent_put_recursion_context(*rctxp);
 | |
| err_recursion:
 | |
| 	local_irq_restore(*irq_flags);
 | |
| 	return NULL;
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(ftrace_perf_buf_prepare);
 |