2
0
mirror of https://github.com/edk2-porting/linux-next.git synced 2024-12-22 20:23:57 +08:00

ftrace: move ftrace_special to trace.c

Move the ftrace_special out of sched_switch to trace.c.

Signed-off-by: Steven Rostedt <srostedt@redhat.com>
Cc: pq@iki.fi
Cc: proski@gnu.org
Cc: sandmann@redhat.com
Cc: a.p.zijlstra@chello.nl
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
This commit is contained in:
Steven Rostedt 2008-05-22 00:22:18 -04:00 committed by Thomas Gleixner
parent 19384c0314
commit 4902f8849d
2 changed files with 25 additions and 26 deletions

View File

@ -941,6 +941,30 @@ tracing_sched_wakeup_trace(struct trace_array *tr,
trace_wake_up();
}
void
ftrace_special(unsigned long arg1, unsigned long arg2, unsigned long arg3)
{
struct trace_array *tr = &global_trace;
struct trace_array_cpu *data;
unsigned long flags;
long disabled;
int cpu;
if (tracing_disabled || current_trace == &no_tracer || !tr->ctrl)
return;
local_irq_save(flags);
cpu = raw_smp_processor_id();
data = tr->data[cpu];
disabled = atomic_inc_return(&data->disabled);
if (likely(disabled == 1))
__trace_special(tr, data, arg1, arg2, arg3);
atomic_dec(&data->disabled);
local_irq_restore(flags);
}
#ifdef CONFIG_FTRACE
static void
function_trace_call(unsigned long ip, unsigned long parent_ip)
@ -2941,8 +2965,6 @@ __init static int tracer_alloc_buffers(void)
int ret = -ENOMEM;
int i;
global_trace.ctrl = tracer_enabled;
/* TODO: make the number of buffers hot pluggable with CPUS */
tracing_nr_buffers = num_possible_cpus();
tracing_buffer_mask = cpu_possible_map;
@ -3012,6 +3034,7 @@ __init static int tracer_alloc_buffers(void)
current_trace = &no_tracer;
/* All seems OK, enable tracing */
global_trace.ctrl = tracer_enabled;
tracing_disabled = 0;
return 0;

View File

@ -125,30 +125,6 @@ wake_up_callback(void *probe_data, void *call_data,
wakeup_func(probe_data, __rq, task, curr);
}
void
ftrace_special(unsigned long arg1, unsigned long arg2, unsigned long arg3)
{
struct trace_array *tr = ctx_trace;
struct trace_array_cpu *data;
unsigned long flags;
long disabled;
int cpu;
if (!tracer_enabled)
return;
local_irq_save(flags);
cpu = raw_smp_processor_id();
data = tr->data[cpu];
disabled = atomic_inc_return(&data->disabled);
if (likely(disabled == 1))
__trace_special(tr, data, arg1, arg2, arg3);
atomic_dec(&data->disabled);
local_irq_restore(flags);
}
static void sched_switch_reset(struct trace_array *tr)
{
int cpu;